Results 1 to 8 of 8

Thread: copy website - kde

  1. #1

    Question copy website - kde

    I installed httrack. This opens a new tab then stalls and won't continue,

    Also, I found you can do this with wget. https://www.linuxjournal.com/content...-web-site-wget

    Need help here.

  2. #2
    Join Date
    Jun 2008
    Location
    West Yorkshire, UK
    Posts
    3,493

    Default Re: copy website - kde

    View>View Document Source (Ctrl-U) allows you to save the source code of a page.

  3. #3

    Default Re: copy website - kde

    Quote Originally Posted by john_hudson View Post
    View>View Document Source (Ctrl-U) allows you to save the source code of a page.
    FYI, This site hasn't visited in 6 months.

    I need download the whole site at once.

  4. #4
    Join Date
    Jun 2008
    Location
    San Diego, Ca, USA
    Posts
    12,004
    Blog Entries
    2

    Default Re: copy website - kde

    Are you able to post details,
    eg

    When running httrack, have you tried modifying options? eg more or fewer simultaneous threads, resource allocation, etc.

    If you've tried wget and run into a problem, describe in detail.

    Personally,
    I've used both for different circumstances.
    If I can't download the entire website in one session or need fine grained control of options, then I'll use httrack although it can take more time to set up and run than wget.

    I use wget if I want something simple and fast, particularly if I'm using the same configuration against multiple URLs.

    TSU
    Beginner Wiki Quickstart - https://en.opensuse.org/User:Tsu2/Quickstart_Wiki
    Solved a problem recently? Create a wiki page for future personal reference!
    Learn something new?
    Attended a computing event?
    Post and Share!

  5. #5
    Join Date
    Jun 2008
    Location
    Netherlands
    Posts
    25,924

    Default Re: copy website - kde

    Be aware of the pitfalls here.

    Downloading a "whole web site" is possible if the pages are linked to each other using URLs because then the download tool can try to follow those links (and avoid pages that are already found because many links on many pages can point to a page).

    But nowadays many web sites are very dynamic and when scripts (e.g.Java) decide what URLs are to be asked from the server depending on all sorts of parameters the tool might not be able to trigger that.
    Henk van Velden

  6. #6

    Default copy website - kde

    Quote Originally Posted by tsu2 View Post
    Are you able to post details,
    eg

    When running httrack, have you tried modifying options? eg more or fewer simultaneous threads, resource allocation, etc.

    If you've tried wget and run into a problem, describe in detail.

    Personally,
    I've used both for different circumstances.
    If I can't download the entire website in one session or need fine grained control of options, then I'll use httrack although it can take more time to set up and run than wget.

    I use wget if I want something simple and fast, particularly if I'm using the same configuration against multiple URLs.

    TSU
    Installed httrack through YAST - software management.
    Click on deskop menu symbol. Search httrack. Click on httrack website copier.
    New firefox tab opens with address linux-xp0f:8080 Stalls --> connection timed out


    I found a website showing a long sequence to download a whole website.
    https://www.linuxjournal.com/content...-web-site-wget

    This looks the code to download a website correctly. However, i'll ask permission first to download the whole website from the owner.

    Is the above link correct? Does it need any more?

  7. #7

    Default Re: copy website - kde

    Quote Originally Posted by hcvv View Post
    Be aware of the pitfalls here.

    Downloading a "whole web site" is possible if the pages are linked to each other using URLs because then the download tool can try to follow those links (and avoid pages that are already found because many links on many pages can point to a page).

    But nowadays many web sites are very dynamic and when scripts (e.g.Java) decide what URLs are to be asked from the server depending on all sorts of parameters the tool might not be able to trigger that.
    I'll keep thar in mind.

  8. #8

    Default Re: copy website - kde

    Whatever linux bug was causing it, seems have been been fixed.

    Thanks to all for the help.

Tags for this Thread

Posting Permissions

  • You may not post new threads
  • You may not post replies
  • You may not post attachments
  • You may not edit your posts
  •