How do you archive an entire website for offline viewing?

后端 未结 9 1213
渐次进展
渐次进展 2021-01-29 21:02

We actually have burned static/archived copies of our asp.net websites for customers many times. We have used WebZip until now but we have had endless problems with crashes, do

相关标签:
9条回答
  • 2021-01-29 21:26

    I just use: wget -m <url>.

    0 讨论(0)
  • 2021-01-29 21:35

    In Windows, you can look at HTTrack. It's very configurable allowing you to set the speed of the downloads. But you can just point it at a website and run it too with no configuration at all.

    In my experience it's been a really good tool and works well. Some of the things I like about HTTrack are:

    • Open Source license
    • Resumes stopped downloads
    • Can update an existing archive
    • You can configure it to be non-aggressive when it downloads so it doesn't waste your bandwidth and the bandwidth of the site.
    0 讨论(0)
  • 2021-01-29 21:36

    If your customers are archiving for compliance issues, you want to ensure that the content can be authenticated. The options listed are fine for simple viewing, but they aren't legally admissible. In that case, you're looking for timestamps and digital signatures. Much more complicated if you're doing it yourself. I'd suggest a service such as PageFreezer.

    0 讨论(0)
  • 2021-01-29 21:38

    You could use wget:

    wget -m -k -K -E http://url/of/web/site
    
    0 讨论(0)
  • 2021-01-29 21:38

    I've been using HTTrack for several years now. It handles all of the inter-page linking, etc. just fine. My only complaint is that I haven't found a good way to keep it limited to a sub-site very well. For instance, if there is a site www.foo.com/steve that I want to archive, it will likely follow links to www.foo.com/rowe and archive that too. Otherwise it's great. Highly configurable and reliable.

    0 讨论(0)
  • 2021-01-29 21:39

    I use Blue Crab on OSX and WebCopier on Windows.

    0 讨论(0)
提交回复
热议问题