I've used HTTrack in the past
It worked well for the websites that I tried it on and it's pretty easy to run. You could give it a go?
1. Posts must be related to the discussion of digital piracy
2. Don't request invites, trade, sell, or self-promote
3. Don't request or link to specific pirated titles, including DMs
4. Don't submit low-quality posts, be entitled, or harass others
📜 c/Piracy Wiki (Community Edition):
💰 Please help cover server costs.
Ko-fi | Liberapay |
I've used HTTrack in the past
It worked well for the websites that I tried it on and it's pretty easy to run. You could give it a go?
I've got a container set up of this. Drops the output on the NAS and can be accessed from any box on the local net. The only issue is it has a tendency to need the container recycled every so often, like it just gets board sitting there and quits. 🤔
SingleFile extension can save a web page into a single html file where all media are neatly inlined inside the file. You'll have to do this manually on each page though, so it's not ideal for saving the whole website.
If you're comfortable running commands in terminal, you can use SingleFile CLI to crawl the whole website, e.g.: single-file https://www.wikipedia.org --crawl-links=true --crawl-inner-links-only=true --crawl-max-depth=1 --crawl-rewrite-rule="^(.*)\\?.*$ $1"
I use grab-site (unmaintained) for full site archival and wget -p -k
for simple non-javascript single pages
I've heard good things about HTTrack, SingleFile and Archivebox but don't have any experience with them.
Archivebox looks the most modern and intuitive but is hosted on docker