This is an automated archive made by the Lemmit Bot.

The original was posted on /r/datahoarder by /u/Mhanz97 on 2024-03-18 17:00:31.


Hi everyone, like title said, why the hell its so hard to download a complete website for offline view?

I was trying to download a fandom wiki (the entire wiki about a videogame), i tried lot of tools and i always got some problems…

i tried:

  • Wget: got problems when downloading images…lot of them was not downloaded…
  • httrack: takes forever/ super slow, and not downloading all the images too + even with depth levels restriction keep download useless outside-domain websites
  • offline explorer: maybe the worst since everything was messed up after download + no all the images
  • Cyotek web copy: same as offline explorer
  • Wikiteam software (dumpgenerator.py): ultra messy, super hard to install and didnt worked on my windows

Basically the only thing that at least download all the text + images its the chrome ctrl+s (save page), but i need to manually full load and save page by page…and when i read in offline mode its a bit messed up, but at least i have all the thing saved…