Rip and save static copy of website using wget


Download a complete static version of a website:

wget -m -p -E -k example.com

For a comprehensive download, you might use wget like this. This will download the entire website from www.example.com/tutorials/html/.

wget --recursive --no-clobber --page-requisites --html-extension --convert-links --restrict-file-names=windows --domains example.com --no-parent www.example.com/tutorials/html/

Explanation of the options:

  • --recursive: Downloads the whole website.
  • --domains website.org: Limits downloads to website.org, ignoring external links.
  • --no-parent: Avoids retrieving anything from parent directories.
  • --page-requisites: Downloads all necessary elements for the page (like images and CSS).
  • --html-extension: Saves files with the .html extension.
  • --convert-links: Adjusts links to work locally, for offline viewing.
  • --restrict-file-names=windows: Ensures filenames are compatible with Windows.
  • --no-clobber: Prevents overwriting existing files, useful for resuming downloads.
comments powered byDisqus