Advertisements
The following command will try to download a full website with all pages it can find through public links.
wget --wait=2 --limit-rate=200K --recursive --no-clobber --page-requisites --convert-links --domains example.com http://example.com/
Parameters:
--wait
Wait the specified number of seconds between the retrievals. We use this option to lighten the server load by making the requests less frequent.--limit-rate
Limit the download speed to amount bytes per second. We use this option to lighten the server load and to reduce the bandwidth we consume on our own network.--recursive
Turn on recursive retrieving.--no-clobber
If a file is downloaded more than once in the same directory, we prevent multiple version saving.--page-requisites
This option causes Wget to download all the files that are necessary to properly display a given HTML page.--convert-links
After the download is complete, convert the links in the document to make them suitable for local viewing.--domains
Set domains to be followed. It accepts a domain-list as a comma-separated list of domains.