On Fri, 17 Nov 2023, Michael Barnes wrote:

I have used this command string successfully in the past to download
complete websites.

$ wget      --recursive      --no-clobber      --page-requisites
--html-extension      --convert-links      --restrict-file-names=windows
  --domains website.com      --no-parent website.com

Michael,

This returned only the site's index page:
$ wget -r --no-clobber --page-requisites  --html-extension --convert-links 
--restrict-file-names=windows --domains http://ph-public-data.com --no-parent 
ph-public-data.com
Both --no-clobber and --convert-links were specified, only --convert-links will 
be used.
--2023-11-17 13:16:41--  http://ph-public-data.com/
Resolving ph-public-data.com... 138.68.58.192
Connecting to ph-public-data.com|138.68.58.192|:80... connected.
HTTP request sent, awaiting response... 200 OK
Length: 19486 (19K) [text/html]
Saving to: ‘ph-public-data.com/index.html’

ph-public-data.com/index 100%[==================================>]  19.03K  
--.-KB/s    in 0.02s

2023-11-17 13:16:41 (904 KB/s) - ‘ph-public-data.com/index.html’ saved 
[19486/19486]

FINISHED --2023-11-17 13:16:41--
Total wall clock time: 0.2s
Downloaded: 1 files, 19K in 0.02s (904 KB/s)
Converting links in ph-public-data.com/index.html... 0-52
Converted links in 1 files in 0.001 seconds.

Regards,

Rich

Reply via email to