On Thu, 10 May 2007 16:04:41 -0500 (CDT)
Steven M. Schweda wrote:
> From: R Kimber
>
> > Yes there's a web page. I usually know what I want.
>
> There's a difference between knowing what you want and being able
> to describe what you want so that it makes sense to someone who does
> not know what you want.
Well I was wondering if wget had a way of allowing me to specify it.
> > But won't a recursive get get more than just those files? Indeed,
> > won't it get everything at that level? The accept/reject options
> > seem to assume you know what's there and can list them to exclude
> > them. I only know what I want. [...]
>
> Are you trying to say that you have a list of URLs, and would like
> to use one wget command for all instead of one wget command per URL?
> Around here:
>
> ALP $ wget -h
> GNU Wget 1.10.2c, a non-interactive network retriever.
> Usage: alp$dka0:[utility]wget.exe;13 [OPTION]... [URL]...
> [...]
>
> That "[URL]..." was supposed to suggest that you can supply more than
> one URL on the command line. Subject to possible command-line length
> limitations, this should allow any number of URLs to be specified at
> once.
>
> There's also "-1" ("--input-file=FILE"). No bets, but it looks as
> if you can specify "-" for FILE, and it'll read the URLs from stdin,
> so you could pipe them in from anything.
Thanks, but my point is I don't know the full URL, just the pattern.
What I'm trying to download is what I might express as:
http://www.stirling.gov.uk/*.pdf
but I guess that's not possible. I just wondered if it was possible
for wget to filter out everything except *.pdf - i.e. wget would look
at a site, or a directory on a site, and just accept those files that
match a pattern.
- Richard
--
Richard Kimber
http://www.psr.keele.ac.uk/