On 2005-03-21 15:32, [EMAIL PROTECTED] wrote:
> *** This is not problem of wget, but your filesystem. Try to do 
> 
> touch 
> search.ebay.de/wget_W0QQcatrefZ3DC6QQcoactionZ3DcompareQQcoentrypageZ3DsearchQQcopagenumZ3D1QQdfeZ3D20050024QQdfsZ3D20050024QQdfteZ3DQ2d1QQdftsZ3DQ2d1QQfltZ3D9QQfromZ3DR9QQfsooZ3D2QQfsopZ3D2QQsaetmZ3D1111396614QQsojsZ3D1QQsspagenameZ3DADMEQ3aBQ3aSSQ3aDEQ3a21QQversionZ3D2.html

I'm very sure that my file system has some limits somewhere - but I
suppose a web server may create virtual URLs which will be too long or
will include illegal characters for almost any file system around.


The file name here might get repaired by some regex, e.g.
wget_?&catref=C6&coaction=compare&coentrypage=search&copagenum=1&dfte=Q2d1&dfts=Q2d1&flt=9&from=R9&fsoo=2&fsop=2&saetm=1111396614&sojs=1&sspagename=ADMEQ3aBQ3aSSQ3aDEQ3a21&version=2.html

However, I'd be comfortable enough with some fixed length or char
limitation, such as a 'trim' extension:

  -tc, --trimcharacter char cut filename after character, such as "_"
  -tl, --trimlength    num  cut filename after num characters
  -ts, --trimsuffix    num  digits used for incremented cut filenames
  -tt, --trimtable     file log trimmed file name and original to file


For the moment I'd be happy enough with saving to a md5.html checksum as
filename instead of a filename too long for my fs.
The output log could tell me about the shrinked and the original
filename.

> > search.ebay.de and then download the links to cgi.ebay.de in one:
> > 
> >   wget -kxrE -l1 -D cgi.ebay.de -H $URL
> 
> *** maybe to create SHA1 sum of the request and store the result in this file
> (but you will not know what was the original request, if you don't create some
> DB of requests). Or do just simple counting
> 
> URL="....."
> sha1sum="$( echo -n "$URL" | sha1sum )"
> echo "$sha1sum $URL" >> SHA1-URL.db
> wget -O sha1sum.html [other options] "$URL"
> 
> or
> 
> URL="...."
> i=0
> echo "$i $URL" >> URL.db
> wget -O search-$i.html "$URL"
> 
> Could be this your solution?

Nice idea - I'll give it a try. However, it does not answer the -D problem
itself. I'm afraid this does require some further awk/sed processing of
the result?

Thanks,
Martin

Reply via email to