[EMAIL PROTECTED] writes:

> Complaint: offline both
> $ wwwoffle SomeDontGetURL
> $ wwwoffle SomeDontCacheURL
> both say
> Requesting ThatURL
> and return $?=0 to the shell,
> even though WWWOFFLE intends to do no such fetching.

You are correct, the WWWOFFLE server doesn't intend to do any
fetching.  But the wwwoffle program is correct because it says that it
is requesting the URL - which it does.

For the DontGet URLs it is difficult because if you have a replacement
URL configured then you will get that.  There is no way to know that
it is the replacement and not the real thing.

For the DontCache URLs it would be possible to recognise the error
header in this case but I don't really see the point of it.

> At least one can do
>   # grep 'not to get' /var/log/syslog
>   wwwoffles[5218]: The URL 'http://example.net/f.jpg' matches one in the
>   list not to get.
> to know about the former, but what about the latter?

You can only do that if the log files are readable though.

> The latter still ends up in http://localhost:8080/index/outgoing/
> Clicking on it there, still here offline, says
>   Your request for URL
>   
> http://en.wikipedia.org/w/index.php?title=List_of_thinking_errors&action=edit
>   failed because it is on the list of hosts and/or paths that are not to
>   be cached and cannot be requested when offline.
> Well, OK, then it should be barred from ending up in outgoing too.

It is a bug that a request is put into the outgoing directory.  This
happens because the request is for the WWWOFFLE refresh URL not the
original URL.  I have fixed this for the next version.

> OK, to check for the latter one would do, after fetching,
> # less +/not\ possible /var/log/syslog
> (note I use maximum debug level for my messages)
> 
> Anyway, if the shell returned 1 and a message for both, one could much
> easier tell which of one's command line requests one had betted make
> other plans for (fetching by hand, as they are on our DontGet and
> DontCache lists), rather that thinking WWWOFFLE will remember to fetch
> them for us when indeed it has no such plans, and that will be last we
> will see of them until six months later when we realize that somehow
> we never read the Plurbitsky article or whatever that we had on our
> reading list.

It wouldn't work to return an error condition because things would
fall over when multiple URLs were specified on the command line and
only some were not to be fetched.  I don't see the point of adding a
feature that only partly works.

> Anyway, it sure is tough to check in a batch job way what will be
> forgotten. And there's no http://localhost:8080/index/lastfailures --
> but one would rather know if something bad will happen right at
> $ wwwoffle $URL||echo Holmes, make other plans for that $URL. It is on \
> one of your Dont lists.
> 
> Just keeping the mailing list warm, here with latest Debian sid wwwoffle 
> 2.9a-2.

Congratulations on the first post of the year.  Shame that you don't
have the latest version of the program rather than one that is nearly
two years old.

-- 
Andrew.
----------------------------------------------------------------------
Andrew M. Bishop                             [EMAIL PROTECTED]
                                      http://www.gedanken.demon.co.uk/

WWWOFFLE users page:
        http://www.gedanken.demon.co.uk/wwwoffle/version-2.9/user.html

Reply via email to