Hi Andrew,

please take a look at the following Debian bug report.
I've written a few comments at the end.
(Please preserve [EMAIL PROTECTED] in the CC: list when
responding, so that your responses can be tracked by the Debian BTS.)

On Sat 25 Nov 2006, Tim Connors wrote:
> Subject: Bug#400329: wwwoffle: lock-files in concurrent downloading broken 
> either way
> From: Tim Connors <[EMAIL PROTECTED]>
> To: Debian Bug Tracking System <[EMAIL PROTECTED]>

> Package: wwwoffle
> Version: 2.9-2
> Severity: grave
> Justification: causes non-serious data loss
> 
> wwwoffle has the setting:
> # lock-files = yes | no
> #         Enable the use of lock files to stop more than one WWWOFFLE process
> #         from downloading the same URL at the same time (default=no).
> 
> Either way this is set, it is broken.  
> 
> If set to yes, it seems that a lockfile only manages to tell the
> second process to give up loading the page at all, giving back a HTTP
> 500 WWWOFFLE Server Error:
> 
> > for i in `seq 1 10 ` ;do lynx -dump http://www.google.com.au & done
> ....
>              WWWOFFLE - World Wide Web Offline Explorer - v2.9
>      _________________________________________________________________
> 
>                            WWWOFFLE Server Error
> 
>                The WWWOFFLE server encountered a fatal error:
> 
>                  Cannot open the spooled web page to read.
>             The program cannot continue to service this request.
> 
> Help
> 
>    This is an internal WWWOFFLE server error that should not occur. There
>    is no other way of handling the error apart from providing this
>    warning page. If this error continues and WWWOFFLE is configured
>    correctly and there are no other computer problems then you should
>    report the error to the WWWOFFLE author.
>      _________________________________________________________________
> 
>                WWWOFFLE - [[1]Welcome Page|[2]FAQ] - WWWOFFLE
> 
> References
> 
>    1. http://scuzzie:8080/Welcome.html
>    2. http://scuzzie:8080/FAQ.html
> ...
> 
> By loading many pages at once as above, some of them end up succeeding
> after the cache has been loaded, but prior to that, all but one of
> them are going to fail -- above I'd see 5 or so error pages, then 5 or
> so successful downloads -- YMW(ill)V as my computer is probably slow
> enough for this test to come out this way.
> 
> If set to no, then the first process to hit the cache gets their
> download, but the second process only retrieves as much as had reached
> the first process at that time.  So the second download ends up
> incomplete.  No error is returned, so it may not even be apparent
> until a later date -- hence dataloss.
> 
> Marked as grave because of dataloss, since any webpage using crappy
> session management, or submition of forms, etc, ends up being broken
> whenever the page is being concurrently loaded in two seperate
> pages/processes with either setting, and one copy will not be complete
> or will fail altogether.  Most of the rest of the time, a reload ought
> to fix it, but this sometimes ends up loading from the browser's own
> cache, which has only been half downloaded (I can't seem to convince
> opera to drop its version of its cache, and reload from the proxy)


I've responded that it's not a grave loss of data, as that's what the
option is for; say "yes" if you want to prevent that.

That said, the error you get when a page is indeed locked, is a bit
unexpected: "This is an internal WWWOFFLE server error that should not
occur."  It's after all a documented option...
IMHO the second process should wait for the completion of the first
download process before proceeding; giving an "500 WWWOFFLE Server
Error" is not the right thing to do here...


thanks,
Paul Slootman


-- 
To UNSUBSCRIBE, email to [EMAIL PROTECTED]
with a subject of "unsubscribe". Trouble? Contact [EMAIL PROTECTED]

Reply via email to