En Tue, 27 Mar 2007 17:41:44 -0300, supercooper <[EMAIL PROTECTED]> escribió:
> On Mar 27, 3:13 pm, "Gabriel Genellina" <[EMAIL PROTECTED]> > wrote: >> En Tue, 27 Mar 2007 16:21:55 -0300, supercooper <[EMAIL PROTECTED]> >> escribió: >> >> > I am downloading images using the script below. Sometimes it will go >> > for 10 mins, sometimes 2 hours before timing out with the following >> > error: >> >> > urllib.urlretrieve(fullurl, localfile) >> > IOError: [Errno socket error] (10060, 'Operation timed out') >> >> > I have searched this forum extensively and tried to avoid timing out, >> > but to no avail. Anyone have any ideas as to why I keep getting a >> > timeout? I thought setting the socket timeout did it, but it didnt. >> >> You should do the opposite: timing out *early* -not waiting 2 hours- and >> handling the error (maybe using a queue to hold pending requests) >> >> -- >> Gabriel Genellina > > Gabriel, thanks for the input. So are you saying there is no way to > realistically *prevent* the timeout from occurring in the first Exactly. The error is out of your control: maybe the server is down, irresponsive, overloaded, a proxy has any problems, any network problem, etc. > place? And by timing out early, do you mean to set the timeout for x > seconds and if and when the timeout occurs, handle the error and start > the process again somehow on the pending requests? Thanks. Exactly! Another option: Python is cool, but there is no need to reinvent the wheel. Use wget instead :) -- Gabriel Genellina -- http://mail.python.org/mailman/listinfo/python-list