On Mar 27, 3:13 pm, "Gabriel Genellina" <[EMAIL PROTECTED]> wrote: > En Tue, 27 Mar 2007 16:21:55 -0300, supercooper <[EMAIL PROTECTED]> > escribió: > > > I am downloading images using the script below. Sometimes it will go > > for 10 mins, sometimes 2 hours before timing out with the following > > error: > > > urllib.urlretrieve(fullurl, localfile) > > IOError: [Errno socket error] (10060, 'Operation timed out') > > > I have searched this forum extensively and tried to avoid timing out, > > but to no avail. Anyone have any ideas as to why I keep getting a > > timeout? I thought setting the socket timeout did it, but it didnt. > > You should do the opposite: timing out *early* -not waiting 2 hours- and > handling the error (maybe using a queue to hold pending requests) > > -- > Gabriel Genellina
Gabriel, thanks for the input. So are you saying there is no way to realistically *prevent* the timeout from occurring in the first place? And by timing out early, do you mean to set the timeout for x seconds and if and when the timeout occurs, handle the error and start the process again somehow on the pending requests? Thanks. chad -- http://mail.python.org/mailman/listinfo/python-list