Hi,
I use urllib2 to grab google.com webpages on my Mac over my Comcast home
network.
I see about 1 error for every 50 pages grabbed. Most exceptions are
ssl.SSLError, very few are socket.error and urllib2.URLError.
The problem is - after a first exception, urllib2 occasionally stalls for upto
an hour (!), at either the urllib2.urlopen or response.read stages.
Apparently the urllib2 and socket timeouts are not effective here - how do I
fix this?
----------------
import urllib2
import socket
from sys import exc_info as sysExc_info
timeout = 2
socket.setdefaulttimeout(timeout)
try :
req = urllib2.Request(url,None,headers)
response = urllib2.urlopen(req,timeout=timeout)
html = response.read()
except :
e = sysExc_info()[0]
open(logfile,'a').write('Exception: %s \n' % e)
< code that follows this : after the first exception, I try again for a few
tries >
--
https://mail.python.org/mailman/listinfo/python-list