Hello everybody, really new to python, so bear with me. I am trying to
do some very basic scraping tool. Bascally it just grabs a page xy
times and tells me how long it took. When I do this once, it is
blazingly fast, but when I increase the number of repetitions, it is
slowing down considerably (1 is like 3 ms, 100 takes 6 seconds). I
have done implementations in couple more languages (php, ruby) and
none of them seems to suffer from a similar problem and it seems, that
it behaves linearly. Maybe it is a known issue in urllib2, or I am
simply using it badly. I am using python 2.4.3, machine has CentOS,
below is the sc. Thanks in advance

import urllib2
from datetime import datetime

def application():
    start = datetime.now()
    req = urllib2.Request("http://127.0.0.1/gdc/about";, None,
{'Accept': 'application/json'})
    for number in range(100):
      response = urllib2.urlopen(req)
    end = datetime.now()
    output = end - start
    print output

application()
-- 
http://mail.python.org/mailman/listinfo/python-list

Reply via email to