Thanks Amo for your fast reply, I meant it ignores the POST requests, I can access the requested page normally but with no post parameters. So, If I want to navigate through the website using POST, I can't and I get only the first page.
Actually I didn't change any configuration !! Squid Version and Configuration options: Squid Cache: Version 3.0.STABLE16 configure options: '--build=i386-redhat-linux-gnu' '--host=i386-redhat-linux-gnu' '--target=i386-redhat-linux-gnu' '--program-prefix=' '--prefix=/usr' '--exec-prefix=/usr' '--bindir=/usr/bin' '--sbindir=/usr/sbin' '--sysconfdir=/etc' '--includedir=/usr/include' '--libdir=/usr/lib' '--libexecdir=/usr/libexec' '--sharedstatedir=/usr/com' '--mandir=/usr/share/man' '--infodir=/usr/share/info' '--exec_prefix=/usr' '--bindir=/usr/sbin' '--libexecdir=/usr/lib/squid' '--localstatedir=/var' '--datadir=/usr/share' '--sysconfdir=/etc/squid' '--disable-dependency-tracking' '--enable-arp-acl' '--enable-auth=basic,digest,ntlm,negotiate' '--enable-basic-auth-helpers=LDAP,MSNT,NCSA,PAM,SMB,YP,getpwnam,multi-domain-NTLM,SASL' '--enable-negotiate-auth-helpers=squid_kerb_auth' '--enable-cache-digests' '--enable-cachemgr-hostname=localhost' '--enable-delay-pools' '--enable-digest-auth-helpers=password' '--enable-epoll' '--enable-external-acl-helpers=ip_user,ldap_group,unix_group,wbinfo_group' '--enable-icap-client' '--enable-ident-lookups' '--with-large-files' '--enable-linux-netfilter' '--enable-ntlm-auth-helpers=SMB,fakeauth' '--enable-referer-log' '--enable-removal-policies=heap,lru' '--enable-snmp' '--enable-ssl' '--enable-storeio=aufs,diskd,null,ufs' '--enable-useragent-log' '--enable-wccpv2' '--with-aio' '--with-default-user=squid' '--with-filedescriptors=16384' '--with-dl' '--with-openssl=/usr/kerberos' '--with-pthreads' 'build_alias=i386-redhat-linux-gnu' 'host_alias=i386-redhat-linux-gnu' 'target_alias=i386-redhat-linux-gnu' 'CFLAGS=-fPIE -Os -g -pipe -fsigned-char -O2 -g -pipe -Wall -Wp,-D_FORTIFY_SOURCE=2 -fexceptions -fstack-protector --param=ssp-buffer-size=4 -m32 -march=i386 -mtune=generic -fasynchronous-unwind-tables' 'LDFLAGS=-pie' 'CXXFLAGS=-fPIE -O2 -g -pipe -Wall -Wp,-D_FORTIFY_SOURCE=2 -fexceptions -fstack-protector --param=ssp-buffer-size=4 -m32 -march=i386 -mtune=generic -fasynchronous-unwind-tables' 'FFLAGS=-O2 -g -pipe -Wall -Wp,-D_FORTIFY_SOURCE=2 -fexceptions -fstack-protector --param=ssp-buffer-size=4 -m32 -march=i386 -mtune=generic -fasynchronous-unwind-tables' But I don't know how to get the HTTP headers of the traffic between squid and the spider. Can squid log such details ? Finally, the spider is trying to access this website, navigate on it using post requests and parse the data. When I use the spider without squid, it works like a charm, once I use squid, the post problem occurs. Thanks again On Tue, May 29, 2012 at 2:29 AM, Amos Jeffries <squ...@treenet.co.nz> wrote: > On 29.05.2012 07:47, Mahmoud Abdel-Fattah wrote: >> >> Hello, >> >> I'm totally new to squid and using it as a proxy for a scrapy (python >> spider ). it works great with all get requests, but it doesn't work >> with large post requests like the requests is generated by ASP.net in >> the following site : http://www.e3050.com/ >> >> So, is there any configuration should I adjust to enable / enlarge >> post requests ? > > > Maybe yes, Maybe no. The answer to this question depends on your definition > of "doesn't work". > > What kind of limitation did you configure to prevent POST working like they > do out of the box? > > We need to know specifics about what is going wrong... > > * software version numbers. > -> check against the latest current release, if its old try with the > current one(s). > > * HTTP headers of the traffic between Squid and the spider if you can get > them. > -> what exactly is the spider trying to do? and how? > --> what does the site do? how is the spider reacting to the response? > > Amos >