In my first case:

Squid request:

-MGET 
/cgi-bin/upload_status.cgi?uid=060950223627&files=:iso-27001-router-security-audit-checklist.xls&ok=1
 HTTP/1.1
Accept: text/html, application/xhtml+xml, */*
Referer: http://xyz/
Accept-Language: de-DE
User-Agent: Mozilla/5.0 (Windows NT 6.1; WOW64; Trident/7.0; rv:11.0) like Gecko
Accept-Encoding: gzip, deflate
Host: xyz
X-Forwarded-For: unknown, unknown
Cache-Control: max-age=0
Connection: keep-alive

Webserver answer:
[-MHTTP/1.1 200 OK
Date: Mon, 25 Nov 2013 12:48:57 GMT
Server: Apache/2.2.22 (Linux/SUSE)
Expires: Mon, 26 Jul 1997 05:00:00 GMT
Pragma: no-cache
Keep-Alive: timeout=15, max=100
Connection: Keep-Alive
Transfer-Encoding: chunked
Content-Type: text/html

> Squid send the first request again and again.

Direct request without squid:

Gm/GET /cgi-bin/upload_status.cgi?uid=318568766743&files=:aukirche.JPG&ok=1 
HTTP/1.1
Accept: text/html, application/xhtml+xml, */*
Referer: http://xyz/
Accept-Language: de-DE
User-Agent: Mozilla/5.0 (Windows NT 6.1; WOW64; Trident/7.0; rv:11.0) like Gecko
Accept-Encoding: gzip, deflate
Host: xyz
DNT: 1
Connection: Keep-Alive

Webserver answer:
GmHTTP/1.1 200 OK
Date: Tue, 26 Nov 2013 10:36:25 GMT
Server: Apache/2.2.22 (Linux/SUSE)
Expires: Mon, 26 Jul 1997 05:00:00 GMT
Pragma: no-cache
Keep-Alive: timeout=15, max=100
Connection: Keep-Alive
Transfer-Encoding: chunked
Content-Type: text/html

>Website gets displayed.



In my second case:

Squid request:

SGET / HTTP/1.1
Accept: text/html, application/xhtml+xml, */*
Accept-Language: de-DE
User-Agent: Mozilla/5.0 (Windows NT 6.1; WOW64; Trident/7.0; rv:11.0) like Gecko
Accept-Encoding: gzip, deflate
If-Modified-Since: Tue, 26 Nov 2013 10:52:01 GMT
DNT: 1
Host: xyz
Pragma: no-cache
X-Forwarded-For: unknown, unknown
Cache-Control: max-age=259200
Connection: keep-alive

> No answer from Host

Direct request without squid:

S       GET / HTTP/1.1
Accept: text/html, application/xhtml+xml, */*
Accept-Language: de-DE
User-Agent: Mozilla/5.0 (Windows NT 6.1; WOW64; Trident/7.0; rv:11.0) like Gecko
Accept-Encoding: gzip, deflate
Host: xyz
If-Modified-Since: Tue, 26 Nov 2013 10:52:01 GMT
DNT: 1
Connection: Keep-Alive

> successful answer from Webserver.

Kind regards marc


-----Ursprüngliche Nachricht-----
Von: Grooz, Marc (regio iT) [mailto:marc.gr...@regioit.de] 
Gesendet: Dienstag, 26. November 2013 11:55
An: Kinkie
Cc: squid-users@squid-cache.org
Betreff: [squid-users] ##palin AW: [squid-users] #Can't access certain webpages

Hi Kinkie,

yes i made a capture but don't see the cause.

I send you my traces.

Kind regards.

Marc

-----Ursprüngliche Nachricht-----
Von: Kinkie [mailto:gkin...@gmail.com] 
Gesendet: Montag, 25. November 2013 15:45
An: Grooz, Marc (regio iT)
Cc: squid-users@squid-cache.org
Betreff: Re: [squid-users] #Can't access certain webpages

On Mon, Nov 25, 2013 at 3:21 PM, Grooz, Marc (regio iT) <marc.gr...@regioit.de> 
wrote:
> Hi,
>
> Currently I use Squid 3.3.8 and I can't use/access two webservers thru squid. 
> If I bypass squid this websites work great.
>
> One of this websites is a fileupload/download website with a generated 
> downloadlink. When I upload a file I receive the following Squidlog Entrys:
>
> TCP_MISS/200 398 GET http://w.y.x.z/cgi-bin/upload_status.cgi?
> .
> .
> TCP_MISS_ABORTED/000 0 GET http:// w.y.x.z/cgi-bin/upload_status.cgi?
> TCP_MISS/200 398 GET http://w.y.x.z/cgi-bin/upload_status.cgi?
>
> And the downloadlink never gets generated.
>
>
> In the second case you never get a webpage back from squid. If I use lynx 
> from the commandline of the squid system the Webpage gets loaded.
> With a tcpdump I see that if squid makes the request then the Webserver 
> didn't answer.

Well, this is consistent with the behavior in squid's logs.
Have you tried accessing the misbehaving server from a client running on the 
squid box, and comparing the differences in the network traces?


-- 
    /kinkie

Attachment: smime.p7s
Description: S/MIME cryptographic signature

Reply via email to