Re: [squid-users] Cache Chrome updates

2014-04-20 Thread Amos Jeffries
On 17/04/2014 7:40 a.m., Eliezer Croitoru wrote:
 Hey Amos,
 
 I have a tiny question which I am not sure about the answer(related to
 the topic).
 What would happen in the case which we deny reply or request headers?
 Would squid look at the Vary (headers as an example) and decide if it's
 a Vary object or it would see the request or\and response without
 the headers?

Squid works on the received headers. The request/reply header denial
directives only change the outgoing sent headers.

If you remove/deny a header with the Squid directives all you do is
screw up the recipients abilities which are based on receiving that header.


 Or, would squid suppose to write the full response headers to the Disk
 and Memory object at the same shape it was received from the server?

Yes it would. The removal is done on delivery of HITs.

 As far as I can tell ICAP removes the headers before caching it to disk.

Yes, ICAP works far earlier in the transaction processing. In particular
ICAP is done before caching the object in order to prevent having to
ICAP process all HITs.

Amos



Re: [squid-users] Cache Chrome updates

2014-04-16 Thread Jasper Van Der Westhuizen

  On Tue, 2014-04-15 at 13:11 +0100, Nick Hill wrote:
  This may the the culprit
 
  hierarchy_stoplist cgi-bin ?
 
  I believe this will prevent caching of any URL containing a ?
 
  
  Should I remove the ? and leave cgi-bin?
 
 You can remove the whole line quite safely.
 
 It prevents cache_peers being sent requests that match the regex
 patterns listed. Since it is now very rare to find a peer that cannot
 support those requests...
 
 Amos

Thanks Amos. I will remove the string and test.

Regards
Jasper


Re: [squid-users] Cache Chrome updates

2014-04-16 Thread Jasper Van Der Westhuizen


On Tue, 2014-04-15 at 14:38 +0100, Nick Hill wrote:
 URLs with query strings have traditionally returned dynamic content.
 Consequently, http caches by default tend not to cache content when
 the URL has a query string.
 
 In recent years, notably Microsoft and indeed many others have adopted
 a habit of putting query strings on static content.
 
 This could be somewhat inconvenient on days where Microsoft push out a
 new 4Gb update for windows 8, and you have many such devices connected
 to your nicely cached network. Each device will download exactly the
 same content, but with it's own query string.
 
 The nett result is generation of a huge amount of network traffic.
 Often for surprisingly minor updates.
 
 I am currently testing a new configuration for squid which identifies
 the SHA1 hash of the windows update in the URL, then returns the bit
 perfect cached content, irrespective of a wide set of URL changes. I
 have it in production in a busy computer repair centre. I am
 monitoring the results. So far, very promising.

Hi Nick

As you rightly said, Windows 8 devices are becoming more and more common
now, specially in the work place. I don't want to download the same 4GB
update multiple times. Would you mind sharing your SHA1 hash
configuration or is it perhaps available somewhere?

Regards
Jasper


Re: [squid-users] Cache Chrome updates

2014-04-16 Thread Nick Hill
Hi Jasper
I have compiled 3.4 to provide the store_id functionality implemented
by Ellizer.

I have it running in a production heterogeneous environment.
I'm still checking for bugs, but seems to work well.

#squid.conf file for Squid Cache: Version 3.4.4
#compiled on Ubuntu with configure options:  '--enable-async-io=8'
'--enable-storeio=ufs,aufs,diskd' '--enable-removal-policies=lru,heap'
#'--enable-delay-pools' '--enable-underscores' '--enable-icap-client'
'--enable-follow-x-forwarded-for' '--with-logdir=/var/log/squid3'
#'--with-pidfile=/var/run/squid3.pid' '--with-filedescriptors=65536'
'--with-large-files' '--with-default-user=proxy'
#'--enable-linux-netfilter' '--enable-storeid-rewrite-helpers=file'

#Recommendations: in full production, you may want to set debug
options from 2 to 1 or 0.
#You may also want to comment out strip_query_terms off for user privacy

logformat squid  %tg.%03tu %6tr %a %Ss/%03Hs %st %rm %ru %[un %Sh/%a %mt

#Explicitly define logs for my compiled version
cache_store_log /var/log/squid3/store.log
access_log /var/log/squid3/access.log
cache_log /var/log/squid3/cache.log


#Lets have a fair bit of debugging info
debug_options ALL,2
#Include query strings in logs
strip_query_terms off

acl all src all
#Which domains do windows updates come from?
acl windowsupdate dstdomain .ws.microsoft.com
acl windowsupdate dstdomain .download.windowsupdate.com

acl QUERY urlpath_regex cgi-bin \?

#I'm  behind a NAT firewall, so I don't need to restrict access
http_access allow all

#Uncomment these if you have web apps on the local server which auth
through local ip
#acl to_localhost dst 127.0.0.0/8 0.0.0.0/32
#http_access deny to_localhost

visible_hostname myclient.hostname.com
http_port 3128

#Always optimise bandwidth over hits
cache_replacement_policy heap LFUDA

#Windows update files are HUGE! I have set this to 6Gb.
#A recent (as of Apr 2014) windows 8 update file is 4Gb
maximum_object_size 6 GB

#Set these according to your file system
cache_dir ufs /home/smb/squid/squid 7 16 256
coredump_dir /home/smb/squid/squid


#Guaranteed static content from Microsoft. Usually fetched with range
requests so lets not revalidate. Underscore, 40 hex(SHA1 hash) .
extension
refresh_pattern _[0-9a-f]{40}\.(cab|exe|esd|psf|zip|msi|appx) 518400
80% 518400 override-lastmod override-expire ignore-reload
ignore-must-revalidate ignore-private
#Otherwise potentially variable
refresh_pattern -i
ws.microsoft.com/.*\.(cab|exe|ms[i|u|f]|asf|wm[v|a]|dat|zip|psf|appx|esd)
43200 80% 43200 reload-into-ims
refresh_pattern -i
download.windowsupdate.com/.*\.(cab|exe|ms[i|u|f]|asf|wm[v|a]|dat|zip|psf|appx|esd)
43200 80% 43200 reload-into-ims
#Default refresh patterns last if no others match
refresh_pattern ^ftp: 1440 20% 10080
refresh_pattern ^gopher: 1440 0% 1440
refresh_pattern . 0 20% 4320

#Directive sets I have been experimenting with
#override-lastmod override-expire ignore-reload ignore-must-revalidate
ignore-private
#reload-into-ims

#Windows updates use a lot of range requests. The only way to deal with this
#in Squid is to fetch the whole file as soon as requested
range_offset_limit -1 windowsupdate
quick_abort_min -1 KB windowsupdate


#My internet connection is not just used for Squid. I want to leave
#responsive bandwidth for other services. This limits D/L speed
delay_pools 1
delay_class 1 1
delay_access 1 allow all
delay_parameters 1 120/120

#We use the store_id helper to convert windows update file hashes to bare URLs.
#This way, any fetch for a given hash embedded in the URL will deliver
the same data
#You must make your own /etc/squid3/storeid_rewrite instructiosn at end.
#change the helper program location from
/usr/local/squid/libexec/storeid_file_rewrite to wherever yours is
#It is written in PERL, so on most Linux systems, put it somewhere
convenient, chmod 755 filename
store_id_program /usr/local/squid/libexec/storeid_file_rewrite
/etc/squid3/storeid_rewrite
store_id_children 10 startup=5 idle=3 concurrency=0
store_id_access allow windowsupdate
store_id_access deny all

#We want to cache windowsupdate URLs which include queries
#but only those queries which act on an installable file.
#we don't want to cache queries on asp files as this is a genuine server
#side query as opposed to just a cache breaker
acl wupdatecachablequery urlpath_regex
(cab|exe|ms[i|u|f]|asf|wm[v|a]|dat|zip|psf|appx|appxbundle|esd)\?

cache allow windowsupdate wupdatecachablequery
cache deny QUERY

#Given windows update is un-cooperative towards third party
#methods to reduce network bandwidth, it is safe to presume
#cache-specific headers or dates significantly differing from
#system date will be unhelpful
reply_header_access Date deny windowsupdate
reply_header_access Age deny windowsupdate

#Put the following line in /etc/squid3/storeid_rewrite ommitting the
starting hash. Tab separates fields
#_([0-9a-z]{40})\.(cab|exe|ms[i|u|f]|asf|wm[v|a]|dat|zip|psf|appx|esd)
   http://wupdate.squid.local/$1


Re: [squid-users] Cache Chrome updates

2014-04-16 Thread Amos Jeffries
Hi Nick,
 could you add a section for the WU SHA1 patterns to our DB of useful
StoreID patterns please:

 http://wiki.squid-cache.org/Features/StoreID/DB

Cheers
Amos


Re: [squid-users] Cache Chrome updates

2014-04-16 Thread Eliezer Croitoru

Hey Amos,

I have a tiny question which I am not sure about the answer(related to 
the topic).
What would happen in the case which we deny reply or request headers? 
Would squid look at the Vary (headers as an example) and decide if it's 
a Vary object or it would see the request or\and response without 
the headers?
Or, would squid suppose to write the full response headers to the Disk 
and Memory object at the same shape it was received from the server?

As far as I can tell ICAP removes the headers before caching it to disk.

Eliezer

On 04/16/2014 09:51 AM, Amos Jeffries wrote:

Hi Nick,
  could you add a section for the WU SHA1 patterns to our DB of useful
StoreID patterns please:

  http://wiki.squid-cache.org/Features/StoreID/DB

Cheers
Amos





Re: [squid-users] Cache Chrome updates

2014-04-15 Thread Jasper Van Der Westhuizen

  
  Hi Pieter
  
  No, that gives me an incorrect regular expression error.
 
 NP: regex has an implied .* prefix and suffix on patterns unless you use
 the ^ and $ endpoint anchors.
 
 
 What are the HTTP headers for these requests and replies?
  The 206 status indicates a Range request taking place. But the first
 and third log entry indicate a 283KB range out of a 1400 byte rediretion
 object being delivered. But weird.
 
  A quick scan by redbot indicates that the top URL is a redirection (302
 as logged first but not second time around) to a URL which responds with
 the worlds most annoying Vary header.
 Vary: * effectively makes this response uncacheable. in redbots words.
 
 
 Amos

Hi Amos

Using squidclient I see a document has moved redirect of sorts. The
second one I get this.

--cut--
HTTP/1.0 200 OK
Accept-Ranges: bytes
Content-Length: 530464
Content-Type: application/x-msdos-program
ETag: 3457d
Server: downloads
Vary: *
X-Content-Type-Options: nosniff
X-Frame-Options: SAMEORIGIN
X-Xss-Protection: 1; mode=block
Date: Fri, 11 Apr 2014 05:42:56 GMT
Last-Modified: Thu, 13 Dec 2012 20:00:00 GMT
Alternate-Protocol: 80:quic
X-Cache: MISS from sles-squid2
X-Cache-Lookup: MISS from sles-squid2:8080
Via: 1.0 sles-squid2 (squid/3.1.12)
Connection: close
--cut--

Not sure if this is what you are referring to?

Regards
Jasper


Re: [squid-users] Cache Chrome updates

2014-04-15 Thread Nick Hill
Hi Jaspar

I use an expression like this, which will work on almost all Limux
machines, Cygwin on windows and I expect Mac OSX or a terminal in
Android so long as you have a version of grep similar to GNU grep.

echo 
http://cache.pack.google.com/edgedl/chrome/win/34.0.1847.116_33.0.1750.154_chrome_updater.exe;
| grep -Pi pack.google.com/.*\.(exe|crx)

This shows a match, so I think your expression is OK.

Do you have any refresh_pattern entries earlier in your squid.conf
which may inadvertently match?

I see some of your requests have a query string attached. Many
squid.conf recipies have the following, which will block caching of
URLs with query strings.
acl QUERY urlpath_regex cgi-bin \?
no_cache deny QUERY

If you want to see the part of your query string after the ?  in your
log, you will have to add this to your squid.conf:
strip_query_terms off

But for privacy reasons you may not want to leave this on.


Re: [squid-users] Cache Chrome updates

2014-04-15 Thread Jasper Van Der Westhuizen


On Tue, 2014-04-15 at 12:09 +0100, Nick Hill wrote:
 Hi Jaspar
 
 I use an expression like this, which will work on almost all Limux
 machines, Cygwin on windows and I expect Mac OSX or a terminal in
 Android so long as you have a version of grep similar to GNU grep.
 
 echo 
 http://cache.pack.google.com/edgedl/chrome/win/34.0.1847.116_33.0.1750.154_chrome_updater.exe;
 | grep -Pi pack.google.com/.*\.(exe|crx)
 
 This shows a match, so I think your expression is OK.
 
 Do you have any refresh_pattern entries earlier in your squid.conf
 which may inadvertently match?
 
 I see some of your requests have a query string attached. Many
 squid.conf recipies have the following, which will block caching of
 URLs with query strings.
 acl QUERY urlpath_regex cgi-bin \?
 no_cache deny QUERY
 
 If you want to see the part of your query string after the ?  in your
 log, you will have to add this to your squid.conf:
 strip_query_terms off
 
 But for privacy reasons you may not want to leave this on.
 

Hi Nick

Thanks for the reply. 

I have the following refresh_patterns in my config file.

--cut--
refresh_pattern -i microsoft.com/.*\.(cab|exe|ms[i|u|f]|[ap]sf|wm[v|a]|
dat|zip) 10080 80% 43200 ignore-no-cache ignore-no-store ignore-reload
reload-into-ims

refresh_pattern -i windowsupdate.com/.*\.(cab|exe|ms[i|u|f]|[ap]sf|wm[v|
a]|dat|zip|psf) 10080 80% 43200 ignore-no-cache ignore-no-store
ignore-reload reload-into-ims

refresh_pattern -i windows.com/.*\.(cab|exe|ms[i|u|f]|[ap]sf|wm[v|a]|
dat|zip) 10080 80% 43200 ignore-no-cache ignore-no-store ignore-reload
reload-into-ims

#---
#Google updates
#---

refresh_pattern -i pack.google.com/.*\.(exe|crx) 10080 80% 43200
override-expire override-lastmod ignore-no-cache  ignore-reload
reload-into-ims ignore-private

#
# Apple
#

refresh_pattern -i apple.com/.*\.(pkg|dmg|dat) 10080 80% 43200
ignore-no-cache  ignore-reload  reload-into-ims

#---
# Cache Adobe updates better
#---

refresh_pattern armdl.adobe.com/.*\.(exe|msp|msi) 10800  80%  43200
ignore-no-store ignore-reload reload-into-ims

--cut--

I also have a list of single file type refresh patterns below this list.

The only cgi-bin entries I have in my squid conf file are:

hierarchy_stoplist cgi-bin ?

and

refresh_pattern -i (/cgi-bin/|\?) 0 0%  0

I added the strip_query_terms off setting to my QA system config and
tested again.

15/Apr/2014:13:58:29 +0200302 xxx.xxx.xxx.xxx TCP_MISS/302 1385 GET
http://cache.pack.google.com/edgedl/chrome/win/34.0.1847.116_33.0.1750.154_chrome_updater.exe?
 - DIRECT/173.194.34.64 text/html
15/Apr/2014:13:58:30 +0200   1028 xxx.xxx.xxx.xxx TCP_MISS/200 8704709
GET
http://r2---sn-uxa3vh-j2ue.c.pack.google.com/edgedl/chrome/win/34.0.1847.116_33.0.1750.154_chrome_updater.exe?cms_redirect=yesexpire=1397577513ip=xxx.xxx.xxx.xxxipbits=0ir=1ms=nvhmt=1397563090mv=msparams=expire,ip,ipbitssignature=47FA99194E00E36C1FB3677DC81ACAE1F684AD12.64E6B64942EA9AFCF58B5F13EDFF452AC0D54F49key=cms1
 - DIRECT/165.165.38.13 application/x-msdos-program
15/Apr/2014:13:58:36 +0200156 xxx.xxx.xxx.xxx TCP_MISS/302 1385 GET
http://cache.pack.google.com/edgedl/chrome/win/34.0.1847.116_33.0.1750.154_chrome_updater.exe?
 - DIRECT/173.194.34.64 text/html
15/Apr/2014:13:58:37 +0200   1363 xxx.xxx.xxx.xxx TCP_MISS/200 8704709
GET
http://r2---sn-uxa3vh-j2ue.c.pack.google.com/edgedl/chrome/win/34.0.1847.116_33.0.1750.154_chrome_updater.exe?cms_redirect=yesexpire=1397577519ip=xxx.xxx.xxx.xxxipbits=0ir=1ms=nvhmt=1397563090mv=msparams=expire,ip,ipbitssignature=3D4FCBE80DCCF3C6CB9689B2B5006A580FACA0F2.63E7CA08C16498F695CEB745A03E1FFF2343B0F7key=cms1
 - DIRECT/165.165.38.13 application/x-msdos-program
15/Apr/2014:13:59:04 +0200155 xxx.xxx.xxx.xxx TCP_MISS/302 1385 GET
http://cache.pack.google.com/edgedl/chrome/win/34.0.1847.116_33.0.1750.154_chrome_updater.exe?
 - DIRECT/173.194.34.64 text/html
15/Apr/2014:13:59:05 +0200962 xxx.xxx.xxx.xxx TCP_MISS/200 8704709
GET
http://r2---sn-uxa3vh-j2ue.c.pack.google.com/edgedl/chrome/win/34.0.1847.116_33.0.1750.154_chrome_updater.exe?cms_redirect=yesexpire=1397577548ip=xxx.xxx.xxx.xxxipbits=0ir=1ms=nvhmt=1397563090mv=msparams=expire,ip,ipbitssignature=4C96A3FEA0F319EEFF155C11212659D682B6F432.1ADE1DE4DA72DAB0333AC1014640C6775BCB5382key=cms1
 - DIRECT/165.165.38.13 application/x-msdos-program


-- 
Kind Regards
Jasper 


Re: [squid-users] Cache Chrome updates

2014-04-15 Thread Nick Hill
This may the the culprit

hierarchy_stoplist cgi-bin ?

I believe this will prevent caching of any URL containing a ?


Re: [squid-users] Cache Chrome updates

2014-04-15 Thread Jasper Van Der Westhuizen
 

On Tue, 2014-04-15 at 13:11 +0100, Nick Hill wrote:
 This may the the culprit
 
 hierarchy_stoplist cgi-bin ?
 
 I believe this will prevent caching of any URL containing a ?
 

Should I remove the ? and leave cgi-bin?

Regards
Jasper


Re: [squid-users] Cache Chrome updates

2014-04-15 Thread Nick Hill
URLs with query strings have traditionally returned dynamic content.
Consequently, http caches by default tend not to cache content when
the URL has a query string.

In recent years, notably Microsoft and indeed many others have adopted
a habit of putting query strings on static content.

This could be somewhat inconvenient on days where Microsoft push out a
new 4Gb update for windows 8, and you have many such devices connected
to your nicely cached network. Each device will download exactly the
same content, but with it's own query string.

The nett result is generation of a huge amount of network traffic.
Often for surprisingly minor updates.

I am currently testing a new configuration for squid which identifies
the SHA1 hash of the windows update in the URL, then returns the bit
perfect cached content, irrespective of a wide set of URL changes. I
have it in production in a busy computer repair centre. I am
monitoring the results. So far, very promising.









On 15 April 2014 14:17, Jasper Van Der Westhuizen
jvdwesth...@shoprite.co.za wrote:


 On Tue, 2014-04-15 at 13:11 +0100, Nick Hill wrote:
 This may the the culprit

 hierarchy_stoplist cgi-bin ?

 I believe this will prevent caching of any URL containing a ?


 Should I remove the ? and leave cgi-bin?

 Regards
 Jasper


Re: [squid-users] Cache Chrome updates

2014-04-15 Thread Amos Jeffries
On 16/04/2014 1:17 a.m., Jasper Van Der Westhuizen wrote:
  
 
 On Tue, 2014-04-15 at 13:11 +0100, Nick Hill wrote:
 This may the the culprit

 hierarchy_stoplist cgi-bin ?

 I believe this will prevent caching of any URL containing a ?

 
 Should I remove the ? and leave cgi-bin?

You can remove the whole line quite safely.

It prevents cache_peers being sent requests that match the regex
patterns listed. Since it is now very rare to find a peer that cannot
support those requests...

Amos


Re: [squid-users] Cache Chrome updates

2014-04-15 Thread Eliezer Croitoru

Hey,

It might be not understood but StoreID in squid 3.4.4 might help with that.
I noticed that you are using SLES which I do not have a build node for yet.
Since there are dynamic variables in the url it is hard to cache it.
Note that there are other options in the request and response that can 
cause the object to be un-cachable.


I can think about a pattern for StoreID to be used.
^http:\/\/[\d\w\-]+\.c\.pack\.google\.com\/([\S]+)\? 
http://google_packs.squid.internal/$1


Which should work for many cases even with no strict refresh_pattern.
One of the issues is that you will need to fetch the full object when a 
206 request is being pulled throw squid.


Eliezer

On 04/14/2014 10:32 AM, Jasper Van Der Westhuizen wrote:

Hi all

I'm trying to cache chrome updates, but I see it always fetches over and
over again.

I have the following refresh pattern in my config.

refresh_pattern -i pack.google.com/.*\.(exe|crx) 10080 80% 43200
override-expire override-lastmod ignore-no-cache  ignore-reload
reload-into-ims ignore-private

I see the following behavior in my logs. This is for the same
client(source). Multiple entries, like it gets downloaded over and over
again.
Logs:

1397459574.511199 xxx.xxx.xxx.xxx TCP_MISS/302 1400 GET
http://cache.pack.google.com/edgedl/chrome/win/34.0.1847.116_33.0.1750.154_chrome_updater.exe
 - DEFAULT_PARENT/xxx.xxx.xxx.xxx text/html
1397459579.924   4794 xxx.xxx.xxx.xxx TCP_MISS/206 141330 GET
http://r2---sn-pn-woce.c.pack.google.com/edgedl/chrome/win/34.0.1847.116_33.0.1750.154_chrome_updater.exe?
 - DEFAULT_PARENT/xxx.xxx.xxx.xxx application/x-msdos-program
1397459591.067548 xxx.xxx.xxx.xxx TCP_MISS/302 1400 GET
http://cache.pack.google.com/edgedl/chrome/win/34.0.1847.116_33.0.1750.154_chrome_updater.exe
 - DEFAULT_PARENT/xxx.xxx.xxx.xxx text/html
1397459596.709   4917 xxx.xxx.xxx.xxx TCP_MISS/206 283744 GET
http://r2---sn-pn-woce.c.pack.google.com/edgedl/chrome/win/34.0.1847.116_33.0.1750.154_chrome_updater.exe?
 - DEFAULT_PARENT/xxx.xxx.xxx.xxx
application/x-msdos-program

Is my refresh pattern incorrect?





[squid-users] Cache Chrome updates

2014-04-14 Thread Jasper Van Der Westhuizen
Hi all

I'm trying to cache chrome updates, but I see it always fetches over and
over again. 

I have the following refresh pattern in my config.

refresh_pattern -i pack.google.com/.*\.(exe|crx) 10080 80% 43200
override-expire override-lastmod ignore-no-cache  ignore-reload
reload-into-ims ignore-private

I see the following behavior in my logs. This is for the same
client(source). Multiple entries, like it gets downloaded over and over
again.
Logs:

1397459574.511199 xxx.xxx.xxx.xxx TCP_MISS/302 1400 GET
http://cache.pack.google.com/edgedl/chrome/win/34.0.1847.116_33.0.1750.154_chrome_updater.exe
 - DEFAULT_PARENT/xxx.xxx.xxx.xxx text/html
1397459579.924   4794 xxx.xxx.xxx.xxx TCP_MISS/206 141330 GET
http://r2---sn-pn-woce.c.pack.google.com/edgedl/chrome/win/34.0.1847.116_33.0.1750.154_chrome_updater.exe?
 - DEFAULT_PARENT/xxx.xxx.xxx.xxx application/x-msdos-program
1397459591.067548 xxx.xxx.xxx.xxx TCP_MISS/302 1400 GET
http://cache.pack.google.com/edgedl/chrome/win/34.0.1847.116_33.0.1750.154_chrome_updater.exe
 - DEFAULT_PARENT/xxx.xxx.xxx.xxx text/html
1397459596.709   4917 xxx.xxx.xxx.xxx TCP_MISS/206 283744 GET
http://r2---sn-pn-woce.c.pack.google.com/edgedl/chrome/win/34.0.1847.116_33.0.1750.154_chrome_updater.exe?
 - DEFAULT_PARENT/xxx.xxx.xxx.xxx 
application/x-msdos-program

Is my refresh pattern incorrect?

-- 
Kind Regards
Jasper 


Re: [squid-users] Cache Chrome updates

2014-04-14 Thread Pieter De Wit

On 14/04/2014 19:32, Jasper Van Der Westhuizen wrote:

Hi all

I'm trying to cache chrome updates, but I see it always fetches over and
over again.

I have the following refresh pattern in my config.

refresh_pattern -i pack.google.com/.*\.(exe|crx) 10080 80% 43200
override-expire override-lastmod ignore-no-cache  ignore-reload
reload-into-ims ignore-private

I see the following behavior in my logs. This is for the same
client(source). Multiple entries, like it gets downloaded over and over
again.
Logs:

1397459574.511199 xxx.xxx.xxx.xxx TCP_MISS/302 1400 GET
http://cache.pack.google.com/edgedl/chrome/win/34.0.1847.116_33.0.1750.154_chrome_updater.exe
 - DEFAULT_PARENT/xxx.xxx.xxx.xxx text/html
1397459579.924   4794 xxx.xxx.xxx.xxx TCP_MISS/206 141330 GET
http://r2---sn-pn-woce.c.pack.google.com/edgedl/chrome/win/34.0.1847.116_33.0.1750.154_chrome_updater.exe?
 - DEFAULT_PARENT/xxx.xxx.xxx.xxx application/x-msdos-program
1397459591.067548 xxx.xxx.xxx.xxx TCP_MISS/302 1400 GET
http://cache.pack.google.com/edgedl/chrome/win/34.0.1847.116_33.0.1750.154_chrome_updater.exe
 - DEFAULT_PARENT/xxx.xxx.xxx.xxx text/html
1397459596.709   4917 xxx.xxx.xxx.xxx TCP_MISS/206 283744 GET
http://r2---sn-pn-woce.c.pack.google.com/edgedl/chrome/win/34.0.1847.116_33.0.1750.154_chrome_updater.exe?
 - DEFAULT_PARENT/xxx.xxx.xxx.xxx
application/x-msdos-program

Is my refresh pattern incorrect?


Dag se Jasper :)

Should it not read *pack.google

Cheers,

Pieter


Re: [squid-users] Cache Chrome updates

2014-04-14 Thread Jasper Van Der Westhuizen

  refresh_pattern -i pack.google.com/.*\.(exe|crx) 10080 80% 43200
  override-expire override-lastmod ignore-no-cache  ignore-reload
  reload-into-ims ignore-private
 
  I see the following behavior in my logs. This is for the same
  client(source). Multiple entries, like it gets downloaded over and over
  again.
  Logs:
 
  1397459574.511199 xxx.xxx.xxx.xxx TCP_MISS/302 1400 GET
  http://cache.pack.google.com/edgedl/chrome/win/34.0.1847.116_33.0.1750.154_chrome_updater.exe
   - DEFAULT_PARENT/xxx.xxx.xxx.xxx text/html
  1397459579.924   4794 xxx.xxx.xxx.xxx TCP_MISS/206 141330 GET
  http://r2---sn-pn-woce.c.pack.google.com/edgedl/chrome/win/34.0.1847.116_33.0.1750.154_chrome_updater.exe?
   - DEFAULT_PARENT/xxx.xxx.xxx.xxx application/x-msdos-program
  1397459591.067548 xxx.xxx.xxx.xxx TCP_MISS/302 1400 GET
  http://cache.pack.google.com/edgedl/chrome/win/34.0.1847.116_33.0.1750.154_chrome_updater.exe
   - DEFAULT_PARENT/xxx.xxx.xxx.xxx text/html
  1397459596.709   4917 xxx.xxx.xxx.xxx TCP_MISS/206 283744 GET
  http://r2---sn-pn-woce.c.pack.google.com/edgedl/chrome/win/34.0.1847.116_33.0.1750.154_chrome_updater.exe?
   - DEFAULT_PARENT/xxx.xxx.xxx.xxx
  application/x-msdos-program
 
  Is my refresh pattern incorrect?
 
 Dag se Jasper :)
 
 Should it not read *pack.google
 
 Cheers,
 
 Pieter
 

Hi Pieter

No, that gives me an incorrect regular expression error.

Regards
Jasper


Re: [squid-users] Cache Chrome updates

2014-04-14 Thread Amos Jeffries
On 14/04/2014 10:50 p.m., Jasper Van Der Westhuizen wrote:
 
 refresh_pattern -i pack.google.com/.*\.(exe|crx) 10080 80% 43200
 override-expire override-lastmod ignore-no-cache  ignore-reload
 reload-into-ims ignore-private

 I see the following behavior in my logs. This is for the same
 client(source). Multiple entries, like it gets downloaded over and over
 again.
 Logs:

 1397459574.511199 xxx.xxx.xxx.xxx TCP_MISS/302 1400 GET
 http://cache.pack.google.com/edgedl/chrome/win/34.0.1847.116_33.0.1750.154_chrome_updater.exe
  - DEFAULT_PARENT/xxx.xxx.xxx.xxx text/html
 1397459579.924   4794 xxx.xxx.xxx.xxx TCP_MISS/206 141330 GET
 http://r2---sn-pn-woce.c.pack.google.com/edgedl/chrome/win/34.0.1847.116_33.0.1750.154_chrome_updater.exe?
  - DEFAULT_PARENT/xxx.xxx.xxx.xxx application/x-msdos-program
 1397459591.067548 xxx.xxx.xxx.xxx TCP_MISS/302 1400 GET
 http://cache.pack.google.com/edgedl/chrome/win/34.0.1847.116_33.0.1750.154_chrome_updater.exe
  - DEFAULT_PARENT/xxx.xxx.xxx.xxx text/html
 1397459596.709   4917 xxx.xxx.xxx.xxx TCP_MISS/206 283744 GET
 http://r2---sn-pn-woce.c.pack.google.com/edgedl/chrome/win/34.0.1847.116_33.0.1750.154_chrome_updater.exe?
  - DEFAULT_PARENT/xxx.xxx.xxx.xxx
 application/x-msdos-program

 Is my refresh pattern incorrect?

 Dag se Jasper :)

 Should it not read *pack.google

 Cheers,

 Pieter

 
 Hi Pieter
 
 No, that gives me an incorrect regular expression error.

NP: regex has an implied .* prefix and suffix on patterns unless you use
the ^ and $ endpoint anchors.


What are the HTTP headers for these requests and replies?
 The 206 status indicates a Range request taking place. But the first
and third log entry indicate a 283KB range out of a 1400 byte rediretion
object being delivered. But weird.

 A quick scan by redbot indicates that the top URL is a redirection (302
as logged first but not second time around) to a URL which responds with
the worlds most annoying Vary header.
Vary: * effectively makes this response uncacheable. in redbots words.


Amos