Re: [squid-users] WARNING: swapfile header inconsistent with available data

2014-06-27 Thread Jasper Van Der Westhuizen

On Fri, 2014-06-27 at 19:06 +1200, Amos Jeffries wrote:
> On 27/06/2014 6:53 p.m., Jasper Van Der Westhuizen wrote:
> > 
> > 
> > On Thu, 2014-06-26 at 18:03 +0300, Eliezer Croitoru wrote:
> >> On 06/25/2014 04:06 PM, Jasper Van Der Westhuizen wrote:
> >>> As a matter of interest, in my cache logs I see many lines like these
> >>>
> >>> 2014/06/25 14:52:58 kid1| WARNING: swapfile header inconsistent with
> >>> available data
> >>> 2014/06/25 14:53:28 kid1| WARNING: swapfile header inconsistent with
> >>> available data
> >> What Storage are you using?
> >> What is the fstab content?
> >>
> >> Eliezer
> >>
> > 
> > Hi Eliezer
> > 
> > These proxies are VM's with the cache and logs sitting on ext3 LV's.
> > 
> > /dev/VG01/SQUIDCACHE  /var/cache/squid ext3   defaults1  1 
> > 
> > Regards
> > Jasper
> > 
> 
> Do any of those cache_dir share an LV?
> 
> Amos
> 

Hi Amos

No, each proxy server has it's own cache_dir on it's own LV.

Regards
Jasper


Re: [squid-users] WARNING: swapfile header inconsistent with available data

2014-06-26 Thread Jasper Van Der Westhuizen


On Thu, 2014-06-26 at 18:03 +0300, Eliezer Croitoru wrote:
> On 06/25/2014 04:06 PM, Jasper Van Der Westhuizen wrote:
> > As a matter of interest, in my cache logs I see many lines like these
> >
> > 2014/06/25 14:52:58 kid1| WARNING: swapfile header inconsistent with
> > available data
> > 2014/06/25 14:53:28 kid1| WARNING: swapfile header inconsistent with
> > available data
> What Storage are you using?
> What is the fstab content?
> 
> Eliezer
> 

Hi Eliezer

These proxies are VM's with the cache and logs sitting on ext3 LV's.

/dev/VG01/SQUIDCACHE  /var/cache/squid ext3   defaults1  1 

Regards
Jasper


Re: [squid-users] WARNING: swapfile header inconsistent with available data

2014-06-26 Thread Jasper Van Der Westhuizen

> 
> Are you using SMP workers with an AUFS, UFS or diskd cache_dir?
>  UFS/AUFS/diskd are not SMP-aware and this is how it shows up when two
> or more workers are over-writing cache disk files and corrupting each
> others records.
> 
> Amos
> 

Hi Amos

No, I don't make use of multiple SMP workers. Nothing in my config
anyway. 

PS from one of the proxies.

root 25678 1  0 11:00 ?
00:00:00 /usr/local/squid/sbin/squid -sY
squid25680 25678  8 11:00 ?00:03:36 (squid-1) -sY
squid25681 25680  0 11:00 ?00:00:10
(logfile-daemon) /var/log/squid//access.log
squid25682 25680  0 11:00 ?00:00:06 (unlinkd)

Kind Regards
Jasper


[squid-users] WARNING: swapfile header inconsistent with available data

2014-06-25 Thread Jasper Van Der Westhuizen
Hi all

I'm running a compiled version of Squid 3.4.4 and I'm having some
strange behavior lately. I have a two node cluster load balancing via a
F5 LB and at times one of the two servers will simply not complete a
connection. Squid is running, the logs keep rolling(although much slower
and most entries are from it's sibling cache, but no connections go
through. A restart of Squid fixes this.

As a matter of interest, in my cache logs I see many lines like these

2014/06/25 14:52:58 kid1| WARNING: swapfile header inconsistent with
available data
2014/06/25 14:53:28 kid1| WARNING: swapfile header inconsistent with
available data

On one of my servers I cleaned out the cache, as well as the swap.state
file but it didn't make a difference.

squid -v output.

Squid Cache: Version 3.4.4
configure options:  '--prefix=/usr/local/squid'
'--sysconfdir=/etc/squid3.4.4/' '--enable-follow-x-forwarded-for'
'--with-logdir=/var/log/squid/' '-with-pidfile=/var/run/squid.pid'
'--with-swapdir=/var/cache/squid/' '--with-large-files'
'--disable-strict-error-checking' '--with-default-user=squid'

Can anyone shed some light on this problem?

Kind Regards
Jasper


Re: [squid-users] Cache Chrome updates

2014-04-15 Thread Jasper Van Der Westhuizen


On Tue, 2014-04-15 at 14:38 +0100, Nick Hill wrote:
> URLs with query strings have traditionally returned dynamic content.
> Consequently, http caches by default tend not to cache content when
> the URL has a query string.
> 
> In recent years, notably Microsoft and indeed many others have adopted
> a habit of putting query strings on static content.
> 
> This could be somewhat inconvenient on days where Microsoft push out a
> new 4Gb update for windows 8, and you have many such devices connected
> to your nicely cached network. Each device will download exactly the
> same content, but with it's own query string.
> 
> The nett result is generation of a huge amount of network traffic.
> Often for surprisingly minor updates.
> 
> I am currently testing a new configuration for squid which identifies
> the SHA1 hash of the windows update in the URL, then returns the bit
> perfect cached content, irrespective of a wide set of URL changes. I
> have it in production in a busy computer repair centre. I am
> monitoring the results. So far, very promising.

Hi Nick

As you rightly said, Windows 8 devices are becoming more and more common
now, specially in the work place. I don't want to download the same 4GB
update multiple times. Would you mind sharing your SHA1 hash
configuration or is it perhaps available somewhere?

Regards
Jasper


Re: [squid-users] Cache Chrome updates

2014-04-15 Thread Jasper Van Der Westhuizen

> > On Tue, 2014-04-15 at 13:11 +0100, Nick Hill wrote:
> >> This may the the culprit
> >>
> >> hierarchy_stoplist cgi-bin ?
> >>
> >> I believe this will prevent caching of any URL containing a ?
> >>
> > 
> > Should I remove the "?" and leave cgi-bin?
> 
> You can remove the whole line quite safely.
> 
> It prevents cache_peers being sent requests that match the regex
> patterns listed. Since it is now very rare to find a peer that cannot
> support those requests...
> 
> Amos

Thanks Amos. I will remove the string and test.

Regards
Jasper


Re: [squid-users] Cache Chrome updates

2014-04-15 Thread Jasper Van Der Westhuizen
 

On Tue, 2014-04-15 at 13:11 +0100, Nick Hill wrote:
> This may the the culprit
> 
> hierarchy_stoplist cgi-bin ?
> 
> I believe this will prevent caching of any URL containing a ?
> 

Should I remove the "?" and leave cgi-bin?

Regards
Jasper


Re: [squid-users] Cache Chrome updates

2014-04-15 Thread Jasper Van Der Westhuizen


On Tue, 2014-04-15 at 12:09 +0100, Nick Hill wrote:
> Hi Jaspar
> 
> I use an expression like this, which will work on almost all Limux
> machines, Cygwin on windows and I expect Mac OSX or a terminal in
> Android so long as you have a version of grep similar to GNU grep.
> 
> echo 
> "http://cache.pack.google.com/edgedl/chrome/win/34.0.1847.116_33.0.1750.154_chrome_updater.exe";
> | grep -Pi "pack.google.com/.*\.(exe|crx)"
> 
> This shows a match, so I think your expression is OK.
> 
> Do you have any refresh_pattern entries earlier in your squid.conf
> which may inadvertently match?
> 
> I see some of your requests have a query string attached. Many
> squid.conf recipies have the following, which will block caching of
> URLs with query strings.
> acl QUERY urlpath_regex cgi-bin \?
> no_cache deny QUERY
> 
> If you want to see the part of your query string after the ?  in your
> log, you will have to add this to your squid.conf:
> strip_query_terms off
> 
> But for privacy reasons you may not want to leave this on.
> 

Hi Nick

Thanks for the reply. 

I have the following refresh_patterns in my config file.

--cut--
refresh_pattern -i microsoft.com/.*\.(cab|exe|ms[i|u|f]|[ap]sf|wm[v|a]|
dat|zip) 10080 80% 43200 ignore-no-cache ignore-no-store ignore-reload
reload-into-ims

refresh_pattern -i windowsupdate.com/.*\.(cab|exe|ms[i|u|f]|[ap]sf|wm[v|
a]|dat|zip|psf) 10080 80% 43200 ignore-no-cache ignore-no-store
ignore-reload reload-into-ims

refresh_pattern -i windows.com/.*\.(cab|exe|ms[i|u|f]|[ap]sf|wm[v|a]|
dat|zip) 10080 80% 43200 ignore-no-cache ignore-no-store ignore-reload
reload-into-ims

#---
#Google updates
#---

refresh_pattern -i pack.google.com/.*\.(exe|crx) 10080 80% 43200
override-expire override-lastmod ignore-no-cache  ignore-reload
reload-into-ims ignore-private

#
# Apple
#

refresh_pattern -i apple.com/.*\.(pkg|dmg|dat) 10080 80% 43200
ignore-no-cache  ignore-reload  reload-into-ims

#---
# Cache Adobe updates better
#---

refresh_pattern armdl.adobe.com/.*\.(exe|msp|msi) 10800  80%  43200
ignore-no-store ignore-reload reload-into-ims

--cut--

I also have a list of single file type refresh patterns below this list.

The only cgi-bin entries I have in my squid conf file are:

hierarchy_stoplist cgi-bin ?

and

refresh_pattern -i (/cgi-bin/|\?) 0 0%  0

I added the strip_query_terms off setting to my QA system config and
tested again.

15/Apr/2014:13:58:29 +0200302 xxx.xxx.xxx.xxx TCP_MISS/302 1385 GET
http://cache.pack.google.com/edgedl/chrome/win/34.0.1847.116_33.0.1750.154_chrome_updater.exe?
 - DIRECT/173.194.34.64 text/html
15/Apr/2014:13:58:30 +0200   1028 xxx.xxx.xxx.xxx TCP_MISS/200 8704709
GET
http://r2---sn-uxa3vh-j2ue.c.pack.google.com/edgedl/chrome/win/34.0.1847.116_33.0.1750.154_chrome_updater.exe?cms_redirect=yes&expire=1397577513&ip=xxx.xxx.xxx.xxx&ipbits=0&ir=1&ms=nvh&mt=1397563090&mv=m&sparams=expire,ip,ipbits&signature=47FA99194E00E36C1FB3677DC81ACAE1F684AD12.64E6B64942EA9AFCF58B5F13EDFF452AC0D54F49&key=cms1
 - DIRECT/165.165.38.13 application/x-msdos-program
15/Apr/2014:13:58:36 +0200156 xxx.xxx.xxx.xxx TCP_MISS/302 1385 GET
http://cache.pack.google.com/edgedl/chrome/win/34.0.1847.116_33.0.1750.154_chrome_updater.exe?
 - DIRECT/173.194.34.64 text/html
15/Apr/2014:13:58:37 +0200   1363 xxx.xxx.xxx.xxx TCP_MISS/200 8704709
GET
http://r2---sn-uxa3vh-j2ue.c.pack.google.com/edgedl/chrome/win/34.0.1847.116_33.0.1750.154_chrome_updater.exe?cms_redirect=yes&expire=1397577519&ip=xxx.xxx.xxx.xxx&ipbits=0&ir=1&ms=nvh&mt=1397563090&mv=m&sparams=expire,ip,ipbits&signature=3D4FCBE80DCCF3C6CB9689B2B5006A580FACA0F2.63E7CA08C16498F695CEB745A03E1FFF2343B0F7&key=cms1
 - DIRECT/165.165.38.13 application/x-msdos-program
15/Apr/2014:13:59:04 +0200155 xxx.xxx.xxx.xxx TCP_MISS/302 1385 GET
http://cache.pack.google.com/edgedl/chrome/win/34.0.1847.116_33.0.1750.154_chrome_updater.exe?
 - DIRECT/173.194.34.64 text/html
15/Apr/2014:13:59:05 +0200962 xxx.xxx.xxx.xxx TCP_MISS/200 8704709
GET
http://r2---sn-uxa3vh-j2ue.c.pack.google.com/edgedl/chrome/win/34.0.1847.116_33.0.1750.154_chrome_updater.exe?cms_redirect=yes&expire=1397577548&ip=xxx.xxx.xxx.xxx&ipbits=0&ir=1&ms=nvh&mt=1397563090&mv=m&sparams=expire,ip,ipbits&signature=4C96A3FEA0F319EEFF155C11212659D682B6F432.1ADE1DE4DA72DAB0333AC1014640C6775BCB5382&key=cms1
 - DIRECT/165.165.38.13 application/x-msdos-program


-- 
Kind Regards
Jasper 


Re: [squid-users] Cache Chrome updates

2014-04-15 Thread Jasper Van Der Westhuizen

> > 
> > Hi Pieter
> > 
> > No, that gives me an incorrect regular expression error.
> 
> NP: regex has an implied .* prefix and suffix on patterns unless you use
> the ^ and $ endpoint anchors.
> 
> 
> What are the HTTP headers for these requests and replies?
>  The 206 status indicates a Range request taking place. But the first
> and third log entry indicate a 283KB range out of a 1400 byte rediretion
> object being delivered. But weird.
> 
>  A quick scan by redbot indicates that the top URL is a redirection (302
> as logged first but not second time around) to a URL which responds with
> the worlds most annoying Vary header.
> "Vary: * effectively makes this response uncacheable." in redbots words.
> 
> 
> Amos

Hi Amos

Using squidclient I see a "document has moved" redirect of sorts. The
second one I get this.

--cut--
HTTP/1.0 200 OK
Accept-Ranges: bytes
Content-Length: 530464
Content-Type: application/x-msdos-program
ETag: "3457d"
Server: downloads
Vary: *
X-Content-Type-Options: nosniff
X-Frame-Options: SAMEORIGIN
X-Xss-Protection: 1; mode=block
Date: Fri, 11 Apr 2014 05:42:56 GMT
Last-Modified: Thu, 13 Dec 2012 20:00:00 GMT
Alternate-Protocol: 80:quic
X-Cache: MISS from sles-squid2
X-Cache-Lookup: MISS from sles-squid2:8080
Via: 1.0 sles-squid2 (squid/3.1.12)
Connection: close
--cut--

Not sure if this is what you are referring to?

Regards
Jasper


Re: [squid-users] Cache Chrome updates

2014-04-14 Thread Jasper Van Der Westhuizen

> > refresh_pattern -i pack.google.com/.*\.(exe|crx) 10080 80% 43200
> > override-expire override-lastmod ignore-no-cache  ignore-reload
> > reload-into-ims ignore-private
> >
> > I see the following behavior in my logs. This is for the same
> > client(source). Multiple entries, like it gets downloaded over and over
> > again.
> > Logs:
> >
> > 1397459574.511199 xxx.xxx.xxx.xxx TCP_MISS/302 1400 GET
> > http://cache.pack.google.com/edgedl/chrome/win/34.0.1847.116_33.0.1750.154_chrome_updater.exe
> >  - DEFAULT_PARENT/xxx.xxx.xxx.xxx text/html
> > 1397459579.924   4794 xxx.xxx.xxx.xxx TCP_MISS/206 141330 GET
> > http://r2---sn-pn-woce.c.pack.google.com/edgedl/chrome/win/34.0.1847.116_33.0.1750.154_chrome_updater.exe?
> >  - DEFAULT_PARENT/xxx.xxx.xxx.xxx application/x-msdos-program
> > 1397459591.067548 xxx.xxx.xxx.xxx TCP_MISS/302 1400 GET
> > http://cache.pack.google.com/edgedl/chrome/win/34.0.1847.116_33.0.1750.154_chrome_updater.exe
> >  - DEFAULT_PARENT/xxx.xxx.xxx.xxx text/html
> > 1397459596.709   4917 xxx.xxx.xxx.xxx TCP_MISS/206 283744 GET
> > http://r2---sn-pn-woce.c.pack.google.com/edgedl/chrome/win/34.0.1847.116_33.0.1750.154_chrome_updater.exe?
> >  - DEFAULT_PARENT/xxx.xxx.xxx.xxx
> > application/x-msdos-program
> >
> > Is my refresh pattern incorrect?
> >
> Dag se Jasper :)
> 
> Should it not read *pack.google
> 
> Cheers,
> 
> Pieter
> 

Hi Pieter

No, that gives me an incorrect regular expression error.

Regards
Jasper


Re: [squid-users] Blank page on first load

2014-04-14 Thread Jasper Van Der Westhuizen

On Wed, 2014-04-09 at 23:57 +0300, Eliezer Croitoru wrote:
> Hey Jasper,
> 
> Just to make sure I understand:
> What is the issue and is it on specific sites?
> take my site for example http://www1.ngtech.co.il/
> 
> Try to browse into the main page.
> For me squid works fine.
> I had an issue which ICAP settings delayed the page loading but what you 
> describe is not a blank page but an error page.
> Can you look at the development console of IE11 and see what is 
> happening in the network layer?
> 
> Eliezer
> 
> On 04/09/2014 01:05 PM, Jasper Van Der Westhuizen wrote:
> > Hi Amos
> >
> > I built a new proxy, stock standard settings, and tested it again. With
> > IE11 and SPDY/3 enabled I still get the initial "page can not be loaded"
> > problem. When doing a refresh immediately afterward, it loads the page.
> >
> > On the plus side, Sharepoint sites now work.:-)
> >
> > Any suggestions?
> >
> > Kind Regards
> > Jasper

Hi Eliezer

The error message I see on IE looks to be a IE generated message. I can
see the client request(one or two entries) in the logs to google, yet on
IE it displays "page can not be found" almost immediately. It's not a
SQUID generated message. I tried Squid 3.2, 3.3 and 3.4 on a Fedora Core
host. All show the same results. Disabling SPDY/3 fixes the problem on
IE11. Disabling HTTP1.1 on IE10 and 9 fixes it there, although I don't
get many complaints on ver 9 and 10. 

Right now SLES 11 SP3 only supports Squid 3.1.12-8.12.1 via SMT repo.
I'm hoping 3.2 and later will be supported by SLES 11 SP3 soon so that I
can fully support HTTP1.1 on these proxies.

Regards
Jasper


[squid-users] Cache Chrome updates

2014-04-14 Thread Jasper Van Der Westhuizen
Hi all

I'm trying to cache chrome updates, but I see it always fetches over and
over again. 

I have the following refresh pattern in my config.

refresh_pattern -i pack.google.com/.*\.(exe|crx) 10080 80% 43200
override-expire override-lastmod ignore-no-cache  ignore-reload
reload-into-ims ignore-private

I see the following behavior in my logs. This is for the same
client(source). Multiple entries, like it gets downloaded over and over
again.
Logs:

1397459574.511199 xxx.xxx.xxx.xxx TCP_MISS/302 1400 GET
http://cache.pack.google.com/edgedl/chrome/win/34.0.1847.116_33.0.1750.154_chrome_updater.exe
 - DEFAULT_PARENT/xxx.xxx.xxx.xxx text/html
1397459579.924   4794 xxx.xxx.xxx.xxx TCP_MISS/206 141330 GET
http://r2---sn-pn-woce.c.pack.google.com/edgedl/chrome/win/34.0.1847.116_33.0.1750.154_chrome_updater.exe?
 - DEFAULT_PARENT/xxx.xxx.xxx.xxx application/x-msdos-program
1397459591.067548 xxx.xxx.xxx.xxx TCP_MISS/302 1400 GET
http://cache.pack.google.com/edgedl/chrome/win/34.0.1847.116_33.0.1750.154_chrome_updater.exe
 - DEFAULT_PARENT/xxx.xxx.xxx.xxx text/html
1397459596.709   4917 xxx.xxx.xxx.xxx TCP_MISS/206 283744 GET
http://r2---sn-pn-woce.c.pack.google.com/edgedl/chrome/win/34.0.1847.116_33.0.1750.154_chrome_updater.exe?
 - DEFAULT_PARENT/xxx.xxx.xxx.xxx 
application/x-msdos-program

Is my refresh pattern incorrect?

-- 
Kind Regards
Jasper 


Re: [squid-users] Blank page on first load

2014-04-09 Thread Jasper Van Der Westhuizen

> > 
> > You could avoid that by upgrading Squid, perferrably to the current
> > supproted release (3.4.4). I have a client running many IE11 with their
> > default settings behind a Squid-3.4 and not seeing problems.
> > 
> > Amos
> > 
> > 
> 
> Thank you Amos. I will go to 3.4 then.
> 

Hi Amos

I built a new proxy, stock standard settings, and tested it again. With
IE11 and SPDY/3 enabled I still get the initial "page can not be loaded"
problem. When doing a refresh immediately afterward, it loads the page.

On the plus side, Sharepoint sites now work. :-)

Any suggestions?

Kind Regards
Jasper


Re: [squid-users] Blank page on first load

2014-04-07 Thread Jasper Van Der Westhuizen

> > With the first two options enabled in IE and SPDY/3 disabled, google
> > loads fine first time. IE9 doesn't have a SPDY/3 setting, but disabling
> > the HTTP1.1 settings work.
> > 
> > So to me it seems that HTTP1.1 is the problem here(as well as the SPDY/3
> > problem). We run Squid 3.1.
> 
> 
> Okay. Squid-3.1 is still mostly HTTP/1.0 software and IE has problems
> using HTTP/1.1 to a 1.0 proxy.
> 
> You could avoid that by upgrading Squid, perferrably to the current
> supproted release (3.4.4). I have a client running many IE11 with their
> default settings behind a Squid-3.4 and not seeing problems.
> 
> Amos
> 
> 

Thank you Amos. I will go to 3.4 then.

Regards
Jasper


Re: [squid-users] Blank page on first load

2014-04-07 Thread Jasper Van Der Westhuizen

> > Do you see anything coming back *from* the webserver?
> >  Is anything being delivered by Squid to the client?
> 
> Hi Amos
> 
> Yes I do see traffic coming back from the server. 
> 
> What I'm found though was that when going to http://www.google.co.za or
> even http://www.google.com, it redirects to https://www.google.co.za or
> https://www.google.com. That then gives the user the blank page. When
> refreshing the page, it then loads properly.

Hi Amos

I have done more testing and found the following. It seems this problem
is HTTP1.1 related. In IE11 Advanced settings(pic attached) there are 3
settings under HTTP Settings. Use HTTP1.1, Use HTTP 1.1 through proxy
connections and Use SPDY/3. If we disable the first two, sites like
google load first time. After more searches on the web we found an
article that suggested that SPDY/3 was problematic and that a patch is
available.
(http://angrytechnician.wordpress.com/2014/01/16/google-not-loading-first-time-in-ie11-via-a-web-proxy-on-windows-8-1-turn-off-spdy-support/)

With the first two options enabled in IE and SPDY/3 disabled, google
loads fine first time. IE9 doesn't have a SPDY/3 setting, but disabling
the HTTP1.1 settings work.

So to me it seems that HTTP1.1 is the problem here(as well as the SPDY/3
problem). We run Squid 3.1.

Regards
Jasper
<>

Re: [squid-users] Blank page on first load

2014-04-07 Thread Jasper Van Der Westhuizen


On Mon, 2014-04-07 at 18:42 +1200, Pieter De Wit wrote:
> >> My setup is 3 servers running squid 3-3.1.12-8.12.1 behind an F5 load
> >> balancer. From there I send all traffic to a ZScaler cache peer. In my
> >> testing I have bypassed the cache peer but without any success.
> >>
> >> Has anyone come across this problem before?
> >>
> >
> Hi Jasper,
> 
> Have you tried bypassing the F5's ? They try and do a bunch of clever 
> things and this can mess with "normal networking/caching"
> 
> Cheers,
> 
> Pieter
> 

Hi Pieter

I also thought the F5 could be the problem, but I have a QA squid server
that I have been testing with and this server does not sit behind a load
balancer. Same results as with prod.

Regards
Jasper


Re: [squid-users] Blank page on first load

2014-04-07 Thread Jasper Van Der Westhuizen

> > In my squid logs I can see the request going to the website. The client
> > just gets a blank page until they reload it.
> 
> Do you see anything coming back *from* the webserver?
>  Is anything being delivered by Squid to the client?

Hi Amos

Yes I do see traffic coming back from the server. 

What I'm found though was that when going to http://www.google.co.za or
even http://www.google.com, it redirects to https://www.google.co.za or
https://www.google.com. That then gives the user the blank page. When
refreshing the page, it then loads properly.

Regards
Jasper


[squid-users] Blank page on first load

2014-04-06 Thread Jasper Van Der Westhuizen
Hi all

I have a problem with some of my users getting blank pages when loading
sites like google and MSN. They would open the site and get a blank
page, but when refreshing it loads. These users mostly use IE11 but have
had it with browsers like Safari. Although I have to say that 98% of the
time it is with IE10 and 11.

In my squid logs I can see the request going to the website. The client
just gets a blank page until they reload it.

My setup is 3 servers running squid 3-3.1.12-8.12.1 behind an F5 load
balancer. From there I send all traffic to a ZScaler cache peer. In my
testing I have bypassed the cache peer but without any success.

Has anyone come across this problem before?

-- 
Kind Regards
Jasper 


RE: [squid-users] Connection pinning (NTLM pass through)

2012-07-26 Thread Jasper Van Der Westhuizen


> -Original Message-
> From: Amos Jeffries [mailto:squ...@treenet.co.nz]
> Sent: Sunday, May 27, 2012 1:22 PM
> To: squid-users@squid-cache.org
> Subject: Re: [squid-users] Connection pinning (NTLM pass through)
> 
> On 26/05/2012 8:31 a.m., Petter Abrahamsson wrote:
> > Hi,
> >
> > I'm trying to get NTLM pass through to work with squid 3.1.19. I have
> > followed the instructions found on the wiki[1] on connection pinning
> > but I just keep receiving 401 status messages.
> > Below is the very simple squid.conf that I'm using for this test.
> >
> > acl manager proto cache_object
> > acl localhost src 127.0.0.1/32 ::1
> > acl to_localhost dst 127.0.0.0/8 0.0.0.0/32 ::1
> > acl localnet src 192.168.0.0/16 # RFC1918 possible internal network
> > acl SSL_ports port 443
> > acl Safe_ports port 80  # http
> > acl Safe_ports port 21  # ftp
> > acl Safe_ports port 443 # https
> > acl Safe_ports port 70  # gopher
> > acl Safe_ports port 210 # wais
> > acl Safe_ports port 1025-65535  # unregistered ports
> > acl Safe_ports port 280 # http-mgmt
> > acl Safe_ports port 488 # gss-http
> > acl Safe_ports port 591 # filemaker
> > acl Safe_ports port 777 # multiling http
> > acl CONNECT method CONNECT
> > http_access allow manager localhost
> > http_access deny manager
> > http_access deny !Safe_ports
> > http_access deny CONNECT !SSL_ports
> > http_access allow localnet
> > http_access allow localhost
> > http_access deny all
> > http_port 8080 connection-auth=on
> > hierarchy_stoplist cgi-bin ?
> > coredump_dir /var/cache/squid
> > refresh_pattern ^ftp:   144020% 10080
> > refresh_pattern ^gopher:14400%  1440
> > refresh_pattern -i (/cgi-bin/|\?) 0 0%  0
> > refresh_pattern .   0   20% 4320
> >
> > And below is the corresponding access.log entries with obfuscated ip
> > addresses and host names.
> >
> > 1337976537.852 63 192.168.12.214 TCP_MISS/401 466 GET
> > http://www.example.net/directory/ - DIRECT/x.x.x.x text/html
> > 1337976550.714 29 192.168.12.214 TCP_MISS/401 1074 GET
> > http://www.example.net/directory/ - DIRECT/x.x.x.x text/html
> > 1337976551.025 57 192.168.12.214 TCP_MISS/401 466 GET
> > http://www.example.net/directory/ - DIRECT/x.x.x.x text/html
> > 1337976554.627 57 192.168.12.214 TCP_MISS/401 1074 GET
> > http://www.example.net/directory/ - DIRECT/x.x.x.x text/html
> > 1337976558.006   3128 192.168.12.214 TCP_MISS/401 466 GET
> > http://www.example.net/directory/ - DIRECT/x.x.x.x text/html
> > 1337976559.462 59 192.168.12.214 TCP_MISS/401 1074 GET
> > http://www.example.net/directory/ - DIRECT/x.x.x.x text/html
> > 1337976559.760 56 192.168.12.214 TCP_MISS/401 466 GET
> > http://www.example.net/directory/ - DIRECT/x.x.x.x text/html
> >
> > I feel like I'm missing something obvious since the instructions on
> > the wiki are quite simple.
> > When I try the same website through a v2.7 squid it lets me login.
> > Let me know if any other information is needed.
> > Any help would be very much appreciated.
> 
> Check the HTTP headers at each point before/after Squid for keep-alive.
> There is something a little strange going on with HTTP/1.1 connections to
> servers and NTLM keep-alive in 3.1.19. If you are able to do some code
> digging that would help as well.
> 
> Amos


Hi Peter.

I'm having the same issues with a IIS portal site. In 3.1 it doesn't work but 
with 2.7 it does. Have you managed to fix this problem by any chance?

Kind Regards
Jasper



RE: [squid-users] Re: FTP through squid

2012-06-26 Thread Jasper Van Der Westhuizen
> To proxy FTP well you need an FTP proxy. + FTP was designed not to be
> proxied. There is one called frox proxy which handles FTP with some tricks.

Thank you for the tip Amos. I will have a look at Frox now. 

There is another requirement though. Ideally I would like, as in the case with 
Squid, to have users authenticate against AD when using the FTP proxy. Some of 
my users are allowed to FTP and some now. I know this goes beyond the scope of 
Squid and this mailing list, but any suggestions would be welcome.


RE: [squid-users] Re: FTP through squid

2012-06-26 Thread Jasper Van Der Westhuizen
I don't think that is the problem.. If I ftp directly from the squid server to 
my test ftp site, it works fine. Via a browser it works fine. Only when using a 
FTP client, such as FileZilla for example, it fails..



> -Original Message-
> From: Jakob Curdes [mailto:j...@info-systems.de]
> Sent: Tuesday, June 26, 2012 10:43 AM
> To: squid-users@squid-cache.org
> Subject: Re: [squid-users] Re: FTP through squid
> 
> Am 26.06.2012 10:29, schrieb Vishal Agarwal:
> > There are port 21 and 22.
> You probably mean port 20? 22 is  SSH.
> 
> JC


RE: [squid-users] Re: FTP through squid

2012-06-26 Thread Jasper Van Der Westhuizen
Hi 

I'm trying to force all FTP connections direct. I have a parent cache and at 
the moment ftp connections via a brower works fine and is sent directly but my 
problem is that when using a client like filezilla it sends the connection to 
the parent cache and not directly.

I have enabled the following settings:

acl FTP proto FTP
always_direct allow FTP
acl Safe_ports port 21
http_access allow CONNECT Safe_ports

Is there anything I missed?

Regards
Jasper


> -Original Message-
> From: Amos Jeffries [mailto:squ...@treenet.co.nz]
> Sent: Wednesday, May 09, 2012 12:31 AM
> To: squid-users@squid-cache.org
> Subject: Re: [squid-users] Re: FTP through squid
> 
> On 09.05.2012 01:18, Hugo Deprez wrote:
> > Hello,
> >
> > thank you for the information.
> >
> > If I sepcify the following parameter in my configuration file  it :
> > ftp_epsv OFF
> >
> > FTP with firefox is working.
> >
> > I think users will needs to use Filezilla or something equivalent, so
> > what kind of Solution Do I Have ?
> 
> With Squid, only that nasty CONNECT open-proxy situation.
> 
> Frox proxy is the one we usually point people at for FTP protocol proxying. It
> does FTP proxy without needing the HTTP wrapping layers like Squid.
> 
> 
> >
> > My main goal is to setup a squid in order to specify bandwith
> > limitation.
> >
> > By the way, does delay_parameters are apply to Safe_ports ?
> 
> delay_parameters applies to requests which past the delay_access ACL set
> for each pool. Safe_ports is just an ACL test.
> 
> > For information I use the following ACL :
> > delay_access 1 allow all
> 
> Then delay pool #1 in your config is applied to *all* traffic.
> 
> Amos


RE: [squid-users] IP based ACL - regex?

2012-06-23 Thread Jasper Van Der Westhuizen


> -Original Message-
> From: Amos Jeffries [mailto:squ...@treenet.co.nz]
> Sent: Saturday, June 23, 2012 12:18 PM
> To: squid-users@squid-cache.org
> Subject: Re: [squid-users] IP based ACL - regex?
> 
> On 22/06/2012 11:30 p.m., Jasper Van Der Westhuizen wrote:
> > Hi all
> >
> > Could anyone give me some pointers on how to set up a ACL based on
> allowing : If I want to set up an ACL that includes all hosts(different 
> subnets)
> that end in .105, how would I go about?
> 
> Strictly speaking you *can't*. Because hosts don't end in numbers. Hosts are
> alphanumeric names.
> 
> I take it you mean IPv4 addresses, (one host has at least 1 MAC address,
> 2 IPv4 addresses, 3 IPv6 addresses - three of which may be used to contact
> Squid).
> 
> 
> A 1980-style netmask does what you are asking:
> 
> acl foo src 0.0.0.105/0.0.0.255
> 
> But Why? Modern networks use CIDR subnet masking /8, /16, /24, /32, /48,
> /56, /64, /128 etc.
> 
> Amos
> 
Thank you Amos. Yes, I should have been more clear. I have to set the ACL based 
on the IP address, with the last octet ending with for example .105.

The problem with your second proposal is that I don't have a list of all the 
networks. We have various networks and as in the example of the .105 IP 
address, every host in our enormous list networks ending in .105 has the same 
function. If I want to allow these hosts access to only a certain list of 
websites, I need to set up an ACL based on the last octet.

I think your "acl foo src 0.0.0.105/0.0.0.255" idea will work for me and I will 
test it. 

Thanks again.



[squid-users] IP based ACL - regex?

2012-06-22 Thread Jasper Van Der Westhuizen
Hi all

Could anyone give me some pointers on how to set up a ACL based on allowing : 
If I want to set up an ACL that includes all hosts(different subnets) that end 
in .105, how would I go about?

Any help is appreciated.

Regards
Jasper



RE: [squid-users] does a match on an ACL stop or continue?

2012-04-04 Thread Jasper Van Der Westhuizen

>-Original Message-
>From: Greg Whynott [mailto:greg.whyn...@gmail.com] 
>Sent: Wednesday, April 04, 2012 5:04 PM
>To: Squid Users
>Subject: [squid-users] does a match on an ACL stop or continue?
>
>If i have a list of 10 ACLs and a client matches on ACL#4,  will ACLs
>#6-10 be considered or does squid stop evaluating the ACLs and preform the 
>actions defined in ACL#4?
>
>example: 
>
>if someone in the network 10.101.10.0/24 attempts to load "badsite.com",  will 
>they be denied with the ACLs below or will the TOS be modified and the site 
>>loaded?
>
>acl execnetwork src 10.10.10.0/24
>tcp_outgoing_tos 0x38 execnetwork
>
>acl badsite url_regex -i badsite
>http_access deny badsite
>
>
>I ask as it appears to not be consistent with my current setup.
>
>thanks for your time,
>greg

Hi Greg

As far as I know it stops when it hits a rule. Rules are "AND'd "or "OR'd" 
together. 

This is AND: 
Deny if URL is in badsite AND connection from execnetwork.

http_access deny badsite execnetwork

This is OR: 
Deny if connection from execnetwork OR URL is in badsite.

http_access deny execnetwork
http_access deny badsite 

Regards
Jasper



RE: [squid-users] Allowing linked sites - NTLM and un-authenticated users

2012-04-04 Thread Jasper Van Der Westhuizen


-Original Message-
From: Jasper Van Der Westhuizen [mailto:javanderwesthui...@shoprite.co.za] 
Sent: Wednesday, April 04, 2012 11:13 AM
To: squid-users@squid-cache.org
Subject: RE: [squid-users] Allowing linked sites - NTLM and un-authenticated 
users


> This allows my un-authenticated users access to the whitelisted domains and 
> blocks any links in the sites that are not whitelisted(like facebook and 
> youtube). It also allows my authenticated users access to all sites, 
> including whitelisted sites, as well as allowing linked sites like facebook 
> etc.
>
> Do you perhaps see any issue with this setup?

> The only problem I forsee is that srcdomain is the clients IP rDNS record. 
> You have to encode into that what group they are in, so its restricted to 
> clients you > have control over rDNS for. In which case you may as well make 
> them static and use src IP checks.

>Amos

> Hi Amos

> I want to change my setup to do authentication for everyone, and based on 
> whether the user is in a specific group or not, allow them access to certain 
> ACL's.

> I have a group in AD that should have full access. All users should 
> authenticate. If the user is not in my Internet group then he gets to access 
> a list of sites. If  >the user is in the Internet group he gets a different 
> ACL to access everything.

>Is this possible with NTLM? I don't think it is. How would I approach this?

What I did now was use wbinfo_group.pl to help with group ACL's.

--cut--
external_acl_type testForNTGroup1 children=5 %LOGIN /usr/sbin/wbinfo_group.pl
external_acl_type testForNTGroup2 children=5 %LOGIN /usr/sbin/wbinfo_group.pl
acl InternetUsers external testForNTGroup1 "/etc/squid/group_auth_all_access"
acl NonInternetUsers external testForNTGroup2 
"/etc/squid/group_auth_limited_access"

---

### Allow Whitelisted domains to all users
http_access allow InternetUsers
http_access allow NonInternetUsers whitelist
http_access deny NonInternetUsers !whitelist all

--cut--

I place my groups that allowed to have full access in group_auth_all_access. 
The file group_auth_limited_access contains the group that covers my entire AD 
domain.

I think this will work fine. I've tested it successfully.



RE: [squid-users] Allowing linked sites - NTLM and un-authenticated users

2012-04-04 Thread Jasper Van Der Westhuizen

> This allows my un-authenticated users access to the whitelisted domains and 
> blocks any links in the sites that are not whitelisted(like facebook and 
> youtube). It also allows my authenticated users access to all sites, 
> including whitelisted sites, as well as allowing linked sites like facebook 
> etc.
>
> Do you perhaps see any issue with this setup?

> The only problem I forsee is that srcdomain is the clients IP rDNS record. 
> You have to encode into that what group they are in, so its restricted to 
> clients you > have control over rDNS for. In which case you may as well make 
> them static and use src IP checks.

>Amos

Hi Amos

I want to change my setup to do authentication for everyone, and based on 
whether the user is in a specific group or not, allow them access to certain 
ACL's.

I have a group in AD that should have full access. All users should 
authenticate. If the user is not in my Internet group then he gets to access a 
list of sites. If the user is in the Internet group he gets a different ACL to 
access everything.

Is this possible with NTLM? I don't think it is. How would I approach this?



RE: [squid-users] Allowing linked sites - NTLM and un-authenticated users

2012-04-03 Thread Jasper Van Der Westhuizen


-Original Message-
From: Amos Jeffries [mailto:squ...@treenet.co.nz] 
Sent: Tuesday, April 03, 2012 8:43 AM
To: squid-users@squid-cache.org
Subject: Re: [squid-users] Allowing linked sites - NTLM and un-authenticated 
users

On 3/04/2012 6:12 p.m., Jasper Van Der Westhuizen wrote:
>
> -Original Message-
> From: Amos Jeffries [mailto:squ...@treenet.co.nz]
> Sent: Monday, April 02, 2012 9:27 AM
> To: squid-users@squid-cache.org
> Subject: Re: [squid-users] Allowing linked sites - NTLM and 
> un-authenticated users
>
> On 2/04/2012 5:54 p.m., Jasper Van Der Westhuizen wrote:
>> -Original Message-
>> From: Amos Jeffries
>>
>> On 30/03/2012 11:45 p.m., Jasper Van Der Westhuizen wrote:
>>>> Hi everyone
>>>>
>>>> I've been struggling to get a very specific setup going.
>>>>
>>>> Some background:  Our users are split into "Internet" users and 
>>>> "Non-Internet" users. Everyone in a specific AD group is allowed to have 
>>>> full internet>>access. I have two SQUID proxies with squidGuard load 
>>>> balanced with NTLM authentication to handle the group authentication. All 
>>>> traffic also then gets>>sent to a cache peer.
>>>>
>>>> This is basically what I need:
>>>> 1. All users(internet and non-internet) must be able to access sites in 
>>>> "/etc/squid/lists/whitelist.txt"
>>>> 2. If a user wants to access any external site that is not in the 
>>>> whitelist then he must be authenticated. Obviously a non-internet user can 
>>>> try until he is blue>>in the face, it won't work.
>>>>
>>>> These two scenarios are working 100%, except for one irritating bit. Most 
>>>> of the whitelisted sites have got linked websites like facebook or twitter 
>>>> or>>yourtube in them that load icons and graphics or adds etc. This causes 
>>>> a auth-prompt for non-internet users. I can see the requests in the logs 
>>>> being0>>DENIED.
>>>>
>>>> The only way I could think of getting rid of these errors was to 
>>>> implement a "http_access deny !whitelist" after the allow. This 
>>>> works great for non-internet users and it blocks all the linked 
>>>> sites without asking to authenticate, but obviously this breaks 
>>>> access to all other sites for authenticated users.(access denied 
>>>> for all sites)
>>> You can use the "all" hack and two login lines:
>>>
>>> http_access allow whitelist# allow authed users, but dont challenge 
>>> if missing auth http_access allow authed all # block access to some 
>>> sites unless already>logged in http_access deny blacklist 
>>> http_access deny !authed
>>>
>>>
>>> The authed users may still have problems logging in if the first site they 
>>> visit is one of the "blacklist" ones. But if they visit another page first 
>>> they can login>and get there.
>>>
>>>
>>> Amos
>> Hi Amos
>>
>> Thank you for the reply.
>>
>> I think I already tried this method but it still fails. In any case I tried 
>> what you suggested and the problem remains that my 
>> unauthenticated(non-internet)>users can get to the whitelisted sites just 
>> fine, but they still get authentication prompts for the linked content like 
>> facebook and youtube that the site>contains. An example of a site is 
>> http://www.triptrack.co.za/ and you will see what I mean. At the bottom 
>> right of the site there are links to facebook and>youtube. Those links cause 
>> a authentication request to the unauthenticated(or non-internet) users. I 
>> can't have these prompts appear for these users. They>have a set list of 
>> sites they can visit, and it should work for them and should not get asked 
>> to authenticate. Only once they try and go directly to sites that are>not in 
>> the whitelist, should they be prompted, and obviously denied since they are 
>> not included in the AD group.
>> The problem of course is that they *are* going "directly" to the blacklisted 
>> sites when they load an object from those sites. Even if the object was 
>> embeded>in some third-party whitelisted sites HTML.
>> HTTP protocol makes no distinctions about how HTML, XML, or Flash document 
>> structures group objects. All Squid sees is a request for an object on a 
>> non->whitelisted site.
>> Current rules:
>> http_ac

RE: [squid-users] Allowing linked sites - NTLM and un-authenticated users

2012-04-02 Thread Jasper Van Der Westhuizen


-Original Message-
From: Amos Jeffries [mailto:squ...@treenet.co.nz] 
Sent: Monday, April 02, 2012 9:27 AM
To: squid-users@squid-cache.org
Subject: Re: [squid-users] Allowing linked sites - NTLM and un-authenticated 
users

On 2/04/2012 5:54 p.m., Jasper Van Der Westhuizen wrote:
>
> -Original Message-
> From: Amos Jeffries
>
> On 30/03/2012 11:45 p.m., Jasper Van Der Westhuizen wrote:
>>> Hi everyone
>>>
>>> I've been struggling to get a very specific setup going.
>>>
>>> Some background:  Our users are split into "Internet" users and 
>>> "Non-Internet" users. Everyone in a specific AD group is allowed to have 
>>> full internet>>access. I have two SQUID proxies with squidGuard load 
>>> balanced with NTLM authentication to handle the group authentication. All 
>>> traffic also then gets>>sent to a cache peer.
>>>
>>> This is basically what I need:
>>> 1. All users(internet and non-internet) must be able to access sites in 
>>> "/etc/squid/lists/whitelist.txt"
>>> 2. If a user wants to access any external site that is not in the whitelist 
>>> then he must be authenticated. Obviously a non-internet user can try until 
>>> he is blue>>in the face, it won't work.
>>>
>>> These two scenarios are working 100%, except for one irritating bit. Most 
>>> of the whitelisted sites have got linked websites like facebook or twitter 
>>> or>>yourtube in them that load icons and graphics or adds etc. This causes 
>>> a auth-prompt for non-internet users. I can see the requests in the logs 
>>> being0>>DENIED.
>>>
>>> The only way I could think of getting rid of these errors was to 
>>> implement a "http_access deny !whitelist" after the allow. This 
>>> works great for non-internet users and it blocks all the linked 
>>> sites without asking to authenticate, but obviously this breaks 
>>> access to all other sites for authenticated users.(access denied for 
>>> all sites)
>> You can use the "all" hack and two login lines:
>>
>> http_access allow whitelist# allow authed users, but dont challenge 
>> if missing auth http_access allow authed all # block access to some 
>> sites unless already>logged in http_access deny blacklist http_access 
>> deny !authed
>>
>>
>> The authed users may still have problems logging in if the first site they 
>> visit is one of the "blacklist" ones. But if they visit another page first 
>> they can login>and get there.
>>
>>
>> Amos
> Hi Amos
>
> Thank you for the reply.
>
> I think I already tried this method but it still fails. In any case I tried 
> what you suggested and the problem remains that my 
> unauthenticated(non-internet)  >users can get to the whitelisted sites just 
> fine, but they still get authentication prompts for the linked content like 
> facebook and youtube that the site >contains. An example of a site is 
> http://www.triptrack.co.za/ and you will see what I mean. At the bottom right 
> of the site there are links to facebook and >youtube. Those links cause a 
> authentication request to the unauthenticated(or non-internet) users. I can't 
> have these prompts appear for these users. They >have a set list of sites 
> they can visit, and it should work for them and should not get asked to 
> authenticate. Only once they try and go directly to sites that are >not in 
> the whitelist, should they be prompted, and obviously denied since they are 
> not included in the AD group.

>The problem of course is that they *are* going "directly" to the blacklisted 
>sites when they load an object from those sites. Even if the object was 
>embeded >in some third-party whitelisted sites HTML.
>HTTP protocol makes no distinctions about how HTML, XML, or Flash document 
>structures group objects. All Squid sees is a request for an object on a 
>non->whitelisted site.

>
> Current rules:
> http_access allow whitelist
> http_access allow authenticated all
> http_access deny blacklist
> http_access deny !authenticated
>
> Kind Regards
> Jasper
>


Something else I've tried was using a cache_peer_access to pass the whitelisted 
domains that everyone should have access to, to another squid instance that 
should only allow access to the whitelisted sites. Nothing else. Again it works 
kind of. I can see that the proxy sends the request to the cache_peer, but it 
only sends the requested site there and again not any sites that are linked 
within it.(like facebook).

Is there a way to send the entire "session" to the cache_peer if a particular 
domain was requested?



RE: [squid-users] Allowing linked sites - NTLM and un-authenticated users

2012-04-01 Thread Jasper Van Der Westhuizen


-Original Message-
From: Amos Jeffries [mailto:squ...@treenet.co.nz] 
Sent: Saturday, March 31, 2012 10:11 AM
To: squid-users@squid-cache.org
Subject: Re: [squid-users] Allowing linked sites - NTLM and un-authenticated 
users

On 30/03/2012 11:45 p.m., Jasper Van Der Westhuizen wrote:
>> Hi everyone
>>
>> I've been struggling to get a very specific setup going.
>>
>> Some background:  Our users are split into "Internet" users and 
>> "Non-Internet" users. Everyone in a specific AD group is allowed to have 
>> full internet >>access. I have two SQUID proxies with squidGuard load 
>> balanced with NTLM authentication to handle the group authentication. All 
>> traffic also then gets >>sent to a cache peer.
>>
>> This is basically what I need:
>> 1. All users(internet and non-internet) must be able to access sites in 
>> "/etc/squid/lists/whitelist.txt"
>> 2. If a user wants to access any external site that is not in the whitelist 
>> then he must be authenticated. Obviously a non-internet user can try until 
>> he is blue >>in the face, it won't work.
>>
>> These two scenarios are working 100%, except for one irritating bit. Most of 
>> the whitelisted sites have got linked websites like facebook or twitter or 
>> >>yourtube in them that load icons and graphics or adds etc. This causes a 
>> auth-prompt for non-internet users. I can see the requests in the logs 
>> being0 >>DENIED.
>>
>> The only way I could think of getting rid of these errors was to 
>> implement a "http_access deny !whitelist" after the allow. This works 
>> great for non-internet users and it blocks all the linked sites 
>> without asking to authenticate, but obviously this breaks access to 
>> all other sites for authenticated users.(access denied for all sites)

> You can use the "all" hack and two login lines:
>
>http_access allow whitelist# allow authed users, but dont challenge if missing 
>auth http_access allow authed all # block access to some sites unless already 
>>logged in http_access deny blacklist http_access deny !authed
>
>
>The authed users may still have problems logging in if the first site they 
>visit is one of the "blacklist" ones. But if they visit another page first 
>they can login >and get there.
>
>
>Amos

Hi Amos

Thank you for the reply. 

I think I already tried this method but it still fails. In any case I tried 
what you suggested and the problem remains that my 
unauthenticated(non-internet) users can get to the whitelisted sites just fine, 
but they still get authentication prompts for the linked content like facebook 
and youtube that the site contains. An example of a site is 
http://www.triptrack.co.za/ and you will see what I mean. At the bottom right 
of the site there are links to facebook and youtube. Those links cause a 
authentication request to the unauthenticated(or non-internet) users. I can't 
have these prompts appear for these users. They have a set list of sites they 
can visit, and it should work for them and should not get asked to 
authenticate. Only once they try and go directly to sites that are not in the 
whitelist, should they be prompted, and obviously denied since they are not 
included in the AD group.

Current rules:
http_access allow whitelist
http_access allow authenticated all
http_access deny blacklist
http_access deny !authenticated

Kind Regards
Jasper



[squid-users] Allowing linked sites - NTLM and un-authenticated users

2012-03-30 Thread Jasper Van Der Westhuizen
Hi everyone

I've been struggling to get a very specific setup going.

Some background:  Our users are split into "Internet" users and "Non-Internet" 
users. Everyone in a specific AD group is allowed to have full internet access. 
I have two SQUID proxies with squidGuard load balanced with NTLM authentication 
to handle the group authentication. All traffic also then gets sent to a cache 
peer.

This is basically what I need:
1. All users(internet and non-internet) must be able to access sites in 
"/etc/squid/lists/whitelist.txt"
2. If a user wants to access any external site that is not in the whitelist 
then he must be authenticated. Obviously a non-internet user can try until he 
is blue in the face, it won't work.

These two scenarios are working 100%, except for one irritating bit. Most of 
the whitelisted sites have got linked websites like facebook or twitter or 
yourtube in them that load icons and graphics or adds etc. This causes a 
auth-prompt for non-internet users. I can see the requests in the logs being 
DENIED.

The only way I could think of getting rid of these errors was to implement a 
"http_access deny !whitelist" after the allow. This works great for 
non-internet users and it blocks all the linked sites without asking to 
authenticate, but obviously this breaks access to all other sites for 
authenticated users.(access denied for all sites)

I thought of placing the whitelisted sites in squidGuard but NTLM 
authentication happens in squid, before it reaches squidGuard.

Here is an excerpt from my config file:
--cut--
### Whitelisted sites for all users
acl whitelist dstdomain "/etc/squid/lists/whitelist.txt"

acl authenticated proxy_auth REQUIRED

http_access allow whitelist
http_access allow authenticated !whitelist
http_access deny all
--cut--

What I basically need is something like "http_access deny !whitelist 
!authenticated" which means deny all sites not in whitelist for users who are 
not authenticated.  Can this be done?

Any suggestions would be greatly appreciated.

Kind Regards
Jasper