RE: [squid-users] Problems Using squid 2.6 as a transparent web cache

2008-06-12 Thread Donoso Gabilondo, Daniel

Hello again,
Thank you very much for your help. 

 I suspect you are trying to do some sort of web mashup involving Squid?
 I've found the best ways to do those is to have squid as the public 
 domain gateway and do the app-linking/routing in the squid config.

I want to use squid to cache all the resources needed by the linux application 
and only download again if they are modified.

I have made the changes that you have indicated me.
I am using firefox to make a test, because with the linux application I can't 
test at this moment. I put squid as the proxy, but always download the resource.

I saw that the store.log file is updating with the asked resources. This is the 
file content:

1213266172.237 RELEASE 00 000F EAEEC8FE1A6E2D8434959FA6301A18A0  200 1213266
171 1194446956-1 video/mpeg 6250477/386763 GET 
http://192.168.240.158:808
0/test/video.mpg
1213266174.770 RELEASE 00 0010 197E8B6BA5687EDF00E293B32088D2E7  200 1213266
174 1194446956-1 video/mpeg 6250477/251763 GET 
http://192.168.240.158:808
0/test/video.mpg

I put maximum_object_size 30 KB because the video.mpg is higher than 8 MB 
(10 MB exactly), but I tried to ask small resources (images) and the results 
are the same.

I read squid configuration and for default squid allow all to be catched. 

What am I doing wrong? 

Thank you again for your help.

Daniel

 


-Mensaje original-
De: Amos Jeffries [mailto:[EMAIL PROTECTED] 
Enviado el: miƩrcoles, 11 de junio de 2008 15:11
Para: Donoso Gabilondo, Daniel
CC: squid-users@squid-cache.org
Asunto: Re: [squid-users] Problems Using squid 2.6 as a transparent web cache

Donoso Gabilondo, Daniel wrote:
 Hello,
 I have an application in linux that uses http resources (videos,
 images..). These resources are in other machine with a http server
 running (under windows).
 
 The linux application always download the resources. I installed and
 configured squid in the linux machine to cache these resources, but the
 linux application always downloads them from the http server. I don't
 know how can I resolve the problem. I need some help, please.

I suspect you are trying to do some sort of web mashup involving Squid?
I've found the best ways to do those is to have squid as the public 
domain gateway and do the app-linking/routing in the squid config.

Anyway on to your various problems

 
 The linux ip address is: 192.168.240.23 and the windows with http server
 ip is: 192.168.233.158
 
 This is my squid.conf file content:
 
 acl all src 0.0.0.0/0.0.0.0
 acl manager proto cache_object
 acl localhost src 127.0.0.1/255.255.255.255
 acl to_localhost dst 127.0.0.0/8
 acl SSL_ports port 443
 acl Safe_ports port 80  # http
 acl Safe_ports port 21  # ftp
 acl Safe_ports port 443 # https
 acl Safe_ports port 70  # gopher
 acl Safe_ports port 210 # wais
 acl Safe_ports port 1025-65535  # unregistered ports
 acl Safe_ports port 280 # http-mgmt
 acl Safe_ports port 488 # gss-http
 acl Safe_ports port 591 # filemaker
 acl Safe_ports port 777 # multiling http
 acl CONNECT method CONNECT
 http_access allow manager localhost
 http_access deny manager
 http_access deny !Safe_ports
 http_access deny CONNECT !SSL_ports
 http_access allow localhost
 http_access deny all

So none of the clients are allowed to make requests?
I'd expect to see a control saying the intercepted network has access 
through.
  acl localnet src 192.168.0.0/16
  http_access deny !localnet

and drop the deny all down a bit

 icp_access allow all

allow all with no port configured? looks like you can kill this.

 hierarchy_stoplist cgi-bin ?
 access_log /var/log/squid/access.log squid
 acl QUERY urlpath_regex cgi-bin \?
 cache deny QUERY
 refresh_pattern ^ftp:   144020% 10080
 refresh_pattern ^gopher:14400%  1440
 refresh_pattern .   0   20% 4320
 acl apache rep_header Server ^Apache
 broken_vary_encoding allow apache
 coredump_dir /var/spool/squid
 cache_dir ufs /var/spool/squid 700 32 512
 http_port 3128 transparent
 icp_port0

 cache_peer  localhost.home.nl parent 8080 0 default
 acl HOMEdstdomain .home.nl

 always_direct  allow all
 never_directallow all

Those lines contradict each other 'everything MUST go direct + nothing 
EVER allowed direct'.

You want just:
   never_direct allow HOME
   never_direct deny all
   cache_peer_access localhost.home.nl allow HOME
   cache_peer_access localhost.home.nl deny all
   http_access allow HOME

  .. the deny I mentioned dropping down goes about here. AFTER the peer 
access config.

 
 
 I executed these commands:
 
 iptables -t nat -A PREROUTING -i eth0 -p tcp --dport 80 -j DNAT --to
 192.168.240.23:3128
 iptables -t nat -A PREROUTING -i eth0 -p tcp --dport 80 -j REDIRECT
 --to-port 3128

Okay so far. What about intercepting the requests clients make directly 
to your web app?
  Since the app knows its running

Re: [squid-users] Problems Using squid 2.6 as a transparent web cache

2008-06-12 Thread Amos Jeffries

Donoso Gabilondo, Daniel wrote:

Hello again,
Thank you very much for your help. 


I suspect you are trying to do some sort of web mashup involving Squid?
I've found the best ways to do those is to have squid as the public 
domain gateway and do the app-linking/routing in the squid config.


I want to use squid to cache all the resources needed by the linux application 
and only download again if they are modified.

I have made the changes that you have indicated me.
I am using firefox to make a test, because with the linux application I can't 
test at this moment. I put squid as the proxy, but always download the resource.

I saw that the store.log file is updating with the asked resources. This is the 
file content:

1213266172.237 RELEASE 00 000F EAEEC8FE1A6E2D8434959FA6301A18A0  200 1213266
171 1194446956-1 video/mpeg 6250477/386763 GET 
http://192.168.240.158:808
0/test/video.mpg
1213266174.770 RELEASE 00 0010 197E8B6BA5687EDF00E293B32088D2E7  200 1213266
174 1194446956-1 video/mpeg 6250477/251763 GET 
http://192.168.240.158:808
0/test/video.mpg

I put maximum_object_size 30 KB because the video.mpg is higher than 8 MB 
(10 MB exactly), but I tried to ask small resources (images) and the results 
are the same.

I read squid configuration and for default squid allow all to be catched. 

What am I doing wrong? 


Thank you again for your help.

Daniel



I think that its the requests that app is making, or possibly the 
headers on the files coming out of the server.


If you can get a trace of the request and response headers before they 
go into squid it would help a lot.


Amos
--
Please use Squid 2.7.STABLE1 or 3.0.STABLE6


RE: [squid-users] Problems Using squid 2.6 as a transparent web cache

2008-06-12 Thread Donoso Gabilondo, Daniel

Here is the trace of the Firefox request:

GET /test/pepe.mpg HTTP/1.0\r\n
Request Method: GET
Request URI: /test/pepe.mpg
Request Version: HTTP/1.0
Host: 192.168.240.22:8080\r\n
User-Agent: Mozilla/5.0 (X11; U; Linux i686; en-US; rv:1.8.1.12)
Gecko/20080208 Fedora/2.0.0.12-1.fc8 Firefox/2.0.0.12\r\n
Accept:
text/xml,application/xml,application/xhtml+xml,text/html;q=0.9,text/plai
n;q=0.8,image/png,*/*;q=0.5\r\n
Accept-Language: en-us,en;q=0.5\r\n
Accept-Encoding: gzip,deflate\r\n
Accept-Charset: ISO-8859-1,utf-8;q=0.7,*;q=0.7\r\n
Via: 1.1 localhost.localdomain:3128 (squid/2.6.STABLE19)\r\n
X-Forwarded-For: 192.168.240.23\r\n
Cache-Control: max-age=259200\r\n
Connection: keep-alive\r\n
\r\n


The answer of the server:
HTTP/1.1 200 OK\r\n
Request Version: HTTP/1.1
Response Code: 200
Server: Apache-Coyote/1.1\r\n
ETag: W/6250477-1194446956686\r\n
Last-Modified: Wed, 07 Nov 2007 14:49:16 GMT\r\n
Content-Type: video/mpeg\r\n
Content-Length: 6250477
Date: Thu, 12 Jun 2008 12:43:11 GMT\r\n
Connection: keep-alive\r\n
\r\n



Here is the trace of linux application request (but I can't set to the
application to use the squid as proxy, I have a problem):

GET /test/pepe.mpg HTTP/1.0\r\n
Request Method: GET
Request URI: /test/pepe.mpg
Request Version: HTTP/1.0
User-Agent: Lavf50.5.0\r\n
Accept: */*\r\n
Host: 192.168.240.22:8080\r\n
Authorization: Basic =\r\n
\r\n

The server answer is the same as the other









 I think that its the requests that app is making, or possibly the 
 headers on the files coming out of the server.

 If you can get a trace of the request and response headers before they

 go into squid it would help a lot.

 Amos
 --
 Please use Squid 2.7.STABLE1 or 3.0.STABLE6
-Mensaje original-
De: Amos Jeffries [mailto:[EMAIL PROTECTED] 
Enviado el: jueves, 12 de junio de 2008 14:21
Para: Donoso Gabilondo, Daniel
CC: squid-users@squid-cache.org
Asunto: Re: [squid-users] Problems Using squid 2.6 as a transparent web
cache

Donoso Gabilondo, Daniel wrote:
 Hello again,
 Thank you very much for your help. 
 
 I suspect you are trying to do some sort of web mashup involving
Squid?
 I've found the best ways to do those is to have squid as the public 
 domain gateway and do the app-linking/routing in the squid config.
 
 I want to use squid to cache all the resources needed by the linux
application and only download again if they are modified.
 
 I have made the changes that you have indicated me.
 I am using firefox to make a test, because with the linux application
I can't test at this moment. I put squid as the proxy, but always
download the resource.
 
 I saw that the store.log file is updating with the asked resources.
This is the file content:
 
 1213266172.237 RELEASE 00 000F EAEEC8FE1A6E2D8434959FA6301A18A0
200 1213266
 171 1194446956-1 video/mpeg 6250477/386763 GET
http://192.168.240.158:808
 0/test/video.mpg
 1213266174.770 RELEASE 00 0010 197E8B6BA5687EDF00E293B32088D2E7
200 1213266
 174 1194446956-1 video/mpeg 6250477/251763 GET
http://192.168.240.158:808
 0/test/video.mpg
 
 I put maximum_object_size 30 KB because the video.mpg is higher
than 8 MB (10 MB exactly), but I tried to ask small resources (images)
and the results are the same.
 
 I read squid configuration and for default squid allow all to be
catched. 
 
 What am I doing wrong? 
 
 Thank you again for your help.
 
 Daniel
 

I think that its the requests that app is making, or possibly the 
headers on the files coming out of the server.

If you can get a trace of the request and response headers before they 
go into squid it would help a lot.

Amos
-- 
Please use Squid 2.7.STABLE1 or 3.0.STABLE6


[squid-users] Problems Using squid 2.6 as a transparent web cache

2008-06-11 Thread Donoso Gabilondo, Daniel
Hello,
I have an application in linux that uses http resources (videos,
images..). These resources are in other machine with a http server
running (under windows).

The linux application always download the resources. I installed and
configured squid in the linux machine to cache these resources, but the
linux application always downloads them from the http server. I don't
know how can I resolve the problem. I need some help, please.

The linux ip address is: 192.168.240.23 and the windows with http server
ip is: 192.168.233.158

This is my squid.conf file content:

acl all src 0.0.0.0/0.0.0.0
acl manager proto cache_object
acl localhost src 127.0.0.1/255.255.255.255
acl to_localhost dst 127.0.0.0/8
acl SSL_ports port 443
acl Safe_ports port 80  # http
acl Safe_ports port 21  # ftp
acl Safe_ports port 443 # https
acl Safe_ports port 70  # gopher
acl Safe_ports port 210 # wais
acl Safe_ports port 1025-65535  # unregistered ports
acl Safe_ports port 280 # http-mgmt
acl Safe_ports port 488 # gss-http
acl Safe_ports port 591 # filemaker
acl Safe_ports port 777 # multiling http
acl CONNECT method CONNECT
http_access allow manager localhost
http_access deny manager
http_access deny !Safe_ports
http_access deny CONNECT !SSL_ports
http_access allow localhost
http_access deny all
icp_access allow all
hierarchy_stoplist cgi-bin ?
access_log /var/log/squid/access.log squid
acl QUERY urlpath_regex cgi-bin \?
cache deny QUERY
refresh_pattern ^ftp:   144020% 10080
refresh_pattern ^gopher:14400%  1440
refresh_pattern .   0   20% 4320
acl apache rep_header Server ^Apache
broken_vary_encoding allow apache
coredump_dir /var/spool/squid
cache_dir ufs /var/spool/squid 700 32 512
http_port 3128 transparent
icp_port0
cache_peer  localhost.home.nl parent 8080 0 default
acl HOMEdstdomain .home.nl
always_direct  allow all
never_directallow all


I executed these commands:

iptables -t nat -A PREROUTING -i eth0 -p tcp --dport 80 -j DNAT --to
192.168.240.23:3128
iptables -t nat -A PREROUTING -i eth0 -p tcp --dport 80 -j REDIRECT
--to-port 3128


The cache.log content is this:

2008/06/11 11:30:52| Starting Squid Cache version 2.6.STABLE19 for
i386-redhat-linux-gnu...
2008/06/11 11:30:52| Process ID 8617
2008/06/11 11:30:52| With 1024 file descriptors available
2008/06/11 11:30:52| Using epoll for the IO loop
2008/06/11 11:30:52| ipcacheAddEntryFromHosts: Bad IP address 'tele1'
2008/06/11 11:30:52| ipcacheAddEntryFromHosts: Bad IP address 'svc1'
2008/06/11 11:30:52| DNS Socket created at 0.0.0.0, port 42897, FD 6
2008/06/11 11:30:52| Adding nameserver 192.168.202.11 from
/etc/resolv.conf
2008/06/11 11:30:52| Adding nameserver 192.168.202.13 from
/etc/resolv.conf
2008/06/11 11:30:52| User-Agent logging is disabled.
2008/06/11 11:30:52| Referer logging is disabled.
2008/06/11 11:30:52| Unlinkd pipe opened on FD 11
2008/06/11 11:30:52| Swap maxSize 716800 KB, estimated 55138 objects
2008/06/11 11:30:52| Target number of buckets: 2756
2008/06/11 11:30:52| Using 8192 Store buckets
2008/06/11 11:30:52| Max Mem  size: 8192 KB
2008/06/11 11:30:52| Max Swap size: 716800 KB
2008/06/11 11:30:52| Local cache digest enabled; rebuild/rewrite every
3600/3600 sec
2008/06/11 11:30:52| Rebuilding storage in /var/spool/squid (CLEAN)
2008/06/11 11:30:52| Using Least Load store dir selection
2008/06/11 11:30:52| Set Current Directory to /var/spool/squid
2008/06/11 11:30:52| Loaded Icons.
2008/06/11 11:30:53| Accepting transparently proxied HTTP connections at
0.0.0.0, port 3128, FD 13.
2008/06/11 11:30:53| WCCP Disabled.
2008/06/11 11:30:53| Ready to serve requests.
2008/06/11 11:30:53| Configuring Parent localhost.home.nl/8080/0
2008/06/11 11:30:53| Done reading /var/spool/squid swaplog (0 entries)
2008/06/11 11:30:53| Finished rebuilding storage from disk.
2008/06/11 11:30:53| 0 Entries scanned
2008/06/11 11:30:53| 0 Invalid entries.
2008/06/11 11:30:53| 0 With invalid flags.
2008/06/11 11:30:53| 0 Objects loaded.
2008/06/11 11:30:53| 0 Objects expired.
2008/06/11 11:30:53| 0 Objects cancelled.
2008/06/11 11:30:53| 0 Duplicate URLs purged.
2008/06/11 11:30:53| 0 Swapfile clashes avoided.
2008/06/11 11:30:53|   Took 0.3 seconds (   0.0 objects/sec).
2008/06/11 11:30:53| Beginning Validation Procedure
2008/06/11 11:30:53|   Completed Validation Procedure
2008/06/11 11:30:53|   Validated 0 Entries
2008/06/11 11:30:53|   store_swap_size = 0k
2008/06/11 11:30:53| storeLateRelease: released 0 objects






Re: [squid-users] Problems Using squid 2.6 as a transparent web cache

2008-06-11 Thread Adrian Chadd

Firstly, the Squid defaults don't allow for very large files to be cached.
maximum_object_size defaults to 4 megabytes.

Secondly, maybe the application and/or http server are not handling caching
logic correctly. Look at the request and response headers.



Adrian

On Wed, Jun 11, 2008, Donoso Gabilondo, Daniel wrote:
 Hello,
 I have an application in linux that uses http resources (videos,
 images..). These resources are in other machine with a http server
 running (under windows).
 
 The linux application always download the resources. I installed and
 configured squid in the linux machine to cache these resources, but the
 linux application always downloads them from the http server. I don't
 know how can I resolve the problem. I need some help, please.
 
 The linux ip address is: 192.168.240.23 and the windows with http server
 ip is: 192.168.233.158
 
 This is my squid.conf file content:
 
 acl all src 0.0.0.0/0.0.0.0
 acl manager proto cache_object
 acl localhost src 127.0.0.1/255.255.255.255
 acl to_localhost dst 127.0.0.0/8
 acl SSL_ports port 443
 acl Safe_ports port 80  # http
 acl Safe_ports port 21  # ftp
 acl Safe_ports port 443 # https
 acl Safe_ports port 70  # gopher
 acl Safe_ports port 210 # wais
 acl Safe_ports port 1025-65535  # unregistered ports
 acl Safe_ports port 280 # http-mgmt
 acl Safe_ports port 488 # gss-http
 acl Safe_ports port 591 # filemaker
 acl Safe_ports port 777 # multiling http
 acl CONNECT method CONNECT
 http_access allow manager localhost
 http_access deny manager
 http_access deny !Safe_ports
 http_access deny CONNECT !SSL_ports
 http_access allow localhost
 http_access deny all
 icp_access allow all
 hierarchy_stoplist cgi-bin ?
 access_log /var/log/squid/access.log squid
 acl QUERY urlpath_regex cgi-bin \?
 cache deny QUERY
 refresh_pattern ^ftp:   144020% 10080
 refresh_pattern ^gopher:14400%  1440
 refresh_pattern .   0   20% 4320
 acl apache rep_header Server ^Apache
 broken_vary_encoding allow apache
 coredump_dir /var/spool/squid
 cache_dir ufs /var/spool/squid 700 32 512
 http_port 3128 transparent
 icp_port0
 cache_peer  localhost.home.nl parent 8080 0 default
 acl HOMEdstdomain .home.nl
 always_direct  allow all
 never_directallow all
 
 
 I executed these commands:
 
 iptables -t nat -A PREROUTING -i eth0 -p tcp --dport 80 -j DNAT --to
 192.168.240.23:3128
 iptables -t nat -A PREROUTING -i eth0 -p tcp --dport 80 -j REDIRECT
 --to-port 3128
 
 
 The cache.log content is this:
 
 2008/06/11 11:30:52| Starting Squid Cache version 2.6.STABLE19 for
 i386-redhat-linux-gnu...
 2008/06/11 11:30:52| Process ID 8617
 2008/06/11 11:30:52| With 1024 file descriptors available
 2008/06/11 11:30:52| Using epoll for the IO loop
 2008/06/11 11:30:52| ipcacheAddEntryFromHosts: Bad IP address 'tele1'
 2008/06/11 11:30:52| ipcacheAddEntryFromHosts: Bad IP address 'svc1'
 2008/06/11 11:30:52| DNS Socket created at 0.0.0.0, port 42897, FD 6
 2008/06/11 11:30:52| Adding nameserver 192.168.202.11 from
 /etc/resolv.conf
 2008/06/11 11:30:52| Adding nameserver 192.168.202.13 from
 /etc/resolv.conf
 2008/06/11 11:30:52| User-Agent logging is disabled.
 2008/06/11 11:30:52| Referer logging is disabled.
 2008/06/11 11:30:52| Unlinkd pipe opened on FD 11
 2008/06/11 11:30:52| Swap maxSize 716800 KB, estimated 55138 objects
 2008/06/11 11:30:52| Target number of buckets: 2756
 2008/06/11 11:30:52| Using 8192 Store buckets
 2008/06/11 11:30:52| Max Mem  size: 8192 KB
 2008/06/11 11:30:52| Max Swap size: 716800 KB
 2008/06/11 11:30:52| Local cache digest enabled; rebuild/rewrite every
 3600/3600 sec
 2008/06/11 11:30:52| Rebuilding storage in /var/spool/squid (CLEAN)
 2008/06/11 11:30:52| Using Least Load store dir selection
 2008/06/11 11:30:52| Set Current Directory to /var/spool/squid
 2008/06/11 11:30:52| Loaded Icons.
 2008/06/11 11:30:53| Accepting transparently proxied HTTP connections at
 0.0.0.0, port 3128, FD 13.
 2008/06/11 11:30:53| WCCP Disabled.
 2008/06/11 11:30:53| Ready to serve requests.
 2008/06/11 11:30:53| Configuring Parent localhost.home.nl/8080/0
 2008/06/11 11:30:53| Done reading /var/spool/squid swaplog (0 entries)
 2008/06/11 11:30:53| Finished rebuilding storage from disk.
 2008/06/11 11:30:53| 0 Entries scanned
 2008/06/11 11:30:53| 0 Invalid entries.
 2008/06/11 11:30:53| 0 With invalid flags.
 2008/06/11 11:30:53| 0 Objects loaded.
 2008/06/11 11:30:53| 0 Objects expired.
 2008/06/11 11:30:53| 0 Objects cancelled.
 2008/06/11 11:30:53| 0 Duplicate URLs purged.
 2008/06/11 11:30:53| 0 Swapfile clashes avoided.
 2008/06/11 11:30:53|   Took 0.3 seconds (   0.0 objects/sec).
 2008/06/11 11:30:53| Beginning Validation Procedure
 2008/06/11 11:30:53|   Completed Validation Procedure
 2008/06/11 11:30:53|   Validated 0 Entries
 2008/06/11 11:30:53|   

Re: [squid-users] Problems Using squid 2.6 as a transparent web cache

2008-06-11 Thread Amos Jeffries

Donoso Gabilondo, Daniel wrote:

Hello,
I have an application in linux that uses http resources (videos,
images..). These resources are in other machine with a http server
running (under windows).

The linux application always download the resources. I installed and
configured squid in the linux machine to cache these resources, but the
linux application always downloads them from the http server. I don't
know how can I resolve the problem. I need some help, please.


I suspect you are trying to do some sort of web mashup involving Squid?
I've found the best ways to do those is to have squid as the public 
domain gateway and do the app-linking/routing in the squid config.


Anyway on to your various problems



The linux ip address is: 192.168.240.23 and the windows with http server
ip is: 192.168.233.158

This is my squid.conf file content:

acl all src 0.0.0.0/0.0.0.0
acl manager proto cache_object
acl localhost src 127.0.0.1/255.255.255.255
acl to_localhost dst 127.0.0.0/8
acl SSL_ports port 443
acl Safe_ports port 80  # http
acl Safe_ports port 21  # ftp
acl Safe_ports port 443 # https
acl Safe_ports port 70  # gopher
acl Safe_ports port 210 # wais
acl Safe_ports port 1025-65535  # unregistered ports
acl Safe_ports port 280 # http-mgmt
acl Safe_ports port 488 # gss-http
acl Safe_ports port 591 # filemaker
acl Safe_ports port 777 # multiling http
acl CONNECT method CONNECT
http_access allow manager localhost
http_access deny manager
http_access deny !Safe_ports
http_access deny CONNECT !SSL_ports
http_access allow localhost
http_access deny all


So none of the clients are allowed to make requests?
I'd expect to see a control saying the intercepted network has access 
through.

 acl localnet src 192.168.0.0/16
 http_access deny !localnet

and drop the deny all down a bit


icp_access allow all


allow all with no port configured? looks like you can kill this.


hierarchy_stoplist cgi-bin ?
access_log /var/log/squid/access.log squid
acl QUERY urlpath_regex cgi-bin \?
cache deny QUERY
refresh_pattern ^ftp:   144020% 10080
refresh_pattern ^gopher:14400%  1440
refresh_pattern .   0   20% 4320
acl apache rep_header Server ^Apache
broken_vary_encoding allow apache
coredump_dir /var/spool/squid
cache_dir ufs /var/spool/squid 700 32 512
http_port 3128 transparent
icp_port0



cache_peer  localhost.home.nl parent 8080 0 default
acl HOMEdstdomain .home.nl



always_direct  allow all
never_directallow all


Those lines contradict each other 'everything MUST go direct + nothing 
EVER allowed direct'.


You want just:
  never_direct allow HOME
  never_direct deny all
  cache_peer_access localhost.home.nl allow HOME
  cache_peer_access localhost.home.nl deny all
  http_access allow HOME

 .. the deny I mentioned dropping down goes about here. AFTER the peer 
access config.





I executed these commands:

iptables -t nat -A PREROUTING -i eth0 -p tcp --dport 80 -j DNAT --to
192.168.240.23:3128
iptables -t nat -A PREROUTING -i eth0 -p tcp --dport 80 -j REDIRECT
--to-port 3128


Okay so far. What about intercepting the requests clients make directly 
to your web app?
 Since the app knows its running on port 8080 it will tell the clients 
that in its URLs, and the 'clients' do not know about Squid they will 
not ask for those objects over port 80.





The cache.log content is this:

2008/06/11 11:30:52| Starting Squid Cache version 2.6.STABLE19 for
i386-redhat-linux-gnu...
2008/06/11 11:30:52| Process ID 8617
2008/06/11 11:30:52| With 1024 file descriptors available
2008/06/11 11:30:52| Using epoll for the IO loop
2008/06/11 11:30:52| ipcacheAddEntryFromHosts: Bad IP address 'tele1'
2008/06/11 11:30:52| ipcacheAddEntryFromHosts: Bad IP address 'svc1'


Your hosts file has corrupt content.

Apart from all that, squid looks to be running fine.


Amos
--
Please use Squid 2.7.STABLE1 or 3.0.STABLE6