On Thu, 2003-06-26 at 08:41, Henrik Nordstrom wrote:

> What is a viable approach is to add a second database for this purpose 
> in parallell to Squid, keeping track of the URLs in the cache. This 
> database can be built automatically by tracking the store.log log 
> file and feeding the data into a database of choice. For tracking the 
> store.log file the per File::Tail module is very suitable, but some 
> database design is probably needed to get a database which can be 
> searched in interesting ways.

Another viable approach is to build / find and reuse a non-blocking (via
the squid io logic probably - or via pipes to a another process) a
pageable Trie, and insert and remove URL's there as squid does it's work
that would allow a fixed size memory usage (say 4Kb per page, 1000 pages
would probably be enough for good locality of reference), and not slow
squid appreciably. It's definately not a 3.0 feature though - we already
have many big changes, and getting the current migrated code stable will
be enough on it's own.

Rob

-- 
GPG key available at: <http://members.aardvark.net.au/lifeless/keys.txt>.

Attachment: signature.asc
Description: This is a digitally signed message part

Reply via email to