For information about analyzing search engines requests,
you might want to look at SEARCHENGINE commands
in the Configuration Files section of the Helper Applications
page (http://www.analog.cx/helpers/index.html#conffiles) and
at the SEARCHENGINE command itself in the Analog docs
(http://www.analog.cx/docs/args.html#SEARCHENGINE).
-- Duke Hillard, University Webmaster, UL Lafayette
Ian Stong wrote:
Perhaps a bit of a side topic - but trying to see what people are doing
relating to Spiders, Crawlers and Search Engines and the logging, blocking,
tracking of them, etc.
Specifically wondering if it's possible to limit spiders/crawlers to only
certain parts of your web site (perhaps with robot.txt files)?
Also found a 3rd party file with Analog called srch.cfg that lists search
engines, etc in the file. I've mentioned it in my analog.cfg file but not
sure how to make use of it. Is it used to generate a seperate report to track
search engine accesses to your site or something else?
Looking for someone who has implemented ways to track spider. crawler, search
engine access to a site and ways to minimize/block that traffic.
Thanks in advance,
Ian
begin:vcard
n:Hillard;Duke
tel;work:337-482-5763
x-mozilla-html:TRUE
url:http://www.louisiana.edu/
org:University of Louisiana at Lafayette;University Computing Support Services
adr:;;P.O. Box 42770;Lafayette;LA;70504-2770;USA
version:2.1
email;internet:[EMAIL PROTECTED]
title:Computing Resources Coordinator
fn:Duke Hillard
end:vcard