I have a couple of spider bots hitting my server that I do not wish to have
access to my pages - they ignore robots.txt, so I finally put them on my 'deny
from xxxxx' list. This does deny them access but they persist to keep trying -
trying each page address at least 30 times - several hits per second . Is
there a standard method to forward them to some black hole or the FBI or ...?
Charles
---------------------------------
Be a better friend, newshound, and know-it-all with Yahoo! Mobile. Try it now.