> From: Mathias Walter [mailto:[EMAIL PROTECTED]
> How can I prevent crawler and bots according to their user agent?
>
> I've put a robots.txt in webapps/ROOT, but this file is not
> read again.
So, to check, the crawlers are not reading your robots.txt and are crawling
your site anyway?
> I'd like to stop crawlers by their useragent string.
What do you mean by "stop"? Do you want to return 404s or similar when a
request with a particular user agent string is received? If so, the obvious
approach would be to write a Filter that is placed in front of your webapp, or
a Valve that is placed in the request processing chain, that examines the user
agent string in the request and returns an appropriate response if you don't
like the agent.
- Peter
---------------------------------------------------------------------
To start a new topic, e-mail: [email protected]
To unsubscribe, e-mail: [EMAIL PROTECTED]
For additional commands, e-mail: [EMAIL PROTECTED]