Re: limit_req for spiders only

2013-10-14 Thread Toni Mueller
H Francis, On Mon, Oct 14, 2013 at 03:23:03PM +0100, Francis Daly wrote: > In your map, let $is_spider be empty if is not a spider ("default", > presumably), and be something else if it is a spider (possibly > $binary_remote_addr if every client should be counted individually, > or something els

Re: limit_req for spiders only

2013-10-14 Thread Francis Daly
On Mon, Oct 14, 2013 at 01:59:23PM +0200, Toni Mueller wrote: Hi there, This is untested, but follows the docs at http://nginx.org/r/limit_req_zone: > I therefore constructed a map to > identify spiders, which works well, and then tried to > > limit_req_zone $binary_remote_addr zone=slow:10m ..

Re: limit_req for spiders only

2013-10-14 Thread Toni Mueller
Hello, On Mon, Oct 14, 2013 at 09:25:24AM -0400, Sylvia wrote: > Doesnt robots.txt "Crawl-Delay" directive satisfy your needs? I have it already there, but I don't know how long it takes for such a directive, or any changes to robots.txt for that matter, to take effect. Observing the logs, I'd

Re: limit_req for spiders only

2013-10-14 Thread Sylvia
Hello. Doesnt robots.txt "Crawl-Delay" directive satisfy your needs? Normal spiders should obey robots.txt, if they dont - they can be banned. Posted at Nginx Forum: http://forum.nginx.org/read.php?2,243670,243674#msg-243674 ___ nginx mailing list ng

limit_req for spiders only

2013-10-14 Thread Toni Mueller
Hi, I would like to put a brake on spiders which are hammering a site with dynamic content generation. They should still get to see the content, but only not generate excessive load. I therefore constructed a map to identify spiders, which works well, and then tried to limit_req_zone $binary_rem