Hi Ostapuk Natalia,

Thanks for writing the list. There may be some alternative approaches -
I'll follow up with you off list.

-Adam



On Mon, Oct 24, 2016 at 4:11 AM, Остапук Наталья <natax...@yandex-team.ru>
wrote:

> Hello,
>
> My name is Natalia, I'm a software developer at Yandex (
> https://www.yandex.ru <https://www.yandex.ru/)&>). My team is building a
> large database, which contains objects from real world, their attributes
> and different types of relationships between them.
>
> Wikipedia is one of our main data sources. Until recently we've been
> downloading data from it in a half-manual mode, but now we want to make it
> automatically with our bots. It works fine for all what we need except for
> API pages. As robots.txt states "Friendly, low-speed bots are welcome
> viewing article pages, but not dynamically-generated pages please."
>
> Our bot uses crawl-delay so it won't bother you more often than you allow.
> Moreover, we are planning to make no more than 20-30 requests per 5 minutes.
>
> Is there some way to add an exception to robots.txt for our bot?
>
> Thank you for your time.
>
> Best regards,
> Ostapuk Natalia
>
> _______________________________________________
> Mediawiki-api mailing list
> Mediawiki-api@lists.wikimedia.org
> https://lists.wikimedia.org/mailman/listinfo/mediawiki-api
>
>
_______________________________________________
Mediawiki-api mailing list
Mediawiki-api@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/mediawiki-api

Reply via email to