Михајло Анђелковић wrote: > Hello, > > I would ask for allowance to run a request that can be resource > consuming if not properly scaled: > > SELECT page.page_title as title, rev_user_text as user, rev_timestamp > as timestamp, rev_len as len FROM revision JOIN page ON page.page_id = > rev_page WHERE rev_id > 0 AND rev_id < [...] AND rev_deleted = 0; > > This is intended to extract basic data about all publicly visible > revisions from 1 to [...]. Info about each revision would be a 4-tuple > title/user name/time/length. I need this data to start generating a > timeline of editing of srwiki, so it is intended to be run only once > for each revision. > > If this is generally allowed to do, my question is how large chunks of > data can I take at once, and how long should be waited between two > takes? > > M
Have you considered generating the early timeline from dumps? _______________________________________________ Toolserver-l mailing list (Toolserver-l@lists.wikimedia.org) https://lists.wikimedia.org/mailman/listinfo/toolserver-l Posting guidelines for this list: https://wiki.toolserver.org/view/Mailing_list_etiquette