Hi Dario and Ellery :) As Nuria said there is a lot of work ongoing on Edit data, but I'll be very interested in discussing how to better productionize / serve the data. Cheers ! Joseph
On Tue, May 3, 2016 at 6:17 AM, Nuria Ruiz <nu...@wikimedia.org> wrote: > >Ellery and I have been talking about the idea of – at the very least – > scheduling the generation of new dumps, if not exposing the data > programmatically. Right now, I am afraid >this is not within my team's > capacity and Analytics has a number of other high-priority areas to focus > on. > > Right. Most of our work in the next couple quarters > revolves around editing data, we need that to be in as good shape as > pageview data is now in the near future so I think we will be able to take > projects like this one again in probably couple quarters. > > I have created a phab task with your awesome use case. > https://phabricator.wikimedia.org/T134231 > > On Mon, May 2, 2016 at 4:51 PM, Dario Taraborelli < > dtarabore...@wikimedia.org> wrote: > >> Hey Thomas, >> >> yes, I agree this dataset is really valuable (just looking at the sheer >> number of downloads [1] and requests for similar data we've received). I >> can see the value of making it more easily accessible via an API. >> >> Ellery and I have been talking about the idea of – at the very least – >> scheduling the generation of new dumps, if not exposing the data >> programmatically. Right now, I am afraid this is not within my team's >> capacity and Analytics has a number of other high-priority areas to focus >> on. We were planning to talk to Joseph et al anyway and decide how to move >> forward (hi Joseph!), we'll report back on the lists as soon as this >> happens. >> >> Dario >> >> [1] https://figshare.altmetric.com/details/3707715 >> >> >> >> On Mon, May 2, 2016 at 3:12 AM, Thomas Steiner <to...@google.com> wrote: >> >>> Hi Dario, >>> >>> This data is super interesting! How realistic is it that your team >>> make it available through the Wikimedia REST API [1]? I would then in >>> turn love to add it to Wikipedia Tools [2], just imagine how amazing >>> it would be to be able to ask a spreadsheet for… >>> >>> =WIKI{OUT|IN}BOUNDTRAFFIC("en:London", TODAY()-2, TODAY()-1) >>> >>> …(or obviously the API respectively) and get the results back >>> immediately without the need to download a dump first. What do you >>> think? >>> >>> Cheers, >>> Tom >>> >>> -- >>> [1] https://wikimedia.org/api/rest_v1/?doc >>> [2] http://bit.ly/wikipedia-tools-add-on >>> >>> -- >>> Dr. Thomas Steiner, Employee (http://blog.tomayac.com, >>> https://twitter.com/tomayac) >>> >>> Google Germany GmbH, ABC-Str. 19, 20354 Hamburg, Germany >>> Managing Directors: Matthew Scott Sucherman, Paul Terence Manicle >>> Registration office and registration number: Hamburg, HRB 86891 >>> >>> -----BEGIN PGP SIGNATURE----- >>> Version: GnuPG v2.0.29 (GNU/Linux) >>> >>> >>> iFy0uwAntT0bE3xtRa5AfeCheCkthAtTh3reSabiGbl0ck0fjumBl3DCharaCTersAttH3b0ttom >>> hTtPs://xKcd.cOm/1181/ >>> -----END PGP SIGNATURE----- >>> >>> _______________________________________________ >>> Analytics mailing list >>> Analytics@lists.wikimedia.org >>> https://lists.wikimedia.org/mailman/listinfo/analytics >>> >> >> >> >> -- >> >> >> *Dario Taraborelli *Head of Research, Wikimedia Foundation >> wikimediafoundation.org • nitens.org • @readermeter >> <http://twitter.com/readermeter> >> >> _______________________________________________ >> Analytics mailing list >> Analytics@lists.wikimedia.org >> https://lists.wikimedia.org/mailman/listinfo/analytics >> >> > > _______________________________________________ > Analytics mailing list > Analytics@lists.wikimedia.org > https://lists.wikimedia.org/mailman/listinfo/analytics > > -- *Joseph Allemandou* Data Engineer @ Wikimedia Foundation IRC: joal
_______________________________________________ Analytics mailing list Analytics@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/analytics