Hi Dario and Ellery :)
As Nuria said there is a lot of work ongoing on Edit data,  but I'll be
very interested in discussing how to better productionize / serve the data.
Cheers !
Joseph



On Tue, May 3, 2016 at 6:17 AM, Nuria Ruiz <nu...@wikimedia.org> wrote:

> >Ellery and I have been talking about the idea of – at the very least –
> scheduling the generation of new dumps, if not exposing the data
> programmatically. Right now, I am afraid >this is not within my team's
> capacity and Analytics has a number of other high-priority areas to focus
> on.
>
> Right. Most of our work in the next couple quarters
> revolves around editing data, we need that to be in as good shape as
> pageview data is now in the near future so I think we will be able to take
> projects like this one again in probably couple quarters.
>
> I have created a phab task with your awesome use case.
> https://phabricator.wikimedia.org/T134231
>
> On Mon, May 2, 2016 at 4:51 PM, Dario Taraborelli <
> dtarabore...@wikimedia.org> wrote:
>
>> Hey Thomas,
>>
>> yes, I agree this dataset is really valuable (just looking at the sheer
>> number of downloads  [1] and requests for similar data we've received). I
>> can see the value of making it more easily accessible via an API.
>>
>> Ellery and I have been talking about the idea of – at the very least –
>> scheduling the generation of new dumps, if not exposing the data
>> programmatically. Right now, I am afraid this is not within my team's
>> capacity and Analytics has a number of other high-priority areas to focus
>> on. We were planning to talk to Joseph et al anyway and decide how to move
>> forward (hi Joseph!), we'll report back on the lists as soon as this
>> happens.
>>
>> Dario
>>
>> [1] https://figshare.altmetric.com/details/3707715
>>
>>
>>
>> On Mon, May 2, 2016 at 3:12 AM, Thomas Steiner <to...@google.com> wrote:
>>
>>> Hi Dario,
>>>
>>> This data is super interesting! How realistic is it that your team
>>> make it available through the Wikimedia REST API [1]? I would then in
>>> turn love to add it to Wikipedia Tools [2], just imagine how amazing
>>> it would be to be able to ask a spreadsheet for…
>>>
>>>   =WIKI{OUT|IN}BOUNDTRAFFIC("en:London", TODAY()-2, TODAY()-1)
>>>
>>> …(or obviously the API respectively) and get the results back
>>> immediately without the need to download a dump first. What do you
>>> think?
>>>
>>> Cheers,
>>> Tom
>>>
>>> --
>>> [1] https://wikimedia.org/api/rest_v1/?doc
>>> [2] http://bit.ly/wikipedia-tools-add-on
>>>
>>> --
>>> Dr. Thomas Steiner, Employee (http://blog.tomayac.com,
>>> https://twitter.com/tomayac)
>>>
>>> Google Germany GmbH, ABC-Str. 19, 20354 Hamburg, Germany
>>> Managing Directors: Matthew Scott Sucherman, Paul Terence Manicle
>>> Registration office and registration number: Hamburg, HRB 86891
>>>
>>> -----BEGIN PGP SIGNATURE-----
>>> Version: GnuPG v2.0.29 (GNU/Linux)
>>>
>>>
>>> iFy0uwAntT0bE3xtRa5AfeCheCkthAtTh3reSabiGbl0ck0fjumBl3DCharaCTersAttH3b0ttom
>>> hTtPs://xKcd.cOm/1181/
>>> -----END PGP SIGNATURE-----
>>>
>>> _______________________________________________
>>> Analytics mailing list
>>> Analytics@lists.wikimedia.org
>>> https://lists.wikimedia.org/mailman/listinfo/analytics
>>>
>>
>>
>>
>> --
>>
>>
>> *Dario Taraborelli  *Head of Research, Wikimedia Foundation
>> wikimediafoundation.org • nitens.org • @readermeter
>> <http://twitter.com/readermeter>
>>
>> _______________________________________________
>> Analytics mailing list
>> Analytics@lists.wikimedia.org
>> https://lists.wikimedia.org/mailman/listinfo/analytics
>>
>>
>
> _______________________________________________
> Analytics mailing list
> Analytics@lists.wikimedia.org
> https://lists.wikimedia.org/mailman/listinfo/analytics
>
>


-- 
*Joseph Allemandou*
Data Engineer @ Wikimedia Foundation
IRC: joal
_______________________________________________
Analytics mailing list
Analytics@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/analytics

Reply via email to