Re: [Wiki-research-l] Fwd: [Wikitech-l] statistics about frequent section titles
Just also wanted to note that these paws-public URLs will break in the near-to-mid future :) On Mon, Mar 7, 2016 at 4:22 PM, Aaron Halfaker wrote: > Got some work done here. I'm using this as an opportunity to test out PAWS. > > See > http://paws-public.wmflabs.org/paws-public/EpochFail/projects/headings/extract_headings.ipynb > > It's still running right now, but I should have an output file that we can > download and/or load into MySQL soon. > > -Aaron > > > ___ > Wiki-research-l mailing list > Wiki-research-l@lists.wikimedia.org > https://lists.wikimedia.org/mailman/listinfo/wiki-research-l > -- Yuvi Panda T http://yuvi.in/blog ___ Wiki-research-l mailing list Wiki-research-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wiki-research-l
Re: [Wiki-research-l] [Analytics] Python client for the new pageview API
On Tue, Dec 15, 2015 at 8:10 AM, Finn Årup Nielsen wrote: > Just to confuse the issue further: > > You can already get views statistics from the ordinary API of an ordinary > MediaWiki instance (where the $wgDisableCounters is set to false). https://www.mediawiki.org/wiki/Manual:$wgDisableCounters Wanted to point out that this feature (counters) was removed from MW recently. ___ Wiki-research-l mailing list Wiki-research-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wiki-research-l
Re: [Wiki-research-l] Fwd: Traffic to the portal from Zero providers
more than Swedish is, and especially more than Faroese is) >>>>> while English is a widely used semi-official language in business, >>>>> government, and education (for example about half of university theses >>>>> are now written in English, and several major companies use it as their >>>>> official workplace language). >>>>> >>>>> I think it's possible to come up with something that better aligns with >>>>> readers' actual preferences, but it's not easy! >>>>> >>>>> -Mark >>>>> >>>>> -- >>>>> Mark J. Nelson >>>>> Anadrome Research >>>>> http://www.kmjn.org >>>>> >>>>> ___ >>>>> Wiki-research-l mailing list >>>>> Wiki-research-l@lists.wikimedia.org >>>>> https://lists.wikimedia.org/mailman/listinfo/wiki-research-l >>>> >>>> ___ >>>> Wiki-research-l mailing list >>>> Wiki-research-l@lists.wikimedia.org >>>> https://lists.wikimedia.org/mailman/listinfo/wiki-research-l >>> >>> >>> >>> -- >>> Oliver Keyes >>> Research Analyst >>> Wikimedia Foundation >>> >>> ___ >>> Wiki-research-l mailing list >>> Wiki-research-l@lists.wikimedia.org >>> https://lists.wikimedia.org/mailman/listinfo/wiki-research-l >> >> ___ >> Wiki-research-l mailing list >> Wiki-research-l@lists.wikimedia.org >> https://lists.wikimedia.org/mailman/listinfo/wiki-research-l > > ___ > Wiki-research-l mailing list > Wiki-research-l@lists.wikimedia.org > https://lists.wikimedia.org/mailman/listinfo/wiki-research-l -- Yuvi Panda T http://yuvi.in/blog ___ Wiki-research-l mailing list Wiki-research-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wiki-research-l
Re: [Wiki-research-l] rc stream
On Tue, Apr 7, 2015 at 12:51 PM, Ed Summers wrote: > That wasn't the most compelling argument for migrating. But thanks for the > response: Heh. The compelling arguments for migration to me are: 1. This gives you structured JSON, no need to futz around with IRC colors. This is a big one, I think 2. Easier to extend from the mediawiki side, so more events should be easier to add. 3. Far easier to consume rcstream than IRC Feeds (Websockets vs IRC) 4. Running our own IRC server is not the most fun thing in the world, and our ops team would like to not have to keep doing that forever. I find (1) and (3) most compelling - look at the client examples on https://wikitech.wikimedia.org/wiki/RCStream#Client, they're vastly simpler than similar examples for IRC -- Yuvi Panda T http://yuvi.in/blog ___ Wiki-research-l mailing list Wiki-research-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wiki-research-l
Re: [Wiki-research-l] rc stream
It is going to be at Some Point In The Far Future, and there will be a large amount of notice given, and possibly a less reliable IRC bridge running on labs. But no concrete plans at the moment. You should still migrate, however :) ___ Wiki-research-l mailing list Wiki-research-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wiki-research-l
Re: [Wiki-research-l] How to track all the diffs in real time?
On Sat, Dec 13, 2014 at 2:34 PM, Yuvi Panda wrote: > If a lot of people are doing this, then perhaps it makes sense to have > an 'augmented real time streaming' interface that is an exact replica > of the streaming interface but with diffs added. Or rather, if I were to build such a thing, would people be interested in using it? -- Yuvi Panda T http://yuvi.in/blog ___ Wiki-research-l mailing list Wiki-research-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wiki-research-l
Re: [Wiki-research-l] How to track all the diffs in real time?
If a lot of people are doing this, then perhaps it makes sense to have an 'augmented real time streaming' interface that is an exact replica of the streaming interface but with diffs added. ___ Wiki-research-l mailing list Wiki-research-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wiki-research-l
Re: [Wiki-research-l] Request for feedback on new data dump formats
Would incremental dumps, as described by brion long time ago (http://leuksman.com/log/2007/10/14/incremental-dumps/) be what you're looking for? On Fri, Apr 1, 2011 at 5:01 AM, Aaron Halfaker wrote: > If periodic update dumps are being considered, information that describes > changes to old data (page deletes, user renames, etc) would be very useful > to have along with new revisions. > > -Aaron > > On Mar 31, 2011 6:27 PM, "Luca de Alfaro" wrote: >> I think I would be very interested in 3, or even, in having every month a >> dump of that month's revisions. As I have built tools for the xml dumps, >> no >> change in format is good for me (and for WikiTrust). >> >> I would find incremental dumps (with occasional, yearly, full dumps) much >> easier to manage than full dumps. >> >> Luca >> >> On Thu, Mar 31, 2011 at 2:27 PM, Yuvi Panda wrote: >> >>> Hi, I'm a student planning on doing GSoC this year on mediawiki. >>> Specifically, I'd like to work on data dumps. >>> >>> I'm writing this to gauge what would be useful to the research >>> community. Several ideas thrown about include: >>> 1. JSON Dumps >>> 2. Sqlite Dumps >>> 3. Daily dumps of revisions in last 24 hours >>> 4. Dumps optimized for very fast import into various external storage >>> and smaller size (diffs) >>> 5. JSON/CSV for Special:Import and Special:Export >>> >>> Would any of these be useful? Or is there anything else that I'm >>> missing, that you would consider much more useful? >>> >>> Feedback would be invaluable :) >>> >>> Thanks :) >>> -- >>> Yuvi Panda T >>> http://yuvi.in/blog >>> >>> ___ >>> Wiki-research-l mailing list >>> Wiki-research-l@lists.wikimedia.org >>> https://lists.wikimedia.org/mailman/listinfo/wiki-research-l >>> > > ___ > Wiki-research-l mailing list > Wiki-research-l@lists.wikimedia.org > https://lists.wikimedia.org/mailman/listinfo/wiki-research-l > > -- Yuvi Panda T http://yuvi.in/blog ___ Wiki-research-l mailing list Wiki-research-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wiki-research-l
[Wiki-research-l] Request for feedback on new data dump formats
Hi, I'm a student planning on doing GSoC this year on mediawiki. Specifically, I'd like to work on data dumps. I'm writing this to gauge what would be useful to the research community. Several ideas thrown about include: 1. JSON Dumps 2. Sqlite Dumps 3. Daily dumps of revisions in last 24 hours 4. Dumps optimized for very fast import into various external storage and smaller size (diffs) 5. JSON/CSV for Special:Import and Special:Export Would any of these be useful? Or is there anything else that I'm missing, that you would consider much more useful? Feedback would be invaluable :) Thanks :) -- Yuvi Panda T http://yuvi.in/blog ___ Wiki-research-l mailing list Wiki-research-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wiki-research-l