Re: [Wikitech-l] [Engineering] Wikimedia REST content API is now available in beta
This is awesome. Congratulations to Gabriel and the rest of the team. I'll surely hope this will provide a stable platform for getting Wikimedia content on even more platforms. -- Hay On Wed, Mar 11, 2015 at 12:20 PM, Marc Ordinas i Llopis marc...@wikimedia.org wrote: Congratulations! After all the hard work that has gone into this, it's great to see it up and running. Besides the improvements it will allow in existing projects, I can't wait to see the new things it will enable. -- Marc On Tue, Mar 10, 2015 at 11:23 PM, Gabriel Wicke gwi...@wikimedia.org wrote: Hello all, I am happy to announce the beta release of the Wikimedia REST Content API at https://rest.wikimedia.org/ Each domain has its own API documentation, which is auto-generated from Swagger API specs. For example, here is the link for the English Wikipedia: https://rest.wikimedia.org/en.wikipedia.org/v1/?doc At present, this API provides convenient and low-latency access to article HTML, page metadata and content conversions between HTML and wikitext. After extensive testing we are confident that these endpoints are ready for production use, but have marked them as 'unstable' until we have also validated this with production users. You can start writing applications that depend on it now, if you aren't afraid of possible minor changes before transitioning to 'stable' status. For the definition of the terms ‘stable’ and ‘unstable’ see https://www.mediawiki.org/wiki/API_versioning . While general and not specific to VisualEditor, the selection of endpoints reflects this release's focus on speeding up VisualEditor. By storing private Parsoid round-trip information separately, we were able to reduce the HTML size by about 40%. This in turn reduces network transfer and processing times, which will make loading and saving with VisualEditor faster. We are also switching from a cache to actual storage, which will eliminate slow VisualEditor loads caused by cache misses. Other users of Parsoid HTML like Flow, HTML dumps, the OCG PDF renderer or Content translation will benefit similarly. But, we are not done yet. In the medium term, we plan to further reduce the HTML size by separating out all read-write metadata. This should allow us to use Parsoid HTML with its semantic markup https://www.mediawiki.org/wiki/Parsoid/MediaWiki_DOM_spec directly for both views and editing without increasing the HTML size over the current output. Combined with performance work in VisualEditor, this has the potential to make switching to visual editing instantaneous and free of any scrolling. We are also investigating a sub-page-level edit API for micro-contributions and very fast VisualEditor saves. HTML saves don't necessarily have to wait for the page to re-render from wikitext, which means that we can potentially make them faster than wikitext saves. For this to work we'll need to minimize network transfer and processing time on both client and server. More generally, this API is intended to be the beginning of a multi-purpose content API. Its implementation (RESTBase http://www.mediawiki.org/wiki/RESTBase) is driven by a declarative Swagger API specification, which helps to make it straightforward to extend the API with new entry points. The same API spec is also used to auto-generate the aforementioned sandbox environment, complete with handy try it buttons. So, please give it a try and let us know what you think! This API is currently unmetered; we recommend that users not perform more than 200 requests per second and may implement limitations if necessary. I also want to use this opportunity to thank all contributors who made this possible: - Marko Obrovac, Eric Evans, James Douglas and Hardik Juneja on the Services team worked hard to build RESTBase, and to make it as extensible and clean as it is now. - Filippo Giunchedi, Alex Kosiaris, Andrew Otto, Faidon Liambotis, Rob Halsell and Mark Bergsma helped to procure and set up the Cassandra storage cluster backing this API. - The Parsoid team with Subbu Sastry, Arlo Breault, C. Scott Ananian and Marc Ordinas i Llopis is solving the extremely difficult task of converting between wikitext and HTML, and built a new API that lets us retrieve and pass in metadata separately. - On the MediaWiki core team, Brad Jorsch quickly created a minimal authorization API that will let us support private wikis, and Aaron Schulz, Alex Monk and Ori Livneh built and extended the VirtualRestService that lets VisualEditor and MediaWiki in general easily access external services. We welcome your feedback here: https://www.mediawiki.org/wiki/Talk:RESTBase - and in Phabricator https://phabricator.wikimedia.org/maniphest/task/create/?projects=RESTBasetitle=Feedback: . Sincerely -- Gabriel Wicke Principal Software Engineer, Wikimedia Foundation ___ Engineering mailing list
Re: [Wikitech-l] Volunteers for Wikimania mobile app?
A full-featured multi-platform app would be pretty shortday now (especially if you want to get it past the app review process on iOS). A nice mobile website would be a better option IMHO. Let me know if you need any help on that. -- Hay On Sat, Jun 2, 2012 at 10:45 AM, Finne Boonen hen...@gmail.com wrote: http://m.fosdem.org/ Is the one that's used by Fosdem which works quite well, and works on most devices. henna On Sat, Jun 2, 2012 at 9:58 AM, Tomasz Finc tf...@wikimedia.org wrote: Certainly take a look at the official Wikipedia and Wiktionary apps on github to see how we've bult them. https://github.com/wikimedia Stop by in #wikimedia-mobile if you need any help. --tomasz On Fri, Jun 1, 2012 at 11:34 PM, Gregory Varnum gregory.var...@gmail.com wrote: Greetings, As someone involved with Wikimania programming - I'm wondering if there are volunteers interested in developing a Wikimania mobile app. It may not be feasible given the short timeline, but I've seen a number of conference apps coming out over the past week and figured it should at least be pondered. :) Essentially the idea would be to include schedule, local info, maps, etc. Things that can be pulled from the Wikimania 2012 wiki. Perhaps something PhoneGap based? Any thoughts or interest? -greg aka varnent ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l -- Maybe you knew early on that your track went from point A to B, but unlike you I wasn't given a map at birth! Alyssa, Chasing Amy ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l
[Wikitech-l] PhotoCommons Wordpress plugin
Hi everyone, some of you might have been aware of the Hackathon that was held in Amsterdam in honor of Wikipedia's 10th birthday in early january. At that event Krinkle and i hacked together a WordPress plugin that makes it possible to easily search and include Wikimedia Commons pictures in your blog posts. It's far from production ready, but we'd like to give you a sneak peek and ask for your input, thoughts and of course bugzilla tickets and patches ;) Installation instructions and a link to the download can be found here: http://www.mediawiki.org/wiki/PhotoCommons Let us know what you think, and feel free to tweet and blog about it, of course using the plugin to find a freely licensed image for your blog post ;) Thanks, -- Hay / Husky ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l
[Wikitech-l] Firesheep
Has anyone seen this? http://codebutler.com/firesheep A new Firefox plugin that makes it trivially easy to hijack cookies from a website that's using HTTP for login over an unencrypted wireless network. Wikipedia isn't in the standard installation as a site (lots of other sites, such as Facebook, Twitter, etc. are). We are using HTTP login by default, so i guess we're vulnerable as well (please say so if we're using some other kind of defensive mechanism i'm not aware of). Might it be a good idea to se HTTPS as the standard login? Gmail has been doing this since april this year. -- Hay ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l
Re: [Wikitech-l] Wikipedia iPhone app official page?
On Sat, Aug 29, 2009 at 3:07 PM, Dmitriy Sintsovques...@rambler.ru wrote: Some local coder told me that GIT is slower and consumes much more RAM on some operations than SVN. I can't confirm that, though, because I never used GIT and still rarely use SVN. But, be warned. Dmitriy I'm not a git expert, but afaik Git was designed from the ground up by Linus Torvalds to be *very fast*, even on large operations. It runs the Linux project, so i guess it can't be *that* slow :) Chad wrote: It's _ok_. The command line usage is pretty solid and I haven't encountered any issues there. The GUI interface sucks and is a complete waste of time. There aren't any real good interfaces for SVN on Mac and Linux too. Windows users are spoiled by having such a great app as Tortoise for free. I guess most Mac and Linux SVN users simply learn to use the command line, which isn't a bad thing at all. -- Hay ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l
Re: [Wikitech-l] Wikipedia iPhone app official page?
I think it would be a good start to spice up the page and add to link to the app description on iTunes too. The description has a line saying 'come and help us if you're good in HTML5/JS!' text, but no link to follow or where to get more information (maybe somewhere in the app itself too?) Also, why is this hosted externally on github on not in the main SVN repo? Right now it's pretty high on the top 25 on the app store, but has a very low rating because it's obviously still in early beta and lacks many of the features found in the already existing native apps. -- Hay On Thu, Aug 27, 2009 at 8:15 PM, Strainustrain...@gmail.com wrote: On Thu, Aug 27, 2009 at 9:06 PM, David Gerarddger...@gmail.com wrote: Is there an official page for the iPhone app, other than the iTunes store link? Some sort of about page, a link to the source, etc.? - d. ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l See http://www.mediawiki.org/wiki/Wikipedia_iPhone_app for some info... Strainu ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l
Re: [Wikitech-l] On templates and programming languages
Javascript might have gotten a bad name in the past because of 14-year olds who used it to display 'Welcome to my website!' alerts on their Geocities homepage, but it's really unfair. Javascript is a very flexible and dynamic language that can be written very elegantly. I urge everyone who still think Javascript is a toy language to read Douglas Crockford's excellent article: http://javascript.crockford.com/javascript.html -- Hay On Wed, Jul 1, 2009 at 10:35 AM, Gregory Maxwellgmaxw...@gmail.com wrote: On Wed, Jul 1, 2009 at 3:50 AM, William Allen Simpsonwilliam.allen.simp...@gmail.com wrote: Javascript, OMG don't go there. Don't be so quick to dismiss Javscript. If we were making a scorecard it would likely meet most of the checkboxes: * Available of reliable battle tested sandboxes (and probably the only option discussed other than x-in-JVM meeting this criteria) * Availability of fast execution engines * Widely known by the existing technical userbase (JS beats the other options hands down here) * Already used by many Mediawiki developers * Doesn't inflate the number of languages used in the operation of the site * Possibility of reuse between server-executed and client-executed (Only JS of the named options meets this criteria) * Can easily write clear and readable code * Modern high level language features (dynamic arrays, hash tables, etc) There may exist great reasons why another language is a better choice, but JS is far from the first thing that should be eliminated. Python is a fine language but it fails all the criteria I listed above except the last two. ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l
Re: [Wikitech-l] On templates and programming languages
I would opt for Javascript. PHP and Python are intended for large and complex applications and come with a huge standard library people probably expect to be available. Security concerns are a problem too, so a subset would probably be necessary So, in essence you get a crippled-down language that isn't really useful for templates. Making our own language, either by 'fixing' the template language or by inventing something new would only mean we introduce a new language that'll be specific to our own platform and nobody knows outside of Mediawiki developers. XSLT is not meant to be written or read by humans. It's a Turing-complete language stuffed into horrendous XML statements. Let's not go down that road. That leaves us to Lua and Javascript, which are both small and efficient languages meant to solve tasks like this. Remember, i'm talking about 'core' Javascript here, not with all DOM methods and stuff. If you strip that all out (take a look at the 1.5. core reference at Mozilla.com: https://developer.mozilla.org/en/Core_JavaScript_1.5_Reference) you get a pretty nice and simple language that isn't very large. Both would require a new parser and/or installed compilers on the server-side. Compared to the disadvantages of other options, that seems like a pretty small loss for a great win. Javascript is a widely understood and implemented language, with lots of efforts to get it even faster in modern browsers. Every Wikipedia user has a copy of it implemented in their browser and can start experimenting without the need for installing a compiler or a web server. Many people program in Javascript, so you have a huge potential number of people who could start programming Mediawiki templates. And it's already closely tied to the web, so you don't have to invent new ways of dealing with web-specific stuff. So, let's choose Javascript as our new template programming language. Regards, -- Hay On Tue, Jun 30, 2009 at 6:16 PM, Brion Vibberbr...@wikimedia.org wrote: As many folks have noted, our current templating system works ok for simple things, but doesn't scale well -- even moderately complex conditionals or text-munging will quickly turn your template source into what appears to be line noise. And we all thought Perl was bad! ;) There's been talk of Lua as an embedded templating language for a while, and there's even an extension implementation. One advantage of Lua over other languages is that its implementation is optimized for use as an embedded language, and it looks kind of pretty. An _inherent_ disadvantage is that it's a fairly rarely-used language, so still requires special learning on potential template programmers' part. An _implementation_ disadvantage is that it currently is dependent on an external Lua binary installation -- something that probably won't be present on third-party installs, meaning Lua templates couldn't be easily copied to non-Wikimedia wikis. There are perhaps three primary alternative contenders that don't involve making up our own scripting language (something I'd dearly like to avoid): * PHP Advantage: Lots of webbish people have some experience with PHP or can easily find references. Advantage: we're pretty much guaranteed to have a PHP interpreter available. :) Disadvantage: PHP is difficult to lock down for secure execution. * JavaScript Advantage: Even more folks have been exposed to JavaScript programming, including Wikipedia power-users. Disadvantage: Server-side interpreter not guaranteed to be present. Like Lua, would either restrict our portability or would require an interpreter reimplementation. :P * Python Advantage: A Python interpreter will be present on most web servers, though not necessarily all. (Windows-based servers especially.) Wash: Python is probably better known than Lua, but not as well as PHP or JS. Disadvantage: Like PHP, Python is difficult to lock down securely. Any thoughts? Does anybody happen to have a PHP implementation of a Lua or JavaScript interpreter? ;) -- brion ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l
Re: [Wikitech-l] Skin JS cleanup and jQuery
That's great news. I've been programming in JavaScript quite a lot the last few years and i think i would've probably gone insane if i hadn't discovered jQuery. Especially for complex and intricate HTML selections it's pretty amazing what you can do with jQuery. Also, the fact that animations are built into the core could mean it might get a little bit more simple to do visually cool things, which might make the UI more intuitive. -- Hay On Wed, Apr 15, 2009 at 11:45 PM, Sergey Chernyshev sergey.chernys...@gmail.com wrote: Guys, It's great to see that you're working in this direction - I'm thinking about working on this for a while, but didn't have a gut to undertake such ambitious project alone ;) Do you have a working instance of ScriptLoader anywhere so I can aim some performance tools at it? Thank you, Sergey -- Sergey Chernyshev http://www.sergeychernyshev.com/ On Wed, Apr 15, 2009 at 5:29 PM, Michael Dale md...@wikimedia.org wrote: These changes will probably result in some minor adjustments to existing skins. (I will try not to completely break compatibility cuz I know there are many custom skins out in the wild that would be no fun to stop working once they update medaiWiki) This consolidation of head includes _may_ result in _some_ un-updated skins referencing the same files twice which I think most browsers genneraly handle oky Enabling $wgEnableScriptLoader will not work so well with skins that have not been updated. Should have a patch soon. more about scriptLoader: http://www.mediawiki.org/wiki/ScriptLoader (We will most likely ship with $wgEnableScriptLoader off by default ) I am also very excited about jQuery making its way into core. Things like the add_media_wizard are much easier to put together with jQuery's nifty abstractions and msg system. More about add media wizard: http://metavid.org/blog/2009/03/27/add-media-wizard-and-firefogg-on-test-wikimediaorg/ peace, michael Brion Vibber wrote: Just a heads-up -- Michael Dale is working on some cleanup of how the various JavaScript bits are loaded by the skins to centralize some of the currently horridly spread-out code and make it easier to integrate in a centralized loader so we can serve more JS together in a single compressed request. Unless there's a strong objection I'd be very happy for this to also include loading up the jQuery core library as a standard component. The minified jQuery core is 19k gzipped, and can simplify other JS code significantly so we can likely chop down wikibits.js, mwsuggest.js, and the site-customized Monobook.js files by a large margin for a net savings. If you've done browser-side JavaScript development without jQuery and wanted to kill yourself, I highly recommend you try jQuery -- it's so nice. :) -- brion ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l
Re: [Wikitech-l] A proposal to de-table Wikipedia infoboxes
I don't know if making such an infobox that does not support IE6 and IE7 is a good idea. If you would take out all inline style elements and replace with them classes that are available in a general stylesheet it would already safe a lot of the cruft in the original code. -- Hay On Tue, Mar 3, 2009 at 1:30 PM, David Gerard dger...@gmail.com wrote: By Hakon Wium Lie of Opera: http://www.princexml.com/howcome/2009/wikipedia/infobox/ What is the likelihood of making as much as possible CSS? How to make infoboxes degrade gracefully for non-CSS browsers and IE users? - d. ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l
Re: [Wikitech-l] A proposal to de-table Wikipedia infoboxes
Maybe it would be an idea to have some kind of fancy extension that adds classes for templates only if they are available on articles that need them? Actually, i'm not sure if its in their job description, but something that would make tables and infoboxes a lot simpler should be a task of the usability team. -- Hay On Tue, Mar 3, 2009 at 3:29 PM, Aryeh Gregor simetrical+wikil...@gmail.com wrote: On Tue, Mar 3, 2009 at 7:42 AM, Hay (Husky) hus...@gmail.com wrote: I don't know if making such an infobox that does not support IE6 and IE7 is a good idea. It doesn't even support Firefox 2 . . . inline-block wasn't implemented in Gecko until 1.9 (Firefox 3). Also: It should be fairly easy to do so, as the HTML code is generated by templates. Has he *looked* at the templates? :) The major reason why inline style is used on Wikipedia is, of course, because ordinary editors don't have the ability to use stylesheets. And while admins do, they can only effectively add markup to *all* pages at once, regardless of whether they contain the exact infobox in question. An awful lot of the provided CSS is nation-box-specific, and so useless in 99.99% of Wikipedia's articles. (Literally: there are about 2.7 million articles, and I'm pretty sure there are less than 270 recognized nations.) But all that CSS would have to be served with all of them. - Show quoted text - ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l
Re: [Wikitech-l] Norwegian Websites Declare War on Internet Explorer 6
Unfortunately, IE6 (and IE7 as well) are problems that all web sites got to live with. IE6 is still used by about 34% of all web users (according to the latest statistics from thecounter.com), so banning those users or not paying attention to problems they might have with certain website elements is pretty bad, especially considering the fact that we have a mission to deliver our content to everyone, regardless of platform or browser. Testing on those platforms is probably not done enough, simply because most developers and editors are using something else than IE. -- Hay On Fri, Feb 20, 2009 at 11:50 AM, Thomas Dalton thomas.dal...@gmail.com wrote: 2009/2/20 Aryeh Gregor simetrical+wikil...@gmail.com: On Fri, Feb 20, 2009 at 5:19 AM, Thomas Dalton thomas.dal...@gmail.com wrote: There are different levels of support. We should certainly make sure things fail gracefully for IE6, but a new feature not working on IE6 shouldn't be a reason not to implement it for everyone else. (I believe that is pretty much the current policy already.) It depends on the type of feature. For instance, when implementing different icons for certain filetypes on external links in Monobook, I used the [$=foo] CSS selector, knowing it would fail in IE6, because it's not going to hurt anyone if it does. On the other hand, it would still be unacceptable at this point for a significant new feature not to work in IE6. It still has 20% of the market, by some figures I've seen. I can't see any significant new features causing a problem that wouldn't be dealt with by the fail gracefully condition. As long as adding the feature doesn't make things worse for IE6 users (so they can still read and edit the sites), then there isn't a big problem. Of course, if you can cater to IE6 easily, then there is no reason not to. ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l
Re: [Wikitech-l] Norwegian Websites Declare War on Internet Explorer 6
On Fri, Feb 20, 2009 at 3:50 PM, Aryeh Gregor simetrical+wikil...@gmail.com wrote: Wikimedia's goal is not to better humanity in some unspecified way. It's to disseminate free knowledge. Pestering users who probably can't fix the problem does nothing to advance that goal. If we're going to try moralizing our users, why don't we go ahead and nag our users to ditch IE entirely and switch to Firefox? IE7 is pretty bug-ridden too. Or hey, why not try getting them all to switch to Linux? Exactly. Also, in many ways upgrading to IE7 doesn't solve many problems at all (except for some very obvious CSS bugs) because it's still riddled with bugs and wrong implementations of W3C specs, especially on the Javascript implementation. We just have to live with the fact that web development means spending 30% of your time on writing hacks for IE6 and IE7 (and probably for IE8 too). -- Hay ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l
Re: [Wikitech-l] image backup and info on wikitech wiki
I think that any initiative of a new backup or download possibility of Commons files should be welcome. If the whole Commons collection is indeed only located on two servers in the whole world that gives me a little bit of a scary feeling... -- Hay On Tue, Dec 16, 2008 at 11:48 AM, Daniel Kinzler dan...@brightbyte.de wrote: grin schrieb: Even if it is, the entry on image backups is highly nonspecific and contains a big warning about being seriously out of date. I'm not a server admin and I don't know the exact details, but as far as I know, the information there is correct in so far as there are two storage servers for media files in tampa, one replicating the other. I think the second one is even located in a different data center, though in the same building. What exactly the status of automatic replication between these servers is, I do not know. What I do know is that the german chapter plans an off-size backup for media files in amsterdam. We will order the hardware we need for that this year still. I don't know how long it will take to set up a working replication process, but I hope that it will not be too long. Having an off-site backup seems a good idea, the next hurricane isn't that far away. We should also think about providing reasonable media bundles for download in some way. My current idea is to create bundles of all media files used or contained in a given category (maybe even including two or three levels of subcategories). Such bundles would have to be created by an off-line job, i suppose. And maybe requesting them should be limited to admins, to avoid flooding. -- daniel ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l