[Wikitech-l] New mediawiki.page_change.v1 event stream publicly available

2023-08-25 Thread Luke Bowmaker
The Wikimedia Data Engineering team is pleased to announce that a new event
stream, mediawiki.page_change.v1, is now publicly available at
stream.wikimedia.org (here
<https://stream.wikimedia.org/v2/ui/#/?streams=mediawiki.page_change.v1>).

The new event stream models page changes using a consolidated changelog
data model, whereas existing streams, like page-create and revision-create,
model each type of page change as a separate stream. With the current
model, you would have to consume multiple streams to understand how a
MediaWiki page changed. With the new stream, when a page is created,
edited, or deleted, an event captures the new state, as well as the state
prior to the change.

For more information on what fields you can expect to see in the events,
see the schema definition here
<https://schema.wikimedia.org/repositories//primary/jsonschema/mediawiki/page/change/latest.yaml>
.

Starting now, new streams will be suffixed with a major version and will
not use hyphens in stream names. For more information, see here
<https://wikitech.wikimedia.org/wiki/Event_Platform/Stream_Configuration#Stream_versioning>
.

Benefits:

   -

   Only one stream to consume (instead of having to consume page-create,
   page-delete, and revision-create streams to have a full picture of page
   changes)
   -

   Events are ordered for a given page_id (delete will not come before
   create)
   -

   Latest event has current and prior state

The existing event streams, such as page-create, revision-create,
page-delete, etc will continue to remain available, for now. We encourage
people to migrate to the new consolidated stream when they can as we plan
to mark the existing streams as deprecated within the next year. We will
send another communication about this when the plan is decided.

If you have any questions or issues please drop a Phabricator ticket here
<https://phabricator.wikimedia.org/project/board/6628/>

--

Luke Bowmaker

Data Engineering - Product Manager
___
Wikitech-l mailing list -- wikitech-l@lists.wikimedia.org
To unsubscribe send an email to wikitech-l-le...@lists.wikimedia.org
https://lists.wikimedia.org/postorius/lists/wikitech-l.lists.wikimedia.org/

[Wikitech-l] Privilege request

2023-01-08 Thread Luke Rigby
Hi,

I'm not entirely sure what to write here, but I believe "bringing up" Privilege 
requests on Gerrit on this mailing list is required, so I'm emailing to do just 
that, bring up my privilege request for the "Sudo" extension, which can be 
found here: https://phabricator.wikimedia.org/T318196

Thanks in advance,

Original Authority (OAuthority on Phab)
___
Wikitech-l mailing list -- wikitech-l@lists.wikimedia.org
To unsubscribe send an email to wikitech-l-le...@lists.wikimedia.org
https://lists.wikimedia.org/postorius/lists/wikitech-l.lists.wikimedia.org/

[Wikitech-l] (no subject)

2019-07-04 Thread Eduardo Luke

___
Wikitech-l mailing list
Wikitech-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikitech-l

Re: [Wikitech-l] [Wikimedia-l] Fwd: [Translators-l] Purodha passed away

2016-09-03 Thread Luke
Would be a nice idea, John. Sadly there isn't a way yet, to make a query 
for tokens (At least I didn't find one).


For people who want to leave a last message to him, we got a condolence 
list at dewiki: 
https://de.wikipedia.org/wiki/Benutzer:Purodha/Kondolenzliste


--
Luke

Am 03.09.2016 um 02:21 schrieb jay...@gmail.com:

On Sat, 3 Sep 2016 02:43 Amir E. Aharoni, <amir.ahar...@mail.huji.ac.il>
wrote:

This is very sad.

He was not only a prolific translator, but a prolific bug reporter, too.

May he rest in peace.

Indeed. Also developed patches.

His activity can be seen here:

https://phabricator.wikimedia.org/p/Purodha/

Can we find the open bugs he voted for/awarded tokens to?
And do the hardest one in his honour?

--
John
___
Wikitech-l mailing list
Wikitech-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikitech-l




___
Wikitech-l mailing list
Wikitech-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikitech-l

[Wikitech-l] question about wikimedia apache module mod_pagespeed

2013-09-11 Thread Luke Frank

Hello,

it's possible to add the Google Page Speed module on the wikipedia apache ?
For speed purpose.
I want to discuss with you of this possibility.
Thanks
Luke


___
Wikitech-l mailing list
Wikitech-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikitech-l

Re: [Wikitech-l] category intersection conversations

2013-05-09 Thread Luke Welling WMF
Without deliberately making it an even longer term plan, as I think it is a
great idea, another long goal solution to the same problem would be (as
Flow gets Wikipedians into the idea of tagging) that categories get largely
replaced by tags.  That way they lose much of their absoluteness and
therefore some of their controversy.

Categories are hard for Wikipedia because compromise is not possible.
 Consensus can be reached on a subtly different compromise version of the
wording of a sentence or paragraph, but there is no compromise on
categories.  A category either exists or does not. A page either goes in or
does not.

With tags, a biography could relatively uncontroversially  be tagged as
Novelist, Woman, Best Selling, American, Blonde Haired, Enjoys Spicy Food
even if nearly everybody agrees that half the tags while true are entirely
unimportant and not relevant to the subject's area of notability.  Whether
some tags like race and appearance should exist at all may still generate
debate, but if they are only ever available modifiers and not hard
categories their offense would be softened.

For some subjects, entirely uncontroversial tags could be extracted from
Wikidata.

It would be content shakeup and therefore perhaps politically difficult,
but it would take a lot of the technical challenge out of joins, even
permitting joins (automatically or manually) with tags translated into
equivalent versions in other languages.

All possible combinations of tag derived categories would then exist, and
it would just be a matter of debate as to whether there is a justification
to add a link from a page to Biography+Novelist+Enjoys Spicy Food or if
that is a meaningless category.  If reverted, the one person interested in
that exact category could still always visit it, it's just that other users
would not be directed to it unless they probe talk page debates.

Luke Welling


On Thu, May 9, 2013 at 12:38 PM, James Forrester
jforres...@wikimedia.orgwrote:

 [I worry we're talking about operational details, which should be a wider
 discussion, rather than a technology/feasibility conversation to which this
 list is more suited. Perhaps moving this on-wiki would be best?]

 On 9 May 2013 09:28, Brad Jorsch bjor...@wikimedia.org wrote:

  On Wed, May 8, 2013 at 10:47 PM, James Forrester
  jforres...@wikimedia.org wrote:
   * Pages are implicitly in the parent categories of their explicit
  categories
   * - Pages in Politicians from the Netherlands are in People from
 the
   Netherlands by profession (its first parent) and People from the
   Netherlands (its first parent's parent) and Politicians (its second
   parent) and People (its second parent's parent) and …
   * - Yes, this poses issues given the sometimes cyclic nature of
   categories' hierarchies, but this is relatively trivial to code around
 
  Category cycles are the least of it. The fact that the existing
  category hierarchy isn't based on any sensible-for-inference ontology
  is a bigger problem.
 
  Let's consider what would happen to one of my favorite examples on
 enwiki:
  * The article for Romania is in Black Sea countries. Ok.
  * And that category is in Black Sea, so Romania is in that too.
  Which is a little strange, but not too bad.
  * And Black Sea is in Seas of Russia and Landforms of Ukraine.
  Huh? Romania doesn't belong in either of those, despite that being
  equivalent to your example where pages in Politicians from the
  Netherlands also end up in People via Politicians.
 
  And it gets worse the further up you go. You would have Romania in
  Liquids a few more levels up.
 
  For this to work, each wiki would have to redo its category hierarchy
  as a real ontology based on is-a relationships, rather than the
  current is-somehow-related-to. Or we would have to introduce some
  magic word or something to tell MediaWiki that Politicians is-a
  People is a valid inference while Black Sea countries is-a Black
  Sea isn't.
 
  In other words, code-wise adding tags to an article is the same as
  categories with inference and querying. But trying to use the existing
  category setup as it exists on something like enwiki as tags for
  inference (or querying, to a lesser extent) seems like GIGO.
 

 Quite - the bit of my proposal where the categories would get created on
 Wikidata from scratch as a synthesis of the needs of the editing community.
 :-)

 Implicitly, these would have clear semantics about the correctitude of
 their usage governed by something analogous to how Wikidata's community are
 managing the roll-out of statements on the system. In terms of tools to
 prevent this becoming an issue, Wikidata's nature means we could easily
 make sure that the domain of a category would be limited (e.g. Fluids
 maps to substances, not instances of substances).



   * Readers can search, querying across categories regardless of whether
   they're implicit or explicit
   * - A search for the intersection of People from the Netherlands

Re: [Wikitech-l] Gerrit actively discourages discussion

2013-04-02 Thread Luke Welling WMF
For my own changes I have a workflow that's a little clunky, but works for
me.  I'll comment or mark done on every comment, so if I see the same
number of drafts as comments, I know I've addressed them all.

The workflow I have not found is a way to efficiently re-review other
people's changes that I've previously reviewed.  It takes a lot of time to
properly review a big change, and I don't know an easy way to say Show me
all my comments on previous patch sets so I can see if they have been
addressed.

Luke Welling


On Tue, Apr 2, 2013 at 9:10 AM, Chad innocentkil...@gmail.com wrote:

 On Tue, Apr 2, 2013 at 9:03 AM, MZMcBride z...@mzmcbride.com wrote:
  Would it be possible to have Gerrit import a JavaScript page from
  MediaWiki.org (e.g., MediaWiki:Gerrit.js)? This might allow dedicated
  users to override some of the default behavior (such as the block-level
  comment auto-collapsing) or add new functionality (such as a reply link
  similar to Bugzilla's).
 

 There's functionality to write custom Javascript and inject it within
 Gerrit.
 So yes, we could easily have our own JS that overrides default behavior
 here. I wouldn't feel entirely comfortable fetching it from MediaWiki.org
 on
 the fly--but we can host the code within Gerrit.

 -Chad

 ___
 Wikitech-l mailing list
 Wikitech-l@lists.wikimedia.org
 https://lists.wikimedia.org/mailman/listinfo/wikitech-l

___
Wikitech-l mailing list
Wikitech-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikitech-l

Re: [Wikitech-l] job queue

2013-03-25 Thread Luke Welling WMF
You probably want to run those jobs via maintenance/runJobs.php that way
you can have more control over when and where they get run.

You can control the type of jobs that get run, so you can effectively
create priority specific behavior or at least move expensive jobs onto
machines that won't slow the rest of your processing.

Luke Welling


On Mon, Mar 25, 2013 at 12:34 PM, dan entous d_ent...@yahoo.com wrote:

 context
 ---
 i’m working on a mediawiki extension,
 http://www.mediawiki.org/wiki/Extension:GWToolset, which has as one of
 its goals, the ability to upload media files to a wiki. the extension,
 among other tasks, will process an xml file that has a list of urls to
 media files and upload those media files to the wiki. our ideal goal is to
 have this extension run on http://commons.wikimedia.org/.


 job queue goals
 ---
 1. setup the processing of the xml file as a job queue job
 2. each media file upload to be setup as a job queue job


 current implementation
 --
 i have been able to achieve goal 2 and will sort out goal 1 shortly.


 issues/questions
 
 1. each of these jobs can take several seconds to complete. i have noticed
 in my local wiki that each of these jobs is picked up with each wiki visit
 and slows down the response of the wiki by however many seconds the job
 takes to run, a sleep in the job shows that if the job takes 15 seconds to
 run the wiki will be slowed down by that amount of time; i don't want this
 to happen on my local wiki or on commons.

a. are jobs on commons run as part of each wiki visit?
b. is there a cron job that takes care of the job queue on commons
 instead of using each wiki visit?
c. if not, is there a way to indicate that the job should only be run
 as a cron job and not with a wiki visit?

 2. if there's no solution to running the job with each wiki visit and
 slowing down the site, what other suggestions are there on processing the
 xml file and individual media file uploads?


 thanks in advance!
 dan
 ___
 Wikitech-l mailing list
 Wikitech-l@lists.wikimedia.org
 https://lists.wikimedia.org/mailman/listinfo/wikitech-l
___
Wikitech-l mailing list
Wikitech-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikitech-l

Re: [Wikitech-l] Gerrit Wars™: a strategy-guide

2013-03-21 Thread Luke Welling WMF
Ori's advice rings true with me.  It's something I need to get better at.

On the email titiel sidetrack, it should not create a 4th way.  Without
verifying them those all look like valid representations of the same data.
 MIME encoded word syntax only has two possible encodings, quoted printable
and base 64.  The one with the Q after UTF-8 should be the quoted printable
version.  The one with the B after UTF-8 has been encoded as base64 instead.

Of course it's possible somebody's client or MTA has munged them but the
point of encoded word is that even if you only speak RFC2822 and not
RFC2047(?) you can still transmit them correctly.

Luke Welling



On Thu, Mar 21, 2013 at 11:19 AM, Mark A. Hershberger m...@everybody.orgwrote:

 On 03/20/2013 10:43 AM, Jasper Wallace wrote:
  On Tue, 19 Mar 2013, MZMcBride wrote:
 
 
  P.S. mailman: there's a non-ASCII character in the subject line. Attack!
 
  Why? It's correctly encoded:

 Because the way the subject line is displayed 3 different ways on the
 archive page:


 http://lists.wikimedia.org/pipermail/wikitech-l/2013-March/thread.html#67742

 There we have:

   Gerrit =?utf-8?Q?Wars=E2=84=A2=3A_?=a strategy-guide
   Gerrit =?UTF-8?B?V2Fyc+KEog==?=: a strategy-guide
   Gerrit Wars™: a strategy-guide

 I wouldn't be surprised if my message creates a fourth way.

 --
 http://hexmode.com/

 [We are] immortal ... because [we have] a soul, a spirit capable of
compassion and sacrifice and endurance.
 -- William Faulker, Nobel Prize acceptance speech

 ___
 Wikitech-l mailing list
 Wikitech-l@lists.wikimedia.org
 https://lists.wikimedia.org/mailman/listinfo/wikitech-l

___
Wikitech-l mailing list
Wikitech-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikitech-l

Re: [Wikitech-l] Encoding subject encoding bike shed

2013-03-21 Thread Luke Welling WMF
Heh, if clients randomly change character sets than I guess there are a
very large number of possible values.

Given that RFC2047 came out in 1996 it's reasonable that people use
non-ascii characters in titles given that the means to do it in a
compatible way has been around for 17 years.

Luke


On Thu, Mar 21, 2013 at 12:04 PM, Mark A. Hershberger m...@everybody.orgwrote:

 On 03/21/2013 11:45 AM, Luke Welling WMF wrote:
  On the email title sidetrack, it should not create a 4th way.

 The pedant in me says there are at least two more ways -- different
 capitalization for UTF-8.  But your subject line shows another way.

 My client displays all of the subjects the same.

 Jasper,
 Mine:  =?utf-8?q?Gerrit_Wars=E2=84=A2=3A_a_strategy-guide?=
 Yours: =?windows-1252?q?Gerrit_Wars=99=3A_a_strategy-guide?=
 MZ's:  Gerrit =?UTF-8?B?V2Fyc+KEog==?=: a strategy-guide
 Ori:   Gerrit =?utf-8?Q?Wars=E2=84=A2=3A_?=a strategy-guide

 Maybe mailman doesn't understand when the encoding doesn't start at the
 first character since those are the ones that don't display correctly.

 --
 http://hexmode.com/

 [We are] immortal ... because [we have] a soul, a spirit capable of
compassion and sacrifice and endurance.
 -- William Faulker, Nobel Prize acceptance speech

 ___
 Wikitech-l mailing list
 Wikitech-l@lists.wikimedia.org
 https://lists.wikimedia.org/mailman/listinfo/wikitech-l
___
Wikitech-l mailing list
Wikitech-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikitech-l

Re: [Wikitech-l] Default policy for third-party cookies in Firefox

2013-03-19 Thread Luke Welling WMF
If you want to play cat and mouse, a good reference for things that work is
http://samy.pl/evercookie/

It's mostly targeted at a single domain stopping users from deleting
cookies, but some of the same things should break cross domain security
too.

I'm not sure that end of web ethics is where we want to go in general but
sleazy is a spectrum and depends on intent so there may be useful
inspiration in it.

Luke


On Tue, Mar 19, 2013 at 12:56 PM, Greg Grossmeier g...@wikimedia.orgwrote:

 quote name=Seb35 date=2013-03-19 time=14:38:40 +0100
  Hello,
 
  According to [1] and [2], Firefox 22 (release June 25, 2013) will
  change the default third-party cookie policy: a third-party cookie
  will be authorized only if there is already a cookie set on the
  third-party website.

 These two bugs are related to this:
 https://bugzilla.wikimedia.org/show_bug.cgi?id=45578

 https://bugzilla.wikimedia.org/show_bug.cgi?id=45452


 --
 | Greg GrossmeierGPG: B2FA 27B1 F7EB D327 6B8E |
 | identi.ca: @gregA18D 1138 8E47 FAC8 1C7D |

 ___
 Wikitech-l mailing list
 Wikitech-l@lists.wikimedia.org
 https://lists.wikimedia.org/mailman/listinfo/wikitech-l

___
Wikitech-l mailing list
Wikitech-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikitech-l

Re: [Wikitech-l] Pronunciation recording tool wanted

2013-03-13 Thread Luke Welling WMF
It would be a good application for mobile too.

In browser would be reasonably easy with Flash, and can be done with
JavaScript in modern browsers but not yet in a consistent way.  There is a
W3 spec but using a library like
https://github.com/jussi-kalliokoski/sink.js/ would be easier than writing
per browser versions to take into account current real world variation.

A mobile app, or a few native apps for dominant platforms presumably expose
a cleaner interface to what is a core device on that hardware, rather than
an optional, variable peripheral on computers.

Luke


On Wed, Mar 13, 2013 at 4:03 AM, Matthew Flaschen
mflasc...@wikimedia.orgwrote:

 On 03/13/2013 03:17 AM, Nikola Smolenski wrote:
  Why CC0 (public domain)?  Your example
  (http://commons.wikimedia.org/wiki/File:Fr-go%C3%BBter.ogg) is CC-BY,
  which is not public domain and requires attribution (which I think all
  Wikimedia projects do for text).  I'd say CC-BY-SA or CC-BY would be a
  better default.
 
  I am not sure about copyrightability of a pronunciation of a single word.

 Neither am I, but if it's licensed under one of those and a court finds
 it's not copyrightable, so be it.  It still seems reasonable to use an
 attribution license.

 Matt Flaschen

 ___
 Wikitech-l mailing list
 Wikitech-l@lists.wikimedia.org
 https://lists.wikimedia.org/mailman/listinfo/wikitech-l

___
Wikitech-l mailing list
Wikitech-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikitech-l

Re: [Wikitech-l] Query performance - run code faster, merge code faster :-)

2013-03-07 Thread Luke Welling WMF
The advice on
https://wikitech.wikimedia.org/wiki/Query_profiling_for_features_developers
sounds
good.

Is there more detail somewhere on how to do this part Test your query
against production slaves prior to full deployment?

Luke


On Wed, Mar 6, 2013 at 8:14 PM, Matthew Flaschen mflasc...@wikimedia.orgwrote:

 On 03/06/2013 04:36 PM, Sumana Harihareswara wrote:
  If you want your code merged, you need to keep your database queries
  efficient.  How can you tell if a query is inefficient? How do you write
  efficient queries, and avoid inefficient ones?  We have some resources
  around:
 
  Roan Kattouw's
 
 https://www.mediawiki.org/wiki/Manual:Database_layout/MySQL_Optimization/Tutorial
  -- slides at
 
 https://commons.wikimedia.org/wiki/File:MediaWikiPerformanceProfiling.pdf
 
  Asher Feldman's
  https://www.mediawiki.org/wiki/File:MediaWiki_Performance_Profiling.ogv
  -- slides at
 https://www.mediawiki.org/wiki/File:SQL_indexing_Tutorial.pdf

 And
 https://wikitech.wikimedia.org/wiki/Query_profiling_for_features_developers

 Matt Flaschen

 ___
 Wikitech-l mailing list
 Wikitech-l@lists.wikimedia.org
 https://lists.wikimedia.org/mailman/listinfo/wikitech-l

___
Wikitech-l mailing list
Wikitech-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikitech-l

Re: [Wikitech-l] Seemingly proprietary Javascript

2013-03-05 Thread Luke Welling WMF
I don't see the purpose of adding a licence string back on to JavaScript
post-minification.  Any recipient wanting to create a derivative work or
redistribute those files is going to go back to the much more readable
source files.

It would be good form to add licence information to all the JS files in the
same way we do for all the PHP files. Many or all of them are missing that
now.  Given they have a consistent licence, making that clear in each file
is just grunt work.

I don't see the need for that to survive minificaiton though. If somebody
wants to auto verify licence status with software, they can run it on the
original JS source before it get's minified. As others have implied
regardless of whether you think satisfying the FSF is important, satisfying
an automated tool is a concern that can be delegated to the tool owner.

The licence status of on wiki user JavaScript is a separate issue, and
possibly much more complicated.  CC-BY-SA-3.0 is not an ideal licence for
software, and it seems likely that there will be code pasted into some
user JavaScript pages that is licensed under an incompatible licence.

Luke Welling


On Tue, Mar 5, 2013 at 10:57 AM, Mark Holmquist mtrac...@member.fsf.orgwrote:

 On Tue, Mar 05, 2013 at 12:56:23PM +0100, Alexander Berntsen wrote:
  GNU LibreJS blocks several Javascript sources around Wikipedia. I was
  sent to this list by Kirk Billund. My issue as well as Kirk's replies
  follows. I hope you are okay to read it in this form.

 https://bugzilla.wikimedia.org/show_bug.cgi?id=36866

 We have this issue reported, it's on our radar, and I, at least, intend to
 fix it in the future.

 The user JavaScript and CSS might be an issue. I'm not sure how to handle
 that. I guess we could indicate in the license headers that some parts of
 the code are under the CC-BY-SA license, or whatever is set to the default
 license for the wiki. That should be possible, if not trivial.

 The minification process, however, does *not* cause a problem. We can
 simply add the comments to the file(s) after the minification. It does mean
 we'll need to include, potentially, multiple license headers in one HTTP
 response, but that shouldn't cause much issue. Alternatively we could use a
 mixed license header, and link to the texts of multiple licenses, or link
 to multiple files' source code.

 See the linked bug (above) for more discussion of the technical problems
 presented, and a few proposed suggestions. It looks like the best way to do
 it would be the bang comment syntax, suggested by Timo (Krinkle), which
 would allow each script to be tagged on its own, and that way each script
 authour would be responsible for their own licensing.

 I hope that helps, and that the bug discussion is a little more kind than
 wikitech has seemed :)

 --
 Mark Holmquist
 Software Engineer
 Wikimedia Foundation
 mtrac...@member.fsf.org
 https://wikimediafoundation.org/wiki/User:MHolmquist


 ___
 Wikitech-l mailing list
 Wikitech-l@lists.wikimedia.org
 https://lists.wikimedia.org/mailman/listinfo/wikitech-l

___
Wikitech-l mailing list
Wikitech-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikitech-l

Re: [Wikitech-l] Seemingly proprietary Javascript

2013-03-05 Thread Luke Welling WMF
Yes.  There seems little value in unqualified people debating if it is
legally required.

The mainstream FOSS licences all predate minification and seem to have been
written with compiled languages in mind, not interpreted languages.  Most
have language that requires the licence in the source version, but not the
binary version.  Deciding whether minified JavaScript is technically or in
spirit a binary form seems like something best left to experts.

My conscience would certainly be clear if we only had a licence in our
source distribution.

Luke


On Tue, Mar 5, 2013 at 12:25 PM, Marc A. Pelletier m...@uberbox.org wrote:

 On 03/05/2013 12:22 PM, Tyler Romeo wrote:

 it is nonetheless *legally
 required* to do so, regardless of the technical aspect of it


 I think that determination needs to be made by Counsel, not on a guess.
  I've quite some knowledge of copyright myself, and I know enough that the
 matter is subtle enough that this declaration is, at best, an
 oversimplification that cannot possibly reflect reality.

 -- Marc



 __**_
 Wikitech-l mailing list
 Wikitech-l@lists.wikimedia.org
 https://lists.wikimedia.org/**mailman/listinfo/wikitech-lhttps://lists.wikimedia.org/mailman/listinfo/wikitech-l

___
Wikitech-l mailing list
Wikitech-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikitech-l

Re: [Wikitech-l] Seemingly proprietary Javascript

2013-03-05 Thread Luke Welling WMF
We should discuss them separately, but this core mediawiki JS is GPL2
https://github.com/wikimedia/mediawiki-core/tree/master/resources

This JS which was mentioned in the forwarded email that started this
discussion is available via a wiki page so is probably under a CC-BY-SA-3.0
as it is submitted, edited and accessed like content.
http://en.wikipedia.org/wiki/Wikipedia:WikiProject_User_scripts/Scripts#Scripts

Luke


On Tue, Mar 5, 2013 at 2:55 PM, Matthew Flaschen mflasc...@wikimedia.orgwrote:

 On 03/05/2013 09:47 AM, Isarra Yos wrote:
  The licensing information is on the page itself, of which the minified
  js winds up a part. For every file or other object that makes up the
  page to all contain the licensing information would be pretty unusual.
 
  It's like taking a file out of a page and then complaining that it has
  no licensing information when said information was in the page text
  right under it.

 What licensing information are you referring to?

 Of course, the code is not under the content license (content license
 being CC-BY-SA currently for Wikimedia).

 Matt Flaschen

 ___
 Wikitech-l mailing list
 Wikitech-l@lists.wikimedia.org
 https://lists.wikimedia.org/mailman/listinfo/wikitech-l

___
Wikitech-l mailing list
Wikitech-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikitech-l

Re: [Wikitech-l] Better non-MySQL db support

2013-02-26 Thread Luke Welling WMF
Do we have an official position on cross database compatibility?

Some of the MediaWiki SQL is in separate files and can be easily directed
at a specific database engine.  A lot of it though is scattered as
fragments though other code and is going to be run on any engine we connect
to.

Specifically, do we use MySQL specific syntax that is more efficient (but
breaks elsewhere) or do we attempt to write lowest common denominator SQL
that will run more places, but not run as efficiently on our primary target?

The only SQL coding standard doc I have seen treats Database and MySQL as
synonyms: http://www.mediawiki.org/wiki/Manual:Coding_conventions/Database

Luke Welling


On Tue, Feb 26, 2013 at 5:37 AM, Dmitriy Sintsov ques...@rambler.ru wrote:



  26 Февраль 2013 г. 14:27:06 пользователь Nikola Smolenski (
 smole...@eunet.rs) написал:


 On 26/02/13 04:18, Matthew Flaschen wrote:
  Sure, for starters. :) Bear in mind, if we want to keep support for all
  these dbs, every change to the database schema has to (at some point)
  result in a change to separate SQL files for each DB (MySQL and SQLite
  use the same ones).For instance, there is a separate active
  oracle/tables.sql.
 I am wondering if it would make sense to give up on SQL, make universal
 table creation functions in PHP, the same way there are for other queries,
 and use that. Has anyone tried this before, is there other software that
 works like this?


 http://stackoverflow.com/**questions/108699/good-php-orm-**libraryhttp://stackoverflow.com/questions/108699/good-php-orm-library

 By the way, MySQL 5.6 is out and it supports fulltext search indexation
 for InnoDB tables. They also promise better peformance on large hardware.
 Still cannot find 12.04 ppa for 5.6.10, though and do not want to go
 troubles installing from source (although I installed MySQL from source
 some years ago)..
 Why going another database engines besides MySQL / MariaDB?
 Dmitriy


 __**_
 Wikitech-l mailing list
 Wikitech-l@lists.wikimedia.org
 https://lists.wikimedia.org/**mailman/listinfo/wikitech-lhttps://lists.wikimedia.org/mailman/listinfo/wikitech-l

___
Wikitech-l mailing list
Wikitech-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikitech-l

Re: [Wikitech-l] DevOps/Continuous Deployment discussion?

2013-02-20 Thread Luke Welling WMF
I am strongly of the opinion that within broad ranges deployment frequency
does not matter.  It really does not matter if you deploy twice an hour or
every second day.

But, having the machinery to make it so that you could deploy twice an hour
if you wanted to is all kinds of valuable.

Putting time into building:
 * Continuous integration with build-on-commit
 * Tests with good coverage
 * A staging environment that reflects production
 * Managed configuration
 * Scripted deployment to a large number of machines
pays dividends in uptime, ops sanity and developer productivity even if you
only use that machinery every few days.

We have some of that, but heading further down that road would be a good
thing even if we chose to keep organized periodic deployments.

Luke Welling


On Wed, Feb 20, 2013 at 2:04 PM, Juliusz Gonera jgon...@wikimedia.orgwrote:

 Sorry for digging up an old thread, but today I also started wondering if
 there's a way of making our deployments simpler and faster.

 I'm not a big fan of special highly orchestrated events when the whole
 team gathers and waits and then looks for regressions after deploying
 dozens of commits at the same time. I've been reading a bit and it's a fact
 that some projects do Continuous Deployment and it works for them:
 http://radar.oreilly.com/2009/**03/continuous-deployment-5-**eas.htmlhttp://radar.oreilly.com/2009/03/continuous-deployment-5-eas.html
 http://timothyfitz.com/2009/**02/10/continuous-deployment-**
 at-imvu-doing-the-impossible-**fifty-times-a-day/http://timothyfitz.com/2009/02/10/continuous-deployment-at-imvu-doing-the-impossible-fifty-times-a-day/
 http://www.slideshare.net/**mikebrittain/mbrittain-**continuous-**
 deploymentalm3publichttp://www.slideshare.net/mikebrittain/mbrittain-continuous-deploymentalm3public

 Is there any interest at WMF in taking this path at some point?

 --
 Juliusz



 On 12/26/2012 09:31 AM, Chris McMahon wrote:

 Hi,

 A number of people I know of have ideas and aspirations pertaining to a
 DevOps-style deployment process, a.k.a Continuous Deployment.  In recent
 times a number of pieces of such a system have become functional:  Zuul,
 Jenkins enhancements for tests, automated acceptance tests, etc.

 But looking at mediawiki.org I don't see any sort of central discussion
 of
 overall approach/design/process for DevOps/Continuous Deployment.

 Is it time to start such a discussion?  Or is this premature?

 -Chris
 __**_
 Wikitech-l mailing list
 Wikitech-l@lists.wikimedia.org
 https://lists.wikimedia.org/**mailman/listinfo/wikitech-lhttps://lists.wikimedia.org/mailman/listinfo/wikitech-l



 __**_
 Wikitech-l mailing list
 Wikitech-l@lists.wikimedia.org
 https://lists.wikimedia.org/**mailman/listinfo/wikitech-lhttps://lists.wikimedia.org/mailman/listinfo/wikitech-l

___
Wikitech-l mailing list
Wikitech-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikitech-l

Re: [Wikitech-l] MediaWiki API at Codecademy?

2013-02-08 Thread Luke Welling WMF
Do you want help?

I don't know much about the API at the moment, but it is my Level-Up
assignment this quarter. Documenting things is a good way to learn them.

I have written a lot of courseware in the past.

Luke Welling


On Thu, Feb 7, 2013 at 6:20 PM, Yuri Astrakhan yuriastrak...@gmail.comwrote:

 I will be happy to part-take in this, as I do have some experience with the
 API :)

 I am heading the API v2 project, and this would be a natural extension of
 that.

 * RFC http://www.mediawiki.org/wiki/Requests_for_comment/API_Future
 * Action/Parameter
 Cleanup
 http://www.mediawiki.org/wiki/Requests_for_comment/API_Future/Naming_Cleanup
 


 On Thu, Feb 7, 2013 at 6:12 PM, Quim Gil q...@wikimedia.org wrote:

  Long story short: the MediaWiki API could be included at
 
  http://www.codecademy.com/**tracks/apis
 http://www.codecademy.com/tracks/apis
 
  if someone wants to do the work. Codecademy is happy to have us there.
 
  I think it is a good idea, in need of someone willing to drive this:
 
  - It is a good excuse to improve our API documentation at mediawiki.org.
  - Codecademy is a good place to reach to more developers.
  - Maybe it is a good chance for someone to take this as a paid job?
 
  The Wikimedia movement wants to spread the 1,3T of content we have, and
  get more free content from as many channels as possible. Our API plays a
  big role on this.
 
  Therefore, I *personally* believe that a project to update the API
  documentation at mediawiki.org and have corresponding exercises at
  Codecademy has a chance to receive a grant if the proposal and the
  candidate(s) are solid.
 
  Interested? Let me help you.
 
  --
  Quim Gil
  Technical Contributor Coordinator @ Wikimedia Foundation
  http://www.mediawiki.org/wiki/**User:Qgil
 http://www.mediawiki.org/wiki/User:Qgil
 
  __**_
  Wikitech-l mailing list
  Wikitech-l@lists.wikimedia.org
  https://lists.wikimedia.org/**mailman/listinfo/wikitech-l
 https://lists.wikimedia.org/mailman/listinfo/wikitech-l
 ___
 Wikitech-l mailing list
 Wikitech-l@lists.wikimedia.org
 https://lists.wikimedia.org/mailman/listinfo/wikitech-l

___
Wikitech-l mailing list
Wikitech-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikitech-l

Re: [Wikitech-l] Why are we still using captchas on WMF sites?

2013-01-23 Thread Luke Welling WMF
I don't know if we are talking at cross purposes, or if I missed it, but
this paper:
http://elie.im/publication/text-based-captcha-strengths-and-weaknesses
does not try to answer my question.

What I want to know is *How many humans get turned away from editing
Wikipedia by a difficult captcha?*

The same authors have:
http://elie.im/publication/how-good-are-humans-at-solving-captchas-a-large-scale-evaluation
which is closer to what I want to know.  They show humans solving different
text based captures with an accuracy rate of 70% to 98%. Unfortunately,
Wikipedia was not one of the captcha schemes they used in that study, and
they don't attempt to measure how many people try again if they fail.

If 2% of people fail on the first try but 90% of the fails reattempt and
only 1% fail a second time that's an inconvenience, but probably worth it
if it reduces the inconvenience of spam.

If 30% of people fail on the first try and 90% of them give up and never
try to edit again, that's a disaster.

Luke




On Wed, Jan 23, 2013 at 12:24 AM, Federico Leva (Nemo)
nemow...@gmail.comwrote:

 Luke, we do not know how many humans are being turned away by the
 difficulty: actually we sort of do, that paper tells this as well. It's
 where their study came from, and gives recommendations on what captcha
 techniques are best for balancing efficacy with difficulty for humans. We
 don't seem to be following any (except waving, which, they say, shouldn't
 be used alone).
 Then, I'm not qualified to say if their recommendations are the best and
 I've not searched other studies, but it's not correct to say that we start
 from zero or that we have to study by ourselves (an unreasonable
 requirement that implies we'll never change anything until we'll be forced
 to make our wikis read-only due to spam, as many MediaWiki users before us).


 Nemo

 __**_
 Wikitech-l mailing list
 Wikitech-l@lists.wikimedia.org
 https://lists.wikimedia.org/**mailman/listinfo/wikitech-lhttps://lists.wikimedia.org/mailman/listinfo/wikitech-l

___
Wikitech-l mailing list
Wikitech-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikitech-l


Re: [Wikitech-l] Why are we still using captchas on WMF sites?

2013-01-22 Thread Luke Welling WMF
Even ignoring openness and privacy, exactly the same problems are present
with reCAPTCHA as with Fancy Captcha.  It's often very hard or impossible
for humans to read, and is a big enough target to have been broken by
various people.

I don't know if it's constructive to brainstorm solutions to a problem
before we measure the extent of the problem, but a viable compromise is
very easy captchas.  Spammers vary a great deal in sophistication but if we
figure that any sophisticated enough to do any OCR are capable of finding
and downloading existing public exploits of ours, then a block capital
impact font captcha is equally easy for them, equally difficult for
unsophisticated spammers and much easier for sighted humans.

Luke


On Tue, Jan 22, 2013 at 9:45 AM, Arthur Richards aricha...@wikimedia.orgwrote:

 On Tue, Jan 22, 2013 at 10:37 AM, vita...@yourcmc.ru wrote:

 
  Maybe you'll just use recaptcha instead of fancycaptcha?
 
 
 /me gets popcorn to watch recaptcha flame war

 There has been discussion on this list in the past about the use of
 recaptcha, but it has generally ended in a down-vote because reCaptcha is
 not open source (even though it supports free culture) nor is it something
 we can host on our own servers.

 --
 Arthur Richards
 Software Engineer, Mobile
 [[User:Awjrichards]]
 IRC: awjr
 +1-415-839-6885 x6687
 ___
 Wikitech-l mailing list
 Wikitech-l@lists.wikimedia.org
 https://lists.wikimedia.org/mailman/listinfo/wikitech-l

___
Wikitech-l mailing list
Wikitech-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikitech-l


Re: [Wikitech-l] Why are we still using captchas on WMF sites?

2013-01-22 Thread Luke Welling WMF
That was not the end of the problem I was referring to. We know our
specific captcha is broken at turning away machines. As far as I am aware
we do not know how many humans are being turned away by the difficulty of
it.  It's a safe bet that it is non-zero given the manual account requests
we get, but given that we have people to do those kinds of experiments it
would make sense to get a number from them before making any drastic
decisions based on a reasonable gut feeling.  I don't think anybody claims
to have a perfect solution to the spam vs usability balancing act, so it's
possible we'll try (and measure) a few approaches.

Luke


On Tue, Jan 22, 2013 at 10:16 AM, Federico Leva (Nemo)
nemow...@gmail.comwrote:

 Luke, sorry for reiterating, but «brainstorm solutions to a problem
 before we measure the extent of the problem» is wrong: it's already been
 measured by others, see the other posts...

 Nemo


 __**_
 Wikitech-l mailing list
 Wikitech-l@lists.wikimedia.org
 https://lists.wikimedia.org/**mailman/listinfo/wikitech-lhttps://lists.wikimedia.org/mailman/listinfo/wikitech-l

___
Wikitech-l mailing list
Wikitech-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikitech-l


Re: [Wikitech-l] The ultimate bikeshed: typos in commit messages

2013-01-16 Thread Luke Welling WMF
I don't mind getting dinged for typos.  If I'm being sloppy it's fair to
point it out.

However, I think the social contract should be that after I fix the typos
you requested then you owe me a real code review where you look at the
merits of the code.  Code review is an awesomely useful but time consuming
thing to provide.  Patrolling for typos is not.

Put it this way, if I concede and agree that the bikeshed can be green, the
people who spent three hours arguing for green should feel obligated to
turn up to the working bee to help with the painting.

Luke Welling


On Wed, Jan 16, 2013 at 10:40 AM, Bryan Tong Minh
bryan.tongm...@gmail.comwrote:

 On Wed, Jan 16, 2013 at 7:26 PM, Jon Robson jdlrob...@gmail.com wrote:

  This is a ridiculous conversation and I can't believe it now spans +20
  messages.
 
 Apparently you don't care, but other people do care. Please do not
 disregard other people's opinions because you believe yours is correct.
 To keep it in the style of this thread; attitudes like these do demotivate
 _me_.



 Also, I fully agree with Maarten's last comment.


 Bryan
 ___
 Wikitech-l mailing list
 Wikitech-l@lists.wikimedia.org
 https://lists.wikimedia.org/mailman/listinfo/wikitech-l

___
Wikitech-l mailing list
Wikitech-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikitech-l


Re: [Wikitech-l] Problems uploading word 2007 .doc files

2013-01-08 Thread Luke Welling WMF
Is this bug the same issue?  It looks like somebody put up a partial fix

https://bugzilla.wikimedia.org/show_bug.cgi?id=38432

- Luke Welling


On Mon, Jan 7, 2013 at 6:30 PM, Aran Dunkley a...@organicdesign.co.nzwrote:

 The file was a .doc, but I've tried changing it to docx and get the same
 result. Some other .doc and .docx files that are word 2007 upload no
 problem. But I don't see how it can complain when I have the MimeType
 verification disabled - completely disabling the verification would be
 no problem since only specific users can upload.

 p.s. this is a MW 1.19.2 on Ubuntu Server 11.10 with PHP 5.3.6

 On 07/01/13 21:21, Matma Rex wrote:
  On Mon, 07 Jan 2013 23:47:58 +0100, Aran Dunkley
  a...@organicdesign.co.nz wrote:
 
  Hello, can someone please help me with this .doc upload problem? I've
  tried everything and even setting  $wgVerifyMimeType to false fails to
  solve it. No matter what I do I keep getting the following error when I
  upload *some* word 2007 .doc files:
 
  The file is a corrupt or otherwise unreadable ZIP file. It cannot be
  properly checked for security.
 
  I don't know how that check can even be happening with $wgVerifyMimeType
  disabled, but still the error occurs?!
 
  Word 2007 uses a .docx format as far as I know, not .doc. Which one
  were you using in your configuration?
 
  Also, .docx files are essentially ZIP files with magic data inside.
 
  ___
  Wikitech-l mailing list
  Wikitech-l@lists.wikimedia.org
  https://lists.wikimedia.org/mailman/listinfo/wikitech-l


 ___
 Wikitech-l mailing list
 Wikitech-l@lists.wikimedia.org
 https://lists.wikimedia.org/mailman/listinfo/wikitech-l

___
Wikitech-l mailing list
Wikitech-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikitech-l


Re: [Wikitech-l] Jenkins: whitespace test!

2012-12-21 Thread Luke Welling WMF
It should be an ignorable warning.

There are occasionally situations where trailing white space is necessary.
MySQL for instance can be funny about combining multi line strings into one
query, although I don't think it presents when called via PHP.

Luke


On Fri, Dec 21, 2012 at 9:59 AM, Chad innocentkil...@gmail.com wrote:

 On Fri, Dec 21, 2012 at 9:43 AM, Antoine Musso hashar+...@free.fr wrote:
  Hello,
 
  For a long time now we have been reviewing changes that contains
  unwanted trailing whitespaces. To make sure the patch submitter is
  warned about it, I have enabled trailing whitespace check on
  mediawiki/core.git
 
  The bug report is:
   https://bugzilla.wikimedia.org/42628
 
  I have only enabled such check on the mediawiki/core.git repository for
  now.  If that went well I will look up at enabling it on other repos.
 

 Is this a warning, or a failure? I really don't think we should *fail*
 patches for something that's almost always harmless. The only
 reason I ask is because you say warning here but failure on the
 bug.

 -Chad

 ___
 Wikitech-l mailing list
 Wikitech-l@lists.wikimedia.org
 https://lists.wikimedia.org/mailman/listinfo/wikitech-l

___
Wikitech-l mailing list
Wikitech-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikitech-l


Re: [Wikitech-l] Jenkins: whitespace test!

2012-12-21 Thread Luke Welling WMF
And PS, as I've been dinged for this a few times this is now in my .vimrc
to highlight trailing whitespace in vim.

highlight ExtraWhitespace ctermbg=red guibg=red
match ExtraWhitespace /\s\+$/
autocmd BufWInEnter * match ExtraWhitespace /\s\+$/
autocmd InsertEnter * match ExtraWhitespace /\s\+\%#\@!$/
autocmd InsertLeave * match ExtraWhitespace /\s\+$/
autocmd BufWInLeave * call clearmatches()



On Fri, Dec 21, 2012 at 10:20 AM, Luke Welling WMF
lwell...@wikimedia.orgwrote:

 It should be an ignorable warning.

 There are occasionally situations where trailing white space is
 necessary.  MySQL for instance can be funny about combining multi line
 strings into one query, although I don't think it presents when called via
 PHP.

 Luke



 On Fri, Dec 21, 2012 at 9:59 AM, Chad innocentkil...@gmail.com wrote:

 On Fri, Dec 21, 2012 at 9:43 AM, Antoine Musso hashar+...@free.fr
 wrote:
  Hello,
 
  For a long time now we have been reviewing changes that contains
  unwanted trailing whitespaces. To make sure the patch submitter is
  warned about it, I have enabled trailing whitespace check on
  mediawiki/core.git
 
  The bug report is:
   https://bugzilla.wikimedia.org/42628
 
  I have only enabled such check on the mediawiki/core.git repository for
  now.  If that went well I will look up at enabling it on other repos.
 

 Is this a warning, or a failure? I really don't think we should *fail*
 patches for something that's almost always harmless. The only
 reason I ask is because you say warning here but failure on the
 bug.

 -Chad

 ___
 Wikitech-l mailing list
 Wikitech-l@lists.wikimedia.org
 https://lists.wikimedia.org/mailman/listinfo/wikitech-l



___
Wikitech-l mailing list
Wikitech-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikitech-l


Re: [Wikitech-l] Any OSCON plans?

2012-12-21 Thread Luke Welling WMF
I've spoken at it the last 13 years, but I was thinking of skipping it this
time.

I've also been on the committee the last few years.  I don't yet know if I
will be this year, but if you want my perspective on  what gets accepted
I'd be happy to share it.  In general state of TechnologyX is tough to
get accepted unless you are Rasmus, Larry or Guido. How we did
challenging new project at high profile site with technologyX is a
much better bet. OSCON always runs some talks along the lines of saving
the world with open culture, so a charismatic speaker with a non-obvious
spin on that is worth a shot.

Luke Welling


On Fri, Dec 21, 2012 at 2:35 PM, Quim Gil q...@wikimedia.org wrote:

 OSCON call for papers is now open:

 July 22-26 at Portland (USA)
 http://www.oscon.com/oscon2013

 Anybody considering? Thoughts?

 If we are interested: OSCON is a tough venue when it comes to get
 proposals accepted. Good planning and coordination does help, and this is
 why I'm asking here  now.

 --
 Quim Gil
 Technical Contributor Coordinator @ Wikimedia Foundation
 http://www.mediawiki.org/wiki/**User:Qgilhttp://www.mediawiki.org/wiki/User:Qgil

 __**_
 Wikitech-l mailing list
 Wikitech-l@lists.wikimedia.org
 https://lists.wikimedia.org/**mailman/listinfo/wikitech-lhttps://lists.wikimedia.org/mailman/listinfo/wikitech-l

___
Wikitech-l mailing list
Wikitech-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikitech-l


Re: [Wikitech-l] Video on mobile: Firefox works, way is paved for more browser support

2012-12-12 Thread Luke Welling
FirefoxOS/Boot2Gecko phones presumably also support Ogg Theora
and WebM formats, but they're not really a market share yet and may never
be in the developed world.

Without trying to downplay the importance of ideological purity, keep in
mind that Mozilla, who have largely the same ideology on the matter have
conceded defeat on the practical side of it after investing significant
effort.

Eg http://appleinsider.
com/articles/12/03/14/mozilla_considers_h264_video_support_after_googles_vp8_fails_to_gain_traction

With Google unwilling to commit the battle was winnable.

There is not an ideologically pure answer that is compatible with the goal
of taking video content and disseminating it effectively and globally.  The
conversation needs to be framed as what shade of grey is an acceptable
compromise.

Luke Welling


On Wed, Dec 12, 2012 at 6:44 AM, Antoine Musso hashar+...@free.fr wrote:

 Le 12/12/12 00:15, Erik Moeller a écrit :
  Since there are multiple potential paths for changing the policy
  (keeping things ideologically pure, allowing conversion on ingestion,
  allowing h.264 but only for mobile, allowing h.264 for all devices,
  etc.), and since these issues are pretty contentious, it seems like a
  good candidate for an RFC which'll help determine if there's an
  obvious consensus path forward.

 Could we host h.264 videos and related transcoders in a country that
 does not recognize software patents?


 Hints:
  - I am not a lawyer
  - WMF has server in Netherlands, EU.


 --
 Antoine hashar Musso


 ___
 Wikitech-l mailing list
 Wikitech-l@lists.wikimedia.org
 https://lists.wikimedia.org/mailman/listinfo/wikitech-l

___
Wikitech-l mailing list
Wikitech-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikitech-l


Re: [Wikitech-l] Video on mobile: Firefox works, way is paved for more browser support

2012-12-12 Thread Luke Welling
Thanks for the link. I'll try and stay out of it until I've had time to
read the old thread, but I think this is an unfair characterization:

On Wed, Dec 12, 2012 at 2:35 PM, David Gerard dger...@gmail.com wrote:

 This proposal is not about anything other than enhancing the shiny for
 owners of iOS devlces.


It's about providing knowledge to the rapidly growing userbase of mobile
device owners who fall outside the tiny segment that is Android users who
have deliberately chosen to replace the stock Android browser with Firefox.

Luke Welling
___
Wikitech-l mailing list
Wikitech-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikitech-l


Re: [Wikitech-l] Escaping in SQL queries

2012-12-04 Thread Luke Welling
I can't speak for all databases, but I know MySQL fairly well, and it will
generally be safe there.

You will be able to find some edge cases where it may not give you what you
expected. For example
select 0x10 + 1;
is not the same as
select '0x10' + 1;

The first will give 17, the second 1 as the string will get converted to 0,
but I don't know that there is a lot of non-decimal math being done here.

In most cases though, type conversion will work as you expect it too, the
following will all give 2
select '1' + '1'
select '1' + 1
select 1+ 1

Also, if we use MySQL's query cache, it is (or at least was) fairly naive
and compared query strings for an exact match so you will get a cache miss
if you run the same query twice but only include redundant quotes one time.

Quoting nulls would be a problem, but it is not doing that.

More importantly, the function strencode() called inside that addQuotes()
function is overridden for different underlying databases.  The MySQL one
looks good.  It is using mysql_real_escape_string and providing the
connection so the correct character set will be used.

It looks to me like the MS SQL version requires the programmer to manually
call encodeBlob() before building their query if they know they are
inserting binary data.

Luke Welling


On Tue, Dec 4, 2012 at 4:52 AM, Daniel Kinzler dan...@brightbyte.de wrote:

 Hi all!

 I recently found that it is less than clear how numbers should be
 quoted/escaped
 in SQL queries. Should DatabaseBase::addQuotes() be used, or rather just
 inval(), to make sure it's really a number? What's the best practice?

 Looking at DatabaseBase::makeList(), it seems that addQuotes() is used on
 all
 values, string or not. So, what does addQuotes() actually do? Does it
 always add
 quotes, turning the value into a string literal, or does it rather apply
 whatever quoting/escaping is appropriate for the given data type?

 addQuotes' documentation sais:

  * If it's a string, adds quotes and backslashes
  * Otherwise returns as-is

 That's a plain LIE. Here's the code:

 if ( $s === null ) {
 return 'NULL';
 } else {
 # This will also quote numeric values. This should be harmless,
 # and protects against weird problems that occur when they really
 # _are_ strings such as article titles and string-number-string
 # conversion is not 1:1.
 return ' . $this-strencode( $s ) . ';
 }

 So, it actually always returns a quoted string literal, unless $s is null.

 But is it true what the comment sais? Is it really always harmless to quote
 numeric values? Will all database engines always magically convert them to
 numbers before executing the query? If not, this may be causing table
 scans.
 That would be bad - but I suppose someone would have noticed by now...

 So... at the very least, addQuotes' documentation needs fixing. And
 perhaps it
 would be nice to have an additional method that only applies the
 appropriate
 quoting, e.g. escapeValue or some such - that's how addQuotes seems to be
 currently used, but that's not what it actually does... What do you think?


 -- daniel


 PS: There's more fun. The DatabaseMssql class overrides addQuotes to
 support
 Blob object. For the case $s is a Blob object, this code is used:

 return ' . $s-fetch( $s ) . ';

 The value is used raw, without any escaping. Looks like if there's a ' in
 the
 blob, fun things will happen. Or am I missing something?


 ___
 Wikitech-l mailing list
 Wikitech-l@lists.wikimedia.org
 https://lists.wikimedia.org/mailman/listinfo/wikitech-l

___
Wikitech-l mailing list
Wikitech-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikitech-l


[Wikitech-l] Newbie licensing questions

2012-11-28 Thread Luke Welling
These are prompted by Gerrit review comments, but it seems like a hidden
place to discuss them.

The specific background:
A task I am on uses an external library licensed under the Apache2.0
license.  Mediawiki core is shipped under a GPL2 license, so I'm guessing
that is the default licence fondation work is released under.

According to http://en.wikipedia.org/wiki/Apache_License#GPL_compatibilityApache
licenses are compatible with GPL3 but not GPL2.

The questions:
Is my assumption that by default we put everything under GPL2 right?
Are other FS/OSS licences OK to use?
How paranoid are we? ie do we make a good faith effort at getting it right,
or do we refer questions to internal counsel for a slower but safer answer?

In this case my inclination is to licence the whole extension (containing
the external library) as Apache2.0 but I'm happy to defer to normal process
if there is one.

Thanks,

Luke Welling
___
Wikitech-l mailing list
Wikitech-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikitech-l


Re: [Wikitech-l] WMF's webpage

2012-11-28 Thread Luke Welling
Is there a reason not to use the Yahoo championed approach of embedding a
version number in all static file names so you can set a very long cache
expires time and just add new versions to the CDN when a change is made?

I don't know how often our CSS, branding images, scripts, and other static
content change, but there would not be much effort in adding that to a
deploy process and there must be developer overhead being incurred in
trying to keep new code backwards compatible.

Luke Welling


On Wed, Nov 28, 2012 at 5:52 PM, Platonides platoni...@gmail.com wrote:

 Code was updated to use h3 while the (cached) skin CSS still had h5

 See

 http://lists.wikimedia.org/pipermail/wikitech-ambassadors/2012-November/93.html

 Note: code was reverted to wmf4, so the problem will appear now in the
 reverse.

 We should use an intermediate CSS with rules appliying to portlets no
 matter if they are h3 or h5. Then migrate again in a few days.



 ___
 Wikitech-l mailing list
 Wikitech-l@lists.wikimedia.org
 https://lists.wikimedia.org/mailman/listinfo/wikitech-l

___
Wikitech-l mailing list
Wikitech-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikitech-l


Re: [Wikitech-l] Proposed new WMF browser support framework for MediaWiki

2012-11-26 Thread Luke Welling
Something that seems to be being partially considered in these models
is browser shares that are growing or shrinking.

For instance it would make more sense to me to support IE10 than IE7
even if IE7 has a far greater market share this month.  If significant
work is needed for IE7, it might have dropped below an age or market
share threshold before that work is complete.

By the same logic, it seems fair that mobile browsers make the list on
a lower raw market share given that that market is likely to continue
to grow for some time.

Putting those rules (particularly for IE) into unambiguous, fair,
numeric cutoffs may be hard. Predicting future IE market share depends
a great deal on the vagaries of auto update policies and corporate IT
departments.  For instance Windows XP users are stuck on IE8 forever
and there are probably quite a lot of them in corporate America.

Luke Welling

On Mon, Nov 26, 2012 at 2:53 PM, Tomasz Finc tf...@wikimedia.org wrote:
 Correction. I was looking at the total Other/Unknown. Opera is actually
 4.66%

 --tomasz


 On Mon, Nov 26, 2012 at 11:00 AM, Tomasz Finc tf...@wikimedia.org wrote:

 On Fri, Nov 23, 2012 at 12:50 PM, James Forrester 
 jforres...@wikimedia.org wrote:

 … which seems to be a little harsh on the mobile and tablet fronts, and
 overly-
 generous on the MSIE side given their exceptional costs to support per
 %age of
 users, but not too terrible.


 I worry less about it being harsh and more that it needs to be lined up
 with  which documents the mobile teams current support matrix [1].

 Removing Opera support as a general rule is not an option for us as it
 contributes a significant amount of readership traffic (6.69%) [2] .

 --tomasz

 [1] - http://www.mediawiki.org/wiki/Mobile/Testing_process
 [2] - http://stats.wikimedia.org/wikimedia/squids/SquidReportDevices.htm


 ___
 Wikitech-l mailing list
 Wikitech-l@lists.wikimedia.org
 https://lists.wikimedia.org/mailman/listinfo/wikitech-l

___
Wikitech-l mailing list
Wikitech-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikitech-l


Re: [Wikitech-l] Wikimedia URL shortener

2012-11-19 Thread Luke Welling
There are some practical reasons for providing site specific domain
shortening beyond owning a cool domain hack.

Discouraging the use of third party shorteners has real value.  The
web getting littered with urls that camouflage their destination, are
tied to loss making commercial entities of unpredictable lifetime (eg
tr.im), or tied to very well funded commercial entities that choose a
ccTLD of unpredictable stability (eg bit.ly) is a bad thing.
Providing the option of a short url that achieves whatever benefit was
sought without breaking the way the web is supposed to work is a good
thing.

Email used to be problematic for url sharing, as line breaks inserted
at 70 characters would mean the recipient needed to notice that some
urls had been chopped  and reassemble them.  This is probably only
true for mailing lists now, as people who deliberately use a plain
text only email client in 2012 will experience obtrusive side effects
from senders only providing an HTML MIME part regularly.  URL chopping
will not be their major annoyance.  Mailing lists still routinely
strip attachments and encodings from mail that they propagate.

Mobile generally and twitter specifically are the most often cited
justifications now. Aside from the obvious message length limits,
cutting and pasting long strings can be hard on a small screen so
people are in the short url habit.

Twitter is an annoying use case, as even if presented with a short url
it currently replaces it with a potentially longer t.co url.

If all the project does is reduces the use of third party services
that can permanently or transiently fail, can hide links to malware,
break search engine rankings and search behaviour, and provide others
with analytic insight into potentially sensitive user click throughs
it is a good thing.

Luke Welling

PS my unobtainable cool domain hack of choice would be en.cy (but
Cyprus don't do top level subdomains and require local presence)


On Mon, Nov 19, 2012 at 5:49 AM, Neil Harris n...@tonal.clara.co.uk wrote:
 On 19/11/12 02:09, MZMcBride wrote:

 Tim Starling wrote:

 Providing a traditional URL shortener with traditional features would
 be useful, at least for some people, and would eliminate the need to
 use third-party URL shorteners internally, which have a commercial
 focus and unknown lifetime. If there was no integration with MW
 required, it would only take a couple of hours to set up.

 Who's using third-party URL shorteners internally? A lot of services would
 be useful to at least some people (Wikimedians), but the cost of setting
 up
 _and maintaining_ such a service can't be overlooked, in my opinion. Yes,
 it
 would take a few hours to set up a URL shortening service (if that), but
 who's going to be responsible for fixing bugs in it, adding features, and
 doing general maintenance to the service for the indefinite future? There
 are already a number of Wikimedia services that struggle for limited
 resources. Before we add another, we must answer the maintenance question.

 MZMcBride



 If a Wikimedia URL shortening service was to be created, it would make sense
 to, at the very least, (a) make the shortened link to real link mappings
 part of the standard Mediwiki XML dumps, so that they can be preserved
 alongside the content to which they refer, for access by future archivists,
 and (b) participate in initiatives such as the Internet Archive's
 301works.org to preserve these links entirely outside the Wikimedia
 universe.

 Also, on a separate but related note, has anyone considered creating DOIs
 for individual wiki page revisions?

 -- Neil



 ___
 Wikitech-l mailing list
 Wikitech-l@lists.wikimedia.org
 https://lists.wikimedia.org/mailman/listinfo/wikitech-l

___
Wikitech-l mailing list
Wikitech-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikitech-l