[Wikitech-l] Fwd: [Wikimedia-l] Notification about problem identified with a recent CentralNotice banner

2018-03-18 Thread Pine W
Forwarding to Wikitech-l for the benefit of Wikitech-l subscribers who do
not subscribe to Wikimedia-l.

Pine
( https://meta.wikimedia.org/wiki/User:Pine )

-- Forwarded message --
From: Gregory Varnum 
Date: Fri, Mar 16, 2018 at 6:57 PM
Subject: [Wikimedia-l] Notification about problem identified with a recent
CentralNotice banner
To: Wikimedia Mailing List 


On 14 March and 15 March 2018, a CentralNotice banner appeared to some
logged-out users viewing English Wikipedia pages. The banner contained
JavaScript hosted by Facebook, which allowed Facebook to collect traffic
data from those who visited a page with a banner. The banner was prepared
by the Wikimedia Foundation. The Foundation turned the banner off as soon
as we learned how the script was running, and its potential scope. We have
also removed all references to the code in question from CentralNotice on
Meta-Wiki.

The code utilized in this banner was based on an unused prototype created
by an outside vendor. Because the prototype was never enabled, the vendor’s
prototype code was not subjected to our standard quality assurance process.
However, we made the mistake of reusing the code for a different purpose,
and implementing it based on recommendations in documentation from Twitter
and Facebook to improve the appearance of shared links. At the time, our
understanding was that the platforms would only receive traffic data if the
user clicked on the link. Although this was true for Twitter, the Facebook
code operated differently.

We discovered the problematic link configurations during our ongoing
monitoring of live banners. The recommended code enhanced not only the
appearance of links, it also enhanced Facebook's ability to collect
information on people visiting non-Facebook sites. As soon as we realized
these banners were sharing information without even having to click the
link, we disabled them and began an investigation. Staff in multiple
departments are collaboratively reviewing the incident as well as
procedural and technical improvements to prevent future incidents.

While this sort of tracking is commonplace today across most of the
internet, it is not consistent with our policies. We are disappointed that
this type of hidden data collection is routinely recommended by major
platforms, without clearer disclosure.

These practices are why we all must regularly take routine steps to
maintain a secure computer and account. As the Wikimedia Foundation
continues to explore ways we can do that within Wikimedia's platform, we
encourage you to consider tools which block unwanted third-party scripts
like the one provided by Facebook.

We apologize for sending this late on a Friday (San Francisco time).
However, we wanted to provide this information as quickly as possible.
___
Wikimedia-l mailing list, guidelines at: https://meta.wikimedia.org/
wiki/Mailing_lists/Guidelines and https://meta.wikimedia.org/
wiki/Wikimedia-l
New messages to: wikimedi...@lists.wikimedia.org
Unsubscribe: https://lists.wikimedia.org/mailman/listinfo/wikimedia-l,

___
Wikitech-l mailing list
Wikitech-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikitech-l

Re: [Wikitech-l] Incoming and outgoing links enquiry

2018-03-18 Thread John
I would second the recommendation of using the dumps for such a large
graphing project. If it's more than a couple hundred pages the API/database
queries can get bulky

On Sun, Mar 18, 2018 at 5:07 PM Brian Wolff  wrote:

> Hi,
>
> You can run longer queries by getting access to toolforge (
> https://wikitech.wikimedia.org/wiki/Portal:Toolforge) and running from the
> command line.
>
> However the query in question might  still take an excessively long time
> (if you are doing all of wikipedia). I would expect that query to result in
> about 150mb of data and maybe take days to complete.
>
> You can also break it down into parts by adding WHERE page_title >='a' AND
> page_title < 'b'
>
> Note, also of interest: full dumps of all the links is available at
>
> https://dumps.wikimedia.org/enwiki/20180301/enwiki-20180301-pagelinks.sql.gz
> (you would also need
> https://dumps.wikimedia.org/enwiki/20180301/enwiki-20180301-page.sql.gz to
> convert page ids to page names)
> --
> Brian
> On Sunday, March 18, 2018, Nick Bell  wrote:
> > Hi there,
> >
> > I'm a final year Mathematics student at the University of Bristol, and
> I'm
> > studying Wikipedia as a graph for my project.
> >
> > I'd like to get data regarding the number of outgoing links on each page,
> > and the number of pages with links to each page. I have already
> > inquired about this with the Analytics Team mailing list, who gave me a
> few
> > suggestions.
> >
> > One of these was to run the code at this link
> https://quarry.wmflabs.org/
> > query/25400
> > with these instructions:
> >
> > "You will have to fork it and remove the "LIMIT 10" to get it to run on
> > all the English Wikipedia articles. It may take too long or produce
> > too much data, in which case please ask on this list for someone who
> > can run it for you."
> >
> > I ran the code as instructed, but the query was killed as it took longer
> > than 30 minutes to run. I asked if anyone on the mailing list could run
> it
> > for me, but no one replied saying they could. The guy who wrote the code
> > suggested I try this mailing list to see if anyone can help.
> >
> > I'm a beginner in programming and coding etc., so any and all help you
> can
> > give me would be greatly appreciated.
> >
> > Many thanks,
> > Nick Bell
> > University of Bristol
> > ___
> > Wikitech-l mailing list
> > Wikitech-l@lists.wikimedia.org
> > https://lists.wikimedia.org/mailman/listinfo/wikitech-l
> ___
> Wikitech-l mailing list
> Wikitech-l@lists.wikimedia.org
> https://lists.wikimedia.org/mailman/listinfo/wikitech-l
___
Wikitech-l mailing list
Wikitech-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikitech-l

Re: [Wikitech-l] Incoming and outgoing links enquiry

2018-03-18 Thread Brian Wolff
Hi,

You can run longer queries by getting access to toolforge (
https://wikitech.wikimedia.org/wiki/Portal:Toolforge) and running from the
command line.

However the query in question might  still take an excessively long time
(if you are doing all of wikipedia). I would expect that query to result in
about 150mb of data and maybe take days to complete.

You can also break it down into parts by adding WHERE page_title >='a' AND
page_title < 'b'

Note, also of interest: full dumps of all the links is available at
https://dumps.wikimedia.org/enwiki/20180301/enwiki-20180301-pagelinks.sql.gz
(you would also need
https://dumps.wikimedia.org/enwiki/20180301/enwiki-20180301-page.sql.gz to
convert page ids to page names)
--
Brian
On Sunday, March 18, 2018, Nick Bell  wrote:
> Hi there,
>
> I'm a final year Mathematics student at the University of Bristol, and I'm
> studying Wikipedia as a graph for my project.
>
> I'd like to get data regarding the number of outgoing links on each page,
> and the number of pages with links to each page. I have already
> inquired about this with the Analytics Team mailing list, who gave me a
few
> suggestions.
>
> One of these was to run the code at this link https://quarry.wmflabs.org/
> query/25400
> with these instructions:
>
> "You will have to fork it and remove the "LIMIT 10" to get it to run on
> all the English Wikipedia articles. It may take too long or produce
> too much data, in which case please ask on this list for someone who
> can run it for you."
>
> I ran the code as instructed, but the query was killed as it took longer
> than 30 minutes to run. I asked if anyone on the mailing list could run it
> for me, but no one replied saying they could. The guy who wrote the code
> suggested I try this mailing list to see if anyone can help.
>
> I'm a beginner in programming and coding etc., so any and all help you can
> give me would be greatly appreciated.
>
> Many thanks,
> Nick Bell
> University of Bristol
> ___
> Wikitech-l mailing list
> Wikitech-l@lists.wikimedia.org
> https://lists.wikimedia.org/mailman/listinfo/wikitech-l
___
Wikitech-l mailing list
Wikitech-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikitech-l

[Wikitech-l] Incoming and outgoing links enquiry

2018-03-18 Thread Nick Bell
Hi there,

I'm a final year Mathematics student at the University of Bristol, and I'm
studying Wikipedia as a graph for my project.

I'd like to get data regarding the number of outgoing links on each page,
and the number of pages with links to each page. I have already
inquired about this with the Analytics Team mailing list, who gave me a few
suggestions.

One of these was to run the code at this link https://quarry.wmflabs.org/
query/25400
with these instructions:

"You will have to fork it and remove the "LIMIT 10" to get it to run on
all the English Wikipedia articles. It may take too long or produce
too much data, in which case please ask on this list for someone who
can run it for you."

I ran the code as instructed, but the query was killed as it took longer
than 30 minutes to run. I asked if anyone on the mailing list could run it
for me, but no one replied saying they could. The guy who wrote the code
suggested I try this mailing list to see if anyone can help.

I'm a beginner in programming and coding etc., so any and all help you can
give me would be greatly appreciated.

Many thanks,
Nick Bell
University of Bristol
___
Wikitech-l mailing list
Wikitech-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikitech-l

[Wikitech-l] ¿Qué te hace feliz esta semana? / What's making you happy this week? (Week of 18 March 2018)

2018-03-18 Thread Pine W
What's making me happy this week:

* French Wikipedia administrator and OTRS volunteer 0x010C
 created a calendar in Lua
that works on MediaWiki. See https://meta.wikimedia.org/wik
i/Template:Events_calendar and https://lists.wikimedia.org/pi
permail/wikimedia-l/2018-March/089886.html.

* Wikimedia Russia's multilingual writing contest regarding cities and
regions in the 2018 FIFA world cup looks interesting. I like the concept of
associating Wikipedia writing campaigns with popular current events. See
https://ru.wikimedia.org/wiki/%D0%9A%D0%BE%D0%BD%D0%BA%D1%83
%D1%80%D1%81%D1%8B/%D0%A3%D0%B7%D0%BD%D0%B0%D0%B9_%D0%A0%
D0%BE%D1%81%D1%81%D0%B8%D1%8E._%D0%93%D0%BE%D1%80%D0%BE%D0%
B4%D0%B0_%D0%B8_%D1%80%D0%B5%D0%B3%D0%B8%D0%BE%D0%BD%D1%8B_
%D0%A7%D0%9C_%D0%BF%D0%BE_%D1%84%D1%83%D1%82%D0%B1%D0%BE%D0%BB%D1%83_2018/en
and https://lists.wikimedia.org/pipermail/wikimedia-l/2018-March/089876.html
.

* Persian Wikipedia administrator Mohammad

quickly reverted (in approximately 7 minutes) an edit which inserted
cryptomining JavaScript into commons.js. See https://lists.wikimedia.org/pi
permail/wikitech-l/2018-March/089629.html.

From the WMF Blog:

* "Confound it!—Supporting languages with multiple writing systems":
https://blog.wikimedia.org/2018/03/12/supporting-languages-
multiple-writing-systems/

* "James Heilman on expanding the reach of Wikipedia’s medical information":
https://blog.wikimedia.org/2018/03/13/offline-access-medical-information/

* "How we’re using machine learning to visually enrich Wikidata":
https://blog.wikimedia.org/2018/03/14/machine-learning-visually-enriching-wikidata/

What's making you happy this week?

Pine
( https://meta.wikimedia.org/wiki/User:Pine )
___
Wikitech-l mailing list
Wikitech-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikitech-l