Re: [Wikidata] Looking for list article editors

2016-06-15 Thread Biyanto Rebin
Dear Lydia,

You can contact me. I'm on project of language, ethnicity, and also
administrative division in Indonesia.
I usually import the data from the idwiki into Wikidata using PetScan too.

Regards,

2016-06-16 3:01 GMT+07:00 Lydia Pintscher :

> Hey folks :)
>
> We're starting concept work for automated list generation on Wikipedia and
> the other Wikimedia projects based on queries to Wikidata. As a first step
> I'd like to get a better understanding of the current state of things. For
> this Jan (our UX person) and me would like to have a chat with a few
> editors from different Wikipedias and other Wikimedia projects who are
> maintaining list articles. If that is you or you know someone who fits
> please let me know.
>
> Cheers
> Lydia
> --
> Lydia Pintscher - http://about.me/lydia.pintscher
> Product Manager for Wikidata
>
> Wikimedia Deutschland e.V.
> Tempelhofer Ufer 23-24
> 10963 Berlin
> www.wikimedia.de
>
> Wikimedia Deutschland - Gesellschaft zur Förderung Freien Wissens e. V.
>
> Eingetragen im Vereinsregister des Amtsgerichts Berlin-Charlottenburg
> unter der Nummer 23855 Nz. Als gemeinnützig anerkannt durch das Finanzamt
> für Körperschaften I Berlin, Steuernummer 27/029/42207.
>
> ___
> Wikidata mailing list
> Wikidata@lists.wikimedia.org
> https://lists.wikimedia.org/mailman/listinfo/wikidata
>
>


-- 

Biyanto Rebin | Ketua Umum (*Chair*) 2016-2018
Wikimedia Indonesia
Nomor Ponsel: +62 8989 037379
Surel: biyanto.re...@wikimedia.or.id


Dukung upaya kami membebaskan pengetahuan:
http://wikimedia.or.id/wiki/Wikimedia_Indonesia:Donasi
___
Wikidata mailing list
Wikidata@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikidata


Re: [Wikidata] Looking for list article editors

2016-06-15 Thread Gerard Meijssen
Hoi,
I have been really active on lists with awards.. I would love it when these
are available from Wikidata. If you care to pick my brain I am there for
you.
Thanks,
 GerardM

On 15 June 2016 at 22:01, Lydia Pintscher 
wrote:

> Hey folks :)
>
> We're starting concept work for automated list generation on Wikipedia and
> the other Wikimedia projects based on queries to Wikidata. As a first step
> I'd like to get a better understanding of the current state of things. For
> this Jan (our UX person) and me would like to have a chat with a few
> editors from different Wikipedias and other Wikimedia projects who are
> maintaining list articles. If that is you or you know someone who fits
> please let me know.
>
> Cheers
> Lydia
> --
> Lydia Pintscher - http://about.me/lydia.pintscher
> Product Manager for Wikidata
>
> Wikimedia Deutschland e.V.
> Tempelhofer Ufer 23-24
> 10963 Berlin
> www.wikimedia.de
>
> Wikimedia Deutschland - Gesellschaft zur Förderung Freien Wissens e. V.
>
> Eingetragen im Vereinsregister des Amtsgerichts Berlin-Charlottenburg
> unter der Nummer 23855 Nz. Als gemeinnützig anerkannt durch das Finanzamt
> für Körperschaften I Berlin, Steuernummer 27/029/42207.
>
> ___
> Wikidata mailing list
> Wikidata@lists.wikimedia.org
> https://lists.wikimedia.org/mailman/listinfo/wikidata
>
>
___
Wikidata mailing list
Wikidata@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikidata


Re: [Wikidata] language fallbacks on Wikipedia and co

2016-06-15 Thread Gerard Meijssen
Hoi,
Wil it work using the #babel templates?
Thanks,
 GerardM

On 15 June 2016 at 21:47, Lydia Pintscher 
wrote:

> Hey folks :)
>
> With the rollout of the ArticlePlaceholder one issue became very apparent:
> language fallbacks for them are needed - especially on the small languages
> - because otherwise you'll see a lot of Qs and Ps... We should have made
> this possible a long time ago but now with the ArticlePlaceholder it became
> more urgent. So starting in the next days you will see more fallbacks
> happening in Lua and the property parser function on Wikipedia and the
> other Wikimedia projects using Wikidata data. It will use the usual
> MediaWiki fallback. It might need a purge for some pages initially to
> actually show up.
> If you see any problems popping up with this please let me know about it.
>
>
> Cheers
> Lydia
> --
> Lydia Pintscher - http://about.me/lydia.pintscher
> Product Manager for Wikidata
>
> Wikimedia Deutschland e.V.
> Tempelhofer Ufer 23-24
> 10963 Berlin
> www.wikimedia.de
>
> Wikimedia Deutschland - Gesellschaft zur Förderung Freien Wissens e. V.
>
> Eingetragen im Vereinsregister des Amtsgerichts Berlin-Charlottenburg
> unter der Nummer 23855 Nz. Als gemeinnützig anerkannt durch das Finanzamt
> für Körperschaften I Berlin, Steuernummer 27/029/42207.
>
> ___
> Wikidata mailing list
> Wikidata@lists.wikimedia.org
> https://lists.wikimedia.org/mailman/listinfo/wikidata
>
>
___
Wikidata mailing list
Wikidata@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikidata


[Wikidata] user script to show Wikidata changes in article history

2016-06-15 Thread Lydia Pintscher
Hey folks :)

H4stings has created a user script to show Wikidata changes in the history
of a Wikipedia article. (It should also work on the other sister projects.)
https://phabricator.wikimedia.org/T42358#2363346
The script is at
https://fr.wikipedia.org/wiki/Utilisateur:H4stings/wef-history.js
I'd love to hear your feedback on it so we can see if that is a direction
we should explore further.

Cheers
Lydia
-- 
Lydia Pintscher - http://about.me/lydia.pintscher
Product Manager for Wikidata

Wikimedia Deutschland e.V.
Tempelhofer Ufer 23-24
10963 Berlin
www.wikimedia.de

Wikimedia Deutschland - Gesellschaft zur Förderung Freien Wissens e. V.

Eingetragen im Vereinsregister des Amtsgerichts Berlin-Charlottenburg unter
der Nummer 23855 Nz. Als gemeinnützig anerkannt durch das Finanzamt für
Körperschaften I Berlin, Steuernummer 27/029/42207.
___
Wikidata mailing list
Wikidata@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikidata


[Wikidata] Looking for list article editors

2016-06-15 Thread Lydia Pintscher
Hey folks :)

We're starting concept work for automated list generation on Wikipedia and
the other Wikimedia projects based on queries to Wikidata. As a first step
I'd like to get a better understanding of the current state of things. For
this Jan (our UX person) and me would like to have a chat with a few
editors from different Wikipedias and other Wikimedia projects who are
maintaining list articles. If that is you or you know someone who fits
please let me know.

Cheers
Lydia
-- 
Lydia Pintscher - http://about.me/lydia.pintscher
Product Manager for Wikidata

Wikimedia Deutschland e.V.
Tempelhofer Ufer 23-24
10963 Berlin
www.wikimedia.de

Wikimedia Deutschland - Gesellschaft zur Förderung Freien Wissens e. V.

Eingetragen im Vereinsregister des Amtsgerichts Berlin-Charlottenburg unter
der Nummer 23855 Nz. Als gemeinnützig anerkannt durch das Finanzamt für
Körperschaften I Berlin, Steuernummer 27/029/42207.
___
Wikidata mailing list
Wikidata@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikidata


[Wikidata] user script for finding items with the same statement

2016-06-15 Thread Lydia Pintscher
Hey :)

Amir has written a very nice little user script that adds a little icon
next to a statement. Clicking on it leads you to a query for items with the
same statement. So if you are on an item about a cat you can click it on
the "instance of" statement and then find all other cats. More details
here:
https://www.wikidata.org/wiki/Wikidata:Project_chat#Do_you_want_to_see_statements_like_this.3F
Please
also leave feedback there. If you like it it can be turned into a gadget
later. (I don't think it is a good idea to put it in the default UI at this
point.)


Cheers
Lydia
-- 
Lydia Pintscher - http://about.me/lydia.pintscher
Product Manager for Wikidata

Wikimedia Deutschland e.V.
Tempelhofer Ufer 23-24
10963 Berlin
www.wikimedia.de

Wikimedia Deutschland - Gesellschaft zur Förderung Freien Wissens e. V.

Eingetragen im Vereinsregister des Amtsgerichts Berlin-Charlottenburg unter
der Nummer 23855 Nz. Als gemeinnützig anerkannt durch das Finanzamt für
Körperschaften I Berlin, Steuernummer 27/029/42207.
___
Wikidata mailing list
Wikidata@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikidata


[Wikidata] language fallbacks on Wikipedia and co

2016-06-15 Thread Lydia Pintscher
Hey folks :)

With the rollout of the ArticlePlaceholder one issue became very apparent:
language fallbacks for them are needed - especially on the small languages
- because otherwise you'll see a lot of Qs and Ps... We should have made
this possible a long time ago but now with the ArticlePlaceholder it became
more urgent. So starting in the next days you will see more fallbacks
happening in Lua and the property parser function on Wikipedia and the
other Wikimedia projects using Wikidata data. It will use the usual
MediaWiki fallback. It might need a purge for some pages initially to
actually show up.
If you see any problems popping up with this please let me know about it.


Cheers
Lydia
-- 
Lydia Pintscher - http://about.me/lydia.pintscher
Product Manager for Wikidata

Wikimedia Deutschland e.V.
Tempelhofer Ufer 23-24
10963 Berlin
www.wikimedia.de

Wikimedia Deutschland - Gesellschaft zur Förderung Freien Wissens e. V.

Eingetragen im Vereinsregister des Amtsgerichts Berlin-Charlottenburg unter
der Nummer 23855 Nz. Als gemeinnützig anerkannt durch das Finanzamt für
Körperschaften I Berlin, Steuernummer 27/029/42207.
___
Wikidata mailing list
Wikidata@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikidata


Re: [Wikidata] [ANNOUNCEMENT] StrepHit 1.0 Beta Release

2016-06-15 Thread Marco Fossati

Hi Ben,

On 6/15/16 18:24, Benjamin Good wrote:

Hi Marco,

Where might we find some statistics on the current accuracy of the
automated claim and reference extractors?  I assume that information
must be in there somewhere, but I had trouble finding it.

The StrepHit pipeline (codebase) is ready, while the project is ongoing.
We are not there yet, and will publish performance values in the final 
report.


This is a very ambitious project covering a very large technical
territory (which I applaud).  It would be great if your results could be
synthesized a bit more clearly so we can understand where the
weak/strong points are and where we might be able to help improve or
make use of what you have done in other domains.

Sure, this will be done in the final report.
Up to now, you can have a look at the midpoint report summary:
https://meta.wikimedia.org/wiki/Grants:IEG/StrepHit:_Wikidata_Statements_Validation_via_References/Midpoint#Summary

Best,

Marco


-Ben


On Wed, Jun 15, 2016 at 9:06 AM, Marco Fossati mailto:foss...@spaziodati.eu>> wrote:

[Feel free to blame me if you read this more than once]

To whom it may interest,

Full of delight, I would like to announce the first beta release of
*StrepHit*:

https://github.com/Wikidata/StrepHit

TL;DR: StrepHit is an intelligent reading agent that understands
text and translates it into *referenced* Wikidata statements.
It is a IEG project funded by the Wikimedia Foundation.

Key features:
-Web spiders to harvest a collection of documents (corpus) from
reliable sources
-automatic corpus analysis to understand the most meaningful verbs
-sentences and semi-structured data extraction
-train a machine learning classifier via crowdsourcing
-*supervised and rule-based fact extraction from text*
-Natural Language Processing utilities
-parallel processing

You can find all the details here:

https://meta.wikimedia.org/wiki/Grants:IEG/StrepHit:_Wikidata_Statements_Validation_via_References

https://meta.wikimedia.org/wiki/Grants:IEG/StrepHit:_Wikidata_Statements_Validation_via_References/Midpoint

If you like it, star it on GitHub!

Best,

Marco

___
Wikidata mailing list
Wikidata@lists.wikimedia.org 
https://lists.wikimedia.org/mailman/listinfo/wikidata




___
Wikidata mailing list
Wikidata@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikidata



___
Wikidata mailing list
Wikidata@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikidata


Re: [Wikidata] RFC - Primary Sources?

2016-06-15 Thread Tom Morris
On Wed, Jun 15, 2016 at 12:49 PM, Lydia Pintscher <
lydia.pintsc...@wikimedia.de> wrote:

> On Wed, Jun 15, 2016 at 7:42 AM Lydia Pintscher <
> lydia.pintsc...@wikimedia.de> wrote:
>
>>
>> On Jun 14, 2016 23:56, "Tom Morris"  wrote:
>> >
>> > Thanks for the reminder. So that solves the "asking" part.
>> >
>> > Does anyone *not* think that the Wikidata engineering team is the
>> correct place for this?
>> >
>> > Lydia - can you assign someone to come up to speed at whatever level
>> Denny requires to feel comfortable making the transfer?
>>
>> I will take care of it with Denny in the next days.
>>
>
> FYI: In-progress now. I'll report back as soon as it is done but it'll
> take a few days to clarify some stuff.
>

Cool. Thanks for the quick action.

Tom
___
Wikidata mailing list
Wikidata@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikidata


Re: [Wikidata] RFC - Primary Sources?

2016-06-15 Thread Lydia Pintscher
On Wed, Jun 15, 2016 at 7:42 AM Lydia Pintscher <
lydia.pintsc...@wikimedia.de> wrote:

>
> On Jun 14, 2016 23:56, "Tom Morris"  wrote:
> >
> > Thanks for the reminder. So that solves the "asking" part.
> >
> > Does anyone *not* think that the Wikidata engineering team is the
> correct place for this?
> >
> > Lydia - can you assign someone to come up to speed at whatever level
> Denny requires to feel comfortable making the transfer?
>
> I will take care of it with Denny in the next days.
>

FYI: In-progress now. I'll report back as soon as it is done but it'll take
a few days to clarify some stuff.


Cheers
Lydia
-- 
Lydia Pintscher - http://about.me/lydia.pintscher
Product Manager for Wikidata

Wikimedia Deutschland e.V.
Tempelhofer Ufer 23-24
10963 Berlin
www.wikimedia.de

Wikimedia Deutschland - Gesellschaft zur Förderung Freien Wissens e. V.

Eingetragen im Vereinsregister des Amtsgerichts Berlin-Charlottenburg unter
der Nummer 23855 Nz. Als gemeinnützig anerkannt durch das Finanzamt für
Körperschaften I Berlin, Steuernummer 27/029/42207.
___
Wikidata mailing list
Wikidata@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikidata


Re: [Wikidata] [ANNOUNCEMENT] StrepHit 1.0 Beta Release

2016-06-15 Thread Benjamin Good
Hi Marco,

Where might we find some statistics on the current accuracy of the
automated claim and reference extractors?  I assume that information must
be in there somewhere, but I had trouble finding it.

This is a very ambitious project covering a very large technical territory
(which I applaud).  It would be great if your results could be synthesized
a bit more clearly so we can understand where the weak/strong points are
and where we might be able to help improve or make use of what you have
done in other domains.

-Ben


On Wed, Jun 15, 2016 at 9:06 AM, Marco Fossati 
wrote:

> [Feel free to blame me if you read this more than once]
>
> To whom it may interest,
>
> Full of delight, I would like to announce the first beta release of
> *StrepHit*:
>
> https://github.com/Wikidata/StrepHit
>
> TL;DR: StrepHit is an intelligent reading agent that understands text and
> translates it into *referenced* Wikidata statements.
> It is a IEG project funded by the Wikimedia Foundation.
>
> Key features:
> -Web spiders to harvest a collection of documents (corpus) from reliable
> sources
> -automatic corpus analysis to understand the most meaningful verbs
> -sentences and semi-structured data extraction
> -train a machine learning classifier via crowdsourcing
> -*supervised and rule-based fact extraction from text*
> -Natural Language Processing utilities
> -parallel processing
>
> You can find all the details here:
>
> https://meta.wikimedia.org/wiki/Grants:IEG/StrepHit:_Wikidata_Statements_Validation_via_References
>
> https://meta.wikimedia.org/wiki/Grants:IEG/StrepHit:_Wikidata_Statements_Validation_via_References/Midpoint
>
> If you like it, star it on GitHub!
>
> Best,
>
> Marco
>
> ___
> Wikidata mailing list
> Wikidata@lists.wikimedia.org
> https://lists.wikimedia.org/mailman/listinfo/wikidata
>
___
Wikidata mailing list
Wikidata@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikidata


[Wikidata] [ANNOUNCEMENT] StrepHit 1.0 Beta Release

2016-06-15 Thread Marco Fossati

[Feel free to blame me if you read this more than once]

To whom it may interest,

Full of delight, I would like to announce the first beta release of 
*StrepHit*:


https://github.com/Wikidata/StrepHit

TL;DR: StrepHit is an intelligent reading agent that understands text 
and translates it into *referenced* Wikidata statements.

It is a IEG project funded by the Wikimedia Foundation.

Key features:
-Web spiders to harvest a collection of documents (corpus) from reliable 
sources

-automatic corpus analysis to understand the most meaningful verbs
-sentences and semi-structured data extraction
-train a machine learning classifier via crowdsourcing
-*supervised and rule-based fact extraction from text*
-Natural Language Processing utilities
-parallel processing

You can find all the details here:
https://meta.wikimedia.org/wiki/Grants:IEG/StrepHit:_Wikidata_Statements_Validation_via_References
https://meta.wikimedia.org/wiki/Grants:IEG/StrepHit:_Wikidata_Statements_Validation_via_References/Midpoint

If you like it, star it on GitHub!

Best,

Marco

___
Wikidata mailing list
Wikidata@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikidata


Re: [Wikidata] RFC - Primary Sources?

2016-06-15 Thread Marco Fossati

Hi Lydia,

On 6/15/16 07:42, Lydia Pintscher wrote:

Lydia - can you assign someone to come up to speed at whatever level

Denny requires to feel comfortable making the transfer?

I will take care of it with Denny in the next days.


Repasting part of a previous message with the list of requirements:

A. a developer to understand the back-end code [1], written in C++;
B. a developer to understand the front-end code [2], written in Javascript;
C. access to the WMF Labs machine to deploy the back-end [3];
D. a Wikidata administrator to deploy the front-end [4];
E. centralized and exhaustive documentation.

As part of the StrepHit project goals [5], my team is striving to help 
with A. (not exactly trivial) and C., but we really need B. and D. to be 
effective.


Best,

Marco

[1] https://github.com/google/primarysources/tree/master/backend
[2] https://github.com/google/primarysources/tree/master/frontend
[3] https://tools.wmflabs.org/wikidata-primary-sources
[4] 
https://github.com/google/primarysources/tree/master/frontend#deployment-on-wikidata
[5] 
https://meta.wikimedia.org/wiki/Grants:IEG/StrepHit:_Wikidata_Statements_Validation_via_References#Project_Goals


___
Wikidata mailing list
Wikidata@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikidata


Re: [Wikidata] Genes, proteins, and bad merges in general

2016-06-15 Thread Lydia Pintscher
On Jun 15, 2016 10:52, "Magnus Manske"  wrote:
>
>
>
> On Tue, Jun 14, 2016 at 6:54 PM Tom Morris  wrote:
>>
>> Bad merges have been mentioned a couple of times recently and I think
one of the contexts with Ben's gene/protein work.
>>
>> I think there are two general issues here which could be improved:
>>
>> 1. Merging is too easy. Because splitting/unmerging is much harder than
merging, particularly after additional edits, the process should be biased
to mark merging more difficult.
>>
> A technical solution could be to prevent merging of two items (or at
least, show a warning), if one of the items links to the other.

That is already happening.

Cheers
Lydia

>
> ___
> Wikidata mailing list
> Wikidata@lists.wikimedia.org
> https://lists.wikimedia.org/mailman/listinfo/wikidata
>
___
Wikidata mailing list
Wikidata@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikidata


Re: [Wikidata] RFC - Primary Sources?

2016-06-15 Thread Marco Fossati

Hi Tom,

On 6/14/16 19:26, Tom Morris wrote:

Marco - Centralizing the discussion is good, but why not pick one of the
three existing channels (issue tracker, project page, this mailing list)
rather than creating a fourth channel?
The RFC is meant to put together low-level technical problems (issue 
tracker), usability discussions (project page), less structured 
discussions (mailing list).

*And* comments on the uploaded datasets.

As much as I love playing and
watching soccer, I'm much more interested in the vast trove of
identifiers and other curated information in Freebase than I am in
improving Wikidata's soccer coverage, but the Primary Sources tool could
be useful for some portions of the Freebase data, if it could be usable.
I guess you are referring to the StrepHit prototype dataset 
'strephit-soccer'.
Why don't you try the 'strephit-testing' one? It deals with biographies 
and has much broader coverage.


Best,

Marco

___
Wikidata mailing list
Wikidata@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikidata


Re: [Wikidata] Genes, proteins, and bad merges in general

2016-06-15 Thread Magnus Manske
On Tue, Jun 14, 2016 at 6:54 PM Tom Morris  wrote:

> Bad merges have been mentioned a couple of times recently and I think one
> of the contexts with Ben's gene/protein work.
>
> I think there are two general issues here which could be improved:
>
> 1. Merging is too easy. Because splitting/unmerging is much harder than
> merging, particularly after additional edits, the process should be biased
> to mark merging more difficult.
>
> A technical solution could be to prevent merging of two items (or at
least, show a warning), if one of the items links to the other.
___
Wikidata mailing list
Wikidata@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikidata