Re: [Wikidata-l] Auto-transliteration of names of humans

2015-05-03 Thread Jo
Once IPA is there it may be easier to provide Text-to-speech automatically.

2015-05-02 16:57 GMT+02:00 Thomas Douillard :

> I'll all with you, on this, unfortunately only a few people reads IPA
> currently :(. I don't, for example.
>
> 2015-05-02 15:19 GMT+02:00 Gerard Meijssen :
>
>> Hoi,
>> When the point is to express how an official name is to be pronounced,
>> IPA is in order not a text in another script.
>> Thanks,
>>  GerardM
>>
>> On 1 May 2015 at 11:04, Bene*  wrote:
>>
>>>  Hi,
>>>
>>> this is what the monolingual text datatype is for. The labels however
>>> are multilingual and should provide users in all languages an idea how the
>>> name is said.
>>>
>>> Best regards
>>> Bene
>>>
>>>
>>> Am 01.05.2015 um 07:14 schrieb Gerard Meijssen:
>>>
>>> Hoi,
>>> It is still a bad idea. An official name exists only in one language.
>>> Thanks,
>>>  GerardM
>>>
>>> On 30 April 2015 at 18:50, Thomas Douillard 
>>> wrote:
>>>
  I meant "add automatically the transliteration", not replace the name.

  This is a good candidate : we know for sure the source and the target
 language (the one of the user) so a good choice for transliteration method
 is always possible, and we don't pretend it should be the way to say orally
 the name in the target language. It's just a transliteration of the
 official name.



 2015-04-30 15:14 GMT+02:00 Gerard Meijssen :

>   Hoi,
>  It does not quality anything. It is plain wrong.
>  Thanks,
>   GerardM
>
> On 30 April 2015 at 15:06, Joe Filceolaire 
> wrote:
>
>> Exactly. The "official name " property always has the name in the
>> original script. But we can and should have the transliteration in a
>> qualifier.
>>
>> Joe
>>   On 30 Apr 2015 06:13, "Gerard Meijssen" 
>> wrote:
>>
>>>   Hoi,
>>>  We transliterate every name from one script to the other.
>>> Transliteration the official name is exactly the one you should not
>>> transliterate.. What is left after transliteration is not official.
>>>  Thanks,
>>>GerardM
>>>
>>> On 29 April 2015 at 18:54, Thomas Douillard <
>>> thomas.douill...@gmail.com> wrote:
>>>
  It's always possible to transliterate the official name
 property. Of course
 this should be done by a gadget, or we may have to find a special 
 treatment
 for the ''name'' properties.

 2015-04-28 23:06 GMT+02:00 Joe Filceolaire :

> I agree up to a point. Transliteration is not appropriate for
> labels for all items.  There are however a few categories of items for
> which transliterated labels are appropriate. For example :
> * English labels for villages and towns
> * English labels for people
> *English labels for bands and albums
> I'm sure there are  others that could use this too.
>
> Joe
>   On 27 Apr 2015 18:09, "Leon Liesener" <
> leon.liese...@wikipedia.de> wrote:
>
>> The problem with ISO is that it's a standard for
>> language-independent
>> transliteration to Latin script. Since labels on Wikidata are
>> language-dependent, making use of ISO does not make sense really.
>> If
>> you use ISO for Russian names in Cyrillic script, the label you
>> get is
>> not in English. It's still in Russian but transliterated to Latin
>> script. ISO thus would only fit as an alias for the Russian
>> interface
>> language, if at all.
>>
>> 2015-04-26 22:39 GMT+02:00 Gerard Meijssen <
>> gerard.meijs...@gmail.com>:
>> > Hoi,
>> >  ISO is a reliable source; it is THE standard 
>> Wikipedia is
>> > definitely not a standard by its own admission.
>> > Thanks,
>> > GerardM
>> >
>> > On 26 April 2015 at 22:37, Yaroslav M. Blanter <
>> pute...@mccme.ru> wrote:
>> >>
>> >> On 2015-04-26 22:33, Gerard Meijssen wrote:
>> >>>
>> >>> Hoi
>> >>> My point is that it is not a given that we should follow any
>> WIkipedia
>> >>> for anything. Also the point of romanisation of Russian is
>> not for the
>> >>> benefit of Russian speakers, it is for the speakers of
>> English.
>> >>> Thanks,
>> >>>   GerardM
>> >>>
>> >>
>> >> On one hand, yes.
>> >>
>> >> On the other hand, no reliable source uses ISO. When NYT
>> writes about a
>> >> Russian person, they do not use ISO, they use what the English
>> Wikipedia
>> >> uses or smth similar. In my passport, they do not use ISO
>> (fortunately), why
>> >> should then ISO be used on Wikidata in an entry

Re: [Wikidata-l] Auto-transliteration of names of humans

2015-05-03 Thread Gerard Meijssen
Hoi,
It is a Wiktionary thing  :( and yes we can and yes we should already
do this/
Thanks,
   GerardM

On 3 May 2015 at 10:43, Jo  wrote:

> Once IPA is there it may be easier to provide Text-to-speech automatically.
>
> 2015-05-02 16:57 GMT+02:00 Thomas Douillard :
>
>> I'll all with you, on this, unfortunately only a few people reads IPA
>> currently :(. I don't, for example.
>>
>> 2015-05-02 15:19 GMT+02:00 Gerard Meijssen :
>>
>>> Hoi,
>>> When the point is to express how an official name is to be pronounced,
>>> IPA is in order not a text in another script.
>>> Thanks,
>>>  GerardM
>>>
>>> On 1 May 2015 at 11:04, Bene*  wrote:
>>>
  Hi,

 this is what the monolingual text datatype is for. The labels however
 are multilingual and should provide users in all languages an idea how the
 name is said.

 Best regards
 Bene


 Am 01.05.2015 um 07:14 schrieb Gerard Meijssen:

 Hoi,
 It is still a bad idea. An official name exists only in one language.
 Thanks,
  GerardM

 On 30 April 2015 at 18:50, Thomas Douillard >>> > wrote:

>  I meant "add automatically the transliteration", not replace the
> name.
>
>  This is a good candidate : we know for sure the source and the target
> language (the one of the user) so a good choice for transliteration method
> is always possible, and we don't pretend it should be the way to say 
> orally
> the name in the target language. It's just a transliteration of the
> official name.
>
>
>
> 2015-04-30 15:14 GMT+02:00 Gerard Meijssen 
> :
>
>>   Hoi,
>>  It does not quality anything. It is plain wrong.
>>  Thanks,
>>   GerardM
>>
>> On 30 April 2015 at 15:06, Joe Filceolaire 
>> wrote:
>>
>>> Exactly. The "official name " property always has the name in the
>>> original script. But we can and should have the transliteration in a
>>> qualifier.
>>>
>>> Joe
>>>   On 30 Apr 2015 06:13, "Gerard Meijssen" 
>>> wrote:
>>>
   Hoi,
  We transliterate every name from one script to the other.
 Transliteration the official name is exactly the one you should not
 transliterate.. What is left after transliteration is not official.
  Thanks,
GerardM

 On 29 April 2015 at 18:54, Thomas Douillard <
 thomas.douill...@gmail.com> wrote:

>  It's always possible to transliterate the official name
> property. Of course
> this should be done by a gadget, or we may have to find a special 
> treatment
> for the ''name'' properties.
>
> 2015-04-28 23:06 GMT+02:00 Joe Filceolaire 
> :
>
>> I agree up to a point. Transliteration is not appropriate for
>> labels for all items.  There are however a few categories of items 
>> for
>> which transliterated labels are appropriate. For example :
>> * English labels for villages and towns
>> * English labels for people
>> *English labels for bands and albums
>> I'm sure there are  others that could use this too.
>>
>> Joe
>>   On 27 Apr 2015 18:09, "Leon Liesener" <
>> leon.liese...@wikipedia.de> wrote:
>>
>>> The problem with ISO is that it's a standard for
>>> language-independent
>>> transliteration to Latin script. Since labels on Wikidata are
>>> language-dependent, making use of ISO does not make sense
>>> really. If
>>> you use ISO for Russian names in Cyrillic script, the label you
>>> get is
>>> not in English. It's still in Russian but transliterated to Latin
>>> script. ISO thus would only fit as an alias for the Russian
>>> interface
>>> language, if at all.
>>>
>>> 2015-04-26 22:39 GMT+02:00 Gerard Meijssen <
>>> gerard.meijs...@gmail.com>:
>>> > Hoi,
>>> >  ISO is a reliable source; it is THE standard 
>>> Wikipedia is
>>> > definitely not a standard by its own admission.
>>> > Thanks,
>>> > GerardM
>>> >
>>> > On 26 April 2015 at 22:37, Yaroslav M. Blanter <
>>> pute...@mccme.ru> wrote:
>>> >>
>>> >> On 2015-04-26 22:33, Gerard Meijssen wrote:
>>> >>>
>>> >>> Hoi
>>> >>> My point is that it is not a given that we should follow any
>>> WIkipedia
>>> >>> for anything. Also the point of romanisation of Russian is
>>> not for the
>>> >>> benefit of Russian speakers, it is for the speakers of
>>> English.
>>> >>> Thanks,
>>> >>>   GerardM
>>> >>>
>>> >>
>>> >> On one hand, yes.
>>> >>
>>> >> On the other hand, n

[Wikidata-l] weekly summary #156

2015-05-03 Thread Lydia Pintscher
Hey folks :)

Here's your summary of what happened around Wikidata over the last week.

Events /Press/Blogs


   - Wikidata was presented at a Swedish Linked Data Network Meet-up
   

   in Gothenburg.
   - The most important Wikipedia pages
   


Other Noteworthy Stuff

   - P107 (main type)
   

   is finally orphaned and deleted, 20 months after deprecation. This was the
   first property used 100 times.
   - The royal baby  was quickly
   updated on Wikidata. (her family tree
   )
   - New tool by Magnus: Find pictures on geography.co.uk, upload them to
   Commons and add them to Wikidata
   
   - All Van Gogh Museum paintings are now on Commons and Wikidata. Go go
   SumOfAllPaintings peeps!
   - List of topics with links in at least X Wikipedias
   
   - All French Senators matched using MixNMatch
   
   \o/

Did you know?

   - Newest properties: represented by
   , Netflix Identifier
   , maximum number of players
   , minimum number of players
   , CERL ID
   , Name Assigning Authority
   Number , Hall of Valor ID
   , ballots cast
   , eligible voters
   , catholic-hierarchy
   diocese ID , Wikidata
   example geographic coordinates
   , Wikidata example
   monolingual text , Wikidata
   example property , Wikidata
   example quantity , Wikidata
   example time , Wikidata
   example URL , Wikidata
   example item value , Wikidata
   example string , Wikidata
   example media file , Wikidata
   property example 

Development

   - We rolled out usage tracking on the first two wikis (French Wikisource
   and Dutch Wikipedia). Users should not notice anything. More wikis will
   follow in the next weeks. This is the remaining step for enabling arbitrary
   access on wikis other than Commons.
   - The students team is working hard to get a first release of the
   improved constraint reports and checks against 3rd party databases out.
   - Ricordisamoa fixed the issue with long descriptions being cut off.
   - We fixed the focus flow in the property selector.
   - We improved the messages on Special:EntityData to make it more
   understandable.

You can see all open bugs related to Wikidata here
.
Monthly Tasks

   - Hack on one of these
   .
   - Help fix these items
    which
   have been flagged using Wikidata - The Game.
   - Help develop the next summary here!
   
   - Contribute to a Showcase item
   
   - Help translate 
   or proofread pages in your own language!
   - Add labels, in your own language, for the new properties listed above.

Anything to add? Please share! :)

Cheers
Lydia

-- 
Lydia Pintscher - http://about.me/lydia.pintscher
Product Manager for Wikidata

Wikimedia Deutschland e.V.
Tempelhofer Ufer 23-24
10963 Berlin
www.wikimedia.de

Wikimedia Deutschland - Gesellschaft zur Förderung Freien Wissens e. V.

Eingetragen im Vereinsregister des Amtsgerichts Berlin-Charlottenburg unter
der Nummer 23855 Nz. Als gemeinnützig anerkannt durch das Finanzamt für
Körperschaften I Berlin, Steuernummer 27/681/51985.
___

Re: [Wikidata-l] Making queries on Wikibase

2015-05-03 Thread Luca Martinelli
Sorry, all the mails were lost in the mare magnum of my full inbox,
and I forgot to thank you all for your answers. :)

L.

2015-04-30 12:07 GMT+02:00 Markus Krötzsch :
> On 30.04.2015 11:25, Jean-Baptiste Pressac wrote:
>>
>> Hello,
>> Does this also means that the RDF data available via for instance
>> http://www.wikidata.org/entity/Q235382.nt or
>> http://www.wikidata.org/entity/Q235382 could not be queried via SPARQL
>> unless you download the .nt file and use for instance Jena ARQ
>>  on your own
>> computer ?
>
>
> Maybe not your computer, but somebody has to download the data in some way
> in order for it to be queried. Query answering is a complicated process that
> does not usually work on the fly, and somebody has to do the math in the
> end. There are some approaches to explore linked data in a query-like manner
> in real time, but it should be clear that this will always take much longer
> than if you have downloaded the data first.
>
>>
>> Does this also means that there is no use to publish RDF data linking to
>> Wikidata like for instance :
>>
>> @prefix mydata:  .
>> @prefix cidoc:  .
>> @prefix wikidata: 
>>
>> mydata:event/1 a cidoc:E67_Birth ;
>> cidoc:P98_brought_into_life mydata:person/80 ;
>> cidoc:P7_took_place_at wikidata:Q235382 ;
>
>
> This is still a useful thing to do for several reasons. First of all, the
> link connects your data and clarifies its meaning. This is useful to
> consumers who find your data. Second, there are linked data crawlers that
> aggregate linked data from many sources to provide you with a query service.
> OpenLink is running one such service, and if they managed to find your data,
> you could use their service to issue queries.
>
> Regards,
>
> Markus
>
>
>>
>> Thanks,
>>
>>
>> Jean-Baptiste Pressac
>>
>> Traitement et analyse de bases de données
>> Production et diffusion de corpus numériques
>>
>> Centre de Recherche Bretonne et Celtique
>> Unité mixte de service (UMS) 3554
>> 20 rue Duquesne
>> CS 93837
>> 29238 Brest cedex 3
>>
>> tel : +33 (0)2 98 01 68 95
>> fax : +33 (0)2 98 01 63 93
>>
>> Le 29/04/2015 21:44, Markus Krötzsch a écrit :
>>>
>>> On 29.04.2015 20:56, Luca Martinelli wrote:

 Dear all,

 I need to know about the possibility of making queries on a Wikibase
 instance. I think it is possible to make queries on data on a
 particular instance only with external tools at the moment, right?
>>>
>>>
>>> Yes, this is correct. The SPARQL query support that we currently offer
>>> is obtained by making an RDF export and loading it into a SPARQL
>>> database (we use Virtuoso but you could also use BlazeGraph, for
>>> example; both have free and open source versions and are not hard to
>>> install overall; if your data is not so large, you could also try
>>> Jena; there are further open source RDF databases, but these are the
>>> most prominent right now I think).
>>>
>>> The RDF export, too, is not currently generated by Wikibase. However,
>>> Wikidata Toolkit, which we use to make the RDF dumps, can be used with
>>> data from any Wikibase installation in theory. In practice, nobody has
>>> asked for this yet and we might have to make a few adjustments to
>>> really get it to work in a convenient way. For a start, I don't know
>>> what kind of export options a standalone Wikibase offers you at the
>>> moment. We can use the usual XML-based page dump if it contains valid
>>> JSON for a change (this was not the case for Wikidata last time I
>>> checked ...). Better yet would be the JSON exports, but I don't know
>>> if you can generate them with vanilla Wikibase or if WMF is using some
>>> special tools for this.
>>>
>>> Anyway, it can't be too hard to make this work and once it is done you
>>> would have a query service that is at the same level as the one of
>>> Wikidata. You could even combine data from more than one Wiki in one
>>> RDF database, e.g., to run queries over data from both.
>>>
>>> Regards,
>>>
>>> Markus
>>>
>>>
>>> ___
>>> Wikidata-l mailing list
>>> Wikidata-l@lists.wikimedia.org
>>> https://lists.wikimedia.org/mailman/listinfo/wikidata-l
>>
>>
>>
>>
>> ___
>> Wikidata-l mailing list
>> Wikidata-l@lists.wikimedia.org
>> https://lists.wikimedia.org/mailman/listinfo/wikidata-l
>>
>
>
> ___
> Wikidata-l mailing list
> Wikidata-l@lists.wikimedia.org
> https://lists.wikimedia.org/mailman/listinfo/wikidata-l



-- 
Luca "Sannita" Martinelli
http://it.wikipedia.org/wiki/Utente:Sannita

___
Wikidata-l mailing list
Wikidata-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikidata-l