Re: [Wikidata] Wikidata - short biographies

2016-01-31 Thread Edgard Marx
Yep,

One more reason to use RDFSlice ;-),

thnks

On Mon, Feb 1, 2016 at 7:25 AM, Stas Malyshev 
wrote:

> Hi!
>
> >
> > ** For all English bios:*
> >
> > SELECT *
> > WHERE
> > {
> >?s  ?o .
> >filter(lang(?o)='en').
> > }
>
> Please don't run this on query.wikidata.org though. Please add LIMIT.
> Otherwise you'd be trying to download several millions of data items,
> which would probably time out anyway. Add something like "LIMIT 10" to it.
>
> Thanks,
> --
> Stas Malyshev
> smalys...@wikimedia.org
>
> ___
> Wikidata mailing list
> Wikidata@lists.wikimedia.org
> https://lists.wikimedia.org/mailman/listinfo/wikidata
>
>
___
Wikidata mailing list
Wikidata@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikidata


Re: [Wikidata] Wikidata - short biographies

2016-01-31 Thread Stas Malyshev
Hi!

> 
> ** For all English bios:*
> 
> SELECT *
> WHERE
> {
>?s  ?o .   
>filter(lang(?o)='en').
> }

Please don't run this on query.wikidata.org though. Please add LIMIT.
Otherwise you'd be trying to download several millions of data items,
which would probably time out anyway. Add something like "LIMIT 10" to it.

Thanks,
-- 
Stas Malyshev
smalys...@wikimedia.org

___
Wikidata mailing list
Wikidata@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikidata


Re: [Wikidata] Wikidata - short biographies

2016-01-31 Thread Hampton Snowball
Thanks. I see it requires constructing a query to only extract the data you
want. E.g. the graph pattern:

 - desired query, e.g. "SELECT * WHERE {?s ?p ?o}" or graph
pattern e.g. "{?s ?p ?o}"

Since I don't know about constructing queries, would you be able to tell me
what would be the proper query to extract from all the pages the short bio,
english wikipedia, maybe other wikipedias?

For example from: https://www.wikidata.org/wiki/Q1652291";

"Turkish female given name"
https://en.wikipedia.org/wiki/H%C3%BClya
and optionally https://de.wikipedia.org/wiki/H%C3%BClya

Thanks in advance!


On Sun, Jan 31, 2016 at 3:53 PM, Edgard Marx  wrote:

> Hey,
> you can simple use RDFSlice (https://bitbucket.org/emarx/rdfslice/overview)
> directly on the dump file (
> https://dumps.wikimedia.org/wikidatawiki/entities/20160125/)
>
> best,
> Edgard
>
> On Sun, Jan 31, 2016 at 7:43 PM, Hampton Snowball <
> hamptonsnowb...@gmail.com> wrote:
>
>> Hello,
>>
>> I am interested in a subset of wikidata and I am trying to find the best
>> way to get it without getting a larger dataset then necessary.
>>
>> Is there a way to just get the "bios" that appear on the wikidata pages
>> below the name of the person/organization, as well as the link to the
>> english wikipedia page / or all wikipedia pages?
>>
>> For example from: https://www.wikidata.org/wiki/Q1652291";
>>
>> "Turkish female given name"
>> https://en.wikipedia.org/wiki/H%C3%BClya
>> and optionally https://de.wikipedia.org/wiki/H%C3%BClya
>>
>> I know there is SPARQL which previously this list helped me construct a
>> query, but I know some requests seem to timeout when looking at a large
>> amount of data so I am not sure this would work.
>>
>> The dumps I know are the full dataset, but I am not sure if there's any
>> other subset dumps available or better way of grabbing this data
>>
>> Thanks in advance,
>> HS
>>
>>
>> ___
>> Wikidata mailing list
>> Wikidata@lists.wikimedia.org
>> https://lists.wikimedia.org/mailman/listinfo/wikidata
>>
>>
>
> ___
> Wikidata mailing list
> Wikidata@lists.wikimedia.org
> https://lists.wikimedia.org/mailman/listinfo/wikidata
>
>
___
Wikidata mailing list
Wikidata@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikidata


Re: [Wikidata] Internal server error after edit / create

2016-01-31 Thread Info WorldUniversity
Bonjour Wikidatans and Jeremy,

Thanks for information about how to install Wikidata.  WUaS is seeking to
install Wikidata soon.

Wikidatans, is there a way to integrate Google's new AI software
"TensorFlow" with Wikidata? CC WUaS is planning to be in a
Wikipedia/Wikidata's ~300 languages, and after this in all 7,938+ languages
as CC wiki schools for open teaching and learning. (See the beginnings of
WUaS MediaWiki in English and German and the main WUAS SUBECT TEMPLATE
which will inform most pages at WUaS here -
http://scott-macleod.blogspot.com/2016/01/vicia-faba-wikidata-installation-re.html).
See too the upcoming free course on Google TensorFlow from Udacity -
https://www.udacity.com/course/deep-learning--ud730 and
http://googleresearch.blogspot.com/2016/01/teach-yourself-deep-learning-with.html
.

Thanks, Scott

http://worlduniversityandschool.org


Hello everyone,

I'm having some issues with my wikidata installation, any help would be
appreciated.

I'm getting stuck on an "Internal server error" after a Mediawiki +
Wikibase repo installation via GIT + update of wikibase via composer, the
wikibase is functional but after adding 1 element or 1 property, any other
add of elements returns an Internal server error, and I have to wait for
several minutes before it's fine again, for only 1 addition of element and
then error again.

I have no particular errors in my php logs or apache logs, just - error 500.

Here are the versions installed :

MediaWiki   1.27alpha (b3a82ca)
PHP 5.4.45-0+deb7u2 (fpm-fcgi)
MySQL   5.5.46-0+deb7u1-log
ICU 4.8.1.1
Wikibase Repository 0.5 alpha (62c267)

I can add some wiki pages (non-wikibase stuff), modify them, no problems.

When I try to add an element or property, the first time I try it works,
then a few seconds later I can't add a second element I get an "error 500 -
internal server error", and I can't get it work until about 30 minutes / 1
hour then I can add another element and then again nothing !

I think I might have a namespace problem but how to configure it and keep
my other wiki pages ?

Thanks in advance !

Regards,
Jeremy
___
Wikidata mailing list
Wikidata@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikidata
___
Wikidata mailing list
Wikidata@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikidata


Re: [Wikidata] Maintenance scripts for clients

2016-01-31 Thread Federico Leva (Nemo)

John Erling Blad, 06/08/2015 17:13:

A couple of guys at nowiki has started using this tool, and if they
continue at present speed the list will be emptied in two weeks time.

Can you please add nnwiki too and I will inform the community there
that there is a tool available.


4941 items in list, looks like that didn't happen. How comes bots are 
slacking? I thought they cared a lot about creating new Wikidata items 
for new articles!


Nemo



On Thu, Aug 6, 2015 at 10:55 AM, Magnus Manske
 wrote:

John: List mode!

https://tools.wmflabs.org/wikidata-todo/duplicity.php?wiki=nowiki&mode=list

On Thu, Aug 6, 2015 at 8:16 AM Zolo ...  wrote:


About missing labels: in frwiki, most Wikidata data are added using
Module:Wikidata. The module adds a generic "to be translated" category when
there is no French label. With Wikidata usage picking up speed, the
community is finally coming to grip with it, as can be seen from that stats
at Catégorie:Page utilisant des données de Wikidata à traduire.

On Tue, Aug 4, 2015 at 7:06 PM, John Erling Blad  wrote:


Nice solution, I'll post a link at Wikipedia:Torget.
It is a bit like making a traffic statistic by using a road cam, so it
wasn't really what I was looking for..

On Tue, Aug 4, 2015 at 5:18 PM, Magnus Manske
 wrote:

I set up one of my tools for you (nowiki) for [1] :
https://tools.wmflabs.org/wikidata-todo/duplicity.php

It doesn't give you a list (though I could add that), rather presents
you
with a random one and tries to find a matching item. Basically, what
you
need to do anyway for due diligence.


Not quite sure what else you need, too much "somehow" in your
description...


On Tue, Aug 4, 2015 at 4:01 PM John Erling Blad 
wrote:


We lack several maintenance scripts for the clients, that is human
readable special pages with reports on which pages lacks special
treatment. In no particular order we need some way to identify
unconnected pages in general (the present one does not work [1]), we
need some way to identify pages that are unconnected but has some
language links, we need to identify items that are used in some
language and lacks labels (almost like [2],but on the client and for
items that are somehow connected to pages on the client), and we need
to identify items that lacks specific claims and the client pages use
a specific template.

There are probably more such maintenance pages, these are those that
are most urgent. Now users start to create categories to hack around
the missing maintenance pages, which create a bunch of categories.[3]
At Norwegian Bokmål there are just a few scripts that utilize data
from Wikidata, still the number of categories starts to grow large.

For us at the "receiving end" this is a show stopper. We can't
convince the users that this is a positive addition to the pages
without the maintenance scripts, because them we more or less are in
the blind when we try to fix errors. We can't use random pages to try
to prod the pages to find something that is wrong, we must be able to
search for the errors and fix them.

This summer we (nowiki) have added about ten (10) properties to the
infobokses, some with scripts and some with the property parser
function. Most of my time I have not been coding, and I have not been
fixing errors. I have been trying to explain to the community why
Wikidata is a good idea. At one point the changes was even reverted
because someone disagree with what we had done. The whole thing
basically revolves around "my article got an Q-id in the infobox and I
don't know how to fix it". We know how to fix it, and I have explained
that to the editors at nowiki several times. They still don't get it,
so we need some way to fix it, and we don't have maintenance scripts
to do it.

Right now we don't need more wild ideas that will swamp the
development for months and years to come, we need maintenance scripts,
and we need them now!

[1] https://no.wikipedia.org/wiki/Spesial:UnconnectedPages
[2] https://www.wikidata.org/wiki/Special:EntitiesWithoutLabel
[3]

https://no.wikipedia.org/wiki/Spesial:Prefiksindeks/Kategori:Artikler_hvor

John Erling Blad
/jeblad

___
Wikidata mailing list
Wikidata@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikidata



___
Wikidata mailing list
Wikidata@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikidata



___
Wikidata mailing list
Wikidata@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikidata



___
Wikidata mailing list
Wikidata@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikidata



___
Wikidata mailing list
Wikidata@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikidata



___
Wikidata mailing list
Wikidata@lists.wiki

Re: [Wikidata] Goal: Establish a framework to engage with data engineers and open data organizations

2016-01-31 Thread Federico Leva (Nemo)

Raphaël Troncy, 04/07/2015 15:55:

Are you aware of the http://wikirate.org/ community? This is a EU
research funded project.


Doesn't seem to contain any objective data, am I mistaken?

Nemo

___
Wikidata mailing list
Wikidata@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikidata


Re: [Wikidata] Wikidata - short biographies

2016-01-31 Thread Edgard Marx
Hey,
you can simple use RDFSlice (https://bitbucket.org/emarx/rdfslice/overview)
directly on the dump file (
https://dumps.wikimedia.org/wikidatawiki/entities/20160125/)

best,
Edgard

On Sun, Jan 31, 2016 at 7:43 PM, Hampton Snowball  wrote:

> Hello,
>
> I am interested in a subset of wikidata and I am trying to find the best
> way to get it without getting a larger dataset then necessary.
>
> Is there a way to just get the "bios" that appear on the wikidata pages
> below the name of the person/organization, as well as the link to the
> english wikipedia page / or all wikipedia pages?
>
> For example from: https://www.wikidata.org/wiki/Q1652291";
>
> "Turkish female given name"
> https://en.wikipedia.org/wiki/H%C3%BClya
> and optionally https://de.wikipedia.org/wiki/H%C3%BClya
>
> I know there is SPARQL which previously this list helped me construct a
> query, but I know some requests seem to timeout when looking at a large
> amount of data so I am not sure this would work.
>
> The dumps I know are the full dataset, but I am not sure if there's any
> other subset dumps available or better way of grabbing this data
>
> Thanks in advance,
> HS
>
>
> ___
> Wikidata mailing list
> Wikidata@lists.wikimedia.org
> https://lists.wikimedia.org/mailman/listinfo/wikidata
>
>
___
Wikidata mailing list
Wikidata@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikidata


Re: [Wikidata] Wikidata - short biographies

2016-01-31 Thread Gerard Meijssen
Hoi,
Magnus created automated descriptions they are a start. Your only problem
is that they are not using sparql
Thanks,
 GerardM

On 31 January 2016 at 19:43, Hampton Snowball 
wrote:

> Hello,
>
> I am interested in a subset of wikidata and I am trying to find the best
> way to get it without getting a larger dataset then necessary.
>
> Is there a way to just get the "bios" that appear on the wikidata pages
> below the name of the person/organization, as well as the link to the
> english wikipedia page / or all wikipedia pages?
>
> For example from: https://www.wikidata.org/wiki/Q1652291";
>
> "Turkish female given name"
> https://en.wikipedia.org/wiki/H%C3%BClya
> and optionally https://de.wikipedia.org/wiki/H%C3%BClya
>
> I know there is SPARQL which previously this list helped me construct a
> query, but I know some requests seem to timeout when looking at a large
> amount of data so I am not sure this would work.
>
> The dumps I know are the full dataset, but I am not sure if there's any
> other subset dumps available or better way of grabbing this data
>
> Thanks in advance,
> HS
>
>
> ___
> Wikidata mailing list
> Wikidata@lists.wikimedia.org
> https://lists.wikimedia.org/mailman/listinfo/wikidata
>
>
___
Wikidata mailing list
Wikidata@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikidata


[Wikidata] Wikidata - short biographies

2016-01-31 Thread Hampton Snowball
Hello,

I am interested in a subset of wikidata and I am trying to find the best
way to get it without getting a larger dataset then necessary.

Is there a way to just get the "bios" that appear on the wikidata pages
below the name of the person/organization, as well as the link to the
english wikipedia page / or all wikipedia pages?

For example from: https://www.wikidata.org/wiki/Q1652291";

"Turkish female given name"
https://en.wikipedia.org/wiki/H%C3%BClya
and optionally https://de.wikipedia.org/wiki/H%C3%BClya

I know there is SPARQL which previously this list helped me construct a
query, but I know some requests seem to timeout when looking at a large
amount of data so I am not sure this would work.

The dumps I know are the full dataset, but I am not sure if there's any
other subset dumps available or better way of grabbing this data

Thanks in advance,
HS
___
Wikidata mailing list
Wikidata@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikidata


Re: [Wikidata] Internal server error after edit / create

2016-01-31 Thread Lydia Pintscher
On Jan 24, 2016 22:05, "Jérémy Kondi"  wrote:
>
> Hello everyone,
>
> I'm having some issues with my wikidata installation, any help would be
appreciated.
>
> I'm getting stuck on an "Internal server error" after a Mediawiki +
Wikibase repo installation via GIT + update of wikibase via composer, the
wikibase is functional but after adding 1 element or 1 property, any other
add of elements returns an Internal server error, and I have to wait for
several minutes before it's fine again, for only 1 addition of element and
then error again.
>
> I have no particular errors in my php logs or apache logs, just - error
500.
>
> Here are the versions installed :
>
> MediaWiki   1.27alpha (b3a82ca)
> PHP 5.4.45-0+deb7u2 (fpm-fcgi)
> MySQL   5.5.46-0+deb7u1-log
> ICU 4.8.1.1
> Wikibase Repository 0.5 alpha (62c267)
>
> I can add some wiki pages (non-wikibase stuff), modify them, no problems.
>
> When I try to add an element or property, the first time I try it works,
then a few seconds later I can't add a second element I get an "error 500 -
internal server error", and I can't get it work until about 30 minutes / 1
hour then I can add another element and then again nothing !
>
> I think I might have a namespace problem but how to configure it and keep
my other wiki pages ?
>
> Thanks in advance !
>
> Regards,
> Jeremy

Sorry for getting back to you only now. I just talked to Marius about it
and he recommends to check the settings for object caching.

Cheers
Lydia
___
> Wikidata mailing list
> Wikidata@lists.wikimedia.org
> https://lists.wikimedia.org/mailman/listinfo/wikidata
___
Wikidata mailing list
Wikidata@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikidata