Re: [Wikidata] Code level documentation / reporting for Wikidata Query Service

2018-01-15 Thread Alessio Melandri
Thanks Guillaume, I really appreciate it!

Reading them I've found that all the links in
https://doc.wikimedia.org/wikidata-query-rdf/parent/ are pointing to
https://doc.wikimedia.org/index.html currently a 404.

Thanks,

Alessio

2018-01-04 10:27 GMT+01:00 Guillaume Lederrey :

> Hello all!
>
> We are now publishing a bunch of documentation / reports on the code
> behind Wikidata Query Service [1]. All this is very developer
> oriented, so skip the rest of this email if you just want to use the
> service.
>
> All the reports are generated via Maven, and reflect the current state
> of the master branch of the project [2]. Reports are organized by
> modules. Feel free to browse through them and see if you find anything
> interesting. A few pointers to get started (links provided to the
> "tools" module, the same reports are available for other modules as
> well):
>
> * JaCoCo code coverage [3]: line and branch unit test coverage
> * JDepend [4]: a few slightly obscure code metrics. Package cycles and
> "distance" are interesting ones.
> * Dependency updates [5]: which of our dependencies could be upgraded
> (yes, we have some work to do on that side)
>
> If you know about other interesting reports that could be added to
> that collection, feel free to reach out to me. And if you want to
> start addressing some of those issues, patches are always welcomed!
>
> Thanks for your time...
>
>Guillaume
>
>
> [1] https://doc.wikimedia.org/wikidata-query-rdf/parent/
> [2] https://github.com/wikimedia/wikidata-query-rdf/
> [3] https://doc.wikimedia.org/wikidata-query-rdf/parent/
> tools/jacoco/index.html
> [4] https://doc.wikimedia.org/wikidata-query-rdf/parent/
> tools/jdepend-report.html
> [5] https://doc.wikimedia.org/wikidata-query-rdf/parent/
> dependency-updates-report.html
>
> --
> Guillaume Lederrey
> Operations Engineer, Discovery
> Wikimedia Foundation
> UTC+2 / CEST
>
> ___
> Wikidata mailing list
> Wikidata@lists.wikimedia.org
> https://lists.wikimedia.org/mailman/listinfo/wikidata
>
___
Wikidata mailing list
Wikidata@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikidata


[Wikidata] weekly summary #295

2018-01-15 Thread Lydia Pintscher
Hey folks :)

Here is your summary of what's been happening around Wikidata over the past
week.

Discussions

   - Open request for adminship: Mahir256
   

   - Closed request for adminship: Jarekt
   

   (as successful)

Events /
Press/Blogs


   - Upcoming: Wikidata workshop in Vienna
   
   - WikiJabber podcast with Andy Mabbett
   

Other Noteworthy Stuff

   - GraFa, a new faceted browser for Wikidata is looking for feedback.
   
   - Analysis: How much are items about scientific articles, genes and
   chemical entities used on the Wikimedia projects?
   

   - New catalogs in Mix'n'match - for example Nobel Prize People Nomination
    (More than 600
   properties now have a corresponding catalog in Mix'n'match.
   

   )
   - Next round of Projects Grants
    is open for
   applications until the end of January (see also this blog post
   )

Did you know?

   - Newest properties
   : National
   Historic Ships certificate no.
   , Dictionnaire des auteurs
   luxembourgeois ID , Line
   Music album ID , Line
   Music artist ID , Elhuyar
   ZTH ID , is proceedings
   from ,
Basketball-Reference.com
   NBDL player ID , animal
   breed , Uniform Resource
   Identifier Scheme ,
OpenEdition
   journal ID , Brooklyn
   Museum artwork ID , Musée
   des Augustins artwork ID
, Yale
   Center for British Art artwork ID
   , Solomon R. Guggenheim
   Museum artwork ID , IBM
   graphic character global identifier
   , IBM coded character set
   ID , IBM code page
   identifier , produced sound
   , IUPAC GoldBook ID
   , The Baseball Cube player
   ID , Mir@bel journal ID
   , INRAN Italian Food ID
   , uBio ID
   , Deutsche Synchronkartei
   dubbing voice actor ID
, Sign@l
   journal ID , C-SPAN
   organization ID , Le
   Maitron person ID , GSMArena
   phone ID , Érudit journal
   ID 
   - Query examples:
  - Gallery of photography techniques
  

  - Street names that exist several times in Berlin
  


[Wikidata] GlobalFactSync

2018-01-15 Thread Magnus Knuth
Dear all,

last year, we applied for a Wikimedia grant to feed qualified data from 
Wikipedia infoboxes (i.e. missing statements with references) via the DBpedia 
software into Wikidata. The evaluation was already quite good, but some parts 
were still missing and we would like to ask for your help and feedback for the 
next round. The new application is here: 
https://meta.wikimedia.org/wiki/Grants:Project/DBpedia/GlobalFactSync

The main purpose of the grant is:

- Wikipedia infoboxes are quite rich, are manually curated and have references. 
DBpedia is already extracting that data quite well (i.e. there is no other 
software that does it better). However, extracting references is not a priority 
on our agenda. They would be very useful to Wikidata, but there are no user 
requests for this from DBpedia users.

- DBpedia also has all the infos of all infoboxes of all Wikipedia editions 
(>10k pages), so we also know quite well, where Wikidata is used already and 
where information is available in Wikidata or one language version and missing 
in another.

- side-goal: bring the Wikidata, Wikipedia and DBpedia communities closer 
together

Here is a diff between the old an new proposal:

- extraction of infobox references will still be a goal of the reworked proposal

- we have been working on the fusion and data comparison engine (the part of 
the budget that came from us) for a while now and there are first results:

6823 birthDate_gain_wiki.nt
3549 deathDate_gain_wiki.nt
  362541 populationTotal_gain_wiki.nt
  372913 total

We only took three properties for now and showed the gain where no Wikidata 
statement was available. birthDate/deathDate is already quite good. Details 
here: 
https://drive.google.com/file/d/1j5GojhzFJxLYTXerLJYz3Ih-K6UtpnG_/view?usp=sharing

Our plan here is to map all Wikidata properties to the DBpedia Ontology and 
then have the info to compare coverage of Wikidata with all infoboxes across 
languages.

- we will remove the text extraction part from the old proposal (which is here 
for you reference: 
https://meta.wikimedia.org/wiki/Grants:Project/DBpedia/CrossWikiFact). This 
will still be a focus during our work in 2018, together with Diffbot and the 
new DBpedia NLP department, but we think that it distracted from the core of 
the proposal. Results from the Wikipedia article text extraction can be added 
later once they are available and discussed separately.

- We proposed to make an extra website that helps to synchronize all Wikipedias 
and Wikidata with DBpedia as its backend. While the external website is not an 
ideal solution, we are lacking alternatives. The Primary Sources Tool is mainly 
for importing data into Wikidata, not so much synchronization. The MediaWiki 
instances of the Wikipedias do not seem to have any good interfaces to provide 
suggestions and pinpoint missing info. Especially to this part, we would like 
to ask for your help and suggestions, either per mail to the list or on the 
talk page: 
https://meta.wikimedia.org/wiki/Grants_talk:Project/DBpedia/GlobalFactSync

We are looking forward to a fruitful collaboration with you and we thank you 
for your feedback!

All the best
Magnus

-- 
Magnus Knuth

Universität Leipzig
Institut für Informatik
Abt. Betriebliche Informationssysteme, AKSW/KILT
Augustusplatz 10
04109 Leipzig DE

mail: kn...@informatik.uni-leipzig.de
tel: +49 177 3277537
webID: http://magnus.13mm.de/


___
Wikidata mailing list
Wikidata@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikidata


Re: [Wikidata] GraFa: Faceted browser for RDF/Wikidata [feedback requested]

2018-01-15 Thread Aidan Hogan

Hi all!

Just a friendly reminder that tomorrow we will close the questionnaire 
so if you have a few minutes to help us out (or are just curious to see 
our faceted search system) please see the links and instructions below.


And many thanks to those who have already provided feedback! :)

Best,
José & Aidan

On 09-01-2018 14:18, Aidan Hogan wrote:

Hey all,

A Masters student of mine (José Moreno in CC) has been working on a 
faceted navigation system for (large-scale) RDF datasets called "GraFa".


The system is available here loaded with a recent version of Wikidata:

http://grafa.dcc.uchile.cl/

Hopefully it is more or less self-explanatory for the moment. :)


If you have a moment to spare, we would hugely appreciate it if you 
could interact with the system for a few minutes and then answer a quick 
questionnaire that should only take a couple more minutes:


https://goo.gl/forms/h07qzn0aNGsRB6ny1

Just for the moment while the questionnaire is open, we would kindly 
request to send feedback to us personally (off-list) to not affect 
others' responses. We will leave the questionnaire open for a week until 
January 16th, 17:00 GMT. After that time of course we would be happy to 
discuss anything you might be interested in on the list. :)


After completing the questionnaire, please also feel free to visit or 
list something you noticed on the Issue Tracker:


https://github.com/joseignm/GraFa/issues


Many thanks,
Aidan and José






___
Wikidata mailing list
Wikidata@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikidata