[Wikidata-l] LoCloud and Wikimedia

2013-08-26 Thread John Erling Blad
Have anyone heard about a LoCloud project in Europeana? Its about
integration of Wikimedia services and content in a cloud service for
local glam institutions. It seems like the project was started
officially in March 2013.

LoCloud is a cloud service for small and medium sized institutions. It
seems like WLM will be part of this through work package 3, and then
as a micro service (!) in a larger system based on SaaS.

John

___
Wikidata-l mailing list
Wikidata-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikidata-l


Re: [Wikidata-l] LoCloud and Wikimedia

2013-08-26 Thread John Erling Blad
There is a discussion on the cultural-partners list from 26th March 2013.
John

On Mon, Aug 26, 2013 at 9:47 AM, John Erling Blad jeb...@gmail.com wrote:
 Have anyone heard about a LoCloud project in Europeana? Its about
 integration of Wikimedia services and content in a cloud service for
 local glam institutions. It seems like the project was started
 officially in March 2013.

 LoCloud is a cloud service for small and medium sized institutions. It
 seems like WLM will be part of this through work package 3, and then
 as a micro service (!) in a larger system based on SaaS.

 John

___
Wikidata-l mailing list
Wikidata-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikidata-l


Re: [Wikidata-l] Cooperation between Wikidata and DBpedia and Wikipedia

2013-08-26 Thread Dimitris Kontokostas
Speaking from DBpedia (not on behalf), we have been always trying to find
ways to contribute data back to wikipedia and If licencing is the only
issue here I am sure we can make any necessary arrangements.

imho the main scepticism so far was trust from the WIkipedia community to
load data in bulk.
However, there are datasets of very high quality in DBpedia that could be
used for that purpose.

Recently we are experimenting in an alternative where people can manually
import single facts in Wikidata from the DBpedia resource interface.
Here's a recent talk about this [1] on the tech list. Any comments on that
are also welcome.

Best,
Dimitris

[1]
http://lists.wikimedia.org/pipermail/wikidata-tech/2013-August/000189.html



On Fri, Aug 23, 2013 at 6:18 PM, David Cuenca dacu...@gmail.com wrote:

 Hi, I will answer with questions with more questions...

 On Fri, Aug 23, 2013 at 10:58 AM, Gerard Meijssen 
 gerard.meijs...@gmail.com wrote:

 Hoi,

 The questions are:

- would we advance a lot when we adopt the DBpedia schema as it is?

 Which schema? All of them? Some? Article classification? Infobox
 extraction? Wikidata is going to be linked to the infoboxes in Wikipedia,
 so the priority is to support those needs, not to replicate any schema.





- Would we be open to include substantially more data?

 Which data? All of it? What is the reliability?




- When we adopt the schema, can we tinker with it to suit our needs?

 Again, could you please give some example of what to import and how
 should it be adapted?


 If the answers to these questions are yes, what is the point in
 procrastinating???


 Do we have already all the datatypes that would be needed? Most of the
 properties that are missing is because of the lack of value or others.





 One other big thing of DBpedia is that it is connected to many external
 resources. This will make it possible to verify our data against these
 other sources. This is imho the more important thing to do with the time of
 our volunteers. Doing the things that have already been done is a waste of
 time.


 The thing is that if those resources already are in dbpedia, we can just
 use dbpedia as a bridge, that is how linked data is supposed to be... no
 need to replicate everything, but of course, if it is worth replicating, we
 can go through case by case.

 Micru

 ___
 Wikidata-l mailing list
 Wikidata-l@lists.wikimedia.org
 https://lists.wikimedia.org/mailman/listinfo/wikidata-l




-- 
Dimitris Kontokostas
Department of Computer Science, University of Leipzig
Research Group: http://aksw.org
Homepage:http://aksw.org/DimitrisKontokostas
___
Wikidata-l mailing list
Wikidata-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikidata-l


Re: [Wikidata-l] Cooperation between Wikidata and DBpedia and Wikipedia

2013-08-26 Thread Gerard Meijssen
Hoi,

Two things to consider; the Wikipedia community and the Wikidata community
are two separate entities. As far as I am concerned, Wikidata needs more
data to get to the tipping point where it becomes useful to users. I am
really vocal about both.

The license has traditionally been a sticking point and the not invented
here aspect of DBpedia is as well. You are effectively outside the
Wikipedia community.. The other part is that the data of Wikidata is
continually updated while DBpedia is not. So in my opinion the best thing
that can happen is when DBpedia DOES update with Wikidata. The point is not
to absorb it without thinking.

What DBpedia has is a set of properties that work on the data that it has
gleaned from the many Wikipedias.  There is undoubtedly a lot of
documentation on it and, it would be good when this is taken into
consideration when accepting and proposing new properties for Wikidata.
Obviously only the properties that are currently supported can be proposed
at this time. I am quite willing to propose properties based on DBpedia
(but so can you).

With the properties in place, we can import data. We can import it from
both DBpedia and from Wikipedia. Sanity checks are needed for both sources
and as far as I am aware we do not have sanity checks at Wikidata.

What we do have is the possibility to compare data with other sources and
that is where the Wikidata community needs to grow and that is why we need
much data in the first place in order to get into this. However, we can
start building the tools to do this. I hope the DBpedia community can help
us with that.
Thanks,
 GerardM



On 26 August 2013 09:16, Dimitris Kontokostas 
kontokos...@informatik.uni-leipzig.de wrote:

 Speaking from DBpedia (not on behalf), we have been always trying to find
 ways to contribute data back to wikipedia and If licencing is the only
 issue here I am sure we can make any necessary arrangements.

 imho the main scepticism so far was trust from the WIkipedia community
 to load data in bulk.
 However, there are datasets of very high quality in DBpedia that could be
 used for that purpose.

 Recently we are experimenting in an alternative where people can manually
 import single facts in Wikidata from the DBpedia resource interface.
 Here's a recent talk about this [1] on the tech list. Any comments on that
 are also welcome.

 Best,
 Dimitris

 [1]
 http://lists.wikimedia.org/pipermail/wikidata-tech/2013-August/000189.html



 On Fri, Aug 23, 2013 at 6:18 PM, David Cuenca dacu...@gmail.com wrote:

 Hi, I will answer with questions with more questions...

 On Fri, Aug 23, 2013 at 10:58 AM, Gerard Meijssen 
 gerard.meijs...@gmail.com wrote:

 Hoi,

 The questions are:

- would we advance a lot when we adopt the DBpedia schema as it is?

 Which schema? All of them? Some? Article classification? Infobox
 extraction? Wikidata is going to be linked to the infoboxes in Wikipedia,
 so the priority is to support those needs, not to replicate any schema.





- Would we be open to include substantially more data?

 Which data? All of it? What is the reliability?




- When we adopt the schema, can we tinker with it to suit our needs?

 Again, could you please give some example of what to import and how
 should it be adapted?


 If the answers to these questions are yes, what is the point in
 procrastinating???


 Do we have already all the datatypes that would be needed? Most of the
 properties that are missing is because of the lack of value or others.





 One other big thing of DBpedia is that it is connected to many external
 resources. This will make it possible to verify our data against these
 other sources. This is imho the more important thing to do with the time of
 our volunteers. Doing the things that have already been done is a waste of
 time.


 The thing is that if those resources already are in dbpedia, we can just
 use dbpedia as a bridge, that is how linked data is supposed to be... no
 need to replicate everything, but of course, if it is worth replicating, we
 can go through case by case.

 Micru

 ___
 Wikidata-l mailing list
 Wikidata-l@lists.wikimedia.org
 https://lists.wikimedia.org/mailman/listinfo/wikidata-l




 --
 Dimitris Kontokostas
 Department of Computer Science, University of Leipzig
 Research Group: http://aksw.org
 Homepage:http://aksw.org/DimitrisKontokostas

 ___
 Wikidata-l mailing list
 Wikidata-l@lists.wikimedia.org
 https://lists.wikimedia.org/mailman/listinfo/wikidata-l


___
Wikidata-l mailing list
Wikidata-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikidata-l


Re: [Wikidata-l] claims Datatypes inconsistency suspicion

2013-08-26 Thread Daniel Kinzler

Am 25.08.2013 19:19, schrieb Markus Krötzsch:

If we have an IRI DV, considering that URLs are special IRIs, it seems clear
that IRI would be the best way of storing them.


The best way of storing them really depends on the storage platform. It may be a 
string or something else.


I think the real issue here is that we are exposing something that is really an 
internal detail (the data value type) instead of the high level information we 
actually should be exposing, namely property type.


I think splitting the two was a mistake, and I think exposing the DV type while 
making the property type all but inaccessible makes things a lot worse.


In my opinion, data should be self-descriptive, so the *semantic* type of the 
property should be included along with the value. People expect this, and assume 
that this is what the DV type is. But it's not, and should not be used or abused 
for this purpose.


Ideally, it should not matter at all to any 3rd party if use use a string or IRI 
DV internally. The (semantic) property type would be URL, and that's all that 
matters.


I'm quite unhappy about the current situation; we are beginning to see the 
backlash of the decision not to include the property type inline. If we don't do 
anything about this now, I fear the confusion is going to get worse.


-- daniel

___
Wikidata-l mailing list
Wikidata-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikidata-l


Re: [Wikidata-l] claims Datatypes inconsistency suspicion

2013-08-26 Thread Markus Krötzsch

Hi Daniel,

if I understand you correctly, you are in favour of equating datavalue 
types and property types. This would solve indeed the problems at hand.


The reason why both kinds of types are distinct in SMW and also in 
Wikidata is that property types are naturally more extensible than 
datavalue types. CommonsMedia is a good example of this: all you need is 
a custom UI and you can handle new data without changing the 
underlying data model. This makes it easy for contributors to add new 
types without far-reaching ramifications in the backend (think of 
numbers, which could be decimal, natural, positive, range-restricted, 
etc. but would still be treated as a number in the backend).


Using fewer datavalue types also improves interoperability. E.g., you 
want to compare two numbers, even if one is a natural number and another 
one is a decimal.


There is no simple rule for deciding how many datavalue types there 
should be. The general guideline is to decide on datavalue types based 
on use cases. I am arguing for diversifying IRIs and strings since there 
are many contexts and applications where this is a crucial difference. 
Conversely, I don't know of any application where it makes sense to keep 
the two similar (this would have to be something where we compare 
strings and IRIs on a data level, e.g., if you were looking for all 
websites with URLs that are alphabetically greater than the postcode of 
a city in England :-p).


In general, however, it will be good to keep the set of basic datavalue 
types small, while allowing the set of property types to grow. The set 
of base datavalue types that we use is based on the experience in SMW as 
well as on existing formats like XSD (which also has many derived types 
but only a few base types).


As for the possible confusion, I think some naming discipline would 
clarify this. In SMW, there is a stronger difference between both kinds 
of types, and a fixed schema for property type ids that makes it easy to 
recognise them.


In any case, using string for IRIs does not seem to solve any problem. 
It does not simplify the type system in general and it does not help 
with the use cases that I mentioned. What I do not agree with are your 
arguments about all of this being internal. We would not have this 
discussion if it were. The data model of Wikidata is the primary 
conceptual model that specifies what Wikidata stores. You might still be 
right that some of the implementation is internal, but the arguments we 
both exchange are not really on the implementation level ;-).


Best wishes

Markus, offline soon for travelling


On 26/08/13 10:35, Daniel Kinzler wrote:

Am 25.08.2013 19:19, schrieb Markus Krötzsch:

If we have an IRI DV, considering that URLs are special IRIs, it seems
clear
that IRI would be the best way of storing them.


The best way of storing them really depends on the storage platform. It
may be a string or something else.

I think the real issue here is that we are exposing something that is
really an internal detail (the data value type) instead of the high
level information we actually should be exposing, namely property type.

I think splitting the two was a mistake, and I think exposing the DV
type while making the property type all but inaccessible makes things a
lot worse.

In my opinion, data should be self-descriptive, so the *semantic* type
of the property should be included along with the value. People expect
this, and assume that this is what the DV type is. But it's not, and
should not be used or abused for this purpose.

Ideally, it should not matter at all to any 3rd party if use use a
string or IRI DV internally. The (semantic) property type would be URL,
and that's all that matters.

I'm quite unhappy about the current situation; we are beginning to see
the backlash of the decision not to include the property type inline. If
we don't do anything about this now, I fear the confusion is going to
get worse.

-- daniel




___
Wikidata-l mailing list
Wikidata-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikidata-l


[Wikidata-l] DNB 11M bibliographic records as CC0

2013-08-26 Thread David Cuenca
Hi,

Maybe this is interesting as an import source for bibliographic info
http://blogs.ifla.org/bibliography/2013/08/06/german-national-library-offers-over-11-million-marc21-records-under-cc0-open-license/

Cheers,
Micru
___
Wikidata-l mailing list
Wikidata-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikidata-l


Re: [Wikidata-l] [Wikisource-l] DNB 11M bibliographic records as CC0

2013-08-26 Thread Andrea Zanni
This seems very interesting: maybe, there will be a time when Wikidata (or
similar) will host the bibliographic records
of thousands of libraries...

Right now, I'm not sure if we want to discuss a massive upload of these
records in WD,
because:
* they are in MARC, which is way more complex than the list of
bibliographic properties  currently in WD
* they will be in German
* do we really need so many records of books, articles and work, at this
early stage?

Aubrey


On Mon, Aug 26, 2013 at 2:32 PM, David Cuenca dacu...@gmail.com wrote:

 Hi,

 Maybe this is interesting as an import source for bibliographic info

 http://blogs.ifla.org/bibliography/2013/08/06/german-national-library-offers-over-11-million-marc21-records-under-cc0-open-license/

 Cheers,
 Micru

 ___
 Wikisource-l mailing list
 wikisourc...@lists.wikimedia.org
 https://lists.wikimedia.org/mailman/listinfo/wikisource-l


___
Wikidata-l mailing list
Wikidata-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikidata-l


Re: [Wikidata-l] [Wikisource-l] DNB 11M bibliographic records as CC0

2013-08-26 Thread Mark MacGillivray
There are many such CC0 national bibliographies available, and other large
datasets, if the decision is made to import them.

See here for a list:

http://datahub.io/group/bibliographic


Here is a direct link to the British National Bibliography page:

http://bnb.bl.uk/


For further information on the open bibliographic project and the
principles of open bibliographic metadata, check here:

http://openbiblio.net
http://openbiblio.net/principles


Mark MacGillivray






On Mon, Aug 26, 2013 at 1:51 PM, David Cuenca dacu...@gmail.com wrote:

 No, we don't need to import them all, but there was always the question if
 we were allowed to import that data from external sources.
 At least for the DNB that question has been settled.

 Cheers,
 Micru




 On Mon, Aug 26, 2013 at 8:37 AM, Andrea Zanni zanni.andre...@gmail.comwrote:

 This seems very interesting: maybe, there will be a time when Wikidata
 (or similar) will host the bibliographic records
 of thousands of libraries...

 Right now, I'm not sure if we want to discuss a massive upload of these
 records in WD,
 because:
 * they are in MARC, which is way more complex than the list of
 bibliographic properties  currently in WD
 * they will be in German
 * do we really need so many records of books, articles and work, at this
 early stage?

 Aubrey


 On Mon, Aug 26, 2013 at 2:32 PM, David Cuenca dacu...@gmail.com wrote:

 Hi,

 Maybe this is interesting as an import source for bibliographic info

 http://blogs.ifla.org/bibliography/2013/08/06/german-national-library-offers-over-11-million-marc21-records-under-cc0-open-license/

 Cheers,
 Micru

 ___
 Wikisource-l mailing list
 wikisourc...@lists.wikimedia.org
 https://lists.wikimedia.org/mailman/listinfo/wikisource-l



 ___
 Wikisource-l mailing list
 wikisourc...@lists.wikimedia.org
 https://lists.wikimedia.org/mailman/listinfo/wikisource-l




 --
 Etiamsi omnes, ego non

 ___
 Wikidata-l mailing list
 Wikidata-l@lists.wikimedia.org
 https://lists.wikimedia.org/mailman/listinfo/wikidata-l


___
Wikidata-l mailing list
Wikidata-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikidata-l


Re: [Wikidata-l] [Wikisource-l] DNB 11M bibliographic records as CC0

2013-08-26 Thread Thomas Douillard
If Wikidata has an ambition to be a really reliable database, we should do
eveything we can to make it easy for users to use any source they want. In
this perspective, if we got datas with guaranted high quality, it make it
easy for Wikidatian to find and use these references for users. Entering a
reference in the database seems to me a highly fastidious, boring, and
easily automated task.

With that in mind, any reference that the user will not have to enter by
hand is something good, and import high quality sources datas should pass
every Wikidata community barriers easily. If there is no problem for the
software to handle that many information, I say we really have no reason
not to do the imports.

Tom
___
Wikidata-l mailing list
Wikidata-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikidata-l


Re: [Wikidata-l] Cooperation between Wikidata and DBpedia and Wikipedia

2013-08-26 Thread Kingsley Idehen

On 8/26/13 4:34 AM, Gerard Meijssen wrote:

Hoi,

Two things to consider; the Wikipedia community and the Wikidata 
community are two separate entities. As far as I am concerned, 
Wikidata needs more data to get to the tipping point where it becomes 
useful to users. I am really vocal about both.


The license has traditionally been a sticking point and the not 
invented here aspect of DBpedia is as well.


What is this not invented here aspect of DBpedia?

You are effectively outside the Wikipedia community.. The other part 
is that the data of Wikidata is continually updated while DBpedia is not.


What do you mean by that? Are you referring to 
schema/ontology/vocabulary evolution or instance data evolution? 
Remember, there are live editions of DBpedia.


So in my opinion the best thing that can happen is when DBpedia DOES 
update with Wikidata. The point is not to absorb it without thinking.


Yes, the first point of clarity would be when Wikidata produces a dump 
that can be ingested by DBpedia and any other data space in the LOD 
cloud. All that's required is data publication in Linked Data form.




What DBpedia has is a set of properties that work on the data that it 
has gleaned from the many Wikipedias.  There is undoubtedly a lot of 
documentation on it and, it would be good when this is taken into 
consideration when accepting and proposing new properties for 
Wikidata. Obviously only the properties that are currently supported 
can be proposed at this time. I am quite willing to propose properties 
based on DBpedia (but so can you).


With the properties in place, we can import data. We can import it 
from both DBpedia and from Wikipedia. Sanity checks are needed for 
both sources and as far as I am aware we do not have sanity checks at 
Wikidata.


What we do have is the possibility to compare data with other sources 
and that is where the Wikidata community needs to grow and that is why 
we need much data in the first place in order to get into this. 
However, we can start building the tools to do this. I hope the 
DBpedia community can help us with that.

Thanks,
 GerardM


As far as I know, DBpedia has always been interested in collaboration 
with Wikidata.



Kingsley




On 26 August 2013 09:16, Dimitris Kontokostas 
kontokos...@informatik.uni-leipzig.de 
mailto:kontokos...@informatik.uni-leipzig.de wrote:


Speaking from DBpedia (not on behalf), we have been always trying
to find ways to contribute data back to wikipedia and If licencing
is the only issue here I am sure we can make any necessary
arrangements.

imho the main scepticism so far was trust from the WIkipedia
community to load data in bulk.
However, there are datasets of very high quality in DBpedia that
could be used for that purpose.

Recently we are experimenting in an alternative where people can
manually import single facts in Wikidata from the DBpedia resource
interface.
Here's a recent talk about this [1] on the tech list. Any comments
on that are also welcome.

Best,
Dimitris

[1]
http://lists.wikimedia.org/pipermail/wikidata-tech/2013-August/000189.html



On Fri, Aug 23, 2013 at 6:18 PM, David Cuenca dacu...@gmail.com
mailto:dacu...@gmail.com wrote:

Hi, I will answer with questions with more questions...

On Fri, Aug 23, 2013 at 10:58 AM, Gerard Meijssen
gerard.meijs...@gmail.com mailto:gerard.meijs...@gmail.com
wrote:

Hoi,

The questions are:

  * would we advance a lot when we adopt the DBpedia
schema as it is?

Which schema? All of them? Some? Article classification?
Infobox extraction? Wikidata is going to be linked to the
infoboxes in Wikipedia, so the priority is to support those
needs, not to replicate any schema.


  * Would we be open to include substantially more data?

Which data? All of it? What is the reliability?


  * When we adopt the schema, can we tinker with it to
suit our needs?

Again, could you please give some example of what to import
and how should it be adapted?

If the answers to these questions are yes, what is the
point in procrastinating???


Do we have already all the datatypes that would be needed?
Most of the properties that are missing is because of the lack
of value or others.



One other big thing of DBpedia is that it is connected to
many external resources. This will make it possible to
verify our data against these other sources. This is imho
the more important thing to do with the time of our
volunteers. Doing the things that have already been done
is a waste of time.


The thing is that if those resources already are in dbpedia,
we can just use dbpedia as a bridge, that is how 

Re: [Wikidata-l] [Wikisource-l] DNB 11M bibliographic records as CC0

2013-08-26 Thread David Cuenca
If the problem is to automate bibliographic data importing, a solution is
what you propose, to import everything. Another one is to have an import
tool to automatically import the data for the item that needs it. In WP
they do that, there is a tool to import book/journal info by ISBN/doi. The
same can be done in WD.

Micru


On Mon, Aug 26, 2013 at 9:23 AM, Thomas Douillard 
thomas.douill...@gmail.com wrote:

 If Wikidata has an ambition to be a really reliable database, we should do
 eveything we can to make it easy for users to use any source they want. In
 this perspective, if we got datas with guaranted high quality, it make it
 easy for Wikidatian to find and use these references for users. Entering a
 reference in the database seems to me a highly fastidious, boring, and
 easily automated task.

 With that in mind, any reference that the user will not have to enter by
 hand is something good, and import high quality sources datas should pass
 every Wikidata community barriers easily. If there is no problem for the
 software to handle that many information, I say we really have no reason
 not to do the imports.

 Tom


 ___
 Wikidata-l mailing list
 Wikidata-l@lists.wikimedia.org
 https://lists.wikimedia.org/mailman/listinfo/wikidata-l




-- 
Etiamsi omnes, ego non
___
Wikidata-l mailing list
Wikidata-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikidata-l


Re: [Wikidata-l] [Wikisource-l] DNB 11M bibliographic records as CC0

2013-08-26 Thread Thomas Douillard
I know, I started a discussion about porting the bot to WIkidata in
scientific Journal Wikiproject. One answer I got : the bot owner had other
things to do in his life than running the bot and was not around very often
any more. Having everiyhing in Wikidata already will be a lot more reliable
and lazier, no tool that works one day but not the other one, no effort to
tell the newbies that they should go to another website, no significant
problem.

Maybe one opposition would be that the data would be vandalised easily, but
maybe we should find a way to deal with imported sourced datas which have
no real reason to be modified, just marked deprecated or updated by another
import from the same source.


2013/8/26 David Cuenca dacu...@gmail.com

 If the problem is to automate bibliographic data importing, a solution is
 what you propose, to import everything. Another one is to have an import
 tool to automatically import the data for the item that needs it. In WP
 they do that, there is a tool to import book/journal info by ISBN/doi. The
 same can be done in WD.

 Micru


 On Mon, Aug 26, 2013 at 9:23 AM, Thomas Douillard 
 thomas.douill...@gmail.com wrote:

 If Wikidata has an ambition to be a really reliable database, we should
 do eveything we can to make it easy for users to use any source they want.
 In this perspective, if we got datas with guaranted high quality, it make
 it easy for Wikidatian to find and use these references for users. Entering
 a reference in the database seems to me a highly fastidious, boring, and
 easily automated task.

 With that in mind, any reference that the user will not have to enter by
 hand is something good, and import high quality sources datas should pass
 every Wikidata community barriers easily. If there is no problem for the
 software to handle that many information, I say we really have no reason
 not to do the imports.

 Tom


 ___
 Wikidata-l mailing list
 Wikidata-l@lists.wikimedia.org
 https://lists.wikimedia.org/mailman/listinfo/wikidata-l




 --
 Etiamsi omnes, ego non

 ___
 Wikidata-l mailing list
 Wikidata-l@lists.wikimedia.org
 https://lists.wikimedia.org/mailman/listinfo/wikidata-l


___
Wikidata-l mailing list
Wikidata-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikidata-l


Re: [Wikidata-l] Cooperation between Wikidata and DBpedia and Wikipedia

2013-08-26 Thread Gerard Meijssen
Hoi,

As you may have noticed there are other people on the CC..

The not invented here refers to much of the Wiki world.. things get done
but much is done over and over again EVEN when cooperation is easy and
obvious. I have largely resigned myself to it, it is probably part of our
persona. However, it is not in the best interest of our users.

The people I originally wrote to are more DBpedia ... and very much
Wikimedia as well. I do not remember that DBpedia has a live component,
love to learn more..

I do know how much the DBpedia people want to reach out and connect in any
positive way with both Wikipedia and Wikidata. It is something that would
work wonders because it can speed up the addition of new properties, the
import of wholesale data AND it may have us add the processes they have
build to curate the data.

grin and additional point would be that we include some academia at the
same time /grin

Thanks,
 GerardM




On 26 August 2013 15:27, Kingsley Idehen kide...@openlinksw.com wrote:

 On 8/26/13 4:34 AM, Gerard Meijssen wrote:

 Hoi,

 Two things to consider; the Wikipedia community and the Wikidata community
 are two separate entities. As far as I am concerned, Wikidata needs more
 data to get to the tipping point where it becomes useful to users. I am
 really vocal about both.

 The license has traditionally been a sticking point and the not invented
 here aspect of DBpedia is as well.


 What is this not invented here aspect of DBpedia?

 You are effectively outside the Wikipedia community.. The other part is
 that the data of Wikidata is continually updated while DBpedia is not.


 What do you mean by that? Are you referring to schema/ontology/vocabulary
 evolution or instance data evolution? Remember, there are live editions of
 DBpedia.

 So in my opinion the best thing that can happen is when DBpedia DOES
 update with Wikidata. The point is not to absorb it without thinking.


 Yes, the first point of clarity would be when Wikidata produces a dump
 that can be ingested by DBpedia and any other data space in the LOD cloud.
 All that's required is data publication in Linked Data form.


 What DBpedia has is a set of properties that work on the data that it has
 gleaned from the many Wikipedias.  There is undoubtedly a lot of
 documentation on it and, it would be good when this is taken into
 consideration when accepting and proposing new properties for Wikidata.
 Obviously only the properties that are currently supported can be proposed
 at this time. I am quite willing to propose properties based on DBpedia
 (but so can you).

 With the properties in place, we can import data. We can import it from
 both DBpedia and from Wikipedia. Sanity checks are needed for both sources
 and as far as I am aware we do not have sanity checks at Wikidata.

 What we do have is the possibility to compare data with other sources and
 that is where the Wikidata community needs to grow and that is why we need
 much data in the first place in order to get into this. However, we can
 start building the tools to do this. I hope the DBpedia community can help
 us with that.
 Thanks,
  GerardM


 As far as I know, DBpedia has always been interested in collaboration with
 Wikidata.


 Kingsley




 On 26 August 2013 09:16, Dimitris Kontokostas 
 kontokos...@informatik.uni-leipzig.de wrote:

 Speaking from DBpedia (not on behalf), we have been always trying to find
 ways to contribute data back to wikipedia and If licencing is the only
 issue here I am sure we can make any necessary arrangements.

 imho the main scepticism so far was trust from the WIkipedia community
 to load data in bulk.
 However, there are datasets of very high quality in DBpedia that could be
 used for that purpose.

 Recently we are experimenting in an alternative where people can manually
 import single facts in Wikidata from the DBpedia resource interface.
 Here's a recent talk about this [1] on the tech list. Any comments on
 that are also welcome.

 Best,
 Dimitris

 [1]
 http://lists.wikimedia.org/pipermail/wikidata-tech/2013-August/000189.html



 On Fri, Aug 23, 2013 at 6:18 PM, David Cuenca dacu...@gmail.com wrote:

 Hi, I will answer with questions with more questions...

 On Fri, Aug 23, 2013 at 10:58 AM, Gerard Meijssen 
 gerard.meijs...@gmail.com wrote:

 Hoi,

 The questions are:

- would we advance a lot when we adopt the DBpedia schema as it is?

 Which schema? All of them? Some? Article classification? Infobox
 extraction? Wikidata is going to be linked to the infoboxes in Wikipedia,
 so the priority is to support those needs, not to replicate any schema.





- Would we be open to include substantially more data?

 Which data? All of it? What is the reliability?




- When we adopt the schema, can we tinker with it to suit our
needs?

 Again, could you please give some example of what to import and how
 should it be adapted?


 If the answers to these questions are yes, what is the point in
 

[Wikidata-l] WikiData DBpedia Dump Release v.0.1

2013-08-26 Thread Hady elsahar
Hello All,

as an output of GSoC2013 project , Wikidata integration inside
DBpediahttp://wiki.dbpedia.org/gsoc2013/ideas/WikidataMappings?v=hz9
we
are happy to announce that an initial RDF DBpedia Dumps for Wikidata Data
is now available. (details and examples for each dump will be in the
download section in the wikipage) this is an initial Dump , so your
Feedback is needed of course to review the exported data and enhance it.

You can find more information and details here:
 
https://github.com/hadyelsahar/extraction-framework/wiki/WikiData-DBpedia-Dump-Release-v.0.1https://github.com/hadyelsahar/extraction-framework/wiki/WikiData-DBpedia-Dump-Release-v.0.1
 (There is also a list of known issues at the bottom of the wikipage)

Thanks

Regards

-
Hady El-Sahar
Research Assistant
Center of Informatics Sciences | Nile Universityhttp://nileuniversity.edu.eg/
___
Wikidata-l mailing list
Wikidata-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikidata-l


[Wikidata-l] Data acccess for Wikivoyage and bugfixes plus testing needed for the URL datatype

2013-08-26 Thread Lydia Pintscher
Heya folks :)

Wikiyovage now has access to the data on Wikidata - meaning they got
phase 2. With this the first sister project is fully supported \o/ The
next sister projects will follow. More information on this soon.

We've also just updated the software here. This brings a number of bug
fixes and other improvements. The ones that are probably most
important for you:
* The copyright warning no longer pops up again when you change your language
* There is now a new special page to list items/properties without a
description in a given language. Thanks for the patch by Bene*.
* The message that pops up when you want to add a link to an item
which is already in use in another item has been improved. Thanks for
the patch by Nemo bis.
* Broken links to set sitelinks for Wikivoyage in the non-JavaScript
version have been fixed. ([[bugzilla:51914]], [[bugzilla:52095]])
* The automatic comments for edits have been improved. They are more
detailed now.
* API: You can now provide custom edit summaries for edits via the API.
* API: You can undo a revision via the API.
* API: Bot edits via the setclaim API module are now marked as such
([[bugzilla:50933]]).
* API: Precision and globe parameters are now required for geocoordinate data.
* Starting in a few days a Wikidata edit that shows up in recent
changes and watchlist on the client (Wikipedia/Wikivoyage) is going to
be marked with a D. Thanks for the patch by umherirrender.

Unfortunately we were not able to put the URL datatype on the test
system on time for this deployment. It didn't get enough testing so we
couldn't deploy it today with a good concious. We know you're waiting
for this but it's better to give it a bit more testing and roll it out
in 2 weeks with the next deployment. The URL datatype is now live for
real on test.wikidata.org for you to try. Please give it a try and
report any bugs you encounter.

Please let me know if there are any issues.

Oh and one more thing: Abraham, Denny and I sat down for an evening
trying to captchure what Wikidata is about in a video. Hope you like
it :) http://www.wikidata.org/wiki/File:Wikidata%27s_World.webm


Cheers
Lydia

-- 
Lydia Pintscher - http://about.me/lydia.pintscher
Community Communications for Technical Projects

Wikimedia Deutschland e.V.
Obentrautstr. 72
10963 Berlin
www.wikimedia.de

Wikimedia Deutschland - Gesellschaft zur Förderung Freien Wissens e. V.

Eingetragen im Vereinsregister des Amtsgerichts Berlin-Charlottenburg
unter der Nummer 23855 Nz. Als gemeinnützig anerkannt durch das
Finanzamt für Körperschaften I Berlin, Steuernummer 27/681/51985.

___
Wikidata-l mailing list
Wikidata-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikidata-l


Re: [Wikidata-l] Data acccess for Wikivoyage and bugfixes plus testing needed for the URL datatype

2013-08-26 Thread Michał Łazowik
Hi,

great news!

Wiadomość napisana przez Lydia Pintscher lydia.pintsc...@wikimedia.de w dniu 
27 sie 2013, o godz. 00:05:

 * Broken links to set sitelinks for Wikivoyage in the non-JavaScript
 version have been fixed. ([[bugzilla:51914]], [[bugzilla:52095]])

Small correction: this affected all sitelinks (of course in non-js editing).

Michał
___
Wikidata-l mailing list
Wikidata-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikidata-l


Re: [Wikidata-l] Data acccess for Wikivoyage and bugfixes plus testing needed for the URL datatype

2013-08-26 Thread Lydia Pintscher
On Tue, Aug 27, 2013 at 12:11 AM, Michał Łazowik mlazo...@me.com wrote:
 Hi,

 great news!

 Wiadomość napisana przez Lydia Pintscher lydia.pintsc...@wikimedia.de w 
 dniu 27 sie 2013, o godz. 00:05:

 * Broken links to set sitelinks for Wikivoyage in the non-JavaScript
 version have been fixed. ([[bugzilla:51914]], [[bugzilla:52095]])

 Small correction: this affected all sitelinks (of course in non-js editing).

Ah thanks! I'll fix the project chat message.


Cheers
Lydia

-- 
Lydia Pintscher - http://about.me/lydia.pintscher
Community Communications for Technical Projects

Wikimedia Deutschland e.V.
Obentrautstr. 72
10963 Berlin
www.wikimedia.de

Wikimedia Deutschland - Gesellschaft zur Förderung Freien Wissens e. V.

Eingetragen im Vereinsregister des Amtsgerichts Berlin-Charlottenburg
unter der Nummer 23855 Nz. Als gemeinnützig anerkannt durch das
Finanzamt für Körperschaften I Berlin, Steuernummer 27/681/51985.

___
Wikidata-l mailing list
Wikidata-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikidata-l