On 07/08/13 10:20, Magnus Manske wrote:
I could offer an interface:
https://toolserver.org/~magnus/thetalkpage/
<https://toolserver.org/%7Emagnus/thetalkpage/>
Yes, Magnus, I would also have suggested this next :-)
Mingli, inspired by Magnus' demo, we have indeed been talking about such
QA interfaces to Wikidata a bit already. But as you also note, there is
still a lot to be done for this to work. I agree that reasoning will
play an important role for this (because we don't want to store all
"grandfather" relations in Wikidata when they can be inferred easily).
We are also in contact with NLP people regarding generic language
interfaces ("understanding" questions instead of indexing generated
questions). A number of groups have been doing work in this area using
DBPedia or Yago, but Wikidata adds new challenges due to the more
complex data model.
Markus
On Wed, Aug 7, 2013 at 10:00 AM, Mingli Yuan <mingli.y...@gmail.com
<mailto:mingli.y...@gmail.com>> wrote:
Thanks, Markus,
About the background:
One is related with my current work and I can not say it too much.
But another story, I can say it publicly.
After playing with wikidata for a while, I realised that we have the
potential to create a WolframAlpha-like application. To achieve
this, what we need are only a indexer, a generator, a reasoner and a
traversor.
Take a look at simplenlg ( https://code.google.com/p/simplenlg/ ),
it can take claim triples and transform them into sentences of
questions. We then index these the answer by question sentences, the
answer could be provided by wikidata claims.
And then we get a QA engine with weak abilities. i.e.
* they only know who are Enistein's father and children, and do
not know that grandfather is the father's father.
* they only know which planets belong to the Sun system, and they
do not know which is the biggest, the farest, etc.
So if we add some axiom into the system and also with a reasoner and
a traversor. We have the potential to enumerate all possible
*simple* questions and answers of human knowledge. Then we get a QA
engine with strong abilities.
These are what in my brain now. I know these kinds of things are
never easy, but they are possible in the near future.
Regards,
Mingli
On Wed, Aug 7, 2013 at 4:14 PM, Markus Krötzsch
<mar...@semantic-mediawiki.org
<mailto:mar...@semantic-mediawiki.org>> wrote:
Hi Mingli,
thanks, this very interesting, but I think I need a bit more
context to understand what you are doing and why.
Is your goal to create a library for accessing Wikidata from
Clojure (like a Clojure API for Wikidata)? Or is your goal to
use logical inference over Wikidata and you just use Clojure as
a tool since it was most convenient?
To your question:
>
> * Do we have a long term plan to evolve wikidata towards a
> semantic-rich dataset?
>
There are no concrete designs for adding reasoning features to
Wikidata so far (if this is what you mean). There are various
open questions, especially related to inferencing over
quantifiers. But there are also important technical questions,
especially regarding performance. I intend to work the theory
out in more detail soon (that is: "How should logical rules over
the Wikidata data model look work in principle?"). The
implementation then is the next step. I don't think that any of
this will be part of the core features of Wikidata soon, but
hopefully we can set up a useful external service for Wikidata
search and analytics (e.g., to check for property constraint
violations in real time instead of using custom code and bots).
Cheers,
Markus
On 05/08/13 17:30, Mingli Yuan wrote:
Hi, folks,
After one night quick work, I had gave a proof-of-concept to
demonstrate
the feasibility that we can combine Wikidata and Clojure logic
programming together.
The source code is at here:
https://github.com/mountain/__knowledge
<https://github.com/mountain/knowledge>
An example of an entity:
https://github.com/mountain/__knowledge/blob/master/src/__entities/albert_einstein.clj
<https://github.com/mountain/knowledge/blob/master/src/entities/albert_einstein.clj>
Example of types:
https://github.com/mountain/__knowledge/blob/master/src/__meta/types.clj
<https://github.com/mountain/knowledge/blob/master/src/meta/types.clj>
Example of predicates:
https://github.com/mountain/__knowledge/blob/master/src/__meta/properties.clj
<https://github.com/mountain/knowledge/blob/master/src/meta/properties.clj>
Example of inference:
https://github.com/mountain/__knowledge/blob/master/test/__knowledge/test.clj
<https://github.com/mountain/knowledge/blob/master/test/knowledge/test.clj>
Also we found it is very easy to get another language
versions of the
data other than English.
So, thanks very much for your guys' great work!
But I found the semantic layer of wikidata is shallow, that
means you
can only knows who are Einstein's father and children, but
it can not be
inferred automatically from wikidata that Einstein's father
is the
grandfather of Einstein's children.
So my question is that:
* Do we have a long term plan to evolve wikidata towards a
semantic-rich dataset?
Regards,
Mingli
_________________________________________________
Wikidata-l mailing list
Wikidata-l@lists.wikimedia.org
<mailto:Wikidata-l@lists.wikimedia.org>
https://lists.wikimedia.org/__mailman/listinfo/wikidata-l
<https://lists.wikimedia.org/mailman/listinfo/wikidata-l>
_______________________________________________
Wikidata-l mailing list
Wikidata-l@lists.wikimedia.org <mailto:Wikidata-l@lists.wikimedia.org>
https://lists.wikimedia.org/mailman/listinfo/wikidata-l
--
undefined
_______________________________________________
Wikidata-l mailing list
Wikidata-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikidata-l