Hi,
I also think you should do some preprocessing using ASCII Folding
techniques.
You could fold labels and add them as additional surface forms for the
entity.
The same process would apply for labels coming from different languages.
I have successfully used this approach in a project where Solr
Hi,
there is no magic in that.
It only happens that wikipedia has got a page Isere (
http://en.wikipedia.org/wiki/Isere) which is actually a mere redirect to
Isère (http://en.wikipedia.org/wiki/Is%C3%A8re).
Hence the framework links the two DBpedia entities together in a triple
- dbpedia:Isere
/mapping_language
For easier mappings you can use the following chrome extension developed
by Andrea Di Menna
https://github.com/dbpedia/mappings_chrome_extension
I installed the extension, but
http://mappings.dbpedia.org/server/statistics/en/?show=10 doesn’t
show the Infobox spaceflight, which
Tim,
2014-05-06 18:02 GMT+02:00 Timothy Lebo le...@rpi.edu:
Andrea,
On May 6, 2014, at 11:52 AM, Andrea Di Menna ninn...@gmail.com wrote:
Hi Tim,
yes you need to have editor rights on the mappings wiki.
Someone will enable you as soon as possible (@Dimitris?).
Eagerly awaiting privs
Hi Bernand,
in this specific case I would make sure the correct ontology property [1]
is used in the mappings where [2] is currently used, that is [3].
After that we should remove [2] as it is mispelled and a correct property
already exists.
Apart from this specific case, I am not sure whether
Hi Sands,
as far as I know there is no strategy for extracting entity types for
elements in Wikipedia lists articles.
In my opinion it would be interesting to develop such an idea.
Main tasks would be:
- correctly identify lists articles (only based on article title?)
- extract elements from
Hi all,
We just pushed a *VERY BASIC* chrome extension in Git to help mapping
editors start a new template mapping.
https://github.com/dbpedia/mappings_chrome_extension
https://github.com/dbpedia/mappings_chrome_extension
It is enough to install the extension from chrome://extensions (Load
Great Dimitris :-)
Uros if you need help with the statistics please let us know. Great to see
mappings in a new language!
On 23 Dec 2013 07:30, Dimitris Kontokostas jimk...@gmail.com wrote:
Hi Uros,
Sorry for the delay but we had some technical problems.
Mappings are now enabled for Serbian
dump files.
Pretty slow, isn't it?
Cheers!
Andrea
On Tue, Dec 3, 2013 at 2:30 PM, Andrea Di Menna ninn...@gmail.com wrote:
I guess Paul wanted to know which book is cited by one wikipedia page
(e.g.
page A cites book x).
If I am not wrong by asking template transclusions you only get
a successful web crawler is keeping it
simple.
On Dec 5, 2013 4:23 AM, Andrea Di Menna ninn...@gmail.com wrote:
2013/12/4 Paul Houle ontolo...@gmail.com
I think I could get this data out of some API, but there are great
HTML 5 parsing libraries now, so a link extractor from HTML can
Hi Dario,
the dataset you are using is extracted by
the org.dbpedia.extraction.mappings.PageLinksExtractor [1].
This extractor collects internal wiki links [2] from Wikipedia content
articles (that is, wikipedia pages which belong to the Main namespace [3])
to other wikipedia pages (please note I
,
the dump branch seems to work.
It would be nice to add the extractor framework to Ivy as well. sbt uses
Ivy if I am not mistaken.
Thanks for the quick help,
Karsten
On 19 November 2013 12:27, Andrea Di Menna ninn...@gmail.com wrote:
Hi Karsten,
the code used to build DBpedia 3.9 (which has
with
Jona when. Or would Andrea be up for the challenge?
Also, is this magic word within the template definitions or within the
disambiguation pages?
On Mon, Oct 21, 2013 at 8:26 AM, Andrea Di Menna ninn...@gmail.comwrote:
On top of this:
Looks like MediaWiki is currently superseding
Hi all,
I have noticed there are a lot of pages in the enwiki which are not
detected as disambiguation pages by the DBpedia EF.
This is because of the not-maintained map of disambiguation templates in [1]
I am working on the suggestion which appears in [1] and parsing
] http://www.mail-archive.com/wikitech-l@lists.wikimedia.org/msg69704.html
2013/10/21 Andrea Di Menna ninn...@gmail.com
Hi all,
I have noticed there are a lot of pages in the enwiki which are not
detected as disambiguation pages by the DBpedia EF.
This is because of the not-maintained map
Hi Yury,
this page [1] will show you which is the current status of the en template
mappings.
There are stats about the percentage of mapped properties.
Hope this helps.
Cheers
Andrea
[1] http://mappings.dbpedia.org/server/statistics/en/?show=10
2013/10/1 Yury Katkov katkov.ju...@gmail.com
Hi guys,
this is really well done :-)
I will take a look at it as soon as possible.
Bravo Hady and mentors :-)
Cheers
Andrea
2013/8/29 Jona Christopher Sahnwaldt j...@sahnwaldt.de
Hi Hady and all,
good stuff!
After a few changes [1][2][3][4], the mappings wiki now renders http
links to
Hi Julien,
in one case it is reading the file from the filesystem in the other case it
is downloading it from the mappings server.
If you check org.dbpedia.extraction.dump.extract.ConfigLoader you will see
that:
//language-independent val
private lazy val _ontology =
{
val
Ciao Julien,
can you please show us an example of wikipedia article with such data?
Thanks
Andrea
2013/6/27 Julien Plu julien@redaction-developpez.com
Hi,
I opened a new bug report on github :
https://github.com/dbpedia/extraction-framework/issues/76
Someone can tell me if it's
Hi Julien,
this should work if you write them with a starting lower case letter, i.e.
purchasingPowerParity
Cheers
Andrea
2013/6/18 Julien Plu julien@redaction-developpez.com
Hi,
When I see the validation page for the french mappings I have these
validation error :
Hi guys,
why is the Infobox_venue marked as ignored in the en mapping?
This infobox is used for arts, cultural and music venues, e.g.
http://en.wikipedia.org/wiki/Odeon_Leicester_Square
Can we remove it from the ignore list?
Cheers
Andrea
Ciao Alessio,
thanks for your suggestions about wrong mappings.
I have made some changes based on your list.
That gave me also the opportunity to notice that more than 100 Turkish
infoboxes are mapped to OntologyClass:Unknown :
That's really cool!
Thanks for letting us know.
Would it be possible to pre-populate template mapping page on the wiki when
hitting create with a skeleton containing current template properties?
That would ease a bit the mapping effort IMHO.
What do you think?
Cheers
Andrea
2013/6/13 Christian
Hi all,
what is the process to insert templates into the ignore list?
I am looking at Italian templates mapping and there are some which should
be added, e.g. Nota disambigua which is for disambiguations, or
Cassetto which just produces expandable boxes inside an article.
Please let me know.
Hi Fabio,
I am also sending SPARQL queries to the dbpedia endpoint using Ajax and I
cannot experience any problem.
There was only a short service interruption today, due to a server update.
Is it still not working?
Regards
Andrea
2013/6/7 fabio valsecchi fabiova...@gmail.com
Hi,
i'm
http://penrite-france.com/vn/vhytlwkmhgxsmnjsh.rjvzqsk
--
Get 100% visibility into Java/.NET code with AppDynamics Lite
It's a free troubleshooting tool designed for production
Get down to code-level detail for
Hi Andrea,
that makes me think of the Fair use policy that applies to the Virtuoso
instance used for SPARQL queries on DBpedia.
Please check [1], [2] and [3]
[1] http://wiki.dbpedia.org/OnlineAccess
[2] http://sourceforge.net/mailarchive/message.php?msg_id=30412765
[3]
a DBPedia live version ?
If yes, where I can found some details on how to use it, and if it's
possible to use it only for a language.
Best.
Julien.
2013/4/15 Andrea Di Menna ninn...@gmail.com
Can you reach the above URL?
Are you behind a proxy which might block the connection
Hi,
from what I understood the problem which will arise for DBpedia with the
introduction of Wikidata, is that actual values will not be available in
Wikipedia dumps anymore.
Instead, we will end up either finding nothing (see InterWiki links) or
finding wikidata parser functions, e.g.
Hi Dimitris,
I am not completely getting your point.
How would you handle the following example? (supposing the following will
be possible with Wikipedia/Wikidata)
Suppose you have
{{Infobox:Test
| name = {{#property:p45}}
}}
and a mapping
{{PropertyMapping | templateProperty = name |
to
canoncialize on a single best link.
Tom
On Mon, Mar 25, 2013 at 6:41 AM, Andrea Di Menna ninn...@gmail.comwrote:
Hi all,
it looks like there are actually some pages in Wikipedia which contain
wrong data, which is where the pages originate from in Freebase, e.g.
http://en.wikipedia.org/wiki
Sorry,
wrong information.
We should use Page Ids (
http://downloads.dbpedia.org/3.8/en/page_ids_en.nt.bz2)
I am going to try something.
Cheers
Andrea
2013/3/25 Andrea Di Menna ninn...@gmail.com
Hi,
we have article numeric ids in the quads file (as oldid parameter).
Jona, do you think
Hi Giosia,
some answers inline (maybe the other guys will give you more appropriate
answers)
2013/3/25 Giosia Gentile giosiagent...@gmail.com
Hi, I installed a local dbpedia lookup with success but now
I want recreate the new Index, using my own local DBpedia mirror.
From the documentation
Hi all,
what is the license for the DBpedia lookup service code hosted on Github
[1]?
Cannot see any license file.
Regards
Andrea
[1] https://github.com/dbpedia/lookup
--
Everyone hates slow websites. So do we.
Make
Hi Jona,
thanks for merging the pull request!
Anyway, couldn't we use percent encoding for Unicode code points which are
not allowed in N-Triples? (namely those outside the [#x20,#7E] range?
In this case we should get UTF-8 bytes and percent encode them.
For example, as far as I can see
if they are uncompressed first. It looks like Stanbol
has a problem with bz2.
https://issues.apache.org/jira/browse/STANBOL-804
http://markmail.org/message/67ivlyoxfqad6xoe
Cheers,
JC
On 21 March 2013 10:20, Andrea Di Menna ninn...@gmail.com wrote:
Hi Jona,
I compressed the nt
Hi Gaurav,
which are the permissions this folder?
/mnt/ebs/perl/framework/dbpedia/target/classes
Maybe not related, but the error message seems to suggest maven cannot
clean that directory.
Also, it might be useful to rerun the maven command with -eX switches.
Regards
Andrea
2013/3/20 gaurav
Hi,
I have been using Stanbol [1] to process DBpedia data files and build a
dbpedia Solr index.
Stanbol is using Jena TDB in order to load DBpedia files into a triple
store.
Unfortunately, almost all the DBpedia N-Triples files must be pre-processed
before being able to import them using Jena
standards etc. Could you give an
example where Jena complains about a DBpedia 3.8 file?
Cheers,
JC
On Wed, Mar 20, 2013 at 6:16 PM, Andrea Di Menna ninn...@gmail.com
wrote:
Hi,
I have been using Stanbol [1] to process DBpedia data files and build a
dbpedia Solr index.
Stanbol
Hi all,
since the Wikidata project's phase I aims to remove interwiki links from
wikipedia articles, I am expecting to start seeing problems with
canonicalized datasets.
From what I can see, DBpedia relies on such links to produces *_en_uris
datasets using the CanonicalizeUris script.
How are
language edition.
of course we'll have to create a new extractor and adapt a few scripts
but, other than I don't see any other big changes for now
Best,
Dimitris
On Fri, Mar 15, 2013 at 12:38 PM, Andrea Di Menna ninn...@gmail.comwrote:
Hi all,
since the Wikidata project's phase I aims
Ciao Amir,
As far as I understood the transitive option is a Virtuoso extension, that
means it is not defined in SPARQL 1.1 (used by Jena).
That means Jena will fail with a parser exception when trying to parse your
query.
You might want to read [1] which could have a solution for you.
Regards
Hi,
no suggestions?
Regards
Andrea
2013/2/20 Andrea Di Menna ninn...@gmail.com
Hi all,
I have been checking the DBpedia Extraction Framework and I am able to
re-run the extractors on wikipedia data.
Anyway I was wondering if I could set up a clone of the Mappings Wiki to
experiment a bit
Hi all,
I have been checking the DBpedia Extraction Framework and I am able to
re-run the extractors on wikipedia data.
Anyway I was wondering if I could set up a clone of the Mappings Wiki to
experiment a bit with the mappings.
Anything you can share?
Regards
Andrea
Hi Robert,
dbpprop properties are a simple 1:1 mapping of Wikipedia Infobox templates
properties, with no specific mapping nor special inference (I think there
are only some checks on data types).
On the contrary, dbpedia-owl properties are extracted from mappings of
Infoboxes defined on
Hi,
the DBpedia mapping seems correct.
From the Infobox documentation [1]:
*allegiance* – *optional* – the country or other power the person served.
Maybe a more appropriate dbpedia-owl property could be used for this
infobox property, e.g. [2]
I am going to change the mapping, but it is also
Hi all,
from this page [1] I know that only the english instance types data have
been loaded (en/instance_types_en) in the public SPARQL endpoint (
dbpedia.org/sparql) [no other languages].
I am seeing inconsistencies between the data in the SPARQL endpoint and the
raw data files.
Example:
check the resource page ([2]) again?
I didn't find any Athlete type definition there.
Best,
Dimitris
On Fri, Feb 8, 2013 at 5:11 PM, Andrea Di Menna ninn...@gmail.com wrote:
Hi all,
from this page [1] I know that only the english instance types data have
been loaded (en/instance_types_en
Hi Dimitris,
it seems everything is fine now :)
Thanks!
Andrea
2013/2/7 Dimitris Kontokostas jimk...@gmail.com
Hi Andrea,
The problem should be fixed now.
Can you please check and let us know if we missed something?
Best,
Dimitris
On Sat, Jan 26, 2013 at 1:32 PM, Andrea Di Menna
Hi Dimistris,
does not seem so:
http://en.wikipedia.org/wiki/Category:British_people_by_occupation?oldid=489570899
http://en.wikipedia.org/wiki/Category:British_people?oldid=494233120
Cheers
Andrea
2013/1/28 Dimitris Kontokostas kontokos...@informatik.uni-leipzig.de
Hi Ning,
Can you please
should try to fix them at the source (which is
Wikipedia) and on the next DBpedia release they will be fixed
Best,
Dimitris
On Mon, Jan 28, 2013 at 4:14 PM, Andrea Di Menna ninn...@gmail.comwrote:
Hi Dimistris,
does not seem so:
http://en.wikipedia.org/wiki
Hi Kingsley,
just wanted to add that sometimes.
The same query:
Thanks Kingsley.
Is the same reason why all the resources are not listed in the
http://dbpedia.org/page/G%C3%BCtersloh page as reverse
dbpedia-owl:birthPlace property?
Thanks
2013/1/28 Kingsley Idehen kide...@openlinksw.com
On 1/28/13 3:53 PM, Andrea Di Menna wrote:
Hi Kingsley,
just
Hi all,
it looks like there are some problems with the mappings website.
1) I have created a mapping for Infobox_comic_book_title [1] adding a few
property mappings.
Anyway they do not show up in the statistics page [2]
2) If I try to validate any of the mappings page nothing happens (using the
Hi all,
are there any guideline on how to debug the extraction framework using
IntelliJ IDEA?
Cannot find any document nor reference.
Many thanks
Andrea
--
Master Visual Studio, SharePoint, SQL, ASP.NET, C# 2012, HTML5,
Hi Nicolas,
As far as I can see the actors in the starring infobox property are
enclosed in a Plainlist template.
Maybe the extractor does not expect to find a template there? (Question for
the dbpedia devs)
Cheers
Andrea
Il giorno 15/gen/2013 18:36, Nicolas Torzec nicolas.tor...@gmail.com ha
Hi Dimitris,
2013/1/10 Dimitris Kontokostas jimk...@gmail.com
Hi Andrea,
On Thu, Jan 10, 2013 at 7:08 PM, Andrea Di Menna ninn...@gmail.comwrote:
Hi,
I have questions regarding group of infoboxes/templates which are
designed to be used together to provide additional data for a resource
Hi,
I have questions regarding group of infoboxes/templates which are designed
to be used together to provide additional data for a resource.
Specifically, I have been checking the Infobox Template:Starbox begin [1]
(and the other templates in the same group).
This template should be used in
Hi Dimitris,
since the wikipedia article has been corrected today, why is the
change not visible on Live DBpedia yet?
Is the system correctly working?
Thanks
Andrea
2013/1/9 Dimitris Kontokostas jimk...@gmail.com:
Hi Yves,
If you look at the correct revision [1] you'll see that the following
Hi Teddy,
you can try with the following
PREFIX owl: http://www.w3.org/2002/07/owl#
PREFIX xsd: http://www.w3.org/2001/XMLSchema#
PREFIX rdfs: http://www.w3.org/2000/01/rdf-schema#
PREFIX rdf: http://www.w3.org/1999/02/22-rdf-syntax-ns#
PREFIX foaf: http://xmlns.com/foaf/0.1/
PREFIX dc:
Hi all,
I noticed there is a problem with redirected infoboxes and the test extraction.
If I create a mapping for infobox A which is redirected from infobox
B, all the entities which actually use infobox B will not get mapped
in the test extraction.
Example:
Hi all,
I have added yesterday a new class OntologyClass:Noble [1] to start
mapping the Infobox_noble [2].
I can see the class here:
http://mappings.dbpedia.org/server/ontology/classes/#Noble
but not here:
http://live.dbpedia.org/ontology/Noble
What am I doing wrong?
Also I have started
Hi all,
my name is Andrea Di Menna (user: ninniuz) and I would like to
contribute to the en mappings project.
I am planning to use dbpedia data for a work project and would like to
refine the mappings as I can spot some errors :-)
Thank you very much,
Andrea
63 matches
Mail list logo