[ 
https://issues.apache.org/jira/browse/STANBOL-1141?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Antonio David Pérez Morales updated STANBOL-1141:
-------------------------------------------------

    Description: 
Cross-document coreference resolution is the task of grouping the entity 
mentions in a collection of documents into sets that each represent a distinct 
entity. It is central to knowledge base construction and also useful for joint 
inference with other NLP components.
Wikilinks is one of the result of this task. Wikilinks dataset comprising of 40 
million mentions over 3 million entities. The method is based on finding 
hyperlinks to Wikipedia from a web crawl and using anchor text as mentions. The 
resource provides URLs of webpages, along with the anchor of the links, and the 
Wikipedia pages they link to. As provided, this dataset can be used to get all 
the surface strings that refer to a Wikipedia page, but further, it can be used 
to download the webpages and extract the context around the webpages

UMass (http://www.iesl.cs.umass.edu/) has created expanded versions of the 
dataset containing the following extra features:

* Complete webpage content (with cleaned DOM structure)
* Extracted context for the mentions
* Alignment to Freebase entities

The expanded dataset can be downloaded from 
http://iesl.cs.umass.edu/downloads/wiki-link/context-only/

A tool is needed for parsing this information and store it in any kind of 
storage consumible later within Stanbol. For the first version, it is possible 
to convert this dataset to RDF and store it in a triple store like JenaTDB. The 
goal of this task is to provide an API on the top of this store for easing the 
retrieval of entities' contextual data. So, "in disambiguation time", we can 
use the URI of the referenced entity to lookup for disambiguation contexts in 
Wikilinks


  was:
Cross-document coreference resolution is the task of grouping the entity 
mentions in a collection of documents into sets that each represent a distinct 
entity. It is central to knowledge base construction and also useful for joint 
inference with other NLP components.

Wikilinks is one of the result of this task. 
Wikilinks dataset comprising of 40 million mentions over 3 million entities. 
The method is based on finding hyperlinks to Wikipedia from a web crawl and 
using anchor text as mentions. In addition to providing large-scale labeled 
data without human effort, we are able to include many styles of text beyond 
newswire and many entity types beyond people.

UMass has created expanded versions of the dataset containing the following 
extra features:

* Complete webpage content (with cleaned DOM structure)
* Extracted context for the mentions
* Alignment to Freebase entities

The expanded dataset can be downloaded from 
http://iesl.cs.umass.edu/downloads/wiki-link/context-only/

A tool is needed in order to parser this information and store it in any type 
of storage like Jena TDB. 

Wikilinks provides information of documents with mentions to Freebase entities 
and this information can be used both to desambiguate and to merge with the 
Freebase information in order to have a large set of valuable data.


    
> Wikilinks Parser and TDB Generator
> ----------------------------------
>
>                 Key: STANBOL-1141
>                 URL: https://issues.apache.org/jira/browse/STANBOL-1141
>             Project: Stanbol
>          Issue Type: Sub-task
>          Components: Enhancer, Entityhub
>            Reporter: Antonio David Pérez Morales
>              Labels: freebase, jenatdb, wikilinks
>   Original Estimate: 336h
>  Remaining Estimate: 336h
>
> Cross-document coreference resolution is the task of grouping the entity 
> mentions in a collection of documents into sets that each represent a 
> distinct entity. It is central to knowledge base construction and also useful 
> for joint inference with other NLP components.
> Wikilinks is one of the result of this task. Wikilinks dataset comprising of 
> 40 million mentions over 3 million entities. The method is based on finding 
> hyperlinks to Wikipedia from a web crawl and using anchor text as mentions. 
> The resource provides URLs of webpages, along with the anchor of the links, 
> and the Wikipedia pages they link to. As provided, this dataset can be used 
> to get all the surface strings that refer to a Wikipedia page, but further, 
> it can be used to download the webpages and extract the context around the 
> webpages
> UMass (http://www.iesl.cs.umass.edu/) has created expanded versions of the 
> dataset containing the following extra features:
> * Complete webpage content (with cleaned DOM structure)
> * Extracted context for the mentions
> * Alignment to Freebase entities
> The expanded dataset can be downloaded from 
> http://iesl.cs.umass.edu/downloads/wiki-link/context-only/
> A tool is needed for parsing this information and store it in any kind of 
> storage consumible later within Stanbol. For the first version, it is 
> possible to convert this dataset to RDF and store it in a triple store like 
> JenaTDB. The goal of this task is to provide an API on the top of this store 
> for easing the retrieval of entities' contextual data. So, "in disambiguation 
> time", we can use the URI of the referenced entity to lookup for 
> disambiguation contexts in Wikilinks

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

Reply via email to