[ 
https://issues.apache.org/jira/browse/STANBOL-652?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13401389#comment-13401389
 ] 

Rupert Westenthaler commented on STANBOL-652:
---------------------------------------------

Hi Pablo

I think the best think to start would be to refactor the benchmarking module so 
that is can return Benchmark results as RDF

I would even suggest to convert the Benchmarks to RDF (but still supporting the 
Text based format currently used)

    urn:benchmark1 a sbc:Benchmark
        sbc:content "The INPUT text"
        sbc:expects urn:annotation1
        sbc:expects urn:annotation2
        sbc:complain urn:annotation3

    urn:annotation1 a sbc:Annotation
        entity <entity-URI>

NOTE: we might add different sbc:Annotation types in future (e.g. to validate 
Named Entities detection by NER tools or the detection of the correct language)


Results of a Benchmark execution

   urn:benchmarkExecution1 a BenchmarkExecution 
        sbc:benchmark urn:benchmark1
        sbc:num-expects 2^^xsd:int
        sbc:succeeded-expects 1^^xsd:int
        sbc:num-complains 1^^xsd:int
        sbc:succeeded-expects 1^^xsd:int
        sbc:content-item <urn:contentitem-{UUID}>
        dc:created ... and other metadata
        
and the detailed results for Annotations

    urn:result a sbc:BenchmarkResult
        sbc:result-of urn:benchmarkExecution1
        sbc:about urn:annotation1
        sbc:state sbc:benchmark-state-succeeded (or sbc:benchmark-state-failed)
        sbc:enhancement <urn:enhancement-{UUID}>, <urn:enhancement-{UUID}>     

Note that a BenchmarkResult may refer to multiple enhancements as a 
expected/complained Entity could be suggested by several EntityAnnotation.


The Idea would be that the "BenchmarkExecution" provides most of the 
information needed for statistical analyses and the BenchmarkResults allow to 
directly go to the Enhancements and look into details.

Batch processing an Analyses could than be easily added - both on the client 
side and the server side.

WDYT
Rupert
                
> Benchmark should report evaluation summary
> ------------------------------------------
>
>                 Key: STANBOL-652
>                 URL: https://issues.apache.org/jira/browse/STANBOL-652
>             Project: Stanbol
>          Issue Type: Improvement
>          Components: Testing
>            Reporter: Pablo Mendes
>            Priority: Minor
>              Labels: benchmark, evaluation
>
>  The SBC is a nice way to perform manual inspection of the behavior of the 
> enhancement chain for different examples in the evaluation dataset. However, 
> for evaluations with several hundreds of examples, it would be interesting to 
> have scores that summarize the performance for the entire
>  dataset. For example, precision, recall and F1. An evaluation dataset is 
> available here in BDL: http://spotlight.dbpedia.org/download/stanbol/

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira

        

Reply via email to