[ 
https://issues.apache.org/jira/browse/OPENNLP-59?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

William Colen resolved OPENNLP-59.
----------------------------------

    Resolution: Fixed

> Bad precision using FMeasure
> ----------------------------
>
>                 Key: OPENNLP-59
>                 URL: https://issues.apache.org/jira/browse/OPENNLP-59
>             Project: OpenNLP
>          Issue Type: Bug
>    Affects Versions: tools-1.5.1-incubating
>            Reporter: William Colen
>            Assignee: William Colen
>             Fix For: tools-1.5.1-incubating
>
>
> I noticed bad precision in FMeasure results. I think the issue is that the 
> current implementation is summing divisions. It computes the precision and 
> recall for every sample, and after adds the results for each sample to 
> compute the overall result. By doing that, the error related to each division 
> are summed and can impact the final result.
> I found the problem while implementing the ChunkerEvaluator. To verify the 
> evaluator I tried to compare the results we get using OpenNLP and the Perl 
> script conlleval available at 
> http://www.cnts.ua.ac.be/conll2000/chunking/output.html. The results were 
> always different if I process more than one sentence, because the 
> implementation was using FMeasure.updateScores() that was summing divisions.
> To solve that and have the same results provided by conll I basically stopped 
> using the Mean class.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.

Reply via email to