Are you using 1.5.3? Can you send a small sample?

Em segunda-feira, 16 de fevereiro de 2015, Richard Head Jr.
<[email protected]> escreveu:

> I ran the command line evaluator several times on tokenized/untokenized
> and large/small input but get no results (see below). The model appears to
> be finding tokens quite well, I'd just like to evaluate *how* well:
>
> opennlp TokenNameFinderEvaluator  -data some-data.txt -model a-model.bin
> Loading Token Name Finder model ... done (0.111s)
>
>
> Average: 104.2 sent/s
> Total: 15 sent
> Runtime: 0.144s
>
> Precision: 0.0
> Recall: 0.0
> F-Measure: -1.0
>
> Now on a larger set of data:
>
> opennlp TokenNameFinderEvaluator -encoding latin1 -data /tmp/db-raw.txt
> -model a-model.bin
> Loading Token Name Finder model ... done (0.156s)
> current: 364.9 sent/s avg: 364.9 sent/s total: 366 sent
> current: 427.4 sent/s avg: 396.1 sent/s total: 793 sent
>
>
> Average: 477.7 sent/s
> Total: 1434 sent
> Runtime: 3.002s
>
> Precision: 0.0
> Recall: 0.0
> F-Measure: -1.0
>
>
>
> What am I doing wrong?
>
>
> Thanks
>


-- 
William Colen

Reply via email to