I ran the command line evaluator several times on tokenized/untokenized and 
large/small input but get no results (see below). The model appears to be 
finding tokens quite well, I'd just like to evaluate *how* well:

opennlp TokenNameFinderEvaluator  -data some-data.txt -model a-model.bin
Loading Token Name Finder model ... done (0.111s)


Average: 104.2 sent/s
Total: 15 sent
Runtime: 0.144s

Precision: 0.0
Recall: 0.0
F-Measure: -1.0

Now on a larger set of data:

opennlp TokenNameFinderEvaluator -encoding latin1 -data /tmp/db-raw.txt -model 
a-model.bin
Loading Token Name Finder model ... done (0.156s)
current: 364.9 sent/s avg: 364.9 sent/s total: 366 sent
current: 427.4 sent/s avg: 396.1 sent/s total: 793 sent


Average: 477.7 sent/s
Total: 1434 sent
Runtime: 3.002s

Precision: 0.0
Recall: 0.0
F-Measure: -1.0



What am I doing wrong?


Thanks

Reply via email to