Are you using 1.5.3? Can you send a small sample? Em segunda-feira, 16 de fevereiro de 2015, Richard Head Jr. <[email protected]> escreveu:
> I ran the command line evaluator several times on tokenized/untokenized > and large/small input but get no results (see below). The model appears to > be finding tokens quite well, I'd just like to evaluate *how* well: > > opennlp TokenNameFinderEvaluator -data some-data.txt -model a-model.bin > Loading Token Name Finder model ... done (0.111s) > > > Average: 104.2 sent/s > Total: 15 sent > Runtime: 0.144s > > Precision: 0.0 > Recall: 0.0 > F-Measure: -1.0 > > Now on a larger set of data: > > opennlp TokenNameFinderEvaluator -encoding latin1 -data /tmp/db-raw.txt > -model a-model.bin > Loading Token Name Finder model ... done (0.156s) > current: 364.9 sent/s avg: 364.9 sent/s total: 366 sent > current: 427.4 sent/s avg: 396.1 sent/s total: 793 sent > > > Average: 477.7 sent/s > Total: 1434 sent > Runtime: 3.002s > > Precision: 0.0 > Recall: 0.0 > F-Measure: -1.0 > > > > What am I doing wrong? > > > Thanks > -- William Colen
