On 2/23/11 2:18 PM, Rohana Rajapakse wrote:
Hi James,

It works for double quotes, but not for single quotes (i.e. fails for
'mistakes'). Is it a training issue then (not having cases with words
enclosed within single/double quotes.

I have noticed that your model file is much smaller than the model file
available to download. Is it because your training data set is smaller?
How does it affect tokenizing overall?

Are there training sets available to download?

Yes and no, you need a file with tokenization information to train the tokenizer.
In OpenNLP we use a sentence per line format, and the non-whitespace
separated tokens are separated by a special tag.
See our documentation for information about the format.

We observed that is is easy to use rules to detokenize correctly tokenized text.
For that reason I implemented a rule based detokenizer.

Now you just need some kind of tokenized text to produce a training file for our tokenizer. You might want to use the reuters corpus, or other freely available english language corpora. If you have access to the reuters corpus I suggest that we go through the steps to train the
tokenizer with it.

Jörn

Reply via email to