I am looking for a library that would do basic to reasonably smart  
tokenization of natural language strings.

Like, if fed something in English or French, it creates tokens for  
the things between the spaces, for Japanese, it deals with the non- 
spaced strings in a rule based fashion.

I think Lucene can do that and so montezuma would be a candidate (?),  
but I wonder if any of you has experience with such tools, especially  
for languages that do not use spaces.

Jean-Christophe Helary




_______________________________________________
Gardeners mailing list
[email protected]
http://www.lispniks.com/mailman/listinfo/gardeners

Reply via email to