On 01/10/2014 18:42, Steve Rowe wrote:
Paul,

Boilerplate upgrade recommendation: consider using the most recent Lucene 
release (4.10.1) - it’s the most stable, performant, and featureful release 
available, and many bugs have been fixed since the 4.1 release.
Yeah sure, I did try this and hit a load of errors but I certainly will do so.
FYI, StandardTokenizer doesn’t find word boundaries for Chinese, Japanese, 
Korean, Thai, and other languages that don’t use whitespace to denote word 
boundaries, except those around punctuation.  Note that Lucene 4.1 does have 
specialized tokenizers for Simplified Chinese and Japanese: the smartcn and 
kuromoji analysis modules, respectively.
So for Chinese, Japanese, Korean, Thai etc its just identifying that the chars are from said language, and then we can do something clever with it with subsequent filters such as CJBigramFilter right ? My big trouble is my code is meant to deal with any language and I dont know what language it in except by looking at the characters themselves AND i also have to deal with stuff that contains symbols, funny punctuation etc
It is possible to construct a tokenizer just based on pure java code - there 
are several examples of this in Lucene 4.1, see e.g. PatternTokenizer, and 
CharTokenizer and its subclasses WhitespaceTokenizer and LetterTokenizer.

Ah yes I discovered this today, what I would really like is a version of the jflex StandardTokenizer but written in pure Java making it easier to tweak it, but I'm a little concerned that If I naively write it from scratch I may create something that doesnt perform very well.

Paul

---------------------------------------------------------------------
To unsubscribe, e-mail: java-user-unsubscr...@lucene.apache.org
For additional commands, e-mail: java-user-h...@lucene.apache.org

Reply via email to