On Tue, 24 Jun 2008 04:54:46 -0700 (PDT)
Otis Gospodnetic <[EMAIL PROTECTED]> wrote:
> One tokenizer is followed by filters. I think this all might be a bit
> clearer if you read the chapter about Analyzers in Lucene in Action if you
> have a copy. I think if you try to break down that "the result of all this
> passed to " into something more concrete and real you will see how things
> (should) work.
thanks Otis, from this and Ryan's previous reply I understand I was mistaken on
how I was seeing the process - i was expecting the filters / tokenizers to work
as processes with the output of one going to the input of the next , in the
order shown in fieldType definition. .. now that I write this i remember
reading some posts on this list about doing something like this ... open-pipe ?
anyway, it makes sense...not what I was hoping for, but it's what I have to
work with.
Now, if only I can get n-gram to work with search terms > minGramSize :P
Thanks for your time, help and recommendation of Lucene in Action.
B
_________________________
{Beto|Norberto|Numard} Meijome
"The greatest dangers to liberty lurk in insidious encroachment by men of zeal,
well-meaning but without understanding." Justice Louis D. Brandeis
I speak for myself, not my employer. Contents may be hot. Slippery when wet.
Reading disclaimers makes you go blind. Writing them is worse. You have been
Warned.