Re: Case Sensitivity - and more

2002-01-25 Thread Michal Plechawski
different token streams in 1.2, that was the real problem in 1.0. Michal - Original Message - From: "Brian Goetz" <[EMAIL PROTECTED]> To: "Lucene Users List" <[EMAIL PROTECTED]> Sent: Friday, January 25, 2002 11:24 AM Subject: Re: Case Sensitivity - and more

Re: Case Sensitivity - and more

2002-01-25 Thread Brian Goetz
> Ok, maybe I misled a point a bit. But Brian's proposal as I see it was to > _group_ two tokenizers that differ in a single thing. I don't think that's what I was proposing... I was recognizing that sometimes the analysis process is a composite one, and I was advocating that the composition be

Re: Case Sensitivity - and more

2002-01-25 Thread Michal Plechawski
> Currently it is easy to use different analyzers for different purposes, no? > I'm not sure how Brian's proposal (bi-modal analyzers: tokenize only & > tokenize+normalize) addresses your needs. Ok, maybe I misled a point a bit. But Brian's proposal as I see it was to _group_ two tokenizers that

RE: Case Sensitivity - and more

2002-01-24 Thread Doug Cutting
> From: Michal Plechawski > > I think that Brian's idea is more flexible and extendable. In my > application, I need three or more kinds of analyzers: for > counting tfidf > statistics, for indexing (compute more, e.g. summaries) and > for document > classification (compute document-to-class ass

Re: Case Sensitivity - and more

2002-01-22 Thread Michal Plechawski
Hi, I have never written anything to the list but in fact, I am doing some development using Lucene. I think that Brian's idea is more flexible and extendable. In my application, I need three or more kinds of analyzers: for counting tfidf statistics, for indexing (compute more, e.g. summaries) an