It seems so. Interestingly I can't find any mentions of WordDelimiterTokenFilter using google. Is it part of Solr codebase? On 13.06.2011, at 21:49, Em wrote:
> Hi, > > sounds like the WordDelimiterTokenFilter from Solr, doesn't it? > > Regards, > Em > > Am 13.06.2011 12:06, schrieb Denis Bazhenov: >> Some time ago I need to tune our home grown search engine based on lucene to >> perform well on product searches. Product search is search where users come >> with part of product name and we should find the product. >> >> The problem here is that users doesn't provide full model name. For instance >> id product model name is "Sony PRS-A9000QF", users frequently search for >> "PRS 9000", "9000QF" etc. >> >> The simple and straightforward solution to this problem is to tokenize model >> names on the different character type boundary. So for "Sony PRS-A9000QF" we >> will have 5 terms: "sony", "prs", "a", "9000" "qf". This solution could >> dramatically increase search sensitive (which is not a good thing in a >> general search), but works well in a specialized indexes. >> >> So a developed such a token filter. My question is there any interest in >> this solution for the community, and does it make sense to contribute it >> back? >> --- >> Denis Bazhenov <dot...@gmail.com> >> >> >> >> >> >> >> --------------------------------------------------------------------- >> To unsubscribe, e-mail: java-user-unsubscr...@lucene.apache.org >> For additional commands, e-mail: java-user-h...@lucene.apache.org >> >> > > --------------------------------------------------------------------- > To unsubscribe, e-mail: java-user-unsubscr...@lucene.apache.org > For additional commands, e-mail: java-user-h...@lucene.apache.org > --- Denis Bazhenov <dot...@gmail.com>