[ 
https://issues.apache.org/jira/browse/LUCENE-400?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steven Rowe updated LUCENE-400:
-------------------------------

    Attachment: LUCENE-400.patch

Repackaged these four files as a patch, with the following modifications to the 
code:

* Renamed files and variables to refer to "n-grams" as "shingles", to avoid 
confusion with the character-level n-gram code already in Lucene's sandbox
* Placed code in the o.a.l.analysis.shingle package
* Converted commons-collections FIFO usages to LinkedLists
* Removed @author from javadocs
* Changed deprecated Lucene API usages to alternate forms; addressed all 
compilation warnings
* Changed code style to conform to Lucene conventions
* Changed field setters to return null instead of a reference to the class 
instance, then changed instantiations to use individual setter calls instead of 
the chained calling style
* Added ASF license to each file

All tests pass.

Although I left in the ShingleAnalyzerWrapper and its test in the patch, no 
other Lucene filter (AFAICT) has such a filter wrapping facility.  My vote is 
to remove these two files.

> NGramFilter -- construct n-grams from a TokenStream
> ---------------------------------------------------
>
>                 Key: LUCENE-400
>                 URL: https://issues.apache.org/jira/browse/LUCENE-400
>             Project: Lucene - Java
>          Issue Type: Improvement
>          Components: Analysis
>    Affects Versions: unspecified
>         Environment: Operating System: All
> Platform: All
>            Reporter: Sebastian Kirsch
>            Priority: Minor
>         Attachments: LUCENE-400.patch, NGramAnalyzerWrapper.java, 
> NGramAnalyzerWrapperTest.java, NGramFilter.java, NGramFilterTest.java
>
>
> This filter constructs n-grams (token combinations up to a fixed size, 
> sometimes
> called "shingles") from a token stream.
> The filter sets start offsets, end offsets and position increments, so
> highlighting and phrase queries should work.
> Position increments > 1 in the input stream are replaced by filler tokens
> (tokens with termText "_" and endOffset - startOffset = 0) in the output
> n-grams. (Position increments > 1 in the input stream are usually caused by
> removing some tokens, eg. stopwords, from a stream.)
> The filter uses CircularFifoBuffer and UnboundedFifoBuffer from Apache
> Commons-Collections.
> Filter, test case and an analyzer are attached.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.


---------------------------------------------------------------------
To unsubscribe, e-mail: [EMAIL PROTECTED]
For additional commands, e-mail: [EMAIL PROTECTED]

Reply via email to