[ 
https://issues.apache.org/jira/browse/LUCENE-1166?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12569641#action_12569641
 ] 

Otis Gospodnetic commented on LUCENE-1166:
------------------------------------------

I haven't looked at the patch.
But I'm wondering if a similar approach could be used for, say, word 
segmentation in Chinese?
That is, iterate through a string of Chinese characters, buffering them and 
looking up the buffered string in a Chinese dictionary.  Once there is a 
dictionary match, and the addition of the following character results in a 
string that has no entry in the dictionary, that previous buffered string can 
be considered a word/token.

I'm not sure if your patch does something like this, but if it does, I am 
wondering if it is general enough that what you did can be used (as the basis 
of) word segmentation for Chinese, and thus for a Chinese Analyzer that's not 
just a dump n-gram Analyzer (which is what we have today).


> A tokenfilter to decompose compound words
> -----------------------------------------
>
>                 Key: LUCENE-1166
>                 URL: https://issues.apache.org/jira/browse/LUCENE-1166
>             Project: Lucene - Java
>          Issue Type: New Feature
>          Components: Analysis
>            Reporter: Thomas Peuss
>         Attachments: CompoundTokenFilter.patch, CompoundTokenFilter.patch, 
> CompoundTokenFilter.patch, de.xml, hyphenation.dtd
>
>
> A tokenfilter to decompose compound words you find in many germanic languages 
> (like German, Swedish, ...) into single tokens.
> An example: Donaudampfschiff would be decomposed to Donau, dampf, schiff so 
> that you can find the word even when you only enter "Schiff".
> I use the hyphenation code from the Apache XML project FOP 
> (http://xmlgraphics.apache.org/fop/) to do the first step of decomposition. 
> Currently I use the FOP jars directly. I only use a handful of classes from 
> the FOP project.
> My question now:
> Would it be OK to copy this classes over to the Lucene project (renaming the 
> packages of course) or should I stick with the dependency to the FOP jars? 
> The FOP code uses the ASF V2 license as well.
> What do you think?

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.


---------------------------------------------------------------------
To unsubscribe, e-mail: [EMAIL PROTECTED]
For additional commands, e-mail: [EMAIL PROTECTED]

Reply via email to