[ 
https://issues.apache.org/jira/browse/LUCENE-1166?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12569708#action_12569708
 ] 

Thomas Peuss commented on LUCENE-1166:
--------------------------------------

bq. But I'm wondering if a similar approach could be used for, say, word 
segmentation in Chinese? That is, iterate through a string of Chinese 
characters, buffering them and looking up the buffered string in a Chinese 
dictionary. Once there is a dictionary match, and the addition of the following 
character results in a string that has no entry in the dictionary, that 
previous buffered string can be considered a word/token. I'm not sure if your 
patch does something like this, but if it does, I am wondering if it is general 
enough that what you did can be used (as the basis of) word segmentation for 
Chinese, and thus for a Chinese Analyzer that's not just a dump n-gram Analyzer 
(which is what we have today).

Currently the code adds a token to the stream when an n-gram from the current 
token in the token stream matches a word in the dictionary (I am only speaking 
about the DumbCompoundWordTokenFilter because I doubt that there exist 
hyphenation patterns for Chinese languages). I don't know much about the 
structure of Chinese characters to answer this questions in detail. You can 
have a look at the test-case in the patch to see how the filters work.




> A tokenfilter to decompose compound words
> -----------------------------------------
>
>                 Key: LUCENE-1166
>                 URL: https://issues.apache.org/jira/browse/LUCENE-1166
>             Project: Lucene - Java
>          Issue Type: New Feature
>          Components: Analysis
>            Reporter: Thomas Peuss
>         Attachments: CompoundTokenFilter.patch, CompoundTokenFilter.patch, 
> CompoundTokenFilter.patch, de.xml, hyphenation.dtd
>
>
> A tokenfilter to decompose compound words you find in many germanic languages 
> (like German, Swedish, ...) into single tokens.
> An example: Donaudampfschiff would be decomposed to Donau, dampf, schiff so 
> that you can find the word even when you only enter "Schiff".
> I use the hyphenation code from the Apache XML project FOP 
> (http://xmlgraphics.apache.org/fop/) to do the first step of decomposition. 
> Currently I use the FOP jars directly. I only use a handful of classes from 
> the FOP project.
> My question now:
> Would it be OK to copy this classes over to the Lucene project (renaming the 
> packages of course) or should I stick with the dependency to the FOP jars? 
> The FOP code uses the ASF V2 license as well.
> What do you think?

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.


---------------------------------------------------------------------
To unsubscribe, e-mail: [EMAIL PROTECTED]
For additional commands, e-mail: [EMAIL PROTECTED]

Reply via email to