add icu-based tokenizer for unicode text segmentation
-----------------------------------------------------

                 Key: LUCENE-2414
                 URL: https://issues.apache.org/jira/browse/LUCENE-2414
             Project: Lucene - Java
          Issue Type: New Feature
          Components: contrib/*
    Affects Versions: 3.1
            Reporter: Robert Muir
             Fix For: 3.1


I pulled out the last part of LUCENE-1488, the tokenizer itself and cleaned it 
up some.

The idea is simple:
* First step is to divide text into writing system boundaries (scripts)
* You supply an ICUTokenizerConfig (or just use the default) which lets you 
tailor segmentation on a per-writing system basis.
* This tailoring can be any BreakIterator, so rule-based or dictionary-based or 
your own.

The default implementation (if you do not customize) is just to do UAX#29, but 
with tailorings for stuff with no clear word division:
* Thai (uses dictionary-based word breaking)
* Khmer, Myanmar, Lao (uses custom rules for syllabification)

Additionally as more of an example i have a tailoring for hebrew that treats 
the punctuation special. (People have asked before
for ways to make standardanalyzer treat dashes differently, etc)


-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.


---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to