Adrien Grand created LUCENE-5042:
------------------------------------

             Summary: Improve NGramTokenizer
                 Key: LUCENE-5042
                 URL: https://issues.apache.org/jira/browse/LUCENE-5042
             Project: Lucene - Core
          Issue Type: Improvement
            Reporter: Adrien Grand
            Assignee: Adrien Grand
             Fix For: 5.0, 4.4


Now that we fixed NGramTokenizer and NGramTokenFilter to not produce corrupt 
token streams, the only way to have "true" offsets for n-grams is to use the 
tokenizer (the filter emits the offsets of the original token).

Yet, our NGramTokenizer has a few flaws, in particular:
 - it doesn't have the ability to pre-tokenize the input stream, for example on 
whitespaces,
 - it doesn't play nice with surrogate pairs.

Since we already broke backward compatibility for it in 4.4, I'd like to also 
fix these issues before we release.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

---------------------------------------------------------------------
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org

Reply via email to