Adrien Grand created LUCENE-5042: ------------------------------------ Summary: Improve NGramTokenizer Key: LUCENE-5042 URL: https://issues.apache.org/jira/browse/LUCENE-5042 Project: Lucene - Core Issue Type: Improvement Reporter: Adrien Grand Assignee: Adrien Grand Fix For: 5.0, 4.4
Now that we fixed NGramTokenizer and NGramTokenFilter to not produce corrupt token streams, the only way to have "true" offsets for n-grams is to use the tokenizer (the filter emits the offsets of the original token). Yet, our NGramTokenizer has a few flaws, in particular: - it doesn't have the ability to pre-tokenize the input stream, for example on whitespaces, - it doesn't play nice with surrogate pairs. Since we already broke backward compatibility for it in 4.4, I'd like to also fix these issues before we release. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira --------------------------------------------------------------------- To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org