NGramTokenizer to handle more than 1024 chars
---------------------------------------------
Key: LUCENE-1227
URL: https://issues.apache.org/jira/browse/LUCENE-1227
Project: Lucene - Java
Issue Type: Improvement
Components: contrib/*
Reporter: Hiroaki Kawai
Attachments: NGramTokenizer.patch
Current NGramTokenizer can't handle character stream that is longer than 1024.
This is too short for non-whitespace-separated languages.
I created a patch for this issues.
--
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.
---------------------------------------------------------------------
To unsubscribe, e-mail: [EMAIL PROTECTED]
For additional commands, e-mail: [EMAIL PROTECTED]