add simple japanese tokenizer, based on tinysegmenter
-----------------------------------------------------
Key: LUCENE-2522
URL: https://issues.apache.org/jira/browse/LUCENE-2522
Project: Lucene - Java
Issue Type: New Feature
Components: contrib/analyzers
Reporter: Robert Muir
Priority: Minor
TinySegmenter (http://www.chasen.org/~taku/software/TinySegmenter/) is a tiny
japanese segmenter.
It was ported to java/lucene by Kohei TAKETA <[email protected]>,
and is under friendly license terms (BSD, some files explicitly disclaim
copyright to the source code, giving a blessing instead)
Koji knows the author, and already contacted about incorporating into lucene:
{noformat}
I've contacted Takeda-san who is the creater of Java version of
TinySegmenter. He said he is happy if his program is part of Lucene.
He is a co-author of my book about Solr published in Japan, BTW. ;-)
{noformat}
--
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]