Hi,
Dave Kor wrote:
>
> > Hm, a dictionary solution, it think.
>
> Fyi, I am currently working on a dictionary library
> (based on Lucene of course) for tokenizing and
> stemming the Chinese language. From what has been
> mentioned in this thread, it may be useful for the
> German language too.
> Hm, a dictionary solution, it think.
Fyi, I am currently working on a dictionary library
(based on Lucene of course) for tokenizing and
stemming the Chinese language. From what has been
mentioned in this thread, it may be useful for the
German language too.
I must warn you that it is not com
Hi,
Dmitry Serebrennikov wrote:
> Stemming by itself couldn't solve this problem, it seems, because I don't think it
>is designed for splitting compound > words. Yet, this seems like a common issue that
>people would run into constantly. So I was wandering:
> - Do German stemmers typically spli
Greetings, everyone.
We have been using Lucene for English texts for some time and that works really well.
But I recently spoke with someone from Germany and they had raised an issue with that
language that I wasn't sure how Lucene would be able to tackle. The example they used
was with a word