Hello, This recent issue is exactly referring to what you need - you may want to continue the discussion there, perhaps ask the commenter to make the code public on github for experiments?
https://github.com/apache/lucene/issues/13065 Dawid On Sun, Feb 11, 2024 at 8:42 AM _ SATNAM <satnamsingh9...@gmail.com> wrote: > Hello, Lucene community > The project I'm working on requires me to include a Hebrew analyzer, > therefore I searched through the OpenSearch and Elasticsearch documents but > couldn't find anything relevant. I would want to know if the Lucene > community has any plans to integrate Hebrew Tockenizer in the future. > Additionally, I would like to know what obstacles some of our users had > while trying to construct a Hebrew analyzer and whether it is possible to > make it work with stopwords, synonyms, lemmitizers, and other features. > Additionally, I discovered a github repository where users claim to have > created a Hebrew plugin for elastic search, but it appears to be out of > date, and when I attempt to add it to my elastic, the links break. What > other option do we have to fulfill this requirement? > Warm regards >