https://bugzilla.wikimedia.org/show_bug.cgi?id=62733

Nik Everett <[email protected]> changed:

           What    |Removed                     |Added
----------------------------------------------------------------------------
           See Also|                            |https://github.com/elastics
                   |                            |earch/elasticsearch/issues/
                   |                            |5648

--- Comment #2 from Nik Everett <[email protected]> ---
I believe this is caused by us not word breaking foo.bar into foo and bar.  The
solution to this, as I see it, is to use the word_break token filter _but_ to
do that I have to rebuild each analyzer with that filter.  That isn't easy
because now what I want the German analyzer I can ask for 
{"analyzer":{"text":{"type":"german"}}}
but to rebuild it I have to do this:
{"analyzer":{"text":{
            "filter": [
                "standard",
                "lowercase",
                "german_stop",
                "german_normalization",
                "light_german_stemmer"
            ],
            "tokenizer": "standard",
            "type": "custom"
}},"filter":{
        "german_stop": {
            "stopwords": [
                "denn",
...
                "eures",
                "dies",
                "bist",
                "kein"
            ],
            "type": "stop"
        }
}}

Except even that doesn't work because german_normalization isn't properly
exposed!  The pull request I've opened upstream exposes all the stuff I'd need
and it creates an endpoint on Elasticsearch designed to spit this back out for
easy customization.

-- 
You are receiving this mail because:
You are the assignee for the bug.
You are on the CC list for the bug.
_______________________________________________
Wikibugs-l mailing list
[email protected]
https://lists.wikimedia.org/mailman/listinfo/wikibugs-l

Reply via email to