https://bugzilla.wikimedia.org/show_bug.cgi?id=62733
Nik Everett <[email protected]> changed: What |Removed |Added ---------------------------------------------------------------------------- See Also| |https://github.com/elastics | |earch/elasticsearch/issues/ | |5648 --- Comment #2 from Nik Everett <[email protected]> --- I believe this is caused by us not word breaking foo.bar into foo and bar. The solution to this, as I see it, is to use the word_break token filter _but_ to do that I have to rebuild each analyzer with that filter. That isn't easy because now what I want the German analyzer I can ask for {"analyzer":{"text":{"type":"german"}}} but to rebuild it I have to do this: {"analyzer":{"text":{ "filter": [ "standard", "lowercase", "german_stop", "german_normalization", "light_german_stemmer" ], "tokenizer": "standard", "type": "custom" }},"filter":{ "german_stop": { "stopwords": [ "denn", ... "eures", "dies", "bist", "kein" ], "type": "stop" } }} Except even that doesn't work because german_normalization isn't properly exposed! The pull request I've opened upstream exposes all the stuff I'd need and it creates an endpoint on Elasticsearch designed to spit this back out for easy customization. -- You are receiving this mail because: You are the assignee for the bug. You are on the CC list for the bug. _______________________________________________ Wikibugs-l mailing list [email protected] https://lists.wikimedia.org/mailman/listinfo/wikibugs-l
