On 12.03.15 10:03, Itamar Syn-Hershko wrote:
See http://www.elastic.co/guide/en/elasticsearch/reference/current/analysis-length-tokenfilter.html

Unfortunately the length token filter also doesn't filter out these immense terms. See my example from https://gist.github.com/Hocdoc/68b5fcf8819a51816b53 : I have created a length filter for terms greater than 5000 (characters? bytes?) but still get the exception when using the icu_normalizer :

|IllegalArgumentException:  Document  contains at least one immense term
in field="message"  (whose UTF8 encoding is longer than the max length32766),|

( length of this message value is 3728 Bytes UTF8-encoded)



--
You received this message because you are subscribed to the Google Groups 
"elasticsearch" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to [email protected].
To view this discussion on the web visit 
https://groups.google.com/d/msgid/elasticsearch/5501637E.2070400%40gmail.com.
For more options, visit https://groups.google.com/d/optout.

Reply via email to