This is actually a change in Lucene -- previously, the long term was 
silently dropped, now it raises an exception, see Lucene 
ticket https://issues.apache.org/jira/browse/LUCENE-5710

You might want to add a `length` filter to your analyzer 
(http://www.elasticsearch.org/guide/en/elasticsearch/reference/current/analysis-length-tokenfilter.html#analysis-length-tokenfilter).

All in all, it hints at some strange data, because such "immense" term 
shouldn't probably be in the index in the first place.

Karel

On Thursday, May 29, 2014 10:47:37 PM UTC+2, Jeff Dupont wrote:
>
> We’re running into a peculiar issue when updating indexes with content for 
> the document.
>
>
> "document contains at least one immense term in (whose utf8 encoding is 
> longer than the max length 32766), all of which were skipped. please 
> correct the analyzer to not produce such terms”
>
>
> I’m hoping that there’s a simple fix or setting that can resolve this.
>

-- 
You received this message because you are subscribed to the Google Groups 
"elasticsearch" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to [email protected].
To view this discussion on the web visit 
https://groups.google.com/d/msgid/elasticsearch/a91895cb-437a-4642-8734-4445bb420125%40googlegroups.com.
For more options, visit https://groups.google.com/d/optout.

Reply via email to