Hi Simon,
I think we hit the same problem on our index.
I have one shard which sometimes give me this answer:
"_shards": {
      "total": 5,
      "successful": 4,
      "failed": 1,
      "failures": [
         {
            "index": "products",
            "shard": 0,
            "status": 500,
            "reason": "QueryPhaseExecutionException[[products][0]: 
query[filtered(filtered(id:finlandek short_title:finlandek title:finlandek 
category_name:finlandek description:finlandek category_name_path:finlandek 
brand_name:finlandek)->cache(stats_income7_global:[0 TO 
*]))->cache(org.elasticsearch.index.search.nested.NonNestedDocsFilter@3977d235)],from[0],size[10]:
 
Query Failed [Failed to execute main query]]; nested: EOFException[seek 
past EOF: 
MMapIndexInput(path=\"/data/es/elasticsearch-1.3.0-SNAPSHOT/data/elasticsearch/nodes/0/indices/products/0/index/_za6_es090_0.doc\")];
 
"
         }
      ]
   },

Is it still under investigation?

Le mercredi 2 avril 2014 14:01:21 UTC+2, simonw a écrit :
>
> hey,
>
> is it possible to look at this index / shard? do you still have it / can 
> you safe it for further investigations? You can ping me directly at simon 
> AT elasticsearch DOT com
>
> On Wednesday, April 2, 2014 11:23:38 AM UTC+2, Paweł Chabierski wrote:
>>
>> Few days ago we found we've got that same error when we search for data. 
>>
>> reason: "FetchPhaseExecutionException[[site_production][1]: 
>> query[ConstantScore(cache(_type:ademail))],from[0],size[648]: Fetch Failed 
>> [Failed to fetch doc id [9615533]]]; nested: EOFException[seek past EOF: 
>> MMapIndexInput(path="/opt/elasticsearch/data/naprawa/nodes/0/indices/site_production/1/index/_573oa.fdt")];
>>  
>>
>> After this error, elasticsearch stop search and don't return all results, 
>> even if they are returned properly in other query. Is there any way to fix 
>> this index? We try checkIndex from Luccene core library, after that we lost 
>> ~20 milions of documents but error still occurs. Also we try to restore 
>> index from snapshot, but error still occurs :/.
>>
>> W dniu sobota, 22 marca 2014 14:04:56 UTC+1 użytkownik Andrey Perminov 
>> napisał:
>>>
>>> We are using a small elasticsearch cluster of three nodes, version 
>>> 1.0.1. Each node has 7 GB RAM. Our software creates daily indexes for 
>>> storing it's data. Daily index is something around 5 GB. Unfortunately, for 
>>> a reason, Elasticsearch eats up all RAM and hangs the node, even though 
>>> heap size is set to 6 GB max. So we decided to use monit to restart it on 
>>> reaching memory limit of 90%. It works, but sometimes we got such errors:
>>>
>>> [2014-03-22 16:56:04,943][DEBUG][action.search.type       ] [es-00] 
>>> [product-22-03-2014][0], node[jbUDVzuvS5GTM7iOG8iwzQ], [P], s[STARTED]: 
>>> Failed to execute [org.elasticsearch.action.search.SearchRequest@687dc039]
>>> org.elasticsearch.search.fetch.FetchPhaseExecutionException: 
>>> [product-22-03-2014][0]: query[filtered(ToParentBlockJoinQuery 
>>> (filtered(history.created:[1392574921000 TO 
>>> *])->cache(_type:__history)))->cache(_type:product)],from[0],size[1000],sort[<custom:"history.created":
>>>  
>>> org.elasticsearch.index.search.nested.NestedFieldComparatorSource@15e4ece9>]:
>>>  
>>> Fetch Failed [Failed to fetch doc id [7263214]]
>>>         at 
>>> org.elasticsearch.search.fetch.FetchPhase.loadStoredFields(FetchPhase.java:230)
>>>         at 
>>> org.elasticsearch.search.fetch.FetchPhase.execute(FetchPhase.java:156)
>>>         at 
>>> org.elasticsearch.search.SearchService.executeFetchPhase(SearchService.java:332)
>>>         at 
>>> org.elasticsearch.search.action.SearchServiceTransportAction.sendExecuteFetch(SearchServiceTransportAction.java:304)
>>>         at 
>>> org.elasticsearch.action.search.type.TransportSearchQueryAndFetchAction$AsyncAction.sendExecuteFirstPhase(TransportSearchQueryAndFetchAction.java:71)
>>>         at 
>>> org.elasticsearch.action.search.type.TransportSearchTypeAction$BaseAsyncAction.performFirstPhase(TransportSearchTypeAction.java:216)
>>>         at 
>>> org.elasticsearch.action.search.type.TransportSearchTypeAction$BaseAsyncAction$4.run(TransportSearchTypeAction.java:292)
>>>         at java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown 
>>> Source)
>>>         at java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown 
>>> Source)
>>>         at java.lang.Thread.run(Unknown Source)
>>> Caused by: java.io.EOFException: seek past EOF: 
>>> MMapIndexInput(path="/opt/elasticsearch/main/nodes/0/indices/product-22-03-2014/0/index/_9lz.fdt")
>>>         at 
>>> org.apache.lucene.store.ByteBufferIndexInput.seek(ByteBufferIndexInput.java:174)
>>>         at 
>>> org.apache.lucene.codecs.compressing.CompressingStoredFieldsReader.visitDocument(CompressingStoredFieldsReader.java:229)
>>>         at 
>>> org.apache.lucene.index.SegmentReader.document(SegmentReader.java:276)
>>>         at 
>>> org.apache.lucene.index.BaseCompositeReader.document(BaseCompositeReader.java:110)
>>>         at 
>>> org.apache.lucene.search.IndexSearcher.doc(IndexSearcher.java:196)
>>>         at 
>>> org.elasticsearch.search.fetch.FetchPhase.loadStoredFields(FetchPhase.java:228)
>>>         ... 9 more
>>> [2014-03-22 16:56:04,944][DEBUG][action.search.type       ] [es-00] All 
>>> shards failed for phase: [query_fetch]
>>>
>>> According to our logs, this might happen when one or two nodes get 
>>> restarted. More strangely, same shard got corrupted on all nodes of 
>>> cluster. Why could this happen? How can we fix it? Can you suggest us how 
>>> to fix memory usage?
>>>
>>

-- 
You received this message because you are subscribed to the Google Groups 
"elasticsearch" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to [email protected].
To view this discussion on the web visit 
https://groups.google.com/d/msgid/elasticsearch/05ea6cde-87f2-4c0d-ac99-7d6aa769383c%40googlegroups.com.
For more options, visit https://groups.google.com/d/optout.

Reply via email to