Can you provide some specs on your cluster, OS, RAM, heap, disk, java and ES versions? Are you using parent/child relationships, TTLs, large facet or other queries?
(Also, your elaborate legalese signature is kind of moot given you're posting to a public mailing list :p) Regards, Mark Walkom Infrastructure Engineer Campaign Monitor email: [email protected] web: www.campaignmonitor.com On 29 May 2014 07:27, Paul Sanwald <[email protected]> wrote: > Hi Everyone, > We are seeing continual OOM exceptions on one of our 1.1.0 > elasticsearch clusters, the index is ~30GB, quite small. I'm trying to work > out the root cause via heap dump analysis, but not having a lot of luck. I > don't want to include a bunch of unnecessary info, but the stacktrace we're > seeing is pasted below. Has anyone seen this before? I've been using the > cluster stats and node stats APIs to try and find a smoking gun, but I'm > not seeing anything that looks out of the ordinary. > > Any ideas? > > 14/05/27 20:37:08 WARN transport.netty: [Strongarm] Failed to send error > message back to client for action [search/phase/query] > java.lang.OutOfMemoryError: GC overhead limit exceeded > 14/05/27 20:37:08 WARN transport.netty: [Strongarm] Actual Exception > org.elasticsearch.search.query.QueryPhaseExecutionException: > [eventdata][2]: q > uery[ConstantScore(*:*)],from[0],size[0]: Query Failed [Failed to execute > main > query] > at > org.elasticsearch.search.query.QueryPhase.execute(QueryPhase.java:1 > 27) > at > org.elasticsearch.search.SearchService.executeQueryPhase(SearchService.java:257) > at > org.elasticsearch.search.action.SearchServiceTransportAction$SearchQueryTransportHandler.messageReceived(SearchServiceTransportAction.java:623) > at > org.elasticsearch.search.action.SearchServiceTransportAction$SearchQueryTransportHandler.messageReceived(SearchServiceTransportAction.java:612) > at > org.elasticsearch.transport.netty.MessageChannelHandler$RequestHandler.run(MessageChannelHandler.java:270) > at > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) > at > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) > at java.lang.Thread.run(Thread.java:722) > Caused by: java.lang.OutOfMemoryError: GC overhead limit exceeded > > *Important Notice:* The information contained in or attached to this > email message is confidential and proprietary information of RedOwl > Analytics, Inc., and by opening this email or any attachment the recipient > agrees to keep such information strictly confidential and not to use or > disclose the information other than as expressly authorized by RedOwl > Analytics, Inc. If you are not the intended recipient, please be aware > that any use, printing, copying, disclosure, dissemination, or the taking > of any act in reliance on this communication or the information contained > herein is strictly prohibited. If you think that you have received this > email message in error, please delete it and notify the sender. > > -- > You received this message because you are subscribed to the Google Groups > "elasticsearch" group. > To unsubscribe from this group and stop receiving emails from it, send an > email to [email protected]. > To view this discussion on the web visit > https://groups.google.com/d/msgid/elasticsearch/e6634ad4-619f-4f24-8287-d3bc97722a88%40googlegroups.com<https://groups.google.com/d/msgid/elasticsearch/e6634ad4-619f-4f24-8287-d3bc97722a88%40googlegroups.com?utm_medium=email&utm_source=footer> > . > For more options, visit https://groups.google.com/d/optout. > -- You received this message because you are subscribed to the Google Groups "elasticsearch" group. To unsubscribe from this group and stop receiving emails from it, send an email to [email protected]. To view this discussion on the web visit https://groups.google.com/d/msgid/elasticsearch/CAEM624b9vRvomsmvN0LVbTNgg4pgtPvc7gRjB-7L0GqJtG02ug%40mail.gmail.com. For more options, visit https://groups.google.com/d/optout.
