Upload your logs on gist.github.com and paste the link here. May be some more details here: http://www.elasticsearch.org/help/
-- David Pilato | Technical Advocate | Elasticsearch.com @dadoonet | @elasticsearchfr Le 14 juin 2014 à 19:06:46, John Luko ([email protected]) a écrit: Sorry I am a n00b when it comes to ES. How would I go about that? On Saturday, June 14, 2014 1:02:45 PM UTC-4, David Pilato wrote: May be you should gist your logs and copy the link here? -- David Pilato | Technical Advocate | Elasticsearch.com @dadoonet | @elasticsearchfr Le 14 juin 2014 à 18:44:07, John Luko ([email protected]) a écrit: Is there a particular log I should look at? In elasticsearch.log I am seeing a lot of parser failures for a timestamp. On Saturday, June 14, 2014 12:26:47 PM UTC-4, David Pilato wrote: First guess is that you are running out of memory or so. What can you see in elasticsearch logs? Don't you see some WARN about GC? -- David ;-) Twitter : @dadoonet / @elasticsearchfr / @scrutmydocs Le 14 juin 2014 à 18:06, John Luko <[email protected]> a écrit : We are running Elasticsearch with three nodes. The problem we keep running into is eventually Kibana begins to run extremely slow and soon after becomes unresponsive. If you hit refresh the wheels just keep spinning and nothing gets displayed. Restarting the cluster appears to correct this issue, but within 24 hours it begins again. Below is our health: { "cluster_name" : "elasticsearch", "status" : "green", "timed_out" : false, "number_of_nodes" : 4, "number_of_data_nodes" : 3, "active_primary_shards" : 341, "active_shards" : 682, "relocating_shards" : 2, "initializing_shards" : 0, "unassigned_shards" : 0 } When we initially set everything up we had a single node and we did have a span of three weeks where we didn't have to restart the Elasticsearch service. When we begin to pump netflow data in we began to have issues. I thought perhaps because we had logstash and ES running on one node that was causing the issue. Thus we added two virtual nodes and had one of them host the logstash for just Netflow. I thought with the clustering the issue would be resolved, but sadly I still have to start and stop the services everyday. When I look back at the data it flows in the entire time until flat lining for a few hours and then picks up again once I restart the services. Thanks in advance! -- You received this message because you are subscribed to the Google Groups "elasticsearch" group. To unsubscribe from this group and stop receiving emails from it, send an email to [email protected]. To view this discussion on the web visit https://groups.google.com/d/msgid/elasticsearch/d088f293-9dae-475a-a15e-076b3b773680%40googlegroups.com. For more options, visit https://groups.google.com/d/optout. -- You received this message because you are subscribed to the Google Groups "elasticsearch" group. To unsubscribe from this group and stop receiving emails from it, send an email to [email protected]. To view this discussion on the web visit https://groups.google.com/d/msgid/elasticsearch/9a357da0-9ac6-4aaf-853d-f76f453c27d0%40googlegroups.com. For more options, visit https://groups.google.com/d/optout. -- You received this message because you are subscribed to the Google Groups "elasticsearch" group. To unsubscribe from this group and stop receiving emails from it, send an email to [email protected]. To view this discussion on the web visit https://groups.google.com/d/msgid/elasticsearch/6908936b-02e7-46f9-873d-9fe3f2d1d152%40googlegroups.com. For more options, visit https://groups.google.com/d/optout. -- You received this message because you are subscribed to the Google Groups "elasticsearch" group. To unsubscribe from this group and stop receiving emails from it, send an email to [email protected]. To view this discussion on the web visit https://groups.google.com/d/msgid/elasticsearch/etPan.539c8371.12200854.a69e%40MacBook-Air-de-David.local. For more options, visit https://groups.google.com/d/optout.
