Sorry I am a n00b when it comes to ES. How would I go about that? On Saturday, June 14, 2014 1:02:45 PM UTC-4, David Pilato wrote: > > May be you should gist your logs and copy the link here? > > -- > *David Pilato* | *Technical Advocate* | *Elasticsearch.com* > @dadoonet <https://twitter.com/dadoonet> | @elasticsearchfr > <https://twitter.com/elasticsearchfr> > > > Le 14 juin 2014 à 18:44:07, John Luko ([email protected] <javascript:>) a > écrit: > > Is there a particular log I should look at? In elasticsearch.log I am > seeing a lot of parser failures for a timestamp. > > On Saturday, June 14, 2014 12:26:47 PM UTC-4, David Pilato wrote: >> >> First guess is that you are running out of memory or so. >> What can you see in elasticsearch logs? >> >> Don't you see some WARN about GC? >> >> -- >> David ;-) >> Twitter : @dadoonet / @elasticsearchfr / @scrutmydocs >> >> >> Le 14 juin 2014 à 18:06, John Luko <[email protected]> a écrit : >> >> We are running Elasticsearch with three nodes. The problem we keep >> running into is eventually Kibana begins to run extremely slow and soon >> after becomes unresponsive. If you hit refresh the wheels just keep >> spinning and nothing gets displayed. Restarting the cluster appears to >> correct this issue, but within 24 hours it begins again. Below is our >> health: >> >> { >> "cluster_name" : "elasticsearch", >> "status" : "green", >> "timed_out" : false, >> "number_of_nodes" : 4, >> "number_of_data_nodes" : 3, >> "active_primary_shards" : 341, >> "active_shards" : 682, >> "relocating_shards" : 2, >> "initializing_shards" : 0, >> "unassigned_shards" : 0 >> } >> >> When we initially set everything up we had a single node and we did have >> a span of three weeks where we didn't have to restart the Elasticsearch >> service. When we begin to pump netflow data in we began to have issues. I >> thought perhaps because we had logstash and ES running on one node that was >> causing the issue. Thus we added two virtual nodes and had one of them >> host the logstash for just Netflow. I thought with the clustering the >> issue would be resolved, but sadly I still have to start and stop the >> services everyday. When I look back at the data it flows in the entire >> time until flat lining for a few hours and then picks up again once I >> restart the services. >> >> Thanks in advance! >> -- >> You received this message because you are subscribed to the Google Groups >> "elasticsearch" group. >> To unsubscribe from this group and stop receiving emails from it, send an >> email to [email protected]. >> To view this discussion on the web visit >> https://groups.google.com/d/msgid/elasticsearch/d088f293-9dae-475a-a15e-076b3b773680%40googlegroups.com >> >> <https://groups.google.com/d/msgid/elasticsearch/d088f293-9dae-475a-a15e-076b3b773680%40googlegroups.com?utm_medium=email&utm_source=footer> >> . >> For more options, visit https://groups.google.com/d/optout. >> > -- > You received this message because you are subscribed to the Google Groups > "elasticsearch" group. > To unsubscribe from this group and stop receiving emails from it, send an > email to [email protected] <javascript:>. > To view this discussion on the web visit > https://groups.google.com/d/msgid/elasticsearch/9a357da0-9ac6-4aaf-853d-f76f453c27d0%40googlegroups.com > > <https://groups.google.com/d/msgid/elasticsearch/9a357da0-9ac6-4aaf-853d-f76f453c27d0%40googlegroups.com?utm_medium=email&utm_source=footer> > . > For more options, visit https://groups.google.com/d/optout. > >
-- You received this message because you are subscribed to the Google Groups "elasticsearch" group. To unsubscribe from this group and stop receiving emails from it, send an email to [email protected]. To view this discussion on the web visit https://groups.google.com/d/msgid/elasticsearch/6908936b-02e7-46f9-873d-9fe3f2d1d152%40googlegroups.com. For more options, visit https://groups.google.com/d/optout.
