Thanks for the help thus far! I'll be posting the logs on Monday when I
get in the office, but in the mean time it has happened again. When I went
to Kibana I got an actual error that is pointing me in the right direction
I believe:
Oops! SearchPhaseExecutionException[Failed to execute phase [query], all
shards failed; shardFailures
{[CN48pxdyTUWBc8Iyft5vpQ][logstash-2014.06.15][2]:
EsRejectedExecutionException[rejected execution (queue capacity 1000) on
org.elasticsearch.action.search.type.TransportSearchTypeAction$BaseAsyncAction$4@b597ac6]}]
We have I believe 24 cores on our main box so I figure I can set the thread
pool size to 120 and the queue size to -1?
On Saturday, June 14, 2014 1:16:52 PM UTC-4, David Pilato wrote:
>
> Upload your logs on gist.github.com and paste the link here.
>
> May be some more details here: http://www.elasticsearch.org/help/
>
>
> --
> *David Pilato* | *Technical Advocate* | *Elasticsearch.com*
> @dadoonet <https://twitter.com/dadoonet> | @elasticsearchfr
> <https://twitter.com/elasticsearchfr>
>
>
> Le 14 juin 2014 à 19:06:46, John Luko ([email protected] <javascript:>) a
> écrit:
>
> Sorry I am a n00b when it comes to ES. How would I go about that?
>
> On Saturday, June 14, 2014 1:02:45 PM UTC-4, David Pilato wrote:
>>
>> May be you should gist your logs and copy the link here?
>>
>> --
>> *David Pilato* | *Technical Advocate* | *Elasticsearch.com*
>> @dadoonet <https://twitter.com/dadoonet> | @elasticsearchfr
>> <https://twitter.com/elasticsearchfr>
>>
>>
>> Le 14 juin 2014 à 18:44:07, John Luko ([email protected]) a écrit:
>>
>> Is there a particular log I should look at? In elasticsearch.log I am
>> seeing a lot of parser failures for a timestamp.
>>
>> On Saturday, June 14, 2014 12:26:47 PM UTC-4, David Pilato wrote:
>>>
>>> First guess is that you are running out of memory or so.
>>> What can you see in elasticsearch logs?
>>>
>>> Don't you see some WARN about GC?
>>>
>>> --
>>> David ;-)
>>> Twitter : @dadoonet / @elasticsearchfr / @scrutmydocs
>>>
>>>
>>> Le 14 juin 2014 à 18:06, John Luko <[email protected]> a écrit :
>>>
>>> We are running Elasticsearch with three nodes. The problem we keep
>>> running into is eventually Kibana begins to run extremely slow and soon
>>> after becomes unresponsive. If you hit refresh the wheels just keep
>>> spinning and nothing gets displayed. Restarting the cluster appears to
>>> correct this issue, but within 24 hours it begins again. Below is our
>>> health:
>>>
>>> {
>>> "cluster_name" : "elasticsearch",
>>> "status" : "green",
>>> "timed_out" : false,
>>> "number_of_nodes" : 4,
>>> "number_of_data_nodes" : 3,
>>> "active_primary_shards" : 341,
>>> "active_shards" : 682,
>>> "relocating_shards" : 2,
>>> "initializing_shards" : 0,
>>> "unassigned_shards" : 0
>>> }
>>>
>>> When we initially set everything up we had a single node and we did have
>>> a span of three weeks where we didn't have to restart the Elasticsearch
>>> service. When we begin to pump netflow data in we began to have issues. I
>>> thought perhaps because we had logstash and ES running on one node that was
>>> causing the issue. Thus we added two virtual nodes and had one of them
>>> host the logstash for just Netflow. I thought with the clustering the
>>> issue would be resolved, but sadly I still have to start and stop the
>>> services everyday. When I look back at the data it flows in the entire
>>> time until flat lining for a few hours and then picks up again once I
>>> restart the services.
>>>
>>> Thanks in advance!
>>> --
>>> You received this message because you are subscribed to the Google
>>> Groups "elasticsearch" group.
>>> To unsubscribe from this group and stop receiving emails from it, send
>>> an email to [email protected].
>>> To view this discussion on the web visit
>>> https://groups.google.com/d/msgid/elasticsearch/d088f293-9dae-475a-a15e-076b3b773680%40googlegroups.com
>>>
>>> <https://groups.google.com/d/msgid/elasticsearch/d088f293-9dae-475a-a15e-076b3b773680%40googlegroups.com?utm_medium=email&utm_source=footer>
>>> .
>>> For more options, visit https://groups.google.com/d/optout.
>>>
>> --
>> You received this message because you are subscribed to the Google Groups
>> "elasticsearch" group.
>> To unsubscribe from this group and stop receiving emails from it, send an
>> email to [email protected].
>> To view this discussion on the web visit
>> https://groups.google.com/d/msgid/elasticsearch/9a357da0-9ac6-4aaf-853d-f76f453c27d0%40googlegroups.com
>>
>> <https://groups.google.com/d/msgid/elasticsearch/9a357da0-9ac6-4aaf-853d-f76f453c27d0%40googlegroups.com?utm_medium=email&utm_source=footer>
>> .
>> For more options, visit https://groups.google.com/d/optout.
>>
>> --
> You received this message because you are subscribed to the Google Groups
> "elasticsearch" group.
> To unsubscribe from this group and stop receiving emails from it, send an
> email to [email protected] <javascript:>.
> To view this discussion on the web visit
> https://groups.google.com/d/msgid/elasticsearch/6908936b-02e7-46f9-873d-9fe3f2d1d152%40googlegroups.com
>
> <https://groups.google.com/d/msgid/elasticsearch/6908936b-02e7-46f9-873d-9fe3f2d1d152%40googlegroups.com?utm_medium=email&utm_source=footer>
> .
> For more options, visit https://groups.google.com/d/optout.
>
>
--
You received this message because you are subscribed to the Google Groups
"elasticsearch" group.
To unsubscribe from this group and stop receiving emails from it, send an email
to [email protected].
To view this discussion on the web visit
https://groups.google.com/d/msgid/elasticsearch/42b28459-4f9a-4a80-9b7d-5d75f63d8fd4%40googlegroups.com.
For more options, visit https://groups.google.com/d/optout.