They surely do. Although it would probably be easiest here to just get them from `oc logs` against the ES pod, especially if we can't trust ES storage.
On Fri, Jul 15, 2016 at 3:26 PM, Peter Portante <pport...@redhat.com> wrote: > Eric, Luke, > > Do the logs from the ES instance itself flow into that ES instance? > > -peter > > On Fri, Jul 15, 2016 at 12:14 PM, Alex Wauck <alexwa...@exosite.com> > wrote: > > I'm not sure that I can. I clicked the "Archive" link for the logging-es > > pod and then changed the query in Kibana to "kubernetes_container_name: > > logging-es-cycd8veb && kubernetes_namespace_name: logging". I got no > > results, instead getting this error: > > > > Index: unrelated-project.92c37428-11f6-11e6-9c83-020b5091df01.2016.07.12 > > Shard: 2 Reason: EsRejectedExecutionException[rejected execution (queue > > capacity 1000) on > > org.elasticsearch.search.action.SearchServiceTransportAction$23@6b1f2699 > ] > > Index: unrelated-project.92c37428-11f6-11e6-9c83-020b5091df01.2016.07.14 > > Shard: 2 Reason: EsRejectedExecutionException[rejected execution (queue > > capacity 1000) on > > org.elasticsearch.search.action.SearchServiceTransportAction$23@66b9a5fb > ] > > Index: unrelated-project.92c37428-11f6-11e6-9c83-020b5091df01.2016.07.15 > > Shard: 2 Reason: EsRejectedExecutionException[rejected execution (queue > > capacity 1000) on > > org.elasticsearch.search.action.SearchServiceTransportAction$23@512820e] > > Index: unrelated-project.f38ac6ff-3e42-11e6-ab71-020b5091df01.2016.06.29 > > Shard: 2 Reason: EsRejectedExecutionException[rejected execution (queue > > capacity 1000) on > > org.elasticsearch.search.action.SearchServiceTransportAction$23@3dce96b9 > ] > > Index: unrelated-project.f38ac6ff-3e42-11e6-ab71-020b5091df01.2016.06.30 > > Shard: 2 Reason: EsRejectedExecutionException[rejected execution (queue > > capacity 1000) on > > org.elasticsearch.search.action.SearchServiceTransportAction$23@2f774477 > ] > > > > When I initially clicked the "Archive" link, I saw a lot of messages with > > the kubernetes_container_name "logging-fluentd", which is not what I > > expected to see. > > > > > > On Fri, Jul 15, 2016 at 10:44 AM, Peter Portante <pport...@redhat.com> > > wrote: > >> > >> Can you go back further in the logs to the point where the errors > started? > >> > >> I am thinking about possible Java HEAP issues, or possibly ES > >> restarting for some reason. > >> > >> -peter > >> > >> On Fri, Jul 15, 2016 at 11:37 AM, Lukáš Vlček <lvl...@redhat.com> > wrote: > >> > Also looking at this. > >> > Alex, is it possible to investigate if you were having some kind of > >> > network connection issues in the ES cluster (I mean between individual > >> > cluster nodes)? > >> > > >> > Regards, > >> > Lukáš > >> > > >> > > >> > > >> > > >> >> On 15 Jul 2016, at 17:08, Peter Portante <pport...@redhat.com> > wrote: > >> >> > >> >> Just catching up on the thread, will get back to you all in a few ... > >> >> > >> >> On Fri, Jul 15, 2016 at 10:08 AM, Eric Wolinetz <ewoli...@redhat.com > > > >> >> wrote: > >> >>> Adding Lukas and Peter > >> >>> > >> >>> On Fri, Jul 15, 2016 at 8:07 AM, Luke Meyer <lme...@redhat.com> > wrote: > >> >>>> > >> >>>> I believe the "queue capacity" there is the number of parallel > >> >>>> searches > >> >>>> that can be queued while the existing search workers operate. It > >> >>>> sounds like > >> >>>> it has plenty of capacity there and it has a different reason for > >> >>>> rejecting > >> >>>> the query. I would guess the data requested is missing given it > >> >>>> couldn't > >> >>>> fetch shards it expected to. > >> >>>> > >> >>>> The number of shards is a multiple (for redundancy) of the number > of > >> >>>> indices, and there is an index created per project per day. So even > >> >>>> for a > >> >>>> small cluster this doesn't sound out of line. > >> >>>> > >> >>>> Can you give a little more information about your logging > deployment? > >> >>>> Have > >> >>>> you deployed multiple ES nodes for redundancy, and what are you > using > >> >>>> for > >> >>>> storage? Could you attach full ES logs? How many OpenShift nodes > and > >> >>>> projects do you have? Any history of events that might have > resulted > >> >>>> in lost > >> >>>> data? > >> >>>> > >> >>>> On Thu, Jul 14, 2016 at 4:06 PM, Alex Wauck <alexwa...@exosite.com > > > >> >>>> wrote: > >> >>>>> > >> >>>>> When doing searches in Kibana, I get error messages similar to > >> >>>>> "Courier > >> >>>>> Fetch: 919 of 2020 shards failed". Deeper inspection reveals > errors > >> >>>>> like > >> >>>>> this: "EsRejectedExecutionException[rejected execution (queue > >> >>>>> capacity 1000) > >> >>>>> on > >> >>>>> > >> >>>>> > org.elasticsearch.search.action.SearchServiceTransportAction$23@14522b8e > ]". > >> >>>>> > >> >>>>> A bit of investigation lead me to conclude that our Elasticsearch > >> >>>>> server > >> >>>>> was not sufficiently powerful, but I spun up a new one with four > >> >>>>> times the > >> >>>>> CPU and RAM of the original one, but the queue capacity is still > >> >>>>> only 1000. > >> >>>>> Also, 2020 seems like a really ridiculous number of shards. Any > >> >>>>> idea what's > >> >>>>> going on here? > >> >>>>> > >> >>>>> -- > >> >>>>> > >> >>>>> Alex Wauck // DevOps Engineer > >> >>>>> > >> >>>>> E X O S I T E > >> >>>>> www.exosite.com > >> >>>>> > >> >>>>> Making Machines More Human. > >> >>>>> > >> >>>>> > >> >>>>> _______________________________________________ > >> >>>>> users mailing list > >> >>>>> users@lists.openshift.redhat.com > >> >>>>> http://lists.openshift.redhat.com/openshiftmm/listinfo/users > >> >>>>> > >> >>>> > >> >>>> > >> >>>> _______________________________________________ > >> >>>> users mailing list > >> >>>> users@lists.openshift.redhat.com > >> >>>> http://lists.openshift.redhat.com/openshiftmm/listinfo/users > >> > > > > > > > > > > > -- > > > > Alex Wauck // DevOps Engineer > > > > E X O S I T E > > www.exosite.com > > > > Making Machines More Human. >
_______________________________________________ users mailing list users@lists.openshift.redhat.com http://lists.openshift.redhat.com/openshiftmm/listinfo/users