I ran rsyslog in debug mode and now I am in the same state where a copy of
the logs are still being written to an omfile but nothing gets reported to
elasticsearch. Rsyslog generated 728MB of debug logs. What would you like
me to do with it? Should I upload it to pastebin or search for something?
The debug logs are of the following format:

1673.866510445:omelasticsearch-event-collector-kafka-rest-http-request.log
que: actionDoRetry:
omelasticsearch-event-collector-kafka-rest-http-request.log enter loop,
iRetries=294
1673.866533581:omelasticsearch-event-collector-kafka-rest-http-request.log
que: omelasticsearch: tryResume called
1673.949301652:omelasticsearch-event-collector-kafka-rest-http-request.log
que: omelasticsearch: checkConn() curl_easy_perform() failed: HTTP response
code said error
1673.949335309:omelasticsearch-event-collector-kafka-rest-http-request.log
que: actionDoRetry:
omelasticsearch-event-collector-kafka-rest-http-request.log
action->tryResume returned -2007
1673.949342989:omelasticsearch-event-collector-kafka-rest-http-request.log
que: actionDoRetry:
omelasticsearch-event-collector-kafka-rest-http-request.log check for max
retries, iResumeRetryCount -1, iRetries 294

Thanks,

Alec

On Fri, May 27, 2016 at 9:48 AM, Alec Swan <[email protected]> wrote:

> I finally conclusively confirmed that all the buffered logs are sitting in
> omelasticsearch-queue.00000001 file. I grabbed the last log message
> reported to ES, then found the next one that should have been reported and
> found it at the top of omelasticsearch-queue.00000001.
> So, it's either that my queues are configured incorrectly (see my original
> post) or not all queue related bugs were fixed in 8.18.
>
> I restarted rsyslog in debug mode. Is there anything else I can do to help
> troubleshoot this?
>
> > ls -lat /var/spool/rsyslog/
> total 17944
> drwxr-xr-x  2 root root     4096 May 27 15:43 .
> -rw-------  1 root root      561 May 27 15:23 omelasticsearch-queue.qi
> -rw-------  1 root root 17170428 May 27 15:23
> omelasticsearch-queue.00000001
>
> > rpm -qa | grep rsyslog
> rsyslog-elasticsearch-8.18.0-1.el6.x86_64
> rsyslog-8.18.0-1.el6.x86_64
> rsyslog-mmnormalize-8.18.0-1.el6.x86_64
>
>
>
> Thanks,
>
> Alec
>
> On Fri, May 27, 2016 at 9:15 AM, Alec Swan <[email protected]> wrote:
>
>> I just found out that the reason why rsyslog stopped writing to ES but
>> continued writing the copy of the log to a separate log file was because
>> puppet ran and overwrote my manual change to write to the separate log file
>> :(
>> So, I am currently in the state where rsyslog is still writing messages
>> to the log file, but not to elasticsearch. Since startmsg.regex applies to
>> both inputs it's definitely not the culprit.
>>
>> Before I restarted rsyslog in debug mode I decided to grab impstats to
>> see omelasticsearch queue state. The last impstats logs were sent to
>> elasticsearch 4 hours ago and are shown below.
>> I see  failed.http=5 failed.httprequests=3 failed.es=0 not log before
>> logs stop being sent to ES. Could this have caused the logs to start
>> buffering?
>>
>> May 27th 2016, 05:27:07.960 message:omelasticsearch-myapp-launcher.log:
>> origin=core.action processed=3979 failed=0 suspended=0 suspended.duration=0
>> resumed=0 @timestamp:May 27th 2016, 05:27:07.960 host:myhost
>> hostip:10.0.0.1 severity:debug facility:syslog syslogtag:rsyslogd-pstats:
>> programname:rsyslogd-pstats logtag:syslog _id:AVTx99mVggzpxkzyDyY9
>> _type:syslog _index:myapp-2016.05.27 _score:
>> May 27th 2016, 05:27:07.960 message:omelasticsearch-myapp-launcher.log
>> queue[DA]: origin=core.queue size=0 enqueued=0 full=0 discarded.full=0
>> discarded.nf=0 maxqsize=0 @timestamp:May 27th 2016, 05:27:07.960
>> host:myhost hostip:10.0.0.1 severity:debug facility:syslog
>> syslogtag:rsyslogd-pstats: programname:rsyslogd-pstats logtag:syslog
>> _id:AVTx99mVggzpxkzyDyZL _type:syslog _index:myapp-2016.05.27 _score:
>> May 27th 2016, 05:27:07.960 message:omelasticsearch-myapp-http-request.log
>> que[DA]: origin=core.queue size=0 enqueued=0 full=0 discarded.full=0
>> discarded.nf=0 maxqsize=0 @timestamp:May 27th 2016, 05:27:07.960
>> host:myhost hostip:10.0.0.1 severity:debug facility:syslog
>> syslogtag:rsyslogd-pstats: programname:rsyslogd-pstats logtag:syslog
>> _id:AVTx99mVggzpxkzyDyZN _type:syslog _index:myapp-2016.05.27 _score:
>> May 27th 2016, 05:27:07.960 message:omelasticsearch-myapp-http-request.log
>> que: origin=core.queue size=995 enqueued=37639 full=0 discarded.full=0
>> discarded.nf=0 maxqsize=20002 @timestamp:May 27th 2016, 05:27:07.960
>> host:myhost hostip:10.0.0.1 severity:debug facility:syslog
>> syslogtag:rsyslogd-pstats: programname:rsyslogd-pstats logtag:syslog
>> _id:AVTx99mVggzpxkzyDyZO _type:syslog _index:myapp-2016.05.27 _score:
>> May 27th 2016, 05:27:07.960 message:omelasticsearch:
>> origin=omelasticsearch submitted=54608 failed.http=5 failed.httprequests=3
>> failed.es=0 @timestamp:May 27th 2016, 05:27:07.960 host:myhost
>> hostip:10.0.0.1 severity:debug facility:syslog syslogtag:rsyslogd-pstats:
>> programname:rsyslogd-pstats logtag:syslog _id:AVTx99hfggzpxkzyDyY3
>> _type:syslog _index:myapp-2016.05.27 _score:
>> May 27th 2016, 05:27:07.960 message:omelasticsearch-myapp-http-request.log:
>> origin=core.action processed=37639 failed=0 suspended=0
>> suspended.duration=0 resumed=0 @timestamp:May 27th 2016, 05:27:07.960
>> host:myhost hostip:10.0.0.1 severity:debug facility:syslog
>> syslogtag:rsyslogd-pstats: programname:rsyslogd-pstats logtag:syslog
>> _id:AVTx99mVggzpxkzyDyZA _type:syslog _index:myapp-2016.05.27 _score:
>> May 27th 2016, 05:27:07.960 message:omelasticsearch-syslog queue[DA]:
>> origin=core.queue size=0 enqueued=0 full=0 discarded.full=0 discarded.nf=0
>> maxqsize=0 @timestamp:May 27th 2016, 05:27:07.960 host:myhost
>> hostip:10.0.0.1 severity:debug facility:syslog syslogtag:rsyslogd-pstats:
>> programname:rsyslogd-pstats logtag:syslog _id:AVTx99mVggzpxkzyDyZJ
>> _type:syslog _index:myapp-2016.05.27 _score:
>> May 27th 2016, 05:27:07.960 message:omelasticsearch-syslog:
>> origin=core.action processed=35488 failed=0 suspended=0
>> suspended.duration=0 resumed=0 @timestamp:May 27th 2016, 05:27:07.960
>> host:myhost hostip:10.0.0.1 severity:debug facility:syslog
>> syslogtag:rsyslogd-pstats: programname:rsyslogd-pstats logtag:syslog
>> _id:AVTx99mVggzpxkzyDyY5 _type:syslog _index:myapp-2016.05.27 _score:
>> May 27th 2016, 05:27:07.960 message:omelasticsearch-syslog queue:
>> origin=core.queue size=20024 enqueued=35491 full=0 discarded.full=0
>> discarded.nf=0 maxqsize=20024 @timestamp:May 27th 2016, 05:27:07.960
>> host:myhost hostip:10.0.0.1 severity:debug facility:syslog
>> syslogtag:rsyslogd-pstats: programname:rsyslogd-pstats logtag:syslog
>> _id:AVTx99mVggzpxkzyDyZK _type:syslog _index:myapp-2016.05.27 _score:
>> May 27th 2016, 05:27:07.960 message:omelasticsearch-myapp-launcher.log
>> queue: origin=core.queue size=1482 enqueued=3979 full=0 discarded.full=0
>> discarded.nf=0 maxqsize=1482 @timestamp:May 27th 2016, 05:27:07.960
>> host:myhost hostip:10.0.0.1 severity:debug facility:syslog
>> syslogtag:rsyslogd-pstats: programname:rsyslogd-pstats logtag:syslog
>> _id:AVTx99mVggzpxkzyDyZM _type:syslog _index:myapp-2016.05.27 _score:
>> May 27th 2016, 05:26:37.931 message:omelasticsearch-syslog queue[DA]:
>> origin=core.queue size=0 enqueued=0 full=0 discarded.full=0 discarded.nf=0
>> maxqsize=0 @timestamp:May 27th 2016, 05:26:37.931 host:myhost
>> hostip:10.0.0.1 severity:debug facility:syslog syslogtag:rsyslogd-pstats:
>> programname:rsyslogd-pstats logtag:syslog _id:AVTx99hfggzpxkzyDyYu
>> _type:syslog _index:myapp-2016.05.27 _score:
>> May 27th 2016, 05:26:37.931 message:omelasticsearch-myapp-launcher.log
>> queue[DA]: origin=core.queue size=0 enqueued=0 full=0 discarded.full=0
>> discarded.nf=0 maxqsize=0 @timestamp:May 27th 2016, 05:26:37.931
>> host:myhost hostip:10.0.0.1 severity:debug facility:syslog
>> syslogtag:rsyslogd-pstats: programname:rsyslogd-pstats logtag:syslog
>> _id:AVTx99hfggzpxkzyDyYw _type:syslog _index:myapp-2016.05.27 _score:
>>
>>
>>
>> Thanks,
>>
>> Alec
>>
>> On Thu, May 26, 2016 at 11:41 PM, Rainer Gerhards <
>> [email protected]> wrote:
>>
>>> 2016-05-27 3:23 GMT+02:00 Peter Portante <[email protected]>:
>>> > On Thu, May 26, 2016 at 2:11 PM, Alec Swan <[email protected]> wrote:
>>> >
>>> >> David,
>>> >>
>>> >> After enabling the omfile action to write to a log file before
>>> sending to
>>> >> elasticsearch (see below) I confirmed that logs stop being written to
>>> >> /var/log/rsyslog/output.log and to ES at the same time even though
>>> the log
>>> >> file being monitored /var/log/file-to-monitor.log keeps getting new
>>> logs.
>>> >> So, the problem seems to be related to imfile input and not to output
>>> >> plugins or queue settings.
>>> >>
>>> >> Maybe it's related to how I am monitoring the log file?
>>> >>
>>> >>
>>> >> *module(load = "imfile")*
>>> >>
>>> >> *input(*
>>> >> *     type = "imfile"*
>>> >> *     File = "/var/log/file-to-monitor.log"*
>>> >> *     Tag = "myapp"*
>>> >>
>>> >> *     startmsg.regex =
>>> "^[[:digit:]]{4}-[[:digit:]]{2}-[[:digit:]]{2}"*
>>> >>
>>> >
>>> > Are you sure the startmsg.regex pattern always matches?
>>>
>>> This might indded be the problem. To gain more insight, I suggest to
>>> create a debug log. That will most probably get us good information to
>>> know what is going on,.
>>>
>>> Essentially, you can do this interactively by stopping the service and
>>> running it via
>>>
>>> $ rsyslogd -dn ..other opts... &> rsyslog.debug
>>>
>>> You can also use the service version. Full details are here:
>>>
>>> http://www.rsyslog.com/doc/v8-stable/troubleshooting/debug.html
>>>
>>> We would then need to see the *full* debug log (no excerpts, please).
>>> You can post it e.g. to pastebin.
>>>
>>> Rainer
>>>
>>> >
>>> >
>>> >> *     Facility = "local2"*
>>> >> *     Ruleset = "myruleset"*
>>> >> * )*
>>> >>
>>> >>
>>> >> *ruleset(name = "myruleset") {*
>>> >> *    action(type = "mmnormalize" rulebase =
>>> >> "/etc/rsyslog.d/rules/myrule.rb")*
>>> >> *    action(type = "omfile" template="es-payload"
>>> >> file="/var/log/rsyslog/output.log" FileCreateMode="0644")*
>>> >> *    action(type = "omelasticsearch" queue.xxx)*
>>> >> *}*
>>> >>
>>> >> Thanks,
>>> >>
>>> >> Alec
>>> >>
>>> >> On Wed, May 25, 2016 at 11:13 AM, Alec Swan <[email protected]>
>>> wrote:
>>> >>
>>> >> > David,
>>> >> >
>>> >> > The rate of delivery is about 1 log per second. If you are
>>> referring to
>>> >> queue.dequeuebatchsize="1000"
>>> >> > batch size, then I would expect the logs to be batched for 15-20
>>> minutes.
>>> >> > However, I am observing delays of multiple hours.
>>> >> >
>>> >> > When I restart rsyslog all buffered logs get sent to elasticsearch.
>>> I had
>>> >> > logging to log file enabled before and could see all logs being
>>> written
>>> >> to
>>> >> > log files correctly. I enabled it again and will keep an eye on it,
>>> but I
>>> >> > am sure the problem is pushing to ES.
>>> >> >
>>> >> > I currently have a host which hasn't sent logs for about 12 hours.
>>> The
>>> >> > following are the last logs I received from that node. Anything I
>>> can do
>>> >> to
>>> >> > troubleshoot while the host is in a bad state?
>>> >> >
>>> >> > May 25th 2016, 05:28:05.880 @timestamp:May 25th 2016, 05:28:05.880
>>> >> > host:myhost hostip:10.0.0.1 severity:debug facility:syslog
>>> >> > syslogtag:rsyslogd-pstats: programname:rsyslogd-pstats logtag:syslog
>>> >> > message:omelasticsearch-myapp-launcher.log: origin=core.action
>>> >> > processed=333 failed=0 suspended=0 suspended.duration=0 resumed=0
>>> >> > _id:AVTnrAgpFO4BDB55DTh2 _type:syslog
>>> >> > _index:logstash-syslog-myapp-2016.05.25 _score:
>>> >> > May 25th 2016, 05:28:05.880 @timestamp:May 25th 2016, 05:28:05.880
>>> >> > host:myhost hostip:10.0.0.1 severity:debug facility:syslog
>>> >> > syslogtag:rsyslogd-pstats: programname:rsyslogd-pstats logtag:syslog
>>> >> > message:action 4: origin=core.action processed=336 failed=0
>>> suspended=0
>>> >> > suspended.duration=0 resumed=0 _id:AVTnrAgpFO4BDB55DTh3 _type:syslog
>>> >> > _index:logstash-syslog-myapp-2016.05.25 _score:
>>> >> > May 25th 2016, 05:28:05.880 @timestamp:May 25th 2016, 05:28:05.880
>>> >> > host:myhost hostip:10.0.0.1 severity:debug facility:syslog
>>> >> > syslogtag:rsyslogd-pstats: programname:rsyslogd-pstats logtag:syslog
>>> >> > message:omelasticsearch-syslog queue[DA]: origin=core.queue size=0
>>> >> > enqueued=0 full=0 discarded.full=0 discarded.nf=0 maxqsize=0
>>> >> > _id:AVTnrAlsFO4BDB55DTiD _type:syslog
>>> >> > _index:logstash-syslog-myapp-2016.05.25 _score:
>>> >> > May 25th 2016, 05:28:05.880 @timestamp:May 25th 2016, 05:28:05.880
>>> >> > host:myhost hostip:10.0.0.1 severity:debug facility:syslog
>>> >> > syslogtag:rsyslogd-pstats: programname:rsyslogd-pstats logtag:syslog
>>> >> > message:omelasticsearch-myapp-launcher.log queue[DA]:
>>> origin=core.queue
>>> >> > size=0 enqueued=0 full=0 discarded.full=0 discarded.nf=0 maxqsize=0
>>> >> > _id:AVTnrAlsFO4BDB55DTiF _type:syslog
>>> >> > _index:logstash-syslog-myapp-2016.05.25 _score:
>>> >> >
>>> >> > Thanks,
>>> >> >
>>> >> > Alec
>>> >> >
>>> >> > On Tue, May 24, 2016 at 10:35 PM, David Lang <[email protected]> wrote:
>>> >> >
>>> >> >> It shouldn't matter, but what is the rate of log delivery? is
>>> there any
>>> >> >> chance that it is waiting to deliver a full batch?
>>> >> >>
>>> >> >> I would consider putting this in a ruleset and moving the queue to
>>> the
>>> >> >> ruleset. I would then have the ruleset contain two items
>>> >> >>
>>> >> >> 1. the output to ES
>>> >> >> 2. a write to a debug log locally (not necessarily the full
>>> messages,
>>> >> >> timestamp would be enough)
>>> >> >>
>>> >> >> you can then see if the local file in growing while things are not
>>> yet
>>> >> >> showing up in ES to see if the issue is on the sending side or on
>>> the
>>> >> >> receiving side.
>>> >> >>
>>> >> >> David Lang
>>> >> >>
>>> >> >> On Tue, 24 May 2016, Alec Swan wrote:
>>> >> >>
>>> >> >> Date: Tue, 24 May 2016 22:17:22 -0600
>>> >> >>> From: Alec Swan <[email protected]>
>>> >> >>> Reply-To: rsyslog-users <[email protected]>
>>> >> >>> To: rsyslog-users <[email protected]>
>>> >> >>> Subject: [rsyslog] Logs are delayed being pushed to Elasticsearch
>>> >> >>>
>>> >> >>>
>>> >> >>> Hello,
>>> >> >>>
>>> >> >>> I recently upgraded to rsyslog 8.18 and was happy to see that on
>>> disk
>>> >> >>> queues no longer cause rsyslog to get in a bad state. However,
>>> now I
>>> >> see
>>> >> >>> very long delays (several hours) of logs being pushed to
>>> Elasticsearch.
>>> >> >>> It
>>> >> >>> seems that somehow the logs are being buffered on the client for
>>> >> several
>>> >> >>> hours because eventually they do show up in Elasticsearch. I
>>> don't see
>>> >> >>> any
>>> >> >>> errors in /var/log/rsyslog/ES-error.log (see config below) or
>>> >> >>> /var/log/messages.
>>> >> >>>
>>> >> >>> I enabled impstats but didn't see any errors related to
>>> >> omelasticsearch.
>>> >> >>> What else can I do to troubleshoot this?
>>> >> >>>
>>> >> >>> Here is my omelasticsearch config:
>>> >> >>>
>>> >> >>> action(
>>> >> >>>         type = "omelasticsearch"
>>> >> >>>         template = "es-payload"
>>> >> >>>         dynSearchIndex = "on"
>>> >> >>>         searchIndex = "logstash-index"
>>> >> >>>         searchType = "syslog"
>>> >> >>>         server = "127.0.0.1"
>>> >> >>>         serverport = "9200"
>>> >> >>>         uid = "xxx"
>>> >> >>>         pwd = "yyy"
>>> >> >>>         errorFile = "/var/log/rsyslog/ES-error.log"
>>> >> >>>         bulkmode = "on"
>>> >> >>>         action.resumeretrycount="-1"  # retry if ES is
>>> unreachable (-1
>>> >> >>> for
>>> >> >>> infinite retries)
>>> >> >>>         action.resumeInterval="60"
>>> >> >>>         queue.dequeuebatchsize="1000"   # ES bulk size
>>> >> >>>         queue.type="linkedlist"
>>> >> >>>         queue.size="100000"
>>> >> >>>         queue.workerthreads="5"
>>> >> >>>         queue.timeoutworkerthreadshutdown="2000"
>>> >> >>>         queue.spoolDirectory="/var/spool/rsyslog"
>>> >> >>>         queue.filename="omelasticsearch-queue"
>>> >> >>>         queue.maxfilesize="100m"
>>> >> >>>         queue.maxdiskspace="1g"
>>> >> >>>         queue.highwatermark="80000" # when to start spilling to
>>> disk
>>> >> >>>         queue.lowwatermark="20000"  # when to stop spilling to
>>> disk
>>> >> >>>         queue.saveonshutdown="on"
>>> >> >>>    )
>>> >> >>>
>>> >> >>> Thanks,
>>> >> >>>
>>> >> >>> Alec
>>> >> >>> _______________________________________________
>>> >> >>> rsyslog mailing list
>>> >> >>> http://lists.adiscon.net/mailman/listinfo/rsyslog
>>> >> >>> http://www.rsyslog.com/professional-services/
>>> >> >>> What's up with rsyslog? Follow https://twitter.com/rgerhards
>>> >> >>> NOTE WELL: This is a PUBLIC mailing list, posts are ARCHIVED by a
>>> >> myriad
>>> >> >>> of sites beyond our control. PLEASE UNSUBSCRIBE and DO NOT POST
>>> if you
>>> >> >>> DON'T LIKE THAT.
>>> >> >>>
>>> >> >>> _______________________________________________
>>> >> >> rsyslog mailing list
>>> >> >> http://lists.adiscon.net/mailman/listinfo/rsyslog
>>> >> >> http://www.rsyslog.com/professional-services/
>>> >> >> What's up with rsyslog? Follow https://twitter.com/rgerhards
>>> >> >> NOTE WELL: This is a PUBLIC mailing list, posts are ARCHIVED by a
>>> myriad
>>> >> >> of sites beyond our control. PLEASE UNSUBSCRIBE and DO NOT POST if
>>> you
>>> >> >> DON'T LIKE THAT.
>>> >> >>
>>> >> >
>>> >> >
>>> >> _______________________________________________
>>> >> rsyslog mailing list
>>> >> http://lists.adiscon.net/mailman/listinfo/rsyslog
>>> >> http://www.rsyslog.com/professional-services/
>>> >> What's up with rsyslog? Follow https://twitter.com/rgerhards
>>> >> NOTE WELL: This is a PUBLIC mailing list, posts are ARCHIVED by a
>>> myriad
>>> >> of sites beyond our control. PLEASE UNSUBSCRIBE and DO NOT POST if you
>>> >> DON'T LIKE THAT.
>>> >>
>>> > _______________________________________________
>>> > rsyslog mailing list
>>> > http://lists.adiscon.net/mailman/listinfo/rsyslog
>>> > http://www.rsyslog.com/professional-services/
>>> > What's up with rsyslog? Follow https://twitter.com/rgerhards
>>> > NOTE WELL: This is a PUBLIC mailing list, posts are ARCHIVED by a
>>> myriad of sites beyond our control. PLEASE UNSUBSCRIBE and DO NOT POST if
>>> you DON'T LIKE THAT.
>>> _______________________________________________
>>> rsyslog mailing list
>>> http://lists.adiscon.net/mailman/listinfo/rsyslog
>>> http://www.rsyslog.com/professional-services/
>>> What's up with rsyslog? Follow https://twitter.com/rgerhards
>>> NOTE WELL: This is a PUBLIC mailing list, posts are ARCHIVED by a myriad
>>> of sites beyond our control. PLEASE UNSUBSCRIBE and DO NOT POST if you
>>> DON'T LIKE THAT.
>>>
>>
>>
>
_______________________________________________
rsyslog mailing list
http://lists.adiscon.net/mailman/listinfo/rsyslog
http://www.rsyslog.com/professional-services/
What's up with rsyslog? Follow https://twitter.com/rgerhards
NOTE WELL: This is a PUBLIC mailing list, posts are ARCHIVED by a myriad of 
sites beyond our control. PLEASE UNSUBSCRIBE and DO NOT POST if you DON'T LIKE 
THAT.

Reply via email to