> It's easy to end up with several copies of the data in memory as it's being processed, so 17K is not impossible. I do parse and send messages in JSON format to ES. I counted the number of messages in omelasticsearch-queue.00000010 file and got about 11,000. Given that the size of the file is 100M each message would be around 9K. So, you are right that if there are multiple copies of messages in memory it could reach 17K on average per message. But why would there be multiple copies of messages?
> no, all the workers work on the same queue. Unless you find that a thread (either on the rsyslog or elasticsearch sice) is maxed out, you should probably reduce the number of threads to allow better batching. How would I find that a thread is maxed out? Also, it would seem that multiple threads would be able to process data faster because some of them will be blocking waiting on ES response while others reading messages from the queue. Why would smaller number help with batching? I just restarted rsyslog while there are 1GB worth of queue files in /var/spool/rsyslog. Rsyslog started back up with 925MB RAM usage. What parameter controls how much of queue data rsyslog will load up from disk on startup? Thanks, Alec On Thu, Jun 2, 2016 at 3:45 PM, David Lang <[email protected]> wrote: > On Thu, 2 Jun 2016, Alec Swan wrote: > > I just tried out 8.19.0 release and it seemed like the problem has gone >> away. I haven't run it long enough to say this conclusively, but so far so >> good. >> >> I noticed, however, that rsyslog is taking up 1.7GB of RAM when it's >> unable >> to send logs to Elasticsearch. My queue size (see original email) is set >> to >> 100,000 which would mean that each item in the queue would be 17K which >> sound awfully big. >> > > hard to say without knowing what your logs look like and if you parse them > (which is a good idea, ElasticSearch is much more useful if you pass it > json data than if you pass it plain text). It's easy to end up with several > copies of the data in memory as it's being processed, so 17K is not > impossible. > > I have queue.workerthreads set to 5. Does each worker thread have its own >> 100,000 element queue? >> > > no, all the workers work on the same queue. Unless you find that a thread > (either on the rsyslog or elasticsearch sice) is maxed out, you should > probably reduce the number of threads to allow better batching. > > But then again according to this doc >> <http://www.rsyslog.com/doc/v8-stable/concepts/queues.html>: >> "Consequently, >> the number of worker threads is fixed to one for action queues and can not >> be changed." >> >> Any other thoughts where 1.7 GB of RAM could be spent? >> > > It's almost certinly the queue. keep in mind that when the action queue > fills up, the main queue starts to fill as well. take a look at the output > for impstats to see how many messages are queued. > > David Lang > > > Thanks, >> >> Alec >> >> On Sat, May 28, 2016 at 2:40 AM, Rainer Gerhards < >> [email protected]> >> wrote: >> >> 2016-05-28 4:36 GMT+02:00 Alec Swan <[email protected]>: >>> >>>> I ran rsyslog in debug mode and now I am in the same state where a copy >>>> >>> of >>> >>>> the logs are still being written to an omfile but nothing gets reported >>>> >>> to >>> >>>> elasticsearch. Rsyslog generated 728MB of debug logs. What would you >>>> like >>>> me to do with it? Should I upload it to pastebin or search for >>>> something? >>>> The debug logs are of the following format: >>>> >>>> >>>> >>> 1673.866510445:omelasticsearch-event-collector-kafka-rest-http-request.log >>> >>>> que: actionDoRetry: >>>> omelasticsearch-event-collector-kafka-rest-http-request.log enter loop, >>>> iRetries=294 >>>> >>>> >>> 1673.866533581:omelasticsearch-event-collector-kafka-rest-http-request.log >>> >>>> que: omelasticsearch: tryResume called >>>> >>>> >>> 1673.949301652:omelasticsearch-event-collector-kafka-rest-http-request.log >>> >>>> que: omelasticsearch: checkConn() curl_easy_perform() failed: HTTP >>>> >>> response >>> >>>> code said error >>>> >>> >>> That's probably sufficient :-) >>> >>> I found a regression that we had since a couple of versions where >>> data-induced errors causes suspensions instead of writing the errors >>> to the error file. This is patched now. I suggest to try git master >>> branch. It is pretty safe to do so now, it will turn int 8.19.0 next >>> tuesday. >>> >>> Rainer >>> >>> >>>> >>> 1673.949335309:omelasticsearch-event-collector-kafka-rest-http-request.log >>> >>>> que: actionDoRetry: >>>> omelasticsearch-event-collector-kafka-rest-http-request.log >>>> action->tryResume returned -2007 >>>> >>>> >>> 1673.949342989:omelasticsearch-event-collector-kafka-rest-http-request.log >>> >>>> que: actionDoRetry: >>>> omelasticsearch-event-collector-kafka-rest-http-request.log check for >>>> max >>>> retries, iResumeRetryCount -1, iRetries 294 >>>> >>>> Thanks, >>>> >>>> Alec >>>> >>>> On Fri, May 27, 2016 at 9:48 AM, Alec Swan <[email protected]> wrote: >>>> >>>> I finally conclusively confirmed that all the buffered logs are sitting >>>>> >>>> in >>> >>>> omelasticsearch-queue.00000001 file. I grabbed the last log message >>>>> reported to ES, then found the next one that should have been reported >>>>> >>>> and >>> >>>> found it at the top of omelasticsearch-queue.00000001. >>>>> So, it's either that my queues are configured incorrectly (see my >>>>> >>>> original >>> >>>> post) or not all queue related bugs were fixed in 8.18. >>>>> >>>>> I restarted rsyslog in debug mode. Is there anything else I can do to >>>>> >>>> help >>> >>>> troubleshoot this? >>>>> >>>>> ls -lat /var/spool/rsyslog/ >>>>>> >>>>> total 17944 >>>>> drwxr-xr-x 2 root root 4096 May 27 15:43 . >>>>> -rw------- 1 root root 561 May 27 15:23 omelasticsearch-queue.qi >>>>> -rw------- 1 root root 17170428 May 27 15:23 >>>>> omelasticsearch-queue.00000001 >>>>> >>>>> rpm -qa | grep rsyslog >>>>>> >>>>> rsyslog-elasticsearch-8.18.0-1.el6.x86_64 >>>>> rsyslog-8.18.0-1.el6.x86_64 >>>>> rsyslog-mmnormalize-8.18.0-1.el6.x86_64 >>>>> >>>>> >>>>> >>>>> Thanks, >>>>> >>>>> Alec >>>>> >>>>> On Fri, May 27, 2016 at 9:15 AM, Alec Swan <[email protected]> wrote: >>>>> >>>>> I just found out that the reason why rsyslog stopped writing to ES but >>>>>> continued writing the copy of the log to a separate log file was >>>>>> >>>>> because >>> >>>> puppet ran and overwrote my manual change to write to the separate log >>>>>> >>>>> file >>> >>>> :( >>>>>> So, I am currently in the state where rsyslog is still writing >>>>>> messages >>>>>> to the log file, but not to elasticsearch. Since startmsg.regex >>>>>> >>>>> applies to >>> >>>> both inputs it's definitely not the culprit. >>>>>> >>>>>> Before I restarted rsyslog in debug mode I decided to grab impstats to >>>>>> see omelasticsearch queue state. The last impstats logs were sent to >>>>>> elasticsearch 4 hours ago and are shown below. >>>>>> I see failed.http=5 failed.httprequests=3 failed.es=0 not log before >>>>>> logs stop being sent to ES. Could this have caused the logs to start >>>>>> buffering? >>>>>> >>>>>> May 27th 2016, 05:27:07.960 >>>>>> message:omelasticsearch-myapp-launcher.log: >>>>>> origin=core.action processed=3979 failed=0 suspended=0 >>>>>> >>>>> suspended.duration=0 >>> >>>> resumed=0 @timestamp:May 27th 2016, 05:27:07.960 host:myhost >>>>>> hostip:10.0.0.1 severity:debug facility:syslog >>>>>> >>>>> syslogtag:rsyslogd-pstats: >>> >>>> programname:rsyslogd-pstats logtag:syslog _id:AVTx99mVggzpxkzyDyY9 >>>>>> _type:syslog _index:myapp-2016.05.27 _score: >>>>>> May 27th 2016, 05:27:07.960 message:omelasticsearch-myapp-launcher.log >>>>>> queue[DA]: origin=core.queue size=0 enqueued=0 full=0 discarded.full=0 >>>>>> discarded.nf=0 maxqsize=0 @timestamp:May 27th 2016, 05:27:07.960 >>>>>> host:myhost hostip:10.0.0.1 severity:debug facility:syslog >>>>>> syslogtag:rsyslogd-pstats: programname:rsyslogd-pstats logtag:syslog >>>>>> _id:AVTx99mVggzpxkzyDyZL _type:syslog _index:myapp-2016.05.27 _score: >>>>>> May 27th 2016, 05:27:07.960 >>>>>> >>>>> message:omelasticsearch-myapp-http-request.log >>> >>>> que[DA]: origin=core.queue size=0 enqueued=0 full=0 discarded.full=0 >>>>>> discarded.nf=0 maxqsize=0 @timestamp:May 27th 2016, 05:27:07.960 >>>>>> host:myhost hostip:10.0.0.1 severity:debug facility:syslog >>>>>> syslogtag:rsyslogd-pstats: programname:rsyslogd-pstats logtag:syslog >>>>>> _id:AVTx99mVggzpxkzyDyZN _type:syslog _index:myapp-2016.05.27 _score: >>>>>> May 27th 2016, 05:27:07.960 >>>>>> >>>>> message:omelasticsearch-myapp-http-request.log >>> >>>> que: origin=core.queue size=995 enqueued=37639 full=0 discarded.full=0 >>>>>> discarded.nf=0 maxqsize=20002 @timestamp:May 27th 2016, 05:27:07.960 >>>>>> host:myhost hostip:10.0.0.1 severity:debug facility:syslog >>>>>> syslogtag:rsyslogd-pstats: programname:rsyslogd-pstats logtag:syslog >>>>>> _id:AVTx99mVggzpxkzyDyZO _type:syslog _index:myapp-2016.05.27 _score: >>>>>> May 27th 2016, 05:27:07.960 message:omelasticsearch: >>>>>> origin=omelasticsearch submitted=54608 failed.http=5 >>>>>> >>>>> failed.httprequests=3 >>> >>>> failed.es=0 @timestamp:May 27th 2016, 05:27:07.960 host:myhost >>>>>> hostip:10.0.0.1 severity:debug facility:syslog >>>>>> >>>>> syslogtag:rsyslogd-pstats: >>> >>>> programname:rsyslogd-pstats logtag:syslog _id:AVTx99hfggzpxkzyDyY3 >>>>>> _type:syslog _index:myapp-2016.05.27 _score: >>>>>> May 27th 2016, 05:27:07.960 >>>>>> >>>>> message:omelasticsearch-myapp-http-request.log: >>> >>>> origin=core.action processed=37639 failed=0 suspended=0 >>>>>> suspended.duration=0 resumed=0 @timestamp:May 27th 2016, 05:27:07.960 >>>>>> host:myhost hostip:10.0.0.1 severity:debug facility:syslog >>>>>> syslogtag:rsyslogd-pstats: programname:rsyslogd-pstats logtag:syslog >>>>>> _id:AVTx99mVggzpxkzyDyZA _type:syslog _index:myapp-2016.05.27 _score: >>>>>> May 27th 2016, 05:27:07.960 message:omelasticsearch-syslog queue[DA]: >>>>>> origin=core.queue size=0 enqueued=0 full=0 discarded.full=0 >>>>>> >>>>> discarded.nf=0 >>> >>>> maxqsize=0 @timestamp:May 27th 2016, 05:27:07.960 host:myhost >>>>>> hostip:10.0.0.1 severity:debug facility:syslog >>>>>> >>>>> syslogtag:rsyslogd-pstats: >>> >>>> programname:rsyslogd-pstats logtag:syslog _id:AVTx99mVggzpxkzyDyZJ >>>>>> _type:syslog _index:myapp-2016.05.27 _score: >>>>>> May 27th 2016, 05:27:07.960 message:omelasticsearch-syslog: >>>>>> origin=core.action processed=35488 failed=0 suspended=0 >>>>>> suspended.duration=0 resumed=0 @timestamp:May 27th 2016, 05:27:07.960 >>>>>> host:myhost hostip:10.0.0.1 severity:debug facility:syslog >>>>>> syslogtag:rsyslogd-pstats: programname:rsyslogd-pstats logtag:syslog >>>>>> _id:AVTx99mVggzpxkzyDyY5 _type:syslog _index:myapp-2016.05.27 _score: >>>>>> May 27th 2016, 05:27:07.960 message:omelasticsearch-syslog queue: >>>>>> origin=core.queue size=20024 enqueued=35491 full=0 discarded.full=0 >>>>>> discarded.nf=0 maxqsize=20024 @timestamp:May 27th 2016, 05:27:07.960 >>>>>> host:myhost hostip:10.0.0.1 severity:debug facility:syslog >>>>>> syslogtag:rsyslogd-pstats: programname:rsyslogd-pstats logtag:syslog >>>>>> _id:AVTx99mVggzpxkzyDyZK _type:syslog _index:myapp-2016.05.27 _score: >>>>>> May 27th 2016, 05:27:07.960 message:omelasticsearch-myapp-launcher.log >>>>>> queue: origin=core.queue size=1482 enqueued=3979 full=0 >>>>>> >>>>> discarded.full=0 >>> >>>> discarded.nf=0 maxqsize=1482 @timestamp:May 27th 2016, 05:27:07.960 >>>>>> host:myhost hostip:10.0.0.1 severity:debug facility:syslog >>>>>> syslogtag:rsyslogd-pstats: programname:rsyslogd-pstats logtag:syslog >>>>>> _id:AVTx99mVggzpxkzyDyZM _type:syslog _index:myapp-2016.05.27 _score: >>>>>> May 27th 2016, 05:26:37.931 message:omelasticsearch-syslog queue[DA]: >>>>>> origin=core.queue size=0 enqueued=0 full=0 discarded.full=0 >>>>>> >>>>> discarded.nf=0 >>> >>>> maxqsize=0 @timestamp:May 27th 2016, 05:26:37.931 host:myhost >>>>>> hostip:10.0.0.1 severity:debug facility:syslog >>>>>> >>>>> syslogtag:rsyslogd-pstats: >>> >>>> programname:rsyslogd-pstats logtag:syslog _id:AVTx99hfggzpxkzyDyYu >>>>>> _type:syslog _index:myapp-2016.05.27 _score: >>>>>> May 27th 2016, 05:26:37.931 message:omelasticsearch-myapp-launcher.log >>>>>> queue[DA]: origin=core.queue size=0 enqueued=0 full=0 discarded.full=0 >>>>>> discarded.nf=0 maxqsize=0 @timestamp:May 27th 2016, 05:26:37.931 >>>>>> host:myhost hostip:10.0.0.1 severity:debug facility:syslog >>>>>> syslogtag:rsyslogd-pstats: programname:rsyslogd-pstats logtag:syslog >>>>>> _id:AVTx99hfggzpxkzyDyYw _type:syslog _index:myapp-2016.05.27 _score: >>>>>> >>>>>> >>>>>> >>>>>> Thanks, >>>>>> >>>>>> Alec >>>>>> >>>>>> On Thu, May 26, 2016 at 11:41 PM, Rainer Gerhards < >>>>>> [email protected]> wrote: >>>>>> >>>>>> 2016-05-27 3:23 GMT+02:00 Peter Portante <[email protected] >>>>>>> >>>>>> : >>>> >>>>> On Thu, May 26, 2016 at 2:11 PM, Alec Swan <[email protected]> >>>>>>>> >>>>>>> wrote: >>> >>>> >>>>>>>> David, >>>>>>>>> >>>>>>>>> After enabling the omfile action to write to a log file before >>>>>>>>> >>>>>>>> sending to >>>>>>> >>>>>>>> elasticsearch (see below) I confirmed that logs stop being written >>>>>>>>> >>>>>>>> to >>> >>>> /var/log/rsyslog/output.log and to ES at the same time even though >>>>>>>>> >>>>>>>> the log >>>>>>> >>>>>>>> file being monitored /var/log/file-to-monitor.log keeps getting new >>>>>>>>> >>>>>>>> logs. >>>>>>> >>>>>>>> So, the problem seems to be related to imfile input and not to >>>>>>>>> >>>>>>>> output >>> >>>> plugins or queue settings. >>>>>>>>> >>>>>>>>> Maybe it's related to how I am monitoring the log file? >>>>>>>>> >>>>>>>>> >>>>>>>>> *module(load = "imfile")* >>>>>>>>> >>>>>>>>> *input(* >>>>>>>>> * type = "imfile"* >>>>>>>>> * File = "/var/log/file-to-monitor.log"* >>>>>>>>> * Tag = "myapp"* >>>>>>>>> >>>>>>>>> * startmsg.regex = >>>>>>>>> >>>>>>>> "^[[:digit:]]{4}-[[:digit:]]{2}-[[:digit:]]{2}"* >>>>>>> >>>>>>>> >>>>>>>>> >>>>>>>> Are you sure the startmsg.regex pattern always matches? >>>>>>>> >>>>>>> >>>>>>> This might indded be the problem. To gain more insight, I suggest to >>>>>>> create a debug log. That will most probably get us good information >>>>>>> to >>>>>>> know what is going on,. >>>>>>> >>>>>>> Essentially, you can do this interactively by stopping the service >>>>>>> and >>>>>>> running it via >>>>>>> >>>>>>> $ rsyslogd -dn ..other opts... &> rsyslog.debug >>>>>>> >>>>>>> You can also use the service version. Full details are here: >>>>>>> >>>>>>> http://www.rsyslog.com/doc/v8-stable/troubleshooting/debug.html >>>>>>> >>>>>>> We would then need to see the *full* debug log (no excerpts, please). >>>>>>> You can post it e.g. to pastebin. >>>>>>> >>>>>>> Rainer >>>>>>> >>>>>>> >>>>>>>> >>>>>>>> * Facility = "local2"* >>>>>>>>> * Ruleset = "myruleset"* >>>>>>>>> * )* >>>>>>>>> >>>>>>>>> >>>>>>>>> *ruleset(name = "myruleset") {* >>>>>>>>> * action(type = "mmnormalize" rulebase = >>>>>>>>> "/etc/rsyslog.d/rules/myrule.rb")* >>>>>>>>> * action(type = "omfile" template="es-payload" >>>>>>>>> file="/var/log/rsyslog/output.log" FileCreateMode="0644")* >>>>>>>>> * action(type = "omelasticsearch" queue.xxx)* >>>>>>>>> *}* >>>>>>>>> >>>>>>>>> Thanks, >>>>>>>>> >>>>>>>>> Alec >>>>>>>>> >>>>>>>>> On Wed, May 25, 2016 at 11:13 AM, Alec Swan <[email protected]> >>>>>>>>> >>>>>>>> wrote: >>>>>>> >>>>>>>> >>>>>>>>> David, >>>>>>>>>> >>>>>>>>>> The rate of delivery is about 1 log per second. If you are >>>>>>>>>> >>>>>>>>> referring to >>>>>>> >>>>>>>> queue.dequeuebatchsize="1000" >>>>>>>>> >>>>>>>>>> batch size, then I would expect the logs to be batched for 15-20 >>>>>>>>>> >>>>>>>>> minutes. >>>>>>> >>>>>>>> However, I am observing delays of multiple hours. >>>>>>>>>> >>>>>>>>>> When I restart rsyslog all buffered logs get sent to >>>>>>>>>> >>>>>>>>> elasticsearch. >>> >>>> I had >>>>>>> >>>>>>>> logging to log file enabled before and could see all logs being >>>>>>>>>> >>>>>>>>> written >>>>>>> >>>>>>>> to >>>>>>>>> >>>>>>>>>> log files correctly. I enabled it again and will keep an eye on >>>>>>>>>> >>>>>>>>> it, >>> >>>> but I >>>>>>> >>>>>>>> am sure the problem is pushing to ES. >>>>>>>>>> >>>>>>>>>> I currently have a host which hasn't sent logs for about 12 >>>>>>>>>> >>>>>>>>> hours. >>> >>>> The >>>>>>> >>>>>>>> following are the last logs I received from that node. Anything I >>>>>>>>>> >>>>>>>>> can do >>>>>>> >>>>>>>> to >>>>>>>>> >>>>>>>>>> troubleshoot while the host is in a bad state? >>>>>>>>>> >>>>>>>>>> May 25th 2016, 05:28:05.880 @timestamp:May 25th 2016, >>>>>>>>>> >>>>>>>>> 05:28:05.880 >>> >>>> host:myhost hostip:10.0.0.1 severity:debug facility:syslog >>>>>>>>>> syslogtag:rsyslogd-pstats: programname:rsyslogd-pstats >>>>>>>>>> >>>>>>>>> logtag:syslog >>> >>>> message:omelasticsearch-myapp-launcher.log: origin=core.action >>>>>>>>>> processed=333 failed=0 suspended=0 suspended.duration=0 resumed=0 >>>>>>>>>> _id:AVTnrAgpFO4BDB55DTh2 _type:syslog >>>>>>>>>> _index:logstash-syslog-myapp-2016.05.25 _score: >>>>>>>>>> May 25th 2016, 05:28:05.880 @timestamp:May 25th 2016, >>>>>>>>>> >>>>>>>>> 05:28:05.880 >>> >>>> host:myhost hostip:10.0.0.1 severity:debug facility:syslog >>>>>>>>>> syslogtag:rsyslogd-pstats: programname:rsyslogd-pstats >>>>>>>>>> >>>>>>>>> logtag:syslog >>> >>>> message:action 4: origin=core.action processed=336 failed=0 >>>>>>>>>> >>>>>>>>> suspended=0 >>>>>>> >>>>>>>> suspended.duration=0 resumed=0 _id:AVTnrAgpFO4BDB55DTh3 >>>>>>>>>> >>>>>>>>> _type:syslog >>> >>>> _index:logstash-syslog-myapp-2016.05.25 _score: >>>>>>>>>> May 25th 2016, 05:28:05.880 @timestamp:May 25th 2016, >>>>>>>>>> >>>>>>>>> 05:28:05.880 >>> >>>> host:myhost hostip:10.0.0.1 severity:debug facility:syslog >>>>>>>>>> syslogtag:rsyslogd-pstats: programname:rsyslogd-pstats >>>>>>>>>> >>>>>>>>> logtag:syslog >>> >>>> message:omelasticsearch-syslog queue[DA]: origin=core.queue >>>>>>>>>> >>>>>>>>> size=0 >>> >>>> enqueued=0 full=0 discarded.full=0 discarded.nf=0 maxqsize=0 >>>>>>>>>> _id:AVTnrAlsFO4BDB55DTiD _type:syslog >>>>>>>>>> _index:logstash-syslog-myapp-2016.05.25 _score: >>>>>>>>>> May 25th 2016, 05:28:05.880 @timestamp:May 25th 2016, >>>>>>>>>> >>>>>>>>> 05:28:05.880 >>> >>>> host:myhost hostip:10.0.0.1 severity:debug facility:syslog >>>>>>>>>> syslogtag:rsyslogd-pstats: programname:rsyslogd-pstats >>>>>>>>>> >>>>>>>>> logtag:syslog >>> >>>> message:omelasticsearch-myapp-launcher.log queue[DA]: >>>>>>>>>> >>>>>>>>> origin=core.queue >>>>>>> >>>>>>>> size=0 enqueued=0 full=0 discarded.full=0 discarded.nf=0 >>>>>>>>>> >>>>>>>>> maxqsize=0 >>> >>>> _id:AVTnrAlsFO4BDB55DTiF _type:syslog >>>>>>>>>> _index:logstash-syslog-myapp-2016.05.25 _score: >>>>>>>>>> >>>>>>>>>> Thanks, >>>>>>>>>> >>>>>>>>>> Alec >>>>>>>>>> >>>>>>>>>> On Tue, May 24, 2016 at 10:35 PM, David Lang <[email protected]> >>>>>>>>>> >>>>>>>>> wrote: >>> >>>> >>>>>>>>>> It shouldn't matter, but what is the rate of log delivery? is >>>>>>>>>>> >>>>>>>>>> there any >>>>>>> >>>>>>>> chance that it is waiting to deliver a full batch? >>>>>>>>>>> >>>>>>>>>>> I would consider putting this in a ruleset and moving the queue >>>>>>>>>>> >>>>>>>>>> to >>> >>>> the >>>>>>> >>>>>>>> ruleset. I would then have the ruleset contain two items >>>>>>>>>>> >>>>>>>>>>> 1. the output to ES >>>>>>>>>>> 2. a write to a debug log locally (not necessarily the full >>>>>>>>>>> >>>>>>>>>> messages, >>>>>>> >>>>>>>> timestamp would be enough) >>>>>>>>>>> >>>>>>>>>>> you can then see if the local file in growing while things are >>>>>>>>>>> >>>>>>>>>> not >>> >>>> yet >>>>>>> >>>>>>>> showing up in ES to see if the issue is on the sending side or >>>>>>>>>>> >>>>>>>>>> on >>> >>>> the >>>>>>> >>>>>>>> receiving side. >>>>>>>>>>> >>>>>>>>>>> David Lang >>>>>>>>>>> >>>>>>>>>>> On Tue, 24 May 2016, Alec Swan wrote: >>>>>>>>>>> >>>>>>>>>>> Date: Tue, 24 May 2016 22:17:22 -0600 >>>>>>>>>>> >>>>>>>>>>>> From: Alec Swan <[email protected]> >>>>>>>>>>>> Reply-To: rsyslog-users <[email protected]> >>>>>>>>>>>> To: rsyslog-users <[email protected]> >>>>>>>>>>>> Subject: [rsyslog] Logs are delayed being pushed to >>>>>>>>>>>> >>>>>>>>>>> Elasticsearch >>> >>>> >>>>>>>>>>>> >>>>>>>>>>>> Hello, >>>>>>>>>>>> >>>>>>>>>>>> I recently upgraded to rsyslog 8.18 and was happy to see that >>>>>>>>>>>> >>>>>>>>>>> on >>> >>>> disk >>>>>>> >>>>>>>> queues no longer cause rsyslog to get in a bad state. However, >>>>>>>>>>>> >>>>>>>>>>> now I >>>>>>> >>>>>>>> see >>>>>>>>> >>>>>>>>>> very long delays (several hours) of logs being pushed to >>>>>>>>>>>> >>>>>>>>>>> Elasticsearch. >>>>>>> >>>>>>>> It >>>>>>>>>>>> seems that somehow the logs are being buffered on the client >>>>>>>>>>>> >>>>>>>>>>> for >>> >>>> several >>>>>>>>> >>>>>>>>>> hours because eventually they do show up in Elasticsearch. I >>>>>>>>>>>> >>>>>>>>>>> don't see >>>>>>> >>>>>>>> any >>>>>>>>>>>> errors in /var/log/rsyslog/ES-error.log (see config below) or >>>>>>>>>>>> /var/log/messages. >>>>>>>>>>>> >>>>>>>>>>>> I enabled impstats but didn't see any errors related to >>>>>>>>>>>> >>>>>>>>>>> omelasticsearch. >>>>>>>>> >>>>>>>>>> What else can I do to troubleshoot this? >>>>>>>>>>>> >>>>>>>>>>>> Here is my omelasticsearch config: >>>>>>>>>>>> >>>>>>>>>>>> action( >>>>>>>>>>>> type = "omelasticsearch" >>>>>>>>>>>> template = "es-payload" >>>>>>>>>>>> dynSearchIndex = "on" >>>>>>>>>>>> searchIndex = "logstash-index" >>>>>>>>>>>> searchType = "syslog" >>>>>>>>>>>> server = "127.0.0.1" >>>>>>>>>>>> serverport = "9200" >>>>>>>>>>>> uid = "xxx" >>>>>>>>>>>> pwd = "yyy" >>>>>>>>>>>> errorFile = "/var/log/rsyslog/ES-error.log" >>>>>>>>>>>> bulkmode = "on" >>>>>>>>>>>> action.resumeretrycount="-1" # retry if ES is >>>>>>>>>>>> >>>>>>>>>>> unreachable (-1 >>>>>>> >>>>>>>> for >>>>>>>>>>>> infinite retries) >>>>>>>>>>>> action.resumeInterval="60" >>>>>>>>>>>> queue.dequeuebatchsize="1000" # ES bulk size >>>>>>>>>>>> queue.type="linkedlist" >>>>>>>>>>>> queue.size="100000" >>>>>>>>>>>> queue.workerthreads="5" >>>>>>>>>>>> queue.timeoutworkerthreadshutdown="2000" >>>>>>>>>>>> queue.spoolDirectory="/var/spool/rsyslog" >>>>>>>>>>>> queue.filename="omelasticsearch-queue" >>>>>>>>>>>> queue.maxfilesize="100m" >>>>>>>>>>>> queue.maxdiskspace="1g" >>>>>>>>>>>> queue.highwatermark="80000" # when to start spilling to >>>>>>>>>>>> >>>>>>>>>>> disk >>>>>>> >>>>>>>> queue.lowwatermark="20000" # when to stop spilling to >>>>>>>>>>>> >>>>>>>>>>> disk >>>>>>> >>>>>>>> queue.saveonshutdown="on" >>>>>>>>>>>> ) >>>>>>>>>>>> >>>>>>>>>>>> Thanks, >>>>>>>>>>>> >>>>>>>>>>>> Alec >>>>>>>>>>>> _______________________________________________ >>>>>>>>>>>> rsyslog mailing list >>>>>>>>>>>> http://lists.adiscon.net/mailman/listinfo/rsyslog >>>>>>>>>>>> http://www.rsyslog.com/professional-services/ >>>>>>>>>>>> What's up with rsyslog? Follow https://twitter.com/rgerhards >>>>>>>>>>>> NOTE WELL: This is a PUBLIC mailing list, posts are ARCHIVED >>>>>>>>>>>> >>>>>>>>>>> by a >>> >>>> myriad >>>>>>>>> >>>>>>>>>> of sites beyond our control. PLEASE UNSUBSCRIBE and DO NOT POST >>>>>>>>>>>> >>>>>>>>>>> if you >>>>>>> >>>>>>>> DON'T LIKE THAT. >>>>>>>>>>>> >>>>>>>>>>>> _______________________________________________ >>>>>>>>>>>> >>>>>>>>>>> rsyslog mailing list >>>>>>>>>>> http://lists.adiscon.net/mailman/listinfo/rsyslog >>>>>>>>>>> http://www.rsyslog.com/professional-services/ >>>>>>>>>>> What's up with rsyslog? Follow https://twitter.com/rgerhards >>>>>>>>>>> NOTE WELL: This is a PUBLIC mailing list, posts are ARCHIVED by >>>>>>>>>>> >>>>>>>>>> a >>> >>>> myriad >>>>>>> >>>>>>>> of sites beyond our control. PLEASE UNSUBSCRIBE and DO NOT POST >>>>>>>>>>> >>>>>>>>>> if >>> >>>> you >>>>>>> >>>>>>>> DON'T LIKE THAT. >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>> >>>>>>>>>> _______________________________________________ >>>>>>>>> rsyslog mailing list >>>>>>>>> http://lists.adiscon.net/mailman/listinfo/rsyslog >>>>>>>>> http://www.rsyslog.com/professional-services/ >>>>>>>>> What's up with rsyslog? Follow https://twitter.com/rgerhards >>>>>>>>> NOTE WELL: This is a PUBLIC mailing list, posts are ARCHIVED by a >>>>>>>>> >>>>>>>> myriad >>>>>>> >>>>>>>> of sites beyond our control. PLEASE UNSUBSCRIBE and DO NOT POST if >>>>>>>>> >>>>>>>> you >>> >>>> DON'T LIKE THAT. >>>>>>>>> >>>>>>>>> _______________________________________________ >>>>>>>> rsyslog mailing list >>>>>>>> http://lists.adiscon.net/mailman/listinfo/rsyslog >>>>>>>> http://www.rsyslog.com/professional-services/ >>>>>>>> What's up with rsyslog? Follow https://twitter.com/rgerhards >>>>>>>> NOTE WELL: This is a PUBLIC mailing list, posts are ARCHIVED by a >>>>>>>> >>>>>>> myriad of sites beyond our control. PLEASE UNSUBSCRIBE and DO NOT >>>>>>> >>>>>> POST if >>> >>>> you DON'T LIKE THAT. >>>>>>> _______________________________________________ >>>>>>> rsyslog mailing list >>>>>>> http://lists.adiscon.net/mailman/listinfo/rsyslog >>>>>>> http://www.rsyslog.com/professional-services/ >>>>>>> What's up with rsyslog? Follow https://twitter.com/rgerhards >>>>>>> NOTE WELL: This is a PUBLIC mailing list, posts are ARCHIVED by a >>>>>>> >>>>>> myriad >>> >>>> of sites beyond our control. PLEASE UNSUBSCRIBE and DO NOT POST if you >>>>>>> DON'T LIKE THAT. >>>>>>> >>>>>>> >>>>>> >>>>>> >>>>> _______________________________________________ >>>> rsyslog mailing list >>>> http://lists.adiscon.net/mailman/listinfo/rsyslog >>>> http://www.rsyslog.com/professional-services/ >>>> What's up with rsyslog? Follow https://twitter.com/rgerhards >>>> NOTE WELL: This is a PUBLIC mailing list, posts are ARCHIVED by a myriad >>>> >>> of sites beyond our control. PLEASE UNSUBSCRIBE and DO NOT POST if you >>> DON'T LIKE THAT. >>> _______________________________________________ >>> rsyslog mailing list >>> http://lists.adiscon.net/mailman/listinfo/rsyslog >>> http://www.rsyslog.com/professional-services/ >>> What's up with rsyslog? Follow https://twitter.com/rgerhards >>> NOTE WELL: This is a PUBLIC mailing list, posts are ARCHIVED by a myriad >>> of sites beyond our control. PLEASE UNSUBSCRIBE and DO NOT POST if you >>> DON'T LIKE THAT. >>> >>> _______________________________________________ >> rsyslog mailing list >> http://lists.adiscon.net/mailman/listinfo/rsyslog >> http://www.rsyslog.com/professional-services/ >> What's up with rsyslog? Follow https://twitter.com/rgerhards >> NOTE WELL: This is a PUBLIC mailing list, posts are ARCHIVED by a myriad >> of sites beyond our control. PLEASE UNSUBSCRIBE and DO NOT POST if you >> DON'T LIKE THAT. >> >> _______________________________________________ > rsyslog mailing list > http://lists.adiscon.net/mailman/listinfo/rsyslog > http://www.rsyslog.com/professional-services/ > What's up with rsyslog? Follow https://twitter.com/rgerhards > NOTE WELL: This is a PUBLIC mailing list, posts are ARCHIVED by a myriad > of sites beyond our control. PLEASE UNSUBSCRIBE and DO NOT POST if you > DON'T LIKE THAT. > _______________________________________________ rsyslog mailing list http://lists.adiscon.net/mailman/listinfo/rsyslog http://www.rsyslog.com/professional-services/ What's up with rsyslog? Follow https://twitter.com/rgerhards NOTE WELL: This is a PUBLIC mailing list, posts are ARCHIVED by a myriad of sites beyond our control. PLEASE UNSUBSCRIBE and DO NOT POST if you DON'T LIKE THAT.

