Great, I'd be happy if 2 things happen: 1. We retry forever until committed 2. We don't loose in-flight batch on restart
On Wed, Apr 20, 2016 at 11:28 PM, Rainer Gerhards <[email protected]> wrote: > 2016-04-21 8:25 GMT+02:00 Kane Kim <[email protected]>: > > What would happen if it will be retried forever at this point? Would it > > affect any other queues or only current action queue. If it will not > affect > > anything else I'd really prefer it to block there until successfully > > committed. > > I need to read the code to be 100% sure. But I would expect that it > holds processing of just the queue where it is running on (98% sure). > > Rainer > > > > On Wed, Apr 20, 2016 at 11:16 PM, Rainer Gerhards < > [email protected]> > > wrote: > > > >> 2016-04-21 3:22 GMT+02:00 David Lang <[email protected]>: > >> > Thanks, I didn't know that you had been able to take any time to think > >> about > >> > this issue yet. > >> > >> Sorry for not being more explicit about it. But there was not yet > >> anything that could be posted. I did this as a side-activity all since > >> this opened. The plan acutally is: > >> > >> 1. get a testbench test to expose that problem -- this greatly eases > >> debugging > >> line of thought is to update omfile, which is needed anyhow, so > >> that we can induce errors via it > >> 2. add missing functionality to write error file > >> this does not directly address the issue, but is useful and > >> ensures we are on the right path > >> 3. check the end transaction endpoint. After review of old notes and > >> some thinking, I think that > >> part of the "not yet implemented" comment is actually right. It's > >> not as easy as suggested by it, > >> but we probably can do better. > >> > >> > > >> > It sounds like you are tenitively agreeing with my speculation that > the > >> end > >> > of the block processing is happening just a little sooner than it > should, > >> > preventing the chance to detect output errors. > >> > >> No, I don't think so. I think it happens as it should, but we are not > >> handling the full situation. Originally, it was planned to write an > >> error file (point 2 above) when we cannot commit at end of > >> transaction. But we probably can give it some retries at that point as > >> well. In v8, we **need** to buffer transactional data until the end of > >> batch. If we wouldn't do that, we would always need to commit after > >> each and every transaction (because we need processing "completed" > >> when we advance to the next action). The, at end of batch, the > >> transaction commit processing happens. Right now, we have one try, and > >> then give up (unfortunately without any error indication). So we need > >> to go into that part of processing and enhance/fix it. > >> > >> I hope this clarifies. > >> Rainer > >> > >> > > >> > David Lang > >> > > >> > > >> > On Wed, 20 Apr 2016, Rainer Gerhards wrote: > >> > > >> >> I should also say that it had a release goal, so it is on my active > >> to-do > >> >> list. > >> >> > >> >> Sent from phone, thus brief. > >> >> Am 20.04.2016 19:40 schrieb "Rainer Gerhards" < > [email protected] > >> >: > >> >> > >> >>> Yup... Look aty work Log - still no time... Was meant for last > release > >> >>> but > >> >>> then came libfastjson work. > >> >>> > >> >>> Sent from phone, thus brief. > >> >>> Am 20.04.2016 19:39 schrieb "David Lang" <[email protected]>: > >> >>> > >> >>>> Rainer, ping on this issue. > >> >>>> > >> >>>> David Lang > >> >>>> > >> >>>> On Tue, 29 Mar 2016, Kane Kim wrote: > >> >>>> > >> >>>> Date: Tue, 29 Mar 2016 10:46:49 -0700 > >> >>>>> > >> >>>>> From: Kane Kim <[email protected]> > >> >>>>> Reply-To: rsyslog-users <[email protected]> > >> >>>>> To: rsyslog-users <[email protected]> > >> >>>>> Subject: Re: [rsyslog] retry if output module returns > >> RS_RET_SUSPENDED > >> >>>>> > >> >>>>> Updated config to: > >> >>>>> module(load="omrelp") > >> >>>>> local0.* action(type="omrelp" > >> >>>>> target="10.0.0.1" > >> >>>>> port="2514" > >> >>>>> name="debug_relp" > >> >>>>> queue.filename="relp_queue" > >> >>>>> queue.maxdiskspace="1G" > >> >>>>> queue.saveonshutdown="on" > >> >>>>> queue.spoolDirectory="/var/spool/rsyslog" > >> >>>>> queue.type="Disk" > >> >>>>> action.resumeRetryCount="-1" > >> >>>>> action.resumeInterval="1" > >> >>>>> ) > >> >>>>> > >> >>>>> Still message is lost: > >> >>>>> > >> >>>>> 3489.658349703:imuxsock.c : Message from UNIX socket: #3 > >> >>>>> 3489.658367582:imuxsock.c : main Q: qqueueAdd: entry added, > size > >> >>>>> now > >> >>>>> log 1, phys 1 entries > >> >>>>> 3489.658372561:imuxsock.c : main Q: EnqueueMsg advised worker > >> start > >> >>>>> 3489.658374985:imuxsock.c : --------imuxsock calling select, > >> active > >> >>>>> file descriptors (max 3): 3 > >> >>>>> 3489.658384201:main Q:Reg/w0 : wti 0x23c8ba0: worker awoke from > idle > >> >>>>> processing > >> >>>>> 3489.658386792:main Q:Reg/w0 : DeleteProcessedBatch: we deleted 0 > >> >>>>> objects > >> >>>>> and enqueued 0 objects > >> >>>>> 3489.658388548:main Q:Reg/w0 : doDeleteBatch: delete batch from > >> store, > >> >>>>> new > >> >>>>> sizes: log 1, phys 1 > >> >>>>> 3489.658391356:main Q:Reg/w0 : processBATCH: batch of 1 elements > >> must > >> >>>>> be > >> >>>>> processed > >> >>>>> 3489.658392786:main Q:Reg/w0 : processBATCH: next msg 0: > <134>Mar 29 > >> >>>>> 17:44:49 logger: test > >> >>>>> 3489.658395027:main Q:Reg/w0 : PRIFILT 'local0.*' > >> >>>>> 3489.658397540:main Q:Reg/w0 : pmask: X X X X X X X > X X > >> >>>>> X X > >> >>>>> X X X X X FF X X X X X X X X X > >> >>>>> 3489.658413235:main Q:Reg/w0 : PRIFILT condition result is 1 > >> >>>>> 3489.658414686:main Q:Reg/w0 : ACTION 0 > >> >>>>> [omrelp:action(type="omrelp" > >> >>>>> ...)] > >> >>>>> 3489.658417648:main Q:Reg/w0 : executing action 0 > >> >>>>> 3489.658419520:main Q:Reg/w0 : action 'debug_relp': called, > logging > >> to > >> >>>>> omrelp (susp 0/0, direct q 0) > >> >>>>> 3489.658433462:main Q:Reg/w0 : strm 0x23b74c0: file > -1(relp_queue) > >> >>>>> flush, > >> >>>>> buflen 410 > >> >>>>> 3489.658435496:main Q:Reg/w0 : strmPhysWrite, stream 0x23b74c0, > len > >> >>>>> 410 > >> >>>>> 3489.658484243:main Q:Reg/w0 : file > >> >>>>> '/var/spool/rsyslog/relp_queue.00000001' opened as #5 with mode > 384 > >> >>>>> 3489.658487347:main Q:Reg/w0 : strm 0x23b74c0: opened file > >> >>>>> '/var/spool/rsyslog/relp_queue.00000001' for WRITE as 5 > >> >>>>> 3489.658500498:main Q:Reg/w0 : strm 0x23b74c0: file 5 write wrote > >> 410 > >> >>>>> bytes > >> >>>>> 3489.658502751:main Q:Reg/w0 : debug_relp queue: write wrote 410 > >> >>>>> octets > >> >>>>> to > >> >>>>> disk, queue disk size now 410 octets, EnqOnly:0 > >> >>>>> 3489.658504568:main Q:Reg/w0 : debug_relp queue: qqueueAdd: entry > >> >>>>> added, > >> >>>>> size now log 1, phys 1 entries > >> >>>>> 3489.658506705:main Q:Reg/w0 : debug_relp queue:Reg: high > activity - > >> >>>>> starting 1 additional worker thread(s). > >> >>>>> 3489.658528951:main Q:Reg/w0 : debug_relp queue:Reg: started with > >> >>>>> state > >> >>>>> 0, > >> >>>>> num workers now 1 > >> >>>>> 3489.658530689:main Q:Reg/w0 : debug_relp queue: EnqueueMsg > advised > >> >>>>> worker > >> >>>>> start > >> >>>>> 3489.658532183:main Q:Reg/w0 : action 'debug_relp': set suspended > >> >>>>> state > >> >>>>> to > >> >>>>> 0 > >> >>>>> 3489.658533672:main Q:Reg/w0 : PRIFILT > >> >>>>> '*.info;mail.none;authpriv.none;cron.none' > >> >>>>> 3489.658536164:main Q:Reg/w0 : pmask: 7F 7F X 7F 7F 7F 7F > 7F 7F > >> >>>>> X X > >> >>>>> 7F 7F 7F 7F 7F 7F 7F 7F 7F 7F 7F 7F 7F 7F 7F > >> >>>>> 3489.658552240:main Q:Reg/w0 : PRIFILT condition result is 1 > >> >>>>> 3489.658553393:main Q:Reg/w0 : ACTION 1 > >> >>>>> [builtin:omfile:/var/log/messages] > >> >>>>> 3489.658556301:main Q:Reg/w0 : executing action 1 > >> >>>>> 3489.658557878:main Q:Reg/w0 : action 'action 1': called, > logging to > >> >>>>> builtin:omfile (susp 0/0, direct q 1) > >> >>>>> 3489.658561265:main Q:Reg/w0 : action 'action 1': is > transactional - > >> >>>>> executing in commit phase > >> >>>>> 3489.658563696:main Q:Reg/w0 : Action 1 transitioned to state: > itx > >> >>>>> 3489.658564902:main Q:Reg/w0 : action 'action 1': set suspended > >> state > >> >>>>> to 0 > >> >>>>> 3489.658566053:main Q:Reg/w0 : PRIFILT 'authpriv.*' > >> >>>>> 3489.658568233:main Q:Reg/w0 : pmask: X X X X X X X > X X > >> >>>>> X > >> >>>>> FF > >> >>>>> X X X X X X X X X X X X X X X > >> >>>>> 3489.658582888:main Q:Reg/w0 : PRIFILT condition result is 0 > >> >>>>> 3489.658584058:main Q:Reg/w0 : PRIFILT 'mail.*' > >> >>>>> 3489.658586243:main Q:Reg/w0 : pmask: X X FF X X X X > X X > >> >>>>> X X > >> >>>>> X X X X X X X X X X X X X X X > >> >>>>> 3489.658600961:main Q:Reg/w0 : PRIFILT condition result is 0 > >> >>>>> 3489.658602136:main Q:Reg/w0 : PRIFILT 'cron.*' > >> >>>>> 3489.658604562:main Q:Reg/w0 : pmask: X X X X X X X > X X > >> >>>>> FF X > >> >>>>> X X X X X X X X X X X X X X X > >> >>>>> 3489.658619159:main Q:Reg/w0 : PRIFILT condition result is 0 > >> >>>>> 3489.658620352:main Q:Reg/w0 : PRIFILT '*.emerg' > >> >>>>> 3489.658622551:main Q:Reg/w0 : pmask: 1 1 1 1 1 1 1 > 1 1 > >> >>>>> 1 1 > >> >>>>> 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 > >> >>>>> 3489.658638435:main Q:Reg/w0 : PRIFILT condition result is 0 > >> >>>>> 3489.658639529:main Q:Reg/w0 : PRIFILT 'uucp,news.crit' > >> >>>>> 3489.658641711:main Q:Reg/w0 : pmask: X X X X X X X > 7 7 > >> >>>>> X X > >> >>>>> X X X X X X X X X X X X X X X > >> >>>>> 3489.658656380:main Q:Reg/w0 : PRIFILT condition result is 0 > >> >>>>> 3489.658657532:main Q:Reg/w0 : PRIFILT 'local7.*' > >> >>>>> 3489.658659794:main Q:Reg/w0 : pmask: X X X X X X X > X X > >> >>>>> X X > >> >>>>> X X X X X X X X X X X X FF X X > >> >>>>> 3489.658674346:main Q:Reg/w0 : PRIFILT condition result is 0 > >> >>>>> 3489.658675686:main Q:Reg/w0 : END batch execution phase, > entering > >> to > >> >>>>> commit phase [processed 1 of 1 messages] > >> >>>>> 3489.658677400:main Q:Reg/w0 : actionCommitAll: action 1, state > 1, > >> nbr > >> >>>>> to > >> >>>>> commit 0 isTransactional 1 > >> >>>>> 3489.658679044:main Q:Reg/w0 : doTransaction: have > commitTransaction > >> >>>>> IF, > >> >>>>> using that, pWrkrInfo 0x23c8ce0 > >> >>>>> 3489.658680668:main Q:Reg/w0 : entering > >> actionCallCommitTransaction(), > >> >>>>> state: itx, actionNbr 1, nMsgs 1 > >> >>>>> 3489.658682704:main Q:Reg/w0 : omfile: write to stream, > pData->pStrm > >> >>>>> 0x7f1ef4002480, lenBuf 42, strt data Mar 29 17:44:49 314959ac8cb0 > >> >>>>> logger: > >> >>>>> test > >> >>>>> > >> >>>>> 3489.658685296:main Q:Reg/w0 : strm 0x7f1ef4002480: file > 4(messages) > >> >>>>> flush, buflen 42 > >> >>>>> 3489.658686866:main Q:Reg/w0 : strmPhysWrite, stream > 0x7f1ef4002480, > >> >>>>> len 42 > >> >>>>> 3489.658696046:main Q:Reg/w0 : strm 0x7f1ef4002480: file 4 write > >> wrote > >> >>>>> 42 > >> >>>>> bytes > >> >>>>> 3489.658697884:main Q:Reg/w0 : Action 1 transitioned to state: > rdy > >> >>>>> 3489.658699301:main Q:Reg/w0 : Action 1 transitioned to state: > itx > >> >>>>> 3489.658700575:main Q:Reg/w0 : Action 1 transitioned to state: > rdy > >> >>>>> 3489.658701773:main Q:Reg/w0 : actionCommit, in retry loop, iRet > 0 > >> >>>>> 3489.658703176:main Q:Reg/w0 : processBATCH: batch of 1 elements > has > >> >>>>> been > >> >>>>> processed > >> >>>>> 3489.658704989:main Q:Reg/w0 : regular consumer finished, iret=0, > >> >>>>> szlog > >> >>>>> 0 > >> >>>>> sz phys 1 > >> >>>>> 3489.658707214:main Q:Reg/w0 : DeleteProcessedBatch: we deleted 1 > >> >>>>> objects > >> >>>>> and enqueued 0 objects > >> >>>>> 3489.658708626:main Q:Reg/w0 : doDeleteBatch: delete batch from > >> store, > >> >>>>> new > >> >>>>> sizes: log 0, phys 0 > >> >>>>> 3489.658710186:main Q:Reg/w0 : regular consumer finished, iret=4, > >> >>>>> szlog > >> >>>>> 0 > >> >>>>> sz phys 0 > >> >>>>> 3489.658711550:main Q:Reg/w0 : main Q:Reg/w0: worker IDLE, > waiting > >> for > >> >>>>> work. > >> >>>>> 3489.660013052:7f1efbb14700: thread created, tid 30310, name > >> >>>>> 'rs:debug_relp > >> >>>>> queue:Reg' > >> >>>>> 3489.660298870:debug_relp queue:Reg/w0: wti 0x23c78b0: worker > >> starting > >> >>>>> 3489.660645336:debug_relp queue:Reg/w0: DeleteProcessedBatch: we > >> >>>>> deleted > >> >>>>> 0 > >> >>>>> objects and enqueued 0 objects > >> >>>>> 3489.660922752:debug_relp queue:Reg/w0: doDeleteBatch: delete > batch > >> >>>>> from > >> >>>>> store, new sizes: log 1, phys 1 > >> >>>>> 3489.660947418:debug_relp queue:Reg/w0: file > >> >>>>> '/var/spool/rsyslog/relp_queue.00000001' opened as #6 with mode > 384 > >> >>>>> 3489.660959920:debug_relp queue:Reg/w0: strm 0x23b7760: opened > file > >> >>>>> '/var/spool/rsyslog/relp_queue.00000001' for READ as 6 > >> >>>>> 3489.660967370:debug_relp queue:Reg/w0: strm 0x23b7760: file 6 > read > >> 410 > >> >>>>> bytes > >> >>>>> 3489.661768120:debug_relp queue:Reg/w0: action 'debug_relp': is > >> >>>>> transactional - executing in commit phase > >> >>>>> 3489.661782408:debug_relp queue:Reg/w0: wti 0x23c78b0: we need to > >> >>>>> create > >> >>>>> a > >> >>>>> new action worker instance for action 0 > >> >>>>> 3489.661789583:debug_relp queue:Reg/w0: relp engine created new > >> client > >> >>>>> 0x7f1ef00027e0 > >> >>>>> 3489.661794073:debug_relp queue:Reg/w0: wti 0x23c78b0: created > action > >> >>>>> worker instance 1 for action 0 > >> >>>>> 3489.661798129:debug_relp queue:Reg/w0: omrelp: beginTransaction > >> >>>>> ^C3497.710058579:main thread : DoDie called. > >> >>>>> 3497.710272055:main thread : exiting on signal 2 > >> >>>>> 3497.710576430:main thread : Terminating input threads... > >> >>>>> 3497.710765753:main thread : request term via SIGTTIN for input > >> >>>>> thread > >> >>>>> 'imuxsock' 0xfcb16700 > >> >>>>> 3497.711033925:imuxsock.c : thrdStarter: usrThrdMain imuxsock > - > >> >>>>> 0x7f1efcb16700 returned with iRet 0, exiting now. > >> >>>>> 3497.711447372:main thread : non-cancel input thread > termination > >> >>>>> succeeded for thread imuxsock 0xfcb16700 > >> >>>>> 3497.713028584:imuxsock.c : destructor for debug call stack > >> >>>>> 0x7f1eec0008c0 called > >> >>>>> 3497.713390963:main thread : imuxsock: unlinking unix socket > >> file[0] > >> >>>>> /dev/log > >> >>>>> 3497.713734513:main thread : main Q: qqueueAdd: entry added, > size > >> >>>>> now > >> >>>>> log 1, phys 1 entries > >> >>>>> 3497.713906635:main thread : main Q: EnqueueMsg advised worker > >> start > >> >>>>> 3497.714257550:main Q:Reg/w0 : wti 0x23c8ba0: worker awoke from > idle > >> >>>>> processing > >> >>>>> 3497.714544345:main Q:Reg/w0 : DeleteProcessedBatch: we deleted 0 > >> >>>>> objects > >> >>>>> and enqueued 0 objects > >> >>>>> 3497.715372923:main Q:Reg/w0 : doDeleteBatch: delete batch from > >> store, > >> >>>>> new > >> >>>>> sizes: log 1, phys 1 > >> >>>>> 3497.715611163:main Q:Reg/w0 : processBATCH: batch of 1 elements > >> must > >> >>>>> be > >> >>>>> processed > >> >>>>> 3497.715762447:main Q:Reg/w0 : processBATCH: next msg 0: [origin > >> >>>>> software="rsyslogd" swVersion="8.18.0.master" x-pid="30306" > x-info=" > >> >>>>> http://www.rsyslog.com"] exiting on signal 2. > >> >>>>> 3497.715965911:main Q:Reg/w0 : PRIFILT 'local0.*' > >> >>>>> 3497.715976154:main Q:Reg/w0 : pmask: X X X X X X X > X X > >> >>>>> X X > >> >>>>> X X X X X FF X X X X X X X X X > >> >>>>> 3497.716332698:main Q:Reg/w0 : PRIFILT condition result is 0 > >> >>>>> 3497.716337713:main Q:Reg/w0 : PRIFILT > >> >>>>> '*.info;mail.none;authpriv.none;cron.none' > >> >>>>> 3497.716347729:main Q:Reg/w0 : pmask: 7F 7F X 7F 7F 7F 7F > 7F 7F > >> >>>>> X X > >> >>>>> 7F 7F 7F 7F 7F 7F 7F 7F 7F 7F 7F 7F 7F 7F 7F > >> >>>>> 3497.716424684:main Q:Reg/w0 : PRIFILT condition result is 1 > >> >>>>> 3497.716429241:main Q:Reg/w0 : ACTION 1 > >> >>>>> [builtin:omfile:/var/log/messages] > >> >>>>> 3497.716439928:main Q:Reg/w0 : executing action 1 > >> >>>>> 3497.716445132:main Q:Reg/w0 : action 'action 1': called, > logging to > >> >>>>> builtin:omfile (susp 0/0, direct q 1) > >> >>>>> 3497.716452142:main Q:Reg/w0 : action 'action 1': is > transactional - > >> >>>>> executing in commit phase > >> >>>>> 3497.716457817:main Q:Reg/w0 : Action 1 transitioned to state: > itx > >> >>>>> 3497.716462371:main Q:Reg/w0 : action 'action 1': set suspended > >> state > >> >>>>> to 0 > >> >>>>> 3497.716466771:main Q:Reg/w0 : PRIFILT 'authpriv.*' > >> >>>>> 3497.716476330:main Q:Reg/w0 : pmask: X X X X X X X > X X > >> >>>>> X > >> >>>>> FF > >> >>>>> X X X X X X X X X X X X X X X > >> >>>>> 3497.716796250:main Q:Reg/w0 : PRIFILT condition result is 0 > >> >>>>> 3497.716802372:main Q:Reg/w0 : PRIFILT 'mail.*' > >> >>>>> 3497.716811926:main Q:Reg/w0 : pmask: X X FF X X X X > X X > >> >>>>> X X > >> >>>>> X X X X X X X X X X X X X X X > >> >>>>> 3497.717436983:main Q:Reg/w0 : PRIFILT condition result is 0 > >> >>>>> 3497.717444114:main Q:Reg/w0 : PRIFILT 'cron.*' > >> >>>>> 3497.717454023:main Q:Reg/w0 : pmask: X X X X X X X > X X > >> >>>>> FF X > >> >>>>> X X X X X X X X X X X X X X X > >> >>>>> 3497.717823841:main Q:Reg/w0 : PRIFILT condition result is 0 > >> >>>>> 3497.717829094:main Q:Reg/w0 : PRIFILT '*.emerg' > >> >>>>> 3497.717838767:main Q:Reg/w0 : pmask: 1 1 1 1 1 1 1 > 1 1 > >> >>>>> 1 1 > >> >>>>> 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 > >> >>>>> 3497.717916499:main Q:Reg/w0 : PRIFILT condition result is 0 > >> >>>>> 3497.717920817:main Q:Reg/w0 : PRIFILT 'uucp,news.crit' > >> >>>>> 3497.717930304:main Q:Reg/w0 : pmask: X X X X X X X > 7 7 > >> >>>>> X X > >> >>>>> X X X X X X X X X X X X X X X > >> >>>>> 3497.718043828:main Q:Reg/w0 : PRIFILT condition result is 0 > >> >>>>> 3497.718044097:main Q:Reg/w0 : PRIFILT 'local7.*' > >> >>>>> 3497.718055458:main Q:Reg/w0 : pmask: X X X X X X X > X X > >> >>>>> X X > >> >>>>> X X X X X X X X X X X X FF X X > >> >>>>> 3497.718838257:main Q:Reg/w0 : PRIFILT condition result is 0 > >> >>>>> 3497.718842803:main Q:Reg/w0 : END batch execution phase, > entering > >> to > >> >>>>> commit phase [processed 1 of 1 messages] > >> >>>>> 3497.718847686:main Q:Reg/w0 : actionCommitAll: action 1, state > 1, > >> nbr > >> >>>>> to > >> >>>>> commit 0 isTransactional 1 > >> >>>>> 3497.718852750:main Q:Reg/w0 : doTransaction: have > commitTransaction > >> >>>>> IF, > >> >>>>> using that, pWrkrInfo 0x23c8ce0 > >> >>>>> 3497.718857172:main Q:Reg/w0 : entering > >> actionCallCommitTransaction(), > >> >>>>> state: itx, actionNbr 1, nMsgs 1 > >> >>>>> 3497.718862516:main Q:Reg/w0 : omfile: write to stream, > pData->pStrm > >> >>>>> 0x7f1ef4002480, lenBuf 161, strt data Mar 29 17:44:57 314959ac8cb0 > >> >>>>> rsyslogd: [origin software="rsyslogd" swVersion="8.18.0.master" > >> >>>>> x-pid="30306" x-info="http://www.rs > >> >>>>> 3497.718868427:main Q:Reg/w0 : strm 0x7f1ef4002480: file > 4(messages) > >> >>>>> flush, buflen 161 > >> >>>>> 3497.718873173:main Q:Reg/w0 : strmPhysWrite, stream > 0x7f1ef4002480, > >> >>>>> len > >> >>>>> 161 > >> >>>>> 3497.718890732:main Q:Reg/w0 : strm 0x7f1ef4002480: file 4 write > >> wrote > >> >>>>> 161 > >> >>>>> bytes > >> >>>>> 3497.718896519:main Q:Reg/w0 : Action 1 transitioned to state: > rdy > >> >>>>> 3497.718901151:main Q:Reg/w0 : Action 1 transitioned to state: > itx > >> >>>>> 3497.718905316:main Q:Reg/w0 : Action 1 transitioned to state: > rdy > >> >>>>> 3497.718909316:main Q:Reg/w0 : actionCommit, in retry loop, iRet > 0 > >> >>>>> 3497.718913541:main Q:Reg/w0 : processBATCH: batch of 1 elements > has > >> >>>>> been > >> >>>>> processed > >> >>>>> 3497.718918158:main Q:Reg/w0 : regular consumer finished, iret=0, > >> >>>>> szlog > >> >>>>> 0 > >> >>>>> sz phys 1 > >> >>>>> 3497.718923739:main Q:Reg/w0 : DeleteProcessedBatch: we deleted 1 > >> >>>>> objects > >> >>>>> and enqueued 0 objects > >> >>>>> 3497.718928090:main Q:Reg/w0 : doDeleteBatch: delete batch from > >> store, > >> >>>>> new > >> >>>>> sizes: log 0, phys 0 > >> >>>>> 3497.718932606:main Q:Reg/w0 : regular consumer finished, iret=4, > >> >>>>> szlog > >> >>>>> 0 > >> >>>>> sz phys 0 > >> >>>>> 3497.718936888:main Q:Reg/w0 : main Q:Reg/w0: worker IDLE, > waiting > >> for > >> >>>>> work. > >> >>>>> 3497.718942403:main thread : Terminating main queue... > >> >>>>> 3497.718946280:main thread : main Q: initiating worker thread > >> >>>>> shutdown > >> >>>>> sequence > >> >>>>> 3497.718950170:main thread : main Q: trying shutdown of regular > >> >>>>> workers > >> >>>>> 3497.718957067:main thread : sent SIGTTIN to worker thread > >> >>>>> 0xfc315700 > >> >>>>> 3497.718961130:main thread : main Q:Reg: waiting 1500ms on > worker > >> >>>>> thread > >> >>>>> termination, 1 still running > >> >>>>> 3497.718968288:main Q:Reg/w0 : wti 0x23c8ba0: worker awoke from > idle > >> >>>>> processing > >> >>>>> 3497.718972108:main Q:Reg/w0 : DeleteProcessedBatch: we deleted 0 > >> >>>>> objects > >> >>>>> and enqueued 0 objects > >> >>>>> 3497.718975498:main Q:Reg/w0 : doDeleteBatch: delete batch from > >> store, > >> >>>>> new > >> >>>>> sizes: log 0, phys 0 > >> >>>>> 3497.718979083:main Q:Reg/w0 : regular consumer finished, iret=4, > >> >>>>> szlog > >> >>>>> 0 > >> >>>>> sz phys 0 > >> >>>>> 3497.718982620:main Q:Reg/w0 : wti 0x23c8ba0: terminating worker > >> >>>>> terminateRet=5, bInactivityTOOccured=0 > >> >>>>> 3497.718986378:main Q:Reg/w0 : DDDD: wti 0x23c8ba0: worker > cleanup > >> >>>>> action > >> >>>>> instances > >> >>>>> 3497.718990027:main Q:Reg/w0 : wti 0x23c8ba0, action 0, ptr (nil) > >> >>>>> 3497.718993327:main Q:Reg/w0 : wti 0x23c8ba0, action 1, ptr > >> >>>>> 0x7f1ef40022a0 > >> >>>>> 3497.718997548:main Q:Reg/w0 : wti 0x23c8ba0, action 2, ptr (nil) > >> >>>>> 3497.719000789:main Q:Reg/w0 : wti 0x23c8ba0, action 3, ptr (nil) > >> >>>>> 3497.719003996:main Q:Reg/w0 : wti 0x23c8ba0, action 4, ptr (nil) > >> >>>>> 3497.719007201:main Q:Reg/w0 : wti 0x23c8ba0, action 5, ptr (nil) > >> >>>>> 3497.719010409:main Q:Reg/w0 : wti 0x23c8ba0, action 6, ptr (nil) > >> >>>>> 3497.719013604:main Q:Reg/w0 : wti 0x23c8ba0, action 7, ptr (nil) > >> >>>>> 3497.719016916:main Q:Reg/w0 : wti 0x23c8ba0: worker exiting > >> >>>>> 3497.719020757:main Q:Reg/w0 : main Q:Reg: Worker thread 23c8ba0, > >> >>>>> terminated, num workers now 0 > >> >>>>> 3497.719090631:main thread : main Q: regular queue workers shut > >> >>>>> down. > >> >>>>> 3497.719097597:main thread : main Q: checking to see if we > need to > >> >>>>> cancel any worker threads of the primary queue > >> >>>>> 3497.719101335:main thread : main Q: worker threads terminated, > >> >>>>> remaining queue size log 0, phys 0. > >> >>>>> 3497.719111115:main thread : main Q: queue (type 0) will lose 0 > >> >>>>> messages, destroying... > >> >>>>> 3497.719501847:main Q:Reg/w0 : destructor for debug call stack > >> >>>>> 0x7f1ef40008c0 called > >> >>>>> 3497.719547524:main thread : Terminating outputs... > >> >>>>> 3497.719553777:main thread : destructing ruleset 0x23ab150, > name > >> >>>>> 0x23ab190 > >> >>>>> 3497.719558276:main thread : debug_relp queue: initiating > worker > >> >>>>> thread > >> >>>>> shutdown sequence > >> >>>>> 3497.719561940:main thread : debug_relp queue: trying shutdown > of > >> >>>>> regular workers > >> >>>>> 3497.719568068:main thread : sent SIGTTIN to worker thread > >> >>>>> 0xfbb14700 > >> >>>>> 3497.719587478:main thread : debug_relp queue:Reg: waiting 0ms > on > >> >>>>> worker > >> >>>>> thread termination, 1 still running > >> >>>>> 3497.719593945:debug_relp queue:Reg/w0: end relpSessConnect, iRet > >> 10014 > >> >>>>> 3497.719599684:debug_relp queue:Reg/w0: Action 0 transitioned to > >> state: > >> >>>>> rtry > >> >>>>> 3497.719603211:debug_relp queue:Reg/w0: doTransaction: action 0, > >> >>>>> currIParam > >> >>>>> 1 > >> >>>>> 3497.719607186:debug_relp queue:Reg/w0: actionDoRetry: debug_relp > >> enter > >> >>>>> loop, iRetries=0 > >> >>>>> 3497.719677407:main thread : debug_relp queue:Reg: timeout > waiting > >> >>>>> on > >> >>>>> worker thread termination > >> >>>>> 3497.719677554:main thread : sent SIGTTIN to worker thread > >> >>>>> 0xfbb14700 > >> >>>>> 3497.719677725:main thread : debug_relp queue: regular shutdown > >> >>>>> timed > >> >>>>> out on primary queue (this is OK) > >> >>>>> 3497.719677897:main thread : debug_relp queue: trying to > shutdown > >> >>>>> workers within Action Timeoutsetting EnqOnly mode > >> >>>>> 3497.719678191:main thread : debug_relp queue: trying immediate > >> >>>>> shutdown > >> >>>>> of regular workers (if any) > >> >>>>> 3497.719678362:main thread : sent SIGTTIN to worker thread > >> >>>>> 0xfbb14700 > >> >>>>> 3497.719678558:main thread : debug_relp queue:Reg: waiting > 1000ms > >> on > >> >>>>> worker thread termination, 1 still running > >> >>>>> 3497.719678754:debug_relp queue:Reg/w0: end relpSessConnect, iRet > >> 10014 > >> >>>>> 3497.719678877:debug_relp queue:Reg/w0: actionDoRetry: debug_relp > >> >>>>> action->tryResume returned -2007 > >> >>>>> 3497.719678999:debug_relp queue:Reg/w0: actionDoRetry: debug_relp > >> check > >> >>>>> for > >> >>>>> max retries, iResumeRetryCount -1, iRetries 0 > >> >>>>> 3498.720321005:debug_relp queue:Reg/w0: actionTryResume: action > >> >>>>> 0x23bebf0 > >> >>>>> state: rtry, next retry (if applicable): 0 [now 1459273498] > >> >>>>> 3498.720769277:debug_relp queue:Reg/w0: actionCommit, in retry > loop, > >> >>>>> iRet > >> >>>>> -2007 > >> >>>>> 3498.721034552:debug_relp queue:Reg/w0: regular consumer finished, > >> >>>>> iret=-2007, szlog 0 sz phys 1 > >> >>>>> 3498.721240322:debug_relp queue:Reg/w0: DeleteProcessedBatch: we > >> >>>>> deleted > >> >>>>> 1 > >> >>>>> objects and enqueued 0 objects > >> >>>>> 3498.721396664:debug_relp queue:Reg/w0: doDeleteBatch: delete > batch > >> >>>>> from > >> >>>>> store, new sizes: log 0, phys 0 > >> >>>>> 3498.721831840:debug_relp queue:Reg/w0: wti 0x23c78b0: terminating > >> >>>>> worker > >> >>>>> because of TERMINATE_NOW mode, del iRet 0 > >> >>>>> 3498.721991265:debug_relp queue:Reg/w0: DDDD: wti 0x23c78b0: > worker > >> >>>>> cleanup > >> >>>>> action instances > >> >>>>> 3498.722274489:debug_relp queue:Reg/w0: wti 0x23c78b0, action 0, > ptr > >> >>>>> 0x7f1ef00020d0 > >> >>>>> 3498.722431811:debug_relp queue:Reg/w0: relp engine destructing > >> client > >> >>>>> 0x7f1ef00027e0 > >> >>>>> 3498.723389457:debug_relp queue:Reg/w0: wti 0x23c78b0, action 1, > ptr > >> >>>>> (nil) > >> >>>>> 3498.723559462:debug_relp queue:Reg/w0: wti 0x23c78b0, action 2, > ptr > >> >>>>> (nil) > >> >>>>> 3498.723794716:debug_relp queue:Reg/w0: wti 0x23c78b0, action 3, > ptr > >> >>>>> (nil) > >> >>>>> 3498.723972736:debug_relp queue:Reg/w0: wti 0x23c78b0, action 4, > ptr > >> >>>>> (nil) > >> >>>>> 3498.724207978:debug_relp queue:Reg/w0: wti 0x23c78b0, action 5, > ptr > >> >>>>> (nil) > >> >>>>> 3498.724215660:debug_relp queue:Reg/w0: wti 0x23c78b0, action 6, > ptr > >> >>>>> (nil) > >> >>>>> 3498.724219766:debug_relp queue:Reg/w0: wti 0x23c78b0, action 7, > ptr > >> >>>>> (nil) > >> >>>>> 3498.724223775:debug_relp queue:Reg/w0: wti 0x23c78b0: worker > exiting > >> >>>>> 3498.724228430:debug_relp queue:Reg/w0: debug_relp queue:Reg: > Worker > >> >>>>> thread > >> >>>>> 23c78b0, terminated, num workers now 0 > >> >>>>> 3498.724243673:debug_relp queue:Reg/w0: destructor for debug call > >> stack > >> >>>>> 0x7f1ef00008c0 called > >> >>>>> 3498.724371931:main thread : debug_relp queue: checking to see > if > >> we > >> >>>>> need to cancel any worker threads of the primary queue > >> >>>>> 3498.724379712:main thread : debug_relp queue: worker threads > >> >>>>> terminated, remaining queue size log 0, phys 0. > >> >>>>> 3498.724386288:main thread : debug_relp queue: persisting > queue to > >> >>>>> disk, > >> >>>>> 0 entries... > >> >>>>> 3498.724391611:main thread : strm 0x23b74c0: file 5(relp_queue) > >> >>>>> closing > >> >>>>> 3498.724396110:main thread : strm 0x23b74c0: file 5(relp_queue) > >> >>>>> flush, > >> >>>>> buflen 0 (no need to flush) > >> >>>>> 3498.724411574:main thread : strm 0x23b7760: file 6(relp_queue) > >> >>>>> closing > >> >>>>> 3498.724418797:main thread : strm 0x23b7a00: file > -1(relp_queue) > >> >>>>> closing > >> >>>>> 3498.724423956:main thread : strmCloseFile: deleting > >> >>>>> '/var/spool/rsyslog/relp_queue.00000001' > >> >>>>> 3498.724490018:main thread : strm 0x7f1ef4002480: file > 4(messages) > >> >>>>> closing > >> >>>>> 3498.724607874:main thread : strm 0x7f1ef4002480: file > 4(messages) > >> >>>>> flush, buflen 0 (no need to flush) > >> >>>>> 3498.724731690:main thread : all primary multi-thread sources > have > >> >>>>> been > >> >>>>> terminated - now doing aux cleanup... > >> >>>>> 3498.724737248:main thread : destructing current config... > >> >>>>> 3498.724741760:main thread : calling freeCnf(0x23a91e0) for > module > >> >>>>> 'builtin:omfile' > >> >>>>> 3498.724745773:main thread : calling freeCnf(0x23a93a0) for > module > >> >>>>> 'builtin:ompipe' > >> >>>>> 3498.724749977:main thread : calling freeCnf(0x23aa3a0) for > module > >> >>>>> 'builtin:omfwd' > >> >>>>> 3498.724754588:main thread : calling freeCnf(0x23b9b00) for > module > >> >>>>> 'imuxsock' > >> >>>>> 3498.724758956:main thread : calling freeCnf(0x23bac70) for > module > >> >>>>> 'imklog' > >> >>>>> 3498.724807014:main thread : destructing parser > 'rsyslog.rfc5424' > >> >>>>> 3498.724871141:main thread : destructing parser > 'rsyslog.rfc3164' > >> >>>>> 3498.724876440:main thread : pmrfc3164: free parser instance > >> >>>>> 0x23aa8b0 > >> >>>>> 3498.724887558:main thread : file rsyslogd.c released module > >> >>>>> 'lmnet', > >> >>>>> reference count now 2 > >> >>>>> 3498.724892004:main thread : module lmnet NOT unloaded because > it > >> >>>>> still > >> >>>>> has a refcount of 2 > >> >>>>> 3498.724895745:main thread : Unloading module builtin:omfile > >> >>>>> 3498.724900191:main thread : module lmnet NOT unloaded because > it > >> >>>>> still > >> >>>>> has a refcount of 2 > >> >>>>> 3498.724904194:main thread : Unloading module builtin:ompipe > >> >>>>> 3498.724908035:main thread : module lmnet NOT unloaded because > it > >> >>>>> still > >> >>>>> has a refcount of 2 > >> >>>>> 3498.724911707:main thread : Unloading module builtin-shell > >> >>>>> 3498.725044125:main thread : module lmnet NOT unloaded because > it > >> >>>>> still > >> >>>>> has a refcount of 2 > >> >>>>> 3498.725050401:main thread : Unloading module builtin:omdiscard > >> >>>>> 3498.725054843:main thread : module lmnet NOT unloaded because > it > >> >>>>> still > >> >>>>> has a refcount of 2 > >> >>>>> 3498.725058772:main thread : Unloading module builtin:omfwd > >> >>>>> 3498.725064464:main thread : file omfwd.c released module > 'lmnet', > >> >>>>> reference count now 1 > >> >>>>> 3498.725068747:main thread : module lmnet NOT unloaded because > it > >> >>>>> still > >> >>>>> has a refcount of 1 > >> >>>>> 3498.725072460:main thread : Unloading module builtin:omusrmsg > >> >>>>> 3498.725076324:main thread : module lmnet NOT unloaded because > it > >> >>>>> still > >> >>>>> has a refcount of 1 > >> >>>>> 3498.725080023:main thread : Unloading module builtin:pmrfc5424 > >> >>>>> 3498.725084039:main thread : module lmnet NOT unloaded because > it > >> >>>>> still > >> >>>>> has a refcount of 1 > >> >>>>> 3498.725087646:main thread : Unloading module builtin:pmrfc3164 > >> >>>>> 3498.725091651:main thread : module lmnet NOT unloaded because > it > >> >>>>> still > >> >>>>> has a refcount of 1 > >> >>>>> 3498.725095255:main thread : Unloading module builtin:smfile > >> >>>>> 3498.725099010:main thread : module lmnet NOT unloaded because > it > >> >>>>> still > >> >>>>> has a refcount of 1 > >> >>>>> 3498.725102601:main thread : Unloading module > builtin:smtradfile > >> >>>>> 3498.725106419:main thread : module lmnet NOT unloaded because > it > >> >>>>> still > >> >>>>> has a refcount of 1 > >> >>>>> 3498.725110003:main thread : Unloading module builtin:smfwd > >> >>>>> 3498.725113689:main thread : module lmnet NOT unloaded because > it > >> >>>>> still > >> >>>>> has a refcount of 1 > >> >>>>> 3498.725117357:main thread : Unloading module builtin:smtradfwd > >> >>>>> 3498.725121132:main thread : module lmnet NOT unloaded because > it > >> >>>>> still > >> >>>>> has a refcount of 1 > >> >>>>> 3498.725124782:main thread : Unloading module imuxsock > >> >>>>> 3498.725157459:main thread : module lmnet NOT unloaded because > it > >> >>>>> still > >> >>>>> has a refcount of 1 > >> >>>>> 3498.725162797:main thread : Unloading module imklog > >> >>>>> 3498.725179208:main thread : module lmnet NOT unloaded because > it > >> >>>>> still > >> >>>>> has a refcount of 1 > >> >>>>> 3498.726112994:main thread : Unloading module omrelp > >> >>>>> 3498.726184405:main thread : module lmnet NOT unloaded because > it > >> >>>>> still > >> >>>>> has a refcount of 1 > >> >>>>> 3498.726189821:main thread : module lmregexp NOT unloaded > because > >> it > >> >>>>> still has a refcount of 1 > >> >>>>> 3498.726195161:main thread : file conf.c released module > 'lmnet', > >> >>>>> reference count now 0 > >> >>>>> 3498.726198620:main thread : module 'lmnet' has zero reference > >> >>>>> count, > >> >>>>> unloading... > >> >>>>> 3498.726201936:main thread : Unloading module lmnet > >> >>>>> 3498.726218655:main thread : rsyslog runtime de-initialized, > >> current > >> >>>>> users 0 > >> >>>>> 3498.726222660:main thread : Clean shutdown completed, bye > >> >>>>> > >> >>>>> > >> >>>>> On Mon, Mar 28, 2016 at 4:41 PM, Kane Kim <[email protected]> > >> >>>>> wrote: > >> >>>>> > >> >>>>> Hello Reiner, > >> >>>>>> > >> >>>>>> > >> >>>>>> I'm seeing the same behavior with omrelp module. I've configured > >> >>>>>> omrelp > >> >>>>>> to > >> >>>>>> non-routable ip address: > >> >>>>>> module(load="omrelp") > >> >>>>>> local0.* action(type="omrelp" > >> >>>>>> target="10.0.0.1" > >> >>>>>> port="2514" > >> >>>>>> name="debug_relp" > >> >>>>>> queue.filename="relp_queue" > >> >>>>>> queue.maxdiskspace="1G" > >> >>>>>> queue.saveonshutdown="on" > >> >>>>>> queue.spoolDirectory="/var/spool/rsyslog" > >> >>>>>> queue.type="Disk" > >> >>>>>> ) > >> >>>>>> Then I try to log a line: > >> >>>>>> [root@314959ac8cb0 prism]# logger -p local0.info test > >> >>>>>> [root@314959ac8cb0 prism]# ls /var/spool/rsyslog/ > >> >>>>>> relp_queue.00000001 > >> >>>>>> # I shutdown rsyslog at this point > >> >>>>>> [root@314959ac8cb0 prism]# ls /var/spool/rsyslog/ > >> >>>>>> # after shutdown queue is deleted and message is lost > >> >>>>>> [root@314959ac8cb0 prism]# > >> >>>>>> > >> >>>>>> Here is rsyslog log, so omrelp starts transaction and then I > >> shutdown > >> >>>>>> rsyslog, it seems at some point queue is deleted: > >> >>>>>> > >> >>>>>> 8184.798028612:debug_relp queue:Reg/w0: omrelp: beginTransaction > >> >>>>>> ^C8194.245122305:main thread : DoDie called. > >> >>>>>> 8194.245301525:main thread : exiting on signal 2 > >> >>>>>> 8194.245532709:main thread : Terminating input threads... > >> >>>>>> 8194.245873683:main thread : request term via SIGTTIN for > input > >> >>>>>> thread > >> >>>>>> 'imuxsock' 0xcb767700 > >> >>>>>> 8194.246079748:imuxsock.c : thrdStarter: usrThrdMain > imuxsock - > >> >>>>>> 0x7f9bcb767700 returned with iRet 0, exiting now. > >> >>>>>> 8194.247556671:main thread : non-cancel input thread > termination > >> >>>>>> succeeded for thread imuxsock 0xcb767700 > >> >>>>>> 8194.249357547:imuxsock.c : destructor for debug call stack > >> >>>>>> 0x7f9bbc0008c0 called > >> >>>>>> 8194.249784271:main thread : imuxsock: unlinking unix socket > >> >>>>>> file[0] > >> >>>>>> /dev/log > >> >>>>>> 8194.250043368:main thread : main Q: qqueueAdd: entry added, > size > >> >>>>>> now > >> >>>>>> log 1, phys 1 entries > >> >>>>>> 8194.250321193:main thread : main Q: EnqueueMsg advised worker > >> >>>>>> start > >> >>>>>> 8194.250493505:main Q:Reg/w0 : wti 0x1c13bb0: worker awoke from > >> idle > >> >>>>>> processing > >> >>>>>> 8194.250735988:main Q:Reg/w0 : DeleteProcessedBatch: we deleted > 0 > >> >>>>>> objects > >> >>>>>> and enqueued 0 objects > >> >>>>>> 8194.250959294:main Q:Reg/w0 : doDeleteBatch: delete batch from > >> >>>>>> store, > >> >>>>>> new sizes: log 1, phys 1 > >> >>>>>> 8194.251913535:main Q:Reg/w0 : processBATCH: batch of 1 elements > >> must > >> >>>>>> be > >> >>>>>> processed > >> >>>>>> 8194.252193683:main Q:Reg/w0 : processBATCH: next msg 0: > [origin > >> >>>>>> software="rsyslogd" swVersion="8.18.0.master" x-pid="30298" > x-info=" > >> >>>>>> http://www.rsyslog.com"] exiting on signal 2. > >> >>>>>> 8194.252377044:main Q:Reg/w0 : PRIFILT 'local0.*' > >> >>>>>> 8194.252669683:main Q:Reg/w0 : pmask: X X X X X X X > X > >> X > >> >>>>>> X > >> >>>>>> X X X X X X FF X X X X X X X X X > >> >>>>>> 8194.252898063:main Q:Reg/w0 : PRIFILT condition result is 0 > >> >>>>>> 8194.252902083:main Q:Reg/w0 : PRIFILT > >> >>>>>> '*.info;mail.none;authpriv.none;cron.none' > >> >>>>>> 8194.252908847:main Q:Reg/w0 : pmask: 7F 7F X 7F 7F 7F 7F > 7F > >> 7F > >> >>>>>> X > >> >>>>>> X 7F 7F 7F 7F 7F 7F 7F 7F 7F 7F 7F 7F 7F 7F 7F > >> >>>>>> 8194.252956209:main Q:Reg/w0 : PRIFILT condition result is 1 > >> >>>>>> 8194.252959849:main Q:Reg/w0 : ACTION 1 > >> >>>>>> [builtin:omfile:/var/log/messages] > >> >>>>>> 8194.252967510:main Q:Reg/w0 : executing action 1 > >> >>>>>> 8194.252971541:main Q:Reg/w0 : action 'action 1': called, > logging > >> to > >> >>>>>> builtin:omfile (susp 0/0, direct q 1) > >> >>>>>> 8194.252977583:main Q:Reg/w0 : action 'action 1': is > transactional > >> - > >> >>>>>> executing in commit phase > >> >>>>>> 8194.252982638:main Q:Reg/w0 : Action 1 transitioned to state: > itx > >> >>>>>> 8194.252986300:main Q:Reg/w0 : action 'action 1': set suspended > >> state > >> >>>>>> to 0 > >> >>>>>> 8194.252989729:main Q:Reg/w0 : PRIFILT 'authpriv.*' > >> >>>>>> 8194.252996951:main Q:Reg/w0 : pmask: X X X X X X X > X > >> X > >> >>>>>> X > >> >>>>>> FF X X X X X X X X X X X X X X X > >> >>>>>> 8194.253051858:main Q:Reg/w0 : PRIFILT condition result is 0 > >> >>>>>> 8194.253056228:main Q:Reg/w0 : PRIFILT 'mail.*' > >> >>>>>> 8194.253065299:main Q:Reg/w0 : pmask: X X FF X X X X > X > >> X > >> >>>>>> X > >> >>>>>> X X X X X X X X X X X X X X X X > >> >>>>>> 8194.253407472:main Q:Reg/w0 : PRIFILT condition result is 0 > >> >>>>>> 8194.253412679:main Q:Reg/w0 : PRIFILT 'cron.*' > >> >>>>>> 8194.253422055:main Q:Reg/w0 : pmask: X X X X X X X > X > >> X > >> >>>>>> FF > >> >>>>>> X X X X X X X X X X X X X X X X > >> >>>>>> 8194.253493551:main Q:Reg/w0 : PRIFILT condition result is 0 > >> >>>>>> 8194.253498038:main Q:Reg/w0 : PRIFILT '*.emerg' > >> >>>>>> 8194.253507065:main Q:Reg/w0 : pmask: 1 1 1 1 1 1 1 > 1 > >> 1 > >> >>>>>> 1 > >> >>>>>> 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 > >> >>>>>> 8194.253580307:main Q:Reg/w0 : PRIFILT condition result is 0 > >> >>>>>> 8194.253584400:main Q:Reg/w0 : PRIFILT 'uucp,news.crit' > >> >>>>>> 8194.253593442:main Q:Reg/w0 : pmask: X X X X X X X > 7 > >> 7 > >> >>>>>> X > >> >>>>>> X X X X X X X X X X X X X X X X > >> >>>>>> 8194.253664984:main Q:Reg/w0 : PRIFILT condition result is 0 > >> >>>>>> 8194.253669184:main Q:Reg/w0 : PRIFILT 'local7.*' > >> >>>>>> 8194.253678130:main Q:Reg/w0 : pmask: X X X X X X X > X > >> X > >> >>>>>> X > >> >>>>>> X X X X X X X X X X X X X FF X X > >> >>>>>> 8194.254445021:main Q:Reg/w0 : PRIFILT condition result is 0 > >> >>>>>> 8194.254450651:main Q:Reg/w0 : END batch execution phase, > entering > >> to > >> >>>>>> commit phase [processed 1 of 1 messages] > >> >>>>>> 8194.254456257:main Q:Reg/w0 : actionCommitAll: action 1, state > 1, > >> >>>>>> nbr > >> >>>>>> to > >> >>>>>> commit 0 isTransactional 1 > >> >>>>>> 8194.254461646:main Q:Reg/w0 : doTransaction: have > >> commitTransaction > >> >>>>>> IF, > >> >>>>>> using that, pWrkrInfo 0x1c13cf0 > >> >>>>>> 8194.254466238:main Q:Reg/w0 : entering > >> >>>>>> actionCallCommitTransaction(), > >> >>>>>> state: itx, actionNbr 1, nMsgs 1 > >> >>>>>> 8194.254471876:main Q:Reg/w0 : omfile: write to stream, > >> pData->pStrm > >> >>>>>> 0x7f9bc4002480, lenBuf 161, strt data Mar 28 23:36:34 > 314959ac8cb0 > >> >>>>>> rsyslogd: [origin software="rsyslogd" swVersion="8.18.0.master" > >> >>>>>> x-pid="30298" x-info="http://www.rs > >> >>>>>> 8194.254478237:main Q:Reg/w0 : strm 0x7f9bc4002480: file > >> 4(messages) > >> >>>>>> flush, buflen 161 > >> >>>>>> 8194.254483210:main Q:Reg/w0 : strmPhysWrite, stream > >> 0x7f9bc4002480, > >> >>>>>> len > >> >>>>>> 161 > >> >>>>>> 8194.254500329:main Q:Reg/w0 : strm 0x7f9bc4002480: file 4 write > >> >>>>>> wrote > >> >>>>>> 161 bytes > >> >>>>>> 8194.254506254:main Q:Reg/w0 : Action 1 transitioned to state: > rdy > >> >>>>>> 8194.254511506:main Q:Reg/w0 : Action 1 transitioned to state: > itx > >> >>>>>> 8194.254515961:main Q:Reg/w0 : Action 1 transitioned to state: > rdy > >> >>>>>> 8194.254520243:main Q:Reg/w0 : actionCommit, in retry loop, > iRet 0 > >> >>>>>> 8194.254524679:main Q:Reg/w0 : processBATCH: batch of 1 elements > >> has > >> >>>>>> been > >> >>>>>> processed > >> >>>>>> 8194.254529731:main Q:Reg/w0 : regular consumer finished, > iret=0, > >> >>>>>> szlog 0 > >> >>>>>> sz phys 1 > >> >>>>>> 8194.254536104:main Q:Reg/w0 : DeleteProcessedBatch: we deleted > 1 > >> >>>>>> objects > >> >>>>>> and enqueued 0 objects > >> >>>>>> 8194.254540861:main Q:Reg/w0 : doDeleteBatch: delete batch from > >> >>>>>> store, > >> >>>>>> new sizes: log 0, phys 0 > >> >>>>>> 8194.254545792:main Q:Reg/w0 : regular consumer finished, > iret=4, > >> >>>>>> szlog 0 > >> >>>>>> sz phys 0 > >> >>>>>> 8194.254550314:main Q:Reg/w0 : main Q:Reg/w0: worker IDLE, > waiting > >> >>>>>> for > >> >>>>>> work. > >> >>>>>> 8194.254556418:main thread : Terminating main queue... > >> >>>>>> 8194.254560776:main thread : main Q: initiating worker thread > >> >>>>>> shutdown > >> >>>>>> sequence > >> >>>>>> 8194.254564653:main thread : main Q: trying shutdown of > regular > >> >>>>>> workers > >> >>>>>> 8194.254572961:main thread : sent SIGTTIN to worker thread > >> >>>>>> 0xcaf66700 > >> >>>>>> 8194.254577260:main thread : main Q:Reg: waiting 1499ms on > worker > >> >>>>>> thread termination, 1 still running > >> >>>>>> 8194.254585507:main Q:Reg/w0 : wti 0x1c13bb0: worker awoke from > >> idle > >> >>>>>> processing > >> >>>>>> 8194.254589465:main Q:Reg/w0 : DeleteProcessedBatch: we deleted > 0 > >> >>>>>> objects > >> >>>>>> and enqueued 0 objects > >> >>>>>> 8194.254593070:main Q:Reg/w0 : doDeleteBatch: delete batch from > >> >>>>>> store, > >> >>>>>> new sizes: log 0, phys 0 > >> >>>>>> 8194.254596813:main Q:Reg/w0 : regular consumer finished, > iret=4, > >> >>>>>> szlog 0 > >> >>>>>> sz phys 0 > >> >>>>>> 8194.254600412:main Q:Reg/w0 : wti 0x1c13bb0: terminating worker > >> >>>>>> terminateRet=5, bInactivityTOOccured=0 > >> >>>>>> 8194.254604469:main Q:Reg/w0 : DDDD: wti 0x1c13bb0: worker > cleanup > >> >>>>>> action > >> >>>>>> instances > >> >>>>>> 8194.254608268:main Q:Reg/w0 : wti 0x1c13bb0, action 0, ptr > (nil) > >> >>>>>> 8194.254611752:main Q:Reg/w0 : wti 0x1c13bb0, action 1, ptr > >> >>>>>> 0x7f9bc40022a0 > >> >>>>>> 8194.254616114:main Q:Reg/w0 : wti 0x1c13bb0, action 2, ptr > (nil) > >> >>>>>> 8194.254619450:main Q:Reg/w0 : wti 0x1c13bb0, action 3, ptr > (nil) > >> >>>>>> 8194.254622767:main Q:Reg/w0 : wti 0x1c13bb0, action 4, ptr > (nil) > >> >>>>>> 8194.254626075:main Q:Reg/w0 : wti 0x1c13bb0, action 5, ptr > (nil) > >> >>>>>> 8194.254629392:main Q:Reg/w0 : wti 0x1c13bb0, action 6, ptr > (nil) > >> >>>>>> 8194.254632707:main Q:Reg/w0 : wti 0x1c13bb0, action 7, ptr > (nil) > >> >>>>>> 8194.254636124:main Q:Reg/w0 : wti 0x1c13bb0: worker exiting > >> >>>>>> 8194.254639920:main Q:Reg/w0 : main Q:Reg: Worker thread > 1c13bb0, > >> >>>>>> terminated, num workers now 0 > >> >>>>>> 8194.254656279:main thread : main Q: regular queue workers > shut > >> >>>>>> down. > >> >>>>>> 8194.254656402:main thread : main Q: checking to see if we > need > >> to > >> >>>>>> cancel any worker threads of the primary queue > >> >>>>>> 8194.254656524:main thread : main Q: worker threads > terminated, > >> >>>>>> remaining queue size log 0, phys 0. > >> >>>>>> 8194.254656647:main thread : main Q: queue (type 0) will lose > 0 > >> >>>>>> messages, destroying... > >> >>>>>> 8194.255021761:main Q:Reg/w0 : destructor for debug call stack > >> >>>>>> 0x7f9bc40008c0 called > >> >>>>>> 8194.255069521:main thread : Terminating outputs... > >> >>>>>> 8194.255084818:main thread : destructing ruleset 0x1bf6150, > name > >> >>>>>> 0x1bf6190 > >> >>>>>> 8194.255089525:main thread : debug_relp queue: initiating > worker > >> >>>>>> thread > >> >>>>>> shutdown sequence > >> >>>>>> 8194.255093230:main thread : debug_relp queue: trying > shutdown of > >> >>>>>> regular workers > >> >>>>>> 8194.255099210:main thread : sent SIGTTIN to worker thread > >> >>>>>> 0xca765700 > >> >>>>>> 8194.255119598:main thread : debug_relp queue:Reg: waiting > 0ms on > >> >>>>>> worker thread termination, 1 still running > >> >>>>>> 8194.255126357:debug_relp queue:Reg/w0: end relpSessConnect, iRet > >> >>>>>> 10014 > >> >>>>>> 8194.255132022:debug_relp queue:Reg/w0: Action 0 transitioned to > >> >>>>>> state: > >> >>>>>> rtry > >> >>>>>> 8194.255135551:debug_relp queue:Reg/w0: doTransaction: action 0, > >> >>>>>> currIParam 1 > >> >>>>>> 8194.255139603:debug_relp queue:Reg/w0: actionDoRetry: debug_relp > >> >>>>>> enter > >> >>>>>> loop, iRetries=0 > >> >>>>>> 8194.255210939:main thread : debug_relp queue:Reg: timeout > >> waiting > >> >>>>>> on > >> >>>>>> worker thread termination > >> >>>>>> 8194.255211086:main thread : sent SIGTTIN to worker thread > >> >>>>>> 0xca765700 > >> >>>>>> 8194.255211257:main thread : debug_relp queue: regular > shutdown > >> >>>>>> timed > >> >>>>>> out on primary queue (this is OK) > >> >>>>>> 8194.255211429:main thread : debug_relp queue: trying to > shutdown > >> >>>>>> workers within Action Timeoutsetting EnqOnly mode > >> >>>>>> 8194.255215492:main thread : debug_relp queue: trying > immediate > >> >>>>>> shutdown of regular workers (if any) > >> >>>>>> 8194.255221059:main thread : sent SIGTTIN to worker thread > >> >>>>>> 0xca765700 > >> >>>>>> 8194.255226792:main thread : debug_relp queue:Reg: waiting > 1000ms > >> >>>>>> on > >> >>>>>> worker thread termination, 1 still running > >> >>>>>> 8194.255232713:debug_relp queue:Reg/w0: end relpSessConnect, iRet > >> >>>>>> 10014 > >> >>>>>> 8194.255237080:debug_relp queue:Reg/w0: actionDoRetry: debug_relp > >> >>>>>> action->tryResume returned -2007 > >> >>>>>> 8194.255240684:debug_relp queue:Reg/w0: actionDoRetry: debug_relp > >> >>>>>> check > >> >>>>>> for max retries, iResumeRetryCount 0, iRetries 0 > >> >>>>>> 8194.255244459:debug_relp queue:Reg/w0: Action 0 transitioned to > >> >>>>>> state: > >> >>>>>> susp > >> >>>>>> 8194.255257975:debug_relp queue:Reg/w0: Called LogMsg, msg: > action > >> >>>>>> 'debug_relp' suspended, next retry is Mon Mar 28 23:37:04 2016 > >> >>>>>> 8194.255266857:debug_relp queue:Reg/w0: submitMsg2() could not > >> submit > >> >>>>>> message - queue does (no longer?) exist - ignored > >> >>>>>> 8194.255271065:debug_relp queue:Reg/w0: action 'debug_relp' > >> suspended, > >> >>>>>> earliest retry=1459208224 (now 1459208194), iNbrResRtry 0, > duration > >> 30 > >> >>>>>> 8194.255275244:debug_relp queue:Reg/w0: actionTryResume: action > >> >>>>>> 0x1c09af0 > >> >>>>>> state: susp, next retry (if applicable): 1459208224 [now > 1459208194] > >> >>>>>> 8194.255278825:debug_relp queue:Reg/w0: actionCommit, in retry > loop, > >> >>>>>> iRet > >> >>>>>> -2123 > >> >>>>>> 8194.255282409:debug_relp queue:Reg/w0: regular consumer > finished, > >> >>>>>> iret=-2123, szlog 0 sz phys 1 > >> >>>>>> 8194.255287286:debug_relp queue:Reg/w0: DeleteProcessedBatch: we > >> >>>>>> deleted 1 > >> >>>>>> objects and enqueued 0 objects > >> >>>>>> 8194.255290952:debug_relp queue:Reg/w0: doDeleteBatch: delete > batch > >> >>>>>> from > >> >>>>>> store, new sizes: log 0, phys 0 > >> >>>>>> 8194.255295204:debug_relp queue:Reg/w0: wti 0x1c12880: > terminating > >> >>>>>> worker > >> >>>>>> because of TERMINATE_NOW mode, del iRet 0 > >> >>>>>> 8194.255298822:debug_relp queue:Reg/w0: DDDD: wti 0x1c12880: > worker > >> >>>>>> cleanup action instances > >> >>>>>> 8194.255302306:debug_relp queue:Reg/w0: wti 0x1c12880, action 0, > ptr > >> >>>>>> 0x7f9bc00020d0 > >> >>>>>> 8194.255305929:debug_relp queue:Reg/w0: relp engine destructing > >> client > >> >>>>>> 0x7f9bc00027e0 > >> >>>>>> 8194.255309843:debug_relp queue:Reg/w0: wti 0x1c12880, action 1, > ptr > >> >>>>>> (nil) > >> >>>>>> 8194.255313226:debug_relp queue:Reg/w0: wti 0x1c12880, action 2, > ptr > >> >>>>>> (nil) > >> >>>>>> 8194.255316479:debug_relp queue:Reg/w0: wti 0x1c12880, action 3, > ptr > >> >>>>>> (nil) > >> >>>>>> 8194.255319704:debug_relp queue:Reg/w0: wti 0x1c12880, action 4, > ptr > >> >>>>>> (nil) > >> >>>>>> 8194.255322934:debug_relp queue:Reg/w0: wti 0x1c12880, action 5, > ptr > >> >>>>>> (nil) > >> >>>>>> 8194.255326175:debug_relp queue:Reg/w0: wti 0x1c12880, action 6, > ptr > >> >>>>>> (nil) > >> >>>>>> 8194.255329409:debug_relp queue:Reg/w0: wti 0x1c12880, action 7, > ptr > >> >>>>>> (nil) > >> >>>>>> 8194.255332655:debug_relp queue:Reg/w0: wti 0x1c12880: worker > >> exiting > >> >>>>>> 8194.255336231:debug_relp queue:Reg/w0: debug_relp queue:Reg: > Worker > >> >>>>>> thread 1c12880, terminated, num workers now 0 > >> >>>>>> 8194.255344540:debug_relp queue:Reg/w0: destructor for debug call > >> >>>>>> stack > >> >>>>>> 0x7f9bc00008c0 called > >> >>>>>> 8194.255371534:main thread : debug_relp queue: checking to > see if > >> >>>>>> we > >> >>>>>> need to cancel any worker threads of the primary queue > >> >>>>>> 8194.255376798:main thread : debug_relp queue: worker threads > >> >>>>>> terminated, remaining queue size log 0, phys 0. > >> >>>>>> 8194.255381256:main thread : debug_relp queue: persisting > queue > >> to > >> >>>>>> disk, 0 entries... > >> >>>>>> 8194.255385525:main thread : strm 0x1c024c0: file > 5(relp_queue) > >> >>>>>> closing > >> >>>>>> 8194.255389222:main thread : strm 0x1c024c0: file > 5(relp_queue) > >> >>>>>> flush, > >> >>>>>> buflen 0 (no need to flush) > >> >>>>>> 8194.255398716:main thread : strm 0x1c02760: file > 6(relp_queue) > >> >>>>>> closing > >> >>>>>> 8194.255404785:main thread : strm 0x1c02a00: file > -1(relp_queue) > >> >>>>>> closing > >> >>>>>> 8194.255409145:main thread : strmCloseFile: deleting > >> >>>>>> '/var/spool/rsyslog/relp_queue.00000001' > >> >>>>>> 8194.255460812:main thread : strm 0x7f9bc4002480: file > >> 4(messages) > >> >>>>>> closing > >> >>>>>> 8194.255466309:main thread : strm 0x7f9bc4002480: file > >> 4(messages) > >> >>>>>> flush, buflen 0 (no need to flush) > >> >>>>>> 8194.255479325:main thread : all primary multi-thread sources > >> have > >> >>>>>> been > >> >>>>>> terminated - now doing aux cleanup... > >> >>>>>> 8194.255482809:main thread : destructing current config... > >> >>>>>> 8194.255486662:main thread : calling freeCnf(0x1bf41e0) for > >> module > >> >>>>>> 'builtin:omfile' > >> >>>>>> 8194.255490098:main thread : calling freeCnf(0x1bf43a0) for > >> module > >> >>>>>> 'builtin:ompipe' > >> >>>>>> 8194.255493809:main thread : calling freeCnf(0x1bf53a0) for > >> module > >> >>>>>> 'builtin:omfwd' > >> >>>>>> 8194.255497793:main thread : calling freeCnf(0x1c04b00) for > >> module > >> >>>>>> 'imuxsock' > >> >>>>>> 8194.255501565:main thread : calling freeCnf(0x1c05c70) for > >> module > >> >>>>>> 'imklog' > >> >>>>>> 8194.255543579:main thread : destructing parser > 'rsyslog.rfc5424' > >> >>>>>> 8194.255547468:main thread : destructing parser > 'rsyslog.rfc3164' > >> >>>>>> 8194.255550784:main thread : pmrfc3164: free parser instance > >> >>>>>> 0x1bf58b0 > >> >>>>>> 8194.255559659:main thread : file rsyslogd.c released module > >> >>>>>> 'lmnet', > >> >>>>>> reference count now 2 > >> >>>>>> 8194.255563470:main thread : module lmnet NOT unloaded > because it > >> >>>>>> still > >> >>>>>> has a refcount of 2 > >> >>>>>> 8194.255566676:main thread : Unloading module builtin:omfile > >> >>>>>> 8194.255570420:main thread : module lmnet NOT unloaded > because it > >> >>>>>> still > >> >>>>>> has a refcount of 2 > >> >>>>>> 8194.255573674:main thread : Unloading module builtin:ompipe > >> >>>>>> 8194.255576857:main thread : module lmnet NOT unloaded > because it > >> >>>>>> still > >> >>>>>> has a refcount of 2 > >> >>>>>> 8194.255579938:main thread : Unloading module builtin-shell > >> >>>>>> 8194.255583060:main thread : module lmnet NOT unloaded > because it > >> >>>>>> still > >> >>>>>> has a refcount of 2 > >> >>>>>> 8194.255586144:main thread : Unloading module > builtin:omdiscard > >> >>>>>> 8194.255589509:main thread : module lmnet NOT unloaded > because it > >> >>>>>> still > >> >>>>>> has a refcount of 2 > >> >>>>>> 8194.255592658:main thread : Unloading module builtin:omfwd > >> >>>>>> 8194.255596441:main thread : file omfwd.c released module > >> 'lmnet', > >> >>>>>> reference count now 1 > >> >>>>>> 8194.255599981:main thread : module lmnet NOT unloaded > because it > >> >>>>>> still > >> >>>>>> has a refcount of 1 > >> >>>>>> 8194.255603068:main thread : Unloading module builtin:omusrmsg > >> >>>>>> 8194.255606240:main thread : module lmnet NOT unloaded > because it > >> >>>>>> still > >> >>>>>> has a refcount of 1 > >> >>>>>> 8194.255609287:main thread : Unloading module > builtin:pmrfc5424 > >> >>>>>> 8194.255612755:main thread : module lmnet NOT unloaded > because it > >> >>>>>> still > >> >>>>>> has a refcount of 1 > >> >>>>>> 8194.255615826:main thread : Unloading module > builtin:pmrfc3164 > >> >>>>>> 8194.255619252:main thread : module lmnet NOT unloaded > because it > >> >>>>>> still > >> >>>>>> has a refcount of 1 > >> >>>>>> 8194.255622315:main thread : Unloading module builtin:smfile > >> >>>>>> 8194.255625406:main thread : module lmnet NOT unloaded > because it > >> >>>>>> still > >> >>>>>> has a refcount of 1 > >> >>>>>> 8194.255628486:main thread : Unloading module > builtin:smtradfile > >> >>>>>> 8194.255631753:main thread : module lmnet NOT unloaded > because it > >> >>>>>> still > >> >>>>>> has a refcount of 1 > >> >>>>>> 8194.255634819:main thread : Unloading module builtin:smfwd > >> >>>>>> 8194.255637916:main thread : module lmnet NOT unloaded > because it > >> >>>>>> still > >> >>>>>> has a refcount of 1 > >> >>>>>> 8194.255640987:main thread : Unloading module > builtin:smtradfwd > >> >>>>>> 8194.255644187:main thread : module lmnet NOT unloaded > because it > >> >>>>>> still > >> >>>>>> has a refcount of 1 > >> >>>>>> 8194.255647267:main thread : Unloading module imuxsock > >> >>>>>> 8194.255676232:main thread : module lmnet NOT unloaded > because it > >> >>>>>> still > >> >>>>>> has a refcount of 1 > >> >>>>>> 8194.255680751:main thread : Unloading module imklog > >> >>>>>> 8194.255694428:main thread : module lmnet NOT unloaded > because it > >> >>>>>> still > >> >>>>>> has a refcount of 1 > >> >>>>>> 8194.255698085:main thread : Unloading module omrelp > >> >>>>>> 8194.255750953:main thread : module lmnet NOT unloaded > because it > >> >>>>>> still > >> >>>>>> has a refcount of 1 > >> >>>>>> 8194.255755236:main thread : module lmregexp NOT unloaded > because > >> >>>>>> it > >> >>>>>> still has a refcount of 1 > >> >>>>>> 8194.255759705:main thread : file conf.c released module > 'lmnet', > >> >>>>>> reference count now 0 > >> >>>>>> 8194.255762875:main thread : module 'lmnet' has zero reference > >> >>>>>> count, > >> >>>>>> unloading... > >> >>>>>> 8194.255765901:main thread : Unloading module lmnet > >> >>>>>> 8194.255780970:main thread : rsyslog runtime de-initialized, > >> >>>>>> current > >> >>>>>> users 0 > >> >>>>>> 8194.255784704:main thread : Clean shutdown completed, bye > >> >>>>>> > >> >>>>>> > >> >>>>>> > >> >>>>>> On Sun, Mar 13, 2016 at 9:30 PM, David Lang <[email protected]> > wrote: > >> >>>>>> > >> >>>>>> On Sun, 13 Mar 2016, Kane Kim wrote: > >> >>>>>>> > >> >>>>>>> > >> >>>>>>> +1to David, having module handle the errors is not 100% > reliable, > >> >>>>>>> what > >> >>>>>>> if > >> >>>>>>> > >> >>>>>>>> module dies before having a chance to save error log, those > >> messages > >> >>>>>>>> will > >> >>>>>>>> be lost. > >> >>>>>>>> > >> >>>>>>>> > >> >>>>>>> well, the module is a thread of rsyslog, so if it dies, rsyslog > has > >> >>>>>>> died. > >> >>>>>>> > >> >>>>>>> But if you are doing large batches (hundreds or thousands), > putting > >> >>>>>>> all > >> >>>>>>> of them in a 'bad messages' file because one has a problem seems > >> like > >> >>>>>>> the > >> >>>>>>> wrong thing to do. > >> >>>>>>> > >> >>>>>>> given that the error is not detected until the batch is being > >> flushed > >> >>>>>>> for > >> >>>>>>> the output, any outputs that have already flushed their batches > >> can't > >> >>>>>>> be > >> >>>>>>> changed, so the question is if it's worse to have messages > resent > >> to > >> >>>>>>> those > >> >>>>>>> outputs as the batch gets re-processed, or having to have the > >> output > >> >>>>>>> go to > >> >>>>>>> a "couldn't deliver" file. > >> >>>>>>> > >> >>>>>>> It's bad enough if the problem is a corrupt message that can > never > >> be > >> >>>>>>> delivered and the error is a permanent one, but if the problem > is > >> >>>>>>> instead a > >> >>>>>>> temporary error (server is down), you will have a LOT of > messages > >> >>>>>>> going > >> >>>>>>> into this 'couldn't deliver' bucket, when what should really be > >> >>>>>>> happening > >> >>>>>>> is that rsyslog should be pausing in the processing of messages, > >> >>>>>>> retrying > >> >>>>>>> after the suspended timeout. > >> >>>>>>> > >> >>>>>>> David Lang > >> >>>>>>> > >> >>>>>>> > >> >>>>>>> On Mar 13, 2016 9:07 PM, "David Lang" <[email protected]> wrote: > >> >>>>>>> > >> >>>>>>>> > >> >>>>>>>> On Sun, 13 Mar 2016, Rainer Gerhards wrote: > >> >>>>>>>> > >> >>>>>>>>> > >> >>>>>>>>> Rainer, can you explain what happens when a batch fails on an > >> >>>>>>>>> action? > >> >>>>>>>>> > >> >>>>>>>>> > >> >>>>>>>>>> > >> >>>>>>>>>>> It depends on when what happens and if we have batches of > one > >> or > >> >>>>>>>>>> > >> >>>>>>>>>> batches of many. For batches of one, the failure is > immediately > >> >>>>>>>>>> detected and acted upon. For batches of many,all action > strings > >> or > >> >>>>>>>>>> stored temporarily. During that time, it is assume that the > >> >>>>>>>>>> messages > >> >>>>>>>>>> could be processed. There is no way around this assumption, > >> >>>>>>>>>> because > >> >>>>>>>>>> we > >> >>>>>>>>>> now run each action completely through the rule engine before > >> the > >> >>>>>>>>>> next > >> >>>>>>>>>> one is begun to process, which means that other parts of the > >> >>>>>>>>>> config > >> >>>>>>>>>> have actually acted on a message, which means we cannot undo > >> that > >> >>>>>>>>>> state. If is needed to act on the "real" state, we need to > >> process > >> >>>>>>>>>> batches of one. There simply is no other solution possible. > So > >> >>>>>>>>>> let's > >> >>>>>>>>>> stay with batches of many. Once the batch is completed, the > >> >>>>>>>>>> temporary > >> >>>>>>>>>> strings are passed to the output's commitTransaction entry > >> point. > >> >>>>>>>>>> It > >> >>>>>>>>>> is currently assumed that the output does inteligently handle > >> the > >> >>>>>>>>>> case, and as it looks most do this well (no regression tests > >> >>>>>>>>>> failed, > >> >>>>>>>>>> no bad reports since we introduced this with the initial v8 > >> >>>>>>>>>> version > >> >>>>>>>>>> (15+ month ago). If the action fails, status is reflected > >> >>>>>>>>>> accordingly. > >> >>>>>>>>>> Also, the core engine *should* (but obviously does not > support > >> >>>>>>>>>> this > >> >>>>>>>>>> right now) write an error file with the messages in error > (just > >> >>>>>>>>>> like > >> >>>>>>>>>> omelasticsearch does, which has proven to be quite good). > >> >>>>>>>>>> > >> >>>>>>>>>> > >> >>>>>>>>>> well, it's not in the specs that modules should have local > error > >> >>>>>>>>> > >> >>>>>>>>> files > >> >>>>>>>>> the > >> >>>>>>>>> way omelasticsearch does. > >> >>>>>>>>> > >> >>>>>>>>> What I would expect if there is a problem when commiting the > >> batch > >> >>>>>>>>> is > >> >>>>>>>>> that > >> >>>>>>>>> the entire batch would be marked as not delivered and try > again > >> >>>>>>>>> with > >> >>>>>>>>> half > >> >>>>>>>>> of it. > >> >>>>>>>>> > >> >>>>>>>>> This will cause messages to other outputs working from the > same > >> >>>>>>>>> queue to > >> >>>>>>>>> get duplciate messages, but it's better than loosing an entire > >> >>>>>>>>> batch > >> >>>>>>>>> of > >> >>>>>>>>> messages because one is corrupt in some way. or there is a > >> >>>>>>>>> temporary > >> >>>>>>>>> problem with the output. > >> >>>>>>>>> > >> >>>>>>>>> David Lang > >> >>>>>>>>> > >> >>>>>>>>> I have managed to look at least briefly at the > commitTransaction > >> >>>>>>>>> core > >> >>>>>>>>> > >> >>>>>>>>> engine code. I think we can even improve it and may get some > >> >>>>>>>>> ability > >> >>>>>>>>>> > >> >>>>>>>>>> to do retries if the action itself does not handle that. At > the > >> >>>>>>>>>> minimum, it should have the ability (which was intended > >> >>>>>>>>>> originally, > >> >>>>>>>>>> but pushed away by more urgent work) to write that error > file. > >> >>>>>>>>>> > >> >>>>>>>>>> I am very grateful that Kane is working on a repro with the > >> >>>>>>>>>> standard > >> >>>>>>>>>> modules. Once we have that, I can go deeper into the code and > >> >>>>>>>>>> check > >> >>>>>>>>>> what I am thinking about. It might be possible that we cannot > >> >>>>>>>>>> manage > >> >>>>>>>>>> to repro this with the standard modules. Then we need to > think > >> >>>>>>>>>> about a > >> >>>>>>>>>> way via the testing module to do the same thing. > >> >>>>>>>>>> > >> >>>>>>>>>> I will try to address this within the 8.18 timeframe, but > cannot > >> >>>>>>>>>> totally commit to it. There has a lot of work been going on, > and > >> >>>>>>>>>> especially all that CI testing has taken up lots of works. > >> >>>>>>>>>> Nevertheless, I've been able to iron out quite some nits, so > >> it's > >> >>>>>>>>>> probably worth it. I would still like to continue to get the > CI > >> >>>>>>>>>> environment into better shape, because now I know what I am > >> doing. > >> >>>>>>>>>> If > >> >>>>>>>>>> I stop again and let it go for a couple of weeks, I would > lose a > >> >>>>>>>>>> lot > >> >>>>>>>>>> of momentum. I have also do to some work for support > customers, > >> >>>>>>>>>> which > >> >>>>>>>>>> keeps me able to work on rsyslog in general, so this, too, > has > >> >>>>>>>>>> priority (and validly has so because without support > customers I > >> >>>>>>>>>> would > >> >>>>>>>>>> be totally unable to work on rsyslog other than as a hobby) > >> >>>>>>>>>> > >> >>>>>>>>>> I hope this explains the current state. > >> >>>>>>>>>> Rainer > >> >>>>>>>>>> > >> >>>>>>>>>> > >> >>>>>>>>>> Kane, can you try and duplicate your problem using standard > >> >>>>>>>>>> modules? > >> >>>>>>>>>> > >> >>>>>>>>>>> I'm > >> >>>>>>>>>>> thinking that if you define a omrelp action that doesn't go > >> >>>>>>>>>>> anywhere, > >> >>>>>>>>>>> it > >> >>>>>>>>>>> should return errors similar to what you are trying to do, > so > >> you > >> >>>>>>>>>>> should > >> >>>>>>>>>>> be > >> >>>>>>>>>>> able to duplicate the problem that way. > >> >>>>>>>>>>> > >> >>>>>>>>>>> David Lang > >> >>>>>>>>>>> > >> >>>>>>>>>>> > >> >>>>>>>>>>> > >> >>>>>>>>>>> > >> >>>>>>>>>>> On Tue, 23 Feb 2016, Kane Kim wrote: > >> >>>>>>>>>>> > >> >>>>>>>>>>> Looking at omkafka module source code it seems that it > relies > >> on > >> >>>>>>>>>>> rsyslog > >> >>>>>>>>>>> > >> >>>>>>>>>>> retries in DoAction, returning RS_RET_SUSPENDED: > >> >>>>>>>>>>>> > >> >>>>>>>>>>>> > >> >>>>>>>>>>>> DBGPRINTF("omkafka: writeKafka returned %d\n", iRet); > >> >>>>>>>>>>>> if(iRet != RS_RET_OK) { > >> >>>>>>>>>>>> iRet = RS_RET_SUSPENDED; > >> >>>>>>>>>>>> } > >> >>>>>>>>>>>> > >> >>>>>>>>>>>> I've tried similar code in DoAction, it seems that action > >> >>>>>>>>>>>> processing > >> >>>>>>>>>>>> is suspended at that point and messages are not retried to > >> >>>>>>>>>>>> module. > >> >>>>>>>>>>>> > >> >>>>>>>>>>>> > >> >>>>>>>>>>>> On Tue, Feb 23, 2016 at 9:06 AM, Kane Kim > >> >>>>>>>>>>>> <[email protected]> > >> >>>>>>>>>>>> wrote: > >> >>>>>>>>>>>> > >> >>>>>>>>>>>> Thanks for your help Rainer, I'll try to debug what's going > >> on, > >> >>>>>>>>>>>> so > >> >>>>>>>>>>>> far > >> >>>>>>>>>>>> > >> >>>>>>>>>>>> it > >> >>>>>>>>>>>>> > >> >>>>>>>>>>>>> seems rsyslog doesn't retry even with batch size 1. It > >> doesn't > >> >>>>>>>>>>>>> retry > >> >>>>>>>>>>>>> if I > >> >>>>>>>>>>>>> return RS_RET_SUSPENDED from DoAction as well. > >> >>>>>>>>>>>>> > >> >>>>>>>>>>>>> On Tue, Feb 23, 2016 at 9:05 AM, Rainer Gerhards > >> >>>>>>>>>>>>> <[email protected] > >> >>>>>>>>>>>>> > >> >>>>>>>>>>>>> > >> >>>>>>>>>>>>> wrote: > >> >>>>>>>>>>>>>> > >> >>>>>>>>>>>>>> > >> >>>>>>>>>>>>>> > >> >>>>>>>>>>>>>> > >> >>>>>>>>>>>>> 2016-02-23 18:03 GMT+01:00 David Lang <[email protected]>: > >> >>>>>>>>>>>>> > >> >>>>>>>>>>>>> > >> >>>>>>>>>>>>>> > >> >>>>>>>>>>>>>> On Tue, 23 Feb 2016, Rainer Gerhards wrote: > >> >>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>> 2016-02-23 17:38 GMT+01:00 Kane Kim < > [email protected] > >> >: > >> >>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>>> Hello Rainer, thanks for the prompt reply! To give you > >> some > >> >>>>>>>>>>>>>>>>> context: > >> >>>>>>>>>>>>>>>>> I > >> >>>>>>>>>>>>>>>>> want > >> >>>>>>>>>>>>>>>>> to write module that both using batching and also > can't > >> >>>>>>>>>>>>>>>>> loose > >> >>>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>>> messages in > >> >>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>> > >> >>>>>>>>>>>>>> any circumstances. Are you saying it is by design that > >> rsyslog > >> >>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>> can't > >> >>>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>>> do > >> >>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>> > >> >>>>>>>>>>>>>> that together? According to documentation rsyslog will > retry > >> >>>>>>>>>>>>>> if > >> >>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>> module > >> >>>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>>> returns any error. Do you plan to fix this in rsyslog > or > >> >>>>>>>>>>>>>>>>> update > >> >>>>>>>>>>>>>>>>> documentation to say batching and retries don't work? > >> >>>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>> It depends on many things. In almost all cases, the > retry > >> >>>>>>>>>>>>>>>> should > >> >>>>>>>>>>>>>>>> work > >> >>>>>>>>>>>>>>>> well (and does so in practice). Unfortunately, I am > pretty > >> >>>>>>>>>>>>>>>> swamped. > >> >>>>>>>>>>>>>>>> I > >> >>>>>>>>>>>>>>>> need to go to a conference tomorrow and have had quite > >> some > >> >>>>>>>>>>>>>>>> unexpected > >> >>>>>>>>>>>>>>>> work today. It would probably be good if you could > ping me > >> >>>>>>>>>>>>>>>> next > >> >>>>>>>>>>>>>>>> week > >> >>>>>>>>>>>>>>>> to see if we can look into more details what is causing > >> you > >> >>>>>>>>>>>>>>>> pain. > >> >>>>>>>>>>>>>>>> But > >> >>>>>>>>>>>>>>>> I can't guarantee that I will be available early next > >> week. > >> >>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>> In general, we cannot handle a fatal error here from an > >> >>>>>>>>>>>>>>>> engine > >> >>>>>>>>>>>>>>>> PoV, > >> >>>>>>>>>>>>>>>> because everything is already processed and we do no > >> longer > >> >>>>>>>>>>>>>>>> have > >> >>>>>>>>>>>>>>>> the > >> >>>>>>>>>>>>>>>> original messages. This is simply needed if you want to > >> >>>>>>>>>>>>>>>> process > >> >>>>>>>>>>>>>>>> messages one after another through the full config (a > goal > >> >>>>>>>>>>>>>>>> for v8 > >> >>>>>>>>>>>>>>>> that > >> >>>>>>>>>>>>>>>> was muuuuch requested). As I said, the solution is to > use > >> >>>>>>>>>>>>>>>> batches of > >> >>>>>>>>>>>>>>>> one, because otherwise we would really need to turn > back > >> >>>>>>>>>>>>>>>> time > >> >>>>>>>>>>>>>>>> and > >> >>>>>>>>>>>>>>>> undo > >> >>>>>>>>>>>>>>>> everything that was already done on the messages in > >> question > >> >>>>>>>>>>>>>>>> by > >> >>>>>>>>>>>>>>>> other > >> >>>>>>>>>>>>>>>> modules (including state advances). > >> >>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>> I thought that if a batch failed, it pushed all the > >> messages > >> >>>>>>>>>>>>>>> back > >> >>>>>>>>>>>>>>> on > >> >>>>>>>>>>>>>>> the > >> >>>>>>>>>>>>>>> queue and retried with a half size batch until it got to > >> the > >> >>>>>>>>>>>>>>> one > >> >>>>>>>>>>>>>>> message > >> >>>>>>>>>>>>>>> that could not be processed and only did a fatal fail on > >> that > >> >>>>>>>>>>>>>>> message. > >> >>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>> Now, there is a big difference between a module giving a > >> hard > >> >>>>>>>>>>>>>>> error > >> >>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>> "this > >> >>>>>>>>>>>>>> > >> >>>>>>>>>>>>>> > >> >>>>>>>>>>>>>> > >> >>>>>>>>>>>>>> message is never going to be able to be processed no > matter > >> >>>>>>>>>>>>>> how > >> >>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>> many > >> >>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>> times > >> >>>>>>>>>>>>>> > >> >>>>>>>>>>>>>> > >> >>>>>>>>>>>>>> > >> >>>>>>>>>>>>>> it's retried" vs a soft error "there is a problem > delivering > >> >>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>> things > >> >>>>>>>>>>>>>>> to > >> >>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>> this > >> >>>>>>>>>>>>>> > >> >>>>>>>>>>>>>> > >> >>>>>>>>>>>>>> > >> >>>>>>>>>>>>>> destination right now, retry later". I thought the batch > >> >>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>> processing > >> >>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>> handled > >> >>>>>>>>>>>>>> > >> >>>>>>>>>>>>>> > >> >>>>>>>>>>>>>> > >> >>>>>>>>>>>>>> these differently. > >> >>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>> That's no longer possible with v8, at least generically. > As > >> I > >> >>>>>>>>>>>>>> said > >> >>>>>>>>>>>>>> above, we would need to turn back time. > >> >>>>>>>>>>>>>> > >> >>>>>>>>>>>>>> But I really run out of time now... > >> >>>>>>>>>>>>>> > >> >>>>>>>>>>>>>> Rainer > >> >>>>>>>>>>>>>> > >> >>>>>>>>>>>>>> > >> >>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>> David Lang > >> >>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>> _______________________________________________ > >> >>>>>>>>>>>>>>> rsyslog mailing list > >> >>>>>>>>>>>>>>> http://lists.adiscon.net/mailman/listinfo/rsyslog > >> >>>>>>>>>>>>>>> http://www.rsyslog.com/professional-services/ > >> >>>>>>>>>>>>>>> What's up with rsyslog? Follow > >> https://twitter.com/rgerhards > >> >>>>>>>>>>>>>>> NOTE WELL: This is a PUBLIC mailing list, posts are > >> ARCHIVED > >> >>>>>>>>>>>>>>> by a > >> >>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>> myriad of > >> >>>>>>>>>>>>>> > >> >>>>>>>>>>>>>> > >> >>>>>>>>>>>>>> > >> >>>>>>>>>>>>>> sites beyond our control. PLEASE UNSUBSCRIBE and DO NOT > POST > >> >>>>>>>>>>>>>> if > >> >>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>> you > >> >>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>> DON'T > >> >>>>>>>>>>>>>> > >> >>>>>>>>>>>>>> > >> >>>>>>>>>>>>>> > >> >>>>>>>>>>>>>> LIKE THAT. > >> >>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>> _______________________________________________ > >> >>>>>>>>>>>>>> > >> >>>>>>>>>>>>>> rsyslog mailing list > >> >>>>>>>>>>>>>> http://lists.adiscon.net/mailman/listinfo/rsyslog > >> >>>>>>>>>>>>>> http://www.rsyslog.com/professional-services/ > >> >>>>>>>>>>>>>> What's up with rsyslog? Follow > >> https://twitter.com/rgerhards > >> >>>>>>>>>>>>>> NOTE WELL: This is a PUBLIC mailing list, posts are > ARCHIVED > >> >>>>>>>>>>>>>> by > >> >>>>>>>>>>>>>> a > >> >>>>>>>>>>>>>> myriad > >> >>>>>>>>>>>>>> of sites beyond our control. PLEASE UNSUBSCRIBE and DO > NOT > >> >>>>>>>>>>>>>> POST > >> >>>>>>>>>>>>>> if > >> >>>>>>>>>>>>>> you > >> >>>>>>>>>>>>>> DON'T LIKE THAT. > >> >>>>>>>>>>>>>> > >> >>>>>>>>>>>>>> > >> >>>>>>>>>>>>>> > >> >>>>>>>>>>>>>> _______________________________________________ > >> >>>>>>>>>>>>> > >> >>>>>>>>>>>>> > >> >>>>>>>>>>>>> rsyslog mailing list > >> >>>>>>>>>>>> > >> >>>>>>>>>>>> http://lists.adiscon.net/mailman/listinfo/rsyslog > >> >>>>>>>>>>>> http://www.rsyslog.com/professional-services/ > >> >>>>>>>>>>>> What's up with rsyslog? Follow > https://twitter.com/rgerhards > >> >>>>>>>>>>>> NOTE WELL: This is a PUBLIC mailing list, posts are > ARCHIVED > >> by > >> >>>>>>>>>>>> a > >> >>>>>>>>>>>> myriad > >> >>>>>>>>>>>> of sites beyond our control. PLEASE UNSUBSCRIBE and DO NOT > >> POST > >> >>>>>>>>>>>> if > >> >>>>>>>>>>>> you > >> >>>>>>>>>>>> DON'T > >> >>>>>>>>>>>> LIKE THAT. > >> >>>>>>>>>>>> > >> >>>>>>>>>>>> _______________________________________________ > >> >>>>>>>>>>>> > >> >>>>>>>>>>>> rsyslog mailing list > >> >>>>>>>>>>> > >> >>>>>>>>>>> http://lists.adiscon.net/mailman/listinfo/rsyslog > >> >>>>>>>>>>> http://www.rsyslog.com/professional-services/ > >> >>>>>>>>>>> What's up with rsyslog? Follow > https://twitter.com/rgerhards > >> >>>>>>>>>>> NOTE WELL: This is a PUBLIC mailing list, posts are ARCHIVED > >> by a > >> >>>>>>>>>>> myriad > >> >>>>>>>>>>> of > >> >>>>>>>>>>> sites beyond our control. PLEASE UNSUBSCRIBE and DO NOT > POST if > >> >>>>>>>>>>> you > >> >>>>>>>>>>> DON'T > >> >>>>>>>>>>> LIKE THAT. > >> >>>>>>>>>>> > >> >>>>>>>>>>> _______________________________________________ > >> >>>>>>>>>>> > >> >>>>>>>>>> rsyslog mailing list > >> >>>>>>>>>> http://lists.adiscon.net/mailman/listinfo/rsyslog > >> >>>>>>>>>> http://www.rsyslog.com/professional-services/ > >> >>>>>>>>>> What's up with rsyslog? Follow https://twitter.com/rgerhards > >> >>>>>>>>>> NOTE WELL: This is a PUBLIC mailing list, posts are ARCHIVED > by > >> a > >> >>>>>>>>>> myriad > >> >>>>>>>>>> of sites beyond our control. PLEASE UNSUBSCRIBE and DO NOT > POST > >> if > >> >>>>>>>>>> you > >> >>>>>>>>>> DON'T LIKE THAT. > >> >>>>>>>>>> > >> >>>>>>>>>> _______________________________________________ > >> >>>>>>>>>> > >> >>>>>>>>>> rsyslog mailing list > >> >>>>>>>>> > >> >>>>>>>>> http://lists.adiscon.net/mailman/listinfo/rsyslog > >> >>>>>>>>> http://www.rsyslog.com/professional-services/ > >> >>>>>>>>> What's up with rsyslog? Follow https://twitter.com/rgerhards > >> >>>>>>>>> NOTE WELL: This is a PUBLIC mailing list, posts are ARCHIVED > by a > >> >>>>>>>>> myriad > >> >>>>>>>>> of sites beyond our control. PLEASE UNSUBSCRIBE and DO NOT > POST > >> if > >> >>>>>>>>> you > >> >>>>>>>>> DON'T LIKE THAT. > >> >>>>>>>>> > >> >>>>>>>>> _______________________________________________ > >> >>>>>>>>> > >> >>>>>>>> rsyslog mailing list > >> >>>>>>>> http://lists.adiscon.net/mailman/listinfo/rsyslog > >> >>>>>>>> http://www.rsyslog.com/professional-services/ > >> >>>>>>>> What's up with rsyslog? Follow https://twitter.com/rgerhards > >> >>>>>>>> NOTE WELL: This is a PUBLIC mailing list, posts are ARCHIVED > by a > >> >>>>>>>> myriad > >> >>>>>>>> of sites beyond our control. PLEASE UNSUBSCRIBE and DO NOT > POST if > >> >>>>>>>> you > >> >>>>>>>> DON'T LIKE THAT. > >> >>>>>>>> > >> >>>>>>>> _______________________________________________ > >> >>>>>>>> > >> >>>>>>> rsyslog mailing list > >> >>>>>>> http://lists.adiscon.net/mailman/listinfo/rsyslog > >> >>>>>>> http://www.rsyslog.com/professional-services/ > >> >>>>>>> What's up with rsyslog? Follow https://twitter.com/rgerhards > >> >>>>>>> NOTE WELL: This is a PUBLIC mailing list, posts are ARCHIVED by > a > >> >>>>>>> myriad > >> >>>>>>> of sites beyond our control. PLEASE UNSUBSCRIBE and DO NOT POST > if > >> >>>>>>> you > >> >>>>>>> DON'T LIKE THAT. > >> >>>>>>> > >> >>>>>>> > >> >>>>>> > >> >>>>>> _______________________________________________ > >> >>>>> > >> >>>>> rsyslog mailing list > >> >>>>> http://lists.adiscon.net/mailman/listinfo/rsyslog > >> >>>>> http://www.rsyslog.com/professional-services/ > >> >>>>> What's up with rsyslog? Follow https://twitter.com/rgerhards > >> >>>>> NOTE WELL: This is a PUBLIC mailing list, posts are ARCHIVED by a > >> >>>>> myriad > >> >>>>> of sites beyond our control. PLEASE UNSUBSCRIBE and DO NOT POST if > >> you > >> >>>>> DON'T LIKE THAT. > >> >>>>> > >> >>>>> _______________________________________________ > >> >>>> > >> >>>> rsyslog mailing list > >> >>>> http://lists.adiscon.net/mailman/listinfo/rsyslog > >> >>>> http://www.rsyslog.com/professional-services/ > >> >>>> What's up with rsyslog? Follow https://twitter.com/rgerhards > >> >>>> NOTE WELL: This is a PUBLIC mailing list, posts are ARCHIVED by a > >> myriad > >> >>>> of sites beyond our control. PLEASE UNSUBSCRIBE and DO NOT POST if > you > >> >>>> DON'T LIKE THAT. > >> >>>> > >> >>> > >> >> _______________________________________________ > >> >> rsyslog mailing list > >> >> http://lists.adiscon.net/mailman/listinfo/rsyslog > >> >> http://www.rsyslog.com/professional-services/ > >> >> What's up with rsyslog? Follow https://twitter.com/rgerhards > >> >> NOTE WELL: This is a PUBLIC mailing list, posts are ARCHIVED by a > myriad > >> >> of sites beyond our control. PLEASE UNSUBSCRIBE and DO NOT POST if > you > >> DON'T > >> >> LIKE THAT. > >> >> > >> > _______________________________________________ > >> > rsyslog mailing list > >> > http://lists.adiscon.net/mailman/listinfo/rsyslog > >> > http://www.rsyslog.com/professional-services/ > >> > What's up with rsyslog? Follow https://twitter.com/rgerhards > >> > NOTE WELL: This is a PUBLIC mailing list, posts are ARCHIVED by a > myriad > >> of > >> > sites beyond our control. PLEASE UNSUBSCRIBE and DO NOT POST if you > DON'T > >> > LIKE THAT. > >> _______________________________________________ > >> rsyslog mailing list > >> http://lists.adiscon.net/mailman/listinfo/rsyslog > >> http://www.rsyslog.com/professional-services/ > >> What's up with rsyslog? Follow https://twitter.com/rgerhards > >> NOTE WELL: This is a PUBLIC mailing list, posts are ARCHIVED by a myriad > >> of sites beyond our control. PLEASE UNSUBSCRIBE and DO NOT POST if you > >> DON'T LIKE THAT. > >> > > _______________________________________________ > > rsyslog mailing list > > http://lists.adiscon.net/mailman/listinfo/rsyslog > > http://www.rsyslog.com/professional-services/ > > What's up with rsyslog? Follow https://twitter.com/rgerhards > > NOTE WELL: This is a PUBLIC mailing list, posts are ARCHIVED by a myriad > of sites beyond our control. PLEASE UNSUBSCRIBE and DO NOT POST if you > DON'T LIKE THAT. > _______________________________________________ > rsyslog mailing list > http://lists.adiscon.net/mailman/listinfo/rsyslog > http://www.rsyslog.com/professional-services/ > What's up with rsyslog? Follow https://twitter.com/rgerhards > NOTE WELL: This is a PUBLIC mailing list, posts are ARCHIVED by a myriad > of sites beyond our control. PLEASE UNSUBSCRIBE and DO NOT POST if you > DON'T LIKE THAT. > _______________________________________________ rsyslog mailing list http://lists.adiscon.net/mailman/listinfo/rsyslog http://www.rsyslog.com/professional-services/ What's up with rsyslog? Follow https://twitter.com/rgerhards NOTE WELL: This is a PUBLIC mailing list, posts are ARCHIVED by a myriad of sites beyond our control. PLEASE UNSUBSCRIBE and DO NOT POST if you DON'T LIKE THAT.

