Hi Mahesh, Can you try with 2258_v7.patch I just sent to you?
I have pulled the latest code on OpenSAF 5.1 branch, re-created the cluster. And it works with the case old active SC-1 (OpenSAF 5.1) and new standby SC-2 (with 2258_v7.patch included in). To apply 2258_v7.patch, please do remove the just pushed ticket "log: implement SaLogFilterSetCallbackT and version handling [#2146]" , I have not rebased the code on that yet. Regards, Vu > -----Original Message----- > From: A V Mahesh [mailto:mahesh.va...@oracle.com] > Sent: Thursday, February 23, 2017 11:45 AM > To: Vu Minh Nguyen <vu.m.ngu...@dektech.com.au>; > lennart.l...@ericsson.com; canh.v.tru...@dektech.com.au > Cc: opensaf-devel@lists.sourceforge.net > Subject: Re: [devel] [PATCH 0 of 3] Review Request for log: add alternative > destinations of log records [#2258] V4 > > Hi Vu/Lennart, > > > In broad WITHOUT the #2258 patch, the same code/setup working fine with > 2 sc node (staging changeset: 8609 ), > as soon as we apply `2258_v5.patch` V5 patch on staging (changeset: > 8609 ) that you have provided yesterday, > on one sc node and try to bring up that in to cluster (in-service test) > we are observing the issue of new node (with #2258 patch) not joining > cluster. > > ============================================================== > ==================================================== > eb 23 10:01:59 SC-1 osafimmnd[15279]: NO Implementer (applier) > connected: 15 (@safAmfService2010f) <127, 2010f> > Feb 23 10:01:59 SC-1 osaflogd[15289]: NO LOGSV_DATA_GROUPNAME not > found > Feb 23 10:01:59 SC-1 osaflogd[15289]: NO LOG root directory is: > "/var/log/opensaf/saflog" > Feb 23 10:01:59 SC-1 osaflogd[15289]: NO LOG data group is: "" > Feb 23 10:01:59 SC-1 osaflogd[15289]: NO LGS_MBCSV_VERSION = 7 > Feb 23 10:01:59 SC-1 osafamfnd[15329]: NO Assigned > 'safSi=SC-2N,safApp=OpenSAF' STANDBY to 'safSu=SC- > 1,safSg=2N,safApp=OpenSAF' > Feb 23 10:01:59 SC-1 opensafd: OpenSAF(5.1.M0 - ) services successfully > started > Feb 23 10:01:59 SC-1 osafamfnd[15329]: NO > 'safComp=LOG,safSu=SC-1,safSg=2N,safApp=OpenSAF' faulted due to > 'avaDown' : Recovery is 'nodeFailfast' > Feb 23 10:01:59 SC-1 osafamfnd[15329]: ER > safComp=LOG,safSu=SC-1,safSg=2N,safApp=OpenSAF Faulted due > to:avaDown > Recovery is:nodeFailfast > Feb 23 10:01:59 SC-1 osafamfnd[15329]: Rebooting OpenSAF NodeId = > 131343 > EE Name = , Reason: Component faulted: recovery is node failfast, > OwnNodeId = 131343, SupervisionTime = 60 > Feb 23 10:01:59 SC-1 opensaf_reboot: Rebooting local node; timeout=60 > Feb 23 10:02:00 SC-1 osafimmnd[15279]: NO Implementer (applier) > connected: 16 (@OpenSafImmReplicatorB) <144, 2010f> > Feb 23 10:01:59 SC-1 opensaf_reboot: Rebooting local node; timeout=60 > ============================================================== > ==================================================== > > So it is evident that in-service upgrade part code of this need to be > corrected. > > Please see my comments as [AVM] and let me know if you need some traces > . > > If you're planing to prepare new V6 patch , please do prepare on top of > today's latest staging. > > On 2/23/2017 9:33 AM, Vu Minh Nguyen wrote: > > Hi Mahesh, > > > > I have done in-service upgrade/downgrade with following cases: > > 1) New Active SC-1 (OpenSAF 5.2 with the attached patch) + old standby > SC-2 > > (OpenSAF 5.1) > > --> Work fine > [AVM] This is not a practical use cause of in-service upgrade , we can > ignore this test further > > > > 2) Old Active SC-1 (OpenSAF 5.1) + new standby SC-2 (with or without > > attached patch) > > --> SC-2 is restarted & not able to join the cluster. > [AVM] This use cause/flow is we do get in in-service upgrade , so we > need to address this. > > > > I got following messages in syslog: > > Feb 23 09:32:42 SC-2 user.notice opensafd: OpenSAF(5.2.M0 - > > 8529:b5addd36e45d:default) services successfully started > > Feb 23 09:32:43 SC-2 local0.warn osafntfimcnd[701]: WA > ntfimcn_imm_init > > saImmOiImplementerSet() returned SA_AIS_ERR_TIMEOUT (5) > > Feb 23 09:32:45 SC-2 local0.warn osafntfimcnd[701]: WA > ntfimcn_imm_init > > saImmOiImplementerSet() returned SA_AIS_ERR_TIMEOUT (5) > > Feb 23 09:32:47 SC-2 local0.warn osafntfimcnd[701]: WA > ntfimcn_imm_init > > saImmOiImplementerSet() returned SA_AIS_ERR_TIMEOUT (5) > > Feb 23 09:32:49 SC-2 local0.warn osafntfimcnd[701]: WA > ntfimcn_imm_init > > saImmOiImplementerSet() returned SA_AIS_ERR_TIMEOUT (5) > > Feb 23 09:32:50 SC-2 local0.err osafmsgnd[592]: ER > saImmOiImplementerSet > > FAILED:5 > > Feb 23 09:32:50 SC-2 local0.err osafmsgnd[592]: ER > saImmOiImplementerSet > > FAILED:5 > > Feb 23 09:32:50 SC-2 local0.notice osafamfnd[496]: NO > > 'safSu=SC-2,safSg=NoRed,safApp=OpenSAF' component restart probation > timer > > started (timeout: 60000000000 ns) > > Feb 23 09:32:50 SC-2 local0.notice osafamfnd[496]: NO Restarting a > component > > of 'safSu=SC-2,safSg=NoRed,safApp=OpenSAF' (comp restart count: 1) > > Feb 23 09:32:50 SC-2 local0.notice osafamfnd[496]: NO > > 'safComp=MQND,safSu=SC-2,safSg=NoRed,safApp=OpenSAF' faulted due > to > > 'avaDown' : Recovery is 'componentRestart' > > Feb 23 09:32:50 SC-2 local0.info osafmsgnd[736]: mkfifo already exists: > > /var/lib/opensaf/osafmsgnd.fifo File exists > > > > And sometimes, on active SC-1 (OpenSAF 5.1), the node is not able to up > > because of following error: > > > > Feb 23 11:00:32 SC-1 local0.err osafclmna[406]: MDTM:TIPC Dsock Socket > > creation failed in MDTM_INIT err :Address family not supported by > protocol > > Feb 23 11:00:32 SC-1 local0.err osafclmna[406]: ER ncs_agents_startup > FAILED > [AVM] No such issues ( with both TCP & TIPC) (staging changeset: 8609 ) > > > > Are you getting similar problem at your side? > > Please note that, the problem is existed WITH or WITHOUT the #2258 > patch. > [AVM] No , problem only if we apply `2258_v5.patch` V5 patch on staging > (changeset: 8609 ) > try to bring up that node in to cluster. > > > -AVM > > > > > I have informed this to IMM to have a look, not sure any problem with MDS > > layer or any problem with my environment setup. > > In the meantime, please have a look at the updated patch, I will continue > > checking the problem. Will keep you updated. > [AVM] I haven't seen any IMM problems > > > > Regards, Vu > > > >> -----Original Message----- > >> From: A V Mahesh [mailto:mahesh.va...@oracle.com] > >> Sent: Wednesday, February 22, 2017 5:36 PM > >> To: Vu Minh Nguyen <vu.m.ngu...@dektech.com.au>; > >> lennart.l...@ericsson.com; canh.v.tru...@dektech.com.au > >> Cc: opensaf-devel@lists.sourceforge.net > >> Subject: Re: [devel] [PATCH 0 of 3] Review Request for log: add > > alternative > >> destinations of log records [#2258] V4 > >> > >> Please see correction New Standby SC-1 ( with patch ) > >> > >> -AVM > >> > >> On 2/22/2017 4:02 PM, A V Mahesh wrote: > >>> Hi Vu, > >>> > >>> With this new patch , we have another issue : > >>> > >>> 1) standby Core by `/usr/lib64/opensaf/osaflogd' issue got resolved . > >>> > >>> 2) In-service upgrade is Not working , I have Old Active SC-2 ( with > >>> out patch ) and New Standby SC-1 ( with patch ) > >>> > >>> the new New Standby SC-1 not joining the cluster ( in-service > >>> upgrade failed ) > >>> > >>> New Standby SC-1 > >>> > >>> > >> > ============================================================== > >> ====================================================== > >>> > >>> Feb 22 15:53:05 SC-1 osafamfnd[15752]: NO > >>> 'safSu=SC-1,safSg=NoRed,safApp=OpenSAF' Presence State > >> INSTANTIATING > >>> => INSTANTIATED > >>> Feb 22 15:53:05 SC-1 osafamfnd[15752]: NO Assigning > >>> 'safSi=NoRed4,safApp=OpenSAF' ACTIVE to > >>> 'safSu=SC-1,safSg=NoRed,safApp=OpenSAF' > >>> Feb 22 15:53:05 SC-1 osafamfnd[15752]: NO Assigned > >>> 'safSi=NoRed4,safApp=OpenSAF' ACTIVE to > >>> 'safSu=SC-1,safSg=NoRed,safApp=OpenSAF' > >>> Feb 22 15:53:05 SC-1 osafsmfd[15889]: Started > >>> Feb 22 15:53:05 SC-1 osafamfnd[15752]: NO > >>> 'safSu=SC-1,safSg=2N,safApp=OpenSAF' Presence State INSTANTIATING > => > >>> INSTANTIATED > >>> Feb 22 15:53:05 SC-1 osafamfnd[15752]: NO Assigning > >>> 'safSi=SC-2N,safApp=OpenSAF' STANDBY to > >>> 'safSu=SC-1,safSg=2N,safApp=OpenSAF' > >>> Feb 22 15:53:05 SC-1 osafrded[15672]: NO RDE role set to STANDBY > >>> Feb 22 15:53:05 SC-1 osafrded[15672]: NO Peer up on node 0x2020f > >>> Feb 22 15:53:05 SC-1 osafrded[15672]: NO Got peer info request from > >>> node 0x2020f with role ACTIVE > >>> Feb 22 15:53:05 SC-1 osafrded[15672]: NO Got peer info response from > >>> node 0x2020f with role ACTIVE > >>> Feb 22 15:53:05 SC-1 osafimmd[15691]: NO MDS event from svc_id 24 > >>> (change:5, dest:13) > >>> Feb 22 15:53:05 SC-1 osafimmd[15691]: NO MDS event from svc_id 24 > >>> (change:3, dest:13) > >>> Feb 22 15:53:05 SC-1 osafimmd[15691]: NO MDS event from svc_id 24 > >>> (change:5, dest:13) > >>> Feb 22 15:53:05 SC-1 osafimmd[15691]: NO MDS event from svc_id 25 > >>> (change:3, dest:567412424453430) > >>> Feb 22 15:53:05 SC-1 osafimmd[15691]: NO MDS event from svc_id 25 > >>> (change:3, dest:565213401202663) > >>> Feb 22 15:53:05 SC-1 osafimmd[15691]: NO MDS event from svc_id 25 > >>> (change:3, dest:566312912825221) > >>> Feb 22 15:53:05 SC-1 osafimmd[15691]: NO MDS event from svc_id 25 > >>> (change:3, dest:564113889574230) > >>> Feb 22 15:53:05 SC-1 osafimmnd[15702]: NO Implementer (applier) > >>> connected: 17 (@safAmfService2010f) <127, 2010f> > >>> Feb 22 15:53:05 SC-1 osaflogd[15712]: NO LOGSV_DATA_GROUPNAME > >> not found > >>> Feb 22 15:53:05 SC-1 osaflogd[15712]: NO LOG root directory is: > >>> "/var/log/opensaf/saflog" > >>> Feb 22 15:53:05 SC-1 osaflogd[15712]: NO LOG data group is: "" > >>> Feb 22 15:53:05 SC-1 osaflogd[15712]: NO LGS_MBCSV_VERSION = 7 > >>> Feb 22 15:53:05 SC-1 osafamfnd[15752]: NO Assigned > >>> 'safSi=SC-2N,safApp=OpenSAF' STANDBY to > >>> 'safSu=SC-1,safSg=2N,safApp=OpenSAF' > >>> Feb 22 15:53:05 SC-1 opensafd: OpenSAF(5.1.M0 - ) services > >>> successfully started > >>> Feb 22 15:53:05 SC-1 osafamfnd[15752]: NO > >>> 'safComp=LOG,safSu=SC-1,safSg=2N,safApp=OpenSAF' faulted due to > >>> 'avaDown' : Recovery is 'nodeFailfast' > >>> Feb 22 15:53:05 SC-1 osafamfnd[15752]: ER > >>> safComp=LOG,safSu=SC-1,safSg=2N,safApp=OpenSAF Faulted due > >> to:avaDown > >>> Recovery is:nodeFailfast > >>> Feb 22 15:53:05 SC-1 osafamfnd[15752]: Rebooting OpenSAF NodeId = > >>> 131343 EE Name = , Reason: Component faulted: recovery is node > >>> failfast, OwnNodeId = 131343, SupervisionTime = 60 > >>> Feb 22 15:53:05 SC-1 opensaf_reboot: Rebooting local node; > timeout=60 > >>> Feb 22 15:53:43 SC-1 syslog-ng[1171]: syslog-ng starting up; > >>> version='2.0.9' > >>> > >>> > >> > ============================================================== > >> ====================================================== > >>> > >>> Old - Active - SC-2 > >>> > >>> > >> > ============================================================== > >> ====================================================== > >>> > >>> Feb 22 15:53:02 SC-2 osafimmnd[16359]: NO NODE STATE-> > >>> IMM_NODE_R_AVAILABLE > >>> Feb 22 15:53:02 SC-2 osafimmloadd: NO Sync starting > >>> Feb 22 15:53:02 SC-2 osafimmloadd: IN Synced 390 objects in total > >>> Feb 22 15:53:02 SC-2 osafimmnd[16359]: NO NODE STATE-> > >>> IMM_NODE_FULLY_AVAILABLE 18511 > >>> Feb 22 15:53:02 SC-2 osafimmnd[16359]: NO Epoch set to 3 in > ImmModel > >>> Feb 22 15:53:02 SC-2 osafimmd[16346]: NO ACT: New Epoch for > IMMND > >>> process at node 2020f old epoch: 2 new epoch:3 > >>> Feb 22 15:53:02 SC-2 osafimmd[16346]: NO ACT: New Epoch for > IMMND > >>> process at node 2040f old epoch: 2 new epoch:3 > >>> Feb 22 15:53:02 SC-2 osafimmd[16346]: NO ACT: New Epoch for > IMMND > >>> process at node 2030f old epoch: 2 new epoch:3 > >>> Feb 22 15:53:02 SC-2 osafimmloadd: NO Sync ending normally > >>> Feb 22 15:53:02 SC-2 osafimmd[16346]: NO ACT: New Epoch for > IMMND > >>> process at node 2010f old epoch: 0 new epoch:3 > >>> Feb 22 15:53:02 SC-2 osafimmnd[16359]: NO SERVER STATE: > >>> IMM_SERVER_SYNC_SERVER --> IMM_SERVER_READY > >>> Feb 22 15:53:03 SC-2 osafamfd[16408]: NO Received node_up from > 2010f: > >>> msg_id 1 > >>> Feb 22 15:53:03 SC-2 osafamfd[16408]: NO Node 'SC-1' joined the > cluster > >>> Feb 22 15:53:03 SC-2 osafimmnd[16359]: NO Implementer connected: > 16 > >>> (MsgQueueService131343) <0, 2010f> > >>> Feb 22 15:53:03 SC-2 osafrded[16327]: NO Peer up on node 0x2010f > >>> Feb 22 15:53:03 SC-2 osafrded[16327]: NO Got peer info request from > >>> node 0x2010f with role STANDBY > >>> Feb 22 15:53:03 SC-2 osafrded[16327]: NO Got peer info response from > >>> node 0x2010f with role STANDBY > >>> Feb 22 15:53:03 SC-2 osafimmd[16346]: NO MDS event from svc_id 24 > >>> (change:5, dest:13) > >>> Feb 22 15:53:03 SC-2 osafimmnd[16359]: NO Implementer (applier) > >>> connected: 17 (@safAmfService2010f) <0, 2010f> > >>> Feb 22 15:53:03 SC-2 osafamfd[16408]: NO Cluster startup is done > >>> Feb 22 15:53:04 SC-2 osafimmnd[16359]: NO Implementer (applier) > >>> connected: 18 (@OpenSafImmReplicatorB) <0, 2010f> > >>> Feb 22 15:53:09 SC-2 osafdtmd[16304]: NO Lost contact with 'SC-1' > >>> Feb 22 15:53:09 SC-2 osaffmd[16336]: NO Node Down event for node id > >>> 2010f: > >>> Feb 22 15:53:09 SC-2 osafimmd[16346]: NO MDS event from svc_id 24 > >>> (change:6, dest:13) > >>> Feb 22 15:53:09 SC-2 osafimmd[16346]: NO MDS event from svc_id 25 > >>> (change:4, dest:564113889574230) > >>> Feb 22 15:53:09 SC-2 osaffmd[16336]: NO Current role: ACTIVE > >>> Feb 22 15:53:09 SC-2 osaffmd[16336]: Rebooting OpenSAF NodeId = > >> 131343 > >>> EE Name = , Reason: Received Node Down for peer controller, > OwnNodeId > >>> = 131599, SupervisionTime = 60 > >>> Feb 22 15:53:09 SC-2 osafrded[16327]: NO Peer down on node 0x2010f > >>> Feb 22 15:53:09 SC-2 osafamfd[16408]: NO Node 'SC-1' left the cluster > >>> Feb 22 15:53:09 SC-2 osaflogd[16372]: NO Failed (2) to send of WRITE > >>> ack to: 2010f00003d6a > >>> Feb 22 15:53:09 SC-2 osaflogd[16372]: NO Failed (2) to send of WRITE > >>> ack to: 2010f00003d6a > >>> Feb 22 15:53:09 SC-2 osaflogd[16372]: NO Failed (2) to send of WRITE > >>> ack to: 2010f00003d74 > >>> Feb 22 15:53:09 SC-2 osafclmd[16398]: NO Node 131343 went down. > Not > >>> sending track callback for agents on that node > >>> Feb 22 15:53:09 SC-2 osafclmd[16398]: NO Node 131343 went down. > Not > >>> sending track callback for agents on that node > >>> Feb 22 15:53:09 SC-2 osafclmd[16398]: NO Node 131343 went down. > Not > >>> sending track callback for agents on that node > >>> Feb 22 15:53:09 SC-2 osafclmd[16398]: NO Node 131343 went down. > Not > >>> sending track callback for agents on that node > >>> Feb 22 15:53:09 SC-2 osafclmd[16398]: NO Node 131343 went down. > Not > >>> sending track callback for agents on that node > >>> Feb 22 15:53:09 SC-2 osafclmd[16398]: NO Node 131343 went down. > Not > >>> sending track callback for agents on that node > >>> Feb 22 15:53:09 SC-2 osafclmd[16398]: NO Node 131343 went down. > Not > >>> sending track callback for agents on that node > >>> Feb 22 15:53:09 SC-2 osafimmd[16346]: WA IMMD lost contact with > peer > >>> IMMD (NCSMDS_RED_DOWN) > >>> Feb 22 15:53:09 SC-2 osafimmnd[16359]: NO Global discard node > received > >>> for nodeId:2010f pid:15702 > >>> Feb 22 15:53:09 SC-2 osafimmnd[16359]: NO Implementer disconnected > 16 > >>> <0, 2010f(down)> (MsgQueueService131343) > >>> Feb 22 15:53:09 SC-2 osafimmnd[16359]: NO Implementer disconnected > 17 > >>> <0, 2010f(down)> (@safAmfService2010f) > >>> Feb 22 15:53:09 SC-2 osafimmnd[16359]: NO Implementer disconnected > 18 > >>> <0, 2010f(down)> (@OpenSafImmReplicatorB) > >>> Feb 22 15:53:09 SC-2 opensaf_reboot: Rebooting remote node in the > >>> absence of PLM is outside the scope of OpenSAF > >>> > >>> > >> > ============================================================== > >> ====================================================== > >>> > >>> -AVM > >>> > >>> On 2/22/2017 3:13 PM, Vu Minh Nguyen wrote: > >>>> Hi Mahesh, > >>>> > >>>> I put all required patches into one. Try to use this and see if you > >>>> still > >>>> have that problem or not. > >>>> > >>>> Regards, Vu > >>>> > >>>>> -----Original Message----- > >>>>> From: A V Mahesh [mailto:mahesh.va...@oracle.com] > >>>>> Sent: Wednesday, February 22, 2017 3:35 PM > >>>>> To: Vu Minh Nguyen <vu.m.ngu...@dektech.com.au>; > >>>>> lennart.l...@ericsson.com; canh.v.tru...@dektech.com.au > >>>>> Cc: opensaf-devel@lists.sourceforge.net > >>>>> Subject: Re: [devel] [PATCH 0 of 3] Review Request for log: add > >>>> alternative > >>>>> destinations of log records [#2258] V4 > >>>>> > >>>>> Hi Vu, > >>>>> > >>>>> I used new #3 and #4 patches , Can you please re-send All the final > >>>>> patch in go, > >>>>> > >>>>> which i need to apply on today`s staging ( if possible publish the > >>>>> with > >>>>> new version ) > >>>>> > >>>>> -AVM > >>>>> > >>>>> > >>>>> On 2/22/2017 1:52 PM, Vu Minh Nguyen wrote: > >>>>>> Hi Mahesh, > >>>>>> > >>>>>>> Core was generated by `/usr/lib64/opensaf/osaflogd'. > >>>>>>> Program terminated with signal 11, Segmentation fault. > >>>>>>> #0 ckpt_proc_cfg_stream(lgs_cb*, void*) () at > >>>>>>> src/log/logd/lgs_mbcsv.cc:2195 > >>>>>>> 2195 src/log/logd/lgs_mbcsv.cc: No such file or directory. > >>>>>>> in src/log/logd/lgs_mbcsv.cc > >>>>>> Backtrace still points to old position (lgs_mbcsv:2195). I guess the > >>>>>> osaflogd binary has not been updated with the fixed patch. > >>>>>> > >>>>>> > >>>>>> Regards, Vu > >>>>>> > >>>>>>> -----Original Message----- > >>>>>>> From: A V Mahesh [mailto:mahesh.va...@oracle.com] > >>>>>>> Sent: Wednesday, February 22, 2017 3:18 PM > >>>>>>> To: Vu Minh Nguyen <vu.m.ngu...@dektech.com.au>; > >>>>>>> lennart.l...@ericsson.com; canh.v.tru...@dektech.com.au > >>>>>>> Cc: opensaf-devel@lists.sourceforge.net > >>>>>>> Subject: Re: [devel] [PATCH 0 of 3] Review Request for log: add > >>>>>> alternative > >>>>>>> destinations of log records [#2258] V4 > >>>>>>> > >>>>>>> Hi Vu, > >>>>>>> > >>>>>>> SC-2 standby osaflogd core dumped still occurs ( Not resolved) , the > >>>> new > >>>>>>> patch only resolved the application (/usr/bin/logtest ) Segmentation > >>>>>>> fault on SC-1 Active. > >>>>>>> > >>>>>>> > >> > ============================================================== > >> > ============================================================== > >>>>>>> # gdb /usr/lib64/opensaf/osaflogd > core_1487751055.osaflogd.4594 > >> GNU > >>>>>>> gdb > >>>>>>> (GDB) SUSE (7.3-0.6.1) > >>>>>>> Copyright (C) 2011 Free Software Foundation, Inc. > >>>>>>> ....... > >>>>>>> Core was generated by `/usr/lib64/opensaf/osaflogd'. > >>>>>>> Program terminated with signal 11, Segmentation fault. > >>>>>>> #0 ckpt_proc_cfg_stream(lgs_cb*, void*) () at > >>>>>>> src/log/logd/lgs_mbcsv.cc:2195 > >>>>>>> 2195 src/log/logd/lgs_mbcsv.cc: No such file or directory. > >>>>>>> in src/log/logd/lgs_mbcsv.cc > >>>>>>> (gdb) bt > >>>>>>> #0 ckpt_proc_cfg_stream(lgs_cb*, void*) () at > >>>>>>> src/log/logd/lgs_mbcsv.cc:2195 > >>>>>>> #1 0x00007f97b026f960 in ckpt_decode_log_struct(lgs_cb*, > >>>>>>> ncs_mbcsv_cb_arg*, void*, void*, unsigned int (*)(edu_hdl_tag*, > >>>>>>> edu_tkn_tag*, void*, unsigned int*, edu_buf_env_tag*, > >> EDP_OP_TYPE, > >>>>>>> EDU_ERR*)) () at src/log/logd/lgs_mbcsv.cc:950 > >>>>>>> #2 0x00007f97b02710dc in ckpt_decode_async_update(lgs_cb*, > >>>>>>> ncs_mbcsv_cb_arg*) () at src/log/logd/lgs_mbcsv.cc:1086 > >>>>>>> #3 0x00007f97b0273941 in mbcsv_callback(ncs_mbcsv_cb_arg*) () > at > >>>>>>> src/log/logd/lgs_mbcsv.cc:880 > >>>>>>> #4 0x00007f97af372596 in ncs_mbscv_rcv_decode () from > >>>>>>> /usr/lib/../lib64/libopensaf_core.so.0 > >>>>>>> #5 0x00007f97af372766 in ncs_mbcsv_rcv_async_update () from > >>>>>>> /usr/lib/../lib64/libopensaf_core.so.0 > >>>>>>> #6 0x00007f97af379370 in mbcsv_process_events () from > >>>>>>> /usr/lib/../lib64/libopensaf_core.so.0 > >>>>>>> #7 0x00007f97af3794db in mbcsv_hdl_dispatch_all () from > >>>>>>> /usr/lib/../lib64/libopensaf_core.so.0 > >>>>>>> #8 0x00007f97af373ce2 in mbcsv_process_dispatch_request () at > >>>>>>> src/mbc/mbcsv_api.c:423 > >>>>>>> #9 0x00007f97b027096e in lgs_mbcsv_dispatch(unsigned int) () at > >>>>>>> src/log/logd/lgs_mbcsv.cc:327 > >>>>>>> #10 0x00007f97b024d9f2 in main () at src/log/logd/lgs_main.cc:583 > >>>>>>> (gdb) bt full > >>>>>>> #0 ckpt_proc_cfg_stream(lgs_cb*, void*) () at > >>>>>>> src/log/logd/lgs_mbcsv.cc:2195 > >>>>>>> ckpt_data_handler = {0x7f97b0270300 > >>>>>>> <ckpt_proc_initialize_client(lgs_cb*, void*)>, 0x7f97b02701c0 > >>>>>>> <ckpt_proc_finalize_client(lgs_cb*, void*)>, > >>>>>>> 0x7f97b0270060 <ckpt_proc_agent_down(lgs_cb*, void*)>, > >>>>>>> 0x7f97b02712f0 <ckpt_proc_log_write(lgs_cb*, void*)>, > >> 0x7f97b0271ab0 > >>>>>>> <ckpt_proc_open_stream(lgs_cb*, void*)>, > >>>>>>> 0x7f97b026fe80 <ckpt_proc_close_stream(lgs_cb*, > > void*)>, > >>>>>>> 0x7f97b0272380 <ckpt_proc_cfg_stream(lgs_cb*, void*)>, > >>>>> 0x7f97b0274800 > >>>>>>> <ckpt_proc_lgs_cfg_v2(lgs_cb*, void*)>, > >>>>>>> 0x7f97b0274e10 <ckpt_proc_lgs_cfg_v3(lgs_cb*, void*)>, > >>>>>>> 0x7f97b02754f0 <ckpt_proc_lgs_cfg_v5(lgs_cb*, void*)>} > >>>>>>> #1 0x00007f97b026f960 in ckpt_decode_log_struct(lgs_cb*, > >>>>>>> ncs_mbcsv_cb_arg*, void*, void*, unsigned int (*)(edu_hdl_tag*, > >>>>>>> edu_tkn_tag*, void*, unsigned int*, edu_buf_env_tag*, > >> EDP_OP_TYPE, > >>>>>>> EDU_ERR*)) () at src/log/logd/lgs_mbcsv.cc:950 > >>>>>>> ckpt_data_handler = {0x7f97b0270300 > >>>>>>> <ckpt_proc_initialize_client(lgs_cb*, void*)>, 0x7f97b02701c0 > >>>>>>> <ckpt_proc_finalize_client(lgs_cb*, void*)>, > >>>>>>> 0x7f97b0270060 <ckpt_proc_agent_down(lgs_cb*, void*)>, > >>>>>>> 0x7f97b02712f0 <ckpt_proc_log_write(lgs_cb*, void*)>, > >> 0x7f97b0271ab0 > >>>>>>> <ckpt_proc_open_stream(lgs_cb*, void*)>, > >>>>>>> 0x7f97b026fe80 <ckpt_proc_close_stream(lgs_cb*, > > void*)>, > >>>>>>> 0x7f97b0272380 <ckpt_proc_cfg_stream(lgs_cb*, void*)>, > >>>>> 0x7f97b0274800 > >>>>>>> <ckpt_proc_lgs_cfg_v2(lgs_cb*, void*)>, > >>>>>>> 0x7f97b0274e10 <ckpt_proc_lgs_cfg_v3(lgs_cb*, void*)>, > >>>>>>> 0x7f97b02754f0 <ckpt_proc_lgs_cfg_v5(lgs_cb*, void*)>} > >>>>>>> #2 0x00007f97b02710dc in ckpt_decode_async_update(lgs_cb*, > >>>>>>> ncs_mbcsv_cb_arg*) () at src/log/logd/lgs_mbcsv.cc:1086 > >>>>>>> ckpt_data_handler = {0x7f97b0270300 > >>>>>>> <ckpt_proc_initialize_client(lgs_cb*, void*)>, 0x7f97b02701c0 > >>>>>>> <ckpt_proc_finalize_client(lgs_cb*, void*)>, > >>>>>>> 0x7f97b0270060 <ckpt_proc_agent_down(lgs_cb*, void*)>, > >>>>>>> 0x7f97b02712f0 <ckpt_proc_log_write(lgs_cb*, void*)>, > >> 0x7f97b0271ab0 > >>>>>>> <ckpt_proc_open_stream(lgs_cb*, void*)>, > >>>>>>> 0x7f97b026fe80 <ckpt_proc_close_stream(lgs_cb*, > > void*)>, > >>>>>>> 0x7f97b0272380 <ckpt_proc_cfg_stream(lgs_cb*, void*)>, > >>>>> 0x7f97b0274800 > >>>>>>> <ckpt_proc_lgs_cfg_v2(lgs_cb*, void*)>, > >>>>>>> 0x7f97b0274e10 <ckpt_proc_lgs_cfg_v3(lgs_cb*, void*)>, > >>>>>>> 0x7f97b02754f0 <ckpt_proc_lgs_cfg_v5(lgs_cb*, void*)>} > >>>>>>> #3 0x00007f97b0273941 in mbcsv_callback(ncs_mbcsv_cb_arg*) () > at > >>>>>>> src/log/logd/lgs_mbcsv.cc:880 > >>>>>>> ckpt_data_handler = {0x7f97b0270300 > >>>>>>> <ckpt_proc_initialize_client(lgs_cb*, void*)>, 0x7f97b02701c0 > >>>>>>> <ckpt_proc_finalize_client(lgs_cb*, void*)>, > >>>>>>> 0x7f97b0270060 <ckpt_proc_agent_down(lgs_cb*, void*)>, > >>>>>>> 0x7f97b02712f0 <ckpt_proc_log_write(lgs_cb*, void*)>, > >> 0x7f97b0271ab0 > >>>>>>> <ckpt_proc_open_stream(lgs_cb*, void*)>, > >>>>>>> 0x7f97b026fe80 <ckpt_proc_close_stream(lgs_cb*, > > void*)>, > >>>>>>> 0x7f97b0272380 <ckpt_proc_cfg_stream(lgs_cb*, void*)>, > >>>>> 0x7f97b0274800 > >>>>>>> <ckpt_proc_lgs_cfg_v2(lgs_cb*, void*)>, > >>>>>>> 0x7f97b0274e10 <ckpt_proc_lgs_cfg_v3(lgs_cb*, void*)>, > >>>>>>> 0x7f97b02754f0 <ckpt_proc_lgs_cfg_v5(lgs_cb*, void*)>} > >>>>>>> #4 0x00007f97af372596 in ncs_mbscv_rcv_decode () from > >>>>>>> /usr/lib/../lib64/libopensaf_core.so.0 > >>>>>>> mbcsv_init_process_req_func = {0x7f97af373630 > >>>>>>> <mbcsv_process_initialize_request>, 0x7f97af374f10 > >>>>>>> <mbcsv_process_get_sel_obj_request>, > >>>>>>> 0x7f97af373be0 <mbcsv_process_dispatch_request>, > >>>>>>> 0x7f97af373af0 <mbcsv_process_finalize_request>, > 0x7f97af373cf0 > >>>>>>> <mbcsv_process_open_request>, > >>>>>>> 0x7f97af374050 <mbcsv_process_close_request>, > >>>> 0x7f97af3741e0 > >>>>>>> <mbcsv_process_chg_role_request>, 0x7f97af3744c0 > >>>>>>> <mbcsv_process_snd_ckpt_request>, > >>>>>>> 0x7f97af3747d0 <mbcsv_process_snd_ntfy_request>, > >>>>>>> 0x7f97af374970 <mbcsv_process_snd_data_req>, 0x7f97af373930 > >>>>>>> <mbcsv_process_get_request>, > >>>>>>> 0x7f97af374bd0 <mbcsv_process_set_request>} > >>>>>>> #5 0x00007f97af372766 in ncs_mbcsv_rcv_async_update () from > >>>>>>> /usr/lib/../lib64/libopensaf_core.so.0 > >>>>>>> mbcsv_init_process_req_func = {0x7f97af373630 > >>>>>>> <mbcsv_process_initialize_request>, 0x7f97af374f10 > >>>>>>> <mbcsv_process_get_sel_obj_request>, > >>>>>>> 0x7f97af373be0 <mbcsv_process_dispatch_request>, > >>>>>>> 0x7f97af373af0 <mbcsv_process_finalize_request>, > 0x7f97af373cf0 > >>>>>>> <mbcsv_process_open_request>, > >>>>>>> 0x7f97af374050 <mbcsv_process_close_request>, > >>>> 0x7f97af3741e0 > >>>>>>> <mbcsv_process_chg_role_request>, 0x7f97af3744c0 > >>>>>>> <mbcsv_process_snd_ckpt_request>, > >>>>>>> 0x7f97af3747d0 <mbcsv_process_snd_ntfy_request>, > >>>>>>> 0x7f97af374970 <mbcsv_process_snd_data_req>, 0x7f97af373930 > >>>>>>> <mbcsv_process_get_request>, > >>>>>>> 0x7f97af374bd0 <mbcsv_process_set_request>} > >>>>>>> #6 0x00007f97af379370 in mbcsv_process_events () from > >>>>>>> /usr/lib/../lib64/libopensaf_core.so.0 > >>>>>>> No symbol table info available. > >>>>>>> #7 0x00007f97af3794db in mbcsv_hdl_dispatch_all () from > >>>>>>> /usr/lib/../lib64/libopensaf_core.so.0 > >>>>>>> No symbol table info available. > >>>>>>> #8 0x00007f97af373ce2 in mbcsv_process_dispatch_request () at > >>>>>>> src/mbc/mbcsv_api.c:423 > >>>>>>> mbcsv_init_process_req_func = {0x7f97af373630 > >>>>>>> <mbcsv_process_initialize_request>, 0x7f97af374f10 > >>>>>>> <mbcsv_process_get_sel_obj_request>, > >>>>>>> 0x7f97af373be0 <mbcsv_process_dispatch_request>, > >>>>>>> 0x7f97af373af0 <mbcsv_process_finalize_request>, > 0x7f97af373cf0 > >>>>>>> <mbcsv_process_open_request>, > >>>>>>> 0x7f97af374050 <mbcsv_process_close_request>, > >>>> 0x7f97af3741e0 > >>>>>>> <mbcsv_process_chg_role_request>, 0x7f97af3744c0 > >>>>>>> <mbcsv_process_snd_ckpt_request>, > >>>>>>> 0x7f97af3747d0 <mbcsv_process_snd_ntfy_request>, > >>>>>>> 0x7f97af374970 <mbcsv_process_snd_data_req>, 0x7f97af373930 > >>>>>>> <mbcsv_process_get_request>, > >>>>>>> 0x7f97af374bd0 <mbcsv_process_set_request>} > >>>>>>> #9 0x00007f97b027096e in lgs_mbcsv_dispatch(unsigned int) () at > >>>>>>> src/log/logd/lgs_mbcsv.cc:327 > >>>>>>> ckpt_data_handler = {0x7f97b0270300 > >>>>>>> <ckpt_proc_initialize_client(lgs_cb*, void*)>, 0x7f97b02701c0 > >>>>>>> <ckpt_proc_finalize_client(lgs_cb*, void*)>, > >>>>>>> 0x7f97b0270060 <ckpt_proc_agent_down(lgs_cb*, void*)>, > >>>>>>> 0x7f97b02712f0 <ckpt_proc_log_write(lgs_cb*, void*)>, > >> 0x7f97b0271ab0 > >>>>>>> <ckpt_proc_open_stream(lgs_cb*, void*)>, > >>>>>>> 0x7f97b026fe80 <ckpt_proc_close_stream(lgs_cb*, > > void*)>, > >>>>>>> 0x7f97b0272380 <ckpt_proc_cfg_stream(lgs_cb*, void*)>, > >>>>> 0x7f97b0274800 > >>>>>>> <ckpt_proc_lgs_cfg_v2(lgs_cb*, void*)>, > >>>>>>> 0x7f97b0274e10 <ckpt_proc_lgs_cfg_v3(lgs_cb*, void*)>, > >>>>>>> 0x7f97b02754f0 <ckpt_proc_lgs_cfg_v5(lgs_cb*, void*)>} > >>>>>>> ---Type <return> to continue, or q <return> to quit--- > >>>>>>> #10 0x00007f97b024d9f2 in main () at src/log/logd/lgs_main.cc:583 > >>>>>>> usr1_sel_obj = {raise_obj = -1, rmv_obj = -1} > >>>>>>> _lgs_cb = {mds_hdl = 65547, mds_role = > V_DEST_RL_STANDBY, > >>>>> vaddr > >>>>>>> = 11, log_version = {releaseCode = 65 'A', majorVersion = 2 '\002', > >>>>>>> minorVersion = 2 '\002'}, client_tree = { > >>>>>>> root_node = {bit = -1, left = 0x7f97b04cf1b0, right = > >>>>>>> 0x7f97b04a2418, key_info = 0x7f97b04b7bd0 ""}, params = > {key_size > >>>>>>> = 4}, > >>>>>>> n_nodes = 8}, comp_name = {_opaque = {46, 24947, > >>>>>>> 17254, 28015, 15728, 20300, 11335, 24947, 21350, > >>>>>>> 15733, > >>>>>>> 17235, 12845, 29484, 26209, 26451, 12861, 11342, 24947, 16742, > >> 28784, > >>>>>>> 20285, 25968, 21358, 17985, > >>>>>>> 0 <repeats 105 times>}}, amf_hdl = 4288675841, > >>>>>>> amfSelectionObject = 15, amf_invocation_id = 0, is_quiesced_set = > >>>> false, > >>>>>>> immOiHandle = 554050912783, immSelectionObject = 21, > >>>>>>> clmSelectionObject = 17, clm_hdl = 4279238657, > >>>>>>> ha_state = > >>>>>>> SA_AMF_HA_STANDBY, last_client_id = 208, async_upd_cnt = 743, > >>>>>>> ckpt_state > >>>>>>> = COLD_SYNC_IDLE, mbcsv_hdl = 4293918753, > >>>>>>> mbcsv_sel_obj = 23, mbcsv_ckpt_hdl = 4292870177, > >>>>>>> mbcsv_peer_version = 7, edu_hdl = {is_inited = true, tree = > >>>>>>> {root_node > >>>> = > >>>>>>> {bit = -1, left = 0x7f97b04cf2e0, > >>>>>>> right = 0x7f97b04a25b8, key_info = 0x7f97b04b7d40 > >>>> ""}, > >>>>>>> params = {key_size = 8}, n_nodes = 12}, to_version = 1}, > >>>>>>> fully_initialized = true, lga_down_list_head = 0x0, > >>>>>>> lga_down_list_tail = 0x0, clm_init_sel_obj = > >>>>>>> {raise_obj = > >>>> -1, > >>>>>>> rmv_obj = -1}, nid_started = true, scAbsenceAllowed = 900, > >>>>>>> lgs_recovery_state = LGS_NORMAL} > >>>>>>> nfds = 7 > >>>>>>> fds = {{fd = 19, events = 1, revents = 0}, {fd = 15, > >>>>>>> events = > >>>>>>> 1, revents = 0}, {fd = 23, events = 1, revents = 1}, {fd = 13, > >>>>>>> events = > >>>>>>> 1, revents = 0}, {fd = -1, events = 1, > >>>>>>> revents = 0}, {fd = 17, events = 1, revents = 0}, > >>>>>>> {fd = > >>>> 21, > >>>>>>> events = 1, revents = 0}} > >>>>>>> mbox_msgs = {0, 0, 0, 0, 0} > >>>>>>> lgs_cb = 0x7f97b04a2400 > >>>>>>> mbox_low = {0, 0, 0, 0, 0} > >>>>>>> lgs_mbox_init_mutex = {__data = {__lock = 0, __count = 0, > >>>>>>> __owner = 0, __nusers = 0, __kind = 0, __spins = 0, __list = > >>>>>>> {__prev = > >>>>>>> 0x0, __next = 0x0}}, > >>>>>>> __size = '\000' <repeats 39 times>, __align = 0} > >>>>>>> lgs_mbx = 4291821569 > >>>>>>> mbox_high = {0, 0, 0, 0, 0} > >>>>>>> mbox_full = {false, false, false, false, false} > >>>>>>> (gdb) > >>>>>>> (gdb) > >>>>>>> > >>>>>>> Feb 22 13:40:55 SC-2 osafimmnd[4584]: NO Ccb 131 COMMITTED > >>>>>>> (immcfg_SC-1_18714) > >>>>>>> Feb 22 13:40:56 SC-2 osafamfnd[4634]: NO > >>>>>>> 'safComp=LOG,safSu=SC-2,safSg=2N,safApp=OpenSAF' faulted due > to > >>>>>>> 'avaDown' : Recovery is 'nodeFailfast' > >>>>>>> Feb 22 13:40:56 SC-2 osafamfnd[4634]: ER > >>>>>>> safComp=LOG,safSu=SC-2,safSg=2N,safApp=OpenSAF Faulted due > >>>>>>> to:avaDown > >>>>>>> Recovery is:nodeFailfast > >>>>>>> Feb 22 13:40:56 SC-2 osafamfnd[4634]: Rebooting OpenSAF NodeId > = > >>>>>>> 131599 > >>>>>>> EE Name = , Reason: Component faulted: recovery is node failfast, > >>>>>>> OwnNodeId = 131599, SupervisionTime = 60 > >>>>>>> Feb 22 13:40:56 SC-2 opensaf_reboot: Rebooting local node; > >> timeout=60 > >>>>>>> > >> > ============================================================== > >> > ============================================================== > >>>>>>> On 2/22/2017 12:23 PM, A V Mahesh wrote: > >>>>>>>> Hi Vu, > >>>>>>>> > >>>>>>>> On 2/22/2017 12:19 PM, Vu Minh Nguyen wrote: > >>>>>>>>> [Vu] I has sent you 02 patches. There is code change in osaflogd > >>>>>>>>> code > >>>>>> that > >>>>>>>>> fix the coredump you have observed. > >>>>>>>>> The other one is test code that fix the logtest coredump. > >>>>>>>> Ok I will re-test , and update you . > >>>>>>>> > >>>>>>>> -AVM > >>>>>>>> > >>>>>>>> On 2/22/2017 12:19 PM, Vu Minh Nguyen wrote: > >>>>>>>>> Hi Mahehs, > >>>>>>>>> > >>>>>>>>> See my reply inline, [Vu]. > >>>>>>>>> > >>>>>>>>> Regards, Vu > >>>>>>>>> > >>>>>>>>>> -----Original Message----- > >>>>>>>>>> From: A V Mahesh [mailto:mahesh.va...@oracle.com] > >>>>>>>>>> Sent: Wednesday, February 22, 2017 1:36 PM > >>>>>>>>>> To: Vu Minh Nguyen <vu.m.ngu...@dektech.com.au>; > >>>>>>>>>> lennart.l...@ericsson.com; canh.v.tru...@dektech.com.au > >>>>>>>>>> Cc: opensaf-devel@lists.sourceforge.net > >>>>>>>>>> Subject: Re: [PATCH 0 of 3] Review Request for log: add > >>>>>>>>>> alternative > >>>>>>>>>> destinations of log records [#2258] V4 > >>>>>>>>>> > >>>>>>>>>> Hi Vu, > >>>>>>>>>> > >>>>>>>>>> > >>>>>>>>>> On 2/22/2017 11:52 AM, Vu Minh Nguyen wrote: > >>>>>>>>>>> Hi Mahesh, > >>>>>>>>>>> > >>>>>>>>>>> Have a code fault in uml test, and other one in checkpoint. > >>>>>>>>>> [AVM] This is Normal Suse 11 VM ( not UML). > >>>>>>>>>>> I have just updated the code. Please re-apply for #3 and #4 > >>>> patches. > >>>>>>>>>> [AVM] is these new patch has function changes or only test > code > >>>>>>> changes ? > >>>>>>>>> [Vu] I has sent you 02 patches. There is code change in osaflogd > >>>>>>>>> code > >>>>>> that > >>>>>>>>> fix the coredump you have observed. > >>>>>>>>> The other one is test code that fix the logtest coredump. > >>>>>>>>>>> Note that, test case #14 of suite 17 should be run on active > >>>>>>>>>>> node, > >>>>>>>>>> otherwise > >>>>>>>>>>> getting failed. > >>>>>>>>>> [AVM] Segmentation fault of /usr/bin/logtest Not a big issue , > >>>>>>>>>> we need to debug why osaflogd core dumped and it is > >>>>>>>>>> critical > >>>>>>>>> [Vu] I found the problem. You can try with the new one to see if > >>>>>>>>> the > >>>>>>>>> coredump is still there or not. > >>>>>>>>>>> I will put condition check to that test case later. > >>>>>>>>>> -AVM > >>>>>>>>>> > >>>>>>>>>> > >>>>>>>>>>> Regards, Vu > >>>>>>>>>>> > >>>>>>>>>>>> -----Original Message----- > >>>>>>>>>>>> From: A V Mahesh [mailto:mahesh.va...@oracle.com] > >>>>>>>>>>>> Sent: Wednesday, February 22, 2017 12:16 PM > >>>>>>>>>>>> To: Vu Minh Nguyen <vu.m.ngu...@dektech.com.au>; > >>>>>>>>>>>> lennart.l...@ericsson.com; canh.v.tru...@dektech.com.au > >>>>>>>>>>>> Cc: opensaf-devel@lists.sourceforge.net > >>>>>>>>>>>> Subject: Re: [PATCH 0 of 3] Review Request for log: add > >>>> alternative > >>>>>>>>>>>> destinations of log records [#2258] V4 > >>>>>>>>>>>> > >>>>>>>>>>>> Hi Vu, > >>>>>>>>>>>> > >>>>>>>>>>>> Thanks , > >>>>>>>>>>>> > >>>>>>>>>>>> While testing /usr/bin/logtest , SC-2 standby osaflogd core > >>>>> dumped > >>>>>>>>> and > >>>>>>>>>>>> /usr/bin/logtest on SC-1 Active > >>>>>>>>>>>> got Segmentation fault , am I missing any other patch ( i am > >>>> using > >>>>>>>>>>>> devel published patch only ) > >>>>>>>>>>>> > >>>>>>>>>>>> Following patches i am using : > >>>>>>>>>>>> > >>>>>>>>>>>> 1) #2293 (sent by Anders Widel, but not yet pushed) > >>>>>>>>>>>> 2) #2258 (v2, sent by Lennart, but not yet pushed yet) > >>>>>>>>>>>> 3) #2258 (v4, sent by Vu, but not yet pushed yet) > >>>>>>>>>>>> > >>>>>>>>>>>> > >> > ============================================================== > >>>>>>>>>>>> ======================================== > >>>>>>>>>>>> > >>>>>>>>>>>> > >>>>>>>>>>>> Core was generated by `/usr/lib64/opensaf/osaflogd'. > >>>>>>>>>>>> Program terminated with signal 11, Segmentation fault. > >>>>>>>>>>>> #0 ckpt_proc_cfg_stream(lgs_cb*, void*) () at > >>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc:2195 > >>>>>>>>>>>> 2195 src/log/logd/lgs_mbcsv.cc: No such file or directory. > >>>>>>>>>>>> in src/log/logd/lgs_mbcsv.cc > >>>>>>>>>>>> (gdb) bt > >>>>>>>>>>>> #0 ckpt_proc_cfg_stream(lgs_cb*, void*) () at > >>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc:2195 > >>>>>>>>>>>> #1 0x00007f12c3e22960 in ckpt_decode_log_struct(lgs_cb*, > >>>>>>>>>>>> ncs_mbcsv_cb_arg*, void*, void*, unsigned int > >> (*)(edu_hdl_tag*, > >>>>>>>>>>>> edu_tkn_tag*, void*, unsigned int*, edu_buf_env_tag*, > >>>>>>> EDP_OP_TYPE, > >>>>>>>>>>>> EDU_ERR*)) () at src/log/logd/lgs_mbcsv.cc:950 > >>>>>>>>>>>> #2 0x00007f12c3e240dc in > >> ckpt_decode_async_update(lgs_cb*, > >>>>>>>>>>>> ncs_mbcsv_cb_arg*) () at src/log/logd/lgs_mbcsv.cc:1086 > >>>>>>>>>>>> #3 0x00007f12c3e26941 in > >> mbcsv_callback(ncs_mbcsv_cb_arg*) () > >>>>> at > >>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc:880 > >>>>>>>>>>>> #4 0x00007f12c2f25596 in ncs_mbscv_rcv_decode () from > >>>>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0 > >>>>>>>>>>>> #5 0x00007f12c2f25766 in ncs_mbcsv_rcv_async_update () > >> from > >>>>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0 > >>>>>>>>>>>> #6 0x00007f12c2f2c370 in mbcsv_process_events () from > >>>>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0 > >>>>>>>>>>>> #7 0x00007f12c2f2c4db in mbcsv_hdl_dispatch_all () from > >>>>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0 > >>>>>>>>>>>> #8 0x00007f12c2f26ce2 in mbcsv_process_dispatch_request > () > >> at > >>>>>>>>>>>> src/mbc/mbcsv_api.c:423 > >>>>>>>>>>>> #9 0x00007f12c3e2396e in lgs_mbcsv_dispatch(unsigned int) > () > >> at > >>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc:327 > >>>>>>>>>>>> #10 0x00007f12c3e009f2 in main () at > >>>>>>>>>>>> src/log/logd/lgs_main.cc:583 > >>>>>>>>>>>> (gdb) > >>>>>>>>>>>> > >>>>>>>>>>>> > >> > ============================================================== > >>>>>>>>>>>> ======================================== > >>>>>>>>>>>> > >>>>>>>>>>>> Feb 22 10:37:06 SC-1 osafimmnd[4020]: NO Invalid error > >> reported > >>>>>>>>>>>> implementer 'safLogService', Ccb 161 will be aborted > >>>>>>>>>>>> Feb 22 10:37:06 SC-1 osafimmnd[4020]: NO Ccb 161 aborted > in > >>>>>>>>>> COMPLETED > >>>>>>>>>>>> processing (validation) > >>>>>>>>>>>> Feb 22 10:37:06 SC-1 osafimmnd[4020]: NO Ccb 161 > ABORTED > >>>>>>>>>> (immcfg_SC- > >>>>>>>>>>>> 1_5394) > >>>>>>>>>>>> Add values Fail > >>>>>>>>>>>> > >>>>>>>>>>>> Program received signal SIGSEGV, Segmentation fault. > >>>>>>>>>>>> 0x000055555556929a in read_and_compare.isra.7 () at > >>>>>>>>>>>> src/log/apitest/tet_LogOiOps.c:1891 > >>>>>>>>>>>> 1891 src/log/apitest/tet_LogOiOps.c: No such file or > >>>>>>>>>>>> directory. > >>>>>>>>>>>> in src/log/apitest/tet_LogOiOps.c > >>>>>>>>>>>> (gdb) Feb 22 10:37:07 SC-1 sshd[5298]: Accepted keyboard- > >>>>>>>>>> interactive/pam > >>>>>>>>>>>> for root from 10.176.178.22 port 51945 ssh2 > >>>>>>>>>>>> bt > >>>>>>>>>>>> #0 0x000055555556929a in read_and_compare.isra.7 () at > >>>>>>>>>>>> src/log/apitest/tet_LogOiOps.c:1891 > >>>>>>>>>>>> #1 0x0000555555569bbb in > >>>>>>>>>>>> check_logRecordDestinationConfigurationEmpty > >>>>>>>>>>>> () at src/log/apitest/tet_LogOiOps.c:2179 > >>>>>>>>>>>> #2 0x0000555555573495 in run_test_case () > >>>>>>>>>>>> #3 0x0000555555573934 in test_run () > >>>>>>>>>>>> #4 0x000055555555c7cd in main () at > >>>>>>>>>>>> src/log/apitest/logtest.c:569 > >>>>>>>>>>>> (gdb) > >>>>>>>>>>>> > >>>>>>>>>>>> > >> > ============================================================== > >>>>>>>>>>>> ======================================== > >>>>>>>>>>>> > >>>>>>>>>>>> -AVM > >>>>>>>>>>>> > >>>>>>>>>>>> On 2/22/2017 9:48 AM, Vu Minh Nguyen wrote: > >>>>>>>>>>>>> Hi Mahesh, > >>>>>>>>>>>>> > >>>>>>>>>>>>> I send them in attachment instead, and name them in the > >> order. > >>>>>>>>>>>>> I just pull the latest code, and apply them without getting > > any > >>>>>> hunk > >>>>>>>>>>> error. > >>>>>>>>>>>>> Please try with them, and let me know if you see any > problem. > >>>>>>>>>>>>> > >>>>>>>>>>>>> Regards, Vu > >>>>>>>>>>>>> > >>>>>>>>>>>>>> -----Original Message----- > >>>>>>>>>>>>>> From: A V Mahesh [mailto:mahesh.va...@oracle.com] > >>>>>>>>>>>>>> Sent: Wednesday, February 22, 2017 11:09 AM > >>>>>>>>>>>>>> To: Vu Minh Nguyen <vu.m.ngu...@dektech.com.au>; > >>>>>>>>>>>>>> lennart.l...@ericsson.com; > canh.v.tru...@dektech.com.au > >>>>>>>>>>>>>> Cc: opensaf-devel@lists.sourceforge.net > >>>>>>>>>>>>>> Subject: Re: [PATCH 0 of 3] Review Request for log: add > >>>>>> alternative > >>>>>>>>>>>>>> destinations of log records [#2258] V4 > >>>>>>>>>>>>>> > >>>>>>>>>>>>>> Hi Vu, > >>>>>>>>>>>>>> > >>>>>>>>>>>>>> I did follow that still i get Hunk #2 FAILED even on today's > >>>>>> staging > >> > ============================================================== > >>>>>>>>>>>>>> ================== > >>>>>>>>>>>>>> > >>>>>>>>>>>>>> [root@dhcp-hyd-scp-5fl-10-176-177-96 staging]# patch - > p1 > >>>>> <2293 > >>>>>>>>>>>>>> patching file src/base/Makefile.am > >>>>>>>>>>>>>> Hunk #1 succeeded at 33 (offset 1 line). > >>>>>>>>>>>>>> Hunk #3 succeeded at 183 (offset 1 line). > >>>>>>>>>>>>>> patching file src/base/file_descriptor.cc > >>>>>>>>>>>>>> patching file src/base/file_descriptor.h > >>>>>>>>>>>>>> patching file src/base/tests/unix_socket_test.cc > >>>>>>>>>>>>>> patching file src/base/unix_client_socket.cc > >>>>>>>>>>>>>> patching file src/base/unix_server_socket.cc > >>>>>>>>>>>>>> patching file src/base/unix_socket.cc > >>>>>>>>>>>>>> patching file src/base/unix_socket.h > >>>>>>>>>>>>>> > >>>>>>>>>>>>>> [root@dhcp-hyd-scp-5fl-10-176-177-96 staging]# patch - > p1 > >>>>> <2258- > >>>>>>> 1 > >>>>>>>>>>>>>> patching file src/log/Makefile.am > >>>>>>>>>>>>>> Hunk #1 succeeded at 71 (offset -1 lines). > >>>>>>>>>>>>>> patching file src/log/config/logsv_classes.xml > >>>>>>>>>>>>>> Hunk #1 FAILED at 147. > >>>>>>>>>>>>>> 1 out of 1 hunk FAILED -- saving rejects to file > >>>>>>>>>>>>>> src/log/config/logsv_classes.xml.rej > >>>>>>>>>>>>>> patching file src/log/logd/lgs_config.cc > >>>>>>>>>>>>>> Hunk #1 succeeded at 35 (offset -5 lines). > >>>>>>>>>>>>>> Hunk #2 FAILED at 705. > >>>>>>>>>>>>>> Hunk #3 FAILED at 971. > >>>>>>>>>>>>>> 2 out of 3 hunks FAILED -- saving rejects to file > >>>>>>>>>>>>>> src/log/logd/lgs_config.cc.rej > >>>>>>>>>>>>>> patching file src/log/logd/lgs_config.h > >>>>>>>>>>>>>> Hunk #1 FAILED at 304. > >>>>>>>>>>>>>> 1 out of 1 hunk FAILED -- saving rejects to file > >>>>>>>>>>>>>> src/log/logd/lgs_config.h.rej > >>>>>>>>>>>>>> patching file src/log/logd/lgs_dest.cc > >>>>>>>>>>>>>> patching file src/log/logd/lgs_dest.h > >>>>>>>>>>>>>> patching file src/log/logd/lgs_evt.cc > >>>>>>>>>>>>>> patching file src/log/logd/lgs_imm.cc > >>>>>>>>>>>>>> Hunk #1 FAILED at 45. > >>>>>>>>>>>>>> Hunk #2 succeeded at 235 (offset -1 lines). > >>>>>>>>>>>>>> Hunk #3 FAILED at 877. > >>>>>>>>>>>>>> Hunk #4 succeeded at 1273 (offset -20 lines). > >>>>>>>>>>>>>> Hunk #5 succeeded at 1404 (offset -1 lines). > >>>>>>>>>>>>>> Hunk #6 succeeded at 1449 (offset -20 lines). > >>>>>>>>>>>>>> Hunk #7 succeeded at 2032 (offset -1 lines). > >>>>>>>>>>>>>> Hunk #8 FAILED at 2181. > >>>>>>>>>>>>>> Hunk #9 succeeded at 2271 (offset -54 lines). > >>>>>>>>>>>>>> Hunk #10 succeeded at 2387 (offset -1 lines). > >>>>>>>>>>>>>> Hunk #11 succeeded at 2377 (offset -54 lines). > >>>>>>>>>>>>>> Hunk #12 succeeded at 2478 (offset -1 lines). > >>>>>>>>>>>>>> Hunk #13 succeeded at 2684 (offset -54 lines). > >>>>>>>>>>>>>> Hunk #14 succeeded at 2821 (offset -1 lines). > >>>>>>>>>>>>>> 3 out of 14 hunks FAILED -- saving rejects to file > >>>>>>>>>>>>>> src/log/logd/lgs_imm.cc.rej > >>>>>>>>>>>>>> patching file src/log/logd/lgs_main.cc > >>>>>>>>>>>>>> patching file src/log/logd/lgs_mbcsv.cc > >>>>>>>>>>>>>> patching file src/log/logd/lgs_mbcsv.h > >>>>>>>>>>>>>> patching file src/log/logd/lgs_mbcsv_v5.cc > >>>>>>>>>>>>>> Hunk #3 succeeded at 133 (offset -1 lines). > >>>>>>>>>>>>>> patching file src/log/logd/lgs_mbcsv_v7.cc > >>>>>>>>>>>>>> patching file src/log/logd/lgs_mbcsv_v7.h > >>>>>>>>>>>>>> patching file src/log/logd/lgs_stream.cc > >>>>>>>>>>>>>> patching file src/log/logd/lgs_stream.h > >>>>>>>>>>>>>> patching file src/log/logd/lgs_util.cc > >>>>>>>>>>>>>> patching file src/log/logd/lgs_util.h > >>>>>>>>>>>>>> > >>>>>>>>>>>>>> [root@dhcp-hyd-scp-5fl-10-176-177-96 staging]# patch - > p1 > >>>>> <2258- > >>>>>>> 2 > >>>>>>>>>>>>>> patching file src/log/Makefile.am > >>>>>>>>>>>>>> Hunk #1 succeeded at 180 (offset -3 lines). > >>>>>>>>>>>>>> patching file src/log/apitest/tet_LogOiOps.c > >>>>>>>>>>>>>> Hunk #1 FAILED at 1923. > >>>>>>>>>>>>>> Hunk #2 FAILED at 1979. > >>>>>>>>>>>>>> Hunk #3 FAILED at 2067. > >>>>>>>>>>>>>> Hunk #4 FAILED at 2094. > >>>>>>>>>>>>>> 4 out of 4 hunks FAILED -- saving rejects to file > >>>>>>>>>>>>>> src/log/apitest/tet_LogOiOps.c.rej > >>>>>>>>>>>>>> patching file src/log/apitest/tet_cfg_destination.c > >>>>>>>>>>>>>> > >>>>>>>>>>>>>> [root@dhcp-hyd-scp-5fl-10-176-177-96 staging]# patch - > p1 > >>>>> <2258- > >>>>>>> 3 > >>>>>>>>>>>>>> patching file src/log/Makefile > >>>>>>>>>>>>>> patching file src/log/Makefile.am > >>>>>>>>>>>>>> Hunk #1 succeeded at 80 (offset -1 lines). > >>>>>>>>>>>>>> Hunk #2 succeeded at 217 (offset -2 lines). > >>>>>>>>>>>>>> patching file src/log/tests/Makefile > >>>>>>>>>>>>>> patching file src/log/tests/lgs_dest_test.cc > >>>>>>>>>>>>>> [root@dhcp-hyd-scp-5fl-10-176-177-96 staging]# vi > >>>>>>>>>>>>>> src/log/apitest/tet_LogOiOps.c.rej > >>>>>>>>>>>>>> [root@dhcp-hyd-scp-5fl-10-176-177-96 staging]# > >>>>>>>>>>>>>> > >>>>>>>>>>>>>> > >> > ============================================================== > >>>>>>>>>>>>>> ======================== > >>>>>>>>>>>>>> > >>>>>>>>>>>>>> -AVM > >>>>>>>>>>>>>> > >>>>>>>>>>>>>> On 2/21/2017 3:53 PM, Vu Minh Nguyen wrote: > >>>>>>>>>>>>>>> Hi Mahesh, > >>>>>>>>>>>>>>> > >>>>>>>>>>>>>>> As I has mentioned in below: > >>>>>>>>>>>>>>>>> To run the test, this patch has dependent on > >>>> following > >>>>>>>>> patches: > >>>>>>>>>>>>>>>>> 1) #2293 (sent by Anders Widel, but not yet > >>>>>>>>>>>>>>>>> pushed) > >>>>>>>>>>>>>>>>> 2) #2258 (v2, sent by Lennart, but not yet pushed > >>>> yet) > >>>>>>>>>>>>>>> So, you need to apply #2293 first, then #2258 which sent > by > >>>>>>> Lennart > >>>>>>>>>>>>>>> yesterday, then mine. > >>>>>>>>>>>>>>> > >>>>>>>>>>>>>>> Regards, Vu > >>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>> -----Original Message----- > >>>>>>>>>>>>>>>> From: A V Mahesh [mailto:mahesh.va...@oracle.com] > >>>>>>>>>>>>>>>> Sent: Tuesday, February 21, 2017 5:10 PM > >>>>>>>>>>>>>>>> To: Vu Minh Nguyen <vu.m.ngu...@dektech.com.au>; > >>>>>>>>>>>>>>>> lennart.l...@ericsson.com; > >> canh.v.tru...@dektech.com.au > >>>>>>>>>>>>>>>> Cc: opensaf-devel@lists.sourceforge.net > >>>>>>>>>>>>>>>> Subject: Re: [PATCH 0 of 3] Review Request for log: add > >>>>>>> alternative > >>>>>>>>>>>>>>>> destinations of log records [#2258] V4 > >>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>> Hi Vu, > >>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>> Is this applies on top of log #2146 - V4 , I see both > >>>>>> #tickets > >>>>>>>>>>>>> has > >>>>>>>>>>>>>>>> version changes ? > >>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>> in which order i need to apply ( #2146 & #2258 ) or > >>>>>>>>>>>>>>>> (#2258 > >>>> & > >>>>>>>>>>>>> #2146). > >> ========================================================= > >>>>>>>>>>>>>>>> patching file src/log/Makefile.am > >>>>>>>>>>>>>>>> Hunk #1 FAILED at 72. > >>>>>>>>>>>>>>>> Hunk #2 FAILED at 120. > >>>>>>>>>>>>>>>> 2 out of 2 hunks FAILED -- saving rejects to file > >>>>>>>>>>>>> src/log/Makefile.am.rej > >>>>>>>>>>>>>>>> patching file src/log/config/logsv_classes.xml > >>>>>>>>>>>>>>>> Hunk #1 FAILED at 147. > >>>>>>>>>>>>>>>> 1 out of 1 hunk FAILED -- saving rejects to file > >>>>>>>>>>>>>>>> src/log/config/logsv_classes.xml.rej > >>>>>>>>>>>>>>>> patching file src/log/logd/lgs_config.cc > >>>>>>>>>>>>>>>> Hunk #1 succeeded at 35 (offset -5 lines). > >>>>>>>>>>>>>>>> Hunk #2 FAILED at 705. > >>>>>>>>>>>>>>>> Hunk #3 FAILED at 971. > >>>>>>>>>>>>>>>> 2 out of 3 hunks FAILED -- saving rejects to file > >>>>>>>>>>>>>>>> src/log/logd/lgs_config.cc.rej > >>>>>>>>>>>>>>>> patching file src/log/logd/lgs_config.h > >>>>>>>>>>>>>>>> Hunk #1 FAILED at 304. > >>>>>>>>>>>>>>>> 1 out of 1 hunk FAILED -- saving rejects to file > >>>>>>>>>>>>>>>> src/log/logd/lgs_config.h.rej > >>>>>>>>>>>>>>>> patching file src/log/logd/lgs_dest.cc > >>>>>>>>>>>>>>>> patching file src/log/logd/lgs_dest.h > >>>>>>>>>>>>>>>> patching file src/log/logd/lgs_evt.cc > >>>>>>>>>>>>>>>> Hunk #1 FAILED at 1. > >>>>>>>>>>>>>>>> Hunk #2 succeeded at 30 with fuzz 2 (offset 2 lines). > >>>>>>>>>>>>>>>> Hunk #3 succeeded at 1282 (offset 45 lines). > >>>>>>>>>>>>>>>> Hunk #4 succeeded at 1300 (offset 2 lines). > >>>>>>>>>>>>>>>> 1 out of 4 hunks FAILED -- saving rejects to file > >>>>>>>>>>>>>>>> src/log/logd/lgs_evt.cc.rej > >>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>> > >> > ============================================================== > >>>>>>>>>>>>>>>> === > >>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>> -AVM > >>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>> On 2/21/2017 3:03 PM, Vu Minh Nguyen wrote: > >>>>>>>>>>>>>>>>> Summary: log: add alternative destinations of log > records > >>>>>>> [#2258] > >>>>>>>>>>>>>>>>> Review request for Trac Ticket(s): #2258 > >>>>>>>>>>>>>>>>> Peer Reviewer(s): Lennart, Canh, Mahesh > >>>>>>>>>>>>>>>>> Pull request to: <<LIST THE PERSON WITH PUSH > ACCESS > >>>>> HERE>> > >>>>>>>>>>>>>>>>> Affected branch(es): Default > >>>>>>>>>>>>>>>>> Development branch: Default > >>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>> -------------------------------- > >>>>>>>>>>>>>>>>> Impacted area Impact y/n > >>>>>>>>>>>>>>>>> -------------------------------- > >>>>>>>>>>>>>>>>> Docs n > >>>>>>>>>>>>>>>>> Build system n > >>>>>>>>>>>>>>>>> RPM/packaging n > >>>>>>>>>>>>>>>>> Configuration files n > >>>>>>>>>>>>>>>>> Startup scripts n > >>>>>>>>>>>>>>>>> SAF services n > >>>>>>>>>>>>>>>>> OpenSAF services y > >>>>>>>>>>>>>>>>> Core libraries n > >>>>>>>>>>>>>>>>> Samples n > >>>>>>>>>>>>>>>>> Tests y > >>>>>>>>>>>>>>>>> Other n > >>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>> Comments (indicate scope for each "y" above): > >>>>>>>>>>>>>>>>> --------------------------------------------- > >>>>>>>>>>>>>>>>> To run the test, this patch has dependent on > >>>> following > >>>>>>>>> patches: > >>>>>>>>>>>>>>>>> 1) #2293 (sent by Anders Widel, but not yet > >>>>>>>>>>>>>>>>> pushed) > >>>>>>>>>>>>>>>>> 2) #2258 (v2, sent by Lennart, but not yet pushed > >>>> yet) > >>>>>>>>>>>>>>>>> changeset > >> d74aaf3025c99cade3165a15831124548f4d85bd > >>>>>>>>>>>>>>>>> Author: Vu Minh Nguyen > >> <vu.m.ngu...@dektech.com.au> > >>>>>>>>>>>>>>>>> Date: Wed, 15 Feb 2017 14:36:00 +0700 > >>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>> log: add alternative destinations of log records > >>>>>>>>>>>>>>>>> [#2258] > >>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>> Here are major info, detailed info will be added to > >>>>>>>>>>>>>>>>> PR doc > >>>>>>>>>>> soon. 1) > >>>>>>>>>>>>>>>> Add > >>>>>>>>>>>>>>>>> attribute "saLogRecordDestination" to log stream. 2) > >>>>>>>>>>>>>>>>> Add > >>>>>>>>>>> Local > >>>>>>>>>>>>>>>> socket > >>>>>>>>>>>>>>>>> destintion handler 3) Integrate into first increment > >>>>>>>>>>>>>>>>> made by > >>>>>>>>>>> Lennart > >>>>>>>>>>>>>>>>> changeset > >> 4bae27a478c235df3058f43c92d3a5483233b01d > >>>>>>>>>>>>>>>>> Author: Vu Minh Nguyen > >> <vu.m.ngu...@dektech.com.au> > >>>>>>>>>>>>>>>>> Date: Wed, 15 Feb 2017 15:07:09 +0700 > >>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>> log: add UML test case to verify alternative > >>>>>>>>>>>>>>>>> destination > >>>>>>>>>>> [#2258] > >>>>>>>>>>>>>>>>> Major changes: 1) Modify Lennart's test cases > because > >>>>>>>>>>> enhancing > >>>>>>>>>>>>>>>> destination > >>>>>>>>>>>>>>>>> configuration validation rules. 2) Add test suite > >>>>>>>>>>>>>>>>> #17 to > >>>>>>>>>>> verify > >>>>>>>>>>>>>>>> alternative > >>>>>>>>>>>>>>>>> destination > >>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>> changeset > bc375725fed22bb4f8cb3ae3df5f96fb9d281efb > >>>>>>>>>>>>>>>>> Author: Vu Minh Nguyen > >> <vu.m.ngu...@dektech.com.au> > >>>>>>>>>>>>>>>>> Date: Thu, 16 Feb 2017 17:22:13 +0700 > >>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>> log: add unit tests to verify interfaces provided by > >>>>>>>>>>> destination > >>>>>>>>>>>>>>>> handler > >>>>>>>>>>>>>>>>> [#2258] > >>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>> Unit tests to verify major interfaces: 1) > >>>>>>>>>>>>>>>>> CfgDestination() > >>>>>>>>>>> 2) > >>>>>>>>>>>>>>>>> WriteToDestination() > >>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>> Added Files: > >>>>>>>>>>>>>>>>> ------------ > >>>>>>>>>>>>>>>>> src/log/apitest/tet_cfg_destination.c > >>>>>>>>>>>>>>>>> src/log/logd/lgs_dest.cc > >>>>>>>>>>>>>>>>> src/log/logd/lgs_dest.h > >>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv_v7.cc > >>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv_v7.h > >>>>>>>>>>>>>>>>> src/log/tests/lgs_dest_test.cc > >>>>>>>>>>>>>>>>> src/log/tests/Makefile > >>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>> Complete diffstat: > >>>>>>>>>>>>>>>>> ------------------ > >>>>>>>>>>>>>>>>> src/log/Makefile | 4 + > >>>>>>>>>>>>>>>>> src/log/Makefile.am | 31 +++++- > >>>>>>>>>>>>>>>>> src/log/apitest/tet_LogOiOps.c | 8 +- > >>>>>>>>>>>>>>>>> src/log/apitest/tet_cfg_destination.c | 483 > >> > ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++ > >>>>>>>>>>>>>>>> ++++++++++++++++++++++++++++++++++ > >>>>>>>>>>>>>>>>> src/log/config/logsv_classes.xml | 7 +- > >>>>>>>>>>>>>>>>> src/log/logd/lgs_config.cc | 169 > >>>>>>>>>>>>>>>> +++++++++++++++++++++++++++++--- > >>>>>>>>>>>>>>>>> src/log/logd/lgs_config.h | 3 +- > >>>>>>>>>>>>>>>>> src/log/logd/lgs_dest.cc | 707 > >> > ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++ > >> > ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++ > >>>>>>>>>>>>>>>> +++++++++++++++++ > >>>>>>>>>>>>>>>>> src/log/logd/lgs_dest.h | 576 > >> > ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++ > >>>>> ++++++++++++++++++++++++++++++++++++++++++++++++++++ > >>>>>>>>>>>>>>>>> src/log/logd/lgs_evt.cc | 33 ++++++ > >>>>>>>>>>>>>>>>> src/log/logd/lgs_imm.cc | 202 > >>>>>>>>>>>>>>>> +++++++++++++++++++++++++++++++++------ > >>>>>>>>>>>>>>>>> src/log/logd/lgs_main.cc | 8 + > >>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc | 103 > >>>>>>>>> ++++++++++++++++++- > >>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv.h | 6 +- > >>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv_v5.cc | 10 + > >>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv_v7.cc | 177 > >>>>>>>>>>>>>>>> +++++++++++++++++++++++++++++++++++ > >>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv_v7.h | 67 > >>>>>> +++++++++++++ > >>>>>>>>>>>>>>>>> src/log/logd/lgs_stream.cc | 60 > >>>>>> +++++++++++- > >>>>>>>>>>>>>>>>> src/log/logd/lgs_stream.h | 16 +++ > >>>>>>>>>>>>>>>>> src/log/logd/lgs_util.cc | 63 > >>>>>> ++++++++++++ > >>>>>>>>>>>>>>>>> src/log/logd/lgs_util.h | 11 +- > >>>>>>>>>>>>>>>>> src/log/tests/Makefile | 20 +++ > >>>>>>>>>>>>>>>>> src/log/tests/lgs_dest_test.cc | 209 > >>>>>>>>>>>>>>>> +++++++++++++++++++++++++++++++++++++++++ > >>>>>>>>>>>>>>>>> 23 files changed, 2896 insertions(+), 77 > >>>>>>>>>>>>>>>>> deletions(-) > >>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>> Testing Commands: > >>>>>>>>>>>>>>>>> ----------------- > >>>>>>>>>>>>>>>>> Run UML test suite #17 > >>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>> Testing, Expected Results: > >>>>>>>>>>>>>>>>> -------------------------- > >>>>>>>>>>>>>>>>> All test passed > >>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>> Conditions of Submission: > >>>>>>>>>>>>>>>>> ------------------------- > >>>>>>>>>>>>>>>>> <<HOW MANY DAYS BEFORE PUSHING, > CONSENSUS > >>>>> ETC>> > >>>>>>>>>>>>>>>>> Arch Built Started Linux distro > >>>>>>>>>>>>>>>>> ------------------------------------------- > >>>>>>>>>>>>>>>>> mips n n > >>>>>>>>>>>>>>>>> mips64 n n > >>>>>>>>>>>>>>>>> x86 n n > >>>>>>>>>>>>>>>>> x86_64 n n > >>>>>>>>>>>>>>>>> powerpc n n > >>>>>>>>>>>>>>>>> powerpc64 n n > >>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>> Reviewer Checklist: > >>>>>>>>>>>>>>>>> ------------------- > >>>>>>>>>>>>>>>>> [Submitters: make sure that your review doesn't trigger > >> any > >>>>>>>>>>>>>> checkmarks!] > >>>>>>>>>>>>>>>>> Your checkin has not passed review because (see > checked > >>>>>>> entries): > >>>>>>>>>>>>>>>>> ___ Your RR template is generally incomplete; it has > too > >>>>> many > >>>>>>>>> blank > >>>>>>>>>>>>>>> entries > >>>>>>>>>>>>>>>>> that need proper data filled in. > >>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>> ___ You have failed to nominate the proper persons for > >>>>> review > >>>>>>> and > >>>>>>>>>>>>> push. > >>>>>>>>>>>>>>>>> ___ Your patches do not have proper short+long > header > >>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>> ___ You have grammar/spelling in your header that is > >>>>>>> unacceptable. > >>>>>>>>>>>>>>>>> ___ You have exceeded a sensible line length in your > >>>>>>>>>>>>>>>> headers/comments/text. > >>>>>>>>>>>>>>>>> ___ You have failed to put in a proper Trac Ticket # into > >>>> your > >>>>>>>>>>>>> commits. > >>>>>>>>>>>>>>>>> ___ You have incorrectly put/left internal data in your > >>>>>>>>>>> comments/files > >>>>>>>>>>>>>>>>> (i.e. internal bug tracking tool IDs, product > >>>> names > >>>>>> etc) > >>>>>>>>>>>>>>>>> ___ You have not given any evidence of testing beyond > >> basic > >>>>>>> build > >>>>>>>>>>>>> tests. > >>>>>>>>>>>>>>>>> Demonstrate some level of runtime or other sanity > >>>>>> testing. > >>>>>>>>>>>>>>>>> ___ You have ^M present in some of your files. These > >>>>>>>>>>>>>>>>> have to > >>>>>>> be > >>>>>>>>>>>>>> removed. > >>>>>>>>>>>>>>>>> ___ You have needlessly changed whitespace or added > >>>>>>> whitespace > >>>>>>>>>>>>>> crimes > >>>>>>>>>>>>>>>>> like trailing spaces, or spaces before tabs. > >>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>> ___ You have mixed real technical changes with > >> whitespace > >>>>> and > >>>>>>>>>> other > >>>>>>>>>>>>>>>>> cosmetic code cleanup changes. These have to be > >>>>>>> separate > >>>>>>>>>>>>> commits. > >>>>>>>>>>>>>>>>> ___ You need to refactor your submission into logical > >>>>>>>>>>>>>>>>> chunks; > >>>>>>>>> there > >>>>>>>>>>> is > >>>>>>>>>>>>>>>>> too much content into a single commit. > >>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>> ___ You have extraneous garbage in your review > (merge > >>>>>>> commits > >>>>>>>>>> etc) > >>>>>>>>>>>>>>>>> ___ You have giant attachments which should never > have > >>>>> been > >>>>>>>>>> sent; > >>>>>>>>>>>>>>>>> Instead you should place your content in a public > >>>>>> tree to > >>>>>>>>> be > >>>>>>>>>>>>>>> pulled. > >>>>>>>>>>>>>>>>> ___ You have too many commits attached to an e-mail; > >>>>> resend > >>>>>>> as > >>>>>>>>>>>>>> threaded > >>>>>>>>>>>>>>>>> commits, or place in a public tree for a pull. > >>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>> ___ You have resent this content multiple times > without a > >>>>> clear > >>>>>>>>>>>>>>> indication > >>>>>>>>>>>>>>>>> of what has changed between each re-send. > >>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>> ___ You have failed to adequately and individually > address > >>>> all > >>>>>> of > >>>>>>>>>>> the > >>>>>>>>>>>>>>>>> comments and change requests that were proposed in > >>>>>> the > >>>>>>>>>>> initial > >>>>>>>>>>>>>>> review. > >>>>>>>>>>>>>>>>> ___ You have a misconfigured ~/.hgrc file (i.e. > username, > >>>>> email > >>>>>>>>> etc) > >>>>>>>>>>>>>>>>> ___ Your computer have a badly configured date and > >> time; > >>>>>>>>>> confusing > >>>>>>>>>>>> the > >>>>>>>>>>>>>>>>> the threaded patch review. > >>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>> ___ Your changes affect IPC mechanism, and you don't > >>>>> present > >>>>>>> any > >>>>>>>>>>>>>> results > >>>>>>>>>>>>>>>>> for in-service upgradability test. > >>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>> ___ Your changes affect user manual and > documentation, > >>>>> your > >>>>>>>>>> patch > >>>>>>>>>>>>>> series > >>>>>>>>>>>>>>>>> do not contain the patch that updates the > >>>>>>>>>>>>>>>>> Doxygen > >>>>>>> manual. > > ---------------------------------------------------------------------------- > >>>>>> -- > >>>>>>>> Check out the vibrant tech community on one of the world's most > >>>>>>>> engaging tech sites, SlashDot.org! http://sdm.link/slashdot > >>>>>>>> _______________________________________________ > >>>>>>>> Opensaf-devel mailing list > >>>>>>>> Opensaf-devel@lists.sourceforge.net > >>>>>>>> https://lists.sourceforge.net/lists/listinfo/opensaf-devel ------------------------------------------------------------------------------ Check out the vibrant tech community on one of the world's most engaging tech sites, SlashDot.org! http://sdm.link/slashdot _______________________________________________ Opensaf-devel mailing list Opensaf-devel@lists.sourceforge.net https://lists.sourceforge.net/lists/listinfo/opensaf-devel