Hi Vu, Not sure what are other change compare to V7 to V9 , New problems got introduced
Both nodes SC-1 & SC-2 ( with 2258_v9.patch ) , trying bring up both SC`s simple node bringup , SC-2 going for reboot with following : ======================================================================================================================================== Feb 23 15:05:32 SC-2 osaflogd[29988]: NO LOGSV_DATA_GROUPNAME not found Feb 23 15:05:32 SC-2 osaflogd[29988]: NO LOG root directory is: "/var/log/opensaf/saflog" Feb 23 15:05:32 SC-2 osaflogd[29988]: NO LOG data group is: "" Feb 23 15:05:32 SC-2 osafimmnd[29978]: NO Implementer (applier) connected: 16 (@safAmfService2020f) <127, 2020f> Feb 23 15:05:32 SC-2 osaflogd[29988]: NO LGS_MBCSV_VERSION = 7 Feb 23 15:05:32 SC-2 osaflogd[29988]: WA FAILED: ncs_patricia_tree_add, client_id 0 Feb 23 15:05:32 SC-2 osafamfnd[30028]: NO Assigned 'safSi=SC-2N,safApp=OpenSAF' STANDBY to 'safSu=SC-2,safSg=2N,safApp=OpenSAF' Feb 23 15:05:32 SC-2 osaflogd[29988]: ER Exiting with message: Could not create new client Feb 23 15:05:32 SC-2 osafamfnd[30028]: NO 'safSu=SC-2,safSg=2N,safApp=OpenSAF' component restart probation timer started (timeout: 60000000000 ns) Feb 23 15:05:32 SC-2 osafamfnd[30028]: NO Restarting a component of 'safSu=SC-2,safSg=2N,safApp=OpenSAF' (comp restart count: 1) Feb 23 15:05:32 SC-2 osafamfnd[30028]: NO 'safComp=LOG,safSu=SC-2,safSg=2N,safApp=OpenSAF' faulted due to 'errorReport' : Recovery is 'componentRestart' Feb 23 15:05:32 SC-2 opensafd[29908]: ER Service LOGD has unexpectedly crashed. Unable to continue, exiting Feb 23 15:05:32 SC-2 osafamfd[30018]: exiting for shutdown Feb 23 15:05:32 SC-2 osafamfnd[30028]: ER AMFD has unexpectedly crashed. Rebooting node Feb 23 15:05:32 SC-2 osafamfnd[30028]: Rebooting OpenSAF NodeId = 131599 EE Name = , Reason: AMFD has unexpectedly crashed. Rebooting node, OwnNodeId = 131599, SupervisionTime = 60 Feb 23 15:05:32 SC-2 opensaf_reboot: Rebooting local node; timeout=60 Feb 23 15:06:04 SC-2 syslog-ng[1180]: syslog-ng starting up; version='2.0.9' ======================================================================================================================================== -AVM On 2/23/2017 2:20 PM, Vu Minh Nguyen wrote: > Hi Mahesh, > > This is the latest code has been rebased on the latest changeset. > > Note that, in the attached patch, I have included one more dependency, > that is on base::Hash() function, the patch sent by Anders [#2266] > > Please review the patch, then comment if any. Thanks. > > Regards, Vu > >> -----Original Message----- >> From: A V Mahesh [mailto:mahesh.va...@oracle.com] >> Sent: Thursday, February 23, 2017 2:03 PM >> To: Vu Minh Nguyen <vu.m.ngu...@dektech.com.au>; >> lennart.l...@ericsson.com; canh.v.tru...@dektech.com.au >> Cc: opensaf-devel@lists.sourceforge.net >> Subject: Re: [devel] [PATCH 0 of 3] Review Request for log: add > alternative >> destinations of log records [#2258] V4 >> >> Hi Vu, >> >> Now we are now able to proceed further with V7 `2258_v7.patch` in >> service upgraded working fine, >> because of Encode/decode changes done in V7 patch. >> >> But we have another small test case issue (/usr/bin/logtest 5 17 >> Segmentation fault), >> once we resolve this also, we can conclude that all the basic >> functionality is working, >> then you can re-publish the V7 patch ( if change occurred in Lennart >> #2258 V2 please do publish that as well ) >> so that I can go for CODE review. >> >> Steps to reproduce the test case issue : >> >> 1) Bring up old node as Active ( with out `2258_v7.patch` ) >> 2) Bring-up new node as Standby ( with `2258_v7.patch` ) >> 3) Do `amf-adm si-swap safSi=SC-2N,safApp=OpenSAF` >> 4) Run `/usr/bin/logtest 5 17 ` on new Active (because of si-swap ) >> >> Note : both nodes has the new XLM attributes populated . >> >> ============================================================== >> ===================== >> >> gdb /usr/bin/logtest >> (gdb) r 5 >> >> 16 PASSED CCB Object Modify, change root directory. Path exist. OK; >> Detaching after fork from child process 13797. >> Set values Fail >> [New Thread 0x7ffff7ff7b00 (LWP 13801)] >> [New Thread 0x7ffff7fc4b00 (LWP 13802)] >> >> Program received signal SIGSEGV, Segmentation fault. >> 0x00005555555688ea in read_and_compare.isra.7 () at >> src/log/apitest/tet_LogOiOps.c:1891 >> 1891 src/log/apitest/tet_LogOiOps.c: No such file or directory. >> in src/log/apitest/tet_LogOiOps.c >> (gdb) bt >> #0 0x00005555555688ea in read_and_compare.isra.7 () at >> src/log/apitest/tet_LogOiOps.c:1891 >> #1 0x0000555555568a4b in check_logRecordDestinationConfigurationAdd () >> at src/log/apitest/tet_LogOiOps.c:1941 >> #2 0x0000555555571b05 in run_test_case () >> #3 0x0000555555571feb in test_run () >> #4 0x000055555555bfad in main () at src/log/apitest/logtest.c:569 >> (gdb) >> >> ============================================================== >> ===================== >> >> >> -AVM >> >> On 2/23/2017 11:44 AM, Vu Minh Nguyen wrote: >>> Hi Mahesh, >>> >>> Maybe it was broken when transmitting. I zipped to a tar file. Please > try it >>> one more. >>> >>> Regards, Vu >>> >>> >>>> -----Original Message----- >>>> From: A V Mahesh [mailto:mahesh.va...@oracle.com] >>>> Sent: Thursday, February 23, 2017 12:54 PM >>>> To: Vu Minh Nguyen <vu.m.ngu...@dektech.com.au>; >>>> lennart.l...@ericsson.com; canh.v.tru...@dektech.com.au >>>> Cc: opensaf-devel@lists.sourceforge.net >>>> Subject: Re: [devel] [PATCH 0 of 3] Review Request for log: add >>> alternative >>>> destinations of log records [#2258] V4 >>>> >>>> Hi Vu, >>>> >>>> On 2/23/2017 10:20 AM, Vu Minh Nguyen wrote: >>>>> Hi Mahesh, >>>>> >>>>> Can you try with 2258_v7.patch I just sent to you? >>>> I striped changeset: 8610 of today's latest staging ( `hg strip >>>> 8610` which removed log: implement SaLogFilterSetCallbackT and version >>>> handling [#2146]) >>>> and try to apply your `2258_v7.patch`, it says `malformed patch at line >>>> 3324`. >>>> >>>> -AVM >>>>> I have pulled the latest code on OpenSAF 5.1 branch, re-created the >>> cluster. >>>>> And it works with the case old active SC-1 (OpenSAF 5.1) and new >> standby >>>>> SC-2 (with 2258_v7.patch included in). >>>>> >>>>> To apply 2258_v7.patch, please do remove the just pushed ticket "log: >>>>> implement SaLogFilterSetCallbackT and version handling [#2146]" , >>>>> I have not rebased the code on that yet. >>>>> >>>>> Regards, Vu >>>>> >>>>>> -----Original Message----- >>>>>> From: A V Mahesh [mailto:mahesh.va...@oracle.com] >>>>>> Sent: Thursday, February 23, 2017 11:45 AM >>>>>> To: Vu Minh Nguyen <vu.m.ngu...@dektech.com.au>; >>>>>> lennart.l...@ericsson.com; canh.v.tru...@dektech.com.au >>>>>> Cc: opensaf-devel@lists.sourceforge.net >>>>>> Subject: Re: [devel] [PATCH 0 of 3] Review Request for log: add >>>>> alternative >>>>>> destinations of log records [#2258] V4 >>>>>> >>>>>> Hi Vu/Lennart, >>>>>> >>>>>> >>>>>> In broad WITHOUT the #2258 patch, the same code/setup working fine >>>> with >>>>>> 2 sc node (staging changeset: 8609 ), >>>>>> as soon as we apply `2258_v5.patch` V5 patch on staging (changeset: >>>>>> 8609 ) that you have provided yesterday, >>>>>> on one sc node and try to bring up that in to cluster (in-service > test) >>>>>> we are observing the issue of new node (with #2258 patch) not joining >>>>>> cluster. >>>>>> >>>>>> >> ============================================================== >>>>>> ==================================================== >>>>>> eb 23 10:01:59 SC-1 osafimmnd[15279]: NO Implementer (applier) >>>>>> connected: 15 (@safAmfService2010f) <127, 2010f> >>>>>> Feb 23 10:01:59 SC-1 osaflogd[15289]: NO LOGSV_DATA_GROUPNAME >>>> not >>>>>> found >>>>>> Feb 23 10:01:59 SC-1 osaflogd[15289]: NO LOG root directory is: >>>>>> "/var/log/opensaf/saflog" >>>>>> Feb 23 10:01:59 SC-1 osaflogd[15289]: NO LOG data group is: "" >>>>>> Feb 23 10:01:59 SC-1 osaflogd[15289]: NO LGS_MBCSV_VERSION = 7 >>>>>> Feb 23 10:01:59 SC-1 osafamfnd[15329]: NO Assigned >>>>>> 'safSi=SC-2N,safApp=OpenSAF' STANDBY to 'safSu=SC- >>>>>> 1,safSg=2N,safApp=OpenSAF' >>>>>> Feb 23 10:01:59 SC-1 opensafd: OpenSAF(5.1.M0 - ) services >> successfully >>>>>> started >>>>>> Feb 23 10:01:59 SC-1 osafamfnd[15329]: NO >>>>>> 'safComp=LOG,safSu=SC-1,safSg=2N,safApp=OpenSAF' faulted due to >>>>>> 'avaDown' : Recovery is 'nodeFailfast' >>>>>> Feb 23 10:01:59 SC-1 osafamfnd[15329]: ER >>>>>> safComp=LOG,safSu=SC-1,safSg=2N,safApp=OpenSAF Faulted due >>>>>> to:avaDown >>>>>> Recovery is:nodeFailfast >>>>>> Feb 23 10:01:59 SC-1 osafamfnd[15329]: Rebooting OpenSAF NodeId = >>>>>> 131343 >>>>>> EE Name = , Reason: Component faulted: recovery is node failfast, >>>>>> OwnNodeId = 131343, SupervisionTime = 60 >>>>>> Feb 23 10:01:59 SC-1 opensaf_reboot: Rebooting local node; >> timeout=60 >>>>>> Feb 23 10:02:00 SC-1 osafimmnd[15279]: NO Implementer (applier) >>>>>> connected: 16 (@OpenSafImmReplicatorB) <144, 2010f> >>>>>> Feb 23 10:01:59 SC-1 opensaf_reboot: Rebooting local node; >> timeout=60 >> ============================================================== >>>>>> ==================================================== >>>>>> >>>>>> So it is evident that in-service upgrade part code of this need to be >>>>>> corrected. >>>>>> >>>>>> Please see my comments as [AVM] and let me know if you need some >>>> traces >>>>>> . >>>>>> >>>>>> If you're planing to prepare new V6 patch , please do prepare on top > of >>>>>> today's latest staging. >>>>>> >>>>>> On 2/23/2017 9:33 AM, Vu Minh Nguyen wrote: >>>>>>> Hi Mahesh, >>>>>>> >>>>>>> I have done in-service upgrade/downgrade with following cases: >>>>>>> 1) New Active SC-1 (OpenSAF 5.2 with the attached patch) + old >> standby >>>>>> SC-2 >>>>>>> (OpenSAF 5.1) >>>>>>> --> Work fine >>>>>> [AVM] This is not a practical use cause of in-service upgrade , we > can >>>>>> ignore this test further >>>>>>> 2) Old Active SC-1 (OpenSAF 5.1) + new standby SC-2 (with or without >>>>>>> attached patch) >>>>>>> --> SC-2 is restarted & not able to join the cluster. >>>>>> [AVM] This use cause/flow is we do get in in-service upgrade , so we >>>>>> need to address this. >>>>>>> I got following messages in syslog: >>>>>>> Feb 23 09:32:42 SC-2 user.notice opensafd: OpenSAF(5.2.M0 - >>>>>>> 8529:b5addd36e45d:default) services successfully started >>>>>>> Feb 23 09:32:43 SC-2 local0.warn osafntfimcnd[701]: WA >>>>>> ntfimcn_imm_init >>>>>>> saImmOiImplementerSet() returned SA_AIS_ERR_TIMEOUT (5) >>>>>>> Feb 23 09:32:45 SC-2 local0.warn osafntfimcnd[701]: WA >>>>>> ntfimcn_imm_init >>>>>>> saImmOiImplementerSet() returned SA_AIS_ERR_TIMEOUT (5) >>>>>>> Feb 23 09:32:47 SC-2 local0.warn osafntfimcnd[701]: WA >>>>>> ntfimcn_imm_init >>>>>>> saImmOiImplementerSet() returned SA_AIS_ERR_TIMEOUT (5) >>>>>>> Feb 23 09:32:49 SC-2 local0.warn osafntfimcnd[701]: WA >>>>>> ntfimcn_imm_init >>>>>>> saImmOiImplementerSet() returned SA_AIS_ERR_TIMEOUT (5) >>>>>>> Feb 23 09:32:50 SC-2 local0.err osafmsgnd[592]: ER >>>>>> saImmOiImplementerSet >>>>>>> FAILED:5 >>>>>>> Feb 23 09:32:50 SC-2 local0.err osafmsgnd[592]: ER >>>>>> saImmOiImplementerSet >>>>>>> FAILED:5 >>>>>>> Feb 23 09:32:50 SC-2 local0.notice osafamfnd[496]: NO >>>>>>> 'safSu=SC-2,safSg=NoRed,safApp=OpenSAF' component restart >>>> probation >>>>>> timer >>>>>>> started (timeout: 60000000000 ns) >>>>>>> Feb 23 09:32:50 SC-2 local0.notice osafamfnd[496]: NO Restarting a >>>>>> component >>>>>>> of 'safSu=SC-2,safSg=NoRed,safApp=OpenSAF' (comp restart count: 1) >>>>>>> Feb 23 09:32:50 SC-2 local0.notice osafamfnd[496]: NO >>>>>>> 'safComp=MQND,safSu=SC-2,safSg=NoRed,safApp=OpenSAF' faulted >>>> due >>>>>> to >>>>>>> 'avaDown' : Recovery is 'componentRestart' >>>>>>> Feb 23 09:32:50 SC-2 local0.info osafmsgnd[736]: mkfifo already >>> exists: >>>>>>> /var/lib/opensaf/osafmsgnd.fifo File exists >>>>>>> >>>>>>> And sometimes, on active SC-1 (OpenSAF 5.1), the node is not able to >>> up >>>>>>> because of following error: >>>>>>> >>>>>>> Feb 23 11:00:32 SC-1 local0.err osafclmna[406]: MDTM:TIPC Dsock >>>> Socket >>>>>>> creation failed in MDTM_INIT err :Address family not supported by >>>>>> protocol >>>>>>> Feb 23 11:00:32 SC-1 local0.err osafclmna[406]: ER >> ncs_agents_startup >>>>>> FAILED >>>>>> [AVM] No such issues ( with both TCP & TIPC) (staging changeset: >>> 8609 >>>>> ) >>>>>>> Are you getting similar problem at your side? >>>>>>> Please note that, the problem is existed WITH or WITHOUT the #2258 >>>>>> patch. >>>>>> [AVM] No , problem only if we apply `2258_v5.patch` V5 patch on >> staging >>>>>> (changeset: 8609 ) >>>>>> try to bring up that node in to cluster. >>>>>> >>>>>> >>>>>> -AVM >>>>>> >>>>>>> I have informed this to IMM to have a look, not sure any problem > with >>>>> MDS >>>>>>> layer or any problem with my environment setup. >>>>>>> In the meantime, please have a look at the updated patch, I will >>>>> continue >>>>>>> checking the problem. Will keep you updated. >>>>>> [AVM] I haven't seen any IMM problems >>>>>>> Regards, Vu >>>>>>> >>>>>>>> -----Original Message----- >>>>>>>> From: A V Mahesh [mailto:mahesh.va...@oracle.com] >>>>>>>> Sent: Wednesday, February 22, 2017 5:36 PM >>>>>>>> To: Vu Minh Nguyen <vu.m.ngu...@dektech.com.au>; >>>>>>>> lennart.l...@ericsson.com; canh.v.tru...@dektech.com.au >>>>>>>> Cc: opensaf-devel@lists.sourceforge.net >>>>>>>> Subject: Re: [devel] [PATCH 0 of 3] Review Request for log: add >>>>>>> alternative >>>>>>>> destinations of log records [#2258] V4 >>>>>>>> >>>>>>>> Please see correction New Standby SC-1 ( with patch ) >>>>>>>> >>>>>>>> -AVM >>>>>>>> >>>>>>>> On 2/22/2017 4:02 PM, A V Mahesh wrote: >>>>>>>>> Hi Vu, >>>>>>>>> >>>>>>>>> With this new patch , we have another issue : >>>>>>>>> >>>>>>>>> 1) standby Core by `/usr/lib64/opensaf/osaflogd' issue got >>> resolved >>>>> . >>>>>>>>> 2) In-service upgrade is Not working , I have Old Active SC-2 ( >>> with >>>>>>>>> out patch ) and New Standby SC-1 ( with patch ) >>>>>>>>> >>>>>>>>> the new New Standby SC-1 not joining the cluster ( > in-service >>>>>>>>> upgrade failed ) >>>>>>>>> >>>>>>>>> New Standby SC-1 >>>>>>>>> >>>>>>>>> >> ============================================================== >>>>>>>> ====================================================== >>>>>>>>> Feb 22 15:53:05 SC-1 osafamfnd[15752]: NO >>>>>>>>> 'safSu=SC-1,safSg=NoRed,safApp=OpenSAF' Presence State >>>>>>>> INSTANTIATING >>>>>>>>> => INSTANTIATED >>>>>>>>> Feb 22 15:53:05 SC-1 osafamfnd[15752]: NO Assigning >>>>>>>>> 'safSi=NoRed4,safApp=OpenSAF' ACTIVE to >>>>>>>>> 'safSu=SC-1,safSg=NoRed,safApp=OpenSAF' >>>>>>>>> Feb 22 15:53:05 SC-1 osafamfnd[15752]: NO Assigned >>>>>>>>> 'safSi=NoRed4,safApp=OpenSAF' ACTIVE to >>>>>>>>> 'safSu=SC-1,safSg=NoRed,safApp=OpenSAF' >>>>>>>>> Feb 22 15:53:05 SC-1 osafsmfd[15889]: Started >>>>>>>>> Feb 22 15:53:05 SC-1 osafamfnd[15752]: NO >>>>>>>>> 'safSu=SC-1,safSg=2N,safApp=OpenSAF' Presence State >>>> INSTANTIATING >>>>>> => >>>>>>>>> INSTANTIATED >>>>>>>>> Feb 22 15:53:05 SC-1 osafamfnd[15752]: NO Assigning >>>>>>>>> 'safSi=SC-2N,safApp=OpenSAF' STANDBY to >>>>>>>>> 'safSu=SC-1,safSg=2N,safApp=OpenSAF' >>>>>>>>> Feb 22 15:53:05 SC-1 osafrded[15672]: NO RDE role set to >> STANDBY >>>>>>>>> Feb 22 15:53:05 SC-1 osafrded[15672]: NO Peer up on node >> 0x2020f >>>>>>>>> Feb 22 15:53:05 SC-1 osafrded[15672]: NO Got peer info request >> from >>>>>>>>> node 0x2020f with role ACTIVE >>>>>>>>> Feb 22 15:53:05 SC-1 osafrded[15672]: NO Got peer info response >>>> from >>>>>>>>> node 0x2020f with role ACTIVE >>>>>>>>> Feb 22 15:53:05 SC-1 osafimmd[15691]: NO MDS event from svc_id >> 24 >>>>>>>>> (change:5, dest:13) >>>>>>>>> Feb 22 15:53:05 SC-1 osafimmd[15691]: NO MDS event from svc_id >> 24 >>>>>>>>> (change:3, dest:13) >>>>>>>>> Feb 22 15:53:05 SC-1 osafimmd[15691]: NO MDS event from svc_id >> 24 >>>>>>>>> (change:5, dest:13) >>>>>>>>> Feb 22 15:53:05 SC-1 osafimmd[15691]: NO MDS event from svc_id >> 25 >>>>>>>>> (change:3, dest:567412424453430) >>>>>>>>> Feb 22 15:53:05 SC-1 osafimmd[15691]: NO MDS event from svc_id >> 25 >>>>>>>>> (change:3, dest:565213401202663) >>>>>>>>> Feb 22 15:53:05 SC-1 osafimmd[15691]: NO MDS event from svc_id >> 25 >>>>>>>>> (change:3, dest:566312912825221) >>>>>>>>> Feb 22 15:53:05 SC-1 osafimmd[15691]: NO MDS event from svc_id >> 25 >>>>>>>>> (change:3, dest:564113889574230) >>>>>>>>> Feb 22 15:53:05 SC-1 osafimmnd[15702]: NO Implementer (applier) >>>>>>>>> connected: 17 (@safAmfService2010f) <127, 2010f> >>>>>>>>> Feb 22 15:53:05 SC-1 osaflogd[15712]: NO >>>> LOGSV_DATA_GROUPNAME >>>>>>>> not found >>>>>>>>> Feb 22 15:53:05 SC-1 osaflogd[15712]: NO LOG root directory is: >>>>>>>>> "/var/log/opensaf/saflog" >>>>>>>>> Feb 22 15:53:05 SC-1 osaflogd[15712]: NO LOG data group is: "" >>>>>>>>> Feb 22 15:53:05 SC-1 osaflogd[15712]: NO LGS_MBCSV_VERSION = >> 7 >>>>>>>>> Feb 22 15:53:05 SC-1 osafamfnd[15752]: NO Assigned >>>>>>>>> 'safSi=SC-2N,safApp=OpenSAF' STANDBY to >>>>>>>>> 'safSu=SC-1,safSg=2N,safApp=OpenSAF' >>>>>>>>> Feb 22 15:53:05 SC-1 opensafd: OpenSAF(5.1.M0 - ) services >>>>>>>>> successfully started >>>>>>>>> Feb 22 15:53:05 SC-1 osafamfnd[15752]: NO >>>>>>>>> 'safComp=LOG,safSu=SC-1,safSg=2N,safApp=OpenSAF' faulted due >> to >>>>>>>>> 'avaDown' : Recovery is 'nodeFailfast' >>>>>>>>> Feb 22 15:53:05 SC-1 osafamfnd[15752]: ER >>>>>>>>> safComp=LOG,safSu=SC-1,safSg=2N,safApp=OpenSAF Faulted due >>>>>>>> to:avaDown >>>>>>>>> Recovery is:nodeFailfast >>>>>>>>> Feb 22 15:53:05 SC-1 osafamfnd[15752]: Rebooting OpenSAF >> NodeId >>>> = >>>>>>>>> 131343 EE Name = , Reason: Component faulted: recovery is node >>>>>>>>> failfast, OwnNodeId = 131343, SupervisionTime = 60 >>>>>>>>> Feb 22 15:53:05 SC-1 opensaf_reboot: Rebooting local node; >>>>>> timeout=60 >>>>>>>>> Feb 22 15:53:43 SC-1 syslog-ng[1171]: syslog-ng starting up; >>>>>>>>> version='2.0.9' >>>>>>>>> >>>>>>>>> >> ============================================================== >>>>>>>> ====================================================== >>>>>>>>> Old - Active - SC-2 >>>>>>>>> >>>>>>>>> >> ============================================================== >>>>>>>> ====================================================== >>>>>>>>> Feb 22 15:53:02 SC-2 osafimmnd[16359]: NO NODE STATE-> >>>>>>>>> IMM_NODE_R_AVAILABLE >>>>>>>>> Feb 22 15:53:02 SC-2 osafimmloadd: NO Sync starting >>>>>>>>> Feb 22 15:53:02 SC-2 osafimmloadd: IN Synced 390 objects in total >>>>>>>>> Feb 22 15:53:02 SC-2 osafimmnd[16359]: NO NODE STATE-> >>>>>>>>> IMM_NODE_FULLY_AVAILABLE 18511 >>>>>>>>> Feb 22 15:53:02 SC-2 osafimmnd[16359]: NO Epoch set to 3 in >>>>>> ImmModel >>>>>>>>> Feb 22 15:53:02 SC-2 osafimmd[16346]: NO ACT: New Epoch for >>>>>> IMMND >>>>>>>>> process at node 2020f old epoch: 2 new epoch:3 >>>>>>>>> Feb 22 15:53:02 SC-2 osafimmd[16346]: NO ACT: New Epoch for >>>>>> IMMND >>>>>>>>> process at node 2040f old epoch: 2 new epoch:3 >>>>>>>>> Feb 22 15:53:02 SC-2 osafimmd[16346]: NO ACT: New Epoch for >>>>>> IMMND >>>>>>>>> process at node 2030f old epoch: 2 new epoch:3 >>>>>>>>> Feb 22 15:53:02 SC-2 osafimmloadd: NO Sync ending normally >>>>>>>>> Feb 22 15:53:02 SC-2 osafimmd[16346]: NO ACT: New Epoch for >>>>>> IMMND >>>>>>>>> process at node 2010f old epoch: 0 new epoch:3 >>>>>>>>> Feb 22 15:53:02 SC-2 osafimmnd[16359]: NO SERVER STATE: >>>>>>>>> IMM_SERVER_SYNC_SERVER --> IMM_SERVER_READY >>>>>>>>> Feb 22 15:53:03 SC-2 osafamfd[16408]: NO Received node_up from >>>>>> 2010f: >>>>>>>>> msg_id 1 >>>>>>>>> Feb 22 15:53:03 SC-2 osafamfd[16408]: NO Node 'SC-1' joined the >>>>>> cluster >>>>>>>>> Feb 22 15:53:03 SC-2 osafimmnd[16359]: NO Implementer >> connected: >>>>>> 16 >>>>>>>>> (MsgQueueService131343) <0, 2010f> >>>>>>>>> Feb 22 15:53:03 SC-2 osafrded[16327]: NO Peer up on node >> 0x2010f >>>>>>>>> Feb 22 15:53:03 SC-2 osafrded[16327]: NO Got peer info request >> from >>>>>>>>> node 0x2010f with role STANDBY >>>>>>>>> Feb 22 15:53:03 SC-2 osafrded[16327]: NO Got peer info response >>>> from >>>>>>>>> node 0x2010f with role STANDBY >>>>>>>>> Feb 22 15:53:03 SC-2 osafimmd[16346]: NO MDS event from svc_id >> 24 >>>>>>>>> (change:5, dest:13) >>>>>>>>> Feb 22 15:53:03 SC-2 osafimmnd[16359]: NO Implementer (applier) >>>>>>>>> connected: 17 (@safAmfService2010f) <0, 2010f> >>>>>>>>> Feb 22 15:53:03 SC-2 osafamfd[16408]: NO Cluster startup is done >>>>>>>>> Feb 22 15:53:04 SC-2 osafimmnd[16359]: NO Implementer (applier) >>>>>>>>> connected: 18 (@OpenSafImmReplicatorB) <0, 2010f> >>>>>>>>> Feb 22 15:53:09 SC-2 osafdtmd[16304]: NO Lost contact with 'SC-1' >>>>>>>>> Feb 22 15:53:09 SC-2 osaffmd[16336]: NO Node Down event for >> node >>>> id >>>>>>>>> 2010f: >>>>>>>>> Feb 22 15:53:09 SC-2 osafimmd[16346]: NO MDS event from svc_id >> 24 >>>>>>>>> (change:6, dest:13) >>>>>>>>> Feb 22 15:53:09 SC-2 osafimmd[16346]: NO MDS event from svc_id >> 25 >>>>>>>>> (change:4, dest:564113889574230) >>>>>>>>> Feb 22 15:53:09 SC-2 osaffmd[16336]: NO Current role: ACTIVE >>>>>>>>> Feb 22 15:53:09 SC-2 osaffmd[16336]: Rebooting OpenSAF NodeId >> = >>>>>>>> 131343 >>>>>>>>> EE Name = , Reason: Received Node Down for peer controller, >>>>>> OwnNodeId >>>>>>>>> = 131599, SupervisionTime = 60 >>>>>>>>> Feb 22 15:53:09 SC-2 osafrded[16327]: NO Peer down on node >>>> 0x2010f >>>>>>>>> Feb 22 15:53:09 SC-2 osafamfd[16408]: NO Node 'SC-1' left the >>>> cluster >>>>>>>>> Feb 22 15:53:09 SC-2 osaflogd[16372]: NO Failed (2) to send of >> WRITE >>>>>>>>> ack to: 2010f00003d6a >>>>>>>>> Feb 22 15:53:09 SC-2 osaflogd[16372]: NO Failed (2) to send of >> WRITE >>>>>>>>> ack to: 2010f00003d6a >>>>>>>>> Feb 22 15:53:09 SC-2 osaflogd[16372]: NO Failed (2) to send of >> WRITE >>>>>>>>> ack to: 2010f00003d74 >>>>>>>>> Feb 22 15:53:09 SC-2 osafclmd[16398]: NO Node 131343 went >> down. >>>>>> Not >>>>>>>>> sending track callback for agents on that node >>>>>>>>> Feb 22 15:53:09 SC-2 osafclmd[16398]: NO Node 131343 went >> down. >>>>>> Not >>>>>>>>> sending track callback for agents on that node >>>>>>>>> Feb 22 15:53:09 SC-2 osafclmd[16398]: NO Node 131343 went >> down. >>>>>> Not >>>>>>>>> sending track callback for agents on that node >>>>>>>>> Feb 22 15:53:09 SC-2 osafclmd[16398]: NO Node 131343 went >> down. >>>>>> Not >>>>>>>>> sending track callback for agents on that node >>>>>>>>> Feb 22 15:53:09 SC-2 osafclmd[16398]: NO Node 131343 went >> down. >>>>>> Not >>>>>>>>> sending track callback for agents on that node >>>>>>>>> Feb 22 15:53:09 SC-2 osafclmd[16398]: NO Node 131343 went >> down. >>>>>> Not >>>>>>>>> sending track callback for agents on that node >>>>>>>>> Feb 22 15:53:09 SC-2 osafclmd[16398]: NO Node 131343 went >> down. >>>>>> Not >>>>>>>>> sending track callback for agents on that node >>>>>>>>> Feb 22 15:53:09 SC-2 osafimmd[16346]: WA IMMD lost contact >> with >>>>>> peer >>>>>>>>> IMMD (NCSMDS_RED_DOWN) >>>>>>>>> Feb 22 15:53:09 SC-2 osafimmnd[16359]: NO Global discard node >>>>>> received >>>>>>>>> for nodeId:2010f pid:15702 >>>>>>>>> Feb 22 15:53:09 SC-2 osafimmnd[16359]: NO Implementer >>>> disconnected >>>>>> 16 >>>>>>>>> <0, 2010f(down)> (MsgQueueService131343) >>>>>>>>> Feb 22 15:53:09 SC-2 osafimmnd[16359]: NO Implementer >>>> disconnected >>>>>> 17 >>>>>>>>> <0, 2010f(down)> (@safAmfService2010f) >>>>>>>>> Feb 22 15:53:09 SC-2 osafimmnd[16359]: NO Implementer >>>> disconnected >>>>>> 18 >>>>>>>>> <0, 2010f(down)> (@OpenSafImmReplicatorB) >>>>>>>>> Feb 22 15:53:09 SC-2 opensaf_reboot: Rebooting remote node in >> the >>>>>>>>> absence of PLM is outside the scope of OpenSAF >>>>>>>>> >>>>>>>>> >> ============================================================== >>>>>>>> ====================================================== >>>>>>>>> -AVM >>>>>>>>> >>>>>>>>> On 2/22/2017 3:13 PM, Vu Minh Nguyen wrote: >>>>>>>>>> Hi Mahesh, >>>>>>>>>> >>>>>>>>>> I put all required patches into one. Try to use this and see if > you >>>>>>>>>> still >>>>>>>>>> have that problem or not. >>>>>>>>>> >>>>>>>>>> Regards, Vu >>>>>>>>>> >>>>>>>>>>> -----Original Message----- >>>>>>>>>>> From: A V Mahesh [mailto:mahesh.va...@oracle.com] >>>>>>>>>>> Sent: Wednesday, February 22, 2017 3:35 PM >>>>>>>>>>> To: Vu Minh Nguyen <vu.m.ngu...@dektech.com.au>; >>>>>>>>>>> lennart.l...@ericsson.com; canh.v.tru...@dektech.com.au >>>>>>>>>>> Cc: opensaf-devel@lists.sourceforge.net >>>>>>>>>>> Subject: Re: [devel] [PATCH 0 of 3] Review Request for log: add >>>>>>>>>> alternative >>>>>>>>>>> destinations of log records [#2258] V4 >>>>>>>>>>> >>>>>>>>>>> Hi Vu, >>>>>>>>>>> >>>>>>>>>>> I used new #3 and #4 patches , Can you please re-send All the >>>>> final >>>>>>>>>>> patch in go, >>>>>>>>>>> >>>>>>>>>>> which i need to apply on today`s staging ( if possible publish >>> the >>>>>>>>>>> with >>>>>>>>>>> new version ) >>>>>>>>>>> >>>>>>>>>>> -AVM >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> On 2/22/2017 1:52 PM, Vu Minh Nguyen wrote: >>>>>>>>>>>> Hi Mahesh, >>>>>>>>>>>> >>>>>>>>>>>>> Core was generated by `/usr/lib64/opensaf/osaflogd'. >>>>>>>>>>>>> Program terminated with signal 11, Segmentation fault. >>>>>>>>>>>>> #0 ckpt_proc_cfg_stream(lgs_cb*, void*) () at >>>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc:2195 >>>>>>>>>>>>> 2195 src/log/logd/lgs_mbcsv.cc: No such file or directory. >>>>>>>>>>>>> in src/log/logd/lgs_mbcsv.cc >>>>>>>>>>>> Backtrace still points to old position (lgs_mbcsv:2195). I > guess >>>>> the >>>>>>>>>>>> osaflogd binary has not been updated with the fixed patch. >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> Regards, Vu >>>>>>>>>>>> >>>>>>>>>>>>> -----Original Message----- >>>>>>>>>>>>> From: A V Mahesh [mailto:mahesh.va...@oracle.com] >>>>>>>>>>>>> Sent: Wednesday, February 22, 2017 3:18 PM >>>>>>>>>>>>> To: Vu Minh Nguyen <vu.m.ngu...@dektech.com.au>; >>>>>>>>>>>>> lennart.l...@ericsson.com; canh.v.tru...@dektech.com.au >>>>>>>>>>>>> Cc: opensaf-devel@lists.sourceforge.net >>>>>>>>>>>>> Subject: Re: [devel] [PATCH 0 of 3] Review Request for log: >> add >>>>>>>>>>>> alternative >>>>>>>>>>>>> destinations of log records [#2258] V4 >>>>>>>>>>>>> >>>>>>>>>>>>> Hi Vu, >>>>>>>>>>>>> >>>>>>>>>>>>> SC-2 standby osaflogd core dumped still occurs ( Not resolved) >> , >>>>> the >>>>>>>>>> new >>>>>>>>>>>>> patch only resolved the application (/usr/bin/logtest ) >>>>> Segmentation >>>>>>>>>>>>> fault on SC-1 Active. >>>>>>>>>>>>> >>>>>>>>>>>>> >> ============================================================== >> ============================================================== >>>>>>>>>>>>> # gdb /usr/lib64/opensaf/osaflogd >>>>>> core_1487751055.osaflogd.4594 >>>>>>>> GNU >>>>>>>>>>>>> gdb >>>>>>>>>>>>> (GDB) SUSE (7.3-0.6.1) >>>>>>>>>>>>> Copyright (C) 2011 Free Software Foundation, Inc. >>>>>>>>>>>>> ....... >>>>>>>>>>>>> Core was generated by `/usr/lib64/opensaf/osaflogd'. >>>>>>>>>>>>> Program terminated with signal 11, Segmentation fault. >>>>>>>>>>>>> #0 ckpt_proc_cfg_stream(lgs_cb*, void*) () at >>>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc:2195 >>>>>>>>>>>>> 2195 src/log/logd/lgs_mbcsv.cc: No such file or directory. >>>>>>>>>>>>> in src/log/logd/lgs_mbcsv.cc >>>>>>>>>>>>> (gdb) bt >>>>>>>>>>>>> #0 ckpt_proc_cfg_stream(lgs_cb*, void*) () at >>>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc:2195 >>>>>>>>>>>>> #1 0x00007f97b026f960 in ckpt_decode_log_struct(lgs_cb*, >>>>>>>>>>>>> ncs_mbcsv_cb_arg*, void*, void*, unsigned int >> (*)(edu_hdl_tag*, >>>>>>>>>>>>> edu_tkn_tag*, void*, unsigned int*, edu_buf_env_tag*, >>>>>>>> EDP_OP_TYPE, >>>>>>>>>>>>> EDU_ERR*)) () at src/log/logd/lgs_mbcsv.cc:950 >>>>>>>>>>>>> #2 0x00007f97b02710dc in >> ckpt_decode_async_update(lgs_cb*, >>>>>>>>>>>>> ncs_mbcsv_cb_arg*) () at src/log/logd/lgs_mbcsv.cc:1086 >>>>>>>>>>>>> #3 0x00007f97b0273941 in >> mbcsv_callback(ncs_mbcsv_cb_arg*) >>>> () >>>>>> at >>>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc:880 >>>>>>>>>>>>> #4 0x00007f97af372596 in ncs_mbscv_rcv_decode () from >>>>>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0 >>>>>>>>>>>>> #5 0x00007f97af372766 in ncs_mbcsv_rcv_async_update () >> from >>>>>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0 >>>>>>>>>>>>> #6 0x00007f97af379370 in mbcsv_process_events () from >>>>>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0 >>>>>>>>>>>>> #7 0x00007f97af3794db in mbcsv_hdl_dispatch_all () from >>>>>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0 >>>>>>>>>>>>> #8 0x00007f97af373ce2 in mbcsv_process_dispatch_request () >>>> at >>>>>>>>>>>>> src/mbc/mbcsv_api.c:423 >>>>>>>>>>>>> #9 0x00007f97b027096e in lgs_mbcsv_dispatch(unsigned int) >> () >>>> at >>>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc:327 >>>>>>>>>>>>> #10 0x00007f97b024d9f2 in main () at >>>> src/log/logd/lgs_main.cc:583 >>>>>>>>>>>>> (gdb) bt full >>>>>>>>>>>>> #0 ckpt_proc_cfg_stream(lgs_cb*, void*) () at >>>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc:2195 >>>>>>>>>>>>> ckpt_data_handler = {0x7f97b0270300 >>>>>>>>>>>>> <ckpt_proc_initialize_client(lgs_cb*, void*)>, 0x7f97b02701c0 >>>>>>>>>>>>> <ckpt_proc_finalize_client(lgs_cb*, void*)>, >>>>>>>>>>>>> 0x7f97b0270060 <ckpt_proc_agent_down(lgs_cb*, >>>>> void*)>, >>>>>>>>>>>>> 0x7f97b02712f0 <ckpt_proc_log_write(lgs_cb*, void*)>, >>>>>>>> 0x7f97b0271ab0 >>>>>>>>>>>>> <ckpt_proc_open_stream(lgs_cb*, void*)>, >>>>>>>>>>>>> 0x7f97b026fe80 > <ckpt_proc_close_stream(lgs_cb*, >>>>>>> void*)>, >>>>>>>>>>>>> 0x7f97b0272380 <ckpt_proc_cfg_stream(lgs_cb*, void*)>, >>>>>>>>>>> 0x7f97b0274800 >>>>>>>>>>>>> <ckpt_proc_lgs_cfg_v2(lgs_cb*, void*)>, >>>>>>>>>>>>> 0x7f97b0274e10 <ckpt_proc_lgs_cfg_v3(lgs_cb*, >>>>> void*)>, >>>>>>>>>>>>> 0x7f97b02754f0 <ckpt_proc_lgs_cfg_v5(lgs_cb*, void*)>} >>>>>>>>>>>>> #1 0x00007f97b026f960 in ckpt_decode_log_struct(lgs_cb*, >>>>>>>>>>>>> ncs_mbcsv_cb_arg*, void*, void*, unsigned int >> (*)(edu_hdl_tag*, >>>>>>>>>>>>> edu_tkn_tag*, void*, unsigned int*, edu_buf_env_tag*, >>>>>>>> EDP_OP_TYPE, >>>>>>>>>>>>> EDU_ERR*)) () at src/log/logd/lgs_mbcsv.cc:950 >>>>>>>>>>>>> ckpt_data_handler = {0x7f97b0270300 >>>>>>>>>>>>> <ckpt_proc_initialize_client(lgs_cb*, void*)>, 0x7f97b02701c0 >>>>>>>>>>>>> <ckpt_proc_finalize_client(lgs_cb*, void*)>, >>>>>>>>>>>>> 0x7f97b0270060 <ckpt_proc_agent_down(lgs_cb*, >>>>> void*)>, >>>>>>>>>>>>> 0x7f97b02712f0 <ckpt_proc_log_write(lgs_cb*, void*)>, >>>>>>>> 0x7f97b0271ab0 >>>>>>>>>>>>> <ckpt_proc_open_stream(lgs_cb*, void*)>, >>>>>>>>>>>>> 0x7f97b026fe80 > <ckpt_proc_close_stream(lgs_cb*, >>>>>>> void*)>, >>>>>>>>>>>>> 0x7f97b0272380 <ckpt_proc_cfg_stream(lgs_cb*, void*)>, >>>>>>>>>>> 0x7f97b0274800 >>>>>>>>>>>>> <ckpt_proc_lgs_cfg_v2(lgs_cb*, void*)>, >>>>>>>>>>>>> 0x7f97b0274e10 <ckpt_proc_lgs_cfg_v3(lgs_cb*, >>>>> void*)>, >>>>>>>>>>>>> 0x7f97b02754f0 <ckpt_proc_lgs_cfg_v5(lgs_cb*, void*)>} >>>>>>>>>>>>> #2 0x00007f97b02710dc in >> ckpt_decode_async_update(lgs_cb*, >>>>>>>>>>>>> ncs_mbcsv_cb_arg*) () at src/log/logd/lgs_mbcsv.cc:1086 >>>>>>>>>>>>> ckpt_data_handler = {0x7f97b0270300 >>>>>>>>>>>>> <ckpt_proc_initialize_client(lgs_cb*, void*)>, 0x7f97b02701c0 >>>>>>>>>>>>> <ckpt_proc_finalize_client(lgs_cb*, void*)>, >>>>>>>>>>>>> 0x7f97b0270060 <ckpt_proc_agent_down(lgs_cb*, >>>>> void*)>, >>>>>>>>>>>>> 0x7f97b02712f0 <ckpt_proc_log_write(lgs_cb*, void*)>, >>>>>>>> 0x7f97b0271ab0 >>>>>>>>>>>>> <ckpt_proc_open_stream(lgs_cb*, void*)>, >>>>>>>>>>>>> 0x7f97b026fe80 > <ckpt_proc_close_stream(lgs_cb*, >>>>>>> void*)>, >>>>>>>>>>>>> 0x7f97b0272380 <ckpt_proc_cfg_stream(lgs_cb*, void*)>, >>>>>>>>>>> 0x7f97b0274800 >>>>>>>>>>>>> <ckpt_proc_lgs_cfg_v2(lgs_cb*, void*)>, >>>>>>>>>>>>> 0x7f97b0274e10 <ckpt_proc_lgs_cfg_v3(lgs_cb*, >>>>> void*)>, >>>>>>>>>>>>> 0x7f97b02754f0 <ckpt_proc_lgs_cfg_v5(lgs_cb*, void*)>} >>>>>>>>>>>>> #3 0x00007f97b0273941 in >> mbcsv_callback(ncs_mbcsv_cb_arg*) >>>> () >>>>>> at >>>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc:880 >>>>>>>>>>>>> ckpt_data_handler = {0x7f97b0270300 >>>>>>>>>>>>> <ckpt_proc_initialize_client(lgs_cb*, void*)>, 0x7f97b02701c0 >>>>>>>>>>>>> <ckpt_proc_finalize_client(lgs_cb*, void*)>, >>>>>>>>>>>>> 0x7f97b0270060 <ckpt_proc_agent_down(lgs_cb*, >>>>> void*)>, >>>>>>>>>>>>> 0x7f97b02712f0 <ckpt_proc_log_write(lgs_cb*, void*)>, >>>>>>>> 0x7f97b0271ab0 >>>>>>>>>>>>> <ckpt_proc_open_stream(lgs_cb*, void*)>, >>>>>>>>>>>>> 0x7f97b026fe80 > <ckpt_proc_close_stream(lgs_cb*, >>>>>>> void*)>, >>>>>>>>>>>>> 0x7f97b0272380 <ckpt_proc_cfg_stream(lgs_cb*, void*)>, >>>>>>>>>>> 0x7f97b0274800 >>>>>>>>>>>>> <ckpt_proc_lgs_cfg_v2(lgs_cb*, void*)>, >>>>>>>>>>>>> 0x7f97b0274e10 <ckpt_proc_lgs_cfg_v3(lgs_cb*, >>>>> void*)>, >>>>>>>>>>>>> 0x7f97b02754f0 <ckpt_proc_lgs_cfg_v5(lgs_cb*, void*)>} >>>>>>>>>>>>> #4 0x00007f97af372596 in ncs_mbscv_rcv_decode () from >>>>>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0 >>>>>>>>>>>>> mbcsv_init_process_req_func = {0x7f97af373630 >>>>>>>>>>>>> <mbcsv_process_initialize_request>, 0x7f97af374f10 >>>>>>>>>>>>> <mbcsv_process_get_sel_obj_request>, >>>>>>>>>>>>> 0x7f97af373be0 > <mbcsv_process_dispatch_request>, >>>>>>>>>>>>> 0x7f97af373af0 <mbcsv_process_finalize_request>, >>>>>> 0x7f97af373cf0 >>>>>>>>>>>>> <mbcsv_process_open_request>, >>>>>>>>>>>>> 0x7f97af374050 <mbcsv_process_close_request>, >>>>>>>>>> 0x7f97af3741e0 >>>>>>>>>>>>> <mbcsv_process_chg_role_request>, 0x7f97af3744c0 >>>>>>>>>>>>> <mbcsv_process_snd_ckpt_request>, >>>>>>>>>>>>> 0x7f97af3747d0 > <mbcsv_process_snd_ntfy_request>, >>>>>>>>>>>>> 0x7f97af374970 <mbcsv_process_snd_data_req>, >>>> 0x7f97af373930 >>>>>>>>>>>>> <mbcsv_process_get_request>, >>>>>>>>>>>>> 0x7f97af374bd0 <mbcsv_process_set_request>} >>>>>>>>>>>>> #5 0x00007f97af372766 in ncs_mbcsv_rcv_async_update () >> from >>>>>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0 >>>>>>>>>>>>> mbcsv_init_process_req_func = {0x7f97af373630 >>>>>>>>>>>>> <mbcsv_process_initialize_request>, 0x7f97af374f10 >>>>>>>>>>>>> <mbcsv_process_get_sel_obj_request>, >>>>>>>>>>>>> 0x7f97af373be0 > <mbcsv_process_dispatch_request>, >>>>>>>>>>>>> 0x7f97af373af0 <mbcsv_process_finalize_request>, >>>>>> 0x7f97af373cf0 >>>>>>>>>>>>> <mbcsv_process_open_request>, >>>>>>>>>>>>> 0x7f97af374050 <mbcsv_process_close_request>, >>>>>>>>>> 0x7f97af3741e0 >>>>>>>>>>>>> <mbcsv_process_chg_role_request>, 0x7f97af3744c0 >>>>>>>>>>>>> <mbcsv_process_snd_ckpt_request>, >>>>>>>>>>>>> 0x7f97af3747d0 > <mbcsv_process_snd_ntfy_request>, >>>>>>>>>>>>> 0x7f97af374970 <mbcsv_process_snd_data_req>, >>>> 0x7f97af373930 >>>>>>>>>>>>> <mbcsv_process_get_request>, >>>>>>>>>>>>> 0x7f97af374bd0 <mbcsv_process_set_request>} >>>>>>>>>>>>> #6 0x00007f97af379370 in mbcsv_process_events () from >>>>>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0 >>>>>>>>>>>>> No symbol table info available. >>>>>>>>>>>>> #7 0x00007f97af3794db in mbcsv_hdl_dispatch_all () from >>>>>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0 >>>>>>>>>>>>> No symbol table info available. >>>>>>>>>>>>> #8 0x00007f97af373ce2 in mbcsv_process_dispatch_request () >>>> at >>>>>>>>>>>>> src/mbc/mbcsv_api.c:423 >>>>>>>>>>>>> mbcsv_init_process_req_func = {0x7f97af373630 >>>>>>>>>>>>> <mbcsv_process_initialize_request>, 0x7f97af374f10 >>>>>>>>>>>>> <mbcsv_process_get_sel_obj_request>, >>>>>>>>>>>>> 0x7f97af373be0 > <mbcsv_process_dispatch_request>, >>>>>>>>>>>>> 0x7f97af373af0 <mbcsv_process_finalize_request>, >>>>>> 0x7f97af373cf0 >>>>>>>>>>>>> <mbcsv_process_open_request>, >>>>>>>>>>>>> 0x7f97af374050 <mbcsv_process_close_request>, >>>>>>>>>> 0x7f97af3741e0 >>>>>>>>>>>>> <mbcsv_process_chg_role_request>, 0x7f97af3744c0 >>>>>>>>>>>>> <mbcsv_process_snd_ckpt_request>, >>>>>>>>>>>>> 0x7f97af3747d0 > <mbcsv_process_snd_ntfy_request>, >>>>>>>>>>>>> 0x7f97af374970 <mbcsv_process_snd_data_req>, >>>> 0x7f97af373930 >>>>>>>>>>>>> <mbcsv_process_get_request>, >>>>>>>>>>>>> 0x7f97af374bd0 <mbcsv_process_set_request>} >>>>>>>>>>>>> #9 0x00007f97b027096e in lgs_mbcsv_dispatch(unsigned int) >> () >>>> at >>>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc:327 >>>>>>>>>>>>> ckpt_data_handler = {0x7f97b0270300 >>>>>>>>>>>>> <ckpt_proc_initialize_client(lgs_cb*, void*)>, 0x7f97b02701c0 >>>>>>>>>>>>> <ckpt_proc_finalize_client(lgs_cb*, void*)>, >>>>>>>>>>>>> 0x7f97b0270060 <ckpt_proc_agent_down(lgs_cb*, >>>>> void*)>, >>>>>>>>>>>>> 0x7f97b02712f0 <ckpt_proc_log_write(lgs_cb*, void*)>, >>>>>>>> 0x7f97b0271ab0 >>>>>>>>>>>>> <ckpt_proc_open_stream(lgs_cb*, void*)>, >>>>>>>>>>>>> 0x7f97b026fe80 > <ckpt_proc_close_stream(lgs_cb*, >>>>>>> void*)>, >>>>>>>>>>>>> 0x7f97b0272380 <ckpt_proc_cfg_stream(lgs_cb*, void*)>, >>>>>>>>>>> 0x7f97b0274800 >>>>>>>>>>>>> <ckpt_proc_lgs_cfg_v2(lgs_cb*, void*)>, >>>>>>>>>>>>> 0x7f97b0274e10 <ckpt_proc_lgs_cfg_v3(lgs_cb*, >>>>> void*)>, >>>>>>>>>>>>> 0x7f97b02754f0 <ckpt_proc_lgs_cfg_v5(lgs_cb*, void*)>} >>>>>>>>>>>>> ---Type <return> to continue, or q <return> to quit--- >>>>>>>>>>>>> #10 0x00007f97b024d9f2 in main () at >>>> src/log/logd/lgs_main.cc:583 >>>>>>>>>>>>> usr1_sel_obj = {raise_obj = -1, rmv_obj = -1} >>>>>>>>>>>>> _lgs_cb = {mds_hdl = 65547, mds_role = >>>>>> V_DEST_RL_STANDBY, >>>>>>>>>>> vaddr >>>>>>>>>>>>> = 11, log_version = {releaseCode = 65 'A', majorVersion = 2 >>>>> '\002', >>>>>>>>>>>>> minorVersion = 2 '\002'}, client_tree = { >>>>>>>>>>>>> root_node = {bit = -1, left = > 0x7f97b04cf1b0, >>>>> right = >>>>>>>>>>>>> 0x7f97b04a2418, key_info = 0x7f97b04b7bd0 ""}, params = >>>>>> {key_size >>>>>>>>>>>>> = 4}, >>>>>>>>>>>>> n_nodes = 8}, comp_name = {_opaque = {46, 24947, >>>>>>>>>>>>> 17254, 28015, 15728, 20300, 11335, 24947, >>> 21350, >>>>>>>>>>>>> 15733, >>>>>>>>>>>>> 17235, 12845, 29484, 26209, 26451, 12861, 11342, 24947, >>>> 16742, >>>>>>>> 28784, >>>>>>>>>>>>> 20285, 25968, 21358, 17985, >>>>>>>>>>>>> 0 <repeats 105 times>}}, amf_hdl = >>> 4288675841, >>>>>>>>>>>>> amfSelectionObject = 15, amf_invocation_id = 0, >> is_quiesced_set >>>> = >>>>>>>>>> false, >>>>>>>>>>>>> immOiHandle = 554050912783, immSelectionObject = 21, >>>>>>>>>>>>> clmSelectionObject = 17, clm_hdl = 4279238657, >>>>>>>>>>>>> ha_state = >>>>>>>>>>>>> SA_AMF_HA_STANDBY, last_client_id = 208, async_upd_cnt = >>>> 743, >>>>>>>>>>>>> ckpt_state >>>>>>>>>>>>> = COLD_SYNC_IDLE, mbcsv_hdl = 4293918753, >>>>>>>>>>>>> mbcsv_sel_obj = 23, mbcsv_ckpt_hdl = > 4292870177, >>>>>>>>>>>>> mbcsv_peer_version = 7, edu_hdl = {is_inited = true, tree = >>>>>>>>>>>>> {root_node >>>>>>>>>> = >>>>>>>>>>>>> {bit = -1, left = 0x7f97b04cf2e0, >>>>>>>>>>>>> right = 0x7f97b04a25b8, key_info = >>>>> 0x7f97b04b7d40 >>>>>>>>>> ""}, >>>>>>>>>>>>> params = {key_size = 8}, n_nodes = 12}, to_version = 1}, >>>>>>>>>>>>> fully_initialized = true, lga_down_list_head = 0x0, >>>>>>>>>>>>> lga_down_list_tail = 0x0, clm_init_sel_obj = >>>>>>>>>>>>> {raise_obj = >>>>>>>>>> -1, >>>>>>>>>>>>> rmv_obj = -1}, nid_started = true, scAbsenceAllowed = 900, >>>>>>>>>>>>> lgs_recovery_state = LGS_NORMAL} >>>>>>>>>>>>> nfds = 7 >>>>>>>>>>>>> fds = {{fd = 19, events = 1, revents = 0}, {fd = >>> 15, >>>>>>>>>>>>> events = >>>>>>>>>>>>> 1, revents = 0}, {fd = 23, events = 1, revents = 1}, {fd = 13, >>>>>>>>>>>>> events = >>>>>>>>>>>>> 1, revents = 0}, {fd = -1, events = 1, >>>>>>>>>>>>> revents = 0}, {fd = 17, events = 1, revents > = >>> 0}, >>>>>>>>>>>>> {fd = >>>>>>>>>> 21, >>>>>>>>>>>>> events = 1, revents = 0}} >>>>>>>>>>>>> mbox_msgs = {0, 0, 0, 0, 0} >>>>>>>>>>>>> lgs_cb = 0x7f97b04a2400 >>>>>>>>>>>>> mbox_low = {0, 0, 0, 0, 0} >>>>>>>>>>>>> lgs_mbox_init_mutex = {__data = {__lock = 0, >>> __count = >>>>> 0, >>>>>>>>>>>>> __owner = 0, __nusers = 0, __kind = 0, __spins = 0, __list = >>>>>>>>>>>>> {__prev = >>>>>>>>>>>>> 0x0, __next = 0x0}}, >>>>>>>>>>>>> __size = '\000' <repeats 39 times>, __align = > 0} >>>>>>>>>>>>> lgs_mbx = 4291821569 >>>>>>>>>>>>> mbox_high = {0, 0, 0, 0, 0} >>>>>>>>>>>>> mbox_full = {false, false, false, false, false} >>>>>>>>>>>>> (gdb) >>>>>>>>>>>>> (gdb) >>>>>>>>>>>>> >>>>>>>>>>>>> Feb 22 13:40:55 SC-2 osafimmnd[4584]: NO Ccb 131 >>>> COMMITTED >>>>>>>>>>>>> (immcfg_SC-1_18714) >>>>>>>>>>>>> Feb 22 13:40:56 SC-2 osafamfnd[4634]: NO >>>>>>>>>>>>> 'safComp=LOG,safSu=SC-2,safSg=2N,safApp=OpenSAF' faulted >>>> due >>>>>> to >>>>>>>>>>>>> 'avaDown' : Recovery is 'nodeFailfast' >>>>>>>>>>>>> Feb 22 13:40:56 SC-2 osafamfnd[4634]: ER >>>>>>>>>>>>> safComp=LOG,safSu=SC-2,safSg=2N,safApp=OpenSAF Faulted >>>> due >>>>>>>>>>>>> to:avaDown >>>>>>>>>>>>> Recovery is:nodeFailfast >>>>>>>>>>>>> Feb 22 13:40:56 SC-2 osafamfnd[4634]: Rebooting OpenSAF >>>> NodeId >>>>>> = >>>>>>>>>>>>> 131599 >>>>>>>>>>>>> EE Name = , Reason: Component faulted: recovery is node >>>> failfast, >>>>>>>>>>>>> OwnNodeId = 131599, SupervisionTime = 60 >>>>>>>>>>>>> Feb 22 13:40:56 SC-2 opensaf_reboot: Rebooting local node; >>>>>>>> timeout=60 >> ============================================================== >> ============================================================== >>>>>>>>>>>>> On 2/22/2017 12:23 PM, A V Mahesh wrote: >>>>>>>>>>>>>> Hi Vu, >>>>>>>>>>>>>> >>>>>>>>>>>>>> On 2/22/2017 12:19 PM, Vu Minh Nguyen wrote: >>>>>>>>>>>>>>> [Vu] I has sent you 02 patches. There is code change in >>>> osaflogd >>>>>>>>>>>>>>> code >>>>>>>>>>>> that >>>>>>>>>>>>>>> fix the coredump you have observed. >>>>>>>>>>>>>>> The other one is test code that fix the logtest coredump. >>>>>>>>>>>>>> Ok I will re-test , and update you . >>>>>>>>>>>>>> >>>>>>>>>>>>>> -AVM >>>>>>>>>>>>>> >>>>>>>>>>>>>> On 2/22/2017 12:19 PM, Vu Minh Nguyen wrote: >>>>>>>>>>>>>>> Hi Mahehs, >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> See my reply inline, [Vu]. >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> Regards, Vu >>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> -----Original Message----- >>>>>>>>>>>>>>>> From: A V Mahesh [mailto:mahesh.va...@oracle.com] >>>>>>>>>>>>>>>> Sent: Wednesday, February 22, 2017 1:36 PM >>>>>>>>>>>>>>>> To: Vu Minh Nguyen <vu.m.ngu...@dektech.com.au>; >>>>>>>>>>>>>>>> lennart.l...@ericsson.com; >> canh.v.tru...@dektech.com.au >>>>>>>>>>>>>>>> Cc: opensaf-devel@lists.sourceforge.net >>>>>>>>>>>>>>>> Subject: Re: [PATCH 0 of 3] Review Request for log: add >>>>>>>>>>>>>>>> alternative >>>>>>>>>>>>>>>> destinations of log records [#2258] V4 >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> Hi Vu, >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> On 2/22/2017 11:52 AM, Vu Minh Nguyen wrote: >>>>>>>>>>>>>>>>> Hi Mahesh, >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> Have a code fault in uml test, and other one in >> checkpoint. >>>>>>>>>>>>>>>> [AVM] This is Normal Suse 11 VM ( not UML). >>>>>>>>>>>>>>>>> I have just updated the code. Please re-apply for #3 and >> #4 >>>>>>>>>> patches. >>>>>>>>>>>>>>>> [AVM] is these new patch has function changes or only test >>>>>> code >>>>>>>>>>>>> changes ? >>>>>>>>>>>>>>> [Vu] I has sent you 02 patches. There is code change in >>>> osaflogd >>>>>>>>>>>>>>> code >>>>>>>>>>>> that >>>>>>>>>>>>>>> fix the coredump you have observed. >>>>>>>>>>>>>>> The other one is test code that fix the logtest coredump. >>>>>>>>>>>>>>>>> Note that, test case #14 of suite 17 should be run on >> active >>>>>>>>>>>>>>>>> node, >>>>>>>>>>>>>>>> otherwise >>>>>>>>>>>>>>>>> getting failed. >>>>>>>>>>>>>>>> [AVM] Segmentation fault of /usr/bin/logtest Not a big >>> issue >>>>> , >>>>>>>>>>>>>>>> we need to debug why osaflogd core dumped and it > is >>>>>>>>>>>>>>>> critical >>>>>>>>>>>>>>> [Vu] I found the problem. You can try with the new one to >> see >>>> if >>>>>>>>>>>>>>> the >>>>>>>>>>>>>>> coredump is still there or not. >>>>>>>>>>>>>>>>> I will put condition check to that test case later. >>>>>>>>>>>>>>>> -AVM >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> Regards, Vu >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> -----Original Message----- >>>>>>>>>>>>>>>>>> From: A V Mahesh [mailto:mahesh.va...@oracle.com] >>>>>>>>>>>>>>>>>> Sent: Wednesday, February 22, 2017 12:16 PM >>>>>>>>>>>>>>>>>> To: Vu Minh Nguyen <vu.m.ngu...@dektech.com.au>; >>>>>>>>>>>>>>>>>> lennart.l...@ericsson.com; >>>> canh.v.tru...@dektech.com.au >>>>>>>>>>>>>>>>>> Cc: opensaf-devel@lists.sourceforge.net >>>>>>>>>>>>>>>>>> Subject: Re: [PATCH 0 of 3] Review Request for log: add >>>>>>>>>> alternative >>>>>>>>>>>>>>>>>> destinations of log records [#2258] V4 >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> Hi Vu, >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> Thanks , >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> While testing /usr/bin/logtest , SC-2 standby osaflogd >>>> core >>>>>>>>>>> dumped >>>>>>>>>>>>>>> and >>>>>>>>>>>>>>>>>> /usr/bin/logtest on SC-1 Active >>>>>>>>>>>>>>>>>> got Segmentation fault , am I missing any other patch ( > i >>>> am >>>>>>>>>> using >>>>>>>>>>>>>>>>>> devel published patch only ) >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> Following patches i am using : >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> 1) #2293 (sent by Anders Widel, but not yet >>> pushed) >>>>>>>>>>>>>>>>>> 2) #2258 (v2, sent by Lennart, but not yet >>> pushed >>>>> yet) >>>>>>>>>>>>>>>>>> 3) #2258 (v4, sent by Vu, but not yet pushed >>> yet) >> ============================================================== >>>>>>>>>>>>>>>>>> ======================================== >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> Core was generated by `/usr/lib64/opensaf/osaflogd'. >>>>>>>>>>>>>>>>>> Program terminated with signal 11, Segmentation fault. >>>>>>>>>>>>>>>>>> #0 ckpt_proc_cfg_stream(lgs_cb*, void*) () at >>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc:2195 >>>>>>>>>>>>>>>>>> 2195 src/log/logd/lgs_mbcsv.cc: No such file or >>>> directory. >>>>>>>>>>>>>>>>>> in src/log/logd/lgs_mbcsv.cc >>>>>>>>>>>>>>>>>> (gdb) bt >>>>>>>>>>>>>>>>>> #0 ckpt_proc_cfg_stream(lgs_cb*, void*) () at >>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc:2195 >>>>>>>>>>>>>>>>>> #1 0x00007f12c3e22960 in >>>> ckpt_decode_log_struct(lgs_cb*, >>>>>>>>>>>>>>>>>> ncs_mbcsv_cb_arg*, void*, void*, unsigned int >>>>>>>> (*)(edu_hdl_tag*, >>>>>>>>>>>>>>>>>> edu_tkn_tag*, void*, unsigned int*, edu_buf_env_tag*, >>>>>>>>>>>>> EDP_OP_TYPE, >>>>>>>>>>>>>>>>>> EDU_ERR*)) () at src/log/logd/lgs_mbcsv.cc:950 >>>>>>>>>>>>>>>>>> #2 0x00007f12c3e240dc in >>>>>>>> ckpt_decode_async_update(lgs_cb*, >>>>>>>>>>>>>>>>>> ncs_mbcsv_cb_arg*) () at >> src/log/logd/lgs_mbcsv.cc:1086 >>>>>>>>>>>>>>>>>> #3 0x00007f12c3e26941 in >>>>>>>> mbcsv_callback(ncs_mbcsv_cb_arg*) () >>>>>>>>>>> at >>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc:880 >>>>>>>>>>>>>>>>>> #4 0x00007f12c2f25596 in ncs_mbscv_rcv_decode () >> from >>>>>>>>>>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0 >>>>>>>>>>>>>>>>>> #5 0x00007f12c2f25766 in >> ncs_mbcsv_rcv_async_update >>>> () >>>>>>>> from >>>>>>>>>>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0 >>>>>>>>>>>>>>>>>> #6 0x00007f12c2f2c370 in mbcsv_process_events () >> from >>>>>>>>>>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0 >>>>>>>>>>>>>>>>>> #7 0x00007f12c2f2c4db in mbcsv_hdl_dispatch_all () >> from >>>>>>>>>>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0 >>>>>>>>>>>>>>>>>> #8 0x00007f12c2f26ce2 in >>>> mbcsv_process_dispatch_request >>>>>> () >>>>>>>> at >>>>>>>>>>>>>>>>>> src/mbc/mbcsv_api.c:423 >>>>>>>>>>>>>>>>>> #9 0x00007f12c3e2396e in >> lgs_mbcsv_dispatch(unsigned >>>> int) >>>>>> () >>>>>>>> at >>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc:327 >>>>>>>>>>>>>>>>>> #10 0x00007f12c3e009f2 in main () at >>>>>>>>>>>>>>>>>> src/log/logd/lgs_main.cc:583 >>>>>>>>>>>>>>>>>> (gdb) >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> >> ============================================================== >>>>>>>>>>>>>>>>>> ======================================== >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> Feb 22 10:37:06 SC-1 osafimmnd[4020]: NO Invalid error >>>>>>>> reported >>>>>>>>>>>>>>>>>> implementer 'safLogService', Ccb 161 will be aborted >>>>>>>>>>>>>>>>>> Feb 22 10:37:06 SC-1 osafimmnd[4020]: NO Ccb 161 >>>> aborted >>>>>> in >>>>>>>>>>>>>>>> COMPLETED >>>>>>>>>>>>>>>>>> processing (validation) >>>>>>>>>>>>>>>>>> Feb 22 10:37:06 SC-1 osafimmnd[4020]: NO Ccb 161 >>>>>> ABORTED >>>>>>>>>>>>>>>> (immcfg_SC- >>>>>>>>>>>>>>>>>> 1_5394) >>>>>>>>>>>>>>>>>> Add values Fail >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> Program received signal SIGSEGV, Segmentation fault. >>>>>>>>>>>>>>>>>> 0x000055555556929a in read_and_compare.isra.7 () at >>>>>>>>>>>>>>>>>> src/log/apitest/tet_LogOiOps.c:1891 >>>>>>>>>>>>>>>>>> 1891 src/log/apitest/tet_LogOiOps.c: No such file or >>>>>>>>>>>>>>>>>> directory. >>>>>>>>>>>>>>>>>> in src/log/apitest/tet_LogOiOps.c >>>>>>>>>>>>>>>>>> (gdb) Feb 22 10:37:07 SC-1 sshd[5298]: Accepted >> keyboard- >>>>>>>>>>>>>>>> interactive/pam >>>>>>>>>>>>>>>>>> for root from 10.176.178.22 port 51945 ssh2 >>>>>>>>>>>>>>>>>> bt >>>>>>>>>>>>>>>>>> #0 0x000055555556929a in read_and_compare.isra.7 () >> at >>>>>>>>>>>>>>>>>> src/log/apitest/tet_LogOiOps.c:1891 >>>>>>>>>>>>>>>>>> #1 0x0000555555569bbb in >>>>>>>>>>>>>>>>>> check_logRecordDestinationConfigurationEmpty >>>>>>>>>>>>>>>>>> () at src/log/apitest/tet_LogOiOps.c:2179 >>>>>>>>>>>>>>>>>> #2 0x0000555555573495 in run_test_case () >>>>>>>>>>>>>>>>>> #3 0x0000555555573934 in test_run () >>>>>>>>>>>>>>>>>> #4 0x000055555555c7cd in main () at >>>>>>>>>>>>>>>>>> src/log/apitest/logtest.c:569 >>>>>>>>>>>>>>>>>> (gdb) >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> >> ============================================================== >>>>>>>>>>>>>>>>>> ======================================== >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> -AVM >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> On 2/22/2017 9:48 AM, Vu Minh Nguyen wrote: >>>>>>>>>>>>>>>>>>> Hi Mahesh, >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> I send them in attachment instead, and name them in >> the >>>>>>>> order. >>>>>>>>>>>>>>>>>>> I just pull the latest code, and apply them without >>> getting >>>>>>> any >>>>>>>>>>>> hunk >>>>>>>>>>>>>>>>> error. >>>>>>>>>>>>>>>>>>> Please try with them, and let me know if you see any >>>>>> problem. >>>>>>>>>>>>>>>>>>> Regards, Vu >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> -----Original Message----- >>>>>>>>>>>>>>>>>>>> From: A V Mahesh [mailto:mahesh.va...@oracle.com] >>>>>>>>>>>>>>>>>>>> Sent: Wednesday, February 22, 2017 11:09 AM >>>>>>>>>>>>>>>>>>>> To: Vu Minh Nguyen >> <vu.m.ngu...@dektech.com.au>; >>>>>>>>>>>>>>>>>>>> lennart.l...@ericsson.com; >>>>>> canh.v.tru...@dektech.com.au >>>>>>>>>>>>>>>>>>>> Cc: opensaf-devel@lists.sourceforge.net >>>>>>>>>>>>>>>>>>>> Subject: Re: [PATCH 0 of 3] Review Request for log: >> add >>>>>>>>>>>> alternative >>>>>>>>>>>>>>>>>>>> destinations of log records [#2258] V4 >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> Hi Vu, >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> I did follow that still i get Hunk #2 FAILED even on >>>>> today's >>>>>>>>>>>> staging >> ============================================================== >>>>>>>>>>>>>>>>>>>> ================== >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> [root@dhcp-hyd-scp-5fl-10-176-177-96 staging]# >> patch >>>> - >>>>>> p1 >>>>>>>>>>> <2293 >>>>>>>>>>>>>>>>>>>> patching file src/base/Makefile.am >>>>>>>>>>>>>>>>>>>> Hunk #1 succeeded at 33 (offset 1 line). >>>>>>>>>>>>>>>>>>>> Hunk #3 succeeded at 183 (offset 1 line). >>>>>>>>>>>>>>>>>>>> patching file src/base/file_descriptor.cc >>>>>>>>>>>>>>>>>>>> patching file src/base/file_descriptor.h >>>>>>>>>>>>>>>>>>>> patching file src/base/tests/unix_socket_test.cc >>>>>>>>>>>>>>>>>>>> patching file src/base/unix_client_socket.cc >>>>>>>>>>>>>>>>>>>> patching file src/base/unix_server_socket.cc >>>>>>>>>>>>>>>>>>>> patching file src/base/unix_socket.cc >>>>>>>>>>>>>>>>>>>> patching file src/base/unix_socket.h >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> [root@dhcp-hyd-scp-5fl-10-176-177-96 staging]# >> patch >>>> - >>>>>> p1 >>>>>>>>>>> <2258- >>>>>>>>>>>>> 1 >>>>>>>>>>>>>>>>>>>> patching file src/log/Makefile.am >>>>>>>>>>>>>>>>>>>> Hunk #1 succeeded at 71 (offset -1 lines). >>>>>>>>>>>>>>>>>>>> patching file src/log/config/logsv_classes.xml >>>>>>>>>>>>>>>>>>>> Hunk #1 FAILED at 147. >>>>>>>>>>>>>>>>>>>> 1 out of 1 hunk FAILED -- saving rejects to file >>>>>>>>>>>>>>>>>>>> src/log/config/logsv_classes.xml.rej >>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_config.cc >>>>>>>>>>>>>>>>>>>> Hunk #1 succeeded at 35 (offset -5 lines). >>>>>>>>>>>>>>>>>>>> Hunk #2 FAILED at 705. >>>>>>>>>>>>>>>>>>>> Hunk #3 FAILED at 971. >>>>>>>>>>>>>>>>>>>> 2 out of 3 hunks FAILED -- saving rejects to file >>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_config.cc.rej >>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_config.h >>>>>>>>>>>>>>>>>>>> Hunk #1 FAILED at 304. >>>>>>>>>>>>>>>>>>>> 1 out of 1 hunk FAILED -- saving rejects to file >>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_config.h.rej >>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_dest.cc >>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_dest.h >>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_evt.cc >>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_imm.cc >>>>>>>>>>>>>>>>>>>> Hunk #1 FAILED at 45. >>>>>>>>>>>>>>>>>>>> Hunk #2 succeeded at 235 (offset -1 lines). >>>>>>>>>>>>>>>>>>>> Hunk #3 FAILED at 877. >>>>>>>>>>>>>>>>>>>> Hunk #4 succeeded at 1273 (offset -20 lines). >>>>>>>>>>>>>>>>>>>> Hunk #5 succeeded at 1404 (offset -1 lines). >>>>>>>>>>>>>>>>>>>> Hunk #6 succeeded at 1449 (offset -20 lines). >>>>>>>>>>>>>>>>>>>> Hunk #7 succeeded at 2032 (offset -1 lines). >>>>>>>>>>>>>>>>>>>> Hunk #8 FAILED at 2181. >>>>>>>>>>>>>>>>>>>> Hunk #9 succeeded at 2271 (offset -54 lines). >>>>>>>>>>>>>>>>>>>> Hunk #10 succeeded at 2387 (offset -1 lines). >>>>>>>>>>>>>>>>>>>> Hunk #11 succeeded at 2377 (offset -54 lines). >>>>>>>>>>>>>>>>>>>> Hunk #12 succeeded at 2478 (offset -1 lines). >>>>>>>>>>>>>>>>>>>> Hunk #13 succeeded at 2684 (offset -54 lines). >>>>>>>>>>>>>>>>>>>> Hunk #14 succeeded at 2821 (offset -1 lines). >>>>>>>>>>>>>>>>>>>> 3 out of 14 hunks FAILED -- saving rejects to file >>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_imm.cc.rej >>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_main.cc >>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_mbcsv.cc >>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_mbcsv.h >>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_mbcsv_v5.cc >>>>>>>>>>>>>>>>>>>> Hunk #3 succeeded at 133 (offset -1 lines). >>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_mbcsv_v7.cc >>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_mbcsv_v7.h >>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_stream.cc >>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_stream.h >>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_util.cc >>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_util.h >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> [root@dhcp-hyd-scp-5fl-10-176-177-96 staging]# >> patch >>>> - >>>>>> p1 >>>>>>>>>>> <2258- >>>>>>>>>>>>> 2 >>>>>>>>>>>>>>>>>>>> patching file src/log/Makefile.am >>>>>>>>>>>>>>>>>>>> Hunk #1 succeeded at 180 (offset -3 lines). >>>>>>>>>>>>>>>>>>>> patching file src/log/apitest/tet_LogOiOps.c >>>>>>>>>>>>>>>>>>>> Hunk #1 FAILED at 1923. >>>>>>>>>>>>>>>>>>>> Hunk #2 FAILED at 1979. >>>>>>>>>>>>>>>>>>>> Hunk #3 FAILED at 2067. >>>>>>>>>>>>>>>>>>>> Hunk #4 FAILED at 2094. >>>>>>>>>>>>>>>>>>>> 4 out of 4 hunks FAILED -- saving rejects to file >>>>>>>>>>>>>>>>>>>> src/log/apitest/tet_LogOiOps.c.rej >>>>>>>>>>>>>>>>>>>> patching file src/log/apitest/tet_cfg_destination.c >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> [root@dhcp-hyd-scp-5fl-10-176-177-96 staging]# >> patch >>>> - >>>>>> p1 >>>>>>>>>>> <2258- >>>>>>>>>>>>> 3 >>>>>>>>>>>>>>>>>>>> patching file src/log/Makefile >>>>>>>>>>>>>>>>>>>> patching file src/log/Makefile.am >>>>>>>>>>>>>>>>>>>> Hunk #1 succeeded at 80 (offset -1 lines). >>>>>>>>>>>>>>>>>>>> Hunk #2 succeeded at 217 (offset -2 lines). >>>>>>>>>>>>>>>>>>>> patching file src/log/tests/Makefile >>>>>>>>>>>>>>>>>>>> patching file src/log/tests/lgs_dest_test.cc >>>>>>>>>>>>>>>>>>>> [root@dhcp-hyd-scp-5fl-10-176-177-96 staging]# vi >>>>>>>>>>>>>>>>>>>> src/log/apitest/tet_LogOiOps.c.rej >>>>>>>>>>>>>>>>>>>> [root@dhcp-hyd-scp-5fl-10-176-177-96 staging]# >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> >> ============================================================== >>>>>>>>>>>>>>>>>>>> ======================== >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> -AVM >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> On 2/21/2017 3:53 PM, Vu Minh Nguyen wrote: >>>>>>>>>>>>>>>>>>>>> Hi Mahesh, >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> As I has mentioned in below: >>>>>>>>>>>>>>>>>>>>>>> To run the test, this patch has > dependent >>> on >>>>>>>>>> following >>>>>>>>>>>>>>> patches: >>>>>>>>>>>>>>>>>>>>>>> 1) #2293 (sent by Anders Widel, but not > yet >>>>>>>>>>>>>>>>>>>>>>> pushed) >>>>>>>>>>>>>>>>>>>>>>> 2) #2258 (v2, sent by Lennart, but not > yet >>>>> pushed >>>>>>>>>> yet) >>>>>>>>>>>>>>>>>>>>> So, you need to apply #2293 first, then #2258 which >>>> sent >>>>>> by >>>>>>>>>>>>> Lennart >>>>>>>>>>>>>>>>>>>>> yesterday, then mine. >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> Regards, Vu >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>> -----Original Message----- >>>>>>>>>>>>>>>>>>>>>> From: A V Mahesh >> [mailto:mahesh.va...@oracle.com] >>>>>>>>>>>>>>>>>>>>>> Sent: Tuesday, February 21, 2017 5:10 PM >>>>>>>>>>>>>>>>>>>>>> To: Vu Minh Nguyen >>>> <vu.m.ngu...@dektech.com.au>; >>>>>>>>>>>>>>>>>>>>>> lennart.l...@ericsson.com; >>>>>>>> canh.v.tru...@dektech.com.au >>>>>>>>>>>>>>>>>>>>>> Cc: opensaf-devel@lists.sourceforge.net >>>>>>>>>>>>>>>>>>>>>> Subject: Re: [PATCH 0 of 3] Review Request for log: >>>> add >>>>>>>>>>>>> alternative >>>>>>>>>>>>>>>>>>>>>> destinations of log records [#2258] V4 >>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>> Hi Vu, >>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>> Is this applies on top of log #2146 - V4 , I see >>>>> both >>>>>>>>>>>> #tickets >>>>>>>>>>>>>>>>>>> has >>>>>>>>>>>>>>>>>>>>>> version changes ? >>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>> in which order i need to apply ( #2146 & #2258 ) >> or >>>>>>>>>>>>>>>>>>>>>> (#2258 >>>>>>>>>> & >>>>>>>>>>>>>>>>>>> #2146). >> ========================================================= >>>>>>>>>>>>>>>>>>>>>> patching file src/log/Makefile.am >>>>>>>>>>>>>>>>>>>>>> Hunk #1 FAILED at 72. >>>>>>>>>>>>>>>>>>>>>> Hunk #2 FAILED at 120. >>>>>>>>>>>>>>>>>>>>>> 2 out of 2 hunks FAILED -- saving rejects to file >>>>>>>>>>>>>>>>>>> src/log/Makefile.am.rej >>>>>>>>>>>>>>>>>>>>>> patching file src/log/config/logsv_classes.xml >>>>>>>>>>>>>>>>>>>>>> Hunk #1 FAILED at 147. >>>>>>>>>>>>>>>>>>>>>> 1 out of 1 hunk FAILED -- saving rejects to file >>>>>>>>>>>>>>>>>>>>>> src/log/config/logsv_classes.xml.rej >>>>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_config.cc >>>>>>>>>>>>>>>>>>>>>> Hunk #1 succeeded at 35 (offset -5 lines). >>>>>>>>>>>>>>>>>>>>>> Hunk #2 FAILED at 705. >>>>>>>>>>>>>>>>>>>>>> Hunk #3 FAILED at 971. >>>>>>>>>>>>>>>>>>>>>> 2 out of 3 hunks FAILED -- saving rejects to file >>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_config.cc.rej >>>>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_config.h >>>>>>>>>>>>>>>>>>>>>> Hunk #1 FAILED at 304. >>>>>>>>>>>>>>>>>>>>>> 1 out of 1 hunk FAILED -- saving rejects to file >>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_config.h.rej >>>>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_dest.cc >>>>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_dest.h >>>>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_evt.cc >>>>>>>>>>>>>>>>>>>>>> Hunk #1 FAILED at 1. >>>>>>>>>>>>>>>>>>>>>> Hunk #2 succeeded at 30 with fuzz 2 (offset 2 lines). >>>>>>>>>>>>>>>>>>>>>> Hunk #3 succeeded at 1282 (offset 45 lines). >>>>>>>>>>>>>>>>>>>>>> Hunk #4 succeeded at 1300 (offset 2 lines). >>>>>>>>>>>>>>>>>>>>>> 1 out of 4 hunks FAILED -- saving rejects to file >>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_evt.cc.rej >>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>> >> ============================================================== >>>>>>>>>>>>>>>>>>>>>> === >>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>> -AVM >>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>> On 2/21/2017 3:03 PM, Vu Minh Nguyen wrote: >>>>>>>>>>>>>>>>>>>>>>> Summary: log: add alternative destinations of log >>>>>> records >>>>>>>>>>>>> [#2258] >>>>>>>>>>>>>>>>>>>>>>> Review request for Trac Ticket(s): #2258 >>>>>>>>>>>>>>>>>>>>>>> Peer Reviewer(s): Lennart, Canh, Mahesh >>>>>>>>>>>>>>>>>>>>>>> Pull request to: <<LIST THE PERSON WITH PUSH >>>>>> ACCESS >>>>>>>>>>> HERE>> >>>>>>>>>>>>>>>>>>>>>>> Affected branch(es): Default >>>>>>>>>>>>>>>>>>>>>>> Development branch: Default >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> -------------------------------- >>>>>>>>>>>>>>>>>>>>>>> Impacted area Impact y/n >>>>>>>>>>>>>>>>>>>>>>> -------------------------------- >>>>>>>>>>>>>>>>>>>>>>> Docs n >>>>>>>>>>>>>>>>>>>>>>> Build system n >>>>>>>>>>>>>>>>>>>>>>> RPM/packaging n >>>>>>>>>>>>>>>>>>>>>>> Configuration files n >>>>>>>>>>>>>>>>>>>>>>> Startup scripts n >>>>>>>>>>>>>>>>>>>>>>> SAF services n >>>>>>>>>>>>>>>>>>>>>>> OpenSAF services y >>>>>>>>>>>>>>>>>>>>>>> Core libraries n >>>>>>>>>>>>>>>>>>>>>>> Samples n >>>>>>>>>>>>>>>>>>>>>>> Tests y >>>>>>>>>>>>>>>>>>>>>>> Other n >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> Comments (indicate scope for each "y" above): >>>>>>>>>>>>>>>>>>>>>>> --------------------------------------------- >>>>>>>>>>>>>>>>>>>>>>> To run the test, this patch has > dependent >>> on >>>>>>>>>> following >>>>>>>>>>>>>>> patches: >>>>>>>>>>>>>>>>>>>>>>> 1) #2293 (sent by Anders Widel, but not > yet >>>>>>>>>>>>>>>>>>>>>>> pushed) >>>>>>>>>>>>>>>>>>>>>>> 2) #2258 (v2, sent by Lennart, but not > yet >>>>> pushed >>>>>>>>>> yet) >>>>>>>>>>>>>>>>>>>>>>> changeset >>>>>>>> d74aaf3025c99cade3165a15831124548f4d85bd >>>>>>>>>>>>>>>>>>>>>>> Author: Vu Minh Nguyen >>>>>>>> <vu.m.ngu...@dektech.com.au> >>>>>>>>>>>>>>>>>>>>>>> Date: Wed, 15 Feb 2017 14:36:00 +0700 >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> log: add alternative destinations of log > records >>>>>>>>>>>>>>>>>>>>>>> [#2258] >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> Here are major info, detailed info will be >> added >>> to >>>>>>>>>>>>>>>>>>>>>>> PR doc >>>>>>>>>>>>>>>>> soon. 1) >>>>>>>>>>>>>>>>>>>>>> Add >>>>>>>>>>>>>>>>>>>>>>> attribute "saLogRecordDestination" to log >>>> stream. >>>>> 2) >>>>>>>>>>>>>>>>>>>>>>> Add >>>>>>>>>>>>>>>>> Local >>>>>>>>>>>>>>>>>>>>>> socket >>>>>>>>>>>>>>>>>>>>>>> destintion handler 3) Integrate into first >>>>> increment >>>>>>>>>>>>>>>>>>>>>>> made by >>>>>>>>>>>>>>>>> Lennart >>>>>>>>>>>>>>>>>>>>>>> changeset >>>>>>>> 4bae27a478c235df3058f43c92d3a5483233b01d >>>>>>>>>>>>>>>>>>>>>>> Author: Vu Minh Nguyen >>>>>>>> <vu.m.ngu...@dektech.com.au> >>>>>>>>>>>>>>>>>>>>>>> Date: Wed, 15 Feb 2017 15:07:09 +0700 >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> log: add UML test case to verify alternative >>>>>>>>>>>>>>>>>>>>>>> destination >>>>>>>>>>>>>>>>> [#2258] >>>>>>>>>>>>>>>>>>>>>>> Major changes: 1) Modify Lennart's test cases >>>>>> because >>>>>>>>>>>>>>>>> enhancing >>>>>>>>>>>>>>>>>>>>>> destination >>>>>>>>>>>>>>>>>>>>>>> configuration validation rules. 2) Add test >>> suite >>>>>>>>>>>>>>>>>>>>>>> #17 to >>>>>>>>>>>>>>>>> verify >>>>>>>>>>>>>>>>>>>>>> alternative >>>>>>>>>>>>>>>>>>>>>>> destination >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> changeset >>>>>> bc375725fed22bb4f8cb3ae3df5f96fb9d281efb >>>>>>>>>>>>>>>>>>>>>>> Author: Vu Minh Nguyen >>>>>>>> <vu.m.ngu...@dektech.com.au> >>>>>>>>>>>>>>>>>>>>>>> Date: Thu, 16 Feb 2017 17:22:13 +0700 >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> log: add unit tests to verify interfaces >>> provided >>>>> by >>>>>>>>>>>>>>>>> destination >>>>>>>>>>>>>>>>>>>>>> handler >>>>>>>>>>>>>>>>>>>>>>> [#2258] >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> Unit tests to verify major interfaces: 1) >>>>>>>>>>>>>>>>>>>>>>> CfgDestination() >>>>>>>>>>>>>>>>> 2) >>>>>>>>>>>>>>>>>>>>>>> WriteToDestination() >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> Added Files: >>>>>>>>>>>>>>>>>>>>>>> ------------ >>>>>>>>>>>>>>>>>>>>>>> src/log/apitest/tet_cfg_destination.c >>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_dest.cc >>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_dest.h >>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv_v7.cc >>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv_v7.h >>>>>>>>>>>>>>>>>>>>>>> src/log/tests/lgs_dest_test.cc >>>>>>>>>>>>>>>>>>>>>>> src/log/tests/Makefile >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> Complete diffstat: >>>>>>>>>>>>>>>>>>>>>>> ------------------ >>>>>>>>>>>>>>>>>>>>>>> src/log/Makefile | 4 + >>>>>>>>>>>>>>>>>>>>>>> src/log/Makefile.am | 31 +++++- >>>>>>>>>>>>>>>>>>>>>>> src/log/apitest/tet_LogOiOps.c | 8 +- >>>>>>>>>>>>>>>>>>>>>>> src/log/apitest/tet_cfg_destination.c | 483 >> ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++ >>>>>>>>>>>>>>>>>>>>>> ++++++++++++++++++++++++++++++++++ >>>>>>>>>>>>>>>>>>>>>>> src/log/config/logsv_classes.xml | 7 +- >>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_config.cc | 169 >>>>>>>>>>>>>>>>>>>>>> +++++++++++++++++++++++++++++--- >>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_config.h | 3 +- >>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_dest.cc | 707 >> ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++ >> ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++ >>>>>>>>>>>>>>>>>>>>>> +++++++++++++++++ >>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_dest.h | 576 >> ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++ >>>>>>>>>>> ++++++++++++++++++++++++++++++++++++++++++++++++++++ >>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_evt.cc | 33 ++++++ >>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_imm.cc | 202 >>>>>>>>>>>>>>>>>>>>>> +++++++++++++++++++++++++++++++++------ >>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_main.cc | 8 + >>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc | 103 >>>>>>>>>>>>>>> ++++++++++++++++++- >>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv.h | 6 +- >>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv_v5.cc | 10 + >>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv_v7.cc | 177 >>>>>>>>>>>>>>>>>>>>>> +++++++++++++++++++++++++++++++++++ >>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv_v7.h | 67 >>>>>>>>>>>> +++++++++++++ >>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_stream.cc | 60 >>>>>>>>>>>> +++++++++++- >>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_stream.h | 16 +++ >>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_util.cc | 63 >>>>>>>>>>>> ++++++++++++ >>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_util.h | 11 +- >>>>>>>>>>>>>>>>>>>>>>> src/log/tests/Makefile | 20 +++ >>>>>>>>>>>>>>>>>>>>>>> src/log/tests/lgs_dest_test.cc | 209 >>>>>>>>>>>>>>>>>>>>>> +++++++++++++++++++++++++++++++++++++++++ >>>>>>>>>>>>>>>>>>>>>>> 23 files changed, 2896 insertions(+), 77 >>>>>>>>>>>>>>>>>>>>>>> deletions(-) >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> Testing Commands: >>>>>>>>>>>>>>>>>>>>>>> ----------------- >>>>>>>>>>>>>>>>>>>>>>> Run UML test suite #17 >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> Testing, Expected Results: >>>>>>>>>>>>>>>>>>>>>>> -------------------------- >>>>>>>>>>>>>>>>>>>>>>> All test passed >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> Conditions of Submission: >>>>>>>>>>>>>>>>>>>>>>> ------------------------- >>>>>>>>>>>>>>>>>>>>>>> <<HOW MANY DAYS BEFORE PUSHING, >>>>>> CONSENSUS >>>>>>>>>>> ETC>> >>>>>>>>>>>>>>>>>>>>>>> Arch Built Started Linux distro >>>>>>>>>>>>>>>>>>>>>>> ------------------------------------------- >>>>>>>>>>>>>>>>>>>>>>> mips n n >>>>>>>>>>>>>>>>>>>>>>> mips64 n n >>>>>>>>>>>>>>>>>>>>>>> x86 n n >>>>>>>>>>>>>>>>>>>>>>> x86_64 n n >>>>>>>>>>>>>>>>>>>>>>> powerpc n n >>>>>>>>>>>>>>>>>>>>>>> powerpc64 n n >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> Reviewer Checklist: >>>>>>>>>>>>>>>>>>>>>>> ------------------- >>>>>>>>>>>>>>>>>>>>>>> [Submitters: make sure that your review doesn't >>>> trigger >>>>>>>> any >>>>>>>>>>>>>>>>>>>> checkmarks!] >>>>>>>>>>>>>>>>>>>>>>> Your checkin has not passed review because (see >>>>>> checked >>>>>>>>>>>>> entries): >>>>>>>>>>>>>>>>>>>>>>> ___ Your RR template is generally incomplete; it >> has >>>>>> too >>>>>>>>>>> many >>>>>>>>>>>>>>> blank >>>>>>>>>>>>>>>>>>>>> entries >>>>>>>>>>>>>>>>>>>>>>> that need proper data filled in. >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> ___ You have failed to nominate the proper >> persons >>>> for >>>>>>>>>>> review >>>>>>>>>>>>> and >>>>>>>>>>>>>>>>>>> push. >>>>>>>>>>>>>>>>>>>>>>> ___ Your patches do not have proper short+long >>>>>> header >>>>>>>>>>>>>>>>>>>>>>> ___ You have grammar/spelling in your header >> that >>>> is >>>>>>>>>>>>> unacceptable. >>>>>>>>>>>>>>>>>>>>>>> ___ You have exceeded a sensible line length in >> your >>>>>>>>>>>>>>>>>>>>>> headers/comments/text. >>>>>>>>>>>>>>>>>>>>>>> ___ You have failed to put in a proper Trac Ticket >> # >>>>> into >>>>>>>>>> your >>>>>>>>>>>>>>>>>>> commits. >>>>>>>>>>>>>>>>>>>>>>> ___ You have incorrectly put/left internal data in >>>> your >>>>>>>>>>>>>>>>> comments/files >>>>>>>>>>>>>>>>>>>>>>> (i.e. internal bug tracking tool IDs, product >>>>>>>>>> names >>>>>>>>>>>> etc) >>>>>>>>>>>>>>>>>>>>>>> ___ You have not given any evidence of testing >>>> beyond >>>>>>>> basic >>>>>>>>>>>>> build >>>>>>>>>>>>>>>>>>> tests. >>>>>>>>>>>>>>>>>>>>>>> Demonstrate some level of runtime or other sanity >>>>>>>>>>>> testing. >>>>>>>>>>>>>>>>>>>>>>> ___ You have ^M present in some of your files. >>>> These >>>>>>>>>>>>>>>>>>>>>>> have to >>>>>>>>>>>>> be >>>>>>>>>>>>>>>>>>>> removed. >>>>>>>>>>>>>>>>>>>>>>> ___ You have needlessly changed whitespace or >>>> added >>>>>>>>>>>>> whitespace >>>>>>>>>>>>>>>>>>>> crimes >>>>>>>>>>>>>>>>>>>>>>> like trailing spaces, or spaces > before >>>>> tabs. >>>>>>>>>>>>>>>>>>>>>>> ___ You have mixed real technical changes with >>>>>>>> whitespace >>>>>>>>>>> and >>>>>>>>>>>>>>>> other >>>>>>>>>>>>>>>>>>>>>>> cosmetic code cleanup changes. These have to be >>>>>>>>>>>>> separate >>>>>>>>>>>>>>>>>>> commits. >>>>>>>>>>>>>>>>>>>>>>> ___ You need to refactor your submission into >>>> logical >>>>>>>>>>>>>>>>>>>>>>> chunks; >>>>>>>>>>>>>>> there >>>>>>>>>>>>>>>>> is >>>>>>>>>>>>>>>>>>>>>>> too much content into a single > commit. >>>>>>>>>>>>>>>>>>>>>>> ___ You have extraneous garbage in your review >>>>>> (merge >>>>>>>>>>>>> commits >>>>>>>>>>>>>>>> etc) >>>>>>>>>>>>>>>>>>>>>>> ___ You have giant attachments which should >> never >>>>>> have >>>>>>>>>>> been >>>>>>>>>>>>>>>> sent; >>>>>>>>>>>>>>>>>>>>>>> Instead you should place your content in a public >>>>>>>>>>>> tree to >>>>>>>>>>>>>>> be >>>>>>>>>>>>>>>>>>>>> pulled. >>>>>>>>>>>>>>>>>>>>>>> ___ You have too many commits attached to an e- >>>> mail; >>>>>>>>>>> resend >>>>>>>>>>>>> as >>>>>>>>>>>>>>>>>>>> threaded >>>>>>>>>>>>>>>>>>>>>>> commits, or place in a public tree for a pull. >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> ___ You have resent this content multiple times >>>>>> without a >>>>>>>>>>> clear >>>>>>>>>>>>>>>>>>>>> indication >>>>>>>>>>>>>>>>>>>>>>> of what has changed between each >>> re-send. >>>>>>>>>>>>>>>>>>>>>>> ___ You have failed to adequately and individually >>>>>> address >>>>>>>>>> all >>>>>>>>>>>> of >>>>>>>>>>>>>>>>> the >>>>>>>>>>>>>>>>>>>>>>> comments and change requests that were >> proposed >>>> in >>>>>>>>>>>> the >>>>>>>>>>>>>>>>> initial >>>>>>>>>>>>>>>>>>>>> review. >>>>>>>>>>>>>>>>>>>>>>> ___ You have a misconfigured ~/.hgrc file (i.e. >>>>>> username, >>>>>>>>>>> email >>>>>>>>>>>>>>> etc) >>>>>>>>>>>>>>>>>>>>>>> ___ Your computer have a badly configured date >> and >>>>>>>> time; >>>>>>>>>>>>>>>> confusing >>>>>>>>>>>>>>>>>> the >>>>>>>>>>>>>>>>>>>>>>> the threaded patch review. >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> ___ Your changes affect IPC mechanism, and you >>>> don't >>>>>>>>>>> present >>>>>>>>>>>>> any >>>>>>>>>>>>>>>>>>>> results >>>>>>>>>>>>>>>>>>>>>>> for in-service upgradability test. >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> ___ Your changes affect user manual and >>>>>> documentation, >>>>>>>>>>> your >>>>>>>>>>>>>>>> patch >>>>>>>>>>>>>>>>>>>> series >>>>>>>>>>>>>>>>>>>>>>> do not contain the patch that updates >>> the >>>>>>>>>>>>>>>>>>>>>>> Doxygen >>>>>>>>>>>>> manual. > ---------------------------------------------------------------------------- >>>>>>>>>>>> -- >>>>>>>>>>>>>> Check out the vibrant tech community on one of the world's >>>> most >>>>>>>>>>>>>> engaging tech sites, SlashDot.org! http://sdm.link/slashdot >>>>>>>>>>>>>> _______________________________________________ >>>>>>>>>>>>>> Opensaf-devel mailing list >>>>>>>>>>>>>> Opensaf-devel@lists.sourceforge.net >>>>>>>>>>>>>> https://lists.sourceforge.net/lists/listinfo/opensaf-devel ------------------------------------------------------------------------------ Check out the vibrant tech community on one of the world's most engaging tech sites, SlashDot.org! http://sdm.link/slashdot _______________________________________________ Opensaf-devel mailing list Opensaf-devel@lists.sourceforge.net https://lists.sourceforge.net/lists/listinfo/opensaf-devel