Hi Mahesh,

Can you try with 2258_v7.patch I just sent to you? 

I have pulled the latest code on OpenSAF 5.1 branch, re-created the cluster.
And it works with the case old active SC-1 (OpenSAF 5.1) and new standby
SC-2 (with 2258_v7.patch included in).

To apply 2258_v7.patch, please do remove the just pushed ticket "log:
implement SaLogFilterSetCallbackT and version handling [#2146]" ,
I have not rebased the code on that yet.

Regards, Vu

> -----Original Message-----
> From: A V Mahesh [mailto:mahesh.va...@oracle.com]
> Sent: Thursday, February 23, 2017 11:45 AM
> To: Vu Minh Nguyen <vu.m.ngu...@dektech.com.au>;
> lennart.l...@ericsson.com; canh.v.tru...@dektech.com.au
> Cc: opensaf-devel@lists.sourceforge.net
> Subject: Re: [devel] [PATCH 0 of 3] Review Request for log: add
alternative
> destinations of log records [#2258] V4
> 
> Hi Vu/Lennart,
> 
> 
> In broad WITHOUT the #2258 patch, the same code/setup working fine with
> 2 sc node (staging  changeset:   8609 ),
> as soon as we apply `2258_v5.patch` V5 patch on staging (changeset:
> 8609 ) that you have provided yesterday,
> on one sc node and try to bring up that in to cluster (in-service test)
> we are observing the issue of new node (with #2258 patch) not joining
> cluster.
> 
> ==============================================================
> ====================================================
> eb 23 10:01:59 SC-1 osafimmnd[15279]: NO Implementer (applier)
> connected: 15 (@safAmfService2010f) <127, 2010f>
> Feb 23 10:01:59 SC-1 osaflogd[15289]: NO LOGSV_DATA_GROUPNAME not
> found
> Feb 23 10:01:59 SC-1 osaflogd[15289]: NO LOG root directory is:
> "/var/log/opensaf/saflog"
> Feb 23 10:01:59 SC-1 osaflogd[15289]: NO LOG data group is: ""
> Feb 23 10:01:59 SC-1 osaflogd[15289]: NO LGS_MBCSV_VERSION = 7
> Feb 23 10:01:59 SC-1 osafamfnd[15329]: NO Assigned
> 'safSi=SC-2N,safApp=OpenSAF' STANDBY to 'safSu=SC-
> 1,safSg=2N,safApp=OpenSAF'
> Feb 23 10:01:59 SC-1 opensafd: OpenSAF(5.1.M0 - ) services successfully
> started
> Feb 23 10:01:59 SC-1 osafamfnd[15329]: NO
> 'safComp=LOG,safSu=SC-1,safSg=2N,safApp=OpenSAF' faulted due to
> 'avaDown' : Recovery is 'nodeFailfast'
> Feb 23 10:01:59 SC-1 osafamfnd[15329]: ER
> safComp=LOG,safSu=SC-1,safSg=2N,safApp=OpenSAF Faulted due
> to:avaDown
> Recovery is:nodeFailfast
> Feb 23 10:01:59 SC-1 osafamfnd[15329]: Rebooting OpenSAF NodeId =
> 131343
> EE Name = , Reason: Component faulted: recovery is node failfast,
> OwnNodeId = 131343, SupervisionTime = 60
> Feb 23 10:01:59 SC-1 opensaf_reboot: Rebooting local node; timeout=60
> Feb 23 10:02:00 SC-1 osafimmnd[15279]: NO Implementer (applier)
> connected: 16 (@OpenSafImmReplicatorB) <144, 2010f>
> Feb 23 10:01:59 SC-1 opensaf_reboot: Rebooting local node; timeout=60
> ==============================================================
> ====================================================
> 
> So it is evident that in-service upgrade part code of this need to be
> corrected.
> 
> Please see my comments as [AVM] and let me know if you need some traces
> .
> 
> If you're planing to prepare new V6 patch , please do prepare on top of
> today's latest staging.
> 
> On 2/23/2017 9:33 AM, Vu Minh Nguyen wrote:
> > Hi Mahesh,
> >
> > I have done in-service upgrade/downgrade with following cases:
> > 1) New Active SC-1 (OpenSAF 5.2 with the attached patch) + old standby
> SC-2
> > (OpenSAF 5.1)
> > --> Work fine
> [AVM] This is not a practical use cause of in-service upgrade , we can
> ignore this test further
> >
> > 2) Old Active SC-1 (OpenSAF 5.1) + new standby SC-2 (with or without
> > attached patch)
> > --> SC-2 is restarted & not able to join the cluster.
> [AVM] This use cause/flow is  we do get in in-service upgrade , so we
> need to address this.
> >
> > I got following messages in syslog:
> > Feb 23 09:32:42 SC-2 user.notice opensafd: OpenSAF(5.2.M0 -
> > 8529:b5addd36e45d:default) services successfully started
> > Feb 23 09:32:43 SC-2 local0.warn osafntfimcnd[701]: WA
> ntfimcn_imm_init
> > saImmOiImplementerSet() returned SA_AIS_ERR_TIMEOUT (5)
> > Feb 23 09:32:45 SC-2 local0.warn osafntfimcnd[701]: WA
> ntfimcn_imm_init
> > saImmOiImplementerSet() returned SA_AIS_ERR_TIMEOUT (5)
> > Feb 23 09:32:47 SC-2 local0.warn osafntfimcnd[701]: WA
> ntfimcn_imm_init
> > saImmOiImplementerSet() returned SA_AIS_ERR_TIMEOUT (5)
> > Feb 23 09:32:49 SC-2 local0.warn osafntfimcnd[701]: WA
> ntfimcn_imm_init
> > saImmOiImplementerSet() returned SA_AIS_ERR_TIMEOUT (5)
> > Feb 23 09:32:50 SC-2 local0.err osafmsgnd[592]: ER
> saImmOiImplementerSet
> > FAILED:5
> > Feb 23 09:32:50 SC-2 local0.err osafmsgnd[592]: ER
> saImmOiImplementerSet
> > FAILED:5
> > Feb 23 09:32:50 SC-2 local0.notice osafamfnd[496]: NO
> > 'safSu=SC-2,safSg=NoRed,safApp=OpenSAF' component restart probation
> timer
> > started (timeout: 60000000000 ns)
> > Feb 23 09:32:50 SC-2 local0.notice osafamfnd[496]: NO Restarting a
> component
> > of 'safSu=SC-2,safSg=NoRed,safApp=OpenSAF' (comp restart count: 1)
> > Feb 23 09:32:50 SC-2 local0.notice osafamfnd[496]: NO
> > 'safComp=MQND,safSu=SC-2,safSg=NoRed,safApp=OpenSAF' faulted due
> to
> > 'avaDown' : Recovery is 'componentRestart'
> > Feb 23 09:32:50 SC-2 local0.info osafmsgnd[736]: mkfifo already exists:
> > /var/lib/opensaf/osafmsgnd.fifo File exists
> >
> > And sometimes, on active SC-1 (OpenSAF 5.1), the node is not able to up
> > because of following error:
> >
> > Feb 23 11:00:32 SC-1 local0.err osafclmna[406]: MDTM:TIPC Dsock Socket
> > creation failed in MDTM_INIT err :Address family not supported by
> protocol
> > Feb 23 11:00:32 SC-1 local0.err osafclmna[406]: ER ncs_agents_startup
> FAILED
> [AVM]  No such issues ( with both TCP & TIPC)  (staging changeset:   8609
)
> >
> > Are you getting similar problem at your side?
> > Please note that, the problem is existed WITH or WITHOUT the #2258
> patch.
> [AVM] No , problem only if we apply `2258_v5.patch` V5 patch on staging
> (changeset:   8609 )
>             try to bring up that node in to cluster.
> 
> 
> -AVM
> 
> >
> > I have informed this to IMM to have a look, not sure any problem with
MDS
> > layer or any problem with my environment setup.
> > In the meantime, please have a look at the updated patch,  I will
continue
> > checking the problem. Will keep you updated.
> [AVM] I haven't seen any IMM problems
> >
> > Regards, Vu
> >
> >> -----Original Message-----
> >> From: A V Mahesh [mailto:mahesh.va...@oracle.com]
> >> Sent: Wednesday, February 22, 2017 5:36 PM
> >> To: Vu Minh Nguyen <vu.m.ngu...@dektech.com.au>;
> >> lennart.l...@ericsson.com; canh.v.tru...@dektech.com.au
> >> Cc: opensaf-devel@lists.sourceforge.net
> >> Subject: Re: [devel] [PATCH 0 of 3] Review Request for log: add
> > alternative
> >> destinations of log records [#2258] V4
> >>
> >> Please see correction New Standby SC-1 ( with patch )
> >>
> >> -AVM
> >>
> >> On 2/22/2017 4:02 PM, A V Mahesh wrote:
> >>> Hi Vu,
> >>>
> >>> With this new patch , we have another issue :
> >>>
> >>> 1)  standby Core by `/usr/lib64/opensaf/osaflogd'  issue got resolved
.
> >>>
> >>> 2) In-service upgrade is Not working , I have Old  Active SC-2 ( with
> >>> out patch )  and New Standby SC-1 ( with patch )
> >>>
> >>>      the new New Standby SC-1 not joining the cluster ( in-service
> >>> upgrade  failed )
> >>>
> >>> New Standby SC-1
> >>>
> >>>
> >>
> ==============================================================
> >> ======================================================
> >>>
> >>> Feb 22 15:53:05 SC-1 osafamfnd[15752]: NO
> >>> 'safSu=SC-1,safSg=NoRed,safApp=OpenSAF' Presence State
> >> INSTANTIATING
> >>> => INSTANTIATED
> >>> Feb 22 15:53:05 SC-1 osafamfnd[15752]: NO Assigning
> >>> 'safSi=NoRed4,safApp=OpenSAF' ACTIVE to
> >>> 'safSu=SC-1,safSg=NoRed,safApp=OpenSAF'
> >>> Feb 22 15:53:05 SC-1 osafamfnd[15752]: NO Assigned
> >>> 'safSi=NoRed4,safApp=OpenSAF' ACTIVE to
> >>> 'safSu=SC-1,safSg=NoRed,safApp=OpenSAF'
> >>> Feb 22 15:53:05 SC-1 osafsmfd[15889]: Started
> >>> Feb 22 15:53:05 SC-1 osafamfnd[15752]: NO
> >>> 'safSu=SC-1,safSg=2N,safApp=OpenSAF' Presence State INSTANTIATING
> =>
> >>> INSTANTIATED
> >>> Feb 22 15:53:05 SC-1 osafamfnd[15752]: NO Assigning
> >>> 'safSi=SC-2N,safApp=OpenSAF' STANDBY to
> >>> 'safSu=SC-1,safSg=2N,safApp=OpenSAF'
> >>> Feb 22 15:53:05 SC-1 osafrded[15672]: NO RDE role set to STANDBY
> >>> Feb 22 15:53:05 SC-1 osafrded[15672]: NO Peer up on node 0x2020f
> >>> Feb 22 15:53:05 SC-1 osafrded[15672]: NO Got peer info request from
> >>> node 0x2020f with role ACTIVE
> >>> Feb 22 15:53:05 SC-1 osafrded[15672]: NO Got peer info response from
> >>> node 0x2020f with role ACTIVE
> >>> Feb 22 15:53:05 SC-1 osafimmd[15691]: NO MDS event from svc_id 24
> >>> (change:5, dest:13)
> >>> Feb 22 15:53:05 SC-1 osafimmd[15691]: NO MDS event from svc_id 24
> >>> (change:3, dest:13)
> >>> Feb 22 15:53:05 SC-1 osafimmd[15691]: NO MDS event from svc_id 24
> >>> (change:5, dest:13)
> >>> Feb 22 15:53:05 SC-1 osafimmd[15691]: NO MDS event from svc_id 25
> >>> (change:3, dest:567412424453430)
> >>> Feb 22 15:53:05 SC-1 osafimmd[15691]: NO MDS event from svc_id 25
> >>> (change:3, dest:565213401202663)
> >>> Feb 22 15:53:05 SC-1 osafimmd[15691]: NO MDS event from svc_id 25
> >>> (change:3, dest:566312912825221)
> >>> Feb 22 15:53:05 SC-1 osafimmd[15691]: NO MDS event from svc_id 25
> >>> (change:3, dest:564113889574230)
> >>> Feb 22 15:53:05 SC-1 osafimmnd[15702]: NO Implementer (applier)
> >>> connected: 17 (@safAmfService2010f) <127, 2010f>
> >>> Feb 22 15:53:05 SC-1 osaflogd[15712]: NO LOGSV_DATA_GROUPNAME
> >> not found
> >>> Feb 22 15:53:05 SC-1 osaflogd[15712]: NO LOG root directory is:
> >>> "/var/log/opensaf/saflog"
> >>> Feb 22 15:53:05 SC-1 osaflogd[15712]: NO LOG data group is: ""
> >>> Feb 22 15:53:05 SC-1 osaflogd[15712]: NO LGS_MBCSV_VERSION = 7
> >>> Feb 22 15:53:05 SC-1 osafamfnd[15752]: NO Assigned
> >>> 'safSi=SC-2N,safApp=OpenSAF' STANDBY to
> >>> 'safSu=SC-1,safSg=2N,safApp=OpenSAF'
> >>> Feb 22 15:53:05 SC-1 opensafd: OpenSAF(5.1.M0 - ) services
> >>> successfully started
> >>> Feb 22 15:53:05 SC-1 osafamfnd[15752]: NO
> >>> 'safComp=LOG,safSu=SC-1,safSg=2N,safApp=OpenSAF' faulted due to
> >>> 'avaDown' : Recovery is 'nodeFailfast'
> >>> Feb 22 15:53:05 SC-1 osafamfnd[15752]: ER
> >>> safComp=LOG,safSu=SC-1,safSg=2N,safApp=OpenSAF Faulted due
> >> to:avaDown
> >>> Recovery is:nodeFailfast
> >>> Feb 22 15:53:05 SC-1 osafamfnd[15752]: Rebooting OpenSAF NodeId =
> >>> 131343 EE Name = , Reason: Component faulted: recovery is node
> >>> failfast, OwnNodeId = 131343, SupervisionTime = 60
> >>> Feb 22 15:53:05 SC-1 opensaf_reboot: Rebooting local node;
> timeout=60
> >>> Feb 22 15:53:43 SC-1 syslog-ng[1171]: syslog-ng starting up;
> >>> version='2.0.9'
> >>>
> >>>
> >>
> ==============================================================
> >> ======================================================
> >>>
> >>> Old - Active - SC-2
> >>>
> >>>
> >>
> ==============================================================
> >> ======================================================
> >>>
> >>> Feb 22 15:53:02 SC-2 osafimmnd[16359]: NO NODE STATE->
> >>> IMM_NODE_R_AVAILABLE
> >>> Feb 22 15:53:02 SC-2 osafimmloadd: NO Sync starting
> >>> Feb 22 15:53:02 SC-2 osafimmloadd: IN Synced 390 objects in total
> >>> Feb 22 15:53:02 SC-2 osafimmnd[16359]: NO NODE STATE->
> >>> IMM_NODE_FULLY_AVAILABLE 18511
> >>> Feb 22 15:53:02 SC-2 osafimmnd[16359]: NO Epoch set to 3 in
> ImmModel
> >>> Feb 22 15:53:02 SC-2 osafimmd[16346]: NO ACT: New Epoch for
> IMMND
> >>> process at node 2020f old epoch: 2  new epoch:3
> >>> Feb 22 15:53:02 SC-2 osafimmd[16346]: NO ACT: New Epoch for
> IMMND
> >>> process at node 2040f old epoch: 2  new epoch:3
> >>> Feb 22 15:53:02 SC-2 osafimmd[16346]: NO ACT: New Epoch for
> IMMND
> >>> process at node 2030f old epoch: 2  new epoch:3
> >>> Feb 22 15:53:02 SC-2 osafimmloadd: NO Sync ending normally
> >>> Feb 22 15:53:02 SC-2 osafimmd[16346]: NO ACT: New Epoch for
> IMMND
> >>> process at node 2010f old epoch: 0  new epoch:3
> >>> Feb 22 15:53:02 SC-2 osafimmnd[16359]: NO SERVER STATE:
> >>> IMM_SERVER_SYNC_SERVER --> IMM_SERVER_READY
> >>> Feb 22 15:53:03 SC-2 osafamfd[16408]: NO Received node_up from
> 2010f:
> >>> msg_id 1
> >>> Feb 22 15:53:03 SC-2 osafamfd[16408]: NO Node 'SC-1' joined the
> cluster
> >>> Feb 22 15:53:03 SC-2 osafimmnd[16359]: NO Implementer connected:
> 16
> >>> (MsgQueueService131343) <0, 2010f>
> >>> Feb 22 15:53:03 SC-2 osafrded[16327]: NO Peer up on node 0x2010f
> >>> Feb 22 15:53:03 SC-2 osafrded[16327]: NO Got peer info request from
> >>> node 0x2010f with role STANDBY
> >>> Feb 22 15:53:03 SC-2 osafrded[16327]: NO Got peer info response from
> >>> node 0x2010f with role STANDBY
> >>> Feb 22 15:53:03 SC-2 osafimmd[16346]: NO MDS event from svc_id 24
> >>> (change:5, dest:13)
> >>> Feb 22 15:53:03 SC-2 osafimmnd[16359]: NO Implementer (applier)
> >>> connected: 17 (@safAmfService2010f) <0, 2010f>
> >>> Feb 22 15:53:03 SC-2 osafamfd[16408]: NO Cluster startup is done
> >>> Feb 22 15:53:04 SC-2 osafimmnd[16359]: NO Implementer (applier)
> >>> connected: 18 (@OpenSafImmReplicatorB) <0, 2010f>
> >>> Feb 22 15:53:09 SC-2 osafdtmd[16304]: NO Lost contact with 'SC-1'
> >>> Feb 22 15:53:09 SC-2 osaffmd[16336]: NO Node Down event for node id
> >>> 2010f:
> >>> Feb 22 15:53:09 SC-2 osafimmd[16346]: NO MDS event from svc_id 24
> >>> (change:6, dest:13)
> >>> Feb 22 15:53:09 SC-2 osafimmd[16346]: NO MDS event from svc_id 25
> >>> (change:4, dest:564113889574230)
> >>> Feb 22 15:53:09 SC-2 osaffmd[16336]: NO Current role: ACTIVE
> >>> Feb 22 15:53:09 SC-2 osaffmd[16336]: Rebooting OpenSAF NodeId =
> >> 131343
> >>> EE Name = , Reason: Received Node Down for peer controller,
> OwnNodeId
> >>> = 131599, SupervisionTime = 60
> >>> Feb 22 15:53:09 SC-2 osafrded[16327]: NO Peer down on node 0x2010f
> >>> Feb 22 15:53:09 SC-2 osafamfd[16408]: NO Node 'SC-1' left the cluster
> >>> Feb 22 15:53:09 SC-2 osaflogd[16372]: NO Failed (2) to send of WRITE
> >>> ack to: 2010f00003d6a
> >>> Feb 22 15:53:09 SC-2 osaflogd[16372]: NO Failed (2) to send of WRITE
> >>> ack to: 2010f00003d6a
> >>> Feb 22 15:53:09 SC-2 osaflogd[16372]: NO Failed (2) to send of WRITE
> >>> ack to: 2010f00003d74
> >>> Feb 22 15:53:09 SC-2 osafclmd[16398]: NO Node 131343 went down.
> Not
> >>> sending track callback for agents on that node
> >>> Feb 22 15:53:09 SC-2 osafclmd[16398]: NO Node 131343 went down.
> Not
> >>> sending track callback for agents on that node
> >>> Feb 22 15:53:09 SC-2 osafclmd[16398]: NO Node 131343 went down.
> Not
> >>> sending track callback for agents on that node
> >>> Feb 22 15:53:09 SC-2 osafclmd[16398]: NO Node 131343 went down.
> Not
> >>> sending track callback for agents on that node
> >>> Feb 22 15:53:09 SC-2 osafclmd[16398]: NO Node 131343 went down.
> Not
> >>> sending track callback for agents on that node
> >>> Feb 22 15:53:09 SC-2 osafclmd[16398]: NO Node 131343 went down.
> Not
> >>> sending track callback for agents on that node
> >>> Feb 22 15:53:09 SC-2 osafclmd[16398]: NO Node 131343 went down.
> Not
> >>> sending track callback for agents on that node
> >>> Feb 22 15:53:09 SC-2 osafimmd[16346]: WA IMMD lost contact with
> peer
> >>> IMMD (NCSMDS_RED_DOWN)
> >>> Feb 22 15:53:09 SC-2 osafimmnd[16359]: NO Global discard node
> received
> >>> for nodeId:2010f pid:15702
> >>> Feb 22 15:53:09 SC-2 osafimmnd[16359]: NO Implementer disconnected
> 16
> >>> <0, 2010f(down)> (MsgQueueService131343)
> >>> Feb 22 15:53:09 SC-2 osafimmnd[16359]: NO Implementer disconnected
> 17
> >>> <0, 2010f(down)> (@safAmfService2010f)
> >>> Feb 22 15:53:09 SC-2 osafimmnd[16359]: NO Implementer disconnected
> 18
> >>> <0, 2010f(down)> (@OpenSafImmReplicatorB)
> >>> Feb 22 15:53:09 SC-2 opensaf_reboot: Rebooting remote node in the
> >>> absence of PLM is outside the scope of OpenSAF
> >>>
> >>>
> >>
> ==============================================================
> >> ======================================================
> >>>
> >>> -AVM
> >>>
> >>> On 2/22/2017 3:13 PM, Vu Minh Nguyen wrote:
> >>>> Hi Mahesh,
> >>>>
> >>>> I put all required patches into one. Try to use this and see if you
> >>>> still
> >>>> have that problem or not.
> >>>>
> >>>> Regards, Vu
> >>>>
> >>>>> -----Original Message-----
> >>>>> From: A V Mahesh [mailto:mahesh.va...@oracle.com]
> >>>>> Sent: Wednesday, February 22, 2017 3:35 PM
> >>>>> To: Vu Minh Nguyen <vu.m.ngu...@dektech.com.au>;
> >>>>> lennart.l...@ericsson.com; canh.v.tru...@dektech.com.au
> >>>>> Cc: opensaf-devel@lists.sourceforge.net
> >>>>> Subject: Re: [devel] [PATCH 0 of 3] Review Request for log: add
> >>>> alternative
> >>>>> destinations of log records [#2258] V4
> >>>>>
> >>>>> Hi Vu,
> >>>>>
> >>>>> I used  new #3 and #4 patches ,  Can you please re-send All the
final
> >>>>> patch in go,
> >>>>>
> >>>>> which i need to apply on today`s  staging ( if possible publish the
> >>>>> with
> >>>>> new version )
> >>>>>
> >>>>> -AVM
> >>>>>
> >>>>>
> >>>>> On 2/22/2017 1:52 PM, Vu Minh Nguyen wrote:
> >>>>>> Hi Mahesh,
> >>>>>>
> >>>>>>> Core was generated by `/usr/lib64/opensaf/osaflogd'.
> >>>>>>> Program terminated with signal 11, Segmentation fault.
> >>>>>>> #0  ckpt_proc_cfg_stream(lgs_cb*, void*) () at
> >>>>>>> src/log/logd/lgs_mbcsv.cc:2195
> >>>>>>> 2195    src/log/logd/lgs_mbcsv.cc: No such file or directory.
> >>>>>>>             in src/log/logd/lgs_mbcsv.cc
> >>>>>> Backtrace still points to old position (lgs_mbcsv:2195). I guess
the
> >>>>>> osaflogd binary has not been updated with the fixed patch.
> >>>>>>
> >>>>>>
> >>>>>> Regards, Vu
> >>>>>>
> >>>>>>> -----Original Message-----
> >>>>>>> From: A V Mahesh [mailto:mahesh.va...@oracle.com]
> >>>>>>> Sent: Wednesday, February 22, 2017 3:18 PM
> >>>>>>> To: Vu Minh Nguyen <vu.m.ngu...@dektech.com.au>;
> >>>>>>> lennart.l...@ericsson.com; canh.v.tru...@dektech.com.au
> >>>>>>> Cc: opensaf-devel@lists.sourceforge.net
> >>>>>>> Subject: Re: [devel] [PATCH 0 of 3] Review Request for log: add
> >>>>>> alternative
> >>>>>>> destinations of log records [#2258] V4
> >>>>>>>
> >>>>>>> Hi Vu,
> >>>>>>>
> >>>>>>> SC-2 standby osaflogd core dumped still occurs ( Not resolved) ,
the
> >>>> new
> >>>>>>> patch only resolved the application (/usr/bin/logtest )
Segmentation
> >>>>>>> fault on SC-1 Active.
> >>>>>>>
> >>>>>>>
> >>
> ==============================================================
> >>
> ==============================================================
> >>>>>>> # gdb /usr/lib64/opensaf/osaflogd
> core_1487751055.osaflogd.4594
> >> GNU
> >>>>>>> gdb
> >>>>>>> (GDB) SUSE (7.3-0.6.1)
> >>>>>>> Copyright (C) 2011 Free Software Foundation, Inc.
> >>>>>>> .......
> >>>>>>> Core was generated by `/usr/lib64/opensaf/osaflogd'.
> >>>>>>> Program terminated with signal 11, Segmentation fault.
> >>>>>>> #0  ckpt_proc_cfg_stream(lgs_cb*, void*) () at
> >>>>>>> src/log/logd/lgs_mbcsv.cc:2195
> >>>>>>> 2195    src/log/logd/lgs_mbcsv.cc: No such file or directory.
> >>>>>>>             in src/log/logd/lgs_mbcsv.cc
> >>>>>>> (gdb) bt
> >>>>>>> #0  ckpt_proc_cfg_stream(lgs_cb*, void*) () at
> >>>>>>> src/log/logd/lgs_mbcsv.cc:2195
> >>>>>>> #1  0x00007f97b026f960 in ckpt_decode_log_struct(lgs_cb*,
> >>>>>>> ncs_mbcsv_cb_arg*, void*, void*, unsigned int (*)(edu_hdl_tag*,
> >>>>>>> edu_tkn_tag*, void*, unsigned int*, edu_buf_env_tag*,
> >> EDP_OP_TYPE,
> >>>>>>> EDU_ERR*)) () at src/log/logd/lgs_mbcsv.cc:950
> >>>>>>> #2  0x00007f97b02710dc in ckpt_decode_async_update(lgs_cb*,
> >>>>>>> ncs_mbcsv_cb_arg*) () at src/log/logd/lgs_mbcsv.cc:1086
> >>>>>>> #3  0x00007f97b0273941 in mbcsv_callback(ncs_mbcsv_cb_arg*) ()
> at
> >>>>>>> src/log/logd/lgs_mbcsv.cc:880
> >>>>>>> #4  0x00007f97af372596 in ncs_mbscv_rcv_decode () from
> >>>>>>> /usr/lib/../lib64/libopensaf_core.so.0
> >>>>>>> #5  0x00007f97af372766 in ncs_mbcsv_rcv_async_update () from
> >>>>>>> /usr/lib/../lib64/libopensaf_core.so.0
> >>>>>>> #6  0x00007f97af379370 in mbcsv_process_events () from
> >>>>>>> /usr/lib/../lib64/libopensaf_core.so.0
> >>>>>>> #7  0x00007f97af3794db in mbcsv_hdl_dispatch_all () from
> >>>>>>> /usr/lib/../lib64/libopensaf_core.so.0
> >>>>>>> #8  0x00007f97af373ce2 in mbcsv_process_dispatch_request () at
> >>>>>>> src/mbc/mbcsv_api.c:423
> >>>>>>> #9  0x00007f97b027096e in lgs_mbcsv_dispatch(unsigned int) () at
> >>>>>>> src/log/logd/lgs_mbcsv.cc:327
> >>>>>>> #10 0x00007f97b024d9f2 in main () at src/log/logd/lgs_main.cc:583
> >>>>>>> (gdb) bt full
> >>>>>>> #0  ckpt_proc_cfg_stream(lgs_cb*, void*) () at
> >>>>>>> src/log/logd/lgs_mbcsv.cc:2195
> >>>>>>>             ckpt_data_handler = {0x7f97b0270300
> >>>>>>> <ckpt_proc_initialize_client(lgs_cb*, void*)>, 0x7f97b02701c0
> >>>>>>> <ckpt_proc_finalize_client(lgs_cb*, void*)>,
> >>>>>>>               0x7f97b0270060 <ckpt_proc_agent_down(lgs_cb*,
void*)>,
> >>>>>>> 0x7f97b02712f0 <ckpt_proc_log_write(lgs_cb*, void*)>,
> >> 0x7f97b0271ab0
> >>>>>>> <ckpt_proc_open_stream(lgs_cb*, void*)>,
> >>>>>>>               0x7f97b026fe80 <ckpt_proc_close_stream(lgs_cb*,
> > void*)>,
> >>>>>>> 0x7f97b0272380 <ckpt_proc_cfg_stream(lgs_cb*, void*)>,
> >>>>> 0x7f97b0274800
> >>>>>>> <ckpt_proc_lgs_cfg_v2(lgs_cb*, void*)>,
> >>>>>>>               0x7f97b0274e10 <ckpt_proc_lgs_cfg_v3(lgs_cb*,
void*)>,
> >>>>>>> 0x7f97b02754f0 <ckpt_proc_lgs_cfg_v5(lgs_cb*, void*)>}
> >>>>>>> #1  0x00007f97b026f960 in ckpt_decode_log_struct(lgs_cb*,
> >>>>>>> ncs_mbcsv_cb_arg*, void*, void*, unsigned int (*)(edu_hdl_tag*,
> >>>>>>> edu_tkn_tag*, void*, unsigned int*, edu_buf_env_tag*,
> >> EDP_OP_TYPE,
> >>>>>>> EDU_ERR*)) () at src/log/logd/lgs_mbcsv.cc:950
> >>>>>>>             ckpt_data_handler = {0x7f97b0270300
> >>>>>>> <ckpt_proc_initialize_client(lgs_cb*, void*)>, 0x7f97b02701c0
> >>>>>>> <ckpt_proc_finalize_client(lgs_cb*, void*)>,
> >>>>>>>               0x7f97b0270060 <ckpt_proc_agent_down(lgs_cb*,
void*)>,
> >>>>>>> 0x7f97b02712f0 <ckpt_proc_log_write(lgs_cb*, void*)>,
> >> 0x7f97b0271ab0
> >>>>>>> <ckpt_proc_open_stream(lgs_cb*, void*)>,
> >>>>>>>               0x7f97b026fe80 <ckpt_proc_close_stream(lgs_cb*,
> > void*)>,
> >>>>>>> 0x7f97b0272380 <ckpt_proc_cfg_stream(lgs_cb*, void*)>,
> >>>>> 0x7f97b0274800
> >>>>>>> <ckpt_proc_lgs_cfg_v2(lgs_cb*, void*)>,
> >>>>>>>               0x7f97b0274e10 <ckpt_proc_lgs_cfg_v3(lgs_cb*,
void*)>,
> >>>>>>> 0x7f97b02754f0 <ckpt_proc_lgs_cfg_v5(lgs_cb*, void*)>}
> >>>>>>> #2  0x00007f97b02710dc in ckpt_decode_async_update(lgs_cb*,
> >>>>>>> ncs_mbcsv_cb_arg*) () at src/log/logd/lgs_mbcsv.cc:1086
> >>>>>>>             ckpt_data_handler = {0x7f97b0270300
> >>>>>>> <ckpt_proc_initialize_client(lgs_cb*, void*)>, 0x7f97b02701c0
> >>>>>>> <ckpt_proc_finalize_client(lgs_cb*, void*)>,
> >>>>>>>               0x7f97b0270060 <ckpt_proc_agent_down(lgs_cb*,
void*)>,
> >>>>>>> 0x7f97b02712f0 <ckpt_proc_log_write(lgs_cb*, void*)>,
> >> 0x7f97b0271ab0
> >>>>>>> <ckpt_proc_open_stream(lgs_cb*, void*)>,
> >>>>>>>               0x7f97b026fe80 <ckpt_proc_close_stream(lgs_cb*,
> > void*)>,
> >>>>>>> 0x7f97b0272380 <ckpt_proc_cfg_stream(lgs_cb*, void*)>,
> >>>>> 0x7f97b0274800
> >>>>>>> <ckpt_proc_lgs_cfg_v2(lgs_cb*, void*)>,
> >>>>>>>               0x7f97b0274e10 <ckpt_proc_lgs_cfg_v3(lgs_cb*,
void*)>,
> >>>>>>> 0x7f97b02754f0 <ckpt_proc_lgs_cfg_v5(lgs_cb*, void*)>}
> >>>>>>> #3  0x00007f97b0273941 in mbcsv_callback(ncs_mbcsv_cb_arg*) ()
> at
> >>>>>>> src/log/logd/lgs_mbcsv.cc:880
> >>>>>>>             ckpt_data_handler = {0x7f97b0270300
> >>>>>>> <ckpt_proc_initialize_client(lgs_cb*, void*)>, 0x7f97b02701c0
> >>>>>>> <ckpt_proc_finalize_client(lgs_cb*, void*)>,
> >>>>>>>               0x7f97b0270060 <ckpt_proc_agent_down(lgs_cb*,
void*)>,
> >>>>>>> 0x7f97b02712f0 <ckpt_proc_log_write(lgs_cb*, void*)>,
> >> 0x7f97b0271ab0
> >>>>>>> <ckpt_proc_open_stream(lgs_cb*, void*)>,
> >>>>>>>               0x7f97b026fe80 <ckpt_proc_close_stream(lgs_cb*,
> > void*)>,
> >>>>>>> 0x7f97b0272380 <ckpt_proc_cfg_stream(lgs_cb*, void*)>,
> >>>>> 0x7f97b0274800
> >>>>>>> <ckpt_proc_lgs_cfg_v2(lgs_cb*, void*)>,
> >>>>>>>               0x7f97b0274e10 <ckpt_proc_lgs_cfg_v3(lgs_cb*,
void*)>,
> >>>>>>> 0x7f97b02754f0 <ckpt_proc_lgs_cfg_v5(lgs_cb*, void*)>}
> >>>>>>> #4  0x00007f97af372596 in ncs_mbscv_rcv_decode () from
> >>>>>>> /usr/lib/../lib64/libopensaf_core.so.0
> >>>>>>>             mbcsv_init_process_req_func = {0x7f97af373630
> >>>>>>> <mbcsv_process_initialize_request>, 0x7f97af374f10
> >>>>>>> <mbcsv_process_get_sel_obj_request>,
> >>>>>>>               0x7f97af373be0 <mbcsv_process_dispatch_request>,
> >>>>>>> 0x7f97af373af0 <mbcsv_process_finalize_request>,
> 0x7f97af373cf0
> >>>>>>> <mbcsv_process_open_request>,
> >>>>>>>               0x7f97af374050 <mbcsv_process_close_request>,
> >>>> 0x7f97af3741e0
> >>>>>>> <mbcsv_process_chg_role_request>, 0x7f97af3744c0
> >>>>>>> <mbcsv_process_snd_ckpt_request>,
> >>>>>>>               0x7f97af3747d0 <mbcsv_process_snd_ntfy_request>,
> >>>>>>> 0x7f97af374970 <mbcsv_process_snd_data_req>, 0x7f97af373930
> >>>>>>> <mbcsv_process_get_request>,
> >>>>>>>               0x7f97af374bd0 <mbcsv_process_set_request>}
> >>>>>>> #5  0x00007f97af372766 in ncs_mbcsv_rcv_async_update () from
> >>>>>>> /usr/lib/../lib64/libopensaf_core.so.0
> >>>>>>>             mbcsv_init_process_req_func = {0x7f97af373630
> >>>>>>> <mbcsv_process_initialize_request>, 0x7f97af374f10
> >>>>>>> <mbcsv_process_get_sel_obj_request>,
> >>>>>>>               0x7f97af373be0 <mbcsv_process_dispatch_request>,
> >>>>>>> 0x7f97af373af0 <mbcsv_process_finalize_request>,
> 0x7f97af373cf0
> >>>>>>> <mbcsv_process_open_request>,
> >>>>>>>               0x7f97af374050 <mbcsv_process_close_request>,
> >>>> 0x7f97af3741e0
> >>>>>>> <mbcsv_process_chg_role_request>, 0x7f97af3744c0
> >>>>>>> <mbcsv_process_snd_ckpt_request>,
> >>>>>>>               0x7f97af3747d0 <mbcsv_process_snd_ntfy_request>,
> >>>>>>> 0x7f97af374970 <mbcsv_process_snd_data_req>, 0x7f97af373930
> >>>>>>> <mbcsv_process_get_request>,
> >>>>>>>               0x7f97af374bd0 <mbcsv_process_set_request>}
> >>>>>>> #6  0x00007f97af379370 in mbcsv_process_events () from
> >>>>>>> /usr/lib/../lib64/libopensaf_core.so.0
> >>>>>>> No symbol table info available.
> >>>>>>> #7  0x00007f97af3794db in mbcsv_hdl_dispatch_all () from
> >>>>>>> /usr/lib/../lib64/libopensaf_core.so.0
> >>>>>>> No symbol table info available.
> >>>>>>> #8  0x00007f97af373ce2 in mbcsv_process_dispatch_request () at
> >>>>>>> src/mbc/mbcsv_api.c:423
> >>>>>>>             mbcsv_init_process_req_func = {0x7f97af373630
> >>>>>>> <mbcsv_process_initialize_request>, 0x7f97af374f10
> >>>>>>> <mbcsv_process_get_sel_obj_request>,
> >>>>>>>               0x7f97af373be0 <mbcsv_process_dispatch_request>,
> >>>>>>> 0x7f97af373af0 <mbcsv_process_finalize_request>,
> 0x7f97af373cf0
> >>>>>>> <mbcsv_process_open_request>,
> >>>>>>>               0x7f97af374050 <mbcsv_process_close_request>,
> >>>> 0x7f97af3741e0
> >>>>>>> <mbcsv_process_chg_role_request>, 0x7f97af3744c0
> >>>>>>> <mbcsv_process_snd_ckpt_request>,
> >>>>>>>               0x7f97af3747d0 <mbcsv_process_snd_ntfy_request>,
> >>>>>>> 0x7f97af374970 <mbcsv_process_snd_data_req>, 0x7f97af373930
> >>>>>>> <mbcsv_process_get_request>,
> >>>>>>>               0x7f97af374bd0 <mbcsv_process_set_request>}
> >>>>>>> #9  0x00007f97b027096e in lgs_mbcsv_dispatch(unsigned int) () at
> >>>>>>> src/log/logd/lgs_mbcsv.cc:327
> >>>>>>>             ckpt_data_handler = {0x7f97b0270300
> >>>>>>> <ckpt_proc_initialize_client(lgs_cb*, void*)>, 0x7f97b02701c0
> >>>>>>> <ckpt_proc_finalize_client(lgs_cb*, void*)>,
> >>>>>>>               0x7f97b0270060 <ckpt_proc_agent_down(lgs_cb*,
void*)>,
> >>>>>>> 0x7f97b02712f0 <ckpt_proc_log_write(lgs_cb*, void*)>,
> >> 0x7f97b0271ab0
> >>>>>>> <ckpt_proc_open_stream(lgs_cb*, void*)>,
> >>>>>>>               0x7f97b026fe80 <ckpt_proc_close_stream(lgs_cb*,
> > void*)>,
> >>>>>>> 0x7f97b0272380 <ckpt_proc_cfg_stream(lgs_cb*, void*)>,
> >>>>> 0x7f97b0274800
> >>>>>>> <ckpt_proc_lgs_cfg_v2(lgs_cb*, void*)>,
> >>>>>>>               0x7f97b0274e10 <ckpt_proc_lgs_cfg_v3(lgs_cb*,
void*)>,
> >>>>>>> 0x7f97b02754f0 <ckpt_proc_lgs_cfg_v5(lgs_cb*, void*)>}
> >>>>>>> ---Type <return> to continue, or q <return> to quit---
> >>>>>>> #10 0x00007f97b024d9f2 in main () at src/log/logd/lgs_main.cc:583
> >>>>>>>             usr1_sel_obj = {raise_obj = -1, rmv_obj = -1}
> >>>>>>>             _lgs_cb = {mds_hdl = 65547, mds_role =
> V_DEST_RL_STANDBY,
> >>>>> vaddr
> >>>>>>> = 11, log_version = {releaseCode = 65 'A', majorVersion = 2
'\002',
> >>>>>>> minorVersion = 2 '\002'}, client_tree = {
> >>>>>>>                 root_node = {bit = -1, left = 0x7f97b04cf1b0,
right =
> >>>>>>> 0x7f97b04a2418, key_info = 0x7f97b04b7bd0 ""}, params =
> {key_size
> >>>>>>> = 4},
> >>>>>>> n_nodes = 8}, comp_name = {_opaque = {46, 24947,
> >>>>>>>                   17254, 28015, 15728, 20300, 11335, 24947, 21350,
> >>>>>>> 15733,
> >>>>>>> 17235, 12845, 29484, 26209, 26451, 12861, 11342, 24947, 16742,
> >> 28784,
> >>>>>>> 20285, 25968, 21358, 17985,
> >>>>>>>                   0 <repeats 105 times>}}, amf_hdl = 4288675841,
> >>>>>>> amfSelectionObject = 15, amf_invocation_id = 0, is_quiesced_set =
> >>>> false,
> >>>>>>> immOiHandle = 554050912783, immSelectionObject = 21,
> >>>>>>>               clmSelectionObject = 17, clm_hdl = 4279238657,
> >>>>>>> ha_state =
> >>>>>>> SA_AMF_HA_STANDBY, last_client_id = 208, async_upd_cnt = 743,
> >>>>>>> ckpt_state
> >>>>>>> = COLD_SYNC_IDLE, mbcsv_hdl = 4293918753,
> >>>>>>>               mbcsv_sel_obj = 23, mbcsv_ckpt_hdl = 4292870177,
> >>>>>>> mbcsv_peer_version = 7, edu_hdl = {is_inited = true, tree =
> >>>>>>> {root_node
> >>>> =
> >>>>>>> {bit = -1, left = 0x7f97b04cf2e0,
> >>>>>>>                     right = 0x7f97b04a25b8, key_info =
0x7f97b04b7d40
> >>>> ""},
> >>>>>>> params = {key_size = 8}, n_nodes = 12}, to_version = 1},
> >>>>>>> fully_initialized = true, lga_down_list_head = 0x0,
> >>>>>>>               lga_down_list_tail = 0x0, clm_init_sel_obj =
> >>>>>>> {raise_obj =
> >>>> -1,
> >>>>>>> rmv_obj = -1}, nid_started = true, scAbsenceAllowed = 900,
> >>>>>>> lgs_recovery_state = LGS_NORMAL}
> >>>>>>>             nfds = 7
> >>>>>>>             fds = {{fd = 19, events = 1, revents = 0}, {fd = 15,
> >>>>>>> events =
> >>>>>>> 1, revents = 0}, {fd = 23, events = 1, revents = 1}, {fd = 13,
> >>>>>>> events =
> >>>>>>> 1, revents = 0}, {fd = -1, events = 1,
> >>>>>>>                 revents = 0}, {fd = 17, events = 1, revents = 0},
> >>>>>>> {fd =
> >>>> 21,
> >>>>>>> events = 1, revents = 0}}
> >>>>>>>             mbox_msgs = {0, 0, 0, 0, 0}
> >>>>>>>             lgs_cb = 0x7f97b04a2400
> >>>>>>>             mbox_low = {0, 0, 0, 0, 0}
> >>>>>>>             lgs_mbox_init_mutex = {__data = {__lock = 0, __count =
0,
> >>>>>>> __owner = 0, __nusers = 0, __kind = 0, __spins = 0, __list =
> >>>>>>> {__prev =
> >>>>>>> 0x0, __next = 0x0}},
> >>>>>>>               __size = '\000' <repeats 39 times>, __align = 0}
> >>>>>>>             lgs_mbx = 4291821569
> >>>>>>>             mbox_high = {0, 0, 0, 0, 0}
> >>>>>>>             mbox_full = {false, false, false, false, false}
> >>>>>>> (gdb)
> >>>>>>> (gdb)
> >>>>>>>
> >>>>>>> Feb 22 13:40:55 SC-2 osafimmnd[4584]: NO Ccb 131 COMMITTED
> >>>>>>> (immcfg_SC-1_18714)
> >>>>>>> Feb 22 13:40:56 SC-2 osafamfnd[4634]: NO
> >>>>>>> 'safComp=LOG,safSu=SC-2,safSg=2N,safApp=OpenSAF' faulted due
> to
> >>>>>>> 'avaDown' : Recovery is 'nodeFailfast'
> >>>>>>> Feb 22 13:40:56 SC-2 osafamfnd[4634]: ER
> >>>>>>> safComp=LOG,safSu=SC-2,safSg=2N,safApp=OpenSAF Faulted due
> >>>>>>> to:avaDown
> >>>>>>> Recovery is:nodeFailfast
> >>>>>>> Feb 22 13:40:56 SC-2 osafamfnd[4634]: Rebooting OpenSAF NodeId
> =
> >>>>>>> 131599
> >>>>>>> EE Name = , Reason: Component faulted: recovery is node failfast,
> >>>>>>> OwnNodeId = 131599, SupervisionTime = 60
> >>>>>>> Feb 22 13:40:56 SC-2 opensaf_reboot: Rebooting local node;
> >> timeout=60
> >>>>>>>
> >>
> ==============================================================
> >>
> ==============================================================
> >>>>>>> On 2/22/2017 12:23 PM, A V Mahesh wrote:
> >>>>>>>> Hi Vu,
> >>>>>>>>
> >>>>>>>> On 2/22/2017 12:19 PM, Vu Minh Nguyen wrote:
> >>>>>>>>> [Vu] I has sent you 02 patches. There is code change in osaflogd
> >>>>>>>>> code
> >>>>>> that
> >>>>>>>>> fix the coredump you have observed.
> >>>>>>>>> The other one is test code that fix the logtest coredump.
> >>>>>>>> Ok I will re-test , and update you .
> >>>>>>>>
> >>>>>>>> -AVM
> >>>>>>>>
> >>>>>>>> On 2/22/2017 12:19 PM, Vu Minh Nguyen wrote:
> >>>>>>>>> Hi Mahehs,
> >>>>>>>>>
> >>>>>>>>> See my reply inline, [Vu].
> >>>>>>>>>
> >>>>>>>>> Regards, Vu
> >>>>>>>>>
> >>>>>>>>>> -----Original Message-----
> >>>>>>>>>> From: A V Mahesh [mailto:mahesh.va...@oracle.com]
> >>>>>>>>>> Sent: Wednesday, February 22, 2017 1:36 PM
> >>>>>>>>>> To: Vu Minh Nguyen <vu.m.ngu...@dektech.com.au>;
> >>>>>>>>>> lennart.l...@ericsson.com; canh.v.tru...@dektech.com.au
> >>>>>>>>>> Cc: opensaf-devel@lists.sourceforge.net
> >>>>>>>>>> Subject: Re: [PATCH 0 of 3] Review Request for log: add
> >>>>>>>>>> alternative
> >>>>>>>>>> destinations of log records [#2258] V4
> >>>>>>>>>>
> >>>>>>>>>> Hi Vu,
> >>>>>>>>>>
> >>>>>>>>>>
> >>>>>>>>>> On 2/22/2017 11:52 AM, Vu Minh Nguyen wrote:
> >>>>>>>>>>> Hi Mahesh,
> >>>>>>>>>>>
> >>>>>>>>>>> Have a code fault in uml test, and other one in checkpoint.
> >>>>>>>>>> [AVM] This is Normal Suse 11 VM ( not  UML).
> >>>>>>>>>>> I have just updated the code. Please re-apply for #3 and #4
> >>>> patches.
> >>>>>>>>>> [AVM] is these new patch has function changes or only test
> code
> >>>>>>> changes ?
> >>>>>>>>> [Vu] I has sent you 02 patches. There is code change in osaflogd
> >>>>>>>>> code
> >>>>>> that
> >>>>>>>>> fix the coredump you have observed.
> >>>>>>>>> The other one is test code that fix the logtest coredump.
> >>>>>>>>>>> Note that, test case #14 of suite 17 should be run on active
> >>>>>>>>>>> node,
> >>>>>>>>>> otherwise
> >>>>>>>>>>> getting failed.
> >>>>>>>>>> [AVM]  Segmentation fault of  /usr/bin/logtest Not a big issue
,
> >>>>>>>>>>        we need to debug why  osaflogd core dumped and it is
> >>>>>>>>>> critical
> >>>>>>>>> [Vu] I found the problem. You can try with the new one to see if
> >>>>>>>>> the
> >>>>>>>>> coredump is still there or not.
> >>>>>>>>>>> I will put condition check to that test case later.
> >>>>>>>>>> -AVM
> >>>>>>>>>>
> >>>>>>>>>>
> >>>>>>>>>>> Regards, Vu
> >>>>>>>>>>>
> >>>>>>>>>>>> -----Original Message-----
> >>>>>>>>>>>> From: A V Mahesh [mailto:mahesh.va...@oracle.com]
> >>>>>>>>>>>> Sent: Wednesday, February 22, 2017 12:16 PM
> >>>>>>>>>>>> To: Vu Minh Nguyen <vu.m.ngu...@dektech.com.au>;
> >>>>>>>>>>>> lennart.l...@ericsson.com; canh.v.tru...@dektech.com.au
> >>>>>>>>>>>> Cc: opensaf-devel@lists.sourceforge.net
> >>>>>>>>>>>> Subject: Re: [PATCH 0 of 3] Review Request for log: add
> >>>> alternative
> >>>>>>>>>>>> destinations of log records [#2258] V4
> >>>>>>>>>>>>
> >>>>>>>>>>>> Hi Vu,
> >>>>>>>>>>>>
> >>>>>>>>>>>> Thanks ,
> >>>>>>>>>>>>
> >>>>>>>>>>>> While testing  /usr/bin/logtest ,  SC-2 standby osaflogd core
> >>>>> dumped
> >>>>>>>>> and
> >>>>>>>>>>>> /usr/bin/logtest on SC-1 Active
> >>>>>>>>>>>> got Segmentation fault , am I missing any other patch (  i am
> >>>> using
> >>>>>>>>>>>> devel published patch only )
> >>>>>>>>>>>>
> >>>>>>>>>>>> Following patches i am using :
> >>>>>>>>>>>>
> >>>>>>>>>>>>         1) #2293 (sent by Anders Widel, but not yet pushed)
> >>>>>>>>>>>>           2) #2258 (v2, sent by Lennart, but not yet pushed
yet)
> >>>>>>>>>>>>           3) #2258 (v4, sent by Vu, but not yet pushed yet)
> >>>>>>>>>>>>
> >>>>>>>>>>>>
> >>
> ==============================================================
> >>>>>>>>>>>> ========================================
> >>>>>>>>>>>>
> >>>>>>>>>>>>
> >>>>>>>>>>>> Core was generated by `/usr/lib64/opensaf/osaflogd'.
> >>>>>>>>>>>> Program terminated with signal 11, Segmentation fault.
> >>>>>>>>>>>> #0  ckpt_proc_cfg_stream(lgs_cb*, void*) () at
> >>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc:2195
> >>>>>>>>>>>> 2195    src/log/logd/lgs_mbcsv.cc: No such file or directory.
> >>>>>>>>>>>>                in src/log/logd/lgs_mbcsv.cc
> >>>>>>>>>>>> (gdb) bt
> >>>>>>>>>>>> #0  ckpt_proc_cfg_stream(lgs_cb*, void*) () at
> >>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc:2195
> >>>>>>>>>>>> #1  0x00007f12c3e22960 in ckpt_decode_log_struct(lgs_cb*,
> >>>>>>>>>>>> ncs_mbcsv_cb_arg*, void*, void*, unsigned int
> >> (*)(edu_hdl_tag*,
> >>>>>>>>>>>> edu_tkn_tag*, void*, unsigned int*, edu_buf_env_tag*,
> >>>>>>> EDP_OP_TYPE,
> >>>>>>>>>>>> EDU_ERR*)) () at src/log/logd/lgs_mbcsv.cc:950
> >>>>>>>>>>>> #2  0x00007f12c3e240dc in
> >> ckpt_decode_async_update(lgs_cb*,
> >>>>>>>>>>>> ncs_mbcsv_cb_arg*) () at src/log/logd/lgs_mbcsv.cc:1086
> >>>>>>>>>>>> #3  0x00007f12c3e26941 in
> >> mbcsv_callback(ncs_mbcsv_cb_arg*) ()
> >>>>> at
> >>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc:880
> >>>>>>>>>>>> #4  0x00007f12c2f25596 in ncs_mbscv_rcv_decode () from
> >>>>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0
> >>>>>>>>>>>> #5  0x00007f12c2f25766 in ncs_mbcsv_rcv_async_update ()
> >> from
> >>>>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0
> >>>>>>>>>>>> #6  0x00007f12c2f2c370 in mbcsv_process_events () from
> >>>>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0
> >>>>>>>>>>>> #7  0x00007f12c2f2c4db in mbcsv_hdl_dispatch_all () from
> >>>>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0
> >>>>>>>>>>>> #8  0x00007f12c2f26ce2 in mbcsv_process_dispatch_request
> ()
> >> at
> >>>>>>>>>>>> src/mbc/mbcsv_api.c:423
> >>>>>>>>>>>> #9  0x00007f12c3e2396e in lgs_mbcsv_dispatch(unsigned int)
> ()
> >> at
> >>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc:327
> >>>>>>>>>>>> #10 0x00007f12c3e009f2 in main () at
> >>>>>>>>>>>> src/log/logd/lgs_main.cc:583
> >>>>>>>>>>>> (gdb)
> >>>>>>>>>>>>
> >>>>>>>>>>>>
> >>
> ==============================================================
> >>>>>>>>>>>> ========================================
> >>>>>>>>>>>>
> >>>>>>>>>>>> Feb 22 10:37:06 SC-1 osafimmnd[4020]: NO Invalid error
> >> reported
> >>>>>>>>>>>> implementer 'safLogService', Ccb 161 will be aborted
> >>>>>>>>>>>> Feb 22 10:37:06 SC-1 osafimmnd[4020]: NO Ccb 161 aborted
> in
> >>>>>>>>>> COMPLETED
> >>>>>>>>>>>> processing (validation)
> >>>>>>>>>>>> Feb 22 10:37:06 SC-1 osafimmnd[4020]: NO Ccb 161
> ABORTED
> >>>>>>>>>> (immcfg_SC-
> >>>>>>>>>>>> 1_5394)
> >>>>>>>>>>>> Add values Fail
> >>>>>>>>>>>>
> >>>>>>>>>>>> Program received signal SIGSEGV, Segmentation fault.
> >>>>>>>>>>>> 0x000055555556929a in read_and_compare.isra.7 () at
> >>>>>>>>>>>> src/log/apitest/tet_LogOiOps.c:1891
> >>>>>>>>>>>> 1891    src/log/apitest/tet_LogOiOps.c: No such file or
> >>>>>>>>>>>> directory.
> >>>>>>>>>>>>                in src/log/apitest/tet_LogOiOps.c
> >>>>>>>>>>>> (gdb) Feb 22 10:37:07 SC-1 sshd[5298]: Accepted keyboard-
> >>>>>>>>>> interactive/pam
> >>>>>>>>>>>> for root from 10.176.178.22 port 51945 ssh2
> >>>>>>>>>>>> bt
> >>>>>>>>>>>> #0  0x000055555556929a in read_and_compare.isra.7 () at
> >>>>>>>>>>>> src/log/apitest/tet_LogOiOps.c:1891
> >>>>>>>>>>>> #1  0x0000555555569bbb in
> >>>>>>>>>>>> check_logRecordDestinationConfigurationEmpty
> >>>>>>>>>>>> () at src/log/apitest/tet_LogOiOps.c:2179
> >>>>>>>>>>>> #2  0x0000555555573495 in run_test_case ()
> >>>>>>>>>>>> #3  0x0000555555573934 in test_run ()
> >>>>>>>>>>>> #4  0x000055555555c7cd in main () at
> >>>>>>>>>>>> src/log/apitest/logtest.c:569
> >>>>>>>>>>>> (gdb)
> >>>>>>>>>>>>
> >>>>>>>>>>>>
> >>
> ==============================================================
> >>>>>>>>>>>> ========================================
> >>>>>>>>>>>>
> >>>>>>>>>>>> -AVM
> >>>>>>>>>>>>
> >>>>>>>>>>>> On 2/22/2017 9:48 AM, Vu Minh Nguyen wrote:
> >>>>>>>>>>>>> Hi Mahesh,
> >>>>>>>>>>>>>
> >>>>>>>>>>>>> I send them in attachment instead, and name them in the
> >> order.
> >>>>>>>>>>>>> I just pull the latest code, and apply them without getting
> > any
> >>>>>> hunk
> >>>>>>>>>>> error.
> >>>>>>>>>>>>> Please try with them, and let me know if you see any
> problem.
> >>>>>>>>>>>>>
> >>>>>>>>>>>>> Regards, Vu
> >>>>>>>>>>>>>
> >>>>>>>>>>>>>> -----Original Message-----
> >>>>>>>>>>>>>> From: A V Mahesh [mailto:mahesh.va...@oracle.com]
> >>>>>>>>>>>>>> Sent: Wednesday, February 22, 2017 11:09 AM
> >>>>>>>>>>>>>> To: Vu Minh Nguyen <vu.m.ngu...@dektech.com.au>;
> >>>>>>>>>>>>>> lennart.l...@ericsson.com;
> canh.v.tru...@dektech.com.au
> >>>>>>>>>>>>>> Cc: opensaf-devel@lists.sourceforge.net
> >>>>>>>>>>>>>> Subject: Re: [PATCH 0 of 3] Review Request for log: add
> >>>>>> alternative
> >>>>>>>>>>>>>> destinations of log records [#2258] V4
> >>>>>>>>>>>>>>
> >>>>>>>>>>>>>> Hi Vu,
> >>>>>>>>>>>>>>
> >>>>>>>>>>>>>> I did follow that still i get Hunk #2 FAILED even on
today's
> >>>>>> staging
> >>
> ==============================================================
> >>>>>>>>>>>>>> ==================
> >>>>>>>>>>>>>>
> >>>>>>>>>>>>>> [root@dhcp-hyd-scp-5fl-10-176-177-96 staging]# patch -
> p1
> >>>>> <2293
> >>>>>>>>>>>>>> patching file src/base/Makefile.am
> >>>>>>>>>>>>>> Hunk #1 succeeded at 33 (offset 1 line).
> >>>>>>>>>>>>>> Hunk #3 succeeded at 183 (offset 1 line).
> >>>>>>>>>>>>>> patching file src/base/file_descriptor.cc
> >>>>>>>>>>>>>> patching file src/base/file_descriptor.h
> >>>>>>>>>>>>>> patching file src/base/tests/unix_socket_test.cc
> >>>>>>>>>>>>>> patching file src/base/unix_client_socket.cc
> >>>>>>>>>>>>>> patching file src/base/unix_server_socket.cc
> >>>>>>>>>>>>>> patching file src/base/unix_socket.cc
> >>>>>>>>>>>>>> patching file src/base/unix_socket.h
> >>>>>>>>>>>>>>
> >>>>>>>>>>>>>> [root@dhcp-hyd-scp-5fl-10-176-177-96 staging]# patch -
> p1
> >>>>> <2258-
> >>>>>>> 1
> >>>>>>>>>>>>>> patching file src/log/Makefile.am
> >>>>>>>>>>>>>> Hunk #1 succeeded at 71 (offset -1 lines).
> >>>>>>>>>>>>>> patching file src/log/config/logsv_classes.xml
> >>>>>>>>>>>>>> Hunk #1 FAILED at 147.
> >>>>>>>>>>>>>> 1 out of 1 hunk FAILED -- saving rejects to file
> >>>>>>>>>>>>>> src/log/config/logsv_classes.xml.rej
> >>>>>>>>>>>>>> patching file src/log/logd/lgs_config.cc
> >>>>>>>>>>>>>> Hunk #1 succeeded at 35 (offset -5 lines).
> >>>>>>>>>>>>>> Hunk #2 FAILED at 705.
> >>>>>>>>>>>>>> Hunk #3 FAILED at 971.
> >>>>>>>>>>>>>> 2 out of 3 hunks FAILED -- saving rejects to file
> >>>>>>>>>>>>>> src/log/logd/lgs_config.cc.rej
> >>>>>>>>>>>>>> patching file src/log/logd/lgs_config.h
> >>>>>>>>>>>>>> Hunk #1 FAILED at 304.
> >>>>>>>>>>>>>> 1 out of 1 hunk FAILED -- saving rejects to file
> >>>>>>>>>>>>>> src/log/logd/lgs_config.h.rej
> >>>>>>>>>>>>>> patching file src/log/logd/lgs_dest.cc
> >>>>>>>>>>>>>> patching file src/log/logd/lgs_dest.h
> >>>>>>>>>>>>>> patching file src/log/logd/lgs_evt.cc
> >>>>>>>>>>>>>> patching file src/log/logd/lgs_imm.cc
> >>>>>>>>>>>>>> Hunk #1 FAILED at 45.
> >>>>>>>>>>>>>> Hunk #2 succeeded at 235 (offset -1 lines).
> >>>>>>>>>>>>>> Hunk #3 FAILED at 877.
> >>>>>>>>>>>>>> Hunk #4 succeeded at 1273 (offset -20 lines).
> >>>>>>>>>>>>>> Hunk #5 succeeded at 1404 (offset -1 lines).
> >>>>>>>>>>>>>> Hunk #6 succeeded at 1449 (offset -20 lines).
> >>>>>>>>>>>>>> Hunk #7 succeeded at 2032 (offset -1 lines).
> >>>>>>>>>>>>>> Hunk #8 FAILED at 2181.
> >>>>>>>>>>>>>> Hunk #9 succeeded at 2271 (offset -54 lines).
> >>>>>>>>>>>>>> Hunk #10 succeeded at 2387 (offset -1 lines).
> >>>>>>>>>>>>>> Hunk #11 succeeded at 2377 (offset -54 lines).
> >>>>>>>>>>>>>> Hunk #12 succeeded at 2478 (offset -1 lines).
> >>>>>>>>>>>>>> Hunk #13 succeeded at 2684 (offset -54 lines).
> >>>>>>>>>>>>>> Hunk #14 succeeded at 2821 (offset -1 lines).
> >>>>>>>>>>>>>> 3 out of 14 hunks FAILED -- saving rejects to file
> >>>>>>>>>>>>>> src/log/logd/lgs_imm.cc.rej
> >>>>>>>>>>>>>> patching file src/log/logd/lgs_main.cc
> >>>>>>>>>>>>>> patching file src/log/logd/lgs_mbcsv.cc
> >>>>>>>>>>>>>> patching file src/log/logd/lgs_mbcsv.h
> >>>>>>>>>>>>>> patching file src/log/logd/lgs_mbcsv_v5.cc
> >>>>>>>>>>>>>> Hunk #3 succeeded at 133 (offset -1 lines).
> >>>>>>>>>>>>>> patching file src/log/logd/lgs_mbcsv_v7.cc
> >>>>>>>>>>>>>> patching file src/log/logd/lgs_mbcsv_v7.h
> >>>>>>>>>>>>>> patching file src/log/logd/lgs_stream.cc
> >>>>>>>>>>>>>> patching file src/log/logd/lgs_stream.h
> >>>>>>>>>>>>>> patching file src/log/logd/lgs_util.cc
> >>>>>>>>>>>>>> patching file src/log/logd/lgs_util.h
> >>>>>>>>>>>>>>
> >>>>>>>>>>>>>> [root@dhcp-hyd-scp-5fl-10-176-177-96 staging]# patch -
> p1
> >>>>> <2258-
> >>>>>>> 2
> >>>>>>>>>>>>>> patching file src/log/Makefile.am
> >>>>>>>>>>>>>> Hunk #1 succeeded at 180 (offset -3 lines).
> >>>>>>>>>>>>>> patching file src/log/apitest/tet_LogOiOps.c
> >>>>>>>>>>>>>> Hunk #1 FAILED at 1923.
> >>>>>>>>>>>>>> Hunk #2 FAILED at 1979.
> >>>>>>>>>>>>>> Hunk #3 FAILED at 2067.
> >>>>>>>>>>>>>> Hunk #4 FAILED at 2094.
> >>>>>>>>>>>>>> 4 out of 4 hunks FAILED -- saving rejects to file
> >>>>>>>>>>>>>> src/log/apitest/tet_LogOiOps.c.rej
> >>>>>>>>>>>>>> patching file src/log/apitest/tet_cfg_destination.c
> >>>>>>>>>>>>>>
> >>>>>>>>>>>>>> [root@dhcp-hyd-scp-5fl-10-176-177-96 staging]# patch -
> p1
> >>>>> <2258-
> >>>>>>> 3
> >>>>>>>>>>>>>> patching file src/log/Makefile
> >>>>>>>>>>>>>> patching file src/log/Makefile.am
> >>>>>>>>>>>>>> Hunk #1 succeeded at 80 (offset -1 lines).
> >>>>>>>>>>>>>> Hunk #2 succeeded at 217 (offset -2 lines).
> >>>>>>>>>>>>>> patching file src/log/tests/Makefile
> >>>>>>>>>>>>>> patching file src/log/tests/lgs_dest_test.cc
> >>>>>>>>>>>>>> [root@dhcp-hyd-scp-5fl-10-176-177-96 staging]# vi
> >>>>>>>>>>>>>> src/log/apitest/tet_LogOiOps.c.rej
> >>>>>>>>>>>>>> [root@dhcp-hyd-scp-5fl-10-176-177-96 staging]#
> >>>>>>>>>>>>>>
> >>>>>>>>>>>>>>
> >>
> ==============================================================
> >>>>>>>>>>>>>> ========================
> >>>>>>>>>>>>>>
> >>>>>>>>>>>>>> -AVM
> >>>>>>>>>>>>>>
> >>>>>>>>>>>>>> On 2/21/2017 3:53 PM, Vu Minh Nguyen wrote:
> >>>>>>>>>>>>>>> Hi Mahesh,
> >>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>> As I has mentioned in below:
> >>>>>>>>>>>>>>>>>           To run the test, this patch has dependent on
> >>>> following
> >>>>>>>>> patches:
> >>>>>>>>>>>>>>>>>           1) #2293 (sent by Anders Widel, but not yet
> >>>>>>>>>>>>>>>>> pushed)
> >>>>>>>>>>>>>>>>>           2) #2258 (v2, sent by Lennart, but not yet
pushed
> >>>> yet)
> >>>>>>>>>>>>>>> So, you need to apply #2293 first, then #2258 which sent
> by
> >>>>>>> Lennart
> >>>>>>>>>>>>>>> yesterday, then mine.
> >>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>> Regards, Vu
> >>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>> -----Original Message-----
> >>>>>>>>>>>>>>>> From: A V Mahesh [mailto:mahesh.va...@oracle.com]
> >>>>>>>>>>>>>>>> Sent: Tuesday, February 21, 2017 5:10 PM
> >>>>>>>>>>>>>>>> To: Vu Minh Nguyen <vu.m.ngu...@dektech.com.au>;
> >>>>>>>>>>>>>>>> lennart.l...@ericsson.com;
> >> canh.v.tru...@dektech.com.au
> >>>>>>>>>>>>>>>> Cc: opensaf-devel@lists.sourceforge.net
> >>>>>>>>>>>>>>>> Subject: Re: [PATCH 0 of 3] Review Request for log: add
> >>>>>>> alternative
> >>>>>>>>>>>>>>>> destinations of log records [#2258] V4
> >>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>> Hi Vu,
> >>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>> Is this applies on top of   log #2146  - V4  , I see
both
> >>>>>> #tickets
> >>>>>>>>>>>>> has
> >>>>>>>>>>>>>>>> version changes ?
> >>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>> in which order i need to apply  ( #2146 & #2258 )    or
> >>>>>>>>>>>>>>>> (#2258
> >>>> &
> >>>>>>>>>>>>> #2146).
> >> =========================================================
> >>>>>>>>>>>>>>>> patching file src/log/Makefile.am
> >>>>>>>>>>>>>>>> Hunk #1 FAILED at 72.
> >>>>>>>>>>>>>>>> Hunk #2 FAILED at 120.
> >>>>>>>>>>>>>>>> 2 out of 2 hunks FAILED -- saving rejects to file
> >>>>>>>>>>>>> src/log/Makefile.am.rej
> >>>>>>>>>>>>>>>> patching file src/log/config/logsv_classes.xml
> >>>>>>>>>>>>>>>> Hunk #1 FAILED at 147.
> >>>>>>>>>>>>>>>> 1 out of 1 hunk FAILED -- saving rejects to file
> >>>>>>>>>>>>>>>> src/log/config/logsv_classes.xml.rej
> >>>>>>>>>>>>>>>> patching file src/log/logd/lgs_config.cc
> >>>>>>>>>>>>>>>> Hunk #1 succeeded at 35 (offset -5 lines).
> >>>>>>>>>>>>>>>> Hunk #2 FAILED at 705.
> >>>>>>>>>>>>>>>> Hunk #3 FAILED at 971.
> >>>>>>>>>>>>>>>> 2 out of 3 hunks FAILED -- saving rejects to file
> >>>>>>>>>>>>>>>> src/log/logd/lgs_config.cc.rej
> >>>>>>>>>>>>>>>> patching file src/log/logd/lgs_config.h
> >>>>>>>>>>>>>>>> Hunk #1 FAILED at 304.
> >>>>>>>>>>>>>>>> 1 out of 1 hunk FAILED -- saving rejects to file
> >>>>>>>>>>>>>>>> src/log/logd/lgs_config.h.rej
> >>>>>>>>>>>>>>>> patching file src/log/logd/lgs_dest.cc
> >>>>>>>>>>>>>>>> patching file src/log/logd/lgs_dest.h
> >>>>>>>>>>>>>>>> patching file src/log/logd/lgs_evt.cc
> >>>>>>>>>>>>>>>> Hunk #1 FAILED at 1.
> >>>>>>>>>>>>>>>> Hunk #2 succeeded at 30 with fuzz 2 (offset 2 lines).
> >>>>>>>>>>>>>>>> Hunk #3 succeeded at 1282 (offset 45 lines).
> >>>>>>>>>>>>>>>> Hunk #4 succeeded at 1300 (offset 2 lines).
> >>>>>>>>>>>>>>>> 1 out of 4 hunks FAILED -- saving rejects to file
> >>>>>>>>>>>>>>>> src/log/logd/lgs_evt.cc.rej
> >>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>
> >>
> ==============================================================
> >>>>>>>>>>>>>>>> ===
> >>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>> -AVM
> >>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>> On 2/21/2017 3:03 PM, Vu Minh Nguyen wrote:
> >>>>>>>>>>>>>>>>> Summary: log: add alternative destinations of log
> records
> >>>>>>> [#2258]
> >>>>>>>>>>>>>>>>> Review request for Trac Ticket(s): #2258
> >>>>>>>>>>>>>>>>> Peer Reviewer(s): Lennart, Canh, Mahesh
> >>>>>>>>>>>>>>>>> Pull request to: <<LIST THE PERSON WITH PUSH
> ACCESS
> >>>>> HERE>>
> >>>>>>>>>>>>>>>>> Affected branch(es): Default
> >>>>>>>>>>>>>>>>> Development branch: Default
> >>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>> --------------------------------
> >>>>>>>>>>>>>>>>> Impacted area       Impact y/n
> >>>>>>>>>>>>>>>>> --------------------------------
> >>>>>>>>>>>>>>>>>           Docs                    n
> >>>>>>>>>>>>>>>>>           Build system            n
> >>>>>>>>>>>>>>>>>           RPM/packaging           n
> >>>>>>>>>>>>>>>>>           Configuration files     n
> >>>>>>>>>>>>>>>>>           Startup scripts         n
> >>>>>>>>>>>>>>>>>           SAF services            n
> >>>>>>>>>>>>>>>>>           OpenSAF services        y
> >>>>>>>>>>>>>>>>>           Core libraries          n
> >>>>>>>>>>>>>>>>>           Samples                 n
> >>>>>>>>>>>>>>>>>           Tests                   y
> >>>>>>>>>>>>>>>>>           Other                   n
> >>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>> Comments (indicate scope for each "y" above):
> >>>>>>>>>>>>>>>>> ---------------------------------------------
> >>>>>>>>>>>>>>>>>           To run the test, this patch has dependent on
> >>>> following
> >>>>>>>>> patches:
> >>>>>>>>>>>>>>>>>           1) #2293 (sent by Anders Widel, but not yet
> >>>>>>>>>>>>>>>>> pushed)
> >>>>>>>>>>>>>>>>>           2) #2258 (v2, sent by Lennart, but not yet
pushed
> >>>> yet)
> >>>>>>>>>>>>>>>>> changeset
> >> d74aaf3025c99cade3165a15831124548f4d85bd
> >>>>>>>>>>>>>>>>> Author:    Vu Minh Nguyen
> >> <vu.m.ngu...@dektech.com.au>
> >>>>>>>>>>>>>>>>> Date:    Wed, 15 Feb 2017 14:36:00 +0700
> >>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>      log: add alternative destinations of log records
> >>>>>>>>>>>>>>>>> [#2258]
> >>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>      Here are major info, detailed info will be added to
> >>>>>>>>>>>>>>>>> PR doc
> >>>>>>>>>>> soon. 1)
> >>>>>>>>>>>>>>>> Add
> >>>>>>>>>>>>>>>>>      attribute "saLogRecordDestination" to log stream.
2)
> >>>>>>>>>>>>>>>>> Add
> >>>>>>>>>>> Local
> >>>>>>>>>>>>>>>> socket
> >>>>>>>>>>>>>>>>>      destintion handler 3) Integrate into first
increment
> >>>>>>>>>>>>>>>>> made by
> >>>>>>>>>>> Lennart
> >>>>>>>>>>>>>>>>> changeset
> >> 4bae27a478c235df3058f43c92d3a5483233b01d
> >>>>>>>>>>>>>>>>> Author:    Vu Minh Nguyen
> >> <vu.m.ngu...@dektech.com.au>
> >>>>>>>>>>>>>>>>> Date:    Wed, 15 Feb 2017 15:07:09 +0700
> >>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>      log: add UML test case to verify alternative
> >>>>>>>>>>>>>>>>> destination
> >>>>>>>>>>> [#2258]
> >>>>>>>>>>>>>>>>>      Major changes: 1) Modify Lennart's test cases
> because
> >>>>>>>>>>> enhancing
> >>>>>>>>>>>>>>>> destination
> >>>>>>>>>>>>>>>>>      configuration validation rules. 2) Add test suite
> >>>>>>>>>>>>>>>>> #17 to
> >>>>>>>>>>> verify
> >>>>>>>>>>>>>>>> alternative
> >>>>>>>>>>>>>>>>>      destination
> >>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>> changeset
> bc375725fed22bb4f8cb3ae3df5f96fb9d281efb
> >>>>>>>>>>>>>>>>> Author:    Vu Minh Nguyen
> >> <vu.m.ngu...@dektech.com.au>
> >>>>>>>>>>>>>>>>> Date:    Thu, 16 Feb 2017 17:22:13 +0700
> >>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>      log: add unit tests to verify interfaces provided
by
> >>>>>>>>>>> destination
> >>>>>>>>>>>>>>>> handler
> >>>>>>>>>>>>>>>>>      [#2258]
> >>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>      Unit tests to verify major interfaces: 1)
> >>>>>>>>>>>>>>>>> CfgDestination()
> >>>>>>>>>>> 2)
> >>>>>>>>>>>>>>>>>      WriteToDestination()
> >>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>> Added Files:
> >>>>>>>>>>>>>>>>> ------------
> >>>>>>>>>>>>>>>>> src/log/apitest/tet_cfg_destination.c
> >>>>>>>>>>>>>>>>>           src/log/logd/lgs_dest.cc
> >>>>>>>>>>>>>>>>>           src/log/logd/lgs_dest.h
> >>>>>>>>>>>>>>>>>           src/log/logd/lgs_mbcsv_v7.cc
> >>>>>>>>>>>>>>>>>           src/log/logd/lgs_mbcsv_v7.h
> >>>>>>>>>>>>>>>>> src/log/tests/lgs_dest_test.cc
> >>>>>>>>>>>>>>>>>           src/log/tests/Makefile
> >>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>> Complete diffstat:
> >>>>>>>>>>>>>>>>> ------------------
> >>>>>>>>>>>>>>>>> src/log/Makefile |    4 +
> >>>>>>>>>>>>>>>>> src/log/Makefile.am |   31 +++++-
> >>>>>>>>>>>>>>>>> src/log/apitest/tet_LogOiOps.c |    8 +-
> >>>>>>>>>>>>>>>>> src/log/apitest/tet_cfg_destination.c |  483
> >>
> ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++
> >>>>>>>>>>>>>>>> ++++++++++++++++++++++++++++++++++
> >>>>>>>>>>>>>>>>> src/log/config/logsv_classes.xml |    7 +-
> >>>>>>>>>>>>>>>>> src/log/logd/lgs_config.cc |  169
> >>>>>>>>>>>>>>>> +++++++++++++++++++++++++++++---
> >>>>>>>>>>>>>>>>> src/log/logd/lgs_config.h |    3 +-
> >>>>>>>>>>>>>>>>> src/log/logd/lgs_dest.cc |  707
> >>
> ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++
> >>
> ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++
> >>>>>>>>>>>>>>>> +++++++++++++++++
> >>>>>>>>>>>>>>>>> src/log/logd/lgs_dest.h |  576
> >>
> ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++
> >>>>> ++++++++++++++++++++++++++++++++++++++++++++++++++++
> >>>>>>>>>>>>>>>>> src/log/logd/lgs_evt.cc |   33 ++++++
> >>>>>>>>>>>>>>>>> src/log/logd/lgs_imm.cc |  202
> >>>>>>>>>>>>>>>> +++++++++++++++++++++++++++++++++------
> >>>>>>>>>>>>>>>>> src/log/logd/lgs_main.cc |    8 +
> >>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc |  103
> >>>>>>>>> ++++++++++++++++++-
> >>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv.h |    6 +-
> >>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv_v5.cc |   10 +
> >>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv_v7.cc |  177
> >>>>>>>>>>>>>>>> +++++++++++++++++++++++++++++++++++
> >>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv_v7.h |   67
> >>>>>> +++++++++++++
> >>>>>>>>>>>>>>>>> src/log/logd/lgs_stream.cc |   60
> >>>>>> +++++++++++-
> >>>>>>>>>>>>>>>>> src/log/logd/lgs_stream.h |   16 +++
> >>>>>>>>>>>>>>>>> src/log/logd/lgs_util.cc |   63
> >>>>>> ++++++++++++
> >>>>>>>>>>>>>>>>> src/log/logd/lgs_util.h |   11 +-
> >>>>>>>>>>>>>>>>> src/log/tests/Makefile |   20 +++
> >>>>>>>>>>>>>>>>> src/log/tests/lgs_dest_test.cc |  209
> >>>>>>>>>>>>>>>> +++++++++++++++++++++++++++++++++++++++++
> >>>>>>>>>>>>>>>>>           23 files changed, 2896 insertions(+), 77
> >>>>>>>>>>>>>>>>> deletions(-)
> >>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>> Testing Commands:
> >>>>>>>>>>>>>>>>> -----------------
> >>>>>>>>>>>>>>>>>           Run UML test suite #17
> >>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>> Testing, Expected Results:
> >>>>>>>>>>>>>>>>> --------------------------
> >>>>>>>>>>>>>>>>>           All test passed
> >>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>> Conditions of Submission:
> >>>>>>>>>>>>>>>>> -------------------------
> >>>>>>>>>>>>>>>>>           <<HOW MANY DAYS BEFORE PUSHING,
> CONSENSUS
> >>>>> ETC>>
> >>>>>>>>>>>>>>>>> Arch      Built     Started    Linux distro
> >>>>>>>>>>>>>>>>> -------------------------------------------
> >>>>>>>>>>>>>>>>> mips        n          n
> >>>>>>>>>>>>>>>>> mips64      n          n
> >>>>>>>>>>>>>>>>> x86         n          n
> >>>>>>>>>>>>>>>>> x86_64      n          n
> >>>>>>>>>>>>>>>>> powerpc     n          n
> >>>>>>>>>>>>>>>>> powerpc64   n          n
> >>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>> Reviewer Checklist:
> >>>>>>>>>>>>>>>>> -------------------
> >>>>>>>>>>>>>>>>> [Submitters: make sure that your review doesn't trigger
> >> any
> >>>>>>>>>>>>>> checkmarks!]
> >>>>>>>>>>>>>>>>> Your checkin has not passed review because (see
> checked
> >>>>>>> entries):
> >>>>>>>>>>>>>>>>> ___ Your RR template is generally incomplete; it has
> too
> >>>>> many
> >>>>>>>>> blank
> >>>>>>>>>>>>>>> entries
> >>>>>>>>>>>>>>>>>              that need proper data filled in.
> >>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>> ___ You have failed to nominate the proper persons for
> >>>>> review
> >>>>>>> and
> >>>>>>>>>>>>> push.
> >>>>>>>>>>>>>>>>> ___ Your patches do not have proper short+long
> header
> >>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>> ___ You have grammar/spelling in your header that is
> >>>>>>> unacceptable.
> >>>>>>>>>>>>>>>>> ___ You have exceeded a sensible line length in your
> >>>>>>>>>>>>>>>> headers/comments/text.
> >>>>>>>>>>>>>>>>> ___ You have failed to put in a proper Trac Ticket #
into
> >>>> your
> >>>>>>>>>>>>> commits.
> >>>>>>>>>>>>>>>>> ___ You have incorrectly put/left internal data in your
> >>>>>>>>>>> comments/files
> >>>>>>>>>>>>>>>>> (i.e. internal bug tracking tool IDs, product
> >>>> names
> >>>>>> etc)
> >>>>>>>>>>>>>>>>> ___ You have not given any evidence of testing beyond
> >> basic
> >>>>>>> build
> >>>>>>>>>>>>> tests.
> >>>>>>>>>>>>>>>>> Demonstrate some level of runtime or other sanity
> >>>>>> testing.
> >>>>>>>>>>>>>>>>> ___ You have ^M present in some of your files. These
> >>>>>>>>>>>>>>>>> have to
> >>>>>>> be
> >>>>>>>>>>>>>> removed.
> >>>>>>>>>>>>>>>>> ___ You have needlessly changed whitespace or added
> >>>>>>> whitespace
> >>>>>>>>>>>>>> crimes
> >>>>>>>>>>>>>>>>>              like trailing spaces, or spaces before
tabs.
> >>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>> ___ You have mixed real technical changes with
> >> whitespace
> >>>>> and
> >>>>>>>>>> other
> >>>>>>>>>>>>>>>>> cosmetic code cleanup changes. These have to be
> >>>>>>> separate
> >>>>>>>>>>>>> commits.
> >>>>>>>>>>>>>>>>> ___ You need to refactor your submission into logical
> >>>>>>>>>>>>>>>>> chunks;
> >>>>>>>>> there
> >>>>>>>>>>> is
> >>>>>>>>>>>>>>>>>              too much content into a single commit.
> >>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>> ___ You have extraneous garbage in your review
> (merge
> >>>>>>> commits
> >>>>>>>>>> etc)
> >>>>>>>>>>>>>>>>> ___ You have giant attachments which should never
> have
> >>>>> been
> >>>>>>>>>> sent;
> >>>>>>>>>>>>>>>>> Instead you should place your content in a public
> >>>>>> tree to
> >>>>>>>>> be
> >>>>>>>>>>>>>>> pulled.
> >>>>>>>>>>>>>>>>> ___ You have too many commits attached to an e-mail;
> >>>>> resend
> >>>>>>> as
> >>>>>>>>>>>>>> threaded
> >>>>>>>>>>>>>>>>> commits, or place in a public tree for a pull.
> >>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>> ___ You have resent this content multiple times
> without a
> >>>>> clear
> >>>>>>>>>>>>>>> indication
> >>>>>>>>>>>>>>>>>              of what has changed between each re-send.
> >>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>> ___ You have failed to adequately and individually
> address
> >>>> all
> >>>>>> of
> >>>>>>>>>>> the
> >>>>>>>>>>>>>>>>> comments and change requests that were proposed in
> >>>>>> the
> >>>>>>>>>>> initial
> >>>>>>>>>>>>>>> review.
> >>>>>>>>>>>>>>>>> ___ You have a misconfigured ~/.hgrc file (i.e.
> username,
> >>>>> email
> >>>>>>>>> etc)
> >>>>>>>>>>>>>>>>> ___ Your computer have a badly configured date and
> >> time;
> >>>>>>>>>> confusing
> >>>>>>>>>>>> the
> >>>>>>>>>>>>>>>>>              the threaded patch review.
> >>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>> ___ Your changes affect IPC mechanism, and you don't
> >>>>> present
> >>>>>>> any
> >>>>>>>>>>>>>> results
> >>>>>>>>>>>>>>>>>              for in-service upgradability test.
> >>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>> ___ Your changes affect user manual and
> documentation,
> >>>>> your
> >>>>>>>>>> patch
> >>>>>>>>>>>>>> series
> >>>>>>>>>>>>>>>>>              do not contain the patch that updates the
> >>>>>>>>>>>>>>>>> Doxygen
> >>>>>>> manual.
> >
----------------------------------------------------------------------------
> >>>>>> --
> >>>>>>>> Check out the vibrant tech community on one of the world's most
> >>>>>>>> engaging tech sites, SlashDot.org! http://sdm.link/slashdot
> >>>>>>>> _______________________________________________
> >>>>>>>> Opensaf-devel mailing list
> >>>>>>>> Opensaf-devel@lists.sourceforge.net
> >>>>>>>> https://lists.sourceforge.net/lists/listinfo/opensaf-devel



------------------------------------------------------------------------------
Check out the vibrant tech community on one of the world's most
engaging tech sites, SlashDot.org! http://sdm.link/slashdot
_______________________________________________
Opensaf-devel mailing list
Opensaf-devel@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/opensaf-devel

Reply via email to