Hi Vu,

On 2/23/2017 10:20 AM, Vu Minh Nguyen wrote:
> Hi Mahesh,
>
> Can you try with 2258_v7.patch I just sent to you?
I striped    changeset:   8610 of today's latest staging  (  `hg strip 
8610` which  removed log: implement SaLogFilterSetCallbackT and version 
handling [#2146])
and try to apply your `2258_v7.patch`, it says `malformed patch at line 
3324`.

-AVM
>
> I have pulled the latest code on OpenSAF 5.1 branch, re-created the cluster.
> And it works with the case old active SC-1 (OpenSAF 5.1) and new standby
> SC-2 (with 2258_v7.patch included in).
>
> To apply 2258_v7.patch, please do remove the just pushed ticket "log:
> implement SaLogFilterSetCallbackT and version handling [#2146]" ,
> I have not rebased the code on that yet.
>
> Regards, Vu
>
>> -----Original Message-----
>> From: A V Mahesh [mailto:mahesh.va...@oracle.com]
>> Sent: Thursday, February 23, 2017 11:45 AM
>> To: Vu Minh Nguyen <vu.m.ngu...@dektech.com.au>;
>> lennart.l...@ericsson.com; canh.v.tru...@dektech.com.au
>> Cc: opensaf-devel@lists.sourceforge.net
>> Subject: Re: [devel] [PATCH 0 of 3] Review Request for log: add
> alternative
>> destinations of log records [#2258] V4
>>
>> Hi Vu/Lennart,
>>
>>
>> In broad WITHOUT the #2258 patch, the same code/setup working fine with
>> 2 sc node (staging  changeset:   8609 ),
>> as soon as we apply `2258_v5.patch` V5 patch on staging (changeset:
>> 8609 ) that you have provided yesterday,
>> on one sc node and try to bring up that in to cluster (in-service test)
>> we are observing the issue of new node (with #2258 patch) not joining
>> cluster.
>>
>> ==============================================================
>> ====================================================
>> eb 23 10:01:59 SC-1 osafimmnd[15279]: NO Implementer (applier)
>> connected: 15 (@safAmfService2010f) <127, 2010f>
>> Feb 23 10:01:59 SC-1 osaflogd[15289]: NO LOGSV_DATA_GROUPNAME not
>> found
>> Feb 23 10:01:59 SC-1 osaflogd[15289]: NO LOG root directory is:
>> "/var/log/opensaf/saflog"
>> Feb 23 10:01:59 SC-1 osaflogd[15289]: NO LOG data group is: ""
>> Feb 23 10:01:59 SC-1 osaflogd[15289]: NO LGS_MBCSV_VERSION = 7
>> Feb 23 10:01:59 SC-1 osafamfnd[15329]: NO Assigned
>> 'safSi=SC-2N,safApp=OpenSAF' STANDBY to 'safSu=SC-
>> 1,safSg=2N,safApp=OpenSAF'
>> Feb 23 10:01:59 SC-1 opensafd: OpenSAF(5.1.M0 - ) services successfully
>> started
>> Feb 23 10:01:59 SC-1 osafamfnd[15329]: NO
>> 'safComp=LOG,safSu=SC-1,safSg=2N,safApp=OpenSAF' faulted due to
>> 'avaDown' : Recovery is 'nodeFailfast'
>> Feb 23 10:01:59 SC-1 osafamfnd[15329]: ER
>> safComp=LOG,safSu=SC-1,safSg=2N,safApp=OpenSAF Faulted due
>> to:avaDown
>> Recovery is:nodeFailfast
>> Feb 23 10:01:59 SC-1 osafamfnd[15329]: Rebooting OpenSAF NodeId =
>> 131343
>> EE Name = , Reason: Component faulted: recovery is node failfast,
>> OwnNodeId = 131343, SupervisionTime = 60
>> Feb 23 10:01:59 SC-1 opensaf_reboot: Rebooting local node; timeout=60
>> Feb 23 10:02:00 SC-1 osafimmnd[15279]: NO Implementer (applier)
>> connected: 16 (@OpenSafImmReplicatorB) <144, 2010f>
>> Feb 23 10:01:59 SC-1 opensaf_reboot: Rebooting local node; timeout=60
>> ==============================================================
>> ====================================================
>>
>> So it is evident that in-service upgrade part code of this need to be
>> corrected.
>>
>> Please see my comments as [AVM] and let me know if you need some traces
>> .
>>
>> If you're planing to prepare new V6 patch , please do prepare on top of
>> today's latest staging.
>>
>> On 2/23/2017 9:33 AM, Vu Minh Nguyen wrote:
>>> Hi Mahesh,
>>>
>>> I have done in-service upgrade/downgrade with following cases:
>>> 1) New Active SC-1 (OpenSAF 5.2 with the attached patch) + old standby
>> SC-2
>>> (OpenSAF 5.1)
>>> --> Work fine
>> [AVM] This is not a practical use cause of in-service upgrade , we can
>> ignore this test further
>>> 2) Old Active SC-1 (OpenSAF 5.1) + new standby SC-2 (with or without
>>> attached patch)
>>> --> SC-2 is restarted & not able to join the cluster.
>> [AVM] This use cause/flow is  we do get in in-service upgrade , so we
>> need to address this.
>>> I got following messages in syslog:
>>> Feb 23 09:32:42 SC-2 user.notice opensafd: OpenSAF(5.2.M0 -
>>> 8529:b5addd36e45d:default) services successfully started
>>> Feb 23 09:32:43 SC-2 local0.warn osafntfimcnd[701]: WA
>> ntfimcn_imm_init
>>> saImmOiImplementerSet() returned SA_AIS_ERR_TIMEOUT (5)
>>> Feb 23 09:32:45 SC-2 local0.warn osafntfimcnd[701]: WA
>> ntfimcn_imm_init
>>> saImmOiImplementerSet() returned SA_AIS_ERR_TIMEOUT (5)
>>> Feb 23 09:32:47 SC-2 local0.warn osafntfimcnd[701]: WA
>> ntfimcn_imm_init
>>> saImmOiImplementerSet() returned SA_AIS_ERR_TIMEOUT (5)
>>> Feb 23 09:32:49 SC-2 local0.warn osafntfimcnd[701]: WA
>> ntfimcn_imm_init
>>> saImmOiImplementerSet() returned SA_AIS_ERR_TIMEOUT (5)
>>> Feb 23 09:32:50 SC-2 local0.err osafmsgnd[592]: ER
>> saImmOiImplementerSet
>>> FAILED:5
>>> Feb 23 09:32:50 SC-2 local0.err osafmsgnd[592]: ER
>> saImmOiImplementerSet
>>> FAILED:5
>>> Feb 23 09:32:50 SC-2 local0.notice osafamfnd[496]: NO
>>> 'safSu=SC-2,safSg=NoRed,safApp=OpenSAF' component restart probation
>> timer
>>> started (timeout: 60000000000 ns)
>>> Feb 23 09:32:50 SC-2 local0.notice osafamfnd[496]: NO Restarting a
>> component
>>> of 'safSu=SC-2,safSg=NoRed,safApp=OpenSAF' (comp restart count: 1)
>>> Feb 23 09:32:50 SC-2 local0.notice osafamfnd[496]: NO
>>> 'safComp=MQND,safSu=SC-2,safSg=NoRed,safApp=OpenSAF' faulted due
>> to
>>> 'avaDown' : Recovery is 'componentRestart'
>>> Feb 23 09:32:50 SC-2 local0.info osafmsgnd[736]: mkfifo already exists:
>>> /var/lib/opensaf/osafmsgnd.fifo File exists
>>>
>>> And sometimes, on active SC-1 (OpenSAF 5.1), the node is not able to up
>>> because of following error:
>>>
>>> Feb 23 11:00:32 SC-1 local0.err osafclmna[406]: MDTM:TIPC Dsock Socket
>>> creation failed in MDTM_INIT err :Address family not supported by
>> protocol
>>> Feb 23 11:00:32 SC-1 local0.err osafclmna[406]: ER ncs_agents_startup
>> FAILED
>> [AVM]  No such issues ( with both TCP & TIPC)  (staging changeset:   8609
> )
>>> Are you getting similar problem at your side?
>>> Please note that, the problem is existed WITH or WITHOUT the #2258
>> patch.
>> [AVM] No , problem only if we apply `2258_v5.patch` V5 patch on staging
>> (changeset:   8609 )
>>              try to bring up that node in to cluster.
>>
>>
>> -AVM
>>
>>> I have informed this to IMM to have a look, not sure any problem with
> MDS
>>> layer or any problem with my environment setup.
>>> In the meantime, please have a look at the updated patch,  I will
> continue
>>> checking the problem. Will keep you updated.
>> [AVM] I haven't seen any IMM problems
>>> Regards, Vu
>>>
>>>> -----Original Message-----
>>>> From: A V Mahesh [mailto:mahesh.va...@oracle.com]
>>>> Sent: Wednesday, February 22, 2017 5:36 PM
>>>> To: Vu Minh Nguyen <vu.m.ngu...@dektech.com.au>;
>>>> lennart.l...@ericsson.com; canh.v.tru...@dektech.com.au
>>>> Cc: opensaf-devel@lists.sourceforge.net
>>>> Subject: Re: [devel] [PATCH 0 of 3] Review Request for log: add
>>> alternative
>>>> destinations of log records [#2258] V4
>>>>
>>>> Please see correction New Standby SC-1 ( with patch )
>>>>
>>>> -AVM
>>>>
>>>> On 2/22/2017 4:02 PM, A V Mahesh wrote:
>>>>> Hi Vu,
>>>>>
>>>>> With this new patch , we have another issue :
>>>>>
>>>>> 1)  standby Core by `/usr/lib64/opensaf/osaflogd'  issue got resolved
> .
>>>>> 2) In-service upgrade is Not working , I have Old  Active SC-2 ( with
>>>>> out patch )  and New Standby SC-1 ( with patch )
>>>>>
>>>>>       the new New Standby SC-1 not joining the cluster ( in-service
>>>>> upgrade  failed )
>>>>>
>>>>> New Standby SC-1
>>>>>
>>>>>
>> ==============================================================
>>>> ======================================================
>>>>> Feb 22 15:53:05 SC-1 osafamfnd[15752]: NO
>>>>> 'safSu=SC-1,safSg=NoRed,safApp=OpenSAF' Presence State
>>>> INSTANTIATING
>>>>> => INSTANTIATED
>>>>> Feb 22 15:53:05 SC-1 osafamfnd[15752]: NO Assigning
>>>>> 'safSi=NoRed4,safApp=OpenSAF' ACTIVE to
>>>>> 'safSu=SC-1,safSg=NoRed,safApp=OpenSAF'
>>>>> Feb 22 15:53:05 SC-1 osafamfnd[15752]: NO Assigned
>>>>> 'safSi=NoRed4,safApp=OpenSAF' ACTIVE to
>>>>> 'safSu=SC-1,safSg=NoRed,safApp=OpenSAF'
>>>>> Feb 22 15:53:05 SC-1 osafsmfd[15889]: Started
>>>>> Feb 22 15:53:05 SC-1 osafamfnd[15752]: NO
>>>>> 'safSu=SC-1,safSg=2N,safApp=OpenSAF' Presence State INSTANTIATING
>> =>
>>>>> INSTANTIATED
>>>>> Feb 22 15:53:05 SC-1 osafamfnd[15752]: NO Assigning
>>>>> 'safSi=SC-2N,safApp=OpenSAF' STANDBY to
>>>>> 'safSu=SC-1,safSg=2N,safApp=OpenSAF'
>>>>> Feb 22 15:53:05 SC-1 osafrded[15672]: NO RDE role set to STANDBY
>>>>> Feb 22 15:53:05 SC-1 osafrded[15672]: NO Peer up on node 0x2020f
>>>>> Feb 22 15:53:05 SC-1 osafrded[15672]: NO Got peer info request from
>>>>> node 0x2020f with role ACTIVE
>>>>> Feb 22 15:53:05 SC-1 osafrded[15672]: NO Got peer info response from
>>>>> node 0x2020f with role ACTIVE
>>>>> Feb 22 15:53:05 SC-1 osafimmd[15691]: NO MDS event from svc_id 24
>>>>> (change:5, dest:13)
>>>>> Feb 22 15:53:05 SC-1 osafimmd[15691]: NO MDS event from svc_id 24
>>>>> (change:3, dest:13)
>>>>> Feb 22 15:53:05 SC-1 osafimmd[15691]: NO MDS event from svc_id 24
>>>>> (change:5, dest:13)
>>>>> Feb 22 15:53:05 SC-1 osafimmd[15691]: NO MDS event from svc_id 25
>>>>> (change:3, dest:567412424453430)
>>>>> Feb 22 15:53:05 SC-1 osafimmd[15691]: NO MDS event from svc_id 25
>>>>> (change:3, dest:565213401202663)
>>>>> Feb 22 15:53:05 SC-1 osafimmd[15691]: NO MDS event from svc_id 25
>>>>> (change:3, dest:566312912825221)
>>>>> Feb 22 15:53:05 SC-1 osafimmd[15691]: NO MDS event from svc_id 25
>>>>> (change:3, dest:564113889574230)
>>>>> Feb 22 15:53:05 SC-1 osafimmnd[15702]: NO Implementer (applier)
>>>>> connected: 17 (@safAmfService2010f) <127, 2010f>
>>>>> Feb 22 15:53:05 SC-1 osaflogd[15712]: NO LOGSV_DATA_GROUPNAME
>>>> not found
>>>>> Feb 22 15:53:05 SC-1 osaflogd[15712]: NO LOG root directory is:
>>>>> "/var/log/opensaf/saflog"
>>>>> Feb 22 15:53:05 SC-1 osaflogd[15712]: NO LOG data group is: ""
>>>>> Feb 22 15:53:05 SC-1 osaflogd[15712]: NO LGS_MBCSV_VERSION = 7
>>>>> Feb 22 15:53:05 SC-1 osafamfnd[15752]: NO Assigned
>>>>> 'safSi=SC-2N,safApp=OpenSAF' STANDBY to
>>>>> 'safSu=SC-1,safSg=2N,safApp=OpenSAF'
>>>>> Feb 22 15:53:05 SC-1 opensafd: OpenSAF(5.1.M0 - ) services
>>>>> successfully started
>>>>> Feb 22 15:53:05 SC-1 osafamfnd[15752]: NO
>>>>> 'safComp=LOG,safSu=SC-1,safSg=2N,safApp=OpenSAF' faulted due to
>>>>> 'avaDown' : Recovery is 'nodeFailfast'
>>>>> Feb 22 15:53:05 SC-1 osafamfnd[15752]: ER
>>>>> safComp=LOG,safSu=SC-1,safSg=2N,safApp=OpenSAF Faulted due
>>>> to:avaDown
>>>>> Recovery is:nodeFailfast
>>>>> Feb 22 15:53:05 SC-1 osafamfnd[15752]: Rebooting OpenSAF NodeId =
>>>>> 131343 EE Name = , Reason: Component faulted: recovery is node
>>>>> failfast, OwnNodeId = 131343, SupervisionTime = 60
>>>>> Feb 22 15:53:05 SC-1 opensaf_reboot: Rebooting local node;
>> timeout=60
>>>>> Feb 22 15:53:43 SC-1 syslog-ng[1171]: syslog-ng starting up;
>>>>> version='2.0.9'
>>>>>
>>>>>
>> ==============================================================
>>>> ======================================================
>>>>> Old - Active - SC-2
>>>>>
>>>>>
>> ==============================================================
>>>> ======================================================
>>>>> Feb 22 15:53:02 SC-2 osafimmnd[16359]: NO NODE STATE->
>>>>> IMM_NODE_R_AVAILABLE
>>>>> Feb 22 15:53:02 SC-2 osafimmloadd: NO Sync starting
>>>>> Feb 22 15:53:02 SC-2 osafimmloadd: IN Synced 390 objects in total
>>>>> Feb 22 15:53:02 SC-2 osafimmnd[16359]: NO NODE STATE->
>>>>> IMM_NODE_FULLY_AVAILABLE 18511
>>>>> Feb 22 15:53:02 SC-2 osafimmnd[16359]: NO Epoch set to 3 in
>> ImmModel
>>>>> Feb 22 15:53:02 SC-2 osafimmd[16346]: NO ACT: New Epoch for
>> IMMND
>>>>> process at node 2020f old epoch: 2  new epoch:3
>>>>> Feb 22 15:53:02 SC-2 osafimmd[16346]: NO ACT: New Epoch for
>> IMMND
>>>>> process at node 2040f old epoch: 2  new epoch:3
>>>>> Feb 22 15:53:02 SC-2 osafimmd[16346]: NO ACT: New Epoch for
>> IMMND
>>>>> process at node 2030f old epoch: 2  new epoch:3
>>>>> Feb 22 15:53:02 SC-2 osafimmloadd: NO Sync ending normally
>>>>> Feb 22 15:53:02 SC-2 osafimmd[16346]: NO ACT: New Epoch for
>> IMMND
>>>>> process at node 2010f old epoch: 0  new epoch:3
>>>>> Feb 22 15:53:02 SC-2 osafimmnd[16359]: NO SERVER STATE:
>>>>> IMM_SERVER_SYNC_SERVER --> IMM_SERVER_READY
>>>>> Feb 22 15:53:03 SC-2 osafamfd[16408]: NO Received node_up from
>> 2010f:
>>>>> msg_id 1
>>>>> Feb 22 15:53:03 SC-2 osafamfd[16408]: NO Node 'SC-1' joined the
>> cluster
>>>>> Feb 22 15:53:03 SC-2 osafimmnd[16359]: NO Implementer connected:
>> 16
>>>>> (MsgQueueService131343) <0, 2010f>
>>>>> Feb 22 15:53:03 SC-2 osafrded[16327]: NO Peer up on node 0x2010f
>>>>> Feb 22 15:53:03 SC-2 osafrded[16327]: NO Got peer info request from
>>>>> node 0x2010f with role STANDBY
>>>>> Feb 22 15:53:03 SC-2 osafrded[16327]: NO Got peer info response from
>>>>> node 0x2010f with role STANDBY
>>>>> Feb 22 15:53:03 SC-2 osafimmd[16346]: NO MDS event from svc_id 24
>>>>> (change:5, dest:13)
>>>>> Feb 22 15:53:03 SC-2 osafimmnd[16359]: NO Implementer (applier)
>>>>> connected: 17 (@safAmfService2010f) <0, 2010f>
>>>>> Feb 22 15:53:03 SC-2 osafamfd[16408]: NO Cluster startup is done
>>>>> Feb 22 15:53:04 SC-2 osafimmnd[16359]: NO Implementer (applier)
>>>>> connected: 18 (@OpenSafImmReplicatorB) <0, 2010f>
>>>>> Feb 22 15:53:09 SC-2 osafdtmd[16304]: NO Lost contact with 'SC-1'
>>>>> Feb 22 15:53:09 SC-2 osaffmd[16336]: NO Node Down event for node id
>>>>> 2010f:
>>>>> Feb 22 15:53:09 SC-2 osafimmd[16346]: NO MDS event from svc_id 24
>>>>> (change:6, dest:13)
>>>>> Feb 22 15:53:09 SC-2 osafimmd[16346]: NO MDS event from svc_id 25
>>>>> (change:4, dest:564113889574230)
>>>>> Feb 22 15:53:09 SC-2 osaffmd[16336]: NO Current role: ACTIVE
>>>>> Feb 22 15:53:09 SC-2 osaffmd[16336]: Rebooting OpenSAF NodeId =
>>>> 131343
>>>>> EE Name = , Reason: Received Node Down for peer controller,
>> OwnNodeId
>>>>> = 131599, SupervisionTime = 60
>>>>> Feb 22 15:53:09 SC-2 osafrded[16327]: NO Peer down on node 0x2010f
>>>>> Feb 22 15:53:09 SC-2 osafamfd[16408]: NO Node 'SC-1' left the cluster
>>>>> Feb 22 15:53:09 SC-2 osaflogd[16372]: NO Failed (2) to send of WRITE
>>>>> ack to: 2010f00003d6a
>>>>> Feb 22 15:53:09 SC-2 osaflogd[16372]: NO Failed (2) to send of WRITE
>>>>> ack to: 2010f00003d6a
>>>>> Feb 22 15:53:09 SC-2 osaflogd[16372]: NO Failed (2) to send of WRITE
>>>>> ack to: 2010f00003d74
>>>>> Feb 22 15:53:09 SC-2 osafclmd[16398]: NO Node 131343 went down.
>> Not
>>>>> sending track callback for agents on that node
>>>>> Feb 22 15:53:09 SC-2 osafclmd[16398]: NO Node 131343 went down.
>> Not
>>>>> sending track callback for agents on that node
>>>>> Feb 22 15:53:09 SC-2 osafclmd[16398]: NO Node 131343 went down.
>> Not
>>>>> sending track callback for agents on that node
>>>>> Feb 22 15:53:09 SC-2 osafclmd[16398]: NO Node 131343 went down.
>> Not
>>>>> sending track callback for agents on that node
>>>>> Feb 22 15:53:09 SC-2 osafclmd[16398]: NO Node 131343 went down.
>> Not
>>>>> sending track callback for agents on that node
>>>>> Feb 22 15:53:09 SC-2 osafclmd[16398]: NO Node 131343 went down.
>> Not
>>>>> sending track callback for agents on that node
>>>>> Feb 22 15:53:09 SC-2 osafclmd[16398]: NO Node 131343 went down.
>> Not
>>>>> sending track callback for agents on that node
>>>>> Feb 22 15:53:09 SC-2 osafimmd[16346]: WA IMMD lost contact with
>> peer
>>>>> IMMD (NCSMDS_RED_DOWN)
>>>>> Feb 22 15:53:09 SC-2 osafimmnd[16359]: NO Global discard node
>> received
>>>>> for nodeId:2010f pid:15702
>>>>> Feb 22 15:53:09 SC-2 osafimmnd[16359]: NO Implementer disconnected
>> 16
>>>>> <0, 2010f(down)> (MsgQueueService131343)
>>>>> Feb 22 15:53:09 SC-2 osafimmnd[16359]: NO Implementer disconnected
>> 17
>>>>> <0, 2010f(down)> (@safAmfService2010f)
>>>>> Feb 22 15:53:09 SC-2 osafimmnd[16359]: NO Implementer disconnected
>> 18
>>>>> <0, 2010f(down)> (@OpenSafImmReplicatorB)
>>>>> Feb 22 15:53:09 SC-2 opensaf_reboot: Rebooting remote node in the
>>>>> absence of PLM is outside the scope of OpenSAF
>>>>>
>>>>>
>> ==============================================================
>>>> ======================================================
>>>>> -AVM
>>>>>
>>>>> On 2/22/2017 3:13 PM, Vu Minh Nguyen wrote:
>>>>>> Hi Mahesh,
>>>>>>
>>>>>> I put all required patches into one. Try to use this and see if you
>>>>>> still
>>>>>> have that problem or not.
>>>>>>
>>>>>> Regards, Vu
>>>>>>
>>>>>>> -----Original Message-----
>>>>>>> From: A V Mahesh [mailto:mahesh.va...@oracle.com]
>>>>>>> Sent: Wednesday, February 22, 2017 3:35 PM
>>>>>>> To: Vu Minh Nguyen <vu.m.ngu...@dektech.com.au>;
>>>>>>> lennart.l...@ericsson.com; canh.v.tru...@dektech.com.au
>>>>>>> Cc: opensaf-devel@lists.sourceforge.net
>>>>>>> Subject: Re: [devel] [PATCH 0 of 3] Review Request for log: add
>>>>>> alternative
>>>>>>> destinations of log records [#2258] V4
>>>>>>>
>>>>>>> Hi Vu,
>>>>>>>
>>>>>>> I used  new #3 and #4 patches ,  Can you please re-send All the
> final
>>>>>>> patch in go,
>>>>>>>
>>>>>>> which i need to apply on today`s  staging ( if possible publish the
>>>>>>> with
>>>>>>> new version )
>>>>>>>
>>>>>>> -AVM
>>>>>>>
>>>>>>>
>>>>>>> On 2/22/2017 1:52 PM, Vu Minh Nguyen wrote:
>>>>>>>> Hi Mahesh,
>>>>>>>>
>>>>>>>>> Core was generated by `/usr/lib64/opensaf/osaflogd'.
>>>>>>>>> Program terminated with signal 11, Segmentation fault.
>>>>>>>>> #0  ckpt_proc_cfg_stream(lgs_cb*, void*) () at
>>>>>>>>> src/log/logd/lgs_mbcsv.cc:2195
>>>>>>>>> 2195    src/log/logd/lgs_mbcsv.cc: No such file or directory.
>>>>>>>>>              in src/log/logd/lgs_mbcsv.cc
>>>>>>>> Backtrace still points to old position (lgs_mbcsv:2195). I guess
> the
>>>>>>>> osaflogd binary has not been updated with the fixed patch.
>>>>>>>>
>>>>>>>>
>>>>>>>> Regards, Vu
>>>>>>>>
>>>>>>>>> -----Original Message-----
>>>>>>>>> From: A V Mahesh [mailto:mahesh.va...@oracle.com]
>>>>>>>>> Sent: Wednesday, February 22, 2017 3:18 PM
>>>>>>>>> To: Vu Minh Nguyen <vu.m.ngu...@dektech.com.au>;
>>>>>>>>> lennart.l...@ericsson.com; canh.v.tru...@dektech.com.au
>>>>>>>>> Cc: opensaf-devel@lists.sourceforge.net
>>>>>>>>> Subject: Re: [devel] [PATCH 0 of 3] Review Request for log: add
>>>>>>>> alternative
>>>>>>>>> destinations of log records [#2258] V4
>>>>>>>>>
>>>>>>>>> Hi Vu,
>>>>>>>>>
>>>>>>>>> SC-2 standby osaflogd core dumped still occurs ( Not resolved) ,
> the
>>>>>> new
>>>>>>>>> patch only resolved the application (/usr/bin/logtest )
> Segmentation
>>>>>>>>> fault on SC-1 Active.
>>>>>>>>>
>>>>>>>>>
>> ==============================================================
>> ==============================================================
>>>>>>>>> # gdb /usr/lib64/opensaf/osaflogd
>> core_1487751055.osaflogd.4594
>>>> GNU
>>>>>>>>> gdb
>>>>>>>>> (GDB) SUSE (7.3-0.6.1)
>>>>>>>>> Copyright (C) 2011 Free Software Foundation, Inc.
>>>>>>>>> .......
>>>>>>>>> Core was generated by `/usr/lib64/opensaf/osaflogd'.
>>>>>>>>> Program terminated with signal 11, Segmentation fault.
>>>>>>>>> #0  ckpt_proc_cfg_stream(lgs_cb*, void*) () at
>>>>>>>>> src/log/logd/lgs_mbcsv.cc:2195
>>>>>>>>> 2195    src/log/logd/lgs_mbcsv.cc: No such file or directory.
>>>>>>>>>              in src/log/logd/lgs_mbcsv.cc
>>>>>>>>> (gdb) bt
>>>>>>>>> #0  ckpt_proc_cfg_stream(lgs_cb*, void*) () at
>>>>>>>>> src/log/logd/lgs_mbcsv.cc:2195
>>>>>>>>> #1  0x00007f97b026f960 in ckpt_decode_log_struct(lgs_cb*,
>>>>>>>>> ncs_mbcsv_cb_arg*, void*, void*, unsigned int (*)(edu_hdl_tag*,
>>>>>>>>> edu_tkn_tag*, void*, unsigned int*, edu_buf_env_tag*,
>>>> EDP_OP_TYPE,
>>>>>>>>> EDU_ERR*)) () at src/log/logd/lgs_mbcsv.cc:950
>>>>>>>>> #2  0x00007f97b02710dc in ckpt_decode_async_update(lgs_cb*,
>>>>>>>>> ncs_mbcsv_cb_arg*) () at src/log/logd/lgs_mbcsv.cc:1086
>>>>>>>>> #3  0x00007f97b0273941 in mbcsv_callback(ncs_mbcsv_cb_arg*) ()
>> at
>>>>>>>>> src/log/logd/lgs_mbcsv.cc:880
>>>>>>>>> #4  0x00007f97af372596 in ncs_mbscv_rcv_decode () from
>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0
>>>>>>>>> #5  0x00007f97af372766 in ncs_mbcsv_rcv_async_update () from
>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0
>>>>>>>>> #6  0x00007f97af379370 in mbcsv_process_events () from
>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0
>>>>>>>>> #7  0x00007f97af3794db in mbcsv_hdl_dispatch_all () from
>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0
>>>>>>>>> #8  0x00007f97af373ce2 in mbcsv_process_dispatch_request () at
>>>>>>>>> src/mbc/mbcsv_api.c:423
>>>>>>>>> #9  0x00007f97b027096e in lgs_mbcsv_dispatch(unsigned int) () at
>>>>>>>>> src/log/logd/lgs_mbcsv.cc:327
>>>>>>>>> #10 0x00007f97b024d9f2 in main () at src/log/logd/lgs_main.cc:583
>>>>>>>>> (gdb) bt full
>>>>>>>>> #0  ckpt_proc_cfg_stream(lgs_cb*, void*) () at
>>>>>>>>> src/log/logd/lgs_mbcsv.cc:2195
>>>>>>>>>              ckpt_data_handler = {0x7f97b0270300
>>>>>>>>> <ckpt_proc_initialize_client(lgs_cb*, void*)>, 0x7f97b02701c0
>>>>>>>>> <ckpt_proc_finalize_client(lgs_cb*, void*)>,
>>>>>>>>>                0x7f97b0270060 <ckpt_proc_agent_down(lgs_cb*,
> void*)>,
>>>>>>>>> 0x7f97b02712f0 <ckpt_proc_log_write(lgs_cb*, void*)>,
>>>> 0x7f97b0271ab0
>>>>>>>>> <ckpt_proc_open_stream(lgs_cb*, void*)>,
>>>>>>>>>                0x7f97b026fe80 <ckpt_proc_close_stream(lgs_cb*,
>>> void*)>,
>>>>>>>>> 0x7f97b0272380 <ckpt_proc_cfg_stream(lgs_cb*, void*)>,
>>>>>>> 0x7f97b0274800
>>>>>>>>> <ckpt_proc_lgs_cfg_v2(lgs_cb*, void*)>,
>>>>>>>>>                0x7f97b0274e10 <ckpt_proc_lgs_cfg_v3(lgs_cb*,
> void*)>,
>>>>>>>>> 0x7f97b02754f0 <ckpt_proc_lgs_cfg_v5(lgs_cb*, void*)>}
>>>>>>>>> #1  0x00007f97b026f960 in ckpt_decode_log_struct(lgs_cb*,
>>>>>>>>> ncs_mbcsv_cb_arg*, void*, void*, unsigned int (*)(edu_hdl_tag*,
>>>>>>>>> edu_tkn_tag*, void*, unsigned int*, edu_buf_env_tag*,
>>>> EDP_OP_TYPE,
>>>>>>>>> EDU_ERR*)) () at src/log/logd/lgs_mbcsv.cc:950
>>>>>>>>>              ckpt_data_handler = {0x7f97b0270300
>>>>>>>>> <ckpt_proc_initialize_client(lgs_cb*, void*)>, 0x7f97b02701c0
>>>>>>>>> <ckpt_proc_finalize_client(lgs_cb*, void*)>,
>>>>>>>>>                0x7f97b0270060 <ckpt_proc_agent_down(lgs_cb*,
> void*)>,
>>>>>>>>> 0x7f97b02712f0 <ckpt_proc_log_write(lgs_cb*, void*)>,
>>>> 0x7f97b0271ab0
>>>>>>>>> <ckpt_proc_open_stream(lgs_cb*, void*)>,
>>>>>>>>>                0x7f97b026fe80 <ckpt_proc_close_stream(lgs_cb*,
>>> void*)>,
>>>>>>>>> 0x7f97b0272380 <ckpt_proc_cfg_stream(lgs_cb*, void*)>,
>>>>>>> 0x7f97b0274800
>>>>>>>>> <ckpt_proc_lgs_cfg_v2(lgs_cb*, void*)>,
>>>>>>>>>                0x7f97b0274e10 <ckpt_proc_lgs_cfg_v3(lgs_cb*,
> void*)>,
>>>>>>>>> 0x7f97b02754f0 <ckpt_proc_lgs_cfg_v5(lgs_cb*, void*)>}
>>>>>>>>> #2  0x00007f97b02710dc in ckpt_decode_async_update(lgs_cb*,
>>>>>>>>> ncs_mbcsv_cb_arg*) () at src/log/logd/lgs_mbcsv.cc:1086
>>>>>>>>>              ckpt_data_handler = {0x7f97b0270300
>>>>>>>>> <ckpt_proc_initialize_client(lgs_cb*, void*)>, 0x7f97b02701c0
>>>>>>>>> <ckpt_proc_finalize_client(lgs_cb*, void*)>,
>>>>>>>>>                0x7f97b0270060 <ckpt_proc_agent_down(lgs_cb*,
> void*)>,
>>>>>>>>> 0x7f97b02712f0 <ckpt_proc_log_write(lgs_cb*, void*)>,
>>>> 0x7f97b0271ab0
>>>>>>>>> <ckpt_proc_open_stream(lgs_cb*, void*)>,
>>>>>>>>>                0x7f97b026fe80 <ckpt_proc_close_stream(lgs_cb*,
>>> void*)>,
>>>>>>>>> 0x7f97b0272380 <ckpt_proc_cfg_stream(lgs_cb*, void*)>,
>>>>>>> 0x7f97b0274800
>>>>>>>>> <ckpt_proc_lgs_cfg_v2(lgs_cb*, void*)>,
>>>>>>>>>                0x7f97b0274e10 <ckpt_proc_lgs_cfg_v3(lgs_cb*,
> void*)>,
>>>>>>>>> 0x7f97b02754f0 <ckpt_proc_lgs_cfg_v5(lgs_cb*, void*)>}
>>>>>>>>> #3  0x00007f97b0273941 in mbcsv_callback(ncs_mbcsv_cb_arg*) ()
>> at
>>>>>>>>> src/log/logd/lgs_mbcsv.cc:880
>>>>>>>>>              ckpt_data_handler = {0x7f97b0270300
>>>>>>>>> <ckpt_proc_initialize_client(lgs_cb*, void*)>, 0x7f97b02701c0
>>>>>>>>> <ckpt_proc_finalize_client(lgs_cb*, void*)>,
>>>>>>>>>                0x7f97b0270060 <ckpt_proc_agent_down(lgs_cb*,
> void*)>,
>>>>>>>>> 0x7f97b02712f0 <ckpt_proc_log_write(lgs_cb*, void*)>,
>>>> 0x7f97b0271ab0
>>>>>>>>> <ckpt_proc_open_stream(lgs_cb*, void*)>,
>>>>>>>>>                0x7f97b026fe80 <ckpt_proc_close_stream(lgs_cb*,
>>> void*)>,
>>>>>>>>> 0x7f97b0272380 <ckpt_proc_cfg_stream(lgs_cb*, void*)>,
>>>>>>> 0x7f97b0274800
>>>>>>>>> <ckpt_proc_lgs_cfg_v2(lgs_cb*, void*)>,
>>>>>>>>>                0x7f97b0274e10 <ckpt_proc_lgs_cfg_v3(lgs_cb*,
> void*)>,
>>>>>>>>> 0x7f97b02754f0 <ckpt_proc_lgs_cfg_v5(lgs_cb*, void*)>}
>>>>>>>>> #4  0x00007f97af372596 in ncs_mbscv_rcv_decode () from
>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0
>>>>>>>>>              mbcsv_init_process_req_func = {0x7f97af373630
>>>>>>>>> <mbcsv_process_initialize_request>, 0x7f97af374f10
>>>>>>>>> <mbcsv_process_get_sel_obj_request>,
>>>>>>>>>                0x7f97af373be0 <mbcsv_process_dispatch_request>,
>>>>>>>>> 0x7f97af373af0 <mbcsv_process_finalize_request>,
>> 0x7f97af373cf0
>>>>>>>>> <mbcsv_process_open_request>,
>>>>>>>>>                0x7f97af374050 <mbcsv_process_close_request>,
>>>>>> 0x7f97af3741e0
>>>>>>>>> <mbcsv_process_chg_role_request>, 0x7f97af3744c0
>>>>>>>>> <mbcsv_process_snd_ckpt_request>,
>>>>>>>>>                0x7f97af3747d0 <mbcsv_process_snd_ntfy_request>,
>>>>>>>>> 0x7f97af374970 <mbcsv_process_snd_data_req>, 0x7f97af373930
>>>>>>>>> <mbcsv_process_get_request>,
>>>>>>>>>                0x7f97af374bd0 <mbcsv_process_set_request>}
>>>>>>>>> #5  0x00007f97af372766 in ncs_mbcsv_rcv_async_update () from
>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0
>>>>>>>>>              mbcsv_init_process_req_func = {0x7f97af373630
>>>>>>>>> <mbcsv_process_initialize_request>, 0x7f97af374f10
>>>>>>>>> <mbcsv_process_get_sel_obj_request>,
>>>>>>>>>                0x7f97af373be0 <mbcsv_process_dispatch_request>,
>>>>>>>>> 0x7f97af373af0 <mbcsv_process_finalize_request>,
>> 0x7f97af373cf0
>>>>>>>>> <mbcsv_process_open_request>,
>>>>>>>>>                0x7f97af374050 <mbcsv_process_close_request>,
>>>>>> 0x7f97af3741e0
>>>>>>>>> <mbcsv_process_chg_role_request>, 0x7f97af3744c0
>>>>>>>>> <mbcsv_process_snd_ckpt_request>,
>>>>>>>>>                0x7f97af3747d0 <mbcsv_process_snd_ntfy_request>,
>>>>>>>>> 0x7f97af374970 <mbcsv_process_snd_data_req>, 0x7f97af373930
>>>>>>>>> <mbcsv_process_get_request>,
>>>>>>>>>                0x7f97af374bd0 <mbcsv_process_set_request>}
>>>>>>>>> #6  0x00007f97af379370 in mbcsv_process_events () from
>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0
>>>>>>>>> No symbol table info available.
>>>>>>>>> #7  0x00007f97af3794db in mbcsv_hdl_dispatch_all () from
>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0
>>>>>>>>> No symbol table info available.
>>>>>>>>> #8  0x00007f97af373ce2 in mbcsv_process_dispatch_request () at
>>>>>>>>> src/mbc/mbcsv_api.c:423
>>>>>>>>>              mbcsv_init_process_req_func = {0x7f97af373630
>>>>>>>>> <mbcsv_process_initialize_request>, 0x7f97af374f10
>>>>>>>>> <mbcsv_process_get_sel_obj_request>,
>>>>>>>>>                0x7f97af373be0 <mbcsv_process_dispatch_request>,
>>>>>>>>> 0x7f97af373af0 <mbcsv_process_finalize_request>,
>> 0x7f97af373cf0
>>>>>>>>> <mbcsv_process_open_request>,
>>>>>>>>>                0x7f97af374050 <mbcsv_process_close_request>,
>>>>>> 0x7f97af3741e0
>>>>>>>>> <mbcsv_process_chg_role_request>, 0x7f97af3744c0
>>>>>>>>> <mbcsv_process_snd_ckpt_request>,
>>>>>>>>>                0x7f97af3747d0 <mbcsv_process_snd_ntfy_request>,
>>>>>>>>> 0x7f97af374970 <mbcsv_process_snd_data_req>, 0x7f97af373930
>>>>>>>>> <mbcsv_process_get_request>,
>>>>>>>>>                0x7f97af374bd0 <mbcsv_process_set_request>}
>>>>>>>>> #9  0x00007f97b027096e in lgs_mbcsv_dispatch(unsigned int) () at
>>>>>>>>> src/log/logd/lgs_mbcsv.cc:327
>>>>>>>>>              ckpt_data_handler = {0x7f97b0270300
>>>>>>>>> <ckpt_proc_initialize_client(lgs_cb*, void*)>, 0x7f97b02701c0
>>>>>>>>> <ckpt_proc_finalize_client(lgs_cb*, void*)>,
>>>>>>>>>                0x7f97b0270060 <ckpt_proc_agent_down(lgs_cb*,
> void*)>,
>>>>>>>>> 0x7f97b02712f0 <ckpt_proc_log_write(lgs_cb*, void*)>,
>>>> 0x7f97b0271ab0
>>>>>>>>> <ckpt_proc_open_stream(lgs_cb*, void*)>,
>>>>>>>>>                0x7f97b026fe80 <ckpt_proc_close_stream(lgs_cb*,
>>> void*)>,
>>>>>>>>> 0x7f97b0272380 <ckpt_proc_cfg_stream(lgs_cb*, void*)>,
>>>>>>> 0x7f97b0274800
>>>>>>>>> <ckpt_proc_lgs_cfg_v2(lgs_cb*, void*)>,
>>>>>>>>>                0x7f97b0274e10 <ckpt_proc_lgs_cfg_v3(lgs_cb*,
> void*)>,
>>>>>>>>> 0x7f97b02754f0 <ckpt_proc_lgs_cfg_v5(lgs_cb*, void*)>}
>>>>>>>>> ---Type <return> to continue, or q <return> to quit---
>>>>>>>>> #10 0x00007f97b024d9f2 in main () at src/log/logd/lgs_main.cc:583
>>>>>>>>>              usr1_sel_obj = {raise_obj = -1, rmv_obj = -1}
>>>>>>>>>              _lgs_cb = {mds_hdl = 65547, mds_role =
>> V_DEST_RL_STANDBY,
>>>>>>> vaddr
>>>>>>>>> = 11, log_version = {releaseCode = 65 'A', majorVersion = 2
> '\002',
>>>>>>>>> minorVersion = 2 '\002'}, client_tree = {
>>>>>>>>>                  root_node = {bit = -1, left = 0x7f97b04cf1b0,
> right =
>>>>>>>>> 0x7f97b04a2418, key_info = 0x7f97b04b7bd0 ""}, params =
>> {key_size
>>>>>>>>> = 4},
>>>>>>>>> n_nodes = 8}, comp_name = {_opaque = {46, 24947,
>>>>>>>>>                    17254, 28015, 15728, 20300, 11335, 24947, 21350,
>>>>>>>>> 15733,
>>>>>>>>> 17235, 12845, 29484, 26209, 26451, 12861, 11342, 24947, 16742,
>>>> 28784,
>>>>>>>>> 20285, 25968, 21358, 17985,
>>>>>>>>>                    0 <repeats 105 times>}}, amf_hdl = 4288675841,
>>>>>>>>> amfSelectionObject = 15, amf_invocation_id = 0, is_quiesced_set =
>>>>>> false,
>>>>>>>>> immOiHandle = 554050912783, immSelectionObject = 21,
>>>>>>>>>                clmSelectionObject = 17, clm_hdl = 4279238657,
>>>>>>>>> ha_state =
>>>>>>>>> SA_AMF_HA_STANDBY, last_client_id = 208, async_upd_cnt = 743,
>>>>>>>>> ckpt_state
>>>>>>>>> = COLD_SYNC_IDLE, mbcsv_hdl = 4293918753,
>>>>>>>>>                mbcsv_sel_obj = 23, mbcsv_ckpt_hdl = 4292870177,
>>>>>>>>> mbcsv_peer_version = 7, edu_hdl = {is_inited = true, tree =
>>>>>>>>> {root_node
>>>>>> =
>>>>>>>>> {bit = -1, left = 0x7f97b04cf2e0,
>>>>>>>>>                      right = 0x7f97b04a25b8, key_info =
> 0x7f97b04b7d40
>>>>>> ""},
>>>>>>>>> params = {key_size = 8}, n_nodes = 12}, to_version = 1},
>>>>>>>>> fully_initialized = true, lga_down_list_head = 0x0,
>>>>>>>>>                lga_down_list_tail = 0x0, clm_init_sel_obj =
>>>>>>>>> {raise_obj =
>>>>>> -1,
>>>>>>>>> rmv_obj = -1}, nid_started = true, scAbsenceAllowed = 900,
>>>>>>>>> lgs_recovery_state = LGS_NORMAL}
>>>>>>>>>              nfds = 7
>>>>>>>>>              fds = {{fd = 19, events = 1, revents = 0}, {fd = 15,
>>>>>>>>> events =
>>>>>>>>> 1, revents = 0}, {fd = 23, events = 1, revents = 1}, {fd = 13,
>>>>>>>>> events =
>>>>>>>>> 1, revents = 0}, {fd = -1, events = 1,
>>>>>>>>>                  revents = 0}, {fd = 17, events = 1, revents = 0},
>>>>>>>>> {fd =
>>>>>> 21,
>>>>>>>>> events = 1, revents = 0}}
>>>>>>>>>              mbox_msgs = {0, 0, 0, 0, 0}
>>>>>>>>>              lgs_cb = 0x7f97b04a2400
>>>>>>>>>              mbox_low = {0, 0, 0, 0, 0}
>>>>>>>>>              lgs_mbox_init_mutex = {__data = {__lock = 0, __count =
> 0,
>>>>>>>>> __owner = 0, __nusers = 0, __kind = 0, __spins = 0, __list =
>>>>>>>>> {__prev =
>>>>>>>>> 0x0, __next = 0x0}},
>>>>>>>>>                __size = '\000' <repeats 39 times>, __align = 0}
>>>>>>>>>              lgs_mbx = 4291821569
>>>>>>>>>              mbox_high = {0, 0, 0, 0, 0}
>>>>>>>>>              mbox_full = {false, false, false, false, false}
>>>>>>>>> (gdb)
>>>>>>>>> (gdb)
>>>>>>>>>
>>>>>>>>> Feb 22 13:40:55 SC-2 osafimmnd[4584]: NO Ccb 131 COMMITTED
>>>>>>>>> (immcfg_SC-1_18714)
>>>>>>>>> Feb 22 13:40:56 SC-2 osafamfnd[4634]: NO
>>>>>>>>> 'safComp=LOG,safSu=SC-2,safSg=2N,safApp=OpenSAF' faulted due
>> to
>>>>>>>>> 'avaDown' : Recovery is 'nodeFailfast'
>>>>>>>>> Feb 22 13:40:56 SC-2 osafamfnd[4634]: ER
>>>>>>>>> safComp=LOG,safSu=SC-2,safSg=2N,safApp=OpenSAF Faulted due
>>>>>>>>> to:avaDown
>>>>>>>>> Recovery is:nodeFailfast
>>>>>>>>> Feb 22 13:40:56 SC-2 osafamfnd[4634]: Rebooting OpenSAF NodeId
>> =
>>>>>>>>> 131599
>>>>>>>>> EE Name = , Reason: Component faulted: recovery is node failfast,
>>>>>>>>> OwnNodeId = 131599, SupervisionTime = 60
>>>>>>>>> Feb 22 13:40:56 SC-2 opensaf_reboot: Rebooting local node;
>>>> timeout=60
>> ==============================================================
>> ==============================================================
>>>>>>>>> On 2/22/2017 12:23 PM, A V Mahesh wrote:
>>>>>>>>>> Hi Vu,
>>>>>>>>>>
>>>>>>>>>> On 2/22/2017 12:19 PM, Vu Minh Nguyen wrote:
>>>>>>>>>>> [Vu] I has sent you 02 patches. There is code change in osaflogd
>>>>>>>>>>> code
>>>>>>>> that
>>>>>>>>>>> fix the coredump you have observed.
>>>>>>>>>>> The other one is test code that fix the logtest coredump.
>>>>>>>>>> Ok I will re-test , and update you .
>>>>>>>>>>
>>>>>>>>>> -AVM
>>>>>>>>>>
>>>>>>>>>> On 2/22/2017 12:19 PM, Vu Minh Nguyen wrote:
>>>>>>>>>>> Hi Mahehs,
>>>>>>>>>>>
>>>>>>>>>>> See my reply inline, [Vu].
>>>>>>>>>>>
>>>>>>>>>>> Regards, Vu
>>>>>>>>>>>
>>>>>>>>>>>> -----Original Message-----
>>>>>>>>>>>> From: A V Mahesh [mailto:mahesh.va...@oracle.com]
>>>>>>>>>>>> Sent: Wednesday, February 22, 2017 1:36 PM
>>>>>>>>>>>> To: Vu Minh Nguyen <vu.m.ngu...@dektech.com.au>;
>>>>>>>>>>>> lennart.l...@ericsson.com; canh.v.tru...@dektech.com.au
>>>>>>>>>>>> Cc: opensaf-devel@lists.sourceforge.net
>>>>>>>>>>>> Subject: Re: [PATCH 0 of 3] Review Request for log: add
>>>>>>>>>>>> alternative
>>>>>>>>>>>> destinations of log records [#2258] V4
>>>>>>>>>>>>
>>>>>>>>>>>> Hi Vu,
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> On 2/22/2017 11:52 AM, Vu Minh Nguyen wrote:
>>>>>>>>>>>>> Hi Mahesh,
>>>>>>>>>>>>>
>>>>>>>>>>>>> Have a code fault in uml test, and other one in checkpoint.
>>>>>>>>>>>> [AVM] This is Normal Suse 11 VM ( not  UML).
>>>>>>>>>>>>> I have just updated the code. Please re-apply for #3 and #4
>>>>>> patches.
>>>>>>>>>>>> [AVM] is these new patch has function changes or only test
>> code
>>>>>>>>> changes ?
>>>>>>>>>>> [Vu] I has sent you 02 patches. There is code change in osaflogd
>>>>>>>>>>> code
>>>>>>>> that
>>>>>>>>>>> fix the coredump you have observed.
>>>>>>>>>>> The other one is test code that fix the logtest coredump.
>>>>>>>>>>>>> Note that, test case #14 of suite 17 should be run on active
>>>>>>>>>>>>> node,
>>>>>>>>>>>> otherwise
>>>>>>>>>>>>> getting failed.
>>>>>>>>>>>> [AVM]  Segmentation fault of  /usr/bin/logtest Not a big issue
> ,
>>>>>>>>>>>>         we need to debug why  osaflogd core dumped and it is
>>>>>>>>>>>> critical
>>>>>>>>>>> [Vu] I found the problem. You can try with the new one to see if
>>>>>>>>>>> the
>>>>>>>>>>> coredump is still there or not.
>>>>>>>>>>>>> I will put condition check to that test case later.
>>>>>>>>>>>> -AVM
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>> Regards, Vu
>>>>>>>>>>>>>
>>>>>>>>>>>>>> -----Original Message-----
>>>>>>>>>>>>>> From: A V Mahesh [mailto:mahesh.va...@oracle.com]
>>>>>>>>>>>>>> Sent: Wednesday, February 22, 2017 12:16 PM
>>>>>>>>>>>>>> To: Vu Minh Nguyen <vu.m.ngu...@dektech.com.au>;
>>>>>>>>>>>>>> lennart.l...@ericsson.com; canh.v.tru...@dektech.com.au
>>>>>>>>>>>>>> Cc: opensaf-devel@lists.sourceforge.net
>>>>>>>>>>>>>> Subject: Re: [PATCH 0 of 3] Review Request for log: add
>>>>>> alternative
>>>>>>>>>>>>>> destinations of log records [#2258] V4
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> Hi Vu,
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> Thanks ,
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> While testing  /usr/bin/logtest ,  SC-2 standby osaflogd core
>>>>>>> dumped
>>>>>>>>>>> and
>>>>>>>>>>>>>> /usr/bin/logtest on SC-1 Active
>>>>>>>>>>>>>> got Segmentation fault , am I missing any other patch (  i am
>>>>>> using
>>>>>>>>>>>>>> devel published patch only )
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> Following patches i am using :
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>          1) #2293 (sent by Anders Widel, but not yet pushed)
>>>>>>>>>>>>>>            2) #2258 (v2, sent by Lennart, but not yet pushed
> yet)
>>>>>>>>>>>>>>            3) #2258 (v4, sent by Vu, but not yet pushed yet)
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>> ==============================================================
>>>>>>>>>>>>>> ========================================
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> Core was generated by `/usr/lib64/opensaf/osaflogd'.
>>>>>>>>>>>>>> Program terminated with signal 11, Segmentation fault.
>>>>>>>>>>>>>> #0  ckpt_proc_cfg_stream(lgs_cb*, void*) () at
>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc:2195
>>>>>>>>>>>>>> 2195    src/log/logd/lgs_mbcsv.cc: No such file or directory.
>>>>>>>>>>>>>>                 in src/log/logd/lgs_mbcsv.cc
>>>>>>>>>>>>>> (gdb) bt
>>>>>>>>>>>>>> #0  ckpt_proc_cfg_stream(lgs_cb*, void*) () at
>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc:2195
>>>>>>>>>>>>>> #1  0x00007f12c3e22960 in ckpt_decode_log_struct(lgs_cb*,
>>>>>>>>>>>>>> ncs_mbcsv_cb_arg*, void*, void*, unsigned int
>>>> (*)(edu_hdl_tag*,
>>>>>>>>>>>>>> edu_tkn_tag*, void*, unsigned int*, edu_buf_env_tag*,
>>>>>>>>> EDP_OP_TYPE,
>>>>>>>>>>>>>> EDU_ERR*)) () at src/log/logd/lgs_mbcsv.cc:950
>>>>>>>>>>>>>> #2  0x00007f12c3e240dc in
>>>> ckpt_decode_async_update(lgs_cb*,
>>>>>>>>>>>>>> ncs_mbcsv_cb_arg*) () at src/log/logd/lgs_mbcsv.cc:1086
>>>>>>>>>>>>>> #3  0x00007f12c3e26941 in
>>>> mbcsv_callback(ncs_mbcsv_cb_arg*) ()
>>>>>>> at
>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc:880
>>>>>>>>>>>>>> #4  0x00007f12c2f25596 in ncs_mbscv_rcv_decode () from
>>>>>>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0
>>>>>>>>>>>>>> #5  0x00007f12c2f25766 in ncs_mbcsv_rcv_async_update ()
>>>> from
>>>>>>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0
>>>>>>>>>>>>>> #6  0x00007f12c2f2c370 in mbcsv_process_events () from
>>>>>>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0
>>>>>>>>>>>>>> #7  0x00007f12c2f2c4db in mbcsv_hdl_dispatch_all () from
>>>>>>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0
>>>>>>>>>>>>>> #8  0x00007f12c2f26ce2 in mbcsv_process_dispatch_request
>> ()
>>>> at
>>>>>>>>>>>>>> src/mbc/mbcsv_api.c:423
>>>>>>>>>>>>>> #9  0x00007f12c3e2396e in lgs_mbcsv_dispatch(unsigned int)
>> ()
>>>> at
>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc:327
>>>>>>>>>>>>>> #10 0x00007f12c3e009f2 in main () at
>>>>>>>>>>>>>> src/log/logd/lgs_main.cc:583
>>>>>>>>>>>>>> (gdb)
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>> ==============================================================
>>>>>>>>>>>>>> ========================================
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> Feb 22 10:37:06 SC-1 osafimmnd[4020]: NO Invalid error
>>>> reported
>>>>>>>>>>>>>> implementer 'safLogService', Ccb 161 will be aborted
>>>>>>>>>>>>>> Feb 22 10:37:06 SC-1 osafimmnd[4020]: NO Ccb 161 aborted
>> in
>>>>>>>>>>>> COMPLETED
>>>>>>>>>>>>>> processing (validation)
>>>>>>>>>>>>>> Feb 22 10:37:06 SC-1 osafimmnd[4020]: NO Ccb 161
>> ABORTED
>>>>>>>>>>>> (immcfg_SC-
>>>>>>>>>>>>>> 1_5394)
>>>>>>>>>>>>>> Add values Fail
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> Program received signal SIGSEGV, Segmentation fault.
>>>>>>>>>>>>>> 0x000055555556929a in read_and_compare.isra.7 () at
>>>>>>>>>>>>>> src/log/apitest/tet_LogOiOps.c:1891
>>>>>>>>>>>>>> 1891    src/log/apitest/tet_LogOiOps.c: No such file or
>>>>>>>>>>>>>> directory.
>>>>>>>>>>>>>>                 in src/log/apitest/tet_LogOiOps.c
>>>>>>>>>>>>>> (gdb) Feb 22 10:37:07 SC-1 sshd[5298]: Accepted keyboard-
>>>>>>>>>>>> interactive/pam
>>>>>>>>>>>>>> for root from 10.176.178.22 port 51945 ssh2
>>>>>>>>>>>>>> bt
>>>>>>>>>>>>>> #0  0x000055555556929a in read_and_compare.isra.7 () at
>>>>>>>>>>>>>> src/log/apitest/tet_LogOiOps.c:1891
>>>>>>>>>>>>>> #1  0x0000555555569bbb in
>>>>>>>>>>>>>> check_logRecordDestinationConfigurationEmpty
>>>>>>>>>>>>>> () at src/log/apitest/tet_LogOiOps.c:2179
>>>>>>>>>>>>>> #2  0x0000555555573495 in run_test_case ()
>>>>>>>>>>>>>> #3  0x0000555555573934 in test_run ()
>>>>>>>>>>>>>> #4  0x000055555555c7cd in main () at
>>>>>>>>>>>>>> src/log/apitest/logtest.c:569
>>>>>>>>>>>>>> (gdb)
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>> ==============================================================
>>>>>>>>>>>>>> ========================================
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> -AVM
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> On 2/22/2017 9:48 AM, Vu Minh Nguyen wrote:
>>>>>>>>>>>>>>> Hi Mahesh,
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> I send them in attachment instead, and name them in the
>>>> order.
>>>>>>>>>>>>>>> I just pull the latest code, and apply them without getting
>>> any
>>>>>>>> hunk
>>>>>>>>>>>>> error.
>>>>>>>>>>>>>>> Please try with them, and let me know if you see any
>> problem.
>>>>>>>>>>>>>>> Regards, Vu
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> -----Original Message-----
>>>>>>>>>>>>>>>> From: A V Mahesh [mailto:mahesh.va...@oracle.com]
>>>>>>>>>>>>>>>> Sent: Wednesday, February 22, 2017 11:09 AM
>>>>>>>>>>>>>>>> To: Vu Minh Nguyen <vu.m.ngu...@dektech.com.au>;
>>>>>>>>>>>>>>>> lennart.l...@ericsson.com;
>> canh.v.tru...@dektech.com.au
>>>>>>>>>>>>>>>> Cc: opensaf-devel@lists.sourceforge.net
>>>>>>>>>>>>>>>> Subject: Re: [PATCH 0 of 3] Review Request for log: add
>>>>>>>> alternative
>>>>>>>>>>>>>>>> destinations of log records [#2258] V4
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> Hi Vu,
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> I did follow that still i get Hunk #2 FAILED even on
> today's
>>>>>>>> staging
>> ==============================================================
>>>>>>>>>>>>>>>> ==================
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> [root@dhcp-hyd-scp-5fl-10-176-177-96 staging]# patch -
>> p1
>>>>>>> <2293
>>>>>>>>>>>>>>>> patching file src/base/Makefile.am
>>>>>>>>>>>>>>>> Hunk #1 succeeded at 33 (offset 1 line).
>>>>>>>>>>>>>>>> Hunk #3 succeeded at 183 (offset 1 line).
>>>>>>>>>>>>>>>> patching file src/base/file_descriptor.cc
>>>>>>>>>>>>>>>> patching file src/base/file_descriptor.h
>>>>>>>>>>>>>>>> patching file src/base/tests/unix_socket_test.cc
>>>>>>>>>>>>>>>> patching file src/base/unix_client_socket.cc
>>>>>>>>>>>>>>>> patching file src/base/unix_server_socket.cc
>>>>>>>>>>>>>>>> patching file src/base/unix_socket.cc
>>>>>>>>>>>>>>>> patching file src/base/unix_socket.h
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> [root@dhcp-hyd-scp-5fl-10-176-177-96 staging]# patch -
>> p1
>>>>>>> <2258-
>>>>>>>>> 1
>>>>>>>>>>>>>>>> patching file src/log/Makefile.am
>>>>>>>>>>>>>>>> Hunk #1 succeeded at 71 (offset -1 lines).
>>>>>>>>>>>>>>>> patching file src/log/config/logsv_classes.xml
>>>>>>>>>>>>>>>> Hunk #1 FAILED at 147.
>>>>>>>>>>>>>>>> 1 out of 1 hunk FAILED -- saving rejects to file
>>>>>>>>>>>>>>>> src/log/config/logsv_classes.xml.rej
>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_config.cc
>>>>>>>>>>>>>>>> Hunk #1 succeeded at 35 (offset -5 lines).
>>>>>>>>>>>>>>>> Hunk #2 FAILED at 705.
>>>>>>>>>>>>>>>> Hunk #3 FAILED at 971.
>>>>>>>>>>>>>>>> 2 out of 3 hunks FAILED -- saving rejects to file
>>>>>>>>>>>>>>>> src/log/logd/lgs_config.cc.rej
>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_config.h
>>>>>>>>>>>>>>>> Hunk #1 FAILED at 304.
>>>>>>>>>>>>>>>> 1 out of 1 hunk FAILED -- saving rejects to file
>>>>>>>>>>>>>>>> src/log/logd/lgs_config.h.rej
>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_dest.cc
>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_dest.h
>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_evt.cc
>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_imm.cc
>>>>>>>>>>>>>>>> Hunk #1 FAILED at 45.
>>>>>>>>>>>>>>>> Hunk #2 succeeded at 235 (offset -1 lines).
>>>>>>>>>>>>>>>> Hunk #3 FAILED at 877.
>>>>>>>>>>>>>>>> Hunk #4 succeeded at 1273 (offset -20 lines).
>>>>>>>>>>>>>>>> Hunk #5 succeeded at 1404 (offset -1 lines).
>>>>>>>>>>>>>>>> Hunk #6 succeeded at 1449 (offset -20 lines).
>>>>>>>>>>>>>>>> Hunk #7 succeeded at 2032 (offset -1 lines).
>>>>>>>>>>>>>>>> Hunk #8 FAILED at 2181.
>>>>>>>>>>>>>>>> Hunk #9 succeeded at 2271 (offset -54 lines).
>>>>>>>>>>>>>>>> Hunk #10 succeeded at 2387 (offset -1 lines).
>>>>>>>>>>>>>>>> Hunk #11 succeeded at 2377 (offset -54 lines).
>>>>>>>>>>>>>>>> Hunk #12 succeeded at 2478 (offset -1 lines).
>>>>>>>>>>>>>>>> Hunk #13 succeeded at 2684 (offset -54 lines).
>>>>>>>>>>>>>>>> Hunk #14 succeeded at 2821 (offset -1 lines).
>>>>>>>>>>>>>>>> 3 out of 14 hunks FAILED -- saving rejects to file
>>>>>>>>>>>>>>>> src/log/logd/lgs_imm.cc.rej
>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_main.cc
>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_mbcsv.cc
>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_mbcsv.h
>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_mbcsv_v5.cc
>>>>>>>>>>>>>>>> Hunk #3 succeeded at 133 (offset -1 lines).
>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_mbcsv_v7.cc
>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_mbcsv_v7.h
>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_stream.cc
>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_stream.h
>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_util.cc
>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_util.h
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> [root@dhcp-hyd-scp-5fl-10-176-177-96 staging]# patch -
>> p1
>>>>>>> <2258-
>>>>>>>>> 2
>>>>>>>>>>>>>>>> patching file src/log/Makefile.am
>>>>>>>>>>>>>>>> Hunk #1 succeeded at 180 (offset -3 lines).
>>>>>>>>>>>>>>>> patching file src/log/apitest/tet_LogOiOps.c
>>>>>>>>>>>>>>>> Hunk #1 FAILED at 1923.
>>>>>>>>>>>>>>>> Hunk #2 FAILED at 1979.
>>>>>>>>>>>>>>>> Hunk #3 FAILED at 2067.
>>>>>>>>>>>>>>>> Hunk #4 FAILED at 2094.
>>>>>>>>>>>>>>>> 4 out of 4 hunks FAILED -- saving rejects to file
>>>>>>>>>>>>>>>> src/log/apitest/tet_LogOiOps.c.rej
>>>>>>>>>>>>>>>> patching file src/log/apitest/tet_cfg_destination.c
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> [root@dhcp-hyd-scp-5fl-10-176-177-96 staging]# patch -
>> p1
>>>>>>> <2258-
>>>>>>>>> 3
>>>>>>>>>>>>>>>> patching file src/log/Makefile
>>>>>>>>>>>>>>>> patching file src/log/Makefile.am
>>>>>>>>>>>>>>>> Hunk #1 succeeded at 80 (offset -1 lines).
>>>>>>>>>>>>>>>> Hunk #2 succeeded at 217 (offset -2 lines).
>>>>>>>>>>>>>>>> patching file src/log/tests/Makefile
>>>>>>>>>>>>>>>> patching file src/log/tests/lgs_dest_test.cc
>>>>>>>>>>>>>>>> [root@dhcp-hyd-scp-5fl-10-176-177-96 staging]# vi
>>>>>>>>>>>>>>>> src/log/apitest/tet_LogOiOps.c.rej
>>>>>>>>>>>>>>>> [root@dhcp-hyd-scp-5fl-10-176-177-96 staging]#
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>
>> ==============================================================
>>>>>>>>>>>>>>>> ========================
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> -AVM
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> On 2/21/2017 3:53 PM, Vu Minh Nguyen wrote:
>>>>>>>>>>>>>>>>> Hi Mahesh,
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> As I has mentioned in below:
>>>>>>>>>>>>>>>>>>>            To run the test, this patch has dependent on
>>>>>> following
>>>>>>>>>>> patches:
>>>>>>>>>>>>>>>>>>>            1) #2293 (sent by Anders Widel, but not yet
>>>>>>>>>>>>>>>>>>> pushed)
>>>>>>>>>>>>>>>>>>>            2) #2258 (v2, sent by Lennart, but not yet
> pushed
>>>>>> yet)
>>>>>>>>>>>>>>>>> So, you need to apply #2293 first, then #2258 which sent
>> by
>>>>>>>>> Lennart
>>>>>>>>>>>>>>>>> yesterday, then mine.
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> Regards, Vu
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> -----Original Message-----
>>>>>>>>>>>>>>>>>> From: A V Mahesh [mailto:mahesh.va...@oracle.com]
>>>>>>>>>>>>>>>>>> Sent: Tuesday, February 21, 2017 5:10 PM
>>>>>>>>>>>>>>>>>> To: Vu Minh Nguyen <vu.m.ngu...@dektech.com.au>;
>>>>>>>>>>>>>>>>>> lennart.l...@ericsson.com;
>>>> canh.v.tru...@dektech.com.au
>>>>>>>>>>>>>>>>>> Cc: opensaf-devel@lists.sourceforge.net
>>>>>>>>>>>>>>>>>> Subject: Re: [PATCH 0 of 3] Review Request for log: add
>>>>>>>>> alternative
>>>>>>>>>>>>>>>>>> destinations of log records [#2258] V4
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> Hi Vu,
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> Is this applies on top of   log #2146  - V4  , I see
> both
>>>>>>>> #tickets
>>>>>>>>>>>>>>> has
>>>>>>>>>>>>>>>>>> version changes ?
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> in which order i need to apply  ( #2146 & #2258 )    or
>>>>>>>>>>>>>>>>>> (#2258
>>>>>> &
>>>>>>>>>>>>>>> #2146).
>>>> =========================================================
>>>>>>>>>>>>>>>>>> patching file src/log/Makefile.am
>>>>>>>>>>>>>>>>>> Hunk #1 FAILED at 72.
>>>>>>>>>>>>>>>>>> Hunk #2 FAILED at 120.
>>>>>>>>>>>>>>>>>> 2 out of 2 hunks FAILED -- saving rejects to file
>>>>>>>>>>>>>>> src/log/Makefile.am.rej
>>>>>>>>>>>>>>>>>> patching file src/log/config/logsv_classes.xml
>>>>>>>>>>>>>>>>>> Hunk #1 FAILED at 147.
>>>>>>>>>>>>>>>>>> 1 out of 1 hunk FAILED -- saving rejects to file
>>>>>>>>>>>>>>>>>> src/log/config/logsv_classes.xml.rej
>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_config.cc
>>>>>>>>>>>>>>>>>> Hunk #1 succeeded at 35 (offset -5 lines).
>>>>>>>>>>>>>>>>>> Hunk #2 FAILED at 705.
>>>>>>>>>>>>>>>>>> Hunk #3 FAILED at 971.
>>>>>>>>>>>>>>>>>> 2 out of 3 hunks FAILED -- saving rejects to file
>>>>>>>>>>>>>>>>>> src/log/logd/lgs_config.cc.rej
>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_config.h
>>>>>>>>>>>>>>>>>> Hunk #1 FAILED at 304.
>>>>>>>>>>>>>>>>>> 1 out of 1 hunk FAILED -- saving rejects to file
>>>>>>>>>>>>>>>>>> src/log/logd/lgs_config.h.rej
>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_dest.cc
>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_dest.h
>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_evt.cc
>>>>>>>>>>>>>>>>>> Hunk #1 FAILED at 1.
>>>>>>>>>>>>>>>>>> Hunk #2 succeeded at 30 with fuzz 2 (offset 2 lines).
>>>>>>>>>>>>>>>>>> Hunk #3 succeeded at 1282 (offset 45 lines).
>>>>>>>>>>>>>>>>>> Hunk #4 succeeded at 1300 (offset 2 lines).
>>>>>>>>>>>>>>>>>> 1 out of 4 hunks FAILED -- saving rejects to file
>>>>>>>>>>>>>>>>>> src/log/logd/lgs_evt.cc.rej
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>
>> ==============================================================
>>>>>>>>>>>>>>>>>> ===
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> -AVM
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> On 2/21/2017 3:03 PM, Vu Minh Nguyen wrote:
>>>>>>>>>>>>>>>>>>> Summary: log: add alternative destinations of log
>> records
>>>>>>>>> [#2258]
>>>>>>>>>>>>>>>>>>> Review request for Trac Ticket(s): #2258
>>>>>>>>>>>>>>>>>>> Peer Reviewer(s): Lennart, Canh, Mahesh
>>>>>>>>>>>>>>>>>>> Pull request to: <<LIST THE PERSON WITH PUSH
>> ACCESS
>>>>>>> HERE>>
>>>>>>>>>>>>>>>>>>> Affected branch(es): Default
>>>>>>>>>>>>>>>>>>> Development branch: Default
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> --------------------------------
>>>>>>>>>>>>>>>>>>> Impacted area       Impact y/n
>>>>>>>>>>>>>>>>>>> --------------------------------
>>>>>>>>>>>>>>>>>>>            Docs                    n
>>>>>>>>>>>>>>>>>>>            Build system            n
>>>>>>>>>>>>>>>>>>>            RPM/packaging           n
>>>>>>>>>>>>>>>>>>>            Configuration files     n
>>>>>>>>>>>>>>>>>>>            Startup scripts         n
>>>>>>>>>>>>>>>>>>>            SAF services            n
>>>>>>>>>>>>>>>>>>>            OpenSAF services        y
>>>>>>>>>>>>>>>>>>>            Core libraries          n
>>>>>>>>>>>>>>>>>>>            Samples                 n
>>>>>>>>>>>>>>>>>>>            Tests                   y
>>>>>>>>>>>>>>>>>>>            Other                   n
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> Comments (indicate scope for each "y" above):
>>>>>>>>>>>>>>>>>>> ---------------------------------------------
>>>>>>>>>>>>>>>>>>>            To run the test, this patch has dependent on
>>>>>> following
>>>>>>>>>>> patches:
>>>>>>>>>>>>>>>>>>>            1) #2293 (sent by Anders Widel, but not yet
>>>>>>>>>>>>>>>>>>> pushed)
>>>>>>>>>>>>>>>>>>>            2) #2258 (v2, sent by Lennart, but not yet
> pushed
>>>>>> yet)
>>>>>>>>>>>>>>>>>>> changeset
>>>> d74aaf3025c99cade3165a15831124548f4d85bd
>>>>>>>>>>>>>>>>>>> Author:    Vu Minh Nguyen
>>>> <vu.m.ngu...@dektech.com.au>
>>>>>>>>>>>>>>>>>>> Date:    Wed, 15 Feb 2017 14:36:00 +0700
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>       log: add alternative destinations of log records
>>>>>>>>>>>>>>>>>>> [#2258]
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>       Here are major info, detailed info will be added to
>>>>>>>>>>>>>>>>>>> PR doc
>>>>>>>>>>>>> soon. 1)
>>>>>>>>>>>>>>>>>> Add
>>>>>>>>>>>>>>>>>>>       attribute "saLogRecordDestination" to log stream.
> 2)
>>>>>>>>>>>>>>>>>>> Add
>>>>>>>>>>>>> Local
>>>>>>>>>>>>>>>>>> socket
>>>>>>>>>>>>>>>>>>>       destintion handler 3) Integrate into first
> increment
>>>>>>>>>>>>>>>>>>> made by
>>>>>>>>>>>>> Lennart
>>>>>>>>>>>>>>>>>>> changeset
>>>> 4bae27a478c235df3058f43c92d3a5483233b01d
>>>>>>>>>>>>>>>>>>> Author:    Vu Minh Nguyen
>>>> <vu.m.ngu...@dektech.com.au>
>>>>>>>>>>>>>>>>>>> Date:    Wed, 15 Feb 2017 15:07:09 +0700
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>       log: add UML test case to verify alternative
>>>>>>>>>>>>>>>>>>> destination
>>>>>>>>>>>>> [#2258]
>>>>>>>>>>>>>>>>>>>       Major changes: 1) Modify Lennart's test cases
>> because
>>>>>>>>>>>>> enhancing
>>>>>>>>>>>>>>>>>> destination
>>>>>>>>>>>>>>>>>>>       configuration validation rules. 2) Add test suite
>>>>>>>>>>>>>>>>>>> #17 to
>>>>>>>>>>>>> verify
>>>>>>>>>>>>>>>>>> alternative
>>>>>>>>>>>>>>>>>>>       destination
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> changeset
>> bc375725fed22bb4f8cb3ae3df5f96fb9d281efb
>>>>>>>>>>>>>>>>>>> Author:    Vu Minh Nguyen
>>>> <vu.m.ngu...@dektech.com.au>
>>>>>>>>>>>>>>>>>>> Date:    Thu, 16 Feb 2017 17:22:13 +0700
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>       log: add unit tests to verify interfaces provided
> by
>>>>>>>>>>>>> destination
>>>>>>>>>>>>>>>>>> handler
>>>>>>>>>>>>>>>>>>>       [#2258]
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>       Unit tests to verify major interfaces: 1)
>>>>>>>>>>>>>>>>>>> CfgDestination()
>>>>>>>>>>>>> 2)
>>>>>>>>>>>>>>>>>>>       WriteToDestination()
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> Added Files:
>>>>>>>>>>>>>>>>>>> ------------
>>>>>>>>>>>>>>>>>>> src/log/apitest/tet_cfg_destination.c
>>>>>>>>>>>>>>>>>>>            src/log/logd/lgs_dest.cc
>>>>>>>>>>>>>>>>>>>            src/log/logd/lgs_dest.h
>>>>>>>>>>>>>>>>>>>            src/log/logd/lgs_mbcsv_v7.cc
>>>>>>>>>>>>>>>>>>>            src/log/logd/lgs_mbcsv_v7.h
>>>>>>>>>>>>>>>>>>> src/log/tests/lgs_dest_test.cc
>>>>>>>>>>>>>>>>>>>            src/log/tests/Makefile
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> Complete diffstat:
>>>>>>>>>>>>>>>>>>> ------------------
>>>>>>>>>>>>>>>>>>> src/log/Makefile |    4 +
>>>>>>>>>>>>>>>>>>> src/log/Makefile.am |   31 +++++-
>>>>>>>>>>>>>>>>>>> src/log/apitest/tet_LogOiOps.c |    8 +-
>>>>>>>>>>>>>>>>>>> src/log/apitest/tet_cfg_destination.c |  483
>> ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++
>>>>>>>>>>>>>>>>>> ++++++++++++++++++++++++++++++++++
>>>>>>>>>>>>>>>>>>> src/log/config/logsv_classes.xml |    7 +-
>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_config.cc |  169
>>>>>>>>>>>>>>>>>> +++++++++++++++++++++++++++++---
>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_config.h |    3 +-
>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_dest.cc |  707
>> ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++
>> ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++
>>>>>>>>>>>>>>>>>> +++++++++++++++++
>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_dest.h |  576
>> ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++
>>>>>>> ++++++++++++++++++++++++++++++++++++++++++++++++++++
>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_evt.cc |   33 ++++++
>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_imm.cc |  202
>>>>>>>>>>>>>>>>>> +++++++++++++++++++++++++++++++++------
>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_main.cc |    8 +
>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc |  103
>>>>>>>>>>> ++++++++++++++++++-
>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv.h |    6 +-
>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv_v5.cc |   10 +
>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv_v7.cc |  177
>>>>>>>>>>>>>>>>>> +++++++++++++++++++++++++++++++++++
>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv_v7.h |   67
>>>>>>>> +++++++++++++
>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_stream.cc |   60
>>>>>>>> +++++++++++-
>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_stream.h |   16 +++
>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_util.cc |   63
>>>>>>>> ++++++++++++
>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_util.h |   11 +-
>>>>>>>>>>>>>>>>>>> src/log/tests/Makefile |   20 +++
>>>>>>>>>>>>>>>>>>> src/log/tests/lgs_dest_test.cc |  209
>>>>>>>>>>>>>>>>>> +++++++++++++++++++++++++++++++++++++++++
>>>>>>>>>>>>>>>>>>>            23 files changed, 2896 insertions(+), 77
>>>>>>>>>>>>>>>>>>> deletions(-)
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> Testing Commands:
>>>>>>>>>>>>>>>>>>> -----------------
>>>>>>>>>>>>>>>>>>>            Run UML test suite #17
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> Testing, Expected Results:
>>>>>>>>>>>>>>>>>>> --------------------------
>>>>>>>>>>>>>>>>>>>            All test passed
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> Conditions of Submission:
>>>>>>>>>>>>>>>>>>> -------------------------
>>>>>>>>>>>>>>>>>>>            <<HOW MANY DAYS BEFORE PUSHING,
>> CONSENSUS
>>>>>>> ETC>>
>>>>>>>>>>>>>>>>>>> Arch      Built     Started    Linux distro
>>>>>>>>>>>>>>>>>>> -------------------------------------------
>>>>>>>>>>>>>>>>>>> mips        n          n
>>>>>>>>>>>>>>>>>>> mips64      n          n
>>>>>>>>>>>>>>>>>>> x86         n          n
>>>>>>>>>>>>>>>>>>> x86_64      n          n
>>>>>>>>>>>>>>>>>>> powerpc     n          n
>>>>>>>>>>>>>>>>>>> powerpc64   n          n
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> Reviewer Checklist:
>>>>>>>>>>>>>>>>>>> -------------------
>>>>>>>>>>>>>>>>>>> [Submitters: make sure that your review doesn't trigger
>>>> any
>>>>>>>>>>>>>>>> checkmarks!]
>>>>>>>>>>>>>>>>>>> Your checkin has not passed review because (see
>> checked
>>>>>>>>> entries):
>>>>>>>>>>>>>>>>>>> ___ Your RR template is generally incomplete; it has
>> too
>>>>>>> many
>>>>>>>>>>> blank
>>>>>>>>>>>>>>>>> entries
>>>>>>>>>>>>>>>>>>>               that need proper data filled in.
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> ___ You have failed to nominate the proper persons for
>>>>>>> review
>>>>>>>>> and
>>>>>>>>>>>>>>> push.
>>>>>>>>>>>>>>>>>>> ___ Your patches do not have proper short+long
>> header
>>>>>>>>>>>>>>>>>>> ___ You have grammar/spelling in your header that is
>>>>>>>>> unacceptable.
>>>>>>>>>>>>>>>>>>> ___ You have exceeded a sensible line length in your
>>>>>>>>>>>>>>>>>> headers/comments/text.
>>>>>>>>>>>>>>>>>>> ___ You have failed to put in a proper Trac Ticket #
> into
>>>>>> your
>>>>>>>>>>>>>>> commits.
>>>>>>>>>>>>>>>>>>> ___ You have incorrectly put/left internal data in your
>>>>>>>>>>>>> comments/files
>>>>>>>>>>>>>>>>>>> (i.e. internal bug tracking tool IDs, product
>>>>>> names
>>>>>>>> etc)
>>>>>>>>>>>>>>>>>>> ___ You have not given any evidence of testing beyond
>>>> basic
>>>>>>>>> build
>>>>>>>>>>>>>>> tests.
>>>>>>>>>>>>>>>>>>> Demonstrate some level of runtime or other sanity
>>>>>>>> testing.
>>>>>>>>>>>>>>>>>>> ___ You have ^M present in some of your files. These
>>>>>>>>>>>>>>>>>>> have to
>>>>>>>>> be
>>>>>>>>>>>>>>>> removed.
>>>>>>>>>>>>>>>>>>> ___ You have needlessly changed whitespace or added
>>>>>>>>> whitespace
>>>>>>>>>>>>>>>> crimes
>>>>>>>>>>>>>>>>>>>               like trailing spaces, or spaces before
> tabs.
>>>>>>>>>>>>>>>>>>> ___ You have mixed real technical changes with
>>>> whitespace
>>>>>>> and
>>>>>>>>>>>> other
>>>>>>>>>>>>>>>>>>> cosmetic code cleanup changes. These have to be
>>>>>>>>> separate
>>>>>>>>>>>>>>> commits.
>>>>>>>>>>>>>>>>>>> ___ You need to refactor your submission into logical
>>>>>>>>>>>>>>>>>>> chunks;
>>>>>>>>>>> there
>>>>>>>>>>>>> is
>>>>>>>>>>>>>>>>>>>               too much content into a single commit.
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> ___ You have extraneous garbage in your review
>> (merge
>>>>>>>>> commits
>>>>>>>>>>>> etc)
>>>>>>>>>>>>>>>>>>> ___ You have giant attachments which should never
>> have
>>>>>>> been
>>>>>>>>>>>> sent;
>>>>>>>>>>>>>>>>>>> Instead you should place your content in a public
>>>>>>>> tree to
>>>>>>>>>>> be
>>>>>>>>>>>>>>>>> pulled.
>>>>>>>>>>>>>>>>>>> ___ You have too many commits attached to an e-mail;
>>>>>>> resend
>>>>>>>>> as
>>>>>>>>>>>>>>>> threaded
>>>>>>>>>>>>>>>>>>> commits, or place in a public tree for a pull.
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> ___ You have resent this content multiple times
>> without a
>>>>>>> clear
>>>>>>>>>>>>>>>>> indication
>>>>>>>>>>>>>>>>>>>               of what has changed between each re-send.
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> ___ You have failed to adequately and individually
>> address
>>>>>> all
>>>>>>>> of
>>>>>>>>>>>>> the
>>>>>>>>>>>>>>>>>>> comments and change requests that were proposed in
>>>>>>>> the
>>>>>>>>>>>>> initial
>>>>>>>>>>>>>>>>> review.
>>>>>>>>>>>>>>>>>>> ___ You have a misconfigured ~/.hgrc file (i.e.
>> username,
>>>>>>> email
>>>>>>>>>>> etc)
>>>>>>>>>>>>>>>>>>> ___ Your computer have a badly configured date and
>>>> time;
>>>>>>>>>>>> confusing
>>>>>>>>>>>>>> the
>>>>>>>>>>>>>>>>>>>               the threaded patch review.
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> ___ Your changes affect IPC mechanism, and you don't
>>>>>>> present
>>>>>>>>> any
>>>>>>>>>>>>>>>> results
>>>>>>>>>>>>>>>>>>>               for in-service upgradability test.
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> ___ Your changes affect user manual and
>> documentation,
>>>>>>> your
>>>>>>>>>>>> patch
>>>>>>>>>>>>>>>> series
>>>>>>>>>>>>>>>>>>>               do not contain the patch that updates the
>>>>>>>>>>>>>>>>>>> Doxygen
>>>>>>>>> manual.
> ----------------------------------------------------------------------------
>>>>>>>> --
>>>>>>>>>> Check out the vibrant tech community on one of the world's most
>>>>>>>>>> engaging tech sites, SlashDot.org! http://sdm.link/slashdot
>>>>>>>>>> _______________________________________________
>>>>>>>>>> Opensaf-devel mailing list
>>>>>>>>>> Opensaf-devel@lists.sourceforge.net
>>>>>>>>>> https://lists.sourceforge.net/lists/listinfo/opensaf-devel
>


------------------------------------------------------------------------------
Check out the vibrant tech community on one of the world's most
engaging tech sites, SlashDot.org! http://sdm.link/slashdot
_______________________________________________
Opensaf-devel mailing list
Opensaf-devel@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/opensaf-devel

Reply via email to