Hi Vu,

Issue still exist

Feb 23 16:30:04 SC-2 osaflogd[4635]: NO LOG root directory is: 
"/var/log/opensaf/saflog"
Feb 23 16:30:04 SC-2 osaflogd[4635]: NO LOG data group is: ""
Feb 23 16:30:04 SC-2 osaflogd[4635]: NO LGS_MBCSV_VERSION = 7
Feb 23 16:30:04 SC-2 osafamfnd[4675]: NO Assigned 
'safSi=SC-2N,safApp=OpenSAF' STANDBY to 'safSu=SC-2,safSg=2N,safApp=OpenSAF'
Feb 23 16:30:04 SC-2 opensafd: OpenSAF(5.1.M0 - ) services successfully 
started
done
SC-2:/var/crash/opensaf # Feb 23 16:30:04 SC-2 osafimmnd[4625]: NO 
Implementer (applier) connected: 16 (@OpenSafImmReplicatorB) <150, 2020f>
Feb 23 16:30:04 SC-2 osafntfimcnd[4832]: NO Started
Feb 23 16:30:07 SC-2 osafamfd[4665]: NO Cold sync complete!
Feb 23 16:30:07 SC-2 osaflogd[4635]: ER Exiting with message: Client 
attributes differ
Feb 23 16:30:07 SC-2 osafamfnd[4675]: NO 
'safSu=SC-2,safSg=2N,safApp=OpenSAF' component restart probation timer 
started (timeout: 60000000000 ns)
Feb 23 16:30:07 SC-2 osafamfnd[4675]: NO Restarting a component of 
'safSu=SC-2,safSg=2N,safApp=OpenSAF' (comp restart count: 1)
Feb 23 16:30:07 SC-2 osafamfnd[4675]: NO 
'safComp=LOG,safSu=SC-2,safSg=2N,safApp=OpenSAF' faulted due to 
'errorReport' : Recovery is 'componentRes

-AVM


On 2/23/2017 4:09 PM, Vu Minh Nguyen wrote:
> Hi Mahesh,
>
> I found the root cause. It is because in ` log: implement
> SaLogFilterSetCallbackT and version handling [#2146]`,
> Canh introduced MBCSV version #6, but I missed adding that info when
> rebasing.
>
> The attached patch contains the fix. Can you apply it to see if the problem
> still occur? Thanks.
>
> Regards, Vu
>
>> -----Original Message-----
>> From: A V Mahesh [mailto:mahesh.va...@oracle.com]
>> Sent: Thursday, February 23, 2017 5:32 PM
>> To: Vu Minh Nguyen <vu.m.ngu...@dektech.com.au>;
>> lennart.l...@ericsson.com; canh.v.tru...@dektech.com.au
>> Cc: opensaf-devel@lists.sourceforge.net
>> Subject: Re: [devel] [PATCH 0 of 3] Review Request for log: add
> alternative
>> destinations of log records [#2258] V4
>>
>> Hi Vu,
>>
>> Please find attached , below is the time stamp of SC-2
>>
>> ==============================================================
>> ==============================================
>>
>> Feb 23 15:55:30 SC-2 osafimmnd[6978]: NO Implementer (applier)
>> connected: 15 (@safAmfService2020f) <127, 2020f>
>> Feb 23 15:55:30 SC-2 osaflogd[6988]: NO LOGSV_DATA_GROUPNAME not
>> found
>> Feb 23 15:55:30 SC-2 osaflogd[6988]: NO LOG root directory is:
>> "/var/log/opensaf/saflog"
>> Feb 23 15:55:30 SC-2 osaflogd[6988]: NO LOG data group is: ""
>> Feb 23 15:55:30 SC-2 osaflogd[6988]: NO LGS_MBCSV_VERSION = 7
>> Feb 23 15:55:30 SC-2 osafamfnd[7028]: NO Assigned
>> 'safSi=SC-2N,safApp=OpenSAF' STANDBY to 'safSu=SC-
>> 2,safSg=2N,safApp=OpenSAF'
>> Feb 23 15:55:30 SC-2 opensafd: OpenSAF(5.1.M0 - ) services successfully
>> started
>> done
>> SC-2:/var/log/opensaf # Feb 23 15:55:31 SC-2 osafimmnd[6978]: NO
>> Implementer (applier) connected: 16 (@OpenSafImmReplicatorB) <144,
>> 2020f>
>> Feb 23 15:55:31 SC-2 osafntfimcnd[7185]: NO Started
>> Feb 23 15:55:33 SC-2 osafamfd[7018]: NO Cold sync complete!
>> Feb 23 15:55:33 SC-2 osaflogd[6988]: WA FAILED: ncs_patricia_tree_add,
>> client_id 0
>> Feb 23 15:55:33 SC-2 osaflogd[6988]: ER Exiting with message: Could not
>> create new client
>> Feb 23 15:55:33 SC-2 osafamfnd[7028]: NO
>> 'safSu=SC-2,safSg=2N,safApp=OpenSAF' component restart probation timer
>> started (timeout: 60000000000 ns)
>> Feb 23 15:55:33 SC-2 osafamfnd[7028]: NO Restarting a component of
>> 'safSu=SC-2,safSg=2N,safApp=OpenSAF' (comp restart count: 1)
>> Feb 23 15:55:33 SC-2 osafamfnd[7028]: NO
>> 'safComp=LOG,safSu=SC-2,safSg=2N,safApp=OpenSAF' faulted due to
>> 'errorReport' : Recovery is 'componentRestart'
>>
>> ==============================================================
>> ==============================================
>>
>> -AVM
>>
>>
>> On 2/23/2017 3:39 PM, Vu Minh Nguyen wrote:
>>> Hi Mahesh,
>>>
>>> No change in V7 vs V9. Just do rebase the code on latest changeset.
>>>
>>> I have tried to clean up all, and rebuild the cluster to see what you
> are
>>> observing,
>>> and I am not able to reproduce the problem, I have tried several times.
>>>
>>> Can you provide me the osaflogd trace on both SCs node? Thanks.
>>>
>>> Regards, Vu
>>>
>>>> -----Original Message-----
>>>> From: A V Mahesh [mailto:mahesh.va...@oracle.com]
>>>> Sent: Thursday, February 23, 2017 4:48 PM
>>>> To: Vu Minh Nguyen <vu.m.ngu...@dektech.com.au>;
>>>> lennart.l...@ericsson.com; canh.v.tru...@dektech.com.au
>>>> Cc: opensaf-devel@lists.sourceforge.net
>>>> Subject: Re: [devel] [PATCH 0 of 3] Review Request for log: add
>>> alternative
>>>> destinations of log records [#2258] V4
>>>>
>>>> Hi Vu,
>>>>
>>>> On 2/23/2017 3:13 PM, A V Mahesh wrote:
>>>>> Not sure what are other change compare to V7 to V9 , New problems
>> got
>>>>> introduced
>>>>>
>>>>> Both nodes  SC-1 & SC-2 ( with 2258_v9.patch ) , trying bring up both
>>>>> SC`s  simple node bringup  ,
>>>>>
>>>>> SC-2 going for reboot with following :
>>>>>
>>>>>
>> ==============================================================
>> ==============================================================
>>>> ============
>>>>> Feb 23 15:05:32 SC-2 osaflogd[29988]: NO LOGSV_DATA_GROUPNAME
>>>> not found
>>>>> Feb 23 15:05:32 SC-2 osaflogd[29988]: NO LOG root directory is:
>>>>> "/var/log/opensaf/saflog"
>>>>> Feb 23 15:05:32 SC-2 osaflogd[29988]: NO LOG data group is: ""
>>>>> Feb 23 15:05:32 SC-2 osafimmnd[29978]: NO Implementer (applier)
>>>>> connected: 16 (@safAmfService2020f) <127, 2020f>
>>>>> Feb 23 15:05:32 SC-2 osaflogd[29988]: NO LGS_MBCSV_VERSION = 7
>>>>> Feb 23 15:05:32 SC-2 osaflogd[29988]: WA FAILED:
>>>>> ncs_patricia_tree_add, client_id 0
>>>>> Feb 23 15:05:32 SC-2 osafamfnd[30028]: NO Assigned
>>>>> 'safSi=SC-2N,safApp=OpenSAF' STANDBY to
>>>>> 'safSu=SC-2,safSg=2N,safApp=OpenSAF'
>>>>> Feb 23 15:05:32 SC-2 osaflogd[29988]: ER Exiting with message: Could
>>>>> not create new client
>>>>> Feb 23 15:05:32 SC-2 osafamfnd[30028]: NO
>>>>> 'safSu=SC-2,safSg=2N,safApp=OpenSAF' component restart probation
>>>> timer
>>>>> started (timeout: 60000000000 ns)
>>>>> Feb 23 15:05:32 SC-2 osafamfnd[30028]: NO Restarting a component of
>>>>> 'safSu=SC-2,safSg=2N,safApp=OpenSAF' (comp restart count: 1)
>>>>> Feb 23 15:05:32 SC-2 osafamfnd[30028]: NO
>>>>> 'safComp=LOG,safSu=SC-2,safSg=2N,safApp=OpenSAF' faulted due to
>>>>> 'errorReport' : Recovery is 'componentRestart'
>>>>> Feb 23 15:05:32 SC-2 opensafd[29908]: ER Service LOGD has
>> unexpectedly
>>>>> crashed. Unable to continue, exiting
>>>>> Feb 23 15:05:32 SC-2 osafamfd[30018]: exiting for shutdown
>>>>> Feb 23 15:05:32 SC-2 osafamfnd[30028]: ER AMFD has unexpectedly
>>>>> crashed. Rebooting node
>>>>> Feb 23 15:05:32 SC-2 osafamfnd[30028]: Rebooting OpenSAF NodeId =
>>>>> 131599 EE Name = , Reason: AMFD has unexpectedly crashed.
>> Rebooting
>>>>> node, OwnNodeId = 131599, SupervisionTime = 60
>>>>> Feb 23 15:05:32 SC-2 opensaf_reboot: Rebooting local node;
>> timeout=60
>>>>> Feb 23 15:06:04 SC-2 syslog-ng[1180]: syslog-ng starting up;
>>>>> version='2.0.9'
>>>>>
>>>>>
>> ==============================================================
>> ==============================================================
>>>> ============
>>>> Some times :
>>>>
>>>>
>> ==============================================================
>> ==============================================================
>>>> ============
>>>>
>>>> Feb 23 15:15:19 SC-2 osafrded[3858]: NO RDE role set to STANDBY
>>>> Feb 23 15:15:19 SC-2 osafrded[3858]: NO Peer up on node 0x2010f
>>>> Feb 23 15:15:19 SC-2 osafrded[3858]: NO Got peer info request from
>> node
>>>> 0x2010f with role ACTIVE
>>>> Feb 23 15:15:19 SC-2 osafrded[3858]: NO Got peer info response from
>> node
>>>> 0x2010f with role ACTIVE
>>>> Feb 23 15:15:19 SC-2 osafimmd[3877]: NO MDS event from svc_id 24
>>>> (change:3, dest:13)
>>>> Feb 23 15:15:19 SC-2 osafimmd[3877]: NO MDS event from svc_id 24
>>>> (change:5, dest:13)
>>>> Feb 23 15:15:19 SC-2 osafimmd[3877]: NO MDS event from svc_id 24
>>>> (change:5, dest:13)
>>>> Feb 23 15:15:19 SC-2 osafimmd[3877]: NO MDS event from svc_id 25
>>>> (change:3, dest:565217560625168)
>>>> Feb 23 15:15:19 SC-2 osafimmd[3877]: NO MDS event from svc_id 25
>>>> (change:3, dest:564114674417680)
>>>> Feb 23 15:15:19 SC-2 osaflogd[3898]: NO LOGSV_DATA_GROUPNAME
>> not
>>>> found
>>>> Feb 23 15:15:19 SC-2 osaflogd[3898]: NO LOG root directory is:
>>>> "/var/log/opensaf/saflog"
>>>> Feb 23 15:15:19 SC-2 osaflogd[3898]: NO LOG data group is: ""
>>>> Feb 23 15:15:19 SC-2 osafimmnd[3888]: NO Implementer (applier)
>>>> connected: 15 (@safAmfService2020f) <127, 2020f>
>>>> Feb 23 15:15:19 SC-2 osaflogd[3898]: NO LGS_MBCSV_VERSION = 7
>>>> Feb 23 15:15:19 SC-2 osaflogd[3898]: ER Exiting with message: Client
>>>> attributes differ
>>>> Feb 23 15:15:19 SC-2 osafamfnd[3938]: NO
>>>> 'safSu=SC-2,safSg=2N,safApp=OpenSAF' component restart probation
>> timer
>>>> started (timeout: 60000000000 ns)
>>>> Feb 23 15:15:19 SC-2 osafamfnd[3938]: NO Restarting a component of
>>>> 'safSu=SC-2,safSg=2N,safApp=OpenSAF' (comp restart count: 1)
>>>> Feb 23 15:15:19 SC-2 osafamfnd[3938]: NO
>>>> 'safComp=LOG,safSu=SC-2,safSg=2N,safApp=OpenSAF' faulted due to
>>>> 'errorReport' : Recovery is 'componentRestart'
>>>> Feb 23 15:15:19 SC-2 opensafd[3818]: ER Service LOGD has unexpectedly
>>>> crashed. Unable to continue, exiting
>>>> Feb 23 15:15:20 SC-2 osafamfd[3928]: exiting for shutdown
>>>> Feb 23 15:15:20 SC-2 osafamfnd[3938]: ER AMFD has unexpectedly
>> crashed.
>>>> Rebooting node
>>>> Feb 23 15:15:20 SC-2 osafamfnd[3938]: Rebooting OpenSAF NodeId =
>>>> 131599
>>>> EE Name = , Reason: AMFD has unexpectedly crashed. Rebooting node,
>>>> OwnNodeId = 131599, SupervisionTime = 60
>>>> Feb 23 15:15:20 SC-2 osafimmnd[3888]: NO Implementer locally
>>>> disconnected. Marking it as doomed 15 <127, 2020f>
>> (@safAmfService2020f)
>>>> Feb 23 15:15:20 SC-2 osafimmnd[3888]: NO Implementer disconnected 15
>>>> <127, 2020f> (@safAmfService2020f)
>>>> Feb 23 15:15:20 SC-2 opensaf_reboot: Rebooting local node; timeout=60
>>>>
>> ==============================================================
>> ==============================================================
>>>> ============
>>>>
>>>>
>>>>> -AVM
>>>>>
>>>>>
>>>>> On 2/23/2017 2:20 PM, Vu Minh Nguyen wrote:
>>>>>> Hi Mahesh,
>>>>>>
>>>>>> This is the latest code has been rebased on the latest changeset.
>>>>>>
>>>>>> Note that, in the attached patch, I have included one more
>> dependency,
>>>>>> that is on base::Hash() function, the patch sent by Anders [#2266]
>>>>>>
>>>>>> Please review the patch, then comment if any. Thanks.
>>>>>>
>>>>>> Regards, Vu
>>>>>>
>>>>>>> -----Original Message-----
>>>>>>> From: A V Mahesh [mailto:mahesh.va...@oracle.com]
>>>>>>> Sent: Thursday, February 23, 2017 2:03 PM
>>>>>>> To: Vu Minh Nguyen <vu.m.ngu...@dektech.com.au>;
>>>>>>> lennart.l...@ericsson.com; canh.v.tru...@dektech.com.au
>>>>>>> Cc: opensaf-devel@lists.sourceforge.net
>>>>>>> Subject: Re: [devel] [PATCH 0 of 3] Review Request for log: add
>>>>>> alternative
>>>>>>> destinations of log records [#2258] V4
>>>>>>>
>>>>>>> Hi Vu,
>>>>>>>
>>>>>>> Now we are now able to proceed further with  V7 `2258_v7.patch` in
>>>>>>> service upgraded working fine,
>>>>>>> because of Encode/decode changes done in V7 patch.
>>>>>>>
>>>>>>> But we have another small test case issue (/usr/bin/logtest 5 17
>>>>>>> Segmentation fault),
>>>>>>> once we resolve this also, we can conclude that all the basic
>>>>>>> functionality is working,
>>>>>>> then you can re-publish the V7 patch  ( if change occurred in
> Lennart
>>>>>>> #2258 V2 please do publish that as well )
>>>>>>> so that I can go for CODE review.
>>>>>>>
>>>>>>> Steps to reproduce the test case issue :
>>>>>>>
>>>>>>> 1) Bring up old node as Active  ( with out `2258_v7.patch` )
>>>>>>> 2) Bring-up new node as Standby  ( with `2258_v7.patch` )
>>>>>>> 3)  Do `amf-adm si-swap safSi=SC-2N,safApp=OpenSAF`
>>>>>>> 4)  Run `/usr/bin/logtest 5 17 ` on  new Active (because of si-swap
> )
>>>>>>> Note :  both nodes has the new XLM attributes populated .
>>>>>>>
>>>>>>>
>> ==============================================================
>>>>>>> =====================
>>>>>>>
>>>>>>> gdb /usr/bin/logtest
>>>>>>> (gdb) r 5
>>>>>>>
>>>>>>>        16  PASSED   CCB Object Modify, change root directory. Path
>>>>>>> exist. OK;
>>>>>>> Detaching after fork from child process 13797.
>>>>>>> Set values Fail
>>>>>>> [New Thread 0x7ffff7ff7b00 (LWP 13801)]
>>>>>>> [New Thread 0x7ffff7fc4b00 (LWP 13802)]
>>>>>>>
>>>>>>> Program received signal SIGSEGV, Segmentation fault.
>>>>>>> 0x00005555555688ea in read_and_compare.isra.7 () at
>>>>>>> src/log/apitest/tet_LogOiOps.c:1891
>>>>>>> 1891    src/log/apitest/tet_LogOiOps.c: No such file or directory.
>>>>>>>             in src/log/apitest/tet_LogOiOps.c
>>>>>>> (gdb) bt
>>>>>>> #0  0x00005555555688ea in read_and_compare.isra.7 () at
>>>>>>> src/log/apitest/tet_LogOiOps.c:1891
>>>>>>> #1  0x0000555555568a4b in
>>>> check_logRecordDestinationConfigurationAdd ()
>>>>>>> at src/log/apitest/tet_LogOiOps.c:1941
>>>>>>> #2  0x0000555555571b05 in run_test_case ()
>>>>>>> #3  0x0000555555571feb in test_run ()
>>>>>>> #4  0x000055555555bfad in main () at src/log/apitest/logtest.c:569
>>>>>>> (gdb)
>>>>>>>
>>>>>>>
>> ==============================================================
>>>>>>> =====================
>>>>>>>
>>>>>>>
>>>>>>> -AVM
>>>>>>>
>>>>>>> On 2/23/2017 11:44 AM, Vu Minh Nguyen wrote:
>>>>>>>> Hi Mahesh,
>>>>>>>>
>>>>>>>> Maybe it was broken when transmitting. I zipped to a tar file.
> Please
>>>>>> try it
>>>>>>>> one more.
>>>>>>>>
>>>>>>>> Regards, Vu
>>>>>>>>
>>>>>>>>
>>>>>>>>> -----Original Message-----
>>>>>>>>> From: A V Mahesh [mailto:mahesh.va...@oracle.com]
>>>>>>>>> Sent: Thursday, February 23, 2017 12:54 PM
>>>>>>>>> To: Vu Minh Nguyen <vu.m.ngu...@dektech.com.au>;
>>>>>>>>> lennart.l...@ericsson.com; canh.v.tru...@dektech.com.au
>>>>>>>>> Cc: opensaf-devel@lists.sourceforge.net
>>>>>>>>> Subject: Re: [devel] [PATCH 0 of 3] Review Request for log: add
>>>>>>>> alternative
>>>>>>>>> destinations of log records [#2258] V4
>>>>>>>>>
>>>>>>>>> Hi Vu,
>>>>>>>>>
>>>>>>>>> On 2/23/2017 10:20 AM, Vu Minh Nguyen wrote:
>>>>>>>>>> Hi Mahesh,
>>>>>>>>>>
>>>>>>>>>> Can you try with 2258_v7.patch I just sent to you?
>>>>>>>>> I striped    changeset:   8610 of today's latest staging (  `hg
>>> strip
>>>>>>>>> 8610` which  removed log: implement SaLogFilterSetCallbackT and
>>>>>>>>> version
>>>>>>>>> handling [#2146])
>>>>>>>>> and try to apply your `2258_v7.patch`, it says `malformed patch at
>>>>>>>>> line
>>>>>>>>> 3324`.
>>>>>>>>>
>>>>>>>>> -AVM
>>>>>>>>>> I have pulled the latest code on OpenSAF 5.1 branch, re-created
>> the
>>>>>>>> cluster.
>>>>>>>>>> And it works with the case old active SC-1 (OpenSAF 5.1) and new
>>>>>>> standby
>>>>>>>>>> SC-2 (with 2258_v7.patch included in).
>>>>>>>>>>
>>>>>>>>>> To apply 2258_v7.patch, please do remove the just pushed ticket
>>>>>>>>>> "log:
>>>>>>>>>> implement SaLogFilterSetCallbackT and version handling [#2146]" ,
>>>>>>>>>> I have not rebased the code on that yet.
>>>>>>>>>>
>>>>>>>>>> Regards, Vu
>>>>>>>>>>
>>>>>>>>>>> -----Original Message-----
>>>>>>>>>>> From: A V Mahesh [mailto:mahesh.va...@oracle.com]
>>>>>>>>>>> Sent: Thursday, February 23, 2017 11:45 AM
>>>>>>>>>>> To: Vu Minh Nguyen <vu.m.ngu...@dektech.com.au>;
>>>>>>>>>>> lennart.l...@ericsson.com; canh.v.tru...@dektech.com.au
>>>>>>>>>>> Cc: opensaf-devel@lists.sourceforge.net
>>>>>>>>>>> Subject: Re: [devel] [PATCH 0 of 3] Review Request for log: add
>>>>>>>>>> alternative
>>>>>>>>>>> destinations of log records [#2258] V4
>>>>>>>>>>>
>>>>>>>>>>> Hi Vu/Lennart,
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> In broad WITHOUT the #2258 patch, the same code/setup
>> working
>>>> fine
>>>>>>>>> with
>>>>>>>>>>> 2 sc node (staging  changeset: 8609 ),
>>>>>>>>>>> as soon as we apply `2258_v5.patch` V5 patch on staging
>>>> (changeset:
>>>>>>>>>>> 8609 ) that you have provided yesterday,
>>>>>>>>>>> on one sc node and try to bring up that in to cluster
> (in-service
>>>>>> test)
>>>>>>>>>>> we are observing the issue of new node (with #2258 patch) not
>>>>>>>>>>> joining
>>>>>>>>>>> cluster.
>>>>>>>>>>>
>>>>>>>>>>>
>> ==============================================================
>>>>>>>>>>> ====================================================
>>>>>>>>>>> eb 23 10:01:59 SC-1 osafimmnd[15279]: NO Implementer
>> (applier)
>>>>>>>>>>> connected: 15 (@safAmfService2010f) <127, 2010f>
>>>>>>>>>>> Feb 23 10:01:59 SC-1 osaflogd[15289]: NO
>>>> LOGSV_DATA_GROUPNAME
>>>>>>>>> not
>>>>>>>>>>> found
>>>>>>>>>>> Feb 23 10:01:59 SC-1 osaflogd[15289]: NO LOG root directory is:
>>>>>>>>>>> "/var/log/opensaf/saflog"
>>>>>>>>>>> Feb 23 10:01:59 SC-1 osaflogd[15289]: NO LOG data group is: ""
>>>>>>>>>>> Feb 23 10:01:59 SC-1 osaflogd[15289]: NO
>> LGS_MBCSV_VERSION =
>>>> 7
>>>>>>>>>>> Feb 23 10:01:59 SC-1 osafamfnd[15329]: NO Assigned
>>>>>>>>>>> 'safSi=SC-2N,safApp=OpenSAF' STANDBY to 'safSu=SC-
>>>>>>>>>>> 1,safSg=2N,safApp=OpenSAF'
>>>>>>>>>>> Feb 23 10:01:59 SC-1 opensafd: OpenSAF(5.1.M0 - ) services
>>>>>>> successfully
>>>>>>>>>>> started
>>>>>>>>>>> Feb 23 10:01:59 SC-1 osafamfnd[15329]: NO
>>>>>>>>>>> 'safComp=LOG,safSu=SC-1,safSg=2N,safApp=OpenSAF' faulted
>> due
>>>> to
>>>>>>>>>>> 'avaDown' : Recovery is 'nodeFailfast'
>>>>>>>>>>> Feb 23 10:01:59 SC-1 osafamfnd[15329]: ER
>>>>>>>>>>> safComp=LOG,safSu=SC-1,safSg=2N,safApp=OpenSAF Faulted
>> due
>>>>>>>>>>> to:avaDown
>>>>>>>>>>> Recovery is:nodeFailfast
>>>>>>>>>>> Feb 23 10:01:59 SC-1 osafamfnd[15329]: Rebooting OpenSAF
>>>> NodeId =
>>>>>>>>>>> 131343
>>>>>>>>>>> EE Name = , Reason: Component faulted: recovery is node
>> failfast,
>>>>>>>>>>> OwnNodeId = 131343, SupervisionTime = 60
>>>>>>>>>>> Feb 23 10:01:59 SC-1 opensaf_reboot: Rebooting local node;
>>>>>>> timeout=60
>>>>>>>>>>> Feb 23 10:02:00 SC-1 osafimmnd[15279]: NO Implementer
>> (applier)
>>>>>>>>>>> connected: 16 (@OpenSafImmReplicatorB) <144, 2010f>
>>>>>>>>>>> Feb 23 10:01:59 SC-1 opensaf_reboot: Rebooting local node;
>>>>>>> timeout=60
>>>>>>>
>> ==============================================================
>>>>>>>>>>> ====================================================
>>>>>>>>>>>
>>>>>>>>>>> So it is evident that in-service upgrade part code of this need
>>>>>>>>>>> to be
>>>>>>>>>>> corrected.
>>>>>>>>>>>
>>>>>>>>>>> Please see my comments as [AVM] and let me know if you need
>>>> some
>>>>>>>>> traces
>>>>>>>>>>> .
>>>>>>>>>>>
>>>>>>>>>>> If you're planing to prepare new V6 patch , please do prepare on
>>>>>>>>>>> top
>>>>>> of
>>>>>>>>>>> today's latest staging.
>>>>>>>>>>>
>>>>>>>>>>> On 2/23/2017 9:33 AM, Vu Minh Nguyen wrote:
>>>>>>>>>>>> Hi Mahesh,
>>>>>>>>>>>>
>>>>>>>>>>>> I have done in-service upgrade/downgrade with following
>> cases:
>>>>>>>>>>>> 1) New Active SC-1 (OpenSAF 5.2 with the attached patch) + old
>>>>>>> standby
>>>>>>>>>>> SC-2
>>>>>>>>>>>> (OpenSAF 5.1)
>>>>>>>>>>>> --> Work fine
>>>>>>>>>>> [AVM] This is not a practical use cause of in-service upgrade ,
> we
>>>>>> can
>>>>>>>>>>> ignore this test further
>>>>>>>>>>>> 2) Old Active SC-1 (OpenSAF 5.1) + new standby SC-2 (with or
>>>>>>>>>>>> without
>>>>>>>>>>>> attached patch)
>>>>>>>>>>>> --> SC-2 is restarted & not able to join the cluster.
>>>>>>>>>>> [AVM] This use cause/flow is  we do get in in-service upgrade ,
>>>>>>>>>>> so we
>>>>>>>>>>> need to address this.
>>>>>>>>>>>> I got following messages in syslog:
>>>>>>>>>>>> Feb 23 09:32:42 SC-2 user.notice opensafd: OpenSAF(5.2.M0 -
>>>>>>>>>>>> 8529:b5addd36e45d:default) services successfully started
>>>>>>>>>>>> Feb 23 09:32:43 SC-2 local0.warn osafntfimcnd[701]: WA
>>>>>>>>>>> ntfimcn_imm_init
>>>>>>>>>>>> saImmOiImplementerSet() returned SA_AIS_ERR_TIMEOUT (5)
>>>>>>>>>>>> Feb 23 09:32:45 SC-2 local0.warn osafntfimcnd[701]: WA
>>>>>>>>>>> ntfimcn_imm_init
>>>>>>>>>>>> saImmOiImplementerSet() returned SA_AIS_ERR_TIMEOUT (5)
>>>>>>>>>>>> Feb 23 09:32:47 SC-2 local0.warn osafntfimcnd[701]: WA
>>>>>>>>>>> ntfimcn_imm_init
>>>>>>>>>>>> saImmOiImplementerSet() returned SA_AIS_ERR_TIMEOUT (5)
>>>>>>>>>>>> Feb 23 09:32:49 SC-2 local0.warn osafntfimcnd[701]: WA
>>>>>>>>>>> ntfimcn_imm_init
>>>>>>>>>>>> saImmOiImplementerSet() returned SA_AIS_ERR_TIMEOUT (5)
>>>>>>>>>>>> Feb 23 09:32:50 SC-2 local0.err osafmsgnd[592]: ER
>>>>>>>>>>> saImmOiImplementerSet
>>>>>>>>>>>> FAILED:5
>>>>>>>>>>>> Feb 23 09:32:50 SC-2 local0.err osafmsgnd[592]: ER
>>>>>>>>>>> saImmOiImplementerSet
>>>>>>>>>>>> FAILED:5
>>>>>>>>>>>> Feb 23 09:32:50 SC-2 local0.notice osafamfnd[496]: NO
>>>>>>>>>>>> 'safSu=SC-2,safSg=NoRed,safApp=OpenSAF' component restart
>>>>>>>>> probation
>>>>>>>>>>> timer
>>>>>>>>>>>> started (timeout: 60000000000 ns)
>>>>>>>>>>>> Feb 23 09:32:50 SC-2 local0.notice osafamfnd[496]: NO
>> Restarting
>>>> a
>>>>>>>>>>> component
>>>>>>>>>>>> of 'safSu=SC-2,safSg=NoRed,safApp=OpenSAF' (comp restart
>>>> count: 1)
>>>>>>>>>>>> Feb 23 09:32:50 SC-2 local0.notice osafamfnd[496]: NO
>>>>>>>>>>>> 'safComp=MQND,safSu=SC-2,safSg=NoRed,safApp=OpenSAF'
>>>> faulted
>>>>>>>>> due
>>>>>>>>>>> to
>>>>>>>>>>>> 'avaDown' : Recovery is 'componentRestart'
>>>>>>>>>>>> Feb 23 09:32:50 SC-2 local0.info osafmsgnd[736]: mkfifo
>> already
>>>>>>>> exists:
>>>>>>>>>>>> /var/lib/opensaf/osafmsgnd.fifo File exists
>>>>>>>>>>>>
>>>>>>>>>>>> And sometimes, on active SC-1 (OpenSAF 5.1), the node is not
>>>>>>>>>>>> able to
>>>>>>>> up
>>>>>>>>>>>> because of following error:
>>>>>>>>>>>>
>>>>>>>>>>>> Feb 23 11:00:32 SC-1 local0.err osafclmna[406]: MDTM:TIPC
>>>> Dsock
>>>>>>>>> Socket
>>>>>>>>>>>> creation failed in MDTM_INIT err :Address family not supported
>> by
>>>>>>>>>>> protocol
>>>>>>>>>>>> Feb 23 11:00:32 SC-1 local0.err osafclmna[406]: ER
>>>>>>> ncs_agents_startup
>>>>>>>>>>> FAILED
>>>>>>>>>>> [AVM]  No such issues ( with both TCP & TIPC) (staging
>> changeset:
>>>>>>>> 8609
>>>>>>>>>> )
>>>>>>>>>>>> Are you getting similar problem at your side?
>>>>>>>>>>>> Please note that, the problem is existed WITH or WITHOUT the
>>>> #2258
>>>>>>>>>>> patch.
>>>>>>>>>>> [AVM] No , problem only if we apply `2258_v5.patch` V5 patch on
>>>>>>> staging
>>>>>>>>>>> (changeset:   8609 )
>>>>>>>>>>>                  try to bring up that node in to cluster.
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> -AVM
>>>>>>>>>>>
>>>>>>>>>>>> I have informed this to IMM to have a look, not sure any
>> problem
>>>>>> with
>>>>>>>>>> MDS
>>>>>>>>>>>> layer or any problem with my environment setup.
>>>>>>>>>>>> In the meantime, please have a look at the updated patch,  I
> will
>>>>>>>>>> continue
>>>>>>>>>>>> checking the problem. Will keep you updated.
>>>>>>>>>>> [AVM] I haven't seen any IMM problems
>>>>>>>>>>>> Regards, Vu
>>>>>>>>>>>>
>>>>>>>>>>>>> -----Original Message-----
>>>>>>>>>>>>> From: A V Mahesh [mailto:mahesh.va...@oracle.com]
>>>>>>>>>>>>> Sent: Wednesday, February 22, 2017 5:36 PM
>>>>>>>>>>>>> To: Vu Minh Nguyen <vu.m.ngu...@dektech.com.au>;
>>>>>>>>>>>>> lennart.l...@ericsson.com; canh.v.tru...@dektech.com.au
>>>>>>>>>>>>> Cc: opensaf-devel@lists.sourceforge.net
>>>>>>>>>>>>> Subject: Re: [devel] [PATCH 0 of 3] Review Request for log:
>> add
>>>>>>>>>>>> alternative
>>>>>>>>>>>>> destinations of log records [#2258] V4
>>>>>>>>>>>>>
>>>>>>>>>>>>> Please see correction New Standby SC-1 ( with patch )
>>>>>>>>>>>>>
>>>>>>>>>>>>> -AVM
>>>>>>>>>>>>>
>>>>>>>>>>>>> On 2/22/2017 4:02 PM, A V Mahesh wrote:
>>>>>>>>>>>>>> Hi Vu,
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> With this new patch , we have another issue :
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> 1)  standby Core by `/usr/lib64/opensaf/osaflogd'  issue got
>>>>>>>> resolved
>>>>>>>>>> .
>>>>>>>>>>>>>> 2) In-service upgrade is Not working , I have Old  Active
> SC-2
>>> (
>>>>>>>> with
>>>>>>>>>>>>>> out patch )  and New Standby SC-1 ( with patch )
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>           the new New Standby SC-1 not joining the cluster (
>>>>>> in-service
>>>>>>>>>>>>>> upgrade  failed )
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> New Standby SC-1
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>> ==============================================================
>>>> ======================================================
>>>>>>>>>>>>>> Feb 22 15:53:05 SC-1 osafamfnd[15752]: NO
>>>>>>>>>>>>>> 'safSu=SC-1,safSg=NoRed,safApp=OpenSAF' Presence State
>>>>>>>>>>>>> INSTANTIATING
>>>>>>>>>>>>>> => INSTANTIATED
>>>>>>>>>>>>>> Feb 22 15:53:05 SC-1 osafamfnd[15752]: NO Assigning
>>>>>>>>>>>>>> 'safSi=NoRed4,safApp=OpenSAF' ACTIVE to
>>>>>>>>>>>>>> 'safSu=SC-1,safSg=NoRed,safApp=OpenSAF'
>>>>>>>>>>>>>> Feb 22 15:53:05 SC-1 osafamfnd[15752]: NO Assigned
>>>>>>>>>>>>>> 'safSi=NoRed4,safApp=OpenSAF' ACTIVE to
>>>>>>>>>>>>>> 'safSu=SC-1,safSg=NoRed,safApp=OpenSAF'
>>>>>>>>>>>>>> Feb 22 15:53:05 SC-1 osafsmfd[15889]: Started
>>>>>>>>>>>>>> Feb 22 15:53:05 SC-1 osafamfnd[15752]: NO
>>>>>>>>>>>>>> 'safSu=SC-1,safSg=2N,safApp=OpenSAF' Presence State
>>>>>>>>> INSTANTIATING
>>>>>>>>>>> =>
>>>>>>>>>>>>>> INSTANTIATED
>>>>>>>>>>>>>> Feb 22 15:53:05 SC-1 osafamfnd[15752]: NO Assigning
>>>>>>>>>>>>>> 'safSi=SC-2N,safApp=OpenSAF' STANDBY to
>>>>>>>>>>>>>> 'safSu=SC-1,safSg=2N,safApp=OpenSAF'
>>>>>>>>>>>>>> Feb 22 15:53:05 SC-1 osafrded[15672]: NO RDE role set to
>>>>>>> STANDBY
>>>>>>>>>>>>>> Feb 22 15:53:05 SC-1 osafrded[15672]: NO Peer up on node
>>>>>>> 0x2020f
>>>>>>>>>>>>>> Feb 22 15:53:05 SC-1 osafrded[15672]: NO Got peer info
>>>> request
>>>>>>> from
>>>>>>>>>>>>>> node 0x2020f with role ACTIVE
>>>>>>>>>>>>>> Feb 22 15:53:05 SC-1 osafrded[15672]: NO Got peer info
>>>> response
>>>>>>>>> from
>>>>>>>>>>>>>> node 0x2020f with role ACTIVE
>>>>>>>>>>>>>> Feb 22 15:53:05 SC-1 osafimmd[15691]: NO MDS event from
>>>> svc_id
>>>>>>> 24
>>>>>>>>>>>>>> (change:5, dest:13)
>>>>>>>>>>>>>> Feb 22 15:53:05 SC-1 osafimmd[15691]: NO MDS event from
>>>> svc_id
>>>>>>> 24
>>>>>>>>>>>>>> (change:3, dest:13)
>>>>>>>>>>>>>> Feb 22 15:53:05 SC-1 osafimmd[15691]: NO MDS event from
>>>> svc_id
>>>>>>> 24
>>>>>>>>>>>>>> (change:5, dest:13)
>>>>>>>>>>>>>> Feb 22 15:53:05 SC-1 osafimmd[15691]: NO MDS event from
>>>> svc_id
>>>>>>> 25
>>>>>>>>>>>>>> (change:3, dest:567412424453430)
>>>>>>>>>>>>>> Feb 22 15:53:05 SC-1 osafimmd[15691]: NO MDS event from
>>>> svc_id
>>>>>>> 25
>>>>>>>>>>>>>> (change:3, dest:565213401202663)
>>>>>>>>>>>>>> Feb 22 15:53:05 SC-1 osafimmd[15691]: NO MDS event from
>>>> svc_id
>>>>>>> 25
>>>>>>>>>>>>>> (change:3, dest:566312912825221)
>>>>>>>>>>>>>> Feb 22 15:53:05 SC-1 osafimmd[15691]: NO MDS event from
>>>> svc_id
>>>>>>> 25
>>>>>>>>>>>>>> (change:3, dest:564113889574230)
>>>>>>>>>>>>>> Feb 22 15:53:05 SC-1 osafimmnd[15702]: NO Implementer
>>>> (applier)
>>>>>>>>>>>>>> connected: 17 (@safAmfService2010f) <127, 2010f>
>>>>>>>>>>>>>> Feb 22 15:53:05 SC-1 osaflogd[15712]: NO
>>>>>>>>> LOGSV_DATA_GROUPNAME
>>>>>>>>>>>>> not found
>>>>>>>>>>>>>> Feb 22 15:53:05 SC-1 osaflogd[15712]: NO LOG root
>> directory
>>>> is:
>>>>>>>>>>>>>> "/var/log/opensaf/saflog"
>>>>>>>>>>>>>> Feb 22 15:53:05 SC-1 osaflogd[15712]: NO LOG data group is:
>> ""
>>>>>>>>>>>>>> Feb 22 15:53:05 SC-1 osaflogd[15712]: NO
>>>> LGS_MBCSV_VERSION =
>>>>>>> 7
>>>>>>>>>>>>>> Feb 22 15:53:05 SC-1 osafamfnd[15752]: NO Assigned
>>>>>>>>>>>>>> 'safSi=SC-2N,safApp=OpenSAF' STANDBY to
>>>>>>>>>>>>>> 'safSu=SC-1,safSg=2N,safApp=OpenSAF'
>>>>>>>>>>>>>> Feb 22 15:53:05 SC-1 opensafd: OpenSAF(5.1.M0 - ) services
>>>>>>>>>>>>>> successfully started
>>>>>>>>>>>>>> Feb 22 15:53:05 SC-1 osafamfnd[15752]: NO
>>>>>>>>>>>>>> 'safComp=LOG,safSu=SC-1,safSg=2N,safApp=OpenSAF'
>> faulted
>>>> due
>>>>>>> to
>>>>>>>>>>>>>> 'avaDown' : Recovery is 'nodeFailfast'
>>>>>>>>>>>>>> Feb 22 15:53:05 SC-1 osafamfnd[15752]: ER
>>>>>>>>>>>>>> safComp=LOG,safSu=SC-1,safSg=2N,safApp=OpenSAF
>> Faulted
>>>> due
>>>>>>>>>>>>> to:avaDown
>>>>>>>>>>>>>> Recovery is:nodeFailfast
>>>>>>>>>>>>>> Feb 22 15:53:05 SC-1 osafamfnd[15752]: Rebooting OpenSAF
>>>>>>> NodeId
>>>>>>>>> =
>>>>>>>>>>>>>> 131343 EE Name = , Reason: Component faulted: recovery is
>>>> node
>>>>>>>>>>>>>> failfast, OwnNodeId = 131343, SupervisionTime = 60
>>>>>>>>>>>>>> Feb 22 15:53:05 SC-1 opensaf_reboot: Rebooting local node;
>>>>>>>>>>> timeout=60
>>>>>>>>>>>>>> Feb 22 15:53:43 SC-1 syslog-ng[1171]: syslog-ng starting up;
>>>>>>>>>>>>>> version='2.0.9'
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>> ==============================================================
>>>> ======================================================
>>>>>>>>>>>>>> Old - Active - SC-2
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>> ==============================================================
>>>> ======================================================
>>>>>>>>>>>>>> Feb 22 15:53:02 SC-2 osafimmnd[16359]: NO NODE STATE->
>>>>>>>>>>>>>> IMM_NODE_R_AVAILABLE
>>>>>>>>>>>>>> Feb 22 15:53:02 SC-2 osafimmloadd: NO Sync starting
>>>>>>>>>>>>>> Feb 22 15:53:02 SC-2 osafimmloadd: IN Synced 390 objects in
>>>>>>>>>>>>>> total
>>>>>>>>>>>>>> Feb 22 15:53:02 SC-2 osafimmnd[16359]: NO NODE STATE->
>>>>>>>>>>>>>> IMM_NODE_FULLY_AVAILABLE 18511
>>>>>>>>>>>>>> Feb 22 15:53:02 SC-2 osafimmnd[16359]: NO Epoch set to 3
>> in
>>>>>>>>>>> ImmModel
>>>>>>>>>>>>>> Feb 22 15:53:02 SC-2 osafimmd[16346]: NO ACT: New Epoch
>>>> for
>>>>>>>>>>> IMMND
>>>>>>>>>>>>>> process at node 2020f old epoch: 2  new epoch:3
>>>>>>>>>>>>>> Feb 22 15:53:02 SC-2 osafimmd[16346]: NO ACT: New Epoch
>>>> for
>>>>>>>>>>> IMMND
>>>>>>>>>>>>>> process at node 2040f old epoch: 2  new epoch:3
>>>>>>>>>>>>>> Feb 22 15:53:02 SC-2 osafimmd[16346]: NO ACT: New Epoch
>>>> for
>>>>>>>>>>> IMMND
>>>>>>>>>>>>>> process at node 2030f old epoch: 2  new epoch:3
>>>>>>>>>>>>>> Feb 22 15:53:02 SC-2 osafimmloadd: NO Sync ending
>> normally
>>>>>>>>>>>>>> Feb 22 15:53:02 SC-2 osafimmd[16346]: NO ACT: New Epoch
>>>> for
>>>>>>>>>>> IMMND
>>>>>>>>>>>>>> process at node 2010f old epoch: 0  new epoch:3
>>>>>>>>>>>>>> Feb 22 15:53:02 SC-2 osafimmnd[16359]: NO SERVER STATE:
>>>>>>>>>>>>>> IMM_SERVER_SYNC_SERVER --> IMM_SERVER_READY
>>>>>>>>>>>>>> Feb 22 15:53:03 SC-2 osafamfd[16408]: NO Received
>> node_up
>>>> from
>>>>>>>>>>> 2010f:
>>>>>>>>>>>>>> msg_id 1
>>>>>>>>>>>>>> Feb 22 15:53:03 SC-2 osafamfd[16408]: NO Node 'SC-1'
>> joined
>>>> the
>>>>>>>>>>> cluster
>>>>>>>>>>>>>> Feb 22 15:53:03 SC-2 osafimmnd[16359]: NO Implementer
>>>>>>> connected:
>>>>>>>>>>> 16
>>>>>>>>>>>>>> (MsgQueueService131343) <0, 2010f>
>>>>>>>>>>>>>> Feb 22 15:53:03 SC-2 osafrded[16327]: NO Peer up on node
>>>>>>> 0x2010f
>>>>>>>>>>>>>> Feb 22 15:53:03 SC-2 osafrded[16327]: NO Got peer info
>>>> request
>>>>>>> from
>>>>>>>>>>>>>> node 0x2010f with role STANDBY
>>>>>>>>>>>>>> Feb 22 15:53:03 SC-2 osafrded[16327]: NO Got peer info
>>>> response
>>>>>>>>> from
>>>>>>>>>>>>>> node 0x2010f with role STANDBY
>>>>>>>>>>>>>> Feb 22 15:53:03 SC-2 osafimmd[16346]: NO MDS event from
>>>> svc_id
>>>>>>> 24
>>>>>>>>>>>>>> (change:5, dest:13)
>>>>>>>>>>>>>> Feb 22 15:53:03 SC-2 osafimmnd[16359]: NO Implementer
>>>> (applier)
>>>>>>>>>>>>>> connected: 17 (@safAmfService2010f) <0, 2010f>
>>>>>>>>>>>>>> Feb 22 15:53:03 SC-2 osafamfd[16408]: NO Cluster startup is
>>>> done
>>>>>>>>>>>>>> Feb 22 15:53:04 SC-2 osafimmnd[16359]: NO Implementer
>>>> (applier)
>>>>>>>>>>>>>> connected: 18 (@OpenSafImmReplicatorB) <0, 2010f>
>>>>>>>>>>>>>> Feb 22 15:53:09 SC-2 osafdtmd[16304]: NO Lost contact with
>>>>>>>>>>>>>> 'SC-1'
>>>>>>>>>>>>>> Feb 22 15:53:09 SC-2 osaffmd[16336]: NO Node Down event
>>>> for
>>>>>>> node
>>>>>>>>> id
>>>>>>>>>>>>>> 2010f:
>>>>>>>>>>>>>> Feb 22 15:53:09 SC-2 osafimmd[16346]: NO MDS event from
>>>> svc_id
>>>>>>> 24
>>>>>>>>>>>>>> (change:6, dest:13)
>>>>>>>>>>>>>> Feb 22 15:53:09 SC-2 osafimmd[16346]: NO MDS event from
>>>> svc_id
>>>>>>> 25
>>>>>>>>>>>>>> (change:4, dest:564113889574230)
>>>>>>>>>>>>>> Feb 22 15:53:09 SC-2 osaffmd[16336]: NO Current role:
>> ACTIVE
>>>>>>>>>>>>>> Feb 22 15:53:09 SC-2 osaffmd[16336]: Rebooting OpenSAF
>>>> NodeId
>>>>>>> =
>>>>>>>>>>>>> 131343
>>>>>>>>>>>>>> EE Name = , Reason: Received Node Down for peer
>> controller,
>>>>>>>>>>> OwnNodeId
>>>>>>>>>>>>>> = 131599, SupervisionTime = 60
>>>>>>>>>>>>>> Feb 22 15:53:09 SC-2 osafrded[16327]: NO Peer down on
>> node
>>>>>>>>> 0x2010f
>>>>>>>>>>>>>> Feb 22 15:53:09 SC-2 osafamfd[16408]: NO Node 'SC-1' left
>> the
>>>>>>>>> cluster
>>>>>>>>>>>>>> Feb 22 15:53:09 SC-2 osaflogd[16372]: NO Failed (2) to send
>> of
>>>>>>> WRITE
>>>>>>>>>>>>>> ack to: 2010f00003d6a
>>>>>>>>>>>>>> Feb 22 15:53:09 SC-2 osaflogd[16372]: NO Failed (2) to send
>> of
>>>>>>> WRITE
>>>>>>>>>>>>>> ack to: 2010f00003d6a
>>>>>>>>>>>>>> Feb 22 15:53:09 SC-2 osaflogd[16372]: NO Failed (2) to send
>> of
>>>>>>> WRITE
>>>>>>>>>>>>>> ack to: 2010f00003d74
>>>>>>>>>>>>>> Feb 22 15:53:09 SC-2 osafclmd[16398]: NO Node 131343
>> went
>>>>>>> down.
>>>>>>>>>>> Not
>>>>>>>>>>>>>> sending track callback for agents on that node
>>>>>>>>>>>>>> Feb 22 15:53:09 SC-2 osafclmd[16398]: NO Node 131343
>> went
>>>>>>> down.
>>>>>>>>>>> Not
>>>>>>>>>>>>>> sending track callback for agents on that node
>>>>>>>>>>>>>> Feb 22 15:53:09 SC-2 osafclmd[16398]: NO Node 131343
>> went
>>>>>>> down.
>>>>>>>>>>> Not
>>>>>>>>>>>>>> sending track callback for agents on that node
>>>>>>>>>>>>>> Feb 22 15:53:09 SC-2 osafclmd[16398]: NO Node 131343
>> went
>>>>>>> down.
>>>>>>>>>>> Not
>>>>>>>>>>>>>> sending track callback for agents on that node
>>>>>>>>>>>>>> Feb 22 15:53:09 SC-2 osafclmd[16398]: NO Node 131343
>> went
>>>>>>> down.
>>>>>>>>>>> Not
>>>>>>>>>>>>>> sending track callback for agents on that node
>>>>>>>>>>>>>> Feb 22 15:53:09 SC-2 osafclmd[16398]: NO Node 131343
>> went
>>>>>>> down.
>>>>>>>>>>> Not
>>>>>>>>>>>>>> sending track callback for agents on that node
>>>>>>>>>>>>>> Feb 22 15:53:09 SC-2 osafclmd[16398]: NO Node 131343
>> went
>>>>>>> down.
>>>>>>>>>>> Not
>>>>>>>>>>>>>> sending track callback for agents on that node
>>>>>>>>>>>>>> Feb 22 15:53:09 SC-2 osafimmd[16346]: WA IMMD lost
>> contact
>>>>>>> with
>>>>>>>>>>> peer
>>>>>>>>>>>>>> IMMD (NCSMDS_RED_DOWN)
>>>>>>>>>>>>>> Feb 22 15:53:09 SC-2 osafimmnd[16359]: NO Global discard
>>>> node
>>>>>>>>>>> received
>>>>>>>>>>>>>> for nodeId:2010f pid:15702
>>>>>>>>>>>>>> Feb 22 15:53:09 SC-2 osafimmnd[16359]: NO Implementer
>>>>>>>>> disconnected
>>>>>>>>>>> 16
>>>>>>>>>>>>>> <0, 2010f(down)> (MsgQueueService131343)
>>>>>>>>>>>>>> Feb 22 15:53:09 SC-2 osafimmnd[16359]: NO Implementer
>>>>>>>>> disconnected
>>>>>>>>>>> 17
>>>>>>>>>>>>>> <0, 2010f(down)> (@safAmfService2010f)
>>>>>>>>>>>>>> Feb 22 15:53:09 SC-2 osafimmnd[16359]: NO Implementer
>>>>>>>>> disconnected
>>>>>>>>>>> 18
>>>>>>>>>>>>>> <0, 2010f(down)> (@OpenSafImmReplicatorB)
>>>>>>>>>>>>>> Feb 22 15:53:09 SC-2 opensaf_reboot: Rebooting remote
>> node
>>>> in
>>>>>>> the
>>>>>>>>>>>>>> absence of PLM is outside the scope of OpenSAF
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>> ==============================================================
>>>> ======================================================
>>>>>>>>>>>>>> -AVM
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> On 2/22/2017 3:13 PM, Vu Minh Nguyen wrote:
>>>>>>>>>>>>>>> Hi Mahesh,
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> I put all required patches into one. Try to use this and see
>>> if
>>>>>> you
>>>>>>>>>>>>>>> still
>>>>>>>>>>>>>>> have that problem or not.
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> Regards, Vu
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> -----Original Message-----
>>>>>>>>>>>>>>>> From: A V Mahesh [mailto:mahesh.va...@oracle.com]
>>>>>>>>>>>>>>>> Sent: Wednesday, February 22, 2017 3:35 PM
>>>>>>>>>>>>>>>> To: Vu Minh Nguyen <vu.m.ngu...@dektech.com.au>;
>>>>>>>>>>>>>>>> lennart.l...@ericsson.com;
>> canh.v.tru...@dektech.com.au
>>>>>>>>>>>>>>>> Cc: opensaf-devel@lists.sourceforge.net
>>>>>>>>>>>>>>>> Subject: Re: [devel] [PATCH 0 of 3] Review Request for log:
>>>>>>>>>>>>>>>> add
>>>>>>>>>>>>>>> alternative
>>>>>>>>>>>>>>>> destinations of log records [#2258] V4
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> Hi Vu,
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> I used  new #3 and #4 patches ,  Can you please re-send All
>>>>>>>>>>>>>>>> the
>>>>>>>>>> final
>>>>>>>>>>>>>>>> patch in go,
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> which i need to apply on today`s  staging ( if possible
>>>>>>>>>>>>>>>> publish
>>>>>>>> the
>>>>>>>>>>>>>>>> with
>>>>>>>>>>>>>>>> new version )
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> -AVM
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> On 2/22/2017 1:52 PM, Vu Minh Nguyen wrote:
>>>>>>>>>>>>>>>>> Hi Mahesh,
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> Core was generated by `/usr/lib64/opensaf/osaflogd'.
>>>>>>>>>>>>>>>>>> Program terminated with signal 11, Segmentation fault.
>>>>>>>>>>>>>>>>>> #0  ckpt_proc_cfg_stream(lgs_cb*, void*) () at
>>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc:2195
>>>>>>>>>>>>>>>>>> 2195    src/log/logd/lgs_mbcsv.cc: No such file or
>>>>>>>>>>>>>>>>>> directory.
>>>>>>>>>>>>>>>>>>                  in src/log/logd/lgs_mbcsv.cc
>>>>>>>>>>>>>>>>> Backtrace still points to old position (lgs_mbcsv:2195). I
>>>>>> guess
>>>>>>>>>> the
>>>>>>>>>>>>>>>>> osaflogd binary has not been updated with the fixed
>> patch.
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> Regards, Vu
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> -----Original Message-----
>>>>>>>>>>>>>>>>>> From: A V Mahesh [mailto:mahesh.va...@oracle.com]
>>>>>>>>>>>>>>>>>> Sent: Wednesday, February 22, 2017 3:18 PM
>>>>>>>>>>>>>>>>>> To: Vu Minh Nguyen <vu.m.ngu...@dektech.com.au>;
>>>>>>>>>>>>>>>>>> lennart.l...@ericsson.com;
>>>> canh.v.tru...@dektech.com.au
>>>>>>>>>>>>>>>>>> Cc: opensaf-devel@lists.sourceforge.net
>>>>>>>>>>>>>>>>>> Subject: Re: [devel] [PATCH 0 of 3] Review Request for
>> log:
>>>>>>> add
>>>>>>>>>>>>>>>>> alternative
>>>>>>>>>>>>>>>>>> destinations of log records [#2258] V4
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> Hi Vu,
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> SC-2 standby osaflogd core dumped still occurs ( Not
>>>>>>>>>>>>>>>>>> resolved)
>>>>>>> ,
>>>>>>>>>> the
>>>>>>>>>>>>>>> new
>>>>>>>>>>>>>>>>>> patch only resolved the application (/usr/bin/logtest )
>>>>>>>>>> Segmentation
>>>>>>>>>>>>>>>>>> fault on SC-1 Active.
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>
>> ==============================================================
>> ==============================================================
>>>>>>>>>>>>>>>>>> # gdb /usr/lib64/opensaf/osaflogd
>>>>>>>>>>> core_1487751055.osaflogd.4594
>>>>>>>>>>>>> GNU
>>>>>>>>>>>>>>>>>> gdb
>>>>>>>>>>>>>>>>>> (GDB) SUSE (7.3-0.6.1)
>>>>>>>>>>>>>>>>>> Copyright (C) 2011 Free Software Foundation, Inc.
>>>>>>>>>>>>>>>>>> .......
>>>>>>>>>>>>>>>>>> Core was generated by `/usr/lib64/opensaf/osaflogd'.
>>>>>>>>>>>>>>>>>> Program terminated with signal 11, Segmentation fault.
>>>>>>>>>>>>>>>>>> #0  ckpt_proc_cfg_stream(lgs_cb*, void*) () at
>>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc:2195
>>>>>>>>>>>>>>>>>> 2195    src/log/logd/lgs_mbcsv.cc: No such file or
>>>>>>>>>>>>>>>>>> directory.
>>>>>>>>>>>>>>>>>>                  in src/log/logd/lgs_mbcsv.cc
>>>>>>>>>>>>>>>>>> (gdb) bt
>>>>>>>>>>>>>>>>>> #0  ckpt_proc_cfg_stream(lgs_cb*, void*) () at
>>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc:2195
>>>>>>>>>>>>>>>>>> #1  0x00007f97b026f960 in
>>>> ckpt_decode_log_struct(lgs_cb*,
>>>>>>>>>>>>>>>>>> ncs_mbcsv_cb_arg*, void*, void*, unsigned int
>>>>>>> (*)(edu_hdl_tag*,
>>>>>>>>>>>>>>>>>> edu_tkn_tag*, void*, unsigned int*, edu_buf_env_tag*,
>>>>>>>>>>>>> EDP_OP_TYPE,
>>>>>>>>>>>>>>>>>> EDU_ERR*)) () at src/log/logd/lgs_mbcsv.cc:950
>>>>>>>>>>>>>>>>>> #2  0x00007f97b02710dc in
>>>>>>> ckpt_decode_async_update(lgs_cb*,
>>>>>>>>>>>>>>>>>> ncs_mbcsv_cb_arg*) () at
>> src/log/logd/lgs_mbcsv.cc:1086
>>>>>>>>>>>>>>>>>> #3  0x00007f97b0273941 in
>>>>>>> mbcsv_callback(ncs_mbcsv_cb_arg*)
>>>>>>>>> ()
>>>>>>>>>>> at
>>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc:880
>>>>>>>>>>>>>>>>>> #4  0x00007f97af372596 in ncs_mbscv_rcv_decode ()
>> from
>>>>>>>>>>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0
>>>>>>>>>>>>>>>>>> #5  0x00007f97af372766 in
>> ncs_mbcsv_rcv_async_update
>>>> ()
>>>>>>> from
>>>>>>>>>>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0
>>>>>>>>>>>>>>>>>> #6  0x00007f97af379370 in mbcsv_process_events ()
>> from
>>>>>>>>>>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0
>>>>>>>>>>>>>>>>>> #7  0x00007f97af3794db in mbcsv_hdl_dispatch_all ()
>> from
>>>>>>>>>>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0
>>>>>>>>>>>>>>>>>> #8  0x00007f97af373ce2 in
>>>> mbcsv_process_dispatch_request ()
>>>>>>>>> at
>>>>>>>>>>>>>>>>>> src/mbc/mbcsv_api.c:423
>>>>>>>>>>>>>>>>>> #9  0x00007f97b027096e in
>> lgs_mbcsv_dispatch(unsigned
>>>> int)
>>>>>>> ()
>>>>>>>>> at
>>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc:327
>>>>>>>>>>>>>>>>>> #10 0x00007f97b024d9f2 in main () at
>>>>>>>>> src/log/logd/lgs_main.cc:583
>>>>>>>>>>>>>>>>>> (gdb) bt full
>>>>>>>>>>>>>>>>>> #0  ckpt_proc_cfg_stream(lgs_cb*, void*) () at
>>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc:2195
>>>>>>>>>>>>>>>>>>                  ckpt_data_handler = {0x7f97b0270300
>>>>>>>>>>>>>>>>>> <ckpt_proc_initialize_client(lgs_cb*, void*)>,
>>>>>>>>>>>>>>>>>> 0x7f97b02701c0
>>>>>>>>>>>>>>>>>> <ckpt_proc_finalize_client(lgs_cb*, void*)>,
>>>>>>>>>>>>>>>>>>                    0x7f97b0270060
>>>>>>>>>>>>>>>>>> <ckpt_proc_agent_down(lgs_cb*,
>>>>>>>>>> void*)>,
>>>>>>>>>>>>>>>>>> 0x7f97b02712f0 <ckpt_proc_log_write(lgs_cb*, void*)>,
>>>>>>>>>>>>> 0x7f97b0271ab0
>>>>>>>>>>>>>>>>>> <ckpt_proc_open_stream(lgs_cb*, void*)>,
>>>>>>>>>>>>>>>>>>                    0x7f97b026fe80
>>>>>> <ckpt_proc_close_stream(lgs_cb*,
>>>>>>>>>>>> void*)>,
>>>>>>>>>>>>>>>>>> 0x7f97b0272380 <ckpt_proc_cfg_stream(lgs_cb*,
>> void*)>,
>>>>>>>>>>>>>>>> 0x7f97b0274800
>>>>>>>>>>>>>>>>>> <ckpt_proc_lgs_cfg_v2(lgs_cb*, void*)>,
>>>>>>>>>>>>>>>>>>                    0x7f97b0274e10
>>>>>>>>>>>>>>>>>> <ckpt_proc_lgs_cfg_v3(lgs_cb*,
>>>>>>>>>> void*)>,
>>>>>>>>>>>>>>>>>> 0x7f97b02754f0 <ckpt_proc_lgs_cfg_v5(lgs_cb*,
>> void*)>}
>>>>>>>>>>>>>>>>>> #1  0x00007f97b026f960 in
>>>> ckpt_decode_log_struct(lgs_cb*,
>>>>>>>>>>>>>>>>>> ncs_mbcsv_cb_arg*, void*, void*, unsigned int
>>>>>>> (*)(edu_hdl_tag*,
>>>>>>>>>>>>>>>>>> edu_tkn_tag*, void*, unsigned int*, edu_buf_env_tag*,
>>>>>>>>>>>>> EDP_OP_TYPE,
>>>>>>>>>>>>>>>>>> EDU_ERR*)) () at src/log/logd/lgs_mbcsv.cc:950
>>>>>>>>>>>>>>>>>>                  ckpt_data_handler = {0x7f97b0270300
>>>>>>>>>>>>>>>>>> <ckpt_proc_initialize_client(lgs_cb*, void*)>,
>>>>>>>>>>>>>>>>>> 0x7f97b02701c0
>>>>>>>>>>>>>>>>>> <ckpt_proc_finalize_client(lgs_cb*, void*)>,
>>>>>>>>>>>>>>>>>>                    0x7f97b0270060
>>>>>>>>>>>>>>>>>> <ckpt_proc_agent_down(lgs_cb*,
>>>>>>>>>> void*)>,
>>>>>>>>>>>>>>>>>> 0x7f97b02712f0 <ckpt_proc_log_write(lgs_cb*, void*)>,
>>>>>>>>>>>>> 0x7f97b0271ab0
>>>>>>>>>>>>>>>>>> <ckpt_proc_open_stream(lgs_cb*, void*)>,
>>>>>>>>>>>>>>>>>>                    0x7f97b026fe80
>>>>>> <ckpt_proc_close_stream(lgs_cb*,
>>>>>>>>>>>> void*)>,
>>>>>>>>>>>>>>>>>> 0x7f97b0272380 <ckpt_proc_cfg_stream(lgs_cb*,
>> void*)>,
>>>>>>>>>>>>>>>> 0x7f97b0274800
>>>>>>>>>>>>>>>>>> <ckpt_proc_lgs_cfg_v2(lgs_cb*, void*)>,
>>>>>>>>>>>>>>>>>>                    0x7f97b0274e10
>>>>>>>>>>>>>>>>>> <ckpt_proc_lgs_cfg_v3(lgs_cb*,
>>>>>>>>>> void*)>,
>>>>>>>>>>>>>>>>>> 0x7f97b02754f0 <ckpt_proc_lgs_cfg_v5(lgs_cb*,
>> void*)>}
>>>>>>>>>>>>>>>>>> #2  0x00007f97b02710dc in
>>>>>>> ckpt_decode_async_update(lgs_cb*,
>>>>>>>>>>>>>>>>>> ncs_mbcsv_cb_arg*) () at
>> src/log/logd/lgs_mbcsv.cc:1086
>>>>>>>>>>>>>>>>>>                  ckpt_data_handler = {0x7f97b0270300
>>>>>>>>>>>>>>>>>> <ckpt_proc_initialize_client(lgs_cb*, void*)>,
>>>>>>>>>>>>>>>>>> 0x7f97b02701c0
>>>>>>>>>>>>>>>>>> <ckpt_proc_finalize_client(lgs_cb*, void*)>,
>>>>>>>>>>>>>>>>>>                    0x7f97b0270060
>>>>>>>>>>>>>>>>>> <ckpt_proc_agent_down(lgs_cb*,
>>>>>>>>>> void*)>,
>>>>>>>>>>>>>>>>>> 0x7f97b02712f0 <ckpt_proc_log_write(lgs_cb*, void*)>,
>>>>>>>>>>>>> 0x7f97b0271ab0
>>>>>>>>>>>>>>>>>> <ckpt_proc_open_stream(lgs_cb*, void*)>,
>>>>>>>>>>>>>>>>>>                    0x7f97b026fe80
>>>>>> <ckpt_proc_close_stream(lgs_cb*,
>>>>>>>>>>>> void*)>,
>>>>>>>>>>>>>>>>>> 0x7f97b0272380 <ckpt_proc_cfg_stream(lgs_cb*,
>> void*)>,
>>>>>>>>>>>>>>>> 0x7f97b0274800
>>>>>>>>>>>>>>>>>> <ckpt_proc_lgs_cfg_v2(lgs_cb*, void*)>,
>>>>>>>>>>>>>>>>>>                    0x7f97b0274e10
>>>>>>>>>>>>>>>>>> <ckpt_proc_lgs_cfg_v3(lgs_cb*,
>>>>>>>>>> void*)>,
>>>>>>>>>>>>>>>>>> 0x7f97b02754f0 <ckpt_proc_lgs_cfg_v5(lgs_cb*,
>> void*)>}
>>>>>>>>>>>>>>>>>> #3  0x00007f97b0273941 in
>>>>>>> mbcsv_callback(ncs_mbcsv_cb_arg*)
>>>>>>>>> ()
>>>>>>>>>>> at
>>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc:880
>>>>>>>>>>>>>>>>>>                  ckpt_data_handler = {0x7f97b0270300
>>>>>>>>>>>>>>>>>> <ckpt_proc_initialize_client(lgs_cb*, void*)>,
>>>>>>>>>>>>>>>>>> 0x7f97b02701c0
>>>>>>>>>>>>>>>>>> <ckpt_proc_finalize_client(lgs_cb*, void*)>,
>>>>>>>>>>>>>>>>>>                    0x7f97b0270060
>>>>>>>>>>>>>>>>>> <ckpt_proc_agent_down(lgs_cb*,
>>>>>>>>>> void*)>,
>>>>>>>>>>>>>>>>>> 0x7f97b02712f0 <ckpt_proc_log_write(lgs_cb*, void*)>,
>>>>>>>>>>>>> 0x7f97b0271ab0
>>>>>>>>>>>>>>>>>> <ckpt_proc_open_stream(lgs_cb*, void*)>,
>>>>>>>>>>>>>>>>>>                    0x7f97b026fe80
>>>>>> <ckpt_proc_close_stream(lgs_cb*,
>>>>>>>>>>>> void*)>,
>>>>>>>>>>>>>>>>>> 0x7f97b0272380 <ckpt_proc_cfg_stream(lgs_cb*,
>> void*)>,
>>>>>>>>>>>>>>>> 0x7f97b0274800
>>>>>>>>>>>>>>>>>> <ckpt_proc_lgs_cfg_v2(lgs_cb*, void*)>,
>>>>>>>>>>>>>>>>>>                    0x7f97b0274e10
>>>>>>>>>>>>>>>>>> <ckpt_proc_lgs_cfg_v3(lgs_cb*,
>>>>>>>>>> void*)>,
>>>>>>>>>>>>>>>>>> 0x7f97b02754f0 <ckpt_proc_lgs_cfg_v5(lgs_cb*,
>> void*)>}
>>>>>>>>>>>>>>>>>> #4  0x00007f97af372596 in ncs_mbscv_rcv_decode ()
>> from
>>>>>>>>>>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0
>>>>>>>>>>>>>>>>>> mbcsv_init_process_req_func = {0x7f97af373630
>>>>>>>>>>>>>>>>>> <mbcsv_process_initialize_request>, 0x7f97af374f10
>>>>>>>>>>>>>>>>>> <mbcsv_process_get_sel_obj_request>,
>>>>>>>>>>>>>>>>>>                    0x7f97af373be0
>>>>>> <mbcsv_process_dispatch_request>,
>>>>>>>>>>>>>>>>>> 0x7f97af373af0 <mbcsv_process_finalize_request>,
>>>>>>>>>>> 0x7f97af373cf0
>>>>>>>>>>>>>>>>>> <mbcsv_process_open_request>,
>>>>>>>>>>>>>>>>>>                    0x7f97af374050
>>>>>>>>>>>>>>>>>> <mbcsv_process_close_request>,
>>>>>>>>>>>>>>> 0x7f97af3741e0
>>>>>>>>>>>>>>>>>> <mbcsv_process_chg_role_request>, 0x7f97af3744c0
>>>>>>>>>>>>>>>>>> <mbcsv_process_snd_ckpt_request>,
>>>>>>>>>>>>>>>>>>                    0x7f97af3747d0
>>>>>> <mbcsv_process_snd_ntfy_request>,
>>>>>>>>>>>>>>>>>> 0x7f97af374970 <mbcsv_process_snd_data_req>,
>>>>>>>>> 0x7f97af373930
>>>>>>>>>>>>>>>>>> <mbcsv_process_get_request>,
>>>>>>>>>>>>>>>>>>                    0x7f97af374bd0
>>> <mbcsv_process_set_request>}
>>>>>>>>>>>>>>>>>> #5  0x00007f97af372766 in
>> ncs_mbcsv_rcv_async_update
>>>> ()
>>>>>>> from
>>>>>>>>>>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0
>>>>>>>>>>>>>>>>>> mbcsv_init_process_req_func = {0x7f97af373630
>>>>>>>>>>>>>>>>>> <mbcsv_process_initialize_request>, 0x7f97af374f10
>>>>>>>>>>>>>>>>>> <mbcsv_process_get_sel_obj_request>,
>>>>>>>>>>>>>>>>>>                    0x7f97af373be0
>>>>>> <mbcsv_process_dispatch_request>,
>>>>>>>>>>>>>>>>>> 0x7f97af373af0 <mbcsv_process_finalize_request>,
>>>>>>>>>>> 0x7f97af373cf0
>>>>>>>>>>>>>>>>>> <mbcsv_process_open_request>,
>>>>>>>>>>>>>>>>>>                    0x7f97af374050
>>>>>>>>>>>>>>>>>> <mbcsv_process_close_request>,
>>>>>>>>>>>>>>> 0x7f97af3741e0
>>>>>>>>>>>>>>>>>> <mbcsv_process_chg_role_request>, 0x7f97af3744c0
>>>>>>>>>>>>>>>>>> <mbcsv_process_snd_ckpt_request>,
>>>>>>>>>>>>>>>>>>                    0x7f97af3747d0
>>>>>> <mbcsv_process_snd_ntfy_request>,
>>>>>>>>>>>>>>>>>> 0x7f97af374970 <mbcsv_process_snd_data_req>,
>>>>>>>>> 0x7f97af373930
>>>>>>>>>>>>>>>>>> <mbcsv_process_get_request>,
>>>>>>>>>>>>>>>>>>                    0x7f97af374bd0
>>> <mbcsv_process_set_request>}
>>>>>>>>>>>>>>>>>> #6  0x00007f97af379370 in mbcsv_process_events ()
>> from
>>>>>>>>>>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0
>>>>>>>>>>>>>>>>>> No symbol table info available.
>>>>>>>>>>>>>>>>>> #7  0x00007f97af3794db in mbcsv_hdl_dispatch_all ()
>> from
>>>>>>>>>>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0
>>>>>>>>>>>>>>>>>> No symbol table info available.
>>>>>>>>>>>>>>>>>> #8  0x00007f97af373ce2 in
>>>> mbcsv_process_dispatch_request ()
>>>>>>>>> at
>>>>>>>>>>>>>>>>>> src/mbc/mbcsv_api.c:423
>>>>>>>>>>>>>>>>>> mbcsv_init_process_req_func = {0x7f97af373630
>>>>>>>>>>>>>>>>>> <mbcsv_process_initialize_request>, 0x7f97af374f10
>>>>>>>>>>>>>>>>>> <mbcsv_process_get_sel_obj_request>,
>>>>>>>>>>>>>>>>>>                    0x7f97af373be0
>>>>>> <mbcsv_process_dispatch_request>,
>>>>>>>>>>>>>>>>>> 0x7f97af373af0 <mbcsv_process_finalize_request>,
>>>>>>>>>>> 0x7f97af373cf0
>>>>>>>>>>>>>>>>>> <mbcsv_process_open_request>,
>>>>>>>>>>>>>>>>>>                    0x7f97af374050
>>>>>>>>>>>>>>>>>> <mbcsv_process_close_request>,
>>>>>>>>>>>>>>> 0x7f97af3741e0
>>>>>>>>>>>>>>>>>> <mbcsv_process_chg_role_request>, 0x7f97af3744c0
>>>>>>>>>>>>>>>>>> <mbcsv_process_snd_ckpt_request>,
>>>>>>>>>>>>>>>>>>                    0x7f97af3747d0
>>>>>> <mbcsv_process_snd_ntfy_request>,
>>>>>>>>>>>>>>>>>> 0x7f97af374970 <mbcsv_process_snd_data_req>,
>>>>>>>>> 0x7f97af373930
>>>>>>>>>>>>>>>>>> <mbcsv_process_get_request>,
>>>>>>>>>>>>>>>>>>                    0x7f97af374bd0
>>> <mbcsv_process_set_request>}
>>>>>>>>>>>>>>>>>> #9  0x00007f97b027096e in
>> lgs_mbcsv_dispatch(unsigned
>>>> int)
>>>>>>> ()
>>>>>>>>> at
>>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc:327
>>>>>>>>>>>>>>>>>>                  ckpt_data_handler = {0x7f97b0270300
>>>>>>>>>>>>>>>>>> <ckpt_proc_initialize_client(lgs_cb*, void*)>,
>>>>>>>>>>>>>>>>>> 0x7f97b02701c0
>>>>>>>>>>>>>>>>>> <ckpt_proc_finalize_client(lgs_cb*, void*)>,
>>>>>>>>>>>>>>>>>>                    0x7f97b0270060
>>>>>>>>>>>>>>>>>> <ckpt_proc_agent_down(lgs_cb*,
>>>>>>>>>> void*)>,
>>>>>>>>>>>>>>>>>> 0x7f97b02712f0 <ckpt_proc_log_write(lgs_cb*, void*)>,
>>>>>>>>>>>>> 0x7f97b0271ab0
>>>>>>>>>>>>>>>>>> <ckpt_proc_open_stream(lgs_cb*, void*)>,
>>>>>>>>>>>>>>>>>>                    0x7f97b026fe80
>>>>>> <ckpt_proc_close_stream(lgs_cb*,
>>>>>>>>>>>> void*)>,
>>>>>>>>>>>>>>>>>> 0x7f97b0272380 <ckpt_proc_cfg_stream(lgs_cb*,
>> void*)>,
>>>>>>>>>>>>>>>> 0x7f97b0274800
>>>>>>>>>>>>>>>>>> <ckpt_proc_lgs_cfg_v2(lgs_cb*, void*)>,
>>>>>>>>>>>>>>>>>>                    0x7f97b0274e10
>>>>>>>>>>>>>>>>>> <ckpt_proc_lgs_cfg_v3(lgs_cb*,
>>>>>>>>>> void*)>,
>>>>>>>>>>>>>>>>>> 0x7f97b02754f0 <ckpt_proc_lgs_cfg_v5(lgs_cb*,
>> void*)>}
>>>>>>>>>>>>>>>>>> ---Type <return> to continue, or q <return> to quit---
>>>>>>>>>>>>>>>>>> #10 0x00007f97b024d9f2 in main () at
>>>>>>>>> src/log/logd/lgs_main.cc:583
>>>>>>>>>>>>>>>>>> usr1_sel_obj = {raise_obj = -1, rmv_obj = -1}
>>>>>>>>>>>>>>>>>>                  _lgs_cb = {mds_hdl = 65547, mds_role =
>>>>>>>>>>> V_DEST_RL_STANDBY,
>>>>>>>>>>>>>>>> vaddr
>>>>>>>>>>>>>>>>>> = 11, log_version = {releaseCode = 65 'A', majorVersion
>> = 2
>>>>>>>>>> '\002',
>>>>>>>>>>>>>>>>>> minorVersion = 2 '\002'}, client_tree = {
>>>>>>>>>>>>>>>>>>                      root_node = {bit = -1, left =
>>>>>> 0x7f97b04cf1b0,
>>>>>>>>>> right =
>>>>>>>>>>>>>>>>>> 0x7f97b04a2418, key_info = 0x7f97b04b7bd0 ""},
>> params =
>>>>>>>>>>> {key_size
>>>>>>>>>>>>>>>>>> = 4},
>>>>>>>>>>>>>>>>>> n_nodes = 8}, comp_name = {_opaque = {46, 24947,
>>>>>>>>>>>>>>>>>>                        17254, 28015, 15728, 20300, 11335,
>>>>>>>>>>>>>>>>>> 24947,
>>>>>>>> 21350,
>>>>>>>>>>>>>>>>>> 15733,
>>>>>>>>>>>>>>>>>> 17235, 12845, 29484, 26209, 26451, 12861, 11342,
>> 24947,
>>>>>>>>> 16742,
>>>>>>>>>>>>> 28784,
>>>>>>>>>>>>>>>>>> 20285, 25968, 21358, 17985,
>>>>>>>>>>>>>>>>>>                        0 <repeats 105 times>}}, amf_hdl =
>>>>>>>> 4288675841,
>>>>>>>>>>>>>>>>>> amfSelectionObject = 15, amf_invocation_id = 0,
>>>>>>> is_quiesced_set
>>>>>>>>> =
>>>>>>>>>>>>>>> false,
>>>>>>>>>>>>>>>>>> immOiHandle = 554050912783, immSelectionObject =
>> 21,
>>>>>>>>>>>>>>>>>>                    clmSelectionObject = 17, clm_hdl =
>>>>>>>>>>>>>>>>>> 4279238657,
>>>>>>>>>>>>>>>>>> ha_state =
>>>>>>>>>>>>>>>>>> SA_AMF_HA_STANDBY, last_client_id = 208,
>>>> async_upd_cnt =
>>>>>>>>> 743,
>>>>>>>>>>>>>>>>>> ckpt_state
>>>>>>>>>>>>>>>>>> = COLD_SYNC_IDLE, mbcsv_hdl = 4293918753,
>>>>>>>>>>>>>>>>>>                    mbcsv_sel_obj = 23, mbcsv_ckpt_hdl =
>>>>>> 4292870177,
>>>>>>>>>>>>>>>>>> mbcsv_peer_version = 7, edu_hdl = {is_inited = true, tree
>> =
>>>>>>>>>>>>>>>>>> {root_node
>>>>>>>>>>>>>>> =
>>>>>>>>>>>>>>>>>> {bit = -1, left = 0x7f97b04cf2e0,
>>>>>>>>>>>>>>>>>>                          right = 0x7f97b04a25b8, key_info
> =
>>>>>>>>>> 0x7f97b04b7d40
>>>>>>>>>>>>>>> ""},
>>>>>>>>>>>>>>>>>> params = {key_size = 8}, n_nodes = 12}, to_version = 1},
>>>>>>>>>>>>>>>>>> fully_initialized = true, lga_down_list_head = 0x0,
>>>>>>>>>>>>>>>>>>                    lga_down_list_tail = 0x0,
>>>>>>>>>>>>>>>>>> clm_init_sel_obj =
>>>>>>>>>>>>>>>>>> {raise_obj =
>>>>>>>>>>>>>>> -1,
>>>>>>>>>>>>>>>>>> rmv_obj = -1}, nid_started = true, scAbsenceAllowed =
>> 900,
>>>>>>>>>>>>>>>>>> lgs_recovery_state = LGS_NORMAL}
>>>>>>>>>>>>>>>>>>                  nfds = 7
>>>>>>>>>>>>>>>>>>                  fds = {{fd = 19, events = 1, revents =
> 0},
>>>>>>>>>>>>>>>>>> {fd =
>>>>>>>> 15,
>>>>>>>>>>>>>>>>>> events =
>>>>>>>>>>>>>>>>>> 1, revents = 0}, {fd = 23, events = 1, revents = 1}, {fd
>>>>>>>>>>>>>>>>>> = 13,
>>>>>>>>>>>>>>>>>> events =
>>>>>>>>>>>>>>>>>> 1, revents = 0}, {fd = -1, events = 1,
>>>>>>>>>>>>>>>>>>                      revents = 0}, {fd = 17, events = 1,
>>>>>>>>>>>>>>>>>> revents
>>>>>> =
>>>>>>>> 0},
>>>>>>>>>>>>>>>>>> {fd =
>>>>>>>>>>>>>>> 21,
>>>>>>>>>>>>>>>>>> events = 1, revents = 0}}
>>>>>>>>>>>>>>>>>>                  mbox_msgs = {0, 0, 0, 0, 0}
>>>>>>>>>>>>>>>>>>                  lgs_cb = 0x7f97b04a2400
>>>>>>>>>>>>>>>>>>                  mbox_low = {0, 0, 0, 0, 0}
>>>>>>>>>>>>>>>>>>                  lgs_mbox_init_mutex = {__data = {__lock =
> 0,
>>>>>>>> __count =
>>>>>>>>>> 0,
>>>>>>>>>>>>>>>>>> __owner = 0, __nusers = 0, __kind = 0, __spins = 0,
> __list
>>> =
>>>>>>>>>>>>>>>>>> {__prev =
>>>>>>>>>>>>>>>>>> 0x0, __next = 0x0}},
>>>>>>>>>>>>>>>>>>                    __size = '\000' <repeats 39 times>,
>>>>>>>>>>>>>>>>>> __align =
>>>>>> 0}
>>>>>>>>>>>>>>>>>> lgs_mbx = 4291821569
>>>>>>>>>>>>>>>>>>                  mbox_high = {0, 0, 0, 0, 0}
>>>>>>>>>>>>>>>>>>                  mbox_full = {false, false, false, false,
>>>>>>>>>>>>>>>>>> false}
>>>>>>>>>>>>>>>>>> (gdb)
>>>>>>>>>>>>>>>>>> (gdb)
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> Feb 22 13:40:55 SC-2 osafimmnd[4584]: NO Ccb 131
>>>>>>>>> COMMITTED
>>>>>>>>>>>>>>>>>> (immcfg_SC-1_18714)
>>>>>>>>>>>>>>>>>> Feb 22 13:40:56 SC-2 osafamfnd[4634]: NO
>>>>>>>>>>>>>>>>>> 'safComp=LOG,safSu=SC-2,safSg=2N,safApp=OpenSAF'
>>>> faulted
>>>>>>>>> due
>>>>>>>>>>> to
>>>>>>>>>>>>>>>>>> 'avaDown' : Recovery is 'nodeFailfast'
>>>>>>>>>>>>>>>>>> Feb 22 13:40:56 SC-2 osafamfnd[4634]: ER
>>>>>>>>>>>>>>>>>> safComp=LOG,safSu=SC-2,safSg=2N,safApp=OpenSAF
>>>> Faulted
>>>>>>>>> due
>>>>>>>>>>>>>>>>>> to:avaDown
>>>>>>>>>>>>>>>>>> Recovery is:nodeFailfast
>>>>>>>>>>>>>>>>>> Feb 22 13:40:56 SC-2 osafamfnd[4634]: Rebooting
>>>> OpenSAF
>>>>>>>>> NodeId
>>>>>>>>>>> =
>>>>>>>>>>>>>>>>>> 131599
>>>>>>>>>>>>>>>>>> EE Name = , Reason: Component faulted: recovery is
>> node
>>>>>>>>> failfast,
>>>>>>>>>>>>>>>>>> OwnNodeId = 131599, SupervisionTime = 60
>>>>>>>>>>>>>>>>>> Feb 22 13:40:56 SC-2 opensaf_reboot: Rebooting local
>>>> node;
>>>>>>>>>>>>> timeout=60
>> ==============================================================
>> ==============================================================
>>>>>>>>>>>>>>>>>> On 2/22/2017 12:23 PM, A V Mahesh wrote:
>>>>>>>>>>>>>>>>>>> Hi Vu,
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> On 2/22/2017 12:19 PM, Vu Minh Nguyen wrote:
>>>>>>>>>>>>>>>>>>>> [Vu] I has sent you 02 patches. There is code change
>> in
>>>>>>>>> osaflogd
>>>>>>>>>>>>>>>>>>>> code
>>>>>>>>>>>>>>>>> that
>>>>>>>>>>>>>>>>>>>> fix the coredump you have observed.
>>>>>>>>>>>>>>>>>>>> The other one is test code that fix the logtest
>> coredump.
>>>>>>>>>>>>>>>>>>> Ok I will re-test , and update you .
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> -AVM
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> On 2/22/2017 12:19 PM, Vu Minh Nguyen wrote:
>>>>>>>>>>>>>>>>>>>> Hi Mahehs,
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>> See my reply inline, [Vu].
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>> Regards, Vu
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>> -----Original Message-----
>>>>>>>>>>>>>>>>>>>>> From: A V Mahesh
>> [mailto:mahesh.va...@oracle.com]
>>>>>>>>>>>>>>>>>>>>> Sent: Wednesday, February 22, 2017 1:36 PM
>>>>>>>>>>>>>>>>>>>>> To: Vu Minh Nguyen
>> <vu.m.ngu...@dektech.com.au>;
>>>>>>>>>>>>>>>>>>>>> lennart.l...@ericsson.com;
>>>>>>> canh.v.tru...@dektech.com.au
>>>>>>>>>>>>>>>>>>>>> Cc: opensaf-devel@lists.sourceforge.net
>>>>>>>>>>>>>>>>>>>>> Subject: Re: [PATCH 0 of 3] Review Request for log:
>> add
>>>>>>>>>>>>>>>>>>>>> alternative
>>>>>>>>>>>>>>>>>>>>> destinations of log records [#2258] V4
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>> Hi Vu,
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>> On 2/22/2017 11:52 AM, Vu Minh Nguyen wrote:
>>>>>>>>>>>>>>>>>>>>>> Hi Mahesh,
>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>> Have a code fault in uml test, and other one in
>>>>>>> checkpoint.
>>>>>>>>>>>>>>>>>>>>> [AVM] This is Normal Suse 11 VM ( not  UML).
>>>>>>>>>>>>>>>>>>>>>> I have just updated the code. Please re-apply for #3
>>>> and
>>>>>>> #4
>>>>>>>>>>>>>>> patches.
>>>>>>>>>>>>>>>>>>>>> [AVM] is these new patch has function changes or
>> only
>>>>>>>>>>>>>>>>>>>>> test
>>>>>>>>>>> code
>>>>>>>>>>>>>>>>>> changes ?
>>>>>>>>>>>>>>>>>>>> [Vu] I has sent you 02 patches. There is code change
>> in
>>>>>>>>> osaflogd
>>>>>>>>>>>>>>>>>>>> code
>>>>>>>>>>>>>>>>> that
>>>>>>>>>>>>>>>>>>>> fix the coredump you have observed.
>>>>>>>>>>>>>>>>>>>> The other one is test code that fix the logtest
>> coredump.
>>>>>>>>>>>>>>>>>>>>>> Note that, test case #14 of suite 17 should be run
>> on
>>>>>>> active
>>>>>>>>>>>>>>>>>>>>>> node,
>>>>>>>>>>>>>>>>>>>>> otherwise
>>>>>>>>>>>>>>>>>>>>>> getting failed.
>>>>>>>>>>>>>>>>>>>>> [AVM]  Segmentation fault of /usr/bin/logtest Not a
>> big
>>>>>>>> issue
>>>>>>>>>> ,
>>>>>>>>>>>>>>>>>>>>>             we need to debug why  osaflogd core dumped
>>>>>>>>>>>>>>>>>>>>> and it
>>>>>> is
>>>>>>>>>>>>>>>>>>>>> critical
>>>>>>>>>>>>>>>>>>>> [Vu] I found the problem. You can try with the new
>> one
>>>> to
>>>>>>> see
>>>>>>>>> if
>>>>>>>>>>>>>>>>>>>> the
>>>>>>>>>>>>>>>>>>>> coredump is still there or not.
>>>>>>>>>>>>>>>>>>>>>> I will put condition check to that test case later.
>>>>>>>>>>>>>>>>>>>>> -AVM
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>> Regards, Vu
>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> -----Original Message-----
>>>>>>>>>>>>>>>>>>>>>>> From: A V Mahesh
>>>> [mailto:mahesh.va...@oracle.com]
>>>>>>>>>>>>>>>>>>>>>>> Sent: Wednesday, February 22, 2017 12:16 PM
>>>>>>>>>>>>>>>>>>>>>>> To: Vu Minh Nguyen
>>>> <vu.m.ngu...@dektech.com.au>;
>>>>>>>>>>>>>>>>>>>>>>> lennart.l...@ericsson.com;
>>>>>>>>> canh.v.tru...@dektech.com.au
>>>>>>>>>>>>>>>>>>>>>>> Cc: opensaf-devel@lists.sourceforge.net
>>>>>>>>>>>>>>>>>>>>>>> Subject: Re: [PATCH 0 of 3] Review Request for
>> log:
>>>> add
>>>>>>>>>>>>>>> alternative
>>>>>>>>>>>>>>>>>>>>>>> destinations of log records [#2258] V4
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> Hi Vu,
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> Thanks ,
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> While testing /usr/bin/logtest ,  SC-2 standby
>>>> osaflogd
>>>>>>>>> core
>>>>>>>>>>>>>>>> dumped
>>>>>>>>>>>>>>>>>>>> and
>>>>>>>>>>>>>>>>>>>>>>> /usr/bin/logtest on SC-1 Active
>>>>>>>>>>>>>>>>>>>>>>> got Segmentation fault , am I missing any other
>>>> patch (
>>>>>> i
>>>>>>>>> am
>>>>>>>>>>>>>>> using
>>>>>>>>>>>>>>>>>>>>>>> devel published patch only )
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> Following patches i am using :
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>              1) #2293 (sent by Anders Widel, but not
> yet
>>>>>>>> pushed)
>>>>>>>>>>>>>>>>>>>>>>> 2) #2258 (v2, sent by Lennart, but not yet
>>>>>>>> pushed
>>>>>>>>>> yet)
>>>>>>>>>>>>>>>>>>>>>>> 3) #2258 (v4, sent by Vu, but not yet pushed
>>>>>>>> yet)
>> ==============================================================
>>>>>>>>>>>>>>>>>>>>>>> ========================================
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> Core was generated by
>>>> `/usr/lib64/opensaf/osaflogd'.
>>>>>>>>>>>>>>>>>>>>>>> Program terminated with signal 11, Segmentation
>>>> fault.
>>>>>>>>>>>>>>>>>>>>>>> #0 ckpt_proc_cfg_stream(lgs_cb*, void*) () at
>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc:2195
>>>>>>>>>>>>>>>>>>>>>>> 2195 src/log/logd/lgs_mbcsv.cc: No such file or
>>>>>>>>> directory.
>>>>>>>>>>>>>>>>>>>>>>> in src/log/logd/lgs_mbcsv.cc
>>>>>>>>>>>>>>>>>>>>>>> (gdb) bt
>>>>>>>>>>>>>>>>>>>>>>> #0 ckpt_proc_cfg_stream(lgs_cb*, void*) () at
>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc:2195
>>>>>>>>>>>>>>>>>>>>>>> #1  0x00007f12c3e22960 in
>>>>>>>>> ckpt_decode_log_struct(lgs_cb*,
>>>>>>>>>>>>>>>>>>>>>>> ncs_mbcsv_cb_arg*, void*, void*, unsigned int
>>>>>>>>>>>>> (*)(edu_hdl_tag*,
>>>>>>>>>>>>>>>>>>>>>>> edu_tkn_tag*, void*, unsigned int*,
>>>> edu_buf_env_tag*,
>>>>>>>>>>>>>>>>>> EDP_OP_TYPE,
>>>>>>>>>>>>>>>>>>>>>>> EDU_ERR*)) () at src/log/logd/lgs_mbcsv.cc:950
>>>>>>>>>>>>>>>>>>>>>>> #2  0x00007f12c3e240dc in
>>>>>>>>>>>>> ckpt_decode_async_update(lgs_cb*,
>>>>>>>>>>>>>>>>>>>>>>> ncs_mbcsv_cb_arg*) () at
>>>>>>> src/log/logd/lgs_mbcsv.cc:1086
>>>>>>>>>>>>>>>>>>>>>>> #3 0x00007f12c3e26941 in
>>>>>>>>>>>>> mbcsv_callback(ncs_mbcsv_cb_arg*) ()
>>>>>>>>>>>>>>>> at
>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc:880
>>>>>>>>>>>>>>>>>>>>>>> #4  0x00007f12c2f25596 in
>> ncs_mbscv_rcv_decode ()
>>>>>>> from
>>>>>>>>>>>>>>>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0
>>>>>>>>>>>>>>>>>>>>>>> #5  0x00007f12c2f25766 in
>>>>>>> ncs_mbcsv_rcv_async_update
>>>>>>>>> ()
>>>>>>>>>>>>> from
>>>>>>>>>>>>>>>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0
>>>>>>>>>>>>>>>>>>>>>>> #6  0x00007f12c2f2c370 in mbcsv_process_events
>> ()
>>>>>>> from
>>>>>>>>>>>>>>>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0
>>>>>>>>>>>>>>>>>>>>>>> #7  0x00007f12c2f2c4db in
>> mbcsv_hdl_dispatch_all ()
>>>>>>> from
>>>>>>>>>>>>>>>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0
>>>>>>>>>>>>>>>>>>>>>>> #8  0x00007f12c2f26ce2 in
>>>>>>>>> mbcsv_process_dispatch_request
>>>>>>>>>>> ()
>>>>>>>>>>>>> at
>>>>>>>>>>>>>>>>>>>>>>> src/mbc/mbcsv_api.c:423
>>>>>>>>>>>>>>>>>>>>>>> #9  0x00007f12c3e2396e in
>>>>>>> lgs_mbcsv_dispatch(unsigned
>>>>>>>>> int)
>>>>>>>>>>> ()
>>>>>>>>>>>>> at
>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc:327
>>>>>>>>>>>>>>>>>>>>>>> #10 0x00007f12c3e009f2 in main () at
>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_main.cc:583
>>>>>>>>>>>>>>>>>>>>>>> (gdb)
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>
>> ==============================================================
>>>>>>>>>>>>>>>>>>>>>>> ========================================
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> Feb 22 10:37:06 SC-1 osafimmnd[4020]: NO
>> Invalid
>>>> error
>>>>>>>>>>>>> reported
>>>>>>>>>>>>>>>>>>>>>>> implementer 'safLogService', Ccb 161 will be
>> aborted
>>>>>>>>>>>>>>>>>>>>>>> Feb 22 10:37:06 SC-1 osafimmnd[4020]: NO Ccb
>> 161
>>>>>>>>> aborted
>>>>>>>>>>> in
>>>>>>>>>>>>>>>>>>>>> COMPLETED
>>>>>>>>>>>>>>>>>>>>>>> processing (validation)
>>>>>>>>>>>>>>>>>>>>>>> Feb 22 10:37:06 SC-1 osafimmnd[4020]: NO Ccb
>> 161
>>>>>>>>>>> ABORTED
>>>>>>>>>>>>>>>>>>>>> (immcfg_SC-
>>>>>>>>>>>>>>>>>>>>>>> 1_5394)
>>>>>>>>>>>>>>>>>>>>>>> Add values Fail
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> Program received signal SIGSEGV, Segmentation
>>>> fault.
>>>>>>>>>>>>>>>>>>>>>>> 0x000055555556929a in read_and_compare.isra.7
>> ()
>>>> at
>>>>>>>>>>>>>>>>>>>>>>> src/log/apitest/tet_LogOiOps.c:1891
>>>>>>>>>>>>>>>>>>>>>>> 1891 src/log/apitest/tet_LogOiOps.c: No such file
>> or
>>>>>>>>>>>>>>>>>>>>>>> directory.
>>>>>>>>>>>>>>>>>>>>>>>                     in src/log/apitest/tet_LogOiOps.c
>>>>>>>>>>>>>>>>>>>>>>> (gdb) Feb 22 10:37:07 SC-1 sshd[5298]: Accepted
>>>>>>> keyboard-
>>>>>>>>>>>>>>>>>>>>> interactive/pam
>>>>>>>>>>>>>>>>>>>>>>> for root from 10.176.178.22 port 51945 ssh2
>>>>>>>>>>>>>>>>>>>>>>> bt
>>>>>>>>>>>>>>>>>>>>>>> #0  0x000055555556929a in
>>>> read_and_compare.isra.7 ()
>>>>>>> at
>>>>>>>>>>>>>>>>>>>>>>> src/log/apitest/tet_LogOiOps.c:1891
>>>>>>>>>>>>>>>>>>>>>>> #1  0x0000555555569bbb in
>>>>>>>>>>>>>>>>>>>>>>> check_logRecordDestinationConfigurationEmpty
>>>>>>>>>>>>>>>>>>>>>>> () at src/log/apitest/tet_LogOiOps.c:2179
>>>>>>>>>>>>>>>>>>>>>>> #2  0x0000555555573495 in run_test_case ()
>>>>>>>>>>>>>>>>>>>>>>> #3  0x0000555555573934 in test_run ()
>>>>>>>>>>>>>>>>>>>>>>> #4  0x000055555555c7cd in main () at
>>>>>>>>>>>>>>>>>>>>>>> src/log/apitest/logtest.c:569
>>>>>>>>>>>>>>>>>>>>>>> (gdb)
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>
>> ==============================================================
>>>>>>>>>>>>>>>>>>>>>>> ========================================
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> -AVM
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> On 2/22/2017 9:48 AM, Vu Minh Nguyen wrote:
>>>>>>>>>>>>>>>>>>>>>>>> Hi Mahesh,
>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>> I send them in attachment instead, and name
>> them
>>>> in
>>>>>>> the
>>>>>>>>>>>>> order.
>>>>>>>>>>>>>>>>>>>>>>>> I just pull the latest code, and apply them without
>>>>>>>> getting
>>>>>>>>>>>> any
>>>>>>>>>>>>>>>>> hunk
>>>>>>>>>>>>>>>>>>>>>> error.
>>>>>>>>>>>>>>>>>>>>>>>> Please try with them, and let me know if you see
>>>> any
>>>>>>>>>>> problem.
>>>>>>>>>>>>>>>>>>>>>>>> Regards, Vu
>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>> -----Original Message-----
>>>>>>>>>>>>>>>>>>>>>>>>> From: A V Mahesh
>>>> [mailto:mahesh.va...@oracle.com]
>>>>>>>>>>>>>>>>>>>>>>>>> Sent: Wednesday, February 22, 2017 11:09 AM
>>>>>>>>>>>>>>>>>>>>>>>>> To: Vu Minh Nguyen
>>>>>>> <vu.m.ngu...@dektech.com.au>;
>>>>>>>>>>>>>>>>>>>>>>>>> lennart.l...@ericsson.com;
>>>>>>>>>>> canh.v.tru...@dektech.com.au
>>>>>>>>>>>>>>>>>>>>>>>>> Cc: opensaf-devel@lists.sourceforge.net
>>>>>>>>>>>>>>>>>>>>>>>>> Subject: Re: [PATCH 0 of 3] Review Request for
>>>> log:
>>>>>>> add
>>>>>>>>>>>>>>>>> alternative
>>>>>>>>>>>>>>>>>>>>>>>>> destinations of log records [#2258] V4
>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>> Hi Vu,
>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>> I did follow that still i get Hunk #2 FAILED even
>> on
>>>>>>>>>> today's
>>>>>>>>>>>>>>>>> staging
>> ==============================================================
>>>>>>>>>>>>>>>>>>>>>>>>> ==================
>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>> [root@dhcp-hyd-scp-5fl-10-176-177-96
>> staging]#
>>>>>>> patch
>>>>>>>>> -
>>>>>>>>>>> p1
>>>>>>>>>>>>>>>> <2293
>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/base/Makefile.am
>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #1 succeeded at 33 (offset 1 line).
>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #3 succeeded at 183 (offset 1 line).
>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/base/file_descriptor.cc
>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/base/file_descriptor.h
>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/base/tests/unix_socket_test.cc
>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/base/unix_client_socket.cc
>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/base/unix_server_socket.cc
>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/base/unix_socket.cc
>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/base/unix_socket.h
>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>> [root@dhcp-hyd-scp-5fl-10-176-177-96
>> staging]#
>>>>>>> patch
>>>>>>>>> -
>>>>>>>>>>> p1
>>>>>>>>>>>>>>>> <2258-
>>>>>>>>>>>>>>>>>> 1
>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/log/Makefile.am
>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #1 succeeded at 71 (offset -1 lines).
>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/log/config/logsv_classes.xml
>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #1 FAILED at 147.
>>>>>>>>>>>>>>>>>>>>>>>>> 1 out of 1 hunk FAILED -- saving rejects to file
>>>>>>>>>>>>>>>>>>>>>>>>> src/log/config/logsv_classes.xml.rej
>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_config.cc
>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #1 succeeded at 35 (offset -5 lines).
>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #2 FAILED at 705.
>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #3 FAILED at 971.
>>>>>>>>>>>>>>>>>>>>>>>>> 2 out of 3 hunks FAILED -- saving rejects to file
>>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_config.cc.rej
>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_config.h
>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #1 FAILED at 304.
>>>>>>>>>>>>>>>>>>>>>>>>> 1 out of 1 hunk FAILED -- saving rejects to file
>>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_config.h.rej
>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_dest.cc
>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_dest.h
>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_evt.cc
>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_imm.cc
>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #1 FAILED at 45.
>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #2 succeeded at 235 (offset -1 lines).
>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #3 FAILED at 877.
>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #4 succeeded at 1273 (offset -20 lines).
>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #5 succeeded at 1404 (offset -1 lines).
>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #6 succeeded at 1449 (offset -20 lines).
>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #7 succeeded at 2032 (offset -1 lines).
>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #8 FAILED at 2181.
>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #9 succeeded at 2271 (offset -54 lines).
>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #10 succeeded at 2387 (offset -1 lines).
>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #11 succeeded at 2377 (offset -54 lines).
>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #12 succeeded at 2478 (offset -1 lines).
>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #13 succeeded at 2684 (offset -54 lines).
>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #14 succeeded at 2821 (offset -1 lines).
>>>>>>>>>>>>>>>>>>>>>>>>> 3 out of 14 hunks FAILED -- saving rejects to file
>>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_imm.cc.rej
>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_main.cc
>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_mbcsv.cc
>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_mbcsv.h
>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_mbcsv_v5.cc
>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #3 succeeded at 133 (offset -1 lines).
>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_mbcsv_v7.cc
>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_mbcsv_v7.h
>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_stream.cc
>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_stream.h
>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_util.cc
>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_util.h
>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>> [root@dhcp-hyd-scp-5fl-10-176-177-96
>> staging]#
>>>>>>> patch
>>>>>>>>> -
>>>>>>>>>>> p1
>>>>>>>>>>>>>>>> <2258-
>>>>>>>>>>>>>>>>>> 2
>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/log/Makefile.am
>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #1 succeeded at 180 (offset -3 lines).
>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/log/apitest/tet_LogOiOps.c
>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #1 FAILED at 1923.
>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #2 FAILED at 1979.
>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #3 FAILED at 2067.
>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #4 FAILED at 2094.
>>>>>>>>>>>>>>>>>>>>>>>>> 4 out of 4 hunks FAILED -- saving rejects to file
>>>>>>>>>>>>>>>>>>>>>>>>> src/log/apitest/tet_LogOiOps.c.rej
>>>>>>>>>>>>>>>>>>>>>>>>> patching file
>> src/log/apitest/tet_cfg_destination.c
>>>>>>>>>>>>>>>>>>>>>>>>> [root@dhcp-hyd-scp-5fl-10-176-177-96
>> staging]#
>>>>>>> patch
>>>>>>>>> -
>>>>>>>>>>> p1
>>>>>>>>>>>>>>>> <2258-
>>>>>>>>>>>>>>>>>> 3
>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/log/Makefile
>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/log/Makefile.am
>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #1 succeeded at 80 (offset -1 lines).
>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #2 succeeded at 217 (offset -2 lines).
>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/log/tests/Makefile
>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/log/tests/lgs_dest_test.cc
>>>>>>>>>>>>>>>>>>>>>>>>> [root@dhcp-hyd-scp-5fl-10-176-177-96
>> staging]#
>>>> vi
>>>>>>>>>>>>>>>>>>>>>>>>> src/log/apitest/tet_LogOiOps.c.rej
>>>>>>>>>>>>>>>>>>>>>>>>> [root@dhcp-hyd-scp-5fl-10-176-177-96
>> staging]#
>>>>>>>>>>>>>>>>>>>>>>>>>
>> ==============================================================
>>>>>>>>>>>>>>>>>>>>>>>>> ========================
>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>> -AVM
>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>> On 2/21/2017 3:53 PM, Vu Minh Nguyen wrote:
>>>>>>>>>>>>>>>>>>>>>>>>>> Hi Mahesh,
>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>> As I has mentioned in below:
>>>>>>>>>>>>>>>>>>>>>>>>>>>> To run the test, this patch has
>>>>>> dependent
>>>>>>>> on
>>>>>>>>>>>>>>> following
>>>>>>>>>>>>>>>>>>>> patches:
>>>>>>>>>>>>>>>>>>>>>>>>>>>> 1) #2293 (sent by Anders Widel, but not
>>>>>> yet
>>>>>>>>>>>>>>>>>>>>>>>>>>>> pushed)
>>>>>>>>>>>>>>>>>>>>>>>>>>>>                2) #2258 (v2, sent by Lennart,
> but
>>>>>>>>>>>>>>>>>>>>>>>>>>>> not
>>>>>> yet
>>>>>>>>>> pushed
>>>>>>>>>>>>>>> yet)
>>>>>>>>>>>>>>>>>>>>>>>>>> So, you need to apply #2293 first, then #2258
>>>> which
>>>>>>>>> sent
>>>>>>>>>>> by
>>>>>>>>>>>>>>>>>> Lennart
>>>>>>>>>>>>>>>>>>>>>>>>>> yesterday, then mine.
>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>> Regards, Vu
>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>> -----Original Message-----
>>>>>>>>>>>>>>>>>>>>>>>>>>> From: A V Mahesh
>>>>>>> [mailto:mahesh.va...@oracle.com]
>>>>>>>>>>>>>>>>>>>>>>>>>>> Sent: Tuesday, February 21, 2017 5:10 PM
>>>>>>>>>>>>>>>>>>>>>>>>>>> To: Vu Minh Nguyen
>>>>>>>>> <vu.m.ngu...@dektech.com.au>;
>>>>>>>>>>>>>>>>>>>>>>>>>>> lennart.l...@ericsson.com;
>>>>>>>>>>>>> canh.v.tru...@dektech.com.au
>>>>>>>>>>>>>>>>>>>>>>>>>>> Cc: opensaf-devel@lists.sourceforge.net
>>>>>>>>>>>>>>>>>>>>>>>>>>> Subject: Re: [PATCH 0 of 3] Review Request
>> for
>>>> log:
>>>>>>>>> add
>>>>>>>>>>>>>>>>>> alternative
>>>>>>>>>>>>>>>>>>>>>>>>>>> destinations of log records [#2258] V4
>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>> Hi Vu,
>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>> Is this applies on top of   log #2146  - V4 , I
>>> see
>>>>>>>>>> both
>>>>>>>>>>>>>>>>> #tickets
>>>>>>>>>>>>>>>>>>>>>>>> has
>>>>>>>>>>>>>>>>>>>>>>>>>>> version changes ?
>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>> in which order i need to apply  ( #2146 &
>> #2258 )
>>>>>>> or
>>>>>>>>>>>>>>>>>>>>>>>>>>> (#2258
>>>>>>>>>>>>>>> &
>>>>>>>>>>>>>>>>>>>>>>>> #2146).
>>>>>>> =========================================================
>>>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/log/Makefile.am
>>>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #1 FAILED at 72.
>>>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #2 FAILED at 120.
>>>>>>>>>>>>>>>>>>>>>>>>>>> 2 out of 2 hunks FAILED -- saving rejects to
>> file
>>>>>>>>>>>>>>>>>>>>>>>> src/log/Makefile.am.rej
>>>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/log/config/logsv_classes.xml
>>>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #1 FAILED at 147.
>>>>>>>>>>>>>>>>>>>>>>>>>>> 1 out of 1 hunk FAILED -- saving rejects to file
>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/config/logsv_classes.xml.rej
>>>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_config.cc
>>>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #1 succeeded at 35 (offset -5 lines).
>>>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #2 FAILED at 705.
>>>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #3 FAILED at 971.
>>>>>>>>>>>>>>>>>>>>>>>>>>> 2 out of 3 hunks FAILED -- saving rejects to
>> file
>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_config.cc.rej
>>>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_config.h
>>>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #1 FAILED at 304.
>>>>>>>>>>>>>>>>>>>>>>>>>>> 1 out of 1 hunk FAILED -- saving rejects to file
>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_config.h.rej
>>>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_dest.cc
>>>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_dest.h
>>>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_evt.cc
>>>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #1 FAILED at 1.
>>>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #2 succeeded at 30 with fuzz 2 (offset 2
>>>>>>>>>>>>>>>>>>>>>>>>>>> lines).
>>>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #3 succeeded at 1282 (offset 45 lines).
>>>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #4 succeeded at 1300 (offset 2 lines).
>>>>>>>>>>>>>>>>>>>>>>>>>>> 1 out of 4 hunks FAILED -- saving rejects to
>> file
>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_evt.cc.rej
>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>>
>> ==============================================================
>>>>>>>>>>>>>>>>>>>>>>>>>>> ===
>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>> -AVM
>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>> On 2/21/2017 3:03 PM, Vu Minh Nguyen
>> wrote:
>>>>>>>>>>>>>>>>>>>>>>>>>>>> Summary: log: add alternative destinations
>> of
>>>> log
>>>>>>>>>>> records
>>>>>>>>>>>>>>>>>> [#2258]
>>>>>>>>>>>>>>>>>>>>>>>>>>>> Review request for Trac Ticket(s): #2258
>>>>>>>>>>>>>>>>>>>>>>>>>>>> Peer Reviewer(s): Lennart, Canh, Mahesh
>>>>>>>>>>>>>>>>>>>>>>>>>>>> Pull request to: <<LIST THE PERSON WITH
>>>> PUSH
>>>>>>>>>>> ACCESS
>>>>>>>>>>>>>>>> HERE>>
>>>>>>>>>>>>>>>>>>>>>>>>>>>> Affected branch(es): Default
>>>>>>>>>>>>>>>>>>>>>>>>>>>> Development branch: Default
>>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>>> --------------------------------
>>>>>>>>>>>>>>>>>>>>>>>>>>>> Impacted area Impact y/n
>>>>>>>>>>>>>>>>>>>>>>>>>>>> --------------------------------
>>>>>>>>>>>>>>>>>>>>>>>>>>>> Docs n
>>>>>>>>>>>>>>>>>>>>>>>>>>>>                Build system            n
>>>>>>>>>>>>>>>>>>>>>>>>>>>> RPM/packaging n
>>>>>>>>>>>>>>>>>>>>>>>>>>>> Configuration files     n
>>>>>>>>>>>>>>>>>>>>>>>>>>>>                Startup scripts         n
>>>>>>>>>>>>>>>>>>>>>>>>>>>>                SAF services n
>>>>>>>>>>>>>>>>>>>>>>>>>>>>                OpenSAF services        y
>>>>>>>>>>>>>>>>>>>>>>>>>>>>                Core libraries          n
>>>>>>>>>>>>>>>>>>>>>>>>>>>> Samples n
>>>>>>>>>>>>>>>>>>>>>>>>>>>> Tests y
>>>>>>>>>>>>>>>>>>>>>>>>>>>> Other n
>>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>>> Comments (indicate scope for each "y"
>> above):
>>>>>>>>>>>>>>>>>>>>>>>>>>>> ---------------------------------------------
>>>>>>>>>>>>>>>>>>>>>>>>>>>>                To run the test, this patch has
>>>>>> dependent
>>>>>>>> on
>>>>>>>>>>>>>>> following
>>>>>>>>>>>>>>>>>>>> patches:
>>>>>>>>>>>>>>>>>>>>>>>>>>>> 1) #2293 (sent by Anders Widel, but not
>>>>>> yet
>>>>>>>>>>>>>>>>>>>>>>>>>>>> pushed)
>>>>>>>>>>>>>>>>>>>>>>>>>>>>                2) #2258 (v2, sent by Lennart,
> but
>>>>>>>>>>>>>>>>>>>>>>>>>>>> not
>>>>>> yet
>>>>>>>>>> pushed
>>>>>>>>>>>>>>> yet)
>>>>>>>>>>>>>>>>>>>>>>>>>>>> changeset
>>>>>>>>>>>>> d74aaf3025c99cade3165a15831124548f4d85bd
>>>>>>>>>>>>>>>>>>>>>>>>>>>> Author: Vu Minh Nguyen
>>>>>>>>>>>>> <vu.m.ngu...@dektech.com.au>
>>>>>>>>>>>>>>>>>>>>>>>>>>>> Date: Wed, 15 Feb 2017 14:36:00 +0700
>>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>>>           log: add alternative destinations of
> log
>>>>>> records
>>>>>>>>>>>>>>>>>>>>>>>>>>>> [#2258]
>>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>>>           Here are major info, detailed info
> will
>>> be
>>>>>>> added
>>>>>>>> to
>>>>>>>>>>>>>>>>>>>>>>>>>>>> PR doc
>>>>>>>>>>>>>>>>>>>>>> soon. 1)
>>>>>>>>>>>>>>>>>>>>>>>>>>> Add
>>>>>>>>>>>>>>>>>>>>>>>>>>>> attribute "saLogRecordDestination" to log
>>>>>>>>> stream.
>>>>>>>>>> 2)
>>>>>>>>>>>>>>>>>>>>>>>>>>>> Add
>>>>>>>>>>>>>>>>>>>>>> Local
>>>>>>>>>>>>>>>>>>>>>>>>>>> socket
>>>>>>>>>>>>>>>>>>>>>>>>>>>> destintion handler 3) Integrate into first
>>>>>>>>>> increment
>>>>>>>>>>>>>>>>>>>>>>>>>>>> made by
>>>>>>>>>>>>>>>>>>>>>> Lennart
>>>>>>>>>>>>>>>>>>>>>>>>>>>> changeset
>>>>>>>>>>>>> 4bae27a478c235df3058f43c92d3a5483233b01d
>>>>>>>>>>>>>>>>>>>>>>>>>>>> Author: Vu Minh Nguyen
>>>>>>>>>>>>> <vu.m.ngu...@dektech.com.au>
>>>>>>>>>>>>>>>>>>>>>>>>>>>> Date: Wed, 15 Feb 2017 15:07:09 +0700
>>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>>>           log: add UML test case to verify
>>>>>>>>>>>>>>>>>>>>>>>>>>>> alternative
>>>>>>>>>>>>>>>>>>>>>>>>>>>> destination
>>>>>>>>>>>>>>>>>>>>>> [#2258]
>>>>>>>>>>>>>>>>>>>>>>>>>>>> Major changes: 1) Modify Lennart's test
>> cases
>>>>>>>>>>> because
>>>>>>>>>>>>>>>>>>>>>> enhancing
>>>>>>>>>>>>>>>>>>>>>>>>>>> destination
>>>>>>>>>>>>>>>>>>>>>>>>>>>> configuration validation rules. 2) Add test
>>>>>>>> suite
>>>>>>>>>>>>>>>>>>>>>>>>>>>> #17 to
>>>>>>>>>>>>>>>>>>>>>> verify
>>>>>>>>>>>>>>>>>>>>>>>>>>> alternative
>>>>>>>>>>>>>>>>>>>>>>>>>>>> destination
>>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>>> changeset
>>>>>>>>>>> bc375725fed22bb4f8cb3ae3df5f96fb9d281efb
>>>>>>>>>>>>>>>>>>>>>>>>>>>> Author: Vu Minh Nguyen
>>>>>>>>>>>>> <vu.m.ngu...@dektech.com.au>
>>>>>>>>>>>>>>>>>>>>>>>>>>>> Date: Thu, 16 Feb 2017 17:22:13 +0700
>>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>>>           log: add unit tests to verify
> interfaces
>>>>>>>> provided
>>>>>>>>>> by
>>>>>>>>>>>>>>>>>>>>>> destination
>>>>>>>>>>>>>>>>>>>>>>>>>>> handler
>>>>>>>>>>>>>>>>>>>>>>>>>>>> [#2258]
>>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>>>           Unit tests to verify major interfaces:
> 1)
>>>>>>>>>>>>>>>>>>>>>>>>>>>> CfgDestination()
>>>>>>>>>>>>>>>>>>>>>> 2)
>>>>>>>>>>>>>>>>>>>>>>>>>>>> WriteToDestination()
>>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>>> Added Files:
>>>>>>>>>>>>>>>>>>>>>>>>>>>> ------------
>>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/apitest/tet_cfg_destination.c
>>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_dest.cc
>>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_dest.h
>>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv_v7.cc
>>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv_v7.h
>>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/tests/lgs_dest_test.cc
>>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/tests/Makefile
>>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>>> Complete diffstat:
>>>>>>>>>>>>>>>>>>>>>>>>>>>> ------------------
>>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/Makefile |    4 +
>>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/Makefile.am |   31 +++++-
>>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/apitest/tet_LogOiOps.c |    8 +-
>>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/apitest/tet_cfg_destination.c |  483
>> ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++
>>>>>>>>>>>>>>>>>>>>>>>>>>> ++++++++++++++++++++++++++++++++++
>>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/config/logsv_classes.xml |    7 +-
>>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_config.cc |  169
>>>>>>>>>>>>>>>>>>>>>>>>>>> +++++++++++++++++++++++++++++---
>>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_config.h |    3 +-
>>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_dest.cc |  707
>> ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++
>> ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++
>>>>>>>>>>>>>>>>>>>>>>>>>>> +++++++++++++++++
>>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_dest.h |  576
>> ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++
>>>> ++++++++++++++++++++++++++++++++++++++++++++++++++++
>>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_evt.cc |   33 ++++++
>>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_imm.cc |  202
>>>>>>>>>>>>>>>>>>>>>>>>>>> +++++++++++++++++++++++++++++++++------
>>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_main.cc |    8 +
>>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc |  103
>>>>>>>>>>>>>>>>>>>> ++++++++++++++++++-
>>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv.h |    6 +-
>>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv_v5.cc |   10 +
>>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv_v7.cc |  177
>>>>>>>>>>>>>>>>>>>>>>>>>>> +++++++++++++++++++++++++++++++++++
>>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv_v7.h |   67
>>>>>>>>>>>>>>>>> +++++++++++++
>>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_stream.cc |   60
>>>>>>>>>>>>>>>>> +++++++++++-
>>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_stream.h |   16 +++
>>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_util.cc |   63
>>>>>>>>>>>>>>>>> ++++++++++++
>>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_util.h |   11 +-
>>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/tests/Makefile |   20 +++
>>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/tests/lgs_dest_test.cc |  209
>>>> +++++++++++++++++++++++++++++++++++++++++
>>>>>>>>>>>>>>>>>>>>>>>>>>>> 23 files changed, 2896 insertions(+), 77
>>>>>>>>>>>>>>>>>>>>>>>>>>>> deletions(-)
>>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>>> Testing Commands:
>>>>>>>>>>>>>>>>>>>>>>>>>>>> -----------------
>>>>>>>>>>>>>>>>>>>>>>>>>>>>                Run UML test suite #17
>>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>>> Testing, Expected Results:
>>>>>>>>>>>>>>>>>>>>>>>>>>>> --------------------------
>>>>>>>>>>>>>>>>>>>>>>>>>>>>                All test passed
>>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>>> Conditions of Submission:
>>>>>>>>>>>>>>>>>>>>>>>>>>>> -------------------------
>>>>>>>>>>>>>>>>>>>>>>>>>>>> <<HOW MANY DAYS BEFORE PUSHING,
>>>>>>>>>>> CONSENSUS
>>>>>>>>>>>>>>>> ETC>>
>>>>>>>>>>>>>>>>>>>>>>>>>>>> Arch Built     Started Linux distro
>>>>>>>>>>>>>>>>>>>>>>>>>>>> -------------------------------------------
>>>>>>>>>>>>>>>>>>>>>>>>>>>> mips n          n
>>>>>>>>>>>>>>>>>>>>>>>>>>>> mips64 n          n
>>>>>>>>>>>>>>>>>>>>>>>>>>>> x86 n          n
>>>>>>>>>>>>>>>>>>>>>>>>>>>> x86_64 n          n
>>>>>>>>>>>>>>>>>>>>>>>>>>>> powerpc n          n
>>>>>>>>>>>>>>>>>>>>>>>>>>>> powerpc64 n          n
>>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>>> Reviewer Checklist:
>>>>>>>>>>>>>>>>>>>>>>>>>>>> -------------------
>>>>>>>>>>>>>>>>>>>>>>>>>>>> [Submitters: make sure that your review
>>>> doesn't
>>>>>>>>> trigger
>>>>>>>>>>>>> any
>>>>>>>>>>>>>>>>>>>>>>>>> checkmarks!]
>>>>>>>>>>>>>>>>>>>>>>>>>>>> Your checkin has not passed review because
>>>> (see
>>>>>>>>>>> checked
>>>>>>>>>>>>>>>>>> entries):
>>>>>>>>>>>>>>>>>>>>>>>>>>>> ___ Your RR template is generally
>> incomplete;
>>>> it
>>>>>>> has
>>>>>>>>>>> too
>>>>>>>>>>>>>>>> many
>>>>>>>>>>>>>>>>>>>> blank
>>>>>>>>>>>>>>>>>>>>>>>>>> entries
>>>>>>>>>>>>>>>>>>>>>>>>>>>> that need proper data filled in.
>>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>>> ___ You have failed to nominate the proper
>>>>>>> persons
>>>>>>>>> for
>>>>>>>>>>>>>>>> review
>>>>>>>>>>>>>>>>>> and
>>>>>>>>>>>>>>>>>>>>>>>> push.
>>>>>>>>>>>>>>>>>>>>>>>>>>>> ___ Your patches do not have proper
>>>> short+long
>>>>>>>>>>> header
>>>>>>>>>>>>>>>>>>>>>>>>>>>> ___ You have grammar/spelling in your
>> header
>>>>>>> that
>>>>>>>>> is
>>>>>>>>>>>>>>>>>> unacceptable.
>>>>>>>>>>>>>>>>>>>>>>>>>>>> ___ You have exceeded a sensible line
>> length
>>>> in
>>>>>>> your
>>>>>>>>>>>>>>>>>>>>>>>>>>> headers/comments/text.
>>>>>>>>>>>>>>>>>>>>>>>>>>>> ___ You have failed to put in a proper Trac
>>>> Ticket
>>>>>>> #
>>>>>>>>>> into
>>>>>>>>>>>>>>> your
>>>>>>>>>>>>>>>>>>>>>>>> commits.
>>>>>>>>>>>>>>>>>>>>>>>>>>>> ___ You have incorrectly put/left internal
>> data
>>>> in
>>>>>>>>> your
>>>>>>>>>>>>>>>>>>>>>> comments/files
>>>>>>>>>>>>>>>>>>>>>>>>>>>> (i.e. internal bug tracking tool IDs, product
>>>>>>>>>>>>>>> names
>>>>>>>>>>>>>>>>> etc)
>>>>>>>>>>>>>>>>>>>>>>>>>>>> ___ You have not given any evidence of
>> testing
>>>>>>>>> beyond
>>>>>>>>>>>>> basic
>>>>>>>>>>>>>>>>>> build
>>>>>>>>>>>>>>>>>>>>>>>> tests.
>>>>>>>>>>>>>>>>>>>>>>>>>>>> Demonstrate some level of runtime or other
>>>> sanity
>>>>>>>>>>>>>>>>> testing.
>>>>>>>>>>>>>>>>>>>>>>>>>>>> ___ You have ^M present in some of your
>> files.
>>>>>>>>> These
>>>>>>>>>>>>>>>>>>>>>>>>>>>> have to
>>>>>>>>>>>>>>>>>> be
>>>>>>>>>>>>>>>>>>>>>>>>> removed.
>>>>>>>>>>>>>>>>>>>>>>>>>>>> ___ You have needlessly changed
>> whitespace
>>>> or
>>>>>>>>> added
>>>>>>>>>>>>>>>>>> whitespace
>>>>>>>>>>>>>>>>>>>>>>>>> crimes
>>>>>>>>>>>>>>>>>>>>>>>>>>>> like trailing spaces, or spaces
>>>>>> before
>>>>>>>>>> tabs.
>>>>>>>>>>>>>>>>>>>>>>>>>>>> ___ You have mixed real technical changes
>>>> with
>>>>>>>>>>>>> whitespace
>>>>>>>>>>>>>>>> and
>>>>>>>>>>>>>>>>>>>>> other
>>>>>>>>>>>>>>>>>>>>>>>>>>>> cosmetic code cleanup changes. These have
>> to
>>>> be
>>>>>>>>>>>>>>>>>> separate
>>>>>>>>>>>>>>>>>>>>>>>> commits.
>>>>>>>>>>>>>>>>>>>>>>>>>>>> ___ You need to refactor your submission
>> into
>>>>>>>>> logical
>>>>>>>>>>>>>>>>>>>>>>>>>>>> chunks;
>>>>>>>>>>>>>>>>>>>> there
>>>>>>>>>>>>>>>>>>>>>> is
>>>>>>>>>>>>>>>>>>>>>>>>>>>> too much content into a single
>>>>>> commit.
>>>>>>>>>>>>>>>>>>>>>>>>>>>> ___ You have extraneous garbage in your
>>>> review
>>>>>>>>>>> (merge
>>>>>>>>>>>>>>>>>> commits
>>>>>>>>>>>>>>>>>>>>> etc)
>>>>>>>>>>>>>>>>>>>>>>>>>>>> ___ You have giant attachments which
>> should
>>>>>>> never
>>>>>>>>>>> have
>>>>>>>>>>>>>>>> been
>>>>>>>>>>>>>>>>>>>>> sent;
>>>>>>>>>>>>>>>>>>>>>>>>>>>> Instead you should place your content in a
>>>> public
>>>>>>>>>>>>>>>>> tree to
>>>>>>>>>>>>>>>>>>>> be
>>>>>>>>>>>>>>>>>>>>>>>>>> pulled.
>>>>>>>>>>>>>>>>>>>>>>>>>>>> ___ You have too many commits attached
>> to
>>>> an e-
>>>>>>>>> mail;
>>>>>>>>>>>>>>>> resend
>>>>>>>>>>>>>>>>>> as
>>>>>>>>>>>>>>>>>>>>>>>>> threaded
>>>>>>>>>>>>>>>>>>>>>>>>>>>> commits, or place in a public tree for a pull.
>>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>>> ___ You have resent this content multiple
>>>> times
>>>>>>>>>>> without a
>>>>>>>>>>>>>>>> clear
>>>>>>>>>>>>>>>>>>>>>>>>>> indication
>>>>>>>>>>>>>>>>>>>>>>>>>>>> of what has changed between each
>>>>>>>> re-send.
>>>>>>>>>>>>>>>>>>>>>>>>>>>> ___ You have failed to adequately and
>>>> individually
>>>>>>>>>>> address
>>>>>>>>>>>>>>> all
>>>>>>>>>>>>>>>>> of
>>>>>>>>>>>>>>>>>>>>>> the
>>>>>>>>>>>>>>>>>>>>>>>>>>>> comments and change requests that were
>>>>>>> proposed
>>>>>>>>> in
>>>>>>>>>>>>>>>>> the
>>>>>>>>>>>>>>>>>>>>>> initial
>>>>>>>>>>>>>>>>>>>>>>>>>> review.
>>>>>>>>>>>>>>>>>>>>>>>>>>>> ___ You have a misconfigured ~/.hgrc file
>> (i.e.
>>>>>>>>>>> username,
>>>>>>>>>>>>>>>> email
>>>>>>>>>>>>>>>>>>>> etc)
>>>>>>>>>>>>>>>>>>>>>>>>>>>> ___ Your computer have a badly configured
>>>> date
>>>>>>> and
>>>>>>>>>>>>> time;
>>>>>>>>>>>>>>>>>>>>> confusing
>>>>>>>>>>>>>>>>>>>>>>> the
>>>>>>>>>>>>>>>>>>>>>>>>>>>> the threaded patch review.
>>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>>> ___ Your changes affect IPC mechanism,
>> and
>>>> you
>>>>>>>>> don't
>>>>>>>>>>>>>>>> present
>>>>>>>>>>>>>>>>>> any
>>>>>>>>>>>>>>>>>>>>>>>>> results
>>>>>>>>>>>>>>>>>>>>>>>>>>>> for in-service upgradability test.
>>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>>> ___ Your changes affect user manual and
>>>>>>>>>>> documentation,
>>>>>>>>>>>>>>>> your
>>>>>>>>>>>>>>>>>>>>> patch
>>>>>>>>>>>>>>>>>>>>>>>>> series
>>>>>>>>>>>>>>>>>>>>>>>>>>>> do not contain the patch that updates
>>>>>>>> the
>>>>>>>>>>>>>>>>>>>>>>>>>>>> Doxygen
>>>>>>>>>>>>>>>>>> manual.
> ----------------------------------------------------------------------------
>>>>>>>>>>>>>>>>> --
>>>>>>>>>>>>>>>>>>> Check out the vibrant tech community on one of the
>>>> world's
>>>>>>>>> most
>>>>>>>>>>>>>>>>>>> engaging tech sites, SlashDot.org!
>>>> http://sdm.link/slashdot
>> _______________________________________________
>>>>>>>>>>>>>>>>>>> Opensaf-devel mailing list
>>>>>>>>>>>>>>>>>>> Opensaf-devel@lists.sourceforge.net
>>>>>>>>>>>>>>>>>>> https://lists.sourceforge.net/lists/listinfo/opensaf-
>> devel


------------------------------------------------------------------------------
Check out the vibrant tech community on one of the world's most
engaging tech sites, SlashDot.org! http://sdm.link/slashdot
_______________________________________________
Opensaf-devel mailing list
Opensaf-devel@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/opensaf-devel

Reply via email to