Hi Vu,

Issue still exist even with `2258_v10.patch`.
fix all the issue and republish the patch.

==========================================================================================================================================
 
Feb 23 17:09:06 SC-2 osafamfnd[3774]: NO 
'safSu=SC-2,safSg=2N,safApp=OpenSAF' Presence State INSTANTIATING => 
INSTANTIATED Feb 23 17:09:06 SC-2 osafamfnd[3774]: NO Assigning 
'safSi=SC-2N,safApp=OpenSAF' STANDBY to 
'safSu=SC-2,safSg=2N,safApp=OpenSAF' Feb 23 17:09:06 SC-2 
osafrded[3694]: NO RDE role set to STANDBY Feb 23 17:09:06 SC-2 
osafrded[3694]: NO Peer up on node 0x2010f Feb 23 17:09:06 SC-2 
osafrded[3694]: NO Got peer info request from node 0x2010f with role 
ACTIVE Feb 23 17:09:06 SC-2 osafrded[3694]: NO Got peer info response 
from node 0x2010f with role ACTIVE Feb 23 17:09:06 SC-2 osafimmd[3713]: 
NO MDS event from svc_id 24 (change:3, dest:13) Feb 23 17:09:06 SC-2 
osafimmd[3713]: NO MDS event from svc_id 24 (change:5, dest:13) Feb 23 
17:09:06 SC-2 osafimmd[3713]: NO MDS event from svc_id 24 (change:5, 
dest:13) Feb 23 17:09:06 SC-2 osafimmd[3713]: NO MDS event from svc_id 
25 (change:3, dest:565213468688400) Feb 23 17:09:06 SC-2 osafimmd[3713]: 
NO MDS event from svc_id 25 (change:3, dest:564116543111184) Feb 23 
17:09:06 SC-2 osafimmnd[3724]: NO Implementer (applier) connected: 15 
(@safAmfService2020f) <127, 2020f> Feb 23 17:09:06 SC-2 osaflogd[3734]: 
NO LOGSV_DATA_GROUPNAME not found Feb 23 17:09:06 SC-2 osaflogd[3734]: 
NO LOG root directory is: "/var/log/opensaf/saflog" Feb 23 17:09:06 SC-2 
osaflogd[3734]: NO LOG data group is: "" Feb 23 17:09:06 SC-2 
osaflogd[3734]: NO LGS_MBCSV_VERSION = 7 Feb 23 17:09:06 SC-2 
osafamfnd[3774]: NO Assigned 'safSi=SC-2N,safApp=OpenSAF' STANDBY to 
'safSu=SC-2,safSg=2N,safApp=OpenSAF' Feb 23 17:09:06 SC-2 opensafd: 
OpenSAF(5.1.M0 - ) services successfully started done SC-2:~ # Feb 23 
17:09:06 SC-2 osafimmnd[3724]: NO Implementer (applier) connected: 16 
(@OpenSafImmReplicatorB) <150, 2020f> Feb 23 17:09:06 SC-2 
osafntfimcnd[3931]: NO Started Feb 23 17:09:08 SC-2 osafamfd[3764]: NO 
Cold sync complete! Feb 23 17:09:08 SC-2 osaflogd[3734]: WA FAILED: 
ncs_patricia_tree_add, client_id 0 Feb 23 17:09:08 SC-2 osaflogd[3734]: 
ER Exiting with message: Could not create new client Feb 23 17:09:08 
SC-2 osafamfnd[3774]: NO 'safSu=SC-2,safSg=2N,safApp=OpenSAF' component 
restart probation timer started (timeout: 60000000000 ns) Feb 23 
17:09:08 SC-2 osafamfnd[3774]: NO Restarting a component of 
'safSu=SC-2,safSg=2N,safApp=OpenSAF' (comp restart count: 1) Feb 23 
17:09:08 SC-2 osafamfnd[3774]: NO 
'safComp=LOG,safSu=SC-2,safSg=2N,safApp=OpenSAF' faulted due to 
'errorReport' : Recovery is 'componentRestart' 
==========================================================================================================================================


-AVM


On 2/23/2017 4:25 PM, Vu Minh Nguyen wrote:
> Hi Mahesh,
>
> Sorry, I missed fixing other point. See 2258_additional_fix_err.patch.
>
> For your convenience, I pack them all to new version 2258_v10.patch.
>
> Regards, Vu
>
>> -----Original Message-----
>> From: Vu Minh Nguyen [mailto:vu.m.ngu...@dektech.com.au]
>> Sent: Thursday, February 23, 2017 5:39 PM
>> To: 'A V Mahesh' <mahesh.va...@oracle.com>; 'lennart.l...@ericsson.com'
>> <lennart.l...@ericsson.com>; 'canh.v.tru...@dektech.com.au'
>> <canh.v.tru...@dektech.com.au>
>> Cc: 'opensaf-devel@lists.sourceforge.net' <opensaf-
>> de...@lists.sourceforge.net>
>> Subject: RE: [devel] [PATCH 0 of 3] Review Request for log: add
> alternative
>> destinations of log records [#2258] V4
>>
>> Hi Mahesh,
>>
>> I found the root cause. It is because in ` log: implement
>> SaLogFilterSetCallbackT and version handling [#2146]`,
>> Canh introduced MBCSV version #6, but I missed adding that info when
>> rebasing.
>>
>> The attached patch contains the fix. Can you apply it to see if the
> problem
>> still occur? Thanks.
>>
>> Regards, Vu
>>
>>> -----Original Message-----
>>> From: A V Mahesh [mailto:mahesh.va...@oracle.com]
>>> Sent: Thursday, February 23, 2017 5:32 PM
>>> To: Vu Minh Nguyen <vu.m.ngu...@dektech.com.au>;
>>> lennart.l...@ericsson.com; canh.v.tru...@dektech.com.au
>>> Cc: opensaf-devel@lists.sourceforge.net
>>> Subject: Re: [devel] [PATCH 0 of 3] Review Request for log: add
> alternative
>>> destinations of log records [#2258] V4
>>>
>>> Hi Vu,
>>>
>>> Please find attached , below is the time stamp of SC-2
>>>
>>>
>> ==============================================================
>>> ==============================================
>>>
>>> Feb 23 15:55:30 SC-2 osafimmnd[6978]: NO Implementer (applier)
>>> connected: 15 (@safAmfService2020f) <127, 2020f>
>>> Feb 23 15:55:30 SC-2 osaflogd[6988]: NO LOGSV_DATA_GROUPNAME not
>>> found
>>> Feb 23 15:55:30 SC-2 osaflogd[6988]: NO LOG root directory is:
>>> "/var/log/opensaf/saflog"
>>> Feb 23 15:55:30 SC-2 osaflogd[6988]: NO LOG data group is: ""
>>> Feb 23 15:55:30 SC-2 osaflogd[6988]: NO LGS_MBCSV_VERSION = 7
>>> Feb 23 15:55:30 SC-2 osafamfnd[7028]: NO Assigned
>>> 'safSi=SC-2N,safApp=OpenSAF' STANDBY to 'safSu=SC-
>>> 2,safSg=2N,safApp=OpenSAF'
>>> Feb 23 15:55:30 SC-2 opensafd: OpenSAF(5.1.M0 - ) services successfully
>>> started
>>> done
>>> SC-2:/var/log/opensaf # Feb 23 15:55:31 SC-2 osafimmnd[6978]: NO
>>> Implementer (applier) connected: 16 (@OpenSafImmReplicatorB) <144,
>>> 2020f>
>>> Feb 23 15:55:31 SC-2 osafntfimcnd[7185]: NO Started
>>> Feb 23 15:55:33 SC-2 osafamfd[7018]: NO Cold sync complete!
>>> Feb 23 15:55:33 SC-2 osaflogd[6988]: WA FAILED: ncs_patricia_tree_add,
>>> client_id 0
>>> Feb 23 15:55:33 SC-2 osaflogd[6988]: ER Exiting with message: Could not
>>> create new client
>>> Feb 23 15:55:33 SC-2 osafamfnd[7028]: NO
>>> 'safSu=SC-2,safSg=2N,safApp=OpenSAF' component restart probation
>> timer
>>> started (timeout: 60000000000 ns)
>>> Feb 23 15:55:33 SC-2 osafamfnd[7028]: NO Restarting a component of
>>> 'safSu=SC-2,safSg=2N,safApp=OpenSAF' (comp restart count: 1)
>>> Feb 23 15:55:33 SC-2 osafamfnd[7028]: NO
>>> 'safComp=LOG,safSu=SC-2,safSg=2N,safApp=OpenSAF' faulted due to
>>> 'errorReport' : Recovery is 'componentRestart'
>>>
>>>
>> ==============================================================
>>> ==============================================
>>>
>>> -AVM
>>>
>>>
>>> On 2/23/2017 3:39 PM, Vu Minh Nguyen wrote:
>>>> Hi Mahesh,
>>>>
>>>> No change in V7 vs V9. Just do rebase the code on latest changeset.
>>>>
>>>> I have tried to clean up all, and rebuild the cluster to see what you
> are
>>>> observing,
>>>> and I am not able to reproduce the problem, I have tried several
> times.
>>>> Can you provide me the osaflogd trace on both SCs node? Thanks.
>>>>
>>>> Regards, Vu
>>>>
>>>>> -----Original Message-----
>>>>> From: A V Mahesh [mailto:mahesh.va...@oracle.com]
>>>>> Sent: Thursday, February 23, 2017 4:48 PM
>>>>> To: Vu Minh Nguyen <vu.m.ngu...@dektech.com.au>;
>>>>> lennart.l...@ericsson.com; canh.v.tru...@dektech.com.au
>>>>> Cc: opensaf-devel@lists.sourceforge.net
>>>>> Subject: Re: [devel] [PATCH 0 of 3] Review Request for log: add
>>>> alternative
>>>>> destinations of log records [#2258] V4
>>>>>
>>>>> Hi Vu,
>>>>>
>>>>> On 2/23/2017 3:13 PM, A V Mahesh wrote:
>>>>>> Not sure what are other change compare to V7 to V9 , New problems
>>> got
>>>>>> introduced
>>>>>>
>>>>>> Both nodes  SC-1 & SC-2 ( with 2258_v9.patch ) , trying bring up
> both
>>>>>> SC`s  simple node bringup  ,
>>>>>>
>>>>>> SC-2 going for reboot with following :
>>>>>>
>>>>>>
>> ==============================================================
>> ==============================================================
>>>>> ============
>>>>>> Feb 23 15:05:32 SC-2 osaflogd[29988]: NO
>> LOGSV_DATA_GROUPNAME
>>>>> not found
>>>>>> Feb 23 15:05:32 SC-2 osaflogd[29988]: NO LOG root directory is:
>>>>>> "/var/log/opensaf/saflog"
>>>>>> Feb 23 15:05:32 SC-2 osaflogd[29988]: NO LOG data group is: ""
>>>>>> Feb 23 15:05:32 SC-2 osafimmnd[29978]: NO Implementer (applier)
>>>>>> connected: 16 (@safAmfService2020f) <127, 2020f>
>>>>>> Feb 23 15:05:32 SC-2 osaflogd[29988]: NO LGS_MBCSV_VERSION = 7
>>>>>> Feb 23 15:05:32 SC-2 osaflogd[29988]: WA FAILED:
>>>>>> ncs_patricia_tree_add, client_id 0
>>>>>> Feb 23 15:05:32 SC-2 osafamfnd[30028]: NO Assigned
>>>>>> 'safSi=SC-2N,safApp=OpenSAF' STANDBY to
>>>>>> 'safSu=SC-2,safSg=2N,safApp=OpenSAF'
>>>>>> Feb 23 15:05:32 SC-2 osaflogd[29988]: ER Exiting with message: Could
>>>>>> not create new client
>>>>>> Feb 23 15:05:32 SC-2 osafamfnd[30028]: NO
>>>>>> 'safSu=SC-2,safSg=2N,safApp=OpenSAF' component restart probation
>>>>> timer
>>>>>> started (timeout: 60000000000 ns)
>>>>>> Feb 23 15:05:32 SC-2 osafamfnd[30028]: NO Restarting a component
>> of
>>>>>> 'safSu=SC-2,safSg=2N,safApp=OpenSAF' (comp restart count: 1)
>>>>>> Feb 23 15:05:32 SC-2 osafamfnd[30028]: NO
>>>>>> 'safComp=LOG,safSu=SC-2,safSg=2N,safApp=OpenSAF' faulted due to
>>>>>> 'errorReport' : Recovery is 'componentRestart'
>>>>>> Feb 23 15:05:32 SC-2 opensafd[29908]: ER Service LOGD has
>>> unexpectedly
>>>>>> crashed. Unable to continue, exiting
>>>>>> Feb 23 15:05:32 SC-2 osafamfd[30018]: exiting for shutdown
>>>>>> Feb 23 15:05:32 SC-2 osafamfnd[30028]: ER AMFD has unexpectedly
>>>>>> crashed. Rebooting node
>>>>>> Feb 23 15:05:32 SC-2 osafamfnd[30028]: Rebooting OpenSAF NodeId =
>>>>>> 131599 EE Name = , Reason: AMFD has unexpectedly crashed.
>>> Rebooting
>>>>>> node, OwnNodeId = 131599, SupervisionTime = 60
>>>>>> Feb 23 15:05:32 SC-2 opensaf_reboot: Rebooting local node;
>>> timeout=60
>>>>>> Feb 23 15:06:04 SC-2 syslog-ng[1180]: syslog-ng starting up;
>>>>>> version='2.0.9'
>>>>>>
>>>>>>
>> ==============================================================
>> ==============================================================
>>>>> ============
>>>>> Some times :
>>>>>
>>>>>
>> ==============================================================
>> ==============================================================
>>>>> ============
>>>>>
>>>>> Feb 23 15:15:19 SC-2 osafrded[3858]: NO RDE role set to STANDBY
>>>>> Feb 23 15:15:19 SC-2 osafrded[3858]: NO Peer up on node 0x2010f
>>>>> Feb 23 15:15:19 SC-2 osafrded[3858]: NO Got peer info request from
>>> node
>>>>> 0x2010f with role ACTIVE
>>>>> Feb 23 15:15:19 SC-2 osafrded[3858]: NO Got peer info response from
>>> node
>>>>> 0x2010f with role ACTIVE
>>>>> Feb 23 15:15:19 SC-2 osafimmd[3877]: NO MDS event from svc_id 24
>>>>> (change:3, dest:13)
>>>>> Feb 23 15:15:19 SC-2 osafimmd[3877]: NO MDS event from svc_id 24
>>>>> (change:5, dest:13)
>>>>> Feb 23 15:15:19 SC-2 osafimmd[3877]: NO MDS event from svc_id 24
>>>>> (change:5, dest:13)
>>>>> Feb 23 15:15:19 SC-2 osafimmd[3877]: NO MDS event from svc_id 25
>>>>> (change:3, dest:565217560625168)
>>>>> Feb 23 15:15:19 SC-2 osafimmd[3877]: NO MDS event from svc_id 25
>>>>> (change:3, dest:564114674417680)
>>>>> Feb 23 15:15:19 SC-2 osaflogd[3898]: NO LOGSV_DATA_GROUPNAME
>>> not
>>>>> found
>>>>> Feb 23 15:15:19 SC-2 osaflogd[3898]: NO LOG root directory is:
>>>>> "/var/log/opensaf/saflog"
>>>>> Feb 23 15:15:19 SC-2 osaflogd[3898]: NO LOG data group is: ""
>>>>> Feb 23 15:15:19 SC-2 osafimmnd[3888]: NO Implementer (applier)
>>>>> connected: 15 (@safAmfService2020f) <127, 2020f>
>>>>> Feb 23 15:15:19 SC-2 osaflogd[3898]: NO LGS_MBCSV_VERSION = 7
>>>>> Feb 23 15:15:19 SC-2 osaflogd[3898]: ER Exiting with message: Client
>>>>> attributes differ
>>>>> Feb 23 15:15:19 SC-2 osafamfnd[3938]: NO
>>>>> 'safSu=SC-2,safSg=2N,safApp=OpenSAF' component restart probation
>>> timer
>>>>> started (timeout: 60000000000 ns)
>>>>> Feb 23 15:15:19 SC-2 osafamfnd[3938]: NO Restarting a component of
>>>>> 'safSu=SC-2,safSg=2N,safApp=OpenSAF' (comp restart count: 1)
>>>>> Feb 23 15:15:19 SC-2 osafamfnd[3938]: NO
>>>>> 'safComp=LOG,safSu=SC-2,safSg=2N,safApp=OpenSAF' faulted due to
>>>>> 'errorReport' : Recovery is 'componentRestart'
>>>>> Feb 23 15:15:19 SC-2 opensafd[3818]: ER Service LOGD has
>> unexpectedly
>>>>> crashed. Unable to continue, exiting
>>>>> Feb 23 15:15:20 SC-2 osafamfd[3928]: exiting for shutdown
>>>>> Feb 23 15:15:20 SC-2 osafamfnd[3938]: ER AMFD has unexpectedly
>>> crashed.
>>>>> Rebooting node
>>>>> Feb 23 15:15:20 SC-2 osafamfnd[3938]: Rebooting OpenSAF NodeId =
>>>>> 131599
>>>>> EE Name = , Reason: AMFD has unexpectedly crashed. Rebooting node,
>>>>> OwnNodeId = 131599, SupervisionTime = 60
>>>>> Feb 23 15:15:20 SC-2 osafimmnd[3888]: NO Implementer locally
>>>>> disconnected. Marking it as doomed 15 <127, 2020f>
>>> (@safAmfService2020f)
>>>>> Feb 23 15:15:20 SC-2 osafimmnd[3888]: NO Implementer disconnected
>> 15
>>>>> <127, 2020f> (@safAmfService2020f)
>>>>> Feb 23 15:15:20 SC-2 opensaf_reboot: Rebooting local node;
>> timeout=60
>> ==============================================================
>> ==============================================================
>>>>> ============
>>>>>
>>>>>
>>>>>> -AVM
>>>>>>
>>>>>>
>>>>>> On 2/23/2017 2:20 PM, Vu Minh Nguyen wrote:
>>>>>>> Hi Mahesh,
>>>>>>>
>>>>>>> This is the latest code has been rebased on the latest changeset.
>>>>>>>
>>>>>>> Note that, in the attached patch, I have included one more
>>> dependency,
>>>>>>> that is on base::Hash() function, the patch sent by Anders [#2266]
>>>>>>>
>>>>>>> Please review the patch, then comment if any. Thanks.
>>>>>>>
>>>>>>> Regards, Vu
>>>>>>>
>>>>>>>> -----Original Message-----
>>>>>>>> From: A V Mahesh [mailto:mahesh.va...@oracle.com]
>>>>>>>> Sent: Thursday, February 23, 2017 2:03 PM
>>>>>>>> To: Vu Minh Nguyen <vu.m.ngu...@dektech.com.au>;
>>>>>>>> lennart.l...@ericsson.com; canh.v.tru...@dektech.com.au
>>>>>>>> Cc: opensaf-devel@lists.sourceforge.net
>>>>>>>> Subject: Re: [devel] [PATCH 0 of 3] Review Request for log: add
>>>>>>> alternative
>>>>>>>> destinations of log records [#2258] V4
>>>>>>>>
>>>>>>>> Hi Vu,
>>>>>>>>
>>>>>>>> Now we are now able to proceed further with  V7 `2258_v7.patch` in
>>>>>>>> service upgraded working fine,
>>>>>>>> because of Encode/decode changes done in V7 patch.
>>>>>>>>
>>>>>>>> But we have another small test case issue (/usr/bin/logtest 5 17
>>>>>>>> Segmentation fault),
>>>>>>>> once we resolve this also, we can conclude that all the basic
>>>>>>>> functionality is working,
>>>>>>>> then you can re-publish the V7 patch  ( if change occurred in
> Lennart
>>>>>>>> #2258 V2 please do publish that as well )
>>>>>>>> so that I can go for CODE review.
>>>>>>>>
>>>>>>>> Steps to reproduce the test case issue :
>>>>>>>>
>>>>>>>> 1) Bring up old node as Active  ( with out `2258_v7.patch` )
>>>>>>>> 2) Bring-up new node as Standby  ( with `2258_v7.patch` )
>>>>>>>> 3)  Do `amf-adm si-swap safSi=SC-2N,safApp=OpenSAF`
>>>>>>>> 4)  Run `/usr/bin/logtest 5 17 ` on  new Active (because of
> si-swap )
>>>>>>>> Note :  both nodes has the new XLM attributes populated .
>>>>>>>>
>>>>>>>>
>> ==============================================================
>>>>>>>> =====================
>>>>>>>>
>>>>>>>> gdb /usr/bin/logtest
>>>>>>>> (gdb) r 5
>>>>>>>>
>>>>>>>>        16  PASSED   CCB Object Modify, change root directory. Path
>>>>>>>> exist. OK;
>>>>>>>> Detaching after fork from child process 13797.
>>>>>>>> Set values Fail
>>>>>>>> [New Thread 0x7ffff7ff7b00 (LWP 13801)]
>>>>>>>> [New Thread 0x7ffff7fc4b00 (LWP 13802)]
>>>>>>>>
>>>>>>>> Program received signal SIGSEGV, Segmentation fault.
>>>>>>>> 0x00005555555688ea in read_and_compare.isra.7 () at
>>>>>>>> src/log/apitest/tet_LogOiOps.c:1891
>>>>>>>> 1891    src/log/apitest/tet_LogOiOps.c: No such file or directory.
>>>>>>>>             in src/log/apitest/tet_LogOiOps.c
>>>>>>>> (gdb) bt
>>>>>>>> #0  0x00005555555688ea in read_and_compare.isra.7 () at
>>>>>>>> src/log/apitest/tet_LogOiOps.c:1891
>>>>>>>> #1  0x0000555555568a4b in
>>>>> check_logRecordDestinationConfigurationAdd ()
>>>>>>>> at src/log/apitest/tet_LogOiOps.c:1941
>>>>>>>> #2  0x0000555555571b05 in run_test_case ()
>>>>>>>> #3  0x0000555555571feb in test_run ()
>>>>>>>> #4  0x000055555555bfad in main () at src/log/apitest/logtest.c:569
>>>>>>>> (gdb)
>>>>>>>>
>>>>>>>>
>> ==============================================================
>>>>>>>> =====================
>>>>>>>>
>>>>>>>>
>>>>>>>> -AVM
>>>>>>>>
>>>>>>>> On 2/23/2017 11:44 AM, Vu Minh Nguyen wrote:
>>>>>>>>> Hi Mahesh,
>>>>>>>>>
>>>>>>>>> Maybe it was broken when transmitting. I zipped to a tar file.
>> Please
>>>>>>> try it
>>>>>>>>> one more.
>>>>>>>>>
>>>>>>>>> Regards, Vu
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>> -----Original Message-----
>>>>>>>>>> From: A V Mahesh [mailto:mahesh.va...@oracle.com]
>>>>>>>>>> Sent: Thursday, February 23, 2017 12:54 PM
>>>>>>>>>> To: Vu Minh Nguyen <vu.m.ngu...@dektech.com.au>;
>>>>>>>>>> lennart.l...@ericsson.com; canh.v.tru...@dektech.com.au
>>>>>>>>>> Cc: opensaf-devel@lists.sourceforge.net
>>>>>>>>>> Subject: Re: [devel] [PATCH 0 of 3] Review Request for log: add
>>>>>>>>> alternative
>>>>>>>>>> destinations of log records [#2258] V4
>>>>>>>>>>
>>>>>>>>>> Hi Vu,
>>>>>>>>>>
>>>>>>>>>> On 2/23/2017 10:20 AM, Vu Minh Nguyen wrote:
>>>>>>>>>>> Hi Mahesh,
>>>>>>>>>>>
>>>>>>>>>>> Can you try with 2258_v7.patch I just sent to you?
>>>>>>>>>> I striped    changeset:   8610 of today's latest staging (  `hg
>>>> strip
>>>>>>>>>> 8610` which  removed log: implement SaLogFilterSetCallbackT
>> and
>>>>>>>>>> version
>>>>>>>>>> handling [#2146])
>>>>>>>>>> and try to apply your `2258_v7.patch`, it says `malformed patch
> at
>>>>>>>>>> line
>>>>>>>>>> 3324`.
>>>>>>>>>>
>>>>>>>>>> -AVM
>>>>>>>>>>> I have pulled the latest code on OpenSAF 5.1 branch, re-created
>>> the
>>>>>>>>> cluster.
>>>>>>>>>>> And it works with the case old active SC-1 (OpenSAF 5.1) and
>> new
>>>>>>>> standby
>>>>>>>>>>> SC-2 (with 2258_v7.patch included in).
>>>>>>>>>>>
>>>>>>>>>>> To apply 2258_v7.patch, please do remove the just pushed ticket
>>>>>>>>>>> "log:
>>>>>>>>>>> implement SaLogFilterSetCallbackT and version handling
>> [#2146]" ,
>>>>>>>>>>> I have not rebased the code on that yet.
>>>>>>>>>>>
>>>>>>>>>>> Regards, Vu
>>>>>>>>>>>
>>>>>>>>>>>> -----Original Message-----
>>>>>>>>>>>> From: A V Mahesh [mailto:mahesh.va...@oracle.com]
>>>>>>>>>>>> Sent: Thursday, February 23, 2017 11:45 AM
>>>>>>>>>>>> To: Vu Minh Nguyen <vu.m.ngu...@dektech.com.au>;
>>>>>>>>>>>> lennart.l...@ericsson.com; canh.v.tru...@dektech.com.au
>>>>>>>>>>>> Cc: opensaf-devel@lists.sourceforge.net
>>>>>>>>>>>> Subject: Re: [devel] [PATCH 0 of 3] Review Request for log:
> add
>>>>>>>>>>> alternative
>>>>>>>>>>>> destinations of log records [#2258] V4
>>>>>>>>>>>>
>>>>>>>>>>>> Hi Vu/Lennart,
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> In broad WITHOUT the #2258 patch, the same code/setup
>>> working
>>>>> fine
>>>>>>>>>> with
>>>>>>>>>>>> 2 sc node (staging  changeset: 8609 ),
>>>>>>>>>>>> as soon as we apply `2258_v5.patch` V5 patch on staging
>>>>> (changeset:
>>>>>>>>>>>> 8609 ) that you have provided yesterday,
>>>>>>>>>>>> on one sc node and try to bring up that in to cluster
> (in-service
>>>>>>> test)
>>>>>>>>>>>> we are observing the issue of new node (with #2258 patch) not
>>>>>>>>>>>> joining
>>>>>>>>>>>> cluster.
>>>>>>>>>>>>
>>>>>>>>>>>>
>> ==============================================================
>> ====================================================
>>>>>>>>>>>> eb 23 10:01:59 SC-1 osafimmnd[15279]: NO Implementer
>>> (applier)
>>>>>>>>>>>> connected: 15 (@safAmfService2010f) <127, 2010f>
>>>>>>>>>>>> Feb 23 10:01:59 SC-1 osaflogd[15289]: NO
>>>>> LOGSV_DATA_GROUPNAME
>>>>>>>>>> not
>>>>>>>>>>>> found
>>>>>>>>>>>> Feb 23 10:01:59 SC-1 osaflogd[15289]: NO LOG root directory
>> is:
>>>>>>>>>>>> "/var/log/opensaf/saflog"
>>>>>>>>>>>> Feb 23 10:01:59 SC-1 osaflogd[15289]: NO LOG data group is:
>> ""
>>>>>>>>>>>> Feb 23 10:01:59 SC-1 osaflogd[15289]: NO
>>> LGS_MBCSV_VERSION =
>>>>> 7
>>>>>>>>>>>> Feb 23 10:01:59 SC-1 osafamfnd[15329]: NO Assigned
>>>>>>>>>>>> 'safSi=SC-2N,safApp=OpenSAF' STANDBY to 'safSu=SC-
>>>>>>>>>>>> 1,safSg=2N,safApp=OpenSAF'
>>>>>>>>>>>> Feb 23 10:01:59 SC-1 opensafd: OpenSAF(5.1.M0 - ) services
>>>>>>>> successfully
>>>>>>>>>>>> started
>>>>>>>>>>>> Feb 23 10:01:59 SC-1 osafamfnd[15329]: NO
>>>>>>>>>>>> 'safComp=LOG,safSu=SC-1,safSg=2N,safApp=OpenSAF' faulted
>>> due
>>>>> to
>>>>>>>>>>>> 'avaDown' : Recovery is 'nodeFailfast'
>>>>>>>>>>>> Feb 23 10:01:59 SC-1 osafamfnd[15329]: ER
>>>>>>>>>>>> safComp=LOG,safSu=SC-1,safSg=2N,safApp=OpenSAF Faulted
>>> due
>>>>>>>>>>>> to:avaDown
>>>>>>>>>>>> Recovery is:nodeFailfast
>>>>>>>>>>>> Feb 23 10:01:59 SC-1 osafamfnd[15329]: Rebooting OpenSAF
>>>>> NodeId =
>>>>>>>>>>>> 131343
>>>>>>>>>>>> EE Name = , Reason: Component faulted: recovery is node
>>> failfast,
>>>>>>>>>>>> OwnNodeId = 131343, SupervisionTime = 60
>>>>>>>>>>>> Feb 23 10:01:59 SC-1 opensaf_reboot: Rebooting local node;
>>>>>>>> timeout=60
>>>>>>>>>>>> Feb 23 10:02:00 SC-1 osafimmnd[15279]: NO Implementer
>>> (applier)
>>>>>>>>>>>> connected: 16 (@OpenSafImmReplicatorB) <144, 2010f>
>>>>>>>>>>>> Feb 23 10:01:59 SC-1 opensaf_reboot: Rebooting local node;
>>>>>>>> timeout=60
>>>>>>>>
>> ==============================================================
>> ====================================================
>>>>>>>>>>>> So it is evident that in-service upgrade part code of this
> need
>>>>>>>>>>>> to be
>>>>>>>>>>>> corrected.
>>>>>>>>>>>>
>>>>>>>>>>>> Please see my comments as [AVM] and let me know if you need
>>>>> some
>>>>>>>>>> traces
>>>>>>>>>>>> .
>>>>>>>>>>>>
>>>>>>>>>>>> If you're planing to prepare new V6 patch , please do prepare
>> on
>>>>>>>>>>>> top
>>>>>>> of
>>>>>>>>>>>> today's latest staging.
>>>>>>>>>>>>
>>>>>>>>>>>> On 2/23/2017 9:33 AM, Vu Minh Nguyen wrote:
>>>>>>>>>>>>> Hi Mahesh,
>>>>>>>>>>>>>
>>>>>>>>>>>>> I have done in-service upgrade/downgrade with following
>>> cases:
>>>>>>>>>>>>> 1) New Active SC-1 (OpenSAF 5.2 with the attached patch) +
>> old
>>>>>>>> standby
>>>>>>>>>>>> SC-2
>>>>>>>>>>>>> (OpenSAF 5.1)
>>>>>>>>>>>>> --> Work fine
>>>>>>>>>>>> [AVM] This is not a practical use cause of in-service upgrade
> ,
>> we
>>>>>>> can
>>>>>>>>>>>> ignore this test further
>>>>>>>>>>>>> 2) Old Active SC-1 (OpenSAF 5.1) + new standby SC-2 (with or
>>>>>>>>>>>>> without
>>>>>>>>>>>>> attached patch)
>>>>>>>>>>>>> --> SC-2 is restarted & not able to join the cluster.
>>>>>>>>>>>> [AVM] This use cause/flow is  we do get in in-service upgrade
> ,
>>>>>>>>>>>> so we
>>>>>>>>>>>> need to address this.
>>>>>>>>>>>>> I got following messages in syslog:
>>>>>>>>>>>>> Feb 23 09:32:42 SC-2 user.notice opensafd: OpenSAF(5.2.M0 -
>>>>>>>>>>>>> 8529:b5addd36e45d:default) services successfully started
>>>>>>>>>>>>> Feb 23 09:32:43 SC-2 local0.warn osafntfimcnd[701]: WA
>>>>>>>>>>>> ntfimcn_imm_init
>>>>>>>>>>>>> saImmOiImplementerSet() returned SA_AIS_ERR_TIMEOUT (5)
>>>>>>>>>>>>> Feb 23 09:32:45 SC-2 local0.warn osafntfimcnd[701]: WA
>>>>>>>>>>>> ntfimcn_imm_init
>>>>>>>>>>>>> saImmOiImplementerSet() returned SA_AIS_ERR_TIMEOUT (5)
>>>>>>>>>>>>> Feb 23 09:32:47 SC-2 local0.warn osafntfimcnd[701]: WA
>>>>>>>>>>>> ntfimcn_imm_init
>>>>>>>>>>>>> saImmOiImplementerSet() returned SA_AIS_ERR_TIMEOUT (5)
>>>>>>>>>>>>> Feb 23 09:32:49 SC-2 local0.warn osafntfimcnd[701]: WA
>>>>>>>>>>>> ntfimcn_imm_init
>>>>>>>>>>>>> saImmOiImplementerSet() returned SA_AIS_ERR_TIMEOUT (5)
>>>>>>>>>>>>> Feb 23 09:32:50 SC-2 local0.err osafmsgnd[592]: ER
>>>>>>>>>>>> saImmOiImplementerSet
>>>>>>>>>>>>> FAILED:5
>>>>>>>>>>>>> Feb 23 09:32:50 SC-2 local0.err osafmsgnd[592]: ER
>>>>>>>>>>>> saImmOiImplementerSet
>>>>>>>>>>>>> FAILED:5
>>>>>>>>>>>>> Feb 23 09:32:50 SC-2 local0.notice osafamfnd[496]: NO
>>>>>>>>>>>>> 'safSu=SC-2,safSg=NoRed,safApp=OpenSAF' component
>> restart
>>>>>>>>>> probation
>>>>>>>>>>>> timer
>>>>>>>>>>>>> started (timeout: 60000000000 ns)
>>>>>>>>>>>>> Feb 23 09:32:50 SC-2 local0.notice osafamfnd[496]: NO
>>> Restarting
>>>>> a
>>>>>>>>>>>> component
>>>>>>>>>>>>> of 'safSu=SC-2,safSg=NoRed,safApp=OpenSAF' (comp restart
>>>>> count: 1)
>>>>>>>>>>>>> Feb 23 09:32:50 SC-2 local0.notice osafamfnd[496]: NO
>>>>>>>>>>>>> 'safComp=MQND,safSu=SC-2,safSg=NoRed,safApp=OpenSAF'
>>>>> faulted
>>>>>>>>>> due
>>>>>>>>>>>> to
>>>>>>>>>>>>> 'avaDown' : Recovery is 'componentRestart'
>>>>>>>>>>>>> Feb 23 09:32:50 SC-2 local0.info osafmsgnd[736]: mkfifo
>>> already
>>>>>>>>> exists:
>>>>>>>>>>>>> /var/lib/opensaf/osafmsgnd.fifo File exists
>>>>>>>>>>>>>
>>>>>>>>>>>>> And sometimes, on active SC-1 (OpenSAF 5.1), the node is not
>>>>>>>>>>>>> able to
>>>>>>>>> up
>>>>>>>>>>>>> because of following error:
>>>>>>>>>>>>>
>>>>>>>>>>>>> Feb 23 11:00:32 SC-1 local0.err osafclmna[406]: MDTM:TIPC
>>>>> Dsock
>>>>>>>>>> Socket
>>>>>>>>>>>>> creation failed in MDTM_INIT err :Address family not
>> supported
>>> by
>>>>>>>>>>>> protocol
>>>>>>>>>>>>> Feb 23 11:00:32 SC-1 local0.err osafclmna[406]: ER
>>>>>>>> ncs_agents_startup
>>>>>>>>>>>> FAILED
>>>>>>>>>>>> [AVM]  No such issues ( with both TCP & TIPC) (staging
>>> changeset:
>>>>>>>>> 8609
>>>>>>>>>>> )
>>>>>>>>>>>>> Are you getting similar problem at your side?
>>>>>>>>>>>>> Please note that, the problem is existed WITH or WITHOUT
>> the
>>>>> #2258
>>>>>>>>>>>> patch.
>>>>>>>>>>>> [AVM] No , problem only if we apply `2258_v5.patch` V5 patch
>> on
>>>>>>>> staging
>>>>>>>>>>>> (changeset:   8609 )
>>>>>>>>>>>>                  try to bring up that node in to cluster.
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> -AVM
>>>>>>>>>>>>
>>>>>>>>>>>>> I have informed this to IMM to have a look, not sure any
>>> problem
>>>>>>> with
>>>>>>>>>>> MDS
>>>>>>>>>>>>> layer or any problem with my environment setup.
>>>>>>>>>>>>> In the meantime, please have a look at the updated patch,  I
>> will
>>>>>>>>>>> continue
>>>>>>>>>>>>> checking the problem. Will keep you updated.
>>>>>>>>>>>> [AVM] I haven't seen any IMM problems
>>>>>>>>>>>>> Regards, Vu
>>>>>>>>>>>>>
>>>>>>>>>>>>>> -----Original Message-----
>>>>>>>>>>>>>> From: A V Mahesh [mailto:mahesh.va...@oracle.com]
>>>>>>>>>>>>>> Sent: Wednesday, February 22, 2017 5:36 PM
>>>>>>>>>>>>>> To: Vu Minh Nguyen <vu.m.ngu...@dektech.com.au>;
>>>>>>>>>>>>>> lennart.l...@ericsson.com; canh.v.tru...@dektech.com.au
>>>>>>>>>>>>>> Cc: opensaf-devel@lists.sourceforge.net
>>>>>>>>>>>>>> Subject: Re: [devel] [PATCH 0 of 3] Review Request for log:
>>> add
>>>>>>>>>>>>> alternative
>>>>>>>>>>>>>> destinations of log records [#2258] V4
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> Please see correction New Standby SC-1 ( with patch )
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> -AVM
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> On 2/22/2017 4:02 PM, A V Mahesh wrote:
>>>>>>>>>>>>>>> Hi Vu,
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> With this new patch , we have another issue :
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> 1)  standby Core by `/usr/lib64/opensaf/osaflogd'  issue
> got
>>>>>>>>> resolved
>>>>>>>>>>> .
>>>>>>>>>>>>>>> 2) In-service upgrade is Not working , I have Old  Active
> SC-
>> 2
>>>> (
>>>>>>>>> with
>>>>>>>>>>>>>>> out patch )  and New Standby SC-1 ( with patch )
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>           the new New Standby SC-1 not joining the cluster (
>>>>>>> in-service
>>>>>>>>>>>>>>> upgrade  failed )
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> New Standby SC-1
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>> ==============================================================
>>>>> ======================================================
>>>>>>>>>>>>>>> Feb 22 15:53:05 SC-1 osafamfnd[15752]: NO
>>>>>>>>>>>>>>> 'safSu=SC-1,safSg=NoRed,safApp=OpenSAF' Presence State
>>>>>>>>>>>>>> INSTANTIATING
>>>>>>>>>>>>>>> => INSTANTIATED
>>>>>>>>>>>>>>> Feb 22 15:53:05 SC-1 osafamfnd[15752]: NO Assigning
>>>>>>>>>>>>>>> 'safSi=NoRed4,safApp=OpenSAF' ACTIVE to
>>>>>>>>>>>>>>> 'safSu=SC-1,safSg=NoRed,safApp=OpenSAF'
>>>>>>>>>>>>>>> Feb 22 15:53:05 SC-1 osafamfnd[15752]: NO Assigned
>>>>>>>>>>>>>>> 'safSi=NoRed4,safApp=OpenSAF' ACTIVE to
>>>>>>>>>>>>>>> 'safSu=SC-1,safSg=NoRed,safApp=OpenSAF'
>>>>>>>>>>>>>>> Feb 22 15:53:05 SC-1 osafsmfd[15889]: Started
>>>>>>>>>>>>>>> Feb 22 15:53:05 SC-1 osafamfnd[15752]: NO
>>>>>>>>>>>>>>> 'safSu=SC-1,safSg=2N,safApp=OpenSAF' Presence State
>>>>>>>>>> INSTANTIATING
>>>>>>>>>>>> =>
>>>>>>>>>>>>>>> INSTANTIATED
>>>>>>>>>>>>>>> Feb 22 15:53:05 SC-1 osafamfnd[15752]: NO Assigning
>>>>>>>>>>>>>>> 'safSi=SC-2N,safApp=OpenSAF' STANDBY to
>>>>>>>>>>>>>>> 'safSu=SC-1,safSg=2N,safApp=OpenSAF'
>>>>>>>>>>>>>>> Feb 22 15:53:05 SC-1 osafrded[15672]: NO RDE role set to
>>>>>>>> STANDBY
>>>>>>>>>>>>>>> Feb 22 15:53:05 SC-1 osafrded[15672]: NO Peer up on node
>>>>>>>> 0x2020f
>>>>>>>>>>>>>>> Feb 22 15:53:05 SC-1 osafrded[15672]: NO Got peer info
>>>>> request
>>>>>>>> from
>>>>>>>>>>>>>>> node 0x2020f with role ACTIVE
>>>>>>>>>>>>>>> Feb 22 15:53:05 SC-1 osafrded[15672]: NO Got peer info
>>>>> response
>>>>>>>>>> from
>>>>>>>>>>>>>>> node 0x2020f with role ACTIVE
>>>>>>>>>>>>>>> Feb 22 15:53:05 SC-1 osafimmd[15691]: NO MDS event
>> from
>>>>> svc_id
>>>>>>>> 24
>>>>>>>>>>>>>>> (change:5, dest:13)
>>>>>>>>>>>>>>> Feb 22 15:53:05 SC-1 osafimmd[15691]: NO MDS event
>> from
>>>>> svc_id
>>>>>>>> 24
>>>>>>>>>>>>>>> (change:3, dest:13)
>>>>>>>>>>>>>>> Feb 22 15:53:05 SC-1 osafimmd[15691]: NO MDS event
>> from
>>>>> svc_id
>>>>>>>> 24
>>>>>>>>>>>>>>> (change:5, dest:13)
>>>>>>>>>>>>>>> Feb 22 15:53:05 SC-1 osafimmd[15691]: NO MDS event
>> from
>>>>> svc_id
>>>>>>>> 25
>>>>>>>>>>>>>>> (change:3, dest:567412424453430)
>>>>>>>>>>>>>>> Feb 22 15:53:05 SC-1 osafimmd[15691]: NO MDS event
>> from
>>>>> svc_id
>>>>>>>> 25
>>>>>>>>>>>>>>> (change:3, dest:565213401202663)
>>>>>>>>>>>>>>> Feb 22 15:53:05 SC-1 osafimmd[15691]: NO MDS event
>> from
>>>>> svc_id
>>>>>>>> 25
>>>>>>>>>>>>>>> (change:3, dest:566312912825221)
>>>>>>>>>>>>>>> Feb 22 15:53:05 SC-1 osafimmd[15691]: NO MDS event
>> from
>>>>> svc_id
>>>>>>>> 25
>>>>>>>>>>>>>>> (change:3, dest:564113889574230)
>>>>>>>>>>>>>>> Feb 22 15:53:05 SC-1 osafimmnd[15702]: NO Implementer
>>>>> (applier)
>>>>>>>>>>>>>>> connected: 17 (@safAmfService2010f) <127, 2010f>
>>>>>>>>>>>>>>> Feb 22 15:53:05 SC-1 osaflogd[15712]: NO
>>>>>>>>>> LOGSV_DATA_GROUPNAME
>>>>>>>>>>>>>> not found
>>>>>>>>>>>>>>> Feb 22 15:53:05 SC-1 osaflogd[15712]: NO LOG root
>>> directory
>>>>> is:
>>>>>>>>>>>>>>> "/var/log/opensaf/saflog"
>>>>>>>>>>>>>>> Feb 22 15:53:05 SC-1 osaflogd[15712]: NO LOG data group
>> is:
>>> ""
>>>>>>>>>>>>>>> Feb 22 15:53:05 SC-1 osaflogd[15712]: NO
>>>>> LGS_MBCSV_VERSION =
>>>>>>>> 7
>>>>>>>>>>>>>>> Feb 22 15:53:05 SC-1 osafamfnd[15752]: NO Assigned
>>>>>>>>>>>>>>> 'safSi=SC-2N,safApp=OpenSAF' STANDBY to
>>>>>>>>>>>>>>> 'safSu=SC-1,safSg=2N,safApp=OpenSAF'
>>>>>>>>>>>>>>> Feb 22 15:53:05 SC-1 opensafd: OpenSAF(5.1.M0 - ) services
>>>>>>>>>>>>>>> successfully started
>>>>>>>>>>>>>>> Feb 22 15:53:05 SC-1 osafamfnd[15752]: NO
>>>>>>>>>>>>>>> 'safComp=LOG,safSu=SC-1,safSg=2N,safApp=OpenSAF'
>>> faulted
>>>>> due
>>>>>>>> to
>>>>>>>>>>>>>>> 'avaDown' : Recovery is 'nodeFailfast'
>>>>>>>>>>>>>>> Feb 22 15:53:05 SC-1 osafamfnd[15752]: ER
>>>>>>>>>>>>>>> safComp=LOG,safSu=SC-1,safSg=2N,safApp=OpenSAF
>>> Faulted
>>>>> due
>>>>>>>>>>>>>> to:avaDown
>>>>>>>>>>>>>>> Recovery is:nodeFailfast
>>>>>>>>>>>>>>> Feb 22 15:53:05 SC-1 osafamfnd[15752]: Rebooting
>> OpenSAF
>>>>>>>> NodeId
>>>>>>>>>> =
>>>>>>>>>>>>>>> 131343 EE Name = , Reason: Component faulted: recovery
>> is
>>>>> node
>>>>>>>>>>>>>>> failfast, OwnNodeId = 131343, SupervisionTime = 60
>>>>>>>>>>>>>>> Feb 22 15:53:05 SC-1 opensaf_reboot: Rebooting local
>> node;
>>>>>>>>>>>> timeout=60
>>>>>>>>>>>>>>> Feb 22 15:53:43 SC-1 syslog-ng[1171]: syslog-ng starting
>> up;
>>>>>>>>>>>>>>> version='2.0.9'
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>> ==============================================================
>>>>> ======================================================
>>>>>>>>>>>>>>> Old - Active - SC-2
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>> ==============================================================
>>>>> ======================================================
>>>>>>>>>>>>>>> Feb 22 15:53:02 SC-2 osafimmnd[16359]: NO NODE STATE-
>>>>>>>>>>>>>>> IMM_NODE_R_AVAILABLE
>>>>>>>>>>>>>>> Feb 22 15:53:02 SC-2 osafimmloadd: NO Sync starting
>>>>>>>>>>>>>>> Feb 22 15:53:02 SC-2 osafimmloadd: IN Synced 390 objects
>> in
>>>>>>>>>>>>>>> total
>>>>>>>>>>>>>>> Feb 22 15:53:02 SC-2 osafimmnd[16359]: NO NODE STATE-
>>>>>>>>>>>>>>> IMM_NODE_FULLY_AVAILABLE 18511
>>>>>>>>>>>>>>> Feb 22 15:53:02 SC-2 osafimmnd[16359]: NO Epoch set to 3
>>> in
>>>>>>>>>>>> ImmModel
>>>>>>>>>>>>>>> Feb 22 15:53:02 SC-2 osafimmd[16346]: NO ACT: New
>> Epoch
>>>>> for
>>>>>>>>>>>> IMMND
>>>>>>>>>>>>>>> process at node 2020f old epoch: 2  new epoch:3
>>>>>>>>>>>>>>> Feb 22 15:53:02 SC-2 osafimmd[16346]: NO ACT: New
>> Epoch
>>>>> for
>>>>>>>>>>>> IMMND
>>>>>>>>>>>>>>> process at node 2040f old epoch: 2  new epoch:3
>>>>>>>>>>>>>>> Feb 22 15:53:02 SC-2 osafimmd[16346]: NO ACT: New
>> Epoch
>>>>> for
>>>>>>>>>>>> IMMND
>>>>>>>>>>>>>>> process at node 2030f old epoch: 2  new epoch:3
>>>>>>>>>>>>>>> Feb 22 15:53:02 SC-2 osafimmloadd: NO Sync ending
>>> normally
>>>>>>>>>>>>>>> Feb 22 15:53:02 SC-2 osafimmd[16346]: NO ACT: New
>> Epoch
>>>>> for
>>>>>>>>>>>> IMMND
>>>>>>>>>>>>>>> process at node 2010f old epoch: 0  new epoch:3
>>>>>>>>>>>>>>> Feb 22 15:53:02 SC-2 osafimmnd[16359]: NO SERVER
>> STATE:
>>>>>>>>>>>>>>> IMM_SERVER_SYNC_SERVER --> IMM_SERVER_READY
>>>>>>>>>>>>>>> Feb 22 15:53:03 SC-2 osafamfd[16408]: NO Received
>>> node_up
>>>>> from
>>>>>>>>>>>> 2010f:
>>>>>>>>>>>>>>> msg_id 1
>>>>>>>>>>>>>>> Feb 22 15:53:03 SC-2 osafamfd[16408]: NO Node 'SC-1'
>>> joined
>>>>> the
>>>>>>>>>>>> cluster
>>>>>>>>>>>>>>> Feb 22 15:53:03 SC-2 osafimmnd[16359]: NO Implementer
>>>>>>>> connected:
>>>>>>>>>>>> 16
>>>>>>>>>>>>>>> (MsgQueueService131343) <0, 2010f>
>>>>>>>>>>>>>>> Feb 22 15:53:03 SC-2 osafrded[16327]: NO Peer up on node
>>>>>>>> 0x2010f
>>>>>>>>>>>>>>> Feb 22 15:53:03 SC-2 osafrded[16327]: NO Got peer info
>>>>> request
>>>>>>>> from
>>>>>>>>>>>>>>> node 0x2010f with role STANDBY
>>>>>>>>>>>>>>> Feb 22 15:53:03 SC-2 osafrded[16327]: NO Got peer info
>>>>> response
>>>>>>>>>> from
>>>>>>>>>>>>>>> node 0x2010f with role STANDBY
>>>>>>>>>>>>>>> Feb 22 15:53:03 SC-2 osafimmd[16346]: NO MDS event
>> from
>>>>> svc_id
>>>>>>>> 24
>>>>>>>>>>>>>>> (change:5, dest:13)
>>>>>>>>>>>>>>> Feb 22 15:53:03 SC-2 osafimmnd[16359]: NO Implementer
>>>>> (applier)
>>>>>>>>>>>>>>> connected: 17 (@safAmfService2010f) <0, 2010f>
>>>>>>>>>>>>>>> Feb 22 15:53:03 SC-2 osafamfd[16408]: NO Cluster startup
>> is
>>>>> done
>>>>>>>>>>>>>>> Feb 22 15:53:04 SC-2 osafimmnd[16359]: NO Implementer
>>>>> (applier)
>>>>>>>>>>>>>>> connected: 18 (@OpenSafImmReplicatorB) <0, 2010f>
>>>>>>>>>>>>>>> Feb 22 15:53:09 SC-2 osafdtmd[16304]: NO Lost contact
>> with
>>>>>>>>>>>>>>> 'SC-1'
>>>>>>>>>>>>>>> Feb 22 15:53:09 SC-2 osaffmd[16336]: NO Node Down
>> event
>>>>> for
>>>>>>>> node
>>>>>>>>>> id
>>>>>>>>>>>>>>> 2010f:
>>>>>>>>>>>>>>> Feb 22 15:53:09 SC-2 osafimmd[16346]: NO MDS event
>> from
>>>>> svc_id
>>>>>>>> 24
>>>>>>>>>>>>>>> (change:6, dest:13)
>>>>>>>>>>>>>>> Feb 22 15:53:09 SC-2 osafimmd[16346]: NO MDS event
>> from
>>>>> svc_id
>>>>>>>> 25
>>>>>>>>>>>>>>> (change:4, dest:564113889574230)
>>>>>>>>>>>>>>> Feb 22 15:53:09 SC-2 osaffmd[16336]: NO Current role:
>>> ACTIVE
>>>>>>>>>>>>>>> Feb 22 15:53:09 SC-2 osaffmd[16336]: Rebooting OpenSAF
>>>>> NodeId
>>>>>>>> =
>>>>>>>>>>>>>> 131343
>>>>>>>>>>>>>>> EE Name = , Reason: Received Node Down for peer
>>> controller,
>>>>>>>>>>>> OwnNodeId
>>>>>>>>>>>>>>> = 131599, SupervisionTime = 60
>>>>>>>>>>>>>>> Feb 22 15:53:09 SC-2 osafrded[16327]: NO Peer down on
>>> node
>>>>>>>>>> 0x2010f
>>>>>>>>>>>>>>> Feb 22 15:53:09 SC-2 osafamfd[16408]: NO Node 'SC-1' left
>>> the
>>>>>>>>>> cluster
>>>>>>>>>>>>>>> Feb 22 15:53:09 SC-2 osaflogd[16372]: NO Failed (2) to
>> send
>>> of
>>>>>>>> WRITE
>>>>>>>>>>>>>>> ack to: 2010f00003d6a
>>>>>>>>>>>>>>> Feb 22 15:53:09 SC-2 osaflogd[16372]: NO Failed (2) to
>> send
>>> of
>>>>>>>> WRITE
>>>>>>>>>>>>>>> ack to: 2010f00003d6a
>>>>>>>>>>>>>>> Feb 22 15:53:09 SC-2 osaflogd[16372]: NO Failed (2) to
>> send
>>> of
>>>>>>>> WRITE
>>>>>>>>>>>>>>> ack to: 2010f00003d74
>>>>>>>>>>>>>>> Feb 22 15:53:09 SC-2 osafclmd[16398]: NO Node 131343
>>> went
>>>>>>>> down.
>>>>>>>>>>>> Not
>>>>>>>>>>>>>>> sending track callback for agents on that node
>>>>>>>>>>>>>>> Feb 22 15:53:09 SC-2 osafclmd[16398]: NO Node 131343
>>> went
>>>>>>>> down.
>>>>>>>>>>>> Not
>>>>>>>>>>>>>>> sending track callback for agents on that node
>>>>>>>>>>>>>>> Feb 22 15:53:09 SC-2 osafclmd[16398]: NO Node 131343
>>> went
>>>>>>>> down.
>>>>>>>>>>>> Not
>>>>>>>>>>>>>>> sending track callback for agents on that node
>>>>>>>>>>>>>>> Feb 22 15:53:09 SC-2 osafclmd[16398]: NO Node 131343
>>> went
>>>>>>>> down.
>>>>>>>>>>>> Not
>>>>>>>>>>>>>>> sending track callback for agents on that node
>>>>>>>>>>>>>>> Feb 22 15:53:09 SC-2 osafclmd[16398]: NO Node 131343
>>> went
>>>>>>>> down.
>>>>>>>>>>>> Not
>>>>>>>>>>>>>>> sending track callback for agents on that node
>>>>>>>>>>>>>>> Feb 22 15:53:09 SC-2 osafclmd[16398]: NO Node 131343
>>> went
>>>>>>>> down.
>>>>>>>>>>>> Not
>>>>>>>>>>>>>>> sending track callback for agents on that node
>>>>>>>>>>>>>>> Feb 22 15:53:09 SC-2 osafclmd[16398]: NO Node 131343
>>> went
>>>>>>>> down.
>>>>>>>>>>>> Not
>>>>>>>>>>>>>>> sending track callback for agents on that node
>>>>>>>>>>>>>>> Feb 22 15:53:09 SC-2 osafimmd[16346]: WA IMMD lost
>>> contact
>>>>>>>> with
>>>>>>>>>>>> peer
>>>>>>>>>>>>>>> IMMD (NCSMDS_RED_DOWN)
>>>>>>>>>>>>>>> Feb 22 15:53:09 SC-2 osafimmnd[16359]: NO Global discard
>>>>> node
>>>>>>>>>>>> received
>>>>>>>>>>>>>>> for nodeId:2010f pid:15702
>>>>>>>>>>>>>>> Feb 22 15:53:09 SC-2 osafimmnd[16359]: NO Implementer
>>>>>>>>>> disconnected
>>>>>>>>>>>> 16
>>>>>>>>>>>>>>> <0, 2010f(down)> (MsgQueueService131343)
>>>>>>>>>>>>>>> Feb 22 15:53:09 SC-2 osafimmnd[16359]: NO Implementer
>>>>>>>>>> disconnected
>>>>>>>>>>>> 17
>>>>>>>>>>>>>>> <0, 2010f(down)> (@safAmfService2010f)
>>>>>>>>>>>>>>> Feb 22 15:53:09 SC-2 osafimmnd[16359]: NO Implementer
>>>>>>>>>> disconnected
>>>>>>>>>>>> 18
>>>>>>>>>>>>>>> <0, 2010f(down)> (@OpenSafImmReplicatorB)
>>>>>>>>>>>>>>> Feb 22 15:53:09 SC-2 opensaf_reboot: Rebooting remote
>>> node
>>>>> in
>>>>>>>> the
>>>>>>>>>>>>>>> absence of PLM is outside the scope of OpenSAF
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>> ==============================================================
>>>>> ======================================================
>>>>>>>>>>>>>>> -AVM
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> On 2/22/2017 3:13 PM, Vu Minh Nguyen wrote:
>>>>>>>>>>>>>>>> Hi Mahesh,
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> I put all required patches into one. Try to use this and
> see
>>>> if
>>>>>>> you
>>>>>>>>>>>>>>>> still
>>>>>>>>>>>>>>>> have that problem or not.
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> Regards, Vu
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> -----Original Message-----
>>>>>>>>>>>>>>>>> From: A V Mahesh [mailto:mahesh.va...@oracle.com]
>>>>>>>>>>>>>>>>> Sent: Wednesday, February 22, 2017 3:35 PM
>>>>>>>>>>>>>>>>> To: Vu Minh Nguyen <vu.m.ngu...@dektech.com.au>;
>>>>>>>>>>>>>>>>> lennart.l...@ericsson.com;
>>> canh.v.tru...@dektech.com.au
>>>>>>>>>>>>>>>>> Cc: opensaf-devel@lists.sourceforge.net
>>>>>>>>>>>>>>>>> Subject: Re: [devel] [PATCH 0 of 3] Review Request for
>> log:
>>>>>>>>>>>>>>>>> add
>>>>>>>>>>>>>>>> alternative
>>>>>>>>>>>>>>>>> destinations of log records [#2258] V4
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> Hi Vu,
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> I used  new #3 and #4 patches ,  Can you please re-send
>> All
>>>>>>>>>>>>>>>>> the
>>>>>>>>>>> final
>>>>>>>>>>>>>>>>> patch in go,
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> which i need to apply on today`s  staging ( if possible
>>>>>>>>>>>>>>>>> publish
>>>>>>>>> the
>>>>>>>>>>>>>>>>> with
>>>>>>>>>>>>>>>>> new version )
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> -AVM
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> On 2/22/2017 1:52 PM, Vu Minh Nguyen wrote:
>>>>>>>>>>>>>>>>>> Hi Mahesh,
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> Core was generated by `/usr/lib64/opensaf/osaflogd'.
>>>>>>>>>>>>>>>>>>> Program terminated with signal 11, Segmentation
>> fault.
>>>>>>>>>>>>>>>>>>> #0  ckpt_proc_cfg_stream(lgs_cb*, void*) () at
>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc:2195
>>>>>>>>>>>>>>>>>>> 2195    src/log/logd/lgs_mbcsv.cc: No such file or
>>>>>>>>>>>>>>>>>>> directory.
>>>>>>>>>>>>>>>>>>>                  in src/log/logd/lgs_mbcsv.cc
>>>>>>>>>>>>>>>>>> Backtrace still points to old position (lgs_mbcsv:2195).
> I
>>>>>>> guess
>>>>>>>>>>> the
>>>>>>>>>>>>>>>>>> osaflogd binary has not been updated with the fixed
>>> patch.
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> Regards, Vu
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> -----Original Message-----
>>>>>>>>>>>>>>>>>>> From: A V Mahesh [mailto:mahesh.va...@oracle.com]
>>>>>>>>>>>>>>>>>>> Sent: Wednesday, February 22, 2017 3:18 PM
>>>>>>>>>>>>>>>>>>> To: Vu Minh Nguyen
>> <vu.m.ngu...@dektech.com.au>;
>>>>>>>>>>>>>>>>>>> lennart.l...@ericsson.com;
>>>>> canh.v.tru...@dektech.com.au
>>>>>>>>>>>>>>>>>>> Cc: opensaf-devel@lists.sourceforge.net
>>>>>>>>>>>>>>>>>>> Subject: Re: [devel] [PATCH 0 of 3] Review Request for
>>> log:
>>>>>>>> add
>>>>>>>>>>>>>>>>>> alternative
>>>>>>>>>>>>>>>>>>> destinations of log records [#2258] V4
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> Hi Vu,
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> SC-2 standby osaflogd core dumped still occurs ( Not
>>>>>>>>>>>>>>>>>>> resolved)
>>>>>>>> ,
>>>>>>>>>>> the
>>>>>>>>>>>>>>>> new
>>>>>>>>>>>>>>>>>>> patch only resolved the application (/usr/bin/logtest )
>>>>>>>>>>> Segmentation
>>>>>>>>>>>>>>>>>>> fault on SC-1 Active.
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>
>> ==============================================================
>> ==============================================================
>>>>>>>>>>>>>>>>>>> # gdb /usr/lib64/opensaf/osaflogd
>>>>>>>>>>>> core_1487751055.osaflogd.4594
>>>>>>>>>>>>>> GNU
>>>>>>>>>>>>>>>>>>> gdb
>>>>>>>>>>>>>>>>>>> (GDB) SUSE (7.3-0.6.1)
>>>>>>>>>>>>>>>>>>> Copyright (C) 2011 Free Software Foundation, Inc.
>>>>>>>>>>>>>>>>>>> .......
>>>>>>>>>>>>>>>>>>> Core was generated by `/usr/lib64/opensaf/osaflogd'.
>>>>>>>>>>>>>>>>>>> Program terminated with signal 11, Segmentation
>> fault.
>>>>>>>>>>>>>>>>>>> #0  ckpt_proc_cfg_stream(lgs_cb*, void*) () at
>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc:2195
>>>>>>>>>>>>>>>>>>> 2195    src/log/logd/lgs_mbcsv.cc: No such file or
>>>>>>>>>>>>>>>>>>> directory.
>>>>>>>>>>>>>>>>>>>                  in src/log/logd/lgs_mbcsv.cc
>>>>>>>>>>>>>>>>>>> (gdb) bt
>>>>>>>>>>>>>>>>>>> #0  ckpt_proc_cfg_stream(lgs_cb*, void*) () at
>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc:2195
>>>>>>>>>>>>>>>>>>> #1  0x00007f97b026f960 in
>>>>> ckpt_decode_log_struct(lgs_cb*,
>>>>>>>>>>>>>>>>>>> ncs_mbcsv_cb_arg*, void*, void*, unsigned int
>>>>>>>> (*)(edu_hdl_tag*,
>>>>>>>>>>>>>>>>>>> edu_tkn_tag*, void*, unsigned int*,
>> edu_buf_env_tag*,
>>>>>>>>>>>>>> EDP_OP_TYPE,
>>>>>>>>>>>>>>>>>>> EDU_ERR*)) () at src/log/logd/lgs_mbcsv.cc:950
>>>>>>>>>>>>>>>>>>> #2  0x00007f97b02710dc in
>>>>>>>> ckpt_decode_async_update(lgs_cb*,
>>>>>>>>>>>>>>>>>>> ncs_mbcsv_cb_arg*) () at
>>> src/log/logd/lgs_mbcsv.cc:1086
>>>>>>>>>>>>>>>>>>> #3  0x00007f97b0273941 in
>>>>>>>> mbcsv_callback(ncs_mbcsv_cb_arg*)
>>>>>>>>>> ()
>>>>>>>>>>>> at
>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc:880
>>>>>>>>>>>>>>>>>>> #4  0x00007f97af372596 in ncs_mbscv_rcv_decode ()
>>> from
>>>>>>>>>>>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0
>>>>>>>>>>>>>>>>>>> #5  0x00007f97af372766 in
>>> ncs_mbcsv_rcv_async_update
>>>>> ()
>>>>>>>> from
>>>>>>>>>>>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0
>>>>>>>>>>>>>>>>>>> #6  0x00007f97af379370 in mbcsv_process_events ()
>>> from
>>>>>>>>>>>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0
>>>>>>>>>>>>>>>>>>> #7  0x00007f97af3794db in mbcsv_hdl_dispatch_all ()
>>> from
>>>>>>>>>>>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0
>>>>>>>>>>>>>>>>>>> #8  0x00007f97af373ce2 in
>>>>> mbcsv_process_dispatch_request ()
>>>>>>>>>> at
>>>>>>>>>>>>>>>>>>> src/mbc/mbcsv_api.c:423
>>>>>>>>>>>>>>>>>>> #9  0x00007f97b027096e in
>>> lgs_mbcsv_dispatch(unsigned
>>>>> int)
>>>>>>>> ()
>>>>>>>>>> at
>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc:327
>>>>>>>>>>>>>>>>>>> #10 0x00007f97b024d9f2 in main () at
>>>>>>>>>> src/log/logd/lgs_main.cc:583
>>>>>>>>>>>>>>>>>>> (gdb) bt full
>>>>>>>>>>>>>>>>>>> #0  ckpt_proc_cfg_stream(lgs_cb*, void*) () at
>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc:2195
>>>>>>>>>>>>>>>>>>>                  ckpt_data_handler = {0x7f97b0270300
>>>>>>>>>>>>>>>>>>> <ckpt_proc_initialize_client(lgs_cb*, void*)>,
>>>>>>>>>>>>>>>>>>> 0x7f97b02701c0
>>>>>>>>>>>>>>>>>>> <ckpt_proc_finalize_client(lgs_cb*, void*)>,
>>>>>>>>>>>>>>>>>>>                    0x7f97b0270060
>>>>>>>>>>>>>>>>>>> <ckpt_proc_agent_down(lgs_cb*,
>>>>>>>>>>> void*)>,
>>>>>>>>>>>>>>>>>>> 0x7f97b02712f0 <ckpt_proc_log_write(lgs_cb*,
>> void*)>,
>>>>>>>>>>>>>> 0x7f97b0271ab0
>>>>>>>>>>>>>>>>>>> <ckpt_proc_open_stream(lgs_cb*, void*)>,
>>>>>>>>>>>>>>>>>>>                    0x7f97b026fe80
>>>>>>> <ckpt_proc_close_stream(lgs_cb*,
>>>>>>>>>>>>> void*)>,
>>>>>>>>>>>>>>>>>>> 0x7f97b0272380 <ckpt_proc_cfg_stream(lgs_cb*,
>>> void*)>,
>>>>>>>>>>>>>>>>> 0x7f97b0274800
>>>>>>>>>>>>>>>>>>> <ckpt_proc_lgs_cfg_v2(lgs_cb*, void*)>,
>>>>>>>>>>>>>>>>>>>                    0x7f97b0274e10
>>>>>>>>>>>>>>>>>>> <ckpt_proc_lgs_cfg_v3(lgs_cb*,
>>>>>>>>>>> void*)>,
>>>>>>>>>>>>>>>>>>> 0x7f97b02754f0 <ckpt_proc_lgs_cfg_v5(lgs_cb*,
>>> void*)>}
>>>>>>>>>>>>>>>>>>> #1  0x00007f97b026f960 in
>>>>> ckpt_decode_log_struct(lgs_cb*,
>>>>>>>>>>>>>>>>>>> ncs_mbcsv_cb_arg*, void*, void*, unsigned int
>>>>>>>> (*)(edu_hdl_tag*,
>>>>>>>>>>>>>>>>>>> edu_tkn_tag*, void*, unsigned int*,
>> edu_buf_env_tag*,
>>>>>>>>>>>>>> EDP_OP_TYPE,
>>>>>>>>>>>>>>>>>>> EDU_ERR*)) () at src/log/logd/lgs_mbcsv.cc:950
>>>>>>>>>>>>>>>>>>>                  ckpt_data_handler = {0x7f97b0270300
>>>>>>>>>>>>>>>>>>> <ckpt_proc_initialize_client(lgs_cb*, void*)>,
>>>>>>>>>>>>>>>>>>> 0x7f97b02701c0
>>>>>>>>>>>>>>>>>>> <ckpt_proc_finalize_client(lgs_cb*, void*)>,
>>>>>>>>>>>>>>>>>>>                    0x7f97b0270060
>>>>>>>>>>>>>>>>>>> <ckpt_proc_agent_down(lgs_cb*,
>>>>>>>>>>> void*)>,
>>>>>>>>>>>>>>>>>>> 0x7f97b02712f0 <ckpt_proc_log_write(lgs_cb*,
>> void*)>,
>>>>>>>>>>>>>> 0x7f97b0271ab0
>>>>>>>>>>>>>>>>>>> <ckpt_proc_open_stream(lgs_cb*, void*)>,
>>>>>>>>>>>>>>>>>>>                    0x7f97b026fe80
>>>>>>> <ckpt_proc_close_stream(lgs_cb*,
>>>>>>>>>>>>> void*)>,
>>>>>>>>>>>>>>>>>>> 0x7f97b0272380 <ckpt_proc_cfg_stream(lgs_cb*,
>>> void*)>,
>>>>>>>>>>>>>>>>> 0x7f97b0274800
>>>>>>>>>>>>>>>>>>> <ckpt_proc_lgs_cfg_v2(lgs_cb*, void*)>,
>>>>>>>>>>>>>>>>>>>                    0x7f97b0274e10
>>>>>>>>>>>>>>>>>>> <ckpt_proc_lgs_cfg_v3(lgs_cb*,
>>>>>>>>>>> void*)>,
>>>>>>>>>>>>>>>>>>> 0x7f97b02754f0 <ckpt_proc_lgs_cfg_v5(lgs_cb*,
>>> void*)>}
>>>>>>>>>>>>>>>>>>> #2  0x00007f97b02710dc in
>>>>>>>> ckpt_decode_async_update(lgs_cb*,
>>>>>>>>>>>>>>>>>>> ncs_mbcsv_cb_arg*) () at
>>> src/log/logd/lgs_mbcsv.cc:1086
>>>>>>>>>>>>>>>>>>>                  ckpt_data_handler = {0x7f97b0270300
>>>>>>>>>>>>>>>>>>> <ckpt_proc_initialize_client(lgs_cb*, void*)>,
>>>>>>>>>>>>>>>>>>> 0x7f97b02701c0
>>>>>>>>>>>>>>>>>>> <ckpt_proc_finalize_client(lgs_cb*, void*)>,
>>>>>>>>>>>>>>>>>>>                    0x7f97b0270060
>>>>>>>>>>>>>>>>>>> <ckpt_proc_agent_down(lgs_cb*,
>>>>>>>>>>> void*)>,
>>>>>>>>>>>>>>>>>>> 0x7f97b02712f0 <ckpt_proc_log_write(lgs_cb*,
>> void*)>,
>>>>>>>>>>>>>> 0x7f97b0271ab0
>>>>>>>>>>>>>>>>>>> <ckpt_proc_open_stream(lgs_cb*, void*)>,
>>>>>>>>>>>>>>>>>>>                    0x7f97b026fe80
>>>>>>> <ckpt_proc_close_stream(lgs_cb*,
>>>>>>>>>>>>> void*)>,
>>>>>>>>>>>>>>>>>>> 0x7f97b0272380 <ckpt_proc_cfg_stream(lgs_cb*,
>>> void*)>,
>>>>>>>>>>>>>>>>> 0x7f97b0274800
>>>>>>>>>>>>>>>>>>> <ckpt_proc_lgs_cfg_v2(lgs_cb*, void*)>,
>>>>>>>>>>>>>>>>>>>                    0x7f97b0274e10
>>>>>>>>>>>>>>>>>>> <ckpt_proc_lgs_cfg_v3(lgs_cb*,
>>>>>>>>>>> void*)>,
>>>>>>>>>>>>>>>>>>> 0x7f97b02754f0 <ckpt_proc_lgs_cfg_v5(lgs_cb*,
>>> void*)>}
>>>>>>>>>>>>>>>>>>> #3  0x00007f97b0273941 in
>>>>>>>> mbcsv_callback(ncs_mbcsv_cb_arg*)
>>>>>>>>>> ()
>>>>>>>>>>>> at
>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc:880
>>>>>>>>>>>>>>>>>>>                  ckpt_data_handler = {0x7f97b0270300
>>>>>>>>>>>>>>>>>>> <ckpt_proc_initialize_client(lgs_cb*, void*)>,
>>>>>>>>>>>>>>>>>>> 0x7f97b02701c0
>>>>>>>>>>>>>>>>>>> <ckpt_proc_finalize_client(lgs_cb*, void*)>,
>>>>>>>>>>>>>>>>>>>                    0x7f97b0270060
>>>>>>>>>>>>>>>>>>> <ckpt_proc_agent_down(lgs_cb*,
>>>>>>>>>>> void*)>,
>>>>>>>>>>>>>>>>>>> 0x7f97b02712f0 <ckpt_proc_log_write(lgs_cb*,
>> void*)>,
>>>>>>>>>>>>>> 0x7f97b0271ab0
>>>>>>>>>>>>>>>>>>> <ckpt_proc_open_stream(lgs_cb*, void*)>,
>>>>>>>>>>>>>>>>>>>                    0x7f97b026fe80
>>>>>>> <ckpt_proc_close_stream(lgs_cb*,
>>>>>>>>>>>>> void*)>,
>>>>>>>>>>>>>>>>>>> 0x7f97b0272380 <ckpt_proc_cfg_stream(lgs_cb*,
>>> void*)>,
>>>>>>>>>>>>>>>>> 0x7f97b0274800
>>>>>>>>>>>>>>>>>>> <ckpt_proc_lgs_cfg_v2(lgs_cb*, void*)>,
>>>>>>>>>>>>>>>>>>>                    0x7f97b0274e10
>>>>>>>>>>>>>>>>>>> <ckpt_proc_lgs_cfg_v3(lgs_cb*,
>>>>>>>>>>> void*)>,
>>>>>>>>>>>>>>>>>>> 0x7f97b02754f0 <ckpt_proc_lgs_cfg_v5(lgs_cb*,
>>> void*)>}
>>>>>>>>>>>>>>>>>>> #4  0x00007f97af372596 in ncs_mbscv_rcv_decode ()
>>> from
>>>>>>>>>>>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0
>>>>>>>>>>>>>>>>>>> mbcsv_init_process_req_func = {0x7f97af373630
>>>>>>>>>>>>>>>>>>> <mbcsv_process_initialize_request>, 0x7f97af374f10
>>>>>>>>>>>>>>>>>>> <mbcsv_process_get_sel_obj_request>,
>>>>>>>>>>>>>>>>>>>                    0x7f97af373be0
>>>>>>> <mbcsv_process_dispatch_request>,
>>>>>>>>>>>>>>>>>>> 0x7f97af373af0 <mbcsv_process_finalize_request>,
>>>>>>>>>>>> 0x7f97af373cf0
>>>>>>>>>>>>>>>>>>> <mbcsv_process_open_request>,
>>>>>>>>>>>>>>>>>>>                    0x7f97af374050
>>>>>>>>>>>>>>>>>>> <mbcsv_process_close_request>,
>>>>>>>>>>>>>>>> 0x7f97af3741e0
>>>>>>>>>>>>>>>>>>> <mbcsv_process_chg_role_request>, 0x7f97af3744c0
>>>>>>>>>>>>>>>>>>> <mbcsv_process_snd_ckpt_request>,
>>>>>>>>>>>>>>>>>>>                    0x7f97af3747d0
>>>>>>> <mbcsv_process_snd_ntfy_request>,
>>>>>>>>>>>>>>>>>>> 0x7f97af374970 <mbcsv_process_snd_data_req>,
>>>>>>>>>> 0x7f97af373930
>>>>>>>>>>>>>>>>>>> <mbcsv_process_get_request>,
>>>>>>>>>>>>>>>>>>>                    0x7f97af374bd0
>>>> <mbcsv_process_set_request>}
>>>>>>>>>>>>>>>>>>> #5  0x00007f97af372766 in
>>> ncs_mbcsv_rcv_async_update
>>>>> ()
>>>>>>>> from
>>>>>>>>>>>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0
>>>>>>>>>>>>>>>>>>> mbcsv_init_process_req_func = {0x7f97af373630
>>>>>>>>>>>>>>>>>>> <mbcsv_process_initialize_request>, 0x7f97af374f10
>>>>>>>>>>>>>>>>>>> <mbcsv_process_get_sel_obj_request>,
>>>>>>>>>>>>>>>>>>>                    0x7f97af373be0
>>>>>>> <mbcsv_process_dispatch_request>,
>>>>>>>>>>>>>>>>>>> 0x7f97af373af0 <mbcsv_process_finalize_request>,
>>>>>>>>>>>> 0x7f97af373cf0
>>>>>>>>>>>>>>>>>>> <mbcsv_process_open_request>,
>>>>>>>>>>>>>>>>>>>                    0x7f97af374050
>>>>>>>>>>>>>>>>>>> <mbcsv_process_close_request>,
>>>>>>>>>>>>>>>> 0x7f97af3741e0
>>>>>>>>>>>>>>>>>>> <mbcsv_process_chg_role_request>, 0x7f97af3744c0
>>>>>>>>>>>>>>>>>>> <mbcsv_process_snd_ckpt_request>,
>>>>>>>>>>>>>>>>>>>                    0x7f97af3747d0
>>>>>>> <mbcsv_process_snd_ntfy_request>,
>>>>>>>>>>>>>>>>>>> 0x7f97af374970 <mbcsv_process_snd_data_req>,
>>>>>>>>>> 0x7f97af373930
>>>>>>>>>>>>>>>>>>> <mbcsv_process_get_request>,
>>>>>>>>>>>>>>>>>>>                    0x7f97af374bd0
>>>> <mbcsv_process_set_request>}
>>>>>>>>>>>>>>>>>>> #6  0x00007f97af379370 in mbcsv_process_events ()
>>> from
>>>>>>>>>>>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0
>>>>>>>>>>>>>>>>>>> No symbol table info available.
>>>>>>>>>>>>>>>>>>> #7  0x00007f97af3794db in mbcsv_hdl_dispatch_all ()
>>> from
>>>>>>>>>>>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0
>>>>>>>>>>>>>>>>>>> No symbol table info available.
>>>>>>>>>>>>>>>>>>> #8  0x00007f97af373ce2 in
>>>>> mbcsv_process_dispatch_request ()
>>>>>>>>>> at
>>>>>>>>>>>>>>>>>>> src/mbc/mbcsv_api.c:423
>>>>>>>>>>>>>>>>>>> mbcsv_init_process_req_func = {0x7f97af373630
>>>>>>>>>>>>>>>>>>> <mbcsv_process_initialize_request>, 0x7f97af374f10
>>>>>>>>>>>>>>>>>>> <mbcsv_process_get_sel_obj_request>,
>>>>>>>>>>>>>>>>>>>                    0x7f97af373be0
>>>>>>> <mbcsv_process_dispatch_request>,
>>>>>>>>>>>>>>>>>>> 0x7f97af373af0 <mbcsv_process_finalize_request>,
>>>>>>>>>>>> 0x7f97af373cf0
>>>>>>>>>>>>>>>>>>> <mbcsv_process_open_request>,
>>>>>>>>>>>>>>>>>>>                    0x7f97af374050
>>>>>>>>>>>>>>>>>>> <mbcsv_process_close_request>,
>>>>>>>>>>>>>>>> 0x7f97af3741e0
>>>>>>>>>>>>>>>>>>> <mbcsv_process_chg_role_request>, 0x7f97af3744c0
>>>>>>>>>>>>>>>>>>> <mbcsv_process_snd_ckpt_request>,
>>>>>>>>>>>>>>>>>>>                    0x7f97af3747d0
>>>>>>> <mbcsv_process_snd_ntfy_request>,
>>>>>>>>>>>>>>>>>>> 0x7f97af374970 <mbcsv_process_snd_data_req>,
>>>>>>>>>> 0x7f97af373930
>>>>>>>>>>>>>>>>>>> <mbcsv_process_get_request>,
>>>>>>>>>>>>>>>>>>>                    0x7f97af374bd0
>>>> <mbcsv_process_set_request>}
>>>>>>>>>>>>>>>>>>> #9  0x00007f97b027096e in
>>> lgs_mbcsv_dispatch(unsigned
>>>>> int)
>>>>>>>> ()
>>>>>>>>>> at
>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc:327
>>>>>>>>>>>>>>>>>>>                  ckpt_data_handler = {0x7f97b0270300
>>>>>>>>>>>>>>>>>>> <ckpt_proc_initialize_client(lgs_cb*, void*)>,
>>>>>>>>>>>>>>>>>>> 0x7f97b02701c0
>>>>>>>>>>>>>>>>>>> <ckpt_proc_finalize_client(lgs_cb*, void*)>,
>>>>>>>>>>>>>>>>>>>                    0x7f97b0270060
>>>>>>>>>>>>>>>>>>> <ckpt_proc_agent_down(lgs_cb*,
>>>>>>>>>>> void*)>,
>>>>>>>>>>>>>>>>>>> 0x7f97b02712f0 <ckpt_proc_log_write(lgs_cb*,
>> void*)>,
>>>>>>>>>>>>>> 0x7f97b0271ab0
>>>>>>>>>>>>>>>>>>> <ckpt_proc_open_stream(lgs_cb*, void*)>,
>>>>>>>>>>>>>>>>>>>                    0x7f97b026fe80
>>>>>>> <ckpt_proc_close_stream(lgs_cb*,
>>>>>>>>>>>>> void*)>,
>>>>>>>>>>>>>>>>>>> 0x7f97b0272380 <ckpt_proc_cfg_stream(lgs_cb*,
>>> void*)>,
>>>>>>>>>>>>>>>>> 0x7f97b0274800
>>>>>>>>>>>>>>>>>>> <ckpt_proc_lgs_cfg_v2(lgs_cb*, void*)>,
>>>>>>>>>>>>>>>>>>>                    0x7f97b0274e10
>>>>>>>>>>>>>>>>>>> <ckpt_proc_lgs_cfg_v3(lgs_cb*,
>>>>>>>>>>> void*)>,
>>>>>>>>>>>>>>>>>>> 0x7f97b02754f0 <ckpt_proc_lgs_cfg_v5(lgs_cb*,
>>> void*)>}
>>>>>>>>>>>>>>>>>>> ---Type <return> to continue, or q <return> to quit---
>>>>>>>>>>>>>>>>>>> #10 0x00007f97b024d9f2 in main () at
>>>>>>>>>> src/log/logd/lgs_main.cc:583
>>>>>>>>>>>>>>>>>>> usr1_sel_obj = {raise_obj = -1, rmv_obj = -1}
>>>>>>>>>>>>>>>>>>>                  _lgs_cb = {mds_hdl = 65547, mds_role =
>>>>>>>>>>>> V_DEST_RL_STANDBY,
>>>>>>>>>>>>>>>>> vaddr
>>>>>>>>>>>>>>>>>>> = 11, log_version = {releaseCode = 65 'A',
>> majorVersion
>>> = 2
>>>>>>>>>>> '\002',
>>>>>>>>>>>>>>>>>>> minorVersion = 2 '\002'}, client_tree = {
>>>>>>>>>>>>>>>>>>>                      root_node = {bit = -1, left =
>>>>>>> 0x7f97b04cf1b0,
>>>>>>>>>>> right =
>>>>>>>>>>>>>>>>>>> 0x7f97b04a2418, key_info = 0x7f97b04b7bd0 ""},
>>> params =
>>>>>>>>>>>> {key_size
>>>>>>>>>>>>>>>>>>> = 4},
>>>>>>>>>>>>>>>>>>> n_nodes = 8}, comp_name = {_opaque = {46, 24947,
>>>>>>>>>>>>>>>>>>>                        17254, 28015, 15728, 20300,
> 11335,
>>>>>>>>>>>>>>>>>>> 24947,
>>>>>>>>> 21350,
>>>>>>>>>>>>>>>>>>> 15733,
>>>>>>>>>>>>>>>>>>> 17235, 12845, 29484, 26209, 26451, 12861, 11342,
>>> 24947,
>>>>>>>>>> 16742,
>>>>>>>>>>>>>> 28784,
>>>>>>>>>>>>>>>>>>> 20285, 25968, 21358, 17985,
>>>>>>>>>>>>>>>>>>>                        0 <repeats 105 times>}}, amf_hdl
> =
>>>>>>>>> 4288675841,
>>>>>>>>>>>>>>>>>>> amfSelectionObject = 15, amf_invocation_id = 0,
>>>>>>>> is_quiesced_set
>>>>>>>>>> =
>>>>>>>>>>>>>>>> false,
>>>>>>>>>>>>>>>>>>> immOiHandle = 554050912783, immSelectionObject =
>>> 21,
>>>>>>>>>>>>>>>>>>>                    clmSelectionObject = 17, clm_hdl =
>>>>>>>>>>>>>>>>>>> 4279238657,
>>>>>>>>>>>>>>>>>>> ha_state =
>>>>>>>>>>>>>>>>>>> SA_AMF_HA_STANDBY, last_client_id = 208,
>>>>> async_upd_cnt =
>>>>>>>>>> 743,
>>>>>>>>>>>>>>>>>>> ckpt_state
>>>>>>>>>>>>>>>>>>> = COLD_SYNC_IDLE, mbcsv_hdl = 4293918753,
>>>>>>>>>>>>>>>>>>>                    mbcsv_sel_obj = 23, mbcsv_ckpt_hdl =
>>>>>>> 4292870177,
>>>>>>>>>>>>>>>>>>> mbcsv_peer_version = 7, edu_hdl = {is_inited = true,
>> tree
>>> =
>>>>>>>>>>>>>>>>>>> {root_node
>>>>>>>>>>>>>>>> =
>>>>>>>>>>>>>>>>>>> {bit = -1, left = 0x7f97b04cf2e0,
>>>>>>>>>>>>>>>>>>>                          right = 0x7f97b04a25b8,
> key_info =
>>>>>>>>>>> 0x7f97b04b7d40
>>>>>>>>>>>>>>>> ""},
>>>>>>>>>>>>>>>>>>> params = {key_size = 8}, n_nodes = 12}, to_version =
>> 1},
>>>>>>>>>>>>>>>>>>> fully_initialized = true, lga_down_list_head = 0x0,
>>>>>>>>>>>>>>>>>>>                    lga_down_list_tail = 0x0,
>>>>>>>>>>>>>>>>>>> clm_init_sel_obj =
>>>>>>>>>>>>>>>>>>> {raise_obj =
>>>>>>>>>>>>>>>> -1,
>>>>>>>>>>>>>>>>>>> rmv_obj = -1}, nid_started = true, scAbsenceAllowed =
>>> 900,
>>>>>>>>>>>>>>>>>>> lgs_recovery_state = LGS_NORMAL}
>>>>>>>>>>>>>>>>>>>                  nfds = 7
>>>>>>>>>>>>>>>>>>>                  fds = {{fd = 19, events = 1, revents =
> 0},
>>>>>>>>>>>>>>>>>>> {fd =
>>>>>>>>> 15,
>>>>>>>>>>>>>>>>>>> events =
>>>>>>>>>>>>>>>>>>> 1, revents = 0}, {fd = 23, events = 1, revents = 1},
> {fd
>>>>>>>>>>>>>>>>>>> = 13,
>>>>>>>>>>>>>>>>>>> events =
>>>>>>>>>>>>>>>>>>> 1, revents = 0}, {fd = -1, events = 1,
>>>>>>>>>>>>>>>>>>>                      revents = 0}, {fd = 17, events = 1,
>>>>>>>>>>>>>>>>>>> revents
>>>>>>> =
>>>>>>>>> 0},
>>>>>>>>>>>>>>>>>>> {fd =
>>>>>>>>>>>>>>>> 21,
>>>>>>>>>>>>>>>>>>> events = 1, revents = 0}}
>>>>>>>>>>>>>>>>>>>                  mbox_msgs = {0, 0, 0, 0, 0}
>>>>>>>>>>>>>>>>>>>                  lgs_cb = 0x7f97b04a2400
>>>>>>>>>>>>>>>>>>>                  mbox_low = {0, 0, 0, 0, 0}
>>>>>>>>>>>>>>>>>>>                  lgs_mbox_init_mutex = {__data = {__lock
> = 0,
>>>>>>>>> __count =
>>>>>>>>>>> 0,
>>>>>>>>>>>>>>>>>>> __owner = 0, __nusers = 0, __kind = 0, __spins = 0,
>> __list
>>>> =
>>>>>>>>>>>>>>>>>>> {__prev =
>>>>>>>>>>>>>>>>>>> 0x0, __next = 0x0}},
>>>>>>>>>>>>>>>>>>>                    __size = '\000' <repeats 39 times>,
>>>>>>>>>>>>>>>>>>> __align =
>>>>>>> 0}
>>>>>>>>>>>>>>>>>>> lgs_mbx = 4291821569
>>>>>>>>>>>>>>>>>>>                  mbox_high = {0, 0, 0, 0, 0}
>>>>>>>>>>>>>>>>>>>                  mbox_full = {false, false, false,
> false,
>>>>>>>>>>>>>>>>>>> false}
>>>>>>>>>>>>>>>>>>> (gdb)
>>>>>>>>>>>>>>>>>>> (gdb)
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> Feb 22 13:40:55 SC-2 osafimmnd[4584]: NO Ccb 131
>>>>>>>>>> COMMITTED
>>>>>>>>>>>>>>>>>>> (immcfg_SC-1_18714)
>>>>>>>>>>>>>>>>>>> Feb 22 13:40:56 SC-2 osafamfnd[4634]: NO
>>>>>>>>>>>>>>>>>>> 'safComp=LOG,safSu=SC-2,safSg=2N,safApp=OpenSAF'
>>>>> faulted
>>>>>>>>>> due
>>>>>>>>>>>> to
>>>>>>>>>>>>>>>>>>> 'avaDown' : Recovery is 'nodeFailfast'
>>>>>>>>>>>>>>>>>>> Feb 22 13:40:56 SC-2 osafamfnd[4634]: ER
>>>>>>>>>>>>>>>>>>> safComp=LOG,safSu=SC-2,safSg=2N,safApp=OpenSAF
>>>>> Faulted
>>>>>>>>>> due
>>>>>>>>>>>>>>>>>>> to:avaDown
>>>>>>>>>>>>>>>>>>> Recovery is:nodeFailfast
>>>>>>>>>>>>>>>>>>> Feb 22 13:40:56 SC-2 osafamfnd[4634]: Rebooting
>>>>> OpenSAF
>>>>>>>>>> NodeId
>>>>>>>>>>>> =
>>>>>>>>>>>>>>>>>>> 131599
>>>>>>>>>>>>>>>>>>> EE Name = , Reason: Component faulted: recovery is
>>> node
>>>>>>>>>> failfast,
>>>>>>>>>>>>>>>>>>> OwnNodeId = 131599, SupervisionTime = 60
>>>>>>>>>>>>>>>>>>> Feb 22 13:40:56 SC-2 opensaf_reboot: Rebooting local
>>>>> node;
>>>>>>>>>>>>>> timeout=60
>> ==============================================================
>> ==============================================================
>>>>>>>>>>>>>>>>>>> On 2/22/2017 12:23 PM, A V Mahesh wrote:
>>>>>>>>>>>>>>>>>>>> Hi Vu,
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>> On 2/22/2017 12:19 PM, Vu Minh Nguyen wrote:
>>>>>>>>>>>>>>>>>>>>> [Vu] I has sent you 02 patches. There is code change
>>> in
>>>>>>>>>> osaflogd
>>>>>>>>>>>>>>>>>>>>> code
>>>>>>>>>>>>>>>>>> that
>>>>>>>>>>>>>>>>>>>>> fix the coredump you have observed.
>>>>>>>>>>>>>>>>>>>>> The other one is test code that fix the logtest
>>> coredump.
>>>>>>>>>>>>>>>>>>>> Ok I will re-test , and update you .
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>> -AVM
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>> On 2/22/2017 12:19 PM, Vu Minh Nguyen wrote:
>>>>>>>>>>>>>>>>>>>>> Hi Mahehs,
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>> See my reply inline, [Vu].
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>> Regards, Vu
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>> -----Original Message-----
>>>>>>>>>>>>>>>>>>>>>> From: A V Mahesh
>>> [mailto:mahesh.va...@oracle.com]
>>>>>>>>>>>>>>>>>>>>>> Sent: Wednesday, February 22, 2017 1:36 PM
>>>>>>>>>>>>>>>>>>>>>> To: Vu Minh Nguyen
>>> <vu.m.ngu...@dektech.com.au>;
>>>>>>>>>>>>>>>>>>>>>> lennart.l...@ericsson.com;
>>>>>>>> canh.v.tru...@dektech.com.au
>>>>>>>>>>>>>>>>>>>>>> Cc: opensaf-devel@lists.sourceforge.net
>>>>>>>>>>>>>>>>>>>>>> Subject: Re: [PATCH 0 of 3] Review Request for log:
>>> add
>>>>>>>>>>>>>>>>>>>>>> alternative
>>>>>>>>>>>>>>>>>>>>>> destinations of log records [#2258] V4
>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>> Hi Vu,
>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>> On 2/22/2017 11:52 AM, Vu Minh Nguyen wrote:
>>>>>>>>>>>>>>>>>>>>>>> Hi Mahesh,
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> Have a code fault in uml test, and other one in
>>>>>>>> checkpoint.
>>>>>>>>>>>>>>>>>>>>>> [AVM] This is Normal Suse 11 VM ( not  UML).
>>>>>>>>>>>>>>>>>>>>>>> I have just updated the code. Please re-apply for
>> #3
>>>>> and
>>>>>>>> #4
>>>>>>>>>>>>>>>> patches.
>>>>>>>>>>>>>>>>>>>>>> [AVM] is these new patch has function changes or
>>> only
>>>>>>>>>>>>>>>>>>>>>> test
>>>>>>>>>>>> code
>>>>>>>>>>>>>>>>>>> changes ?
>>>>>>>>>>>>>>>>>>>>> [Vu] I has sent you 02 patches. There is code change
>>> in
>>>>>>>>>> osaflogd
>>>>>>>>>>>>>>>>>>>>> code
>>>>>>>>>>>>>>>>>> that
>>>>>>>>>>>>>>>>>>>>> fix the coredump you have observed.
>>>>>>>>>>>>>>>>>>>>> The other one is test code that fix the logtest
>>> coredump.
>>>>>>>>>>>>>>>>>>>>>>> Note that, test case #14 of suite 17 should be run
>>> on
>>>>>>>> active
>>>>>>>>>>>>>>>>>>>>>>> node,
>>>>>>>>>>>>>>>>>>>>>> otherwise
>>>>>>>>>>>>>>>>>>>>>>> getting failed.
>>>>>>>>>>>>>>>>>>>>>> [AVM]  Segmentation fault of /usr/bin/logtest Not
>> a
>>> big
>>>>>>>>> issue
>>>>>>>>>>> ,
>>>>>>>>>>>>>>>>>>>>>>             we need to debug why  osaflogd core
>> dumped
>>>>>>>>>>>>>>>>>>>>>> and it
>>>>>>> is
>>>>>>>>>>>>>>>>>>>>>> critical
>>>>>>>>>>>>>>>>>>>>> [Vu] I found the problem. You can try with the new
>>> one
>>>>> to
>>>>>>>> see
>>>>>>>>>> if
>>>>>>>>>>>>>>>>>>>>> the
>>>>>>>>>>>>>>>>>>>>> coredump is still there or not.
>>>>>>>>>>>>>>>>>>>>>>> I will put condition check to that test case later.
>>>>>>>>>>>>>>>>>>>>>> -AVM
>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> Regards, Vu
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>> -----Original Message-----
>>>>>>>>>>>>>>>>>>>>>>>> From: A V Mahesh
>>>>> [mailto:mahesh.va...@oracle.com]
>>>>>>>>>>>>>>>>>>>>>>>> Sent: Wednesday, February 22, 2017 12:16 PM
>>>>>>>>>>>>>>>>>>>>>>>> To: Vu Minh Nguyen
>>>>> <vu.m.ngu...@dektech.com.au>;
>>>>>>>>>>>>>>>>>>>>>>>> lennart.l...@ericsson.com;
>>>>>>>>>> canh.v.tru...@dektech.com.au
>>>>>>>>>>>>>>>>>>>>>>>> Cc: opensaf-devel@lists.sourceforge.net
>>>>>>>>>>>>>>>>>>>>>>>> Subject: Re: [PATCH 0 of 3] Review Request for
>>> log:
>>>>> add
>>>>>>>>>>>>>>>> alternative
>>>>>>>>>>>>>>>>>>>>>>>> destinations of log records [#2258] V4
>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>> Hi Vu,
>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>> Thanks ,
>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>> While testing /usr/bin/logtest ,  SC-2 standby
>>>>> osaflogd
>>>>>>>>>> core
>>>>>>>>>>>>>>>>> dumped
>>>>>>>>>>>>>>>>>>>>> and
>>>>>>>>>>>>>>>>>>>>>>>> /usr/bin/logtest on SC-1 Active
>>>>>>>>>>>>>>>>>>>>>>>> got Segmentation fault , am I missing any other
>>>>> patch (
>>>>>>> i
>>>>>>>>>> am
>>>>>>>>>>>>>>>> using
>>>>>>>>>>>>>>>>>>>>>>>> devel published patch only )
>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>> Following patches i am using :
>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>              1) #2293 (sent by Anders Widel, but
> not
>> yet
>>>>>>>>> pushed)
>>>>>>>>>>>>>>>>>>>>>>>> 2) #2258 (v2, sent by Lennart, but not yet
>>>>>>>>> pushed
>>>>>>>>>>> yet)
>>>>>>>>>>>>>>>>>>>>>>>> 3) #2258 (v4, sent by Vu, but not yet pushed
>>>>>>>>> yet)
>> ==============================================================
>> ========================================
>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>> Core was generated by
>>>>> `/usr/lib64/opensaf/osaflogd'.
>>>>>>>>>>>>>>>>>>>>>>>> Program terminated with signal 11,
>> Segmentation
>>>>> fault.
>>>>>>>>>>>>>>>>>>>>>>>> #0 ckpt_proc_cfg_stream(lgs_cb*, void*) () at
>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc:2195
>>>>>>>>>>>>>>>>>>>>>>>> 2195 src/log/logd/lgs_mbcsv.cc: No such file or
>>>>>>>>>> directory.
>>>>>>>>>>>>>>>>>>>>>>>> in src/log/logd/lgs_mbcsv.cc
>>>>>>>>>>>>>>>>>>>>>>>> (gdb) bt
>>>>>>>>>>>>>>>>>>>>>>>> #0 ckpt_proc_cfg_stream(lgs_cb*, void*) () at
>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc:2195
>>>>>>>>>>>>>>>>>>>>>>>> #1  0x00007f12c3e22960 in
>>>>>>>>>> ckpt_decode_log_struct(lgs_cb*,
>>>>>>>>>>>>>>>>>>>>>>>> ncs_mbcsv_cb_arg*, void*, void*, unsigned int
>>>>>>>>>>>>>> (*)(edu_hdl_tag*,
>>>>>>>>>>>>>>>>>>>>>>>> edu_tkn_tag*, void*, unsigned int*,
>>>>> edu_buf_env_tag*,
>>>>>>>>>>>>>>>>>>> EDP_OP_TYPE,
>>>>>>>>>>>>>>>>>>>>>>>> EDU_ERR*)) () at src/log/logd/lgs_mbcsv.cc:950
>>>>>>>>>>>>>>>>>>>>>>>> #2  0x00007f12c3e240dc in
>>>>>>>>>>>>>> ckpt_decode_async_update(lgs_cb*,
>>>>>>>>>>>>>>>>>>>>>>>> ncs_mbcsv_cb_arg*) () at
>>>>>>>> src/log/logd/lgs_mbcsv.cc:1086
>>>>>>>>>>>>>>>>>>>>>>>> #3 0x00007f12c3e26941 in
>>>>>>>>>>>>>> mbcsv_callback(ncs_mbcsv_cb_arg*) ()
>>>>>>>>>>>>>>>>> at
>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc:880
>>>>>>>>>>>>>>>>>>>>>>>> #4  0x00007f12c2f25596 in
>>> ncs_mbscv_rcv_decode ()
>>>>>>>> from
>>>>>>>>>>>>>>>>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0
>>>>>>>>>>>>>>>>>>>>>>>> #5  0x00007f12c2f25766 in
>>>>>>>> ncs_mbcsv_rcv_async_update
>>>>>>>>>> ()
>>>>>>>>>>>>>> from
>>>>>>>>>>>>>>>>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0
>>>>>>>>>>>>>>>>>>>>>>>> #6  0x00007f12c2f2c370 in
>> mbcsv_process_events
>>> ()
>>>>>>>> from
>>>>>>>>>>>>>>>>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0
>>>>>>>>>>>>>>>>>>>>>>>> #7  0x00007f12c2f2c4db in
>>> mbcsv_hdl_dispatch_all ()
>>>>>>>> from
>>>>>>>>>>>>>>>>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0
>>>>>>>>>>>>>>>>>>>>>>>> #8  0x00007f12c2f26ce2 in
>>>>>>>>>> mbcsv_process_dispatch_request
>>>>>>>>>>>> ()
>>>>>>>>>>>>>> at
>>>>>>>>>>>>>>>>>>>>>>>> src/mbc/mbcsv_api.c:423
>>>>>>>>>>>>>>>>>>>>>>>> #9  0x00007f12c3e2396e in
>>>>>>>> lgs_mbcsv_dispatch(unsigned
>>>>>>>>>> int)
>>>>>>>>>>>> ()
>>>>>>>>>>>>>> at
>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc:327
>>>>>>>>>>>>>>>>>>>>>>>> #10 0x00007f12c3e009f2 in main () at
>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_main.cc:583
>>>>>>>>>>>>>>>>>>>>>>>> (gdb)
>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>
>> ==============================================================
>> ========================================
>>>>>>>>>>>>>>>>>>>>>>>> Feb 22 10:37:06 SC-1 osafimmnd[4020]: NO
>>> Invalid
>>>>> error
>>>>>>>>>>>>>> reported
>>>>>>>>>>>>>>>>>>>>>>>> implementer 'safLogService', Ccb 161 will be
>>> aborted
>>>>>>>>>>>>>>>>>>>>>>>> Feb 22 10:37:06 SC-1 osafimmnd[4020]: NO Ccb
>>> 161
>>>>>>>>>> aborted
>>>>>>>>>>>> in
>>>>>>>>>>>>>>>>>>>>>> COMPLETED
>>>>>>>>>>>>>>>>>>>>>>>> processing (validation)
>>>>>>>>>>>>>>>>>>>>>>>> Feb 22 10:37:06 SC-1 osafimmnd[4020]: NO Ccb
>>> 161
>>>>>>>>>>>> ABORTED
>>>>>>>>>>>>>>>>>>>>>> (immcfg_SC-
>>>>>>>>>>>>>>>>>>>>>>>> 1_5394)
>>>>>>>>>>>>>>>>>>>>>>>> Add values Fail
>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>> Program received signal SIGSEGV, Segmentation
>>>>> fault.
>>>>>>>>>>>>>>>>>>>>>>>> 0x000055555556929a in
>> read_and_compare.isra.7
>>> ()
>>>>> at
>>>>>>>>>>>>>>>>>>>>>>>> src/log/apitest/tet_LogOiOps.c:1891
>>>>>>>>>>>>>>>>>>>>>>>> 1891 src/log/apitest/tet_LogOiOps.c: No such
>> file
>>> or
>>>>>>>>>>>>>>>>>>>>>>>> directory.
>>>>>>>>>>>>>>>>>>>>>>>>                     in
> src/log/apitest/tet_LogOiOps.c
>>>>>>>>>>>>>>>>>>>>>>>> (gdb) Feb 22 10:37:07 SC-1 sshd[5298]: Accepted
>>>>>>>> keyboard-
>>>>>>>>>>>>>>>>>>>>>> interactive/pam
>>>>>>>>>>>>>>>>>>>>>>>> for root from 10.176.178.22 port 51945 ssh2
>>>>>>>>>>>>>>>>>>>>>>>> bt
>>>>>>>>>>>>>>>>>>>>>>>> #0  0x000055555556929a in
>>>>> read_and_compare.isra.7 ()
>>>>>>>> at
>>>>>>>>>>>>>>>>>>>>>>>> src/log/apitest/tet_LogOiOps.c:1891
>>>>>>>>>>>>>>>>>>>>>>>> #1  0x0000555555569bbb in
>>>>>>>>>>>>>>>>>>>>>>>> check_logRecordDestinationConfigurationEmpty
>>>>>>>>>>>>>>>>>>>>>>>> () at src/log/apitest/tet_LogOiOps.c:2179
>>>>>>>>>>>>>>>>>>>>>>>> #2  0x0000555555573495 in run_test_case ()
>>>>>>>>>>>>>>>>>>>>>>>> #3  0x0000555555573934 in test_run ()
>>>>>>>>>>>>>>>>>>>>>>>> #4  0x000055555555c7cd in main () at
>>>>>>>>>>>>>>>>>>>>>>>> src/log/apitest/logtest.c:569
>>>>>>>>>>>>>>>>>>>>>>>> (gdb)
>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>
>> ==============================================================
>> ========================================
>>>>>>>>>>>>>>>>>>>>>>>> -AVM
>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>> On 2/22/2017 9:48 AM, Vu Minh Nguyen wrote:
>>>>>>>>>>>>>>>>>>>>>>>>> Hi Mahesh,
>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>> I send them in attachment instead, and name
>>> them
>>>>> in
>>>>>>>> the
>>>>>>>>>>>>>> order.
>>>>>>>>>>>>>>>>>>>>>>>>> I just pull the latest code, and apply them
>> without
>>>>>>>>> getting
>>>>>>>>>>>>> any
>>>>>>>>>>>>>>>>>> hunk
>>>>>>>>>>>>>>>>>>>>>>> error.
>>>>>>>>>>>>>>>>>>>>>>>>> Please try with them, and let me know if you
>> see
>>>>> any
>>>>>>>>>>>> problem.
>>>>>>>>>>>>>>>>>>>>>>>>> Regards, Vu
>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>> -----Original Message-----
>>>>>>>>>>>>>>>>>>>>>>>>>> From: A V Mahesh
>>>>> [mailto:mahesh.va...@oracle.com]
>>>>>>>>>>>>>>>>>>>>>>>>>> Sent: Wednesday, February 22, 2017 11:09
>> AM
>>>>>>>>>>>>>>>>>>>>>>>>>> To: Vu Minh Nguyen
>>>>>>>> <vu.m.ngu...@dektech.com.au>;
>>>>>>>>>>>>>>>>>>>>>>>>>> lennart.l...@ericsson.com;
>>>>>>>>>>>> canh.v.tru...@dektech.com.au
>>>>>>>>>>>>>>>>>>>>>>>>>> Cc: opensaf-devel@lists.sourceforge.net
>>>>>>>>>>>>>>>>>>>>>>>>>> Subject: Re: [PATCH 0 of 3] Review Request
>> for
>>>>> log:
>>>>>>>> add
>>>>>>>>>>>>>>>>>> alternative
>>>>>>>>>>>>>>>>>>>>>>>>>> destinations of log records [#2258] V4
>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>> Hi Vu,
>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>> I did follow that still i get Hunk #2 FAILED
> even
>>> on
>>>>>>>>>>> today's
>>>>>>>>>>>>>>>>>> staging
>> ==============================================================
>>>>>>>>>>>>>>>>>>>>>>>>>> ==================
>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>> [root@dhcp-hyd-scp-5fl-10-176-177-96
>>> staging]#
>>>>>>>> patch
>>>>>>>>>> -
>>>>>>>>>>>> p1
>>>>>>>>>>>>>>>>> <2293
>>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/base/Makefile.am
>>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #1 succeeded at 33 (offset 1 line).
>>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #3 succeeded at 183 (offset 1 line).
>>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/base/file_descriptor.cc
>>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/base/file_descriptor.h
>>>>>>>>>>>>>>>>>>>>>>>>>> patching file
>> src/base/tests/unix_socket_test.cc
>>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/base/unix_client_socket.cc
>>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/base/unix_server_socket.cc
>>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/base/unix_socket.cc
>>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/base/unix_socket.h
>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>> [root@dhcp-hyd-scp-5fl-10-176-177-96
>>> staging]#
>>>>>>>> patch
>>>>>>>>>> -
>>>>>>>>>>>> p1
>>>>>>>>>>>>>>>>> <2258-
>>>>>>>>>>>>>>>>>>> 1
>>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/log/Makefile.am
>>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #1 succeeded at 71 (offset -1 lines).
>>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/log/config/logsv_classes.xml
>>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #1 FAILED at 147.
>>>>>>>>>>>>>>>>>>>>>>>>>> 1 out of 1 hunk FAILED -- saving rejects to file
>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/config/logsv_classes.xml.rej
>>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_config.cc
>>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #1 succeeded at 35 (offset -5 lines).
>>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #2 FAILED at 705.
>>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #3 FAILED at 971.
>>>>>>>>>>>>>>>>>>>>>>>>>> 2 out of 3 hunks FAILED -- saving rejects to
> file
>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_config.cc.rej
>>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_config.h
>>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #1 FAILED at 304.
>>>>>>>>>>>>>>>>>>>>>>>>>> 1 out of 1 hunk FAILED -- saving rejects to file
>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_config.h.rej
>>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_dest.cc
>>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_dest.h
>>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_evt.cc
>>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_imm.cc
>>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #1 FAILED at 45.
>>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #2 succeeded at 235 (offset -1 lines).
>>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #3 FAILED at 877.
>>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #4 succeeded at 1273 (offset -20 lines).
>>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #5 succeeded at 1404 (offset -1 lines).
>>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #6 succeeded at 1449 (offset -20 lines).
>>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #7 succeeded at 2032 (offset -1 lines).
>>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #8 FAILED at 2181.
>>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #9 succeeded at 2271 (offset -54 lines).
>>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #10 succeeded at 2387 (offset -1 lines).
>>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #11 succeeded at 2377 (offset -54 lines).
>>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #12 succeeded at 2478 (offset -1 lines).
>>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #13 succeeded at 2684 (offset -54 lines).
>>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #14 succeeded at 2821 (offset -1 lines).
>>>>>>>>>>>>>>>>>>>>>>>>>> 3 out of 14 hunks FAILED -- saving rejects to
>> file
>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_imm.cc.rej
>>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_main.cc
>>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_mbcsv.cc
>>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_mbcsv.h
>>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_mbcsv_v5.cc
>>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #3 succeeded at 133 (offset -1 lines).
>>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_mbcsv_v7.cc
>>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_mbcsv_v7.h
>>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_stream.cc
>>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_stream.h
>>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_util.cc
>>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_util.h
>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>> [root@dhcp-hyd-scp-5fl-10-176-177-96
>>> staging]#
>>>>>>>> patch
>>>>>>>>>> -
>>>>>>>>>>>> p1
>>>>>>>>>>>>>>>>> <2258-
>>>>>>>>>>>>>>>>>>> 2
>>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/log/Makefile.am
>>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #1 succeeded at 180 (offset -3 lines).
>>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/log/apitest/tet_LogOiOps.c
>>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #1 FAILED at 1923.
>>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #2 FAILED at 1979.
>>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #3 FAILED at 2067.
>>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #4 FAILED at 2094.
>>>>>>>>>>>>>>>>>>>>>>>>>> 4 out of 4 hunks FAILED -- saving rejects to
> file
>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/apitest/tet_LogOiOps.c.rej
>>>>>>>>>>>>>>>>>>>>>>>>>> patching file
>>> src/log/apitest/tet_cfg_destination.c
>>>>>>>>>>>>>>>>>>>>>>>>>> [root@dhcp-hyd-scp-5fl-10-176-177-96
>>> staging]#
>>>>>>>> patch
>>>>>>>>>> -
>>>>>>>>>>>> p1
>>>>>>>>>>>>>>>>> <2258-
>>>>>>>>>>>>>>>>>>> 3
>>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/log/Makefile
>>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/log/Makefile.am
>>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #1 succeeded at 80 (offset -1 lines).
>>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #2 succeeded at 217 (offset -2 lines).
>>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/log/tests/Makefile
>>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/log/tests/lgs_dest_test.cc
>>>>>>>>>>>>>>>>>>>>>>>>>> [root@dhcp-hyd-scp-5fl-10-176-177-96
>>> staging]#
>>>>> vi
>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/apitest/tet_LogOiOps.c.rej
>>>>>>>>>>>>>>>>>>>>>>>>>> [root@dhcp-hyd-scp-5fl-10-176-177-96
>>> staging]#
>>>>>>>>>>>>>>>>>>>>>>>>>>
>> ==============================================================
>>>>>>>>>>>>>>>>>>>>>>>>>> ========================
>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>> -AVM
>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>> On 2/21/2017 3:53 PM, Vu Minh Nguyen
>> wrote:
>>>>>>>>>>>>>>>>>>>>>>>>>>> Hi Mahesh,
>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>> As I has mentioned in below:
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> To run the test, this patch has
>>>>>>> dependent
>>>>>>>>> on
>>>>>>>>>>>>>>>> following
>>>>>>>>>>>>>>>>>>>>> patches:
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> 1) #2293 (sent by Anders Widel, but not
>>>>>>> yet
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> pushed)
>>>>>>>>>>>>>>>>>>>>>>>>>>>>>                2) #2258 (v2, sent by Lennart,
> but
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> not
>>>>>>> yet
>>>>>>>>>>> pushed
>>>>>>>>>>>>>>>> yet)
>>>>>>>>>>>>>>>>>>>>>>>>>>> So, you need to apply #2293 first, then #2258
>>>>> which
>>>>>>>>>> sent
>>>>>>>>>>>> by
>>>>>>>>>>>>>>>>>>> Lennart
>>>>>>>>>>>>>>>>>>>>>>>>>>> yesterday, then mine.
>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>> Regards, Vu
>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>>> -----Original Message-----
>>>>>>>>>>>>>>>>>>>>>>>>>>>> From: A V Mahesh
>>>>>>>> [mailto:mahesh.va...@oracle.com]
>>>>>>>>>>>>>>>>>>>>>>>>>>>> Sent: Tuesday, February 21, 2017 5:10 PM
>>>>>>>>>>>>>>>>>>>>>>>>>>>> To: Vu Minh Nguyen
>>>>>>>>>> <vu.m.ngu...@dektech.com.au>;
>>>>>>>>>>>>>>>>>>>>>>>>>>>> lennart.l...@ericsson.com;
>>>>>>>>>>>>>> canh.v.tru...@dektech.com.au
>>>>>>>>>>>>>>>>>>>>>>>>>>>> Cc: opensaf-devel@lists.sourceforge.net
>>>>>>>>>>>>>>>>>>>>>>>>>>>> Subject: Re: [PATCH 0 of 3] Review Request
>>> for
>>>>> log:
>>>>>>>>>> add
>>>>>>>>>>>>>>>>>>> alternative
>>>>>>>>>>>>>>>>>>>>>>>>>>>> destinations of log records [#2258] V4
>>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>>> Hi Vu,
>>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>>> Is this applies on top of   log #2146  - V4 ,
> I
>>>> see
>>>>>>>>>>> both
>>>>>>>>>>>>>>>>>> #tickets
>>>>>>>>>>>>>>>>>>>>>>>>> has
>>>>>>>>>>>>>>>>>>>>>>>>>>>> version changes ?
>>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>>> in which order i need to apply  ( #2146 &
>>> #2258 )
>>>>>>>> or
>>>>>>>>>>>>>>>>>>>>>>>>>>>> (#2258
>>>>>>>>>>>>>>>> &
>>>>>>>>>>>>>>>>>>>>>>>>> #2146).
>> =========================================================
>>>>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/log/Makefile.am
>>>>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #1 FAILED at 72.
>>>>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #2 FAILED at 120.
>>>>>>>>>>>>>>>>>>>>>>>>>>>> 2 out of 2 hunks FAILED -- saving rejects to
>>> file
>>>>>>>>>>>>>>>>>>>>>>>>> src/log/Makefile.am.rej
>>>>>>>>>>>>>>>>>>>>>>>>>>>> patching file
>> src/log/config/logsv_classes.xml
>>>>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #1 FAILED at 147.
>>>>>>>>>>>>>>>>>>>>>>>>>>>> 1 out of 1 hunk FAILED -- saving rejects to
>> file
>>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/config/logsv_classes.xml.rej
>>>>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_config.cc
>>>>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #1 succeeded at 35 (offset -5 lines).
>>>>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #2 FAILED at 705.
>>>>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #3 FAILED at 971.
>>>>>>>>>>>>>>>>>>>>>>>>>>>> 2 out of 3 hunks FAILED -- saving rejects to
>>> file
>>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_config.cc.rej
>>>>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_config.h
>>>>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #1 FAILED at 304.
>>>>>>>>>>>>>>>>>>>>>>>>>>>> 1 out of 1 hunk FAILED -- saving rejects to
>> file
>>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_config.h.rej
>>>>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_dest.cc
>>>>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_dest.h
>>>>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_evt.cc
>>>>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #1 FAILED at 1.
>>>>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #2 succeeded at 30 with fuzz 2 (offset
>> 2
>>>>>>>>>>>>>>>>>>>>>>>>>>>> lines).
>>>>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #3 succeeded at 1282 (offset 45 lines).
>>>>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #4 succeeded at 1300 (offset 2 lines).
>>>>>>>>>>>>>>>>>>>>>>>>>>>> 1 out of 4 hunks FAILED -- saving rejects to
>>> file
>>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_evt.cc.rej
>>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>>>
>> ==============================================================
>>>>>>>>>>>>>>>>>>>>>>>>>>>> ===
>>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>>> -AVM
>>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>>> On 2/21/2017 3:03 PM, Vu Minh Nguyen
>>> wrote:
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> Summary: log: add alternative destinations
>>> of
>>>>> log
>>>>>>>>>>>> records
>>>>>>>>>>>>>>>>>>> [#2258]
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> Review request for Trac Ticket(s): #2258
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> Peer Reviewer(s): Lennart, Canh, Mahesh
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> Pull request to: <<LIST THE PERSON WITH
>>>>> PUSH
>>>>>>>>>>>> ACCESS
>>>>>>>>>>>>>>>>> HERE>>
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> Affected branch(es): Default
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> Development branch: Default
>>>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> --------------------------------
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> Impacted area Impact y/n
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> --------------------------------
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> Docs n
>>>>>>>>>>>>>>>>>>>>>>>>>>>>>                Build system            n
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> RPM/packaging n
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> Configuration files     n
>>>>>>>>>>>>>>>>>>>>>>>>>>>>>                Startup scripts         n
>>>>>>>>>>>>>>>>>>>>>>>>>>>>>                SAF services n
>>>>>>>>>>>>>>>>>>>>>>>>>>>>>                OpenSAF services        y
>>>>>>>>>>>>>>>>>>>>>>>>>>>>>                Core libraries          n
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> Samples n
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> Tests y
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> Other n
>>>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> Comments (indicate scope for each "y"
>>> above):
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> ---------------------------------------------
>>>>>>>>>>>>>>>>>>>>>>>>>>>>>                To run the test, this patch has
>>>>>>> dependent
>>>>>>>>> on
>>>>>>>>>>>>>>>> following
>>>>>>>>>>>>>>>>>>>>> patches:
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> 1) #2293 (sent by Anders Widel, but not
>>>>>>> yet
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> pushed)
>>>>>>>>>>>>>>>>>>>>>>>>>>>>>                2) #2258 (v2, sent by Lennart,
> but
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> not
>>>>>>> yet
>>>>>>>>>>> pushed
>>>>>>>>>>>>>>>> yet)
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> changeset
>>>>>>>>>>>>>> d74aaf3025c99cade3165a15831124548f4d85bd
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> Author: Vu Minh Nguyen
>>>>>>>>>>>>>> <vu.m.ngu...@dektech.com.au>
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> Date: Wed, 15 Feb 2017 14:36:00 +0700
>>>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>>>>           log: add alternative destinations of
>> log
>>>>>>> records
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> [#2258]
>>>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>>>>           Here are major info, detailed info
> will
>>>> be
>>>>>>>> added
>>>>>>>>> to
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> PR doc
>>>>>>>>>>>>>>>>>>>>>>> soon. 1)
>>>>>>>>>>>>>>>>>>>>>>>>>>>> Add
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> attribute "saLogRecordDestination" to log
>>>>>>>>>> stream.
>>>>>>>>>>> 2)
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> Add
>>>>>>>>>>>>>>>>>>>>>>> Local
>>>>>>>>>>>>>>>>>>>>>>>>>>>> socket
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> destintion handler 3) Integrate into first
>>>>>>>>>>> increment
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> made by
>>>>>>>>>>>>>>>>>>>>>>> Lennart
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> changeset
>>>>>>>>>>>>>> 4bae27a478c235df3058f43c92d3a5483233b01d
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> Author: Vu Minh Nguyen
>>>>>>>>>>>>>> <vu.m.ngu...@dektech.com.au>
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> Date: Wed, 15 Feb 2017 15:07:09 +0700
>>>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>>>>           log: add UML test case to verify
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> alternative
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> destination
>>>>>>>>>>>>>>>>>>>>>>> [#2258]
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> Major changes: 1) Modify Lennart's test
>>> cases
>>>>>>>>>>>> because
>>>>>>>>>>>>>>>>>>>>>>> enhancing
>>>>>>>>>>>>>>>>>>>>>>>>>>>> destination
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> configuration validation rules. 2) Add test
>>>>>>>>> suite
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> #17 to
>>>>>>>>>>>>>>>>>>>>>>> verify
>>>>>>>>>>>>>>>>>>>>>>>>>>>> alternative
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> destination
>>>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> changeset
>>>>>>>>>>>> bc375725fed22bb4f8cb3ae3df5f96fb9d281efb
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> Author: Vu Minh Nguyen
>>>>>>>>>>>>>> <vu.m.ngu...@dektech.com.au>
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> Date: Thu, 16 Feb 2017 17:22:13 +0700
>>>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>>>>           log: add unit tests to verify
> interfaces
>>>>>>>>> provided
>>>>>>>>>>> by
>>>>>>>>>>>>>>>>>>>>>>> destination
>>>>>>>>>>>>>>>>>>>>>>>>>>>> handler
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> [#2258]
>>>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>>>>           Unit tests to verify major
> interfaces:
>> 1)
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> CfgDestination()
>>>>>>>>>>>>>>>>>>>>>>> 2)
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> WriteToDestination()
>>>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> Added Files:
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> ------------
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/apitest/tet_cfg_destination.c
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_dest.cc
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_dest.h
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv_v7.cc
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv_v7.h
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/tests/lgs_dest_test.cc
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/tests/Makefile
>>>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> Complete diffstat:
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> ------------------
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/Makefile |    4 +
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/Makefile.am |   31 +++++-
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/apitest/tet_LogOiOps.c |    8 +-
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/apitest/tet_cfg_destination.c |
>> 483
>> ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++
>>>>>>>>>>>>>>>>>>>>>>>>>>>> ++++++++++++++++++++++++++++++++++
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/config/logsv_classes.xml |    7 +-
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_config.cc |  169
>>>>>>>>>>>>>>>>>>>>>>>>>>>> +++++++++++++++++++++++++++++---
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_config.h |    3 +-
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_dest.cc |  707
>> ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++
>> ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++
>>>>>>>>>>>>>>>>>>>>>>>>>>>> +++++++++++++++++
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_dest.h |  576
>> ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++
>>>>> ++++++++++++++++++++++++++++++++++++++++++++++++++++
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_evt.cc |   33 ++++++
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_imm.cc |  202
>>>>>>>>>>>>>>>>>>>>>>>>>>>> +++++++++++++++++++++++++++++++++----
>> --
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_main.cc |    8 +
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc |  103
>>>>>>>>>>>>>>>>>>>>> ++++++++++++++++++-
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv.h |    6 +-
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv_v5.cc |   10 +
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv_v7.cc |  177
>>>>>>>>>>>>>>>>>>>>>>>>>>>> +++++++++++++++++++++++++++++++++++
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv_v7.h |   67
>>>>>>>>>>>>>>>>>> +++++++++++++
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_stream.cc |   60
>>>>>>>>>>>>>>>>>> +++++++++++-
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_stream.h |   16 +++
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_util.cc |   63
>>>>>>>>>>>>>>>>>> ++++++++++++
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_util.h |   11 +-
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/tests/Makefile |   20 +++
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/tests/lgs_dest_test.cc |  209
>>>>> +++++++++++++++++++++++++++++++++++++++++
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> 23 files changed, 2896 insertions(+), 77
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> deletions(-)
>>>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> Testing Commands:
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> -----------------
>>>>>>>>>>>>>>>>>>>>>>>>>>>>>                Run UML test suite #17
>>>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> Testing, Expected Results:
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> --------------------------
>>>>>>>>>>>>>>>>>>>>>>>>>>>>>                All test passed
>>>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> Conditions of Submission:
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> -------------------------
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> <<HOW MANY DAYS BEFORE PUSHING,
>>>>>>>>>>>> CONSENSUS
>>>>>>>>>>>>>>>>> ETC>>
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> Arch Built     Started Linux distro
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> -------------------------------------------
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> mips n          n
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> mips64 n          n
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> x86 n          n
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> x86_64 n          n
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> powerpc n          n
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> powerpc64 n          n
>>>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> Reviewer Checklist:
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> -------------------
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> [Submitters: make sure that your review
>>>>> doesn't
>>>>>>>>>> trigger
>>>>>>>>>>>>>> any
>>>>>>>>>>>>>>>>>>>>>>>>>> checkmarks!]
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> Your checkin has not passed review
>> because
>>>>> (see
>>>>>>>>>>>> checked
>>>>>>>>>>>>>>>>>>> entries):
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> ___ Your RR template is generally
>>> incomplete;
>>>>> it
>>>>>>>> has
>>>>>>>>>>>> too
>>>>>>>>>>>>>>>>> many
>>>>>>>>>>>>>>>>>>>>> blank
>>>>>>>>>>>>>>>>>>>>>>>>>>> entries
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> that need proper data filled in.
>>>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> ___ You have failed to nominate the
>> proper
>>>>>>>> persons
>>>>>>>>>> for
>>>>>>>>>>>>>>>>> review
>>>>>>>>>>>>>>>>>>> and
>>>>>>>>>>>>>>>>>>>>>>>>> push.
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> ___ Your patches do not have proper
>>>>> short+long
>>>>>>>>>>>> header
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> ___ You have grammar/spelling in your
>>> header
>>>>>>>> that
>>>>>>>>>> is
>>>>>>>>>>>>>>>>>>> unacceptable.
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> ___ You have exceeded a sensible line
>>> length
>>>>> in
>>>>>>>> your
>>>>>>>>>>>>>>>>>>>>>>>>>>>> headers/comments/text.
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> ___ You have failed to put in a proper Trac
>>>>> Ticket
>>>>>>>> #
>>>>>>>>>>> into
>>>>>>>>>>>>>>>> your
>>>>>>>>>>>>>>>>>>>>>>>>> commits.
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> ___ You have incorrectly put/left internal
>>> data
>>>>> in
>>>>>>>>>> your
>>>>>>>>>>>>>>>>>>>>>>> comments/files
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> (i.e. internal bug tracking tool IDs, product
>>>>>>>>>>>>>>>> names
>>>>>>>>>>>>>>>>>> etc)
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> ___ You have not given any evidence of
>>> testing
>>>>>>>>>> beyond
>>>>>>>>>>>>>> basic
>>>>>>>>>>>>>>>>>>> build
>>>>>>>>>>>>>>>>>>>>>>>>> tests.
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> Demonstrate some level of runtime or
>> other
>>>>> sanity
>>>>>>>>>>>>>>>>>> testing.
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> ___ You have ^M present in some of your
>>> files.
>>>>>>>>>> These
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> have to
>>>>>>>>>>>>>>>>>>> be
>>>>>>>>>>>>>>>>>>>>>>>>>> removed.
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> ___ You have needlessly changed
>>> whitespace
>>>>> or
>>>>>>>>>> added
>>>>>>>>>>>>>>>>>>> whitespace
>>>>>>>>>>>>>>>>>>>>>>>>>> crimes
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> like trailing spaces, or spaces
>>>>>>> before
>>>>>>>>>>> tabs.
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> ___ You have mixed real technical changes
>>>>> with
>>>>>>>>>>>>>> whitespace
>>>>>>>>>>>>>>>>> and
>>>>>>>>>>>>>>>>>>>>>> other
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> cosmetic code cleanup changes. These
>> have
>>> to
>>>>> be
>>>>>>>>>>>>>>>>>>> separate
>>>>>>>>>>>>>>>>>>>>>>>>> commits.
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> ___ You need to refactor your submission
>>> into
>>>>>>>>>> logical
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> chunks;
>>>>>>>>>>>>>>>>>>>>> there
>>>>>>>>>>>>>>>>>>>>>>> is
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> too much content into a single
>>>>>>> commit.
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> ___ You have extraneous garbage in your
>>>>> review
>>>>>>>>>>>> (merge
>>>>>>>>>>>>>>>>>>> commits
>>>>>>>>>>>>>>>>>>>>>> etc)
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> ___ You have giant attachments which
>>> should
>>>>>>>> never
>>>>>>>>>>>> have
>>>>>>>>>>>>>>>>> been
>>>>>>>>>>>>>>>>>>>>>> sent;
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> Instead you should place your content in a
>>>>> public
>>>>>>>>>>>>>>>>>> tree to
>>>>>>>>>>>>>>>>>>>>> be
>>>>>>>>>>>>>>>>>>>>>>>>>>> pulled.
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> ___ You have too many commits attached
>>> to
>>>>> an e-
>>>>>>>>>> mail;
>>>>>>>>>>>>>>>>> resend
>>>>>>>>>>>>>>>>>>> as
>>>>>>>>>>>>>>>>>>>>>>>>>> threaded
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> commits, or place in a public tree for a
>> pull.
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> ___ You have resent this content multiple
>>>>> times
>>>>>>>>>>>> without a
>>>>>>>>>>>>>>>>> clear
>>>>>>>>>>>>>>>>>>>>>>>>>>> indication
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> of what has changed between each
>>>>>>>>> re-send.
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> ___ You have failed to adequately and
>>>>> individually
>>>>>>>>>>>> address
>>>>>>>>>>>>>>>> all
>>>>>>>>>>>>>>>>>> of
>>>>>>>>>>>>>>>>>>>>>>> the
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> comments and change requests that were
>>>>>>>> proposed
>>>>>>>>>> in
>>>>>>>>>>>>>>>>>> the
>>>>>>>>>>>>>>>>>>>>>>> initial
>>>>>>>>>>>>>>>>>>>>>>>>>>> review.
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> ___ You have a misconfigured ~/.hgrc file
>>> (i.e.
>>>>>>>>>>>> username,
>>>>>>>>>>>>>>>>> email
>>>>>>>>>>>>>>>>>>>>> etc)
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> ___ Your computer have a badly
>> configured
>>>>> date
>>>>>>>> and
>>>>>>>>>>>>>> time;
>>>>>>>>>>>>>>>>>>>>>> confusing
>>>>>>>>>>>>>>>>>>>>>>>> the
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> the threaded patch review.
>>>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> ___ Your changes affect IPC mechanism,
>>> and
>>>>> you
>>>>>>>>>> don't
>>>>>>>>>>>>>>>>> present
>>>>>>>>>>>>>>>>>>> any
>>>>>>>>>>>>>>>>>>>>>>>>>> results
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> for in-service upgradability test.
>>>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> ___ Your changes affect user manual and
>>>>>>>>>>>> documentation,
>>>>>>>>>>>>>>>>> your
>>>>>>>>>>>>>>>>>>>>>> patch
>>>>>>>>>>>>>>>>>>>>>>>>>> series
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> do not contain the patch that updates
>>>>>>>>> the
>>>>>>>>>>>>>>>>>>>>>>>>>>>>> Doxygen
>>>>>>>>>>>>>>>>>>> manual.
> ----------------------------------------------------------------------------
>>>>>>>>>>>>>>>>>> --
>>>>>>>>>>>>>>>>>>>> Check out the vibrant tech community on one of the
>>>>> world's
>>>>>>>>>> most
>>>>>>>>>>>>>>>>>>>> engaging tech sites, SlashDot.org!
>>>>> http://sdm.link/slashdot
>>> _______________________________________________
>>>>>>>>>>>>>>>>>>>> Opensaf-devel mailing list
>>>>>>>>>>>>>>>>>>>> Opensaf-devel@lists.sourceforge.net
>>>>>>>>>>>>>>>>>>>> https://lists.sourceforge.net/lists/listinfo/opensaf-
>>> devel


------------------------------------------------------------------------------
Check out the vibrant tech community on one of the world's most
engaging tech sites, SlashDot.org! http://sdm.link/slashdot
_______________________________________________
Opensaf-devel mailing list
Opensaf-devel@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/opensaf-devel

Reply via email to