Hi Vu,

On 2/23/2017 3:13 PM, A V Mahesh wrote:
>
> Not sure what are other change compare to V7 to V9 , New problems got 
> introduced
>
> Both nodes  SC-1 & SC-2 ( with 2258_v9.patch ) , trying bring up both 
> SC`s  simple node bringup  ,
>
> SC-2 going for reboot with following :
>
> ========================================================================================================================================
>  
>
>
> Feb 23 15:05:32 SC-2 osaflogd[29988]: NO LOGSV_DATA_GROUPNAME not found
> Feb 23 15:05:32 SC-2 osaflogd[29988]: NO LOG root directory is: 
> "/var/log/opensaf/saflog"
> Feb 23 15:05:32 SC-2 osaflogd[29988]: NO LOG data group is: ""
> Feb 23 15:05:32 SC-2 osafimmnd[29978]: NO Implementer (applier) 
> connected: 16 (@safAmfService2020f) <127, 2020f>
> Feb 23 15:05:32 SC-2 osaflogd[29988]: NO LGS_MBCSV_VERSION = 7
> Feb 23 15:05:32 SC-2 osaflogd[29988]: WA FAILED: 
> ncs_patricia_tree_add, client_id 0
> Feb 23 15:05:32 SC-2 osafamfnd[30028]: NO Assigned 
> 'safSi=SC-2N,safApp=OpenSAF' STANDBY to 
> 'safSu=SC-2,safSg=2N,safApp=OpenSAF'
> Feb 23 15:05:32 SC-2 osaflogd[29988]: ER Exiting with message: Could 
> not create new client
> Feb 23 15:05:32 SC-2 osafamfnd[30028]: NO 
> 'safSu=SC-2,safSg=2N,safApp=OpenSAF' component restart probation timer 
> started (timeout: 60000000000 ns)
> Feb 23 15:05:32 SC-2 osafamfnd[30028]: NO Restarting a component of 
> 'safSu=SC-2,safSg=2N,safApp=OpenSAF' (comp restart count: 1)
> Feb 23 15:05:32 SC-2 osafamfnd[30028]: NO 
> 'safComp=LOG,safSu=SC-2,safSg=2N,safApp=OpenSAF' faulted due to 
> 'errorReport' : Recovery is 'componentRestart'
> Feb 23 15:05:32 SC-2 opensafd[29908]: ER Service LOGD has unexpectedly 
> crashed. Unable to continue, exiting
> Feb 23 15:05:32 SC-2 osafamfd[30018]: exiting for shutdown
> Feb 23 15:05:32 SC-2 osafamfnd[30028]: ER AMFD has unexpectedly 
> crashed. Rebooting node
> Feb 23 15:05:32 SC-2 osafamfnd[30028]: Rebooting OpenSAF NodeId = 
> 131599 EE Name = , Reason: AMFD has unexpectedly crashed. Rebooting 
> node, OwnNodeId = 131599, SupervisionTime = 60
> Feb 23 15:05:32 SC-2 opensaf_reboot: Rebooting local node; timeout=60
> Feb 23 15:06:04 SC-2 syslog-ng[1180]: syslog-ng starting up; 
> version='2.0.9'
>
> ========================================================================================================================================
>  
>
Some times :

========================================================================================================================================
 

Feb 23 15:15:19 SC-2 osafrded[3858]: NO RDE role set to STANDBY
Feb 23 15:15:19 SC-2 osafrded[3858]: NO Peer up on node 0x2010f
Feb 23 15:15:19 SC-2 osafrded[3858]: NO Got peer info request from node 
0x2010f with role ACTIVE
Feb 23 15:15:19 SC-2 osafrded[3858]: NO Got peer info response from node 
0x2010f with role ACTIVE
Feb 23 15:15:19 SC-2 osafimmd[3877]: NO MDS event from svc_id 24 
(change:3, dest:13)
Feb 23 15:15:19 SC-2 osafimmd[3877]: NO MDS event from svc_id 24 
(change:5, dest:13)
Feb 23 15:15:19 SC-2 osafimmd[3877]: NO MDS event from svc_id 24 
(change:5, dest:13)
Feb 23 15:15:19 SC-2 osafimmd[3877]: NO MDS event from svc_id 25 
(change:3, dest:565217560625168)
Feb 23 15:15:19 SC-2 osafimmd[3877]: NO MDS event from svc_id 25 
(change:3, dest:564114674417680)
Feb 23 15:15:19 SC-2 osaflogd[3898]: NO LOGSV_DATA_GROUPNAME not found
Feb 23 15:15:19 SC-2 osaflogd[3898]: NO LOG root directory is: 
"/var/log/opensaf/saflog"
Feb 23 15:15:19 SC-2 osaflogd[3898]: NO LOG data group is: ""
Feb 23 15:15:19 SC-2 osafimmnd[3888]: NO Implementer (applier) 
connected: 15 (@safAmfService2020f) <127, 2020f>
Feb 23 15:15:19 SC-2 osaflogd[3898]: NO LGS_MBCSV_VERSION = 7
Feb 23 15:15:19 SC-2 osaflogd[3898]: ER Exiting with message: Client 
attributes differ
Feb 23 15:15:19 SC-2 osafamfnd[3938]: NO 
'safSu=SC-2,safSg=2N,safApp=OpenSAF' component restart probation timer 
started (timeout: 60000000000 ns)
Feb 23 15:15:19 SC-2 osafamfnd[3938]: NO Restarting a component of 
'safSu=SC-2,safSg=2N,safApp=OpenSAF' (comp restart count: 1)
Feb 23 15:15:19 SC-2 osafamfnd[3938]: NO 
'safComp=LOG,safSu=SC-2,safSg=2N,safApp=OpenSAF' faulted due to 
'errorReport' : Recovery is 'componentRestart'
Feb 23 15:15:19 SC-2 opensafd[3818]: ER Service LOGD has unexpectedly 
crashed. Unable to continue, exiting
Feb 23 15:15:20 SC-2 osafamfd[3928]: exiting for shutdown
Feb 23 15:15:20 SC-2 osafamfnd[3938]: ER AMFD has unexpectedly crashed. 
Rebooting node
Feb 23 15:15:20 SC-2 osafamfnd[3938]: Rebooting OpenSAF NodeId = 131599 
EE Name = , Reason: AMFD has unexpectedly crashed. Rebooting node, 
OwnNodeId = 131599, SupervisionTime = 60
Feb 23 15:15:20 SC-2 osafimmnd[3888]: NO Implementer locally 
disconnected. Marking it as doomed 15 <127, 2020f> (@safAmfService2020f)
Feb 23 15:15:20 SC-2 osafimmnd[3888]: NO Implementer disconnected 15 
<127, 2020f> (@safAmfService2020f)
Feb 23 15:15:20 SC-2 opensaf_reboot: Rebooting local node; timeout=60
========================================================================================================================================
 


>
> -AVM
>
>
> On 2/23/2017 2:20 PM, Vu Minh Nguyen wrote:
>> Hi Mahesh,
>>
>> This is the latest code has been rebased on the latest changeset.
>>
>> Note that, in the attached patch, I have included one more dependency,
>> that is on base::Hash() function, the patch sent by Anders [#2266]
>>
>> Please review the patch, then comment if any. Thanks.
>>
>> Regards, Vu
>>
>>> -----Original Message-----
>>> From: A V Mahesh [mailto:mahesh.va...@oracle.com]
>>> Sent: Thursday, February 23, 2017 2:03 PM
>>> To: Vu Minh Nguyen <vu.m.ngu...@dektech.com.au>;
>>> lennart.l...@ericsson.com; canh.v.tru...@dektech.com.au
>>> Cc: opensaf-devel@lists.sourceforge.net
>>> Subject: Re: [devel] [PATCH 0 of 3] Review Request for log: add
>> alternative
>>> destinations of log records [#2258] V4
>>>
>>> Hi Vu,
>>>
>>> Now we are now able to proceed further with  V7 `2258_v7.patch` in
>>> service upgraded working fine,
>>> because of Encode/decode changes done in V7 patch.
>>>
>>> But we have another small test case issue (/usr/bin/logtest 5 17
>>> Segmentation fault),
>>> once we resolve this also, we can conclude that all the basic
>>> functionality is working,
>>> then you can re-publish the V7 patch  ( if change occurred in Lennart
>>> #2258 V2 please do publish that as well )
>>> so that I can go for CODE review.
>>>
>>> Steps to reproduce the test case issue :
>>>
>>> 1) Bring up old node as Active  ( with out `2258_v7.patch` )
>>> 2) Bring-up new node as Standby  ( with `2258_v7.patch` )
>>> 3)  Do `amf-adm si-swap safSi=SC-2N,safApp=OpenSAF`
>>> 4)  Run `/usr/bin/logtest 5 17 ` on  new Active (because of si-swap )
>>>
>>> Note :  both nodes has the new XLM attributes populated .
>>>
>>> ==============================================================
>>> =====================
>>>
>>> gdb /usr/bin/logtest
>>> (gdb) r 5
>>>
>>>      16  PASSED   CCB Object Modify, change root directory. Path 
>>> exist. OK;
>>> Detaching after fork from child process 13797.
>>> Set values Fail
>>> [New Thread 0x7ffff7ff7b00 (LWP 13801)]
>>> [New Thread 0x7ffff7fc4b00 (LWP 13802)]
>>>
>>> Program received signal SIGSEGV, Segmentation fault.
>>> 0x00005555555688ea in read_and_compare.isra.7 () at
>>> src/log/apitest/tet_LogOiOps.c:1891
>>> 1891    src/log/apitest/tet_LogOiOps.c: No such file or directory.
>>>           in src/log/apitest/tet_LogOiOps.c
>>> (gdb) bt
>>> #0  0x00005555555688ea in read_and_compare.isra.7 () at
>>> src/log/apitest/tet_LogOiOps.c:1891
>>> #1  0x0000555555568a4b in check_logRecordDestinationConfigurationAdd ()
>>> at src/log/apitest/tet_LogOiOps.c:1941
>>> #2  0x0000555555571b05 in run_test_case ()
>>> #3  0x0000555555571feb in test_run ()
>>> #4  0x000055555555bfad in main () at src/log/apitest/logtest.c:569
>>> (gdb)
>>>
>>> ==============================================================
>>> =====================
>>>
>>>
>>> -AVM
>>>
>>> On 2/23/2017 11:44 AM, Vu Minh Nguyen wrote:
>>>> Hi Mahesh,
>>>>
>>>> Maybe it was broken when transmitting. I zipped to a tar file. Please
>> try it
>>>> one more.
>>>>
>>>> Regards, Vu
>>>>
>>>>
>>>>> -----Original Message-----
>>>>> From: A V Mahesh [mailto:mahesh.va...@oracle.com]
>>>>> Sent: Thursday, February 23, 2017 12:54 PM
>>>>> To: Vu Minh Nguyen <vu.m.ngu...@dektech.com.au>;
>>>>> lennart.l...@ericsson.com; canh.v.tru...@dektech.com.au
>>>>> Cc: opensaf-devel@lists.sourceforge.net
>>>>> Subject: Re: [devel] [PATCH 0 of 3] Review Request for log: add
>>>> alternative
>>>>> destinations of log records [#2258] V4
>>>>>
>>>>> Hi Vu,
>>>>>
>>>>> On 2/23/2017 10:20 AM, Vu Minh Nguyen wrote:
>>>>>> Hi Mahesh,
>>>>>>
>>>>>> Can you try with 2258_v7.patch I just sent to you?
>>>>> I striped    changeset:   8610 of today's latest staging (  `hg strip
>>>>> 8610` which  removed log: implement SaLogFilterSetCallbackT and 
>>>>> version
>>>>> handling [#2146])
>>>>> and try to apply your `2258_v7.patch`, it says `malformed patch at 
>>>>> line
>>>>> 3324`.
>>>>>
>>>>> -AVM
>>>>>> I have pulled the latest code on OpenSAF 5.1 branch, re-created the
>>>> cluster.
>>>>>> And it works with the case old active SC-1 (OpenSAF 5.1) and new
>>> standby
>>>>>> SC-2 (with 2258_v7.patch included in).
>>>>>>
>>>>>> To apply 2258_v7.patch, please do remove the just pushed ticket 
>>>>>> "log:
>>>>>> implement SaLogFilterSetCallbackT and version handling [#2146]" ,
>>>>>> I have not rebased the code on that yet.
>>>>>>
>>>>>> Regards, Vu
>>>>>>
>>>>>>> -----Original Message-----
>>>>>>> From: A V Mahesh [mailto:mahesh.va...@oracle.com]
>>>>>>> Sent: Thursday, February 23, 2017 11:45 AM
>>>>>>> To: Vu Minh Nguyen <vu.m.ngu...@dektech.com.au>;
>>>>>>> lennart.l...@ericsson.com; canh.v.tru...@dektech.com.au
>>>>>>> Cc: opensaf-devel@lists.sourceforge.net
>>>>>>> Subject: Re: [devel] [PATCH 0 of 3] Review Request for log: add
>>>>>> alternative
>>>>>>> destinations of log records [#2258] V4
>>>>>>>
>>>>>>> Hi Vu/Lennart,
>>>>>>>
>>>>>>>
>>>>>>> In broad WITHOUT the #2258 patch, the same code/setup working fine
>>>>> with
>>>>>>> 2 sc node (staging  changeset: 8609 ),
>>>>>>> as soon as we apply `2258_v5.patch` V5 patch on staging (changeset:
>>>>>>> 8609 ) that you have provided yesterday,
>>>>>>> on one sc node and try to bring up that in to cluster (in-service
>> test)
>>>>>>> we are observing the issue of new node (with #2258 patch) not 
>>>>>>> joining
>>>>>>> cluster.
>>>>>>>
>>>>>>>
>>> ==============================================================
>>>>>>> ====================================================
>>>>>>> eb 23 10:01:59 SC-1 osafimmnd[15279]: NO Implementer (applier)
>>>>>>> connected: 15 (@safAmfService2010f) <127, 2010f>
>>>>>>> Feb 23 10:01:59 SC-1 osaflogd[15289]: NO LOGSV_DATA_GROUPNAME
>>>>> not
>>>>>>> found
>>>>>>> Feb 23 10:01:59 SC-1 osaflogd[15289]: NO LOG root directory is:
>>>>>>> "/var/log/opensaf/saflog"
>>>>>>> Feb 23 10:01:59 SC-1 osaflogd[15289]: NO LOG data group is: ""
>>>>>>> Feb 23 10:01:59 SC-1 osaflogd[15289]: NO LGS_MBCSV_VERSION = 7
>>>>>>> Feb 23 10:01:59 SC-1 osafamfnd[15329]: NO Assigned
>>>>>>> 'safSi=SC-2N,safApp=OpenSAF' STANDBY to 'safSu=SC-
>>>>>>> 1,safSg=2N,safApp=OpenSAF'
>>>>>>> Feb 23 10:01:59 SC-1 opensafd: OpenSAF(5.1.M0 - ) services
>>> successfully
>>>>>>> started
>>>>>>> Feb 23 10:01:59 SC-1 osafamfnd[15329]: NO
>>>>>>> 'safComp=LOG,safSu=SC-1,safSg=2N,safApp=OpenSAF' faulted due to
>>>>>>> 'avaDown' : Recovery is 'nodeFailfast'
>>>>>>> Feb 23 10:01:59 SC-1 osafamfnd[15329]: ER
>>>>>>> safComp=LOG,safSu=SC-1,safSg=2N,safApp=OpenSAF Faulted due
>>>>>>> to:avaDown
>>>>>>> Recovery is:nodeFailfast
>>>>>>> Feb 23 10:01:59 SC-1 osafamfnd[15329]: Rebooting OpenSAF NodeId =
>>>>>>> 131343
>>>>>>> EE Name = , Reason: Component faulted: recovery is node failfast,
>>>>>>> OwnNodeId = 131343, SupervisionTime = 60
>>>>>>> Feb 23 10:01:59 SC-1 opensaf_reboot: Rebooting local node;
>>> timeout=60
>>>>>>> Feb 23 10:02:00 SC-1 osafimmnd[15279]: NO Implementer (applier)
>>>>>>> connected: 16 (@OpenSafImmReplicatorB) <144, 2010f>
>>>>>>> Feb 23 10:01:59 SC-1 opensaf_reboot: Rebooting local node;
>>> timeout=60
>>> ==============================================================
>>>>>>> ====================================================
>>>>>>>
>>>>>>> So it is evident that in-service upgrade part code of this need 
>>>>>>> to be
>>>>>>> corrected.
>>>>>>>
>>>>>>> Please see my comments as [AVM] and let me know if you need some
>>>>> traces
>>>>>>> .
>>>>>>>
>>>>>>> If you're planing to prepare new V6 patch , please do prepare on 
>>>>>>> top
>> of
>>>>>>> today's latest staging.
>>>>>>>
>>>>>>> On 2/23/2017 9:33 AM, Vu Minh Nguyen wrote:
>>>>>>>> Hi Mahesh,
>>>>>>>>
>>>>>>>> I have done in-service upgrade/downgrade with following cases:
>>>>>>>> 1) New Active SC-1 (OpenSAF 5.2 with the attached patch) + old
>>> standby
>>>>>>> SC-2
>>>>>>>> (OpenSAF 5.1)
>>>>>>>> --> Work fine
>>>>>>> [AVM] This is not a practical use cause of in-service upgrade , we
>> can
>>>>>>> ignore this test further
>>>>>>>> 2) Old Active SC-1 (OpenSAF 5.1) + new standby SC-2 (with or 
>>>>>>>> without
>>>>>>>> attached patch)
>>>>>>>> --> SC-2 is restarted & not able to join the cluster.
>>>>>>> [AVM] This use cause/flow is  we do get in in-service upgrade , 
>>>>>>> so we
>>>>>>> need to address this.
>>>>>>>> I got following messages in syslog:
>>>>>>>> Feb 23 09:32:42 SC-2 user.notice opensafd: OpenSAF(5.2.M0 -
>>>>>>>> 8529:b5addd36e45d:default) services successfully started
>>>>>>>> Feb 23 09:32:43 SC-2 local0.warn osafntfimcnd[701]: WA
>>>>>>> ntfimcn_imm_init
>>>>>>>> saImmOiImplementerSet() returned SA_AIS_ERR_TIMEOUT (5)
>>>>>>>> Feb 23 09:32:45 SC-2 local0.warn osafntfimcnd[701]: WA
>>>>>>> ntfimcn_imm_init
>>>>>>>> saImmOiImplementerSet() returned SA_AIS_ERR_TIMEOUT (5)
>>>>>>>> Feb 23 09:32:47 SC-2 local0.warn osafntfimcnd[701]: WA
>>>>>>> ntfimcn_imm_init
>>>>>>>> saImmOiImplementerSet() returned SA_AIS_ERR_TIMEOUT (5)
>>>>>>>> Feb 23 09:32:49 SC-2 local0.warn osafntfimcnd[701]: WA
>>>>>>> ntfimcn_imm_init
>>>>>>>> saImmOiImplementerSet() returned SA_AIS_ERR_TIMEOUT (5)
>>>>>>>> Feb 23 09:32:50 SC-2 local0.err osafmsgnd[592]: ER
>>>>>>> saImmOiImplementerSet
>>>>>>>> FAILED:5
>>>>>>>> Feb 23 09:32:50 SC-2 local0.err osafmsgnd[592]: ER
>>>>>>> saImmOiImplementerSet
>>>>>>>> FAILED:5
>>>>>>>> Feb 23 09:32:50 SC-2 local0.notice osafamfnd[496]: NO
>>>>>>>> 'safSu=SC-2,safSg=NoRed,safApp=OpenSAF' component restart
>>>>> probation
>>>>>>> timer
>>>>>>>> started (timeout: 60000000000 ns)
>>>>>>>> Feb 23 09:32:50 SC-2 local0.notice osafamfnd[496]: NO Restarting a
>>>>>>> component
>>>>>>>> of 'safSu=SC-2,safSg=NoRed,safApp=OpenSAF' (comp restart count: 1)
>>>>>>>> Feb 23 09:32:50 SC-2 local0.notice osafamfnd[496]: NO
>>>>>>>> 'safComp=MQND,safSu=SC-2,safSg=NoRed,safApp=OpenSAF' faulted
>>>>> due
>>>>>>> to
>>>>>>>> 'avaDown' : Recovery is 'componentRestart'
>>>>>>>> Feb 23 09:32:50 SC-2 local0.info osafmsgnd[736]: mkfifo already
>>>> exists:
>>>>>>>> /var/lib/opensaf/osafmsgnd.fifo File exists
>>>>>>>>
>>>>>>>> And sometimes, on active SC-1 (OpenSAF 5.1), the node is not 
>>>>>>>> able to
>>>> up
>>>>>>>> because of following error:
>>>>>>>>
>>>>>>>> Feb 23 11:00:32 SC-1 local0.err osafclmna[406]: MDTM:TIPC Dsock
>>>>> Socket
>>>>>>>> creation failed in MDTM_INIT err :Address family not supported by
>>>>>>> protocol
>>>>>>>> Feb 23 11:00:32 SC-1 local0.err osafclmna[406]: ER
>>> ncs_agents_startup
>>>>>>> FAILED
>>>>>>> [AVM]  No such issues ( with both TCP & TIPC) (staging changeset:
>>>> 8609
>>>>>> )
>>>>>>>> Are you getting similar problem at your side?
>>>>>>>> Please note that, the problem is existed WITH or WITHOUT the #2258
>>>>>>> patch.
>>>>>>> [AVM] No , problem only if we apply `2258_v5.patch` V5 patch on
>>> staging
>>>>>>> (changeset:   8609 )
>>>>>>>                try to bring up that node in to cluster.
>>>>>>>
>>>>>>>
>>>>>>> -AVM
>>>>>>>
>>>>>>>> I have informed this to IMM to have a look, not sure any problem
>> with
>>>>>> MDS
>>>>>>>> layer or any problem with my environment setup.
>>>>>>>> In the meantime, please have a look at the updated patch,  I will
>>>>>> continue
>>>>>>>> checking the problem. Will keep you updated.
>>>>>>> [AVM] I haven't seen any IMM problems
>>>>>>>> Regards, Vu
>>>>>>>>
>>>>>>>>> -----Original Message-----
>>>>>>>>> From: A V Mahesh [mailto:mahesh.va...@oracle.com]
>>>>>>>>> Sent: Wednesday, February 22, 2017 5:36 PM
>>>>>>>>> To: Vu Minh Nguyen <vu.m.ngu...@dektech.com.au>;
>>>>>>>>> lennart.l...@ericsson.com; canh.v.tru...@dektech.com.au
>>>>>>>>> Cc: opensaf-devel@lists.sourceforge.net
>>>>>>>>> Subject: Re: [devel] [PATCH 0 of 3] Review Request for log: add
>>>>>>>> alternative
>>>>>>>>> destinations of log records [#2258] V4
>>>>>>>>>
>>>>>>>>> Please see correction New Standby SC-1 ( with patch )
>>>>>>>>>
>>>>>>>>> -AVM
>>>>>>>>>
>>>>>>>>> On 2/22/2017 4:02 PM, A V Mahesh wrote:
>>>>>>>>>> Hi Vu,
>>>>>>>>>>
>>>>>>>>>> With this new patch , we have another issue :
>>>>>>>>>>
>>>>>>>>>> 1)  standby Core by `/usr/lib64/opensaf/osaflogd'  issue got
>>>> resolved
>>>>>> .
>>>>>>>>>> 2) In-service upgrade is Not working , I have Old  Active SC-2 (
>>>> with
>>>>>>>>>> out patch )  and New Standby SC-1 ( with patch )
>>>>>>>>>>
>>>>>>>>>>         the new New Standby SC-1 not joining the cluster (
>> in-service
>>>>>>>>>> upgrade  failed )
>>>>>>>>>>
>>>>>>>>>> New Standby SC-1
>>>>>>>>>>
>>>>>>>>>>
>>> ==============================================================
>>>>>>>>> ======================================================
>>>>>>>>>> Feb 22 15:53:05 SC-1 osafamfnd[15752]: NO
>>>>>>>>>> 'safSu=SC-1,safSg=NoRed,safApp=OpenSAF' Presence State
>>>>>>>>> INSTANTIATING
>>>>>>>>>> => INSTANTIATED
>>>>>>>>>> Feb 22 15:53:05 SC-1 osafamfnd[15752]: NO Assigning
>>>>>>>>>> 'safSi=NoRed4,safApp=OpenSAF' ACTIVE to
>>>>>>>>>> 'safSu=SC-1,safSg=NoRed,safApp=OpenSAF'
>>>>>>>>>> Feb 22 15:53:05 SC-1 osafamfnd[15752]: NO Assigned
>>>>>>>>>> 'safSi=NoRed4,safApp=OpenSAF' ACTIVE to
>>>>>>>>>> 'safSu=SC-1,safSg=NoRed,safApp=OpenSAF'
>>>>>>>>>> Feb 22 15:53:05 SC-1 osafsmfd[15889]: Started
>>>>>>>>>> Feb 22 15:53:05 SC-1 osafamfnd[15752]: NO
>>>>>>>>>> 'safSu=SC-1,safSg=2N,safApp=OpenSAF' Presence State
>>>>> INSTANTIATING
>>>>>>> =>
>>>>>>>>>> INSTANTIATED
>>>>>>>>>> Feb 22 15:53:05 SC-1 osafamfnd[15752]: NO Assigning
>>>>>>>>>> 'safSi=SC-2N,safApp=OpenSAF' STANDBY to
>>>>>>>>>> 'safSu=SC-1,safSg=2N,safApp=OpenSAF'
>>>>>>>>>> Feb 22 15:53:05 SC-1 osafrded[15672]: NO RDE role set to
>>> STANDBY
>>>>>>>>>> Feb 22 15:53:05 SC-1 osafrded[15672]: NO Peer up on node
>>> 0x2020f
>>>>>>>>>> Feb 22 15:53:05 SC-1 osafrded[15672]: NO Got peer info request
>>> from
>>>>>>>>>> node 0x2020f with role ACTIVE
>>>>>>>>>> Feb 22 15:53:05 SC-1 osafrded[15672]: NO Got peer info response
>>>>> from
>>>>>>>>>> node 0x2020f with role ACTIVE
>>>>>>>>>> Feb 22 15:53:05 SC-1 osafimmd[15691]: NO MDS event from svc_id
>>> 24
>>>>>>>>>> (change:5, dest:13)
>>>>>>>>>> Feb 22 15:53:05 SC-1 osafimmd[15691]: NO MDS event from svc_id
>>> 24
>>>>>>>>>> (change:3, dest:13)
>>>>>>>>>> Feb 22 15:53:05 SC-1 osafimmd[15691]: NO MDS event from svc_id
>>> 24
>>>>>>>>>> (change:5, dest:13)
>>>>>>>>>> Feb 22 15:53:05 SC-1 osafimmd[15691]: NO MDS event from svc_id
>>> 25
>>>>>>>>>> (change:3, dest:567412424453430)
>>>>>>>>>> Feb 22 15:53:05 SC-1 osafimmd[15691]: NO MDS event from svc_id
>>> 25
>>>>>>>>>> (change:3, dest:565213401202663)
>>>>>>>>>> Feb 22 15:53:05 SC-1 osafimmd[15691]: NO MDS event from svc_id
>>> 25
>>>>>>>>>> (change:3, dest:566312912825221)
>>>>>>>>>> Feb 22 15:53:05 SC-1 osafimmd[15691]: NO MDS event from svc_id
>>> 25
>>>>>>>>>> (change:3, dest:564113889574230)
>>>>>>>>>> Feb 22 15:53:05 SC-1 osafimmnd[15702]: NO Implementer (applier)
>>>>>>>>>> connected: 17 (@safAmfService2010f) <127, 2010f>
>>>>>>>>>> Feb 22 15:53:05 SC-1 osaflogd[15712]: NO
>>>>> LOGSV_DATA_GROUPNAME
>>>>>>>>> not found
>>>>>>>>>> Feb 22 15:53:05 SC-1 osaflogd[15712]: NO LOG root directory is:
>>>>>>>>>> "/var/log/opensaf/saflog"
>>>>>>>>>> Feb 22 15:53:05 SC-1 osaflogd[15712]: NO LOG data group is: ""
>>>>>>>>>> Feb 22 15:53:05 SC-1 osaflogd[15712]: NO LGS_MBCSV_VERSION =
>>> 7
>>>>>>>>>> Feb 22 15:53:05 SC-1 osafamfnd[15752]: NO Assigned
>>>>>>>>>> 'safSi=SC-2N,safApp=OpenSAF' STANDBY to
>>>>>>>>>> 'safSu=SC-1,safSg=2N,safApp=OpenSAF'
>>>>>>>>>> Feb 22 15:53:05 SC-1 opensafd: OpenSAF(5.1.M0 - ) services
>>>>>>>>>> successfully started
>>>>>>>>>> Feb 22 15:53:05 SC-1 osafamfnd[15752]: NO
>>>>>>>>>> 'safComp=LOG,safSu=SC-1,safSg=2N,safApp=OpenSAF' faulted due
>>> to
>>>>>>>>>> 'avaDown' : Recovery is 'nodeFailfast'
>>>>>>>>>> Feb 22 15:53:05 SC-1 osafamfnd[15752]: ER
>>>>>>>>>> safComp=LOG,safSu=SC-1,safSg=2N,safApp=OpenSAF Faulted due
>>>>>>>>> to:avaDown
>>>>>>>>>> Recovery is:nodeFailfast
>>>>>>>>>> Feb 22 15:53:05 SC-1 osafamfnd[15752]: Rebooting OpenSAF
>>> NodeId
>>>>> =
>>>>>>>>>> 131343 EE Name = , Reason: Component faulted: recovery is node
>>>>>>>>>> failfast, OwnNodeId = 131343, SupervisionTime = 60
>>>>>>>>>> Feb 22 15:53:05 SC-1 opensaf_reboot: Rebooting local node;
>>>>>>> timeout=60
>>>>>>>>>> Feb 22 15:53:43 SC-1 syslog-ng[1171]: syslog-ng starting up;
>>>>>>>>>> version='2.0.9'
>>>>>>>>>>
>>>>>>>>>>
>>> ==============================================================
>>>>>>>>> ======================================================
>>>>>>>>>> Old - Active - SC-2
>>>>>>>>>>
>>>>>>>>>>
>>> ==============================================================
>>>>>>>>> ======================================================
>>>>>>>>>> Feb 22 15:53:02 SC-2 osafimmnd[16359]: NO NODE STATE->
>>>>>>>>>> IMM_NODE_R_AVAILABLE
>>>>>>>>>> Feb 22 15:53:02 SC-2 osafimmloadd: NO Sync starting
>>>>>>>>>> Feb 22 15:53:02 SC-2 osafimmloadd: IN Synced 390 objects in 
>>>>>>>>>> total
>>>>>>>>>> Feb 22 15:53:02 SC-2 osafimmnd[16359]: NO NODE STATE->
>>>>>>>>>> IMM_NODE_FULLY_AVAILABLE 18511
>>>>>>>>>> Feb 22 15:53:02 SC-2 osafimmnd[16359]: NO Epoch set to 3 in
>>>>>>> ImmModel
>>>>>>>>>> Feb 22 15:53:02 SC-2 osafimmd[16346]: NO ACT: New Epoch for
>>>>>>> IMMND
>>>>>>>>>> process at node 2020f old epoch: 2  new epoch:3
>>>>>>>>>> Feb 22 15:53:02 SC-2 osafimmd[16346]: NO ACT: New Epoch for
>>>>>>> IMMND
>>>>>>>>>> process at node 2040f old epoch: 2  new epoch:3
>>>>>>>>>> Feb 22 15:53:02 SC-2 osafimmd[16346]: NO ACT: New Epoch for
>>>>>>> IMMND
>>>>>>>>>> process at node 2030f old epoch: 2  new epoch:3
>>>>>>>>>> Feb 22 15:53:02 SC-2 osafimmloadd: NO Sync ending normally
>>>>>>>>>> Feb 22 15:53:02 SC-2 osafimmd[16346]: NO ACT: New Epoch for
>>>>>>> IMMND
>>>>>>>>>> process at node 2010f old epoch: 0  new epoch:3
>>>>>>>>>> Feb 22 15:53:02 SC-2 osafimmnd[16359]: NO SERVER STATE:
>>>>>>>>>> IMM_SERVER_SYNC_SERVER --> IMM_SERVER_READY
>>>>>>>>>> Feb 22 15:53:03 SC-2 osafamfd[16408]: NO Received node_up from
>>>>>>> 2010f:
>>>>>>>>>> msg_id 1
>>>>>>>>>> Feb 22 15:53:03 SC-2 osafamfd[16408]: NO Node 'SC-1' joined the
>>>>>>> cluster
>>>>>>>>>> Feb 22 15:53:03 SC-2 osafimmnd[16359]: NO Implementer
>>> connected:
>>>>>>> 16
>>>>>>>>>> (MsgQueueService131343) <0, 2010f>
>>>>>>>>>> Feb 22 15:53:03 SC-2 osafrded[16327]: NO Peer up on node
>>> 0x2010f
>>>>>>>>>> Feb 22 15:53:03 SC-2 osafrded[16327]: NO Got peer info request
>>> from
>>>>>>>>>> node 0x2010f with role STANDBY
>>>>>>>>>> Feb 22 15:53:03 SC-2 osafrded[16327]: NO Got peer info response
>>>>> from
>>>>>>>>>> node 0x2010f with role STANDBY
>>>>>>>>>> Feb 22 15:53:03 SC-2 osafimmd[16346]: NO MDS event from svc_id
>>> 24
>>>>>>>>>> (change:5, dest:13)
>>>>>>>>>> Feb 22 15:53:03 SC-2 osafimmnd[16359]: NO Implementer (applier)
>>>>>>>>>> connected: 17 (@safAmfService2010f) <0, 2010f>
>>>>>>>>>> Feb 22 15:53:03 SC-2 osafamfd[16408]: NO Cluster startup is done
>>>>>>>>>> Feb 22 15:53:04 SC-2 osafimmnd[16359]: NO Implementer (applier)
>>>>>>>>>> connected: 18 (@OpenSafImmReplicatorB) <0, 2010f>
>>>>>>>>>> Feb 22 15:53:09 SC-2 osafdtmd[16304]: NO Lost contact with 
>>>>>>>>>> 'SC-1'
>>>>>>>>>> Feb 22 15:53:09 SC-2 osaffmd[16336]: NO Node Down event for
>>> node
>>>>> id
>>>>>>>>>> 2010f:
>>>>>>>>>> Feb 22 15:53:09 SC-2 osafimmd[16346]: NO MDS event from svc_id
>>> 24
>>>>>>>>>> (change:6, dest:13)
>>>>>>>>>> Feb 22 15:53:09 SC-2 osafimmd[16346]: NO MDS event from svc_id
>>> 25
>>>>>>>>>> (change:4, dest:564113889574230)
>>>>>>>>>> Feb 22 15:53:09 SC-2 osaffmd[16336]: NO Current role: ACTIVE
>>>>>>>>>> Feb 22 15:53:09 SC-2 osaffmd[16336]: Rebooting OpenSAF NodeId
>>> =
>>>>>>>>> 131343
>>>>>>>>>> EE Name = , Reason: Received Node Down for peer controller,
>>>>>>> OwnNodeId
>>>>>>>>>> = 131599, SupervisionTime = 60
>>>>>>>>>> Feb 22 15:53:09 SC-2 osafrded[16327]: NO Peer down on node
>>>>> 0x2010f
>>>>>>>>>> Feb 22 15:53:09 SC-2 osafamfd[16408]: NO Node 'SC-1' left the
>>>>> cluster
>>>>>>>>>> Feb 22 15:53:09 SC-2 osaflogd[16372]: NO Failed (2) to send of
>>> WRITE
>>>>>>>>>> ack to: 2010f00003d6a
>>>>>>>>>> Feb 22 15:53:09 SC-2 osaflogd[16372]: NO Failed (2) to send of
>>> WRITE
>>>>>>>>>> ack to: 2010f00003d6a
>>>>>>>>>> Feb 22 15:53:09 SC-2 osaflogd[16372]: NO Failed (2) to send of
>>> WRITE
>>>>>>>>>> ack to: 2010f00003d74
>>>>>>>>>> Feb 22 15:53:09 SC-2 osafclmd[16398]: NO Node 131343 went
>>> down.
>>>>>>> Not
>>>>>>>>>> sending track callback for agents on that node
>>>>>>>>>> Feb 22 15:53:09 SC-2 osafclmd[16398]: NO Node 131343 went
>>> down.
>>>>>>> Not
>>>>>>>>>> sending track callback for agents on that node
>>>>>>>>>> Feb 22 15:53:09 SC-2 osafclmd[16398]: NO Node 131343 went
>>> down.
>>>>>>> Not
>>>>>>>>>> sending track callback for agents on that node
>>>>>>>>>> Feb 22 15:53:09 SC-2 osafclmd[16398]: NO Node 131343 went
>>> down.
>>>>>>> Not
>>>>>>>>>> sending track callback for agents on that node
>>>>>>>>>> Feb 22 15:53:09 SC-2 osafclmd[16398]: NO Node 131343 went
>>> down.
>>>>>>> Not
>>>>>>>>>> sending track callback for agents on that node
>>>>>>>>>> Feb 22 15:53:09 SC-2 osafclmd[16398]: NO Node 131343 went
>>> down.
>>>>>>> Not
>>>>>>>>>> sending track callback for agents on that node
>>>>>>>>>> Feb 22 15:53:09 SC-2 osafclmd[16398]: NO Node 131343 went
>>> down.
>>>>>>> Not
>>>>>>>>>> sending track callback for agents on that node
>>>>>>>>>> Feb 22 15:53:09 SC-2 osafimmd[16346]: WA IMMD lost contact
>>> with
>>>>>>> peer
>>>>>>>>>> IMMD (NCSMDS_RED_DOWN)
>>>>>>>>>> Feb 22 15:53:09 SC-2 osafimmnd[16359]: NO Global discard node
>>>>>>> received
>>>>>>>>>> for nodeId:2010f pid:15702
>>>>>>>>>> Feb 22 15:53:09 SC-2 osafimmnd[16359]: NO Implementer
>>>>> disconnected
>>>>>>> 16
>>>>>>>>>> <0, 2010f(down)> (MsgQueueService131343)
>>>>>>>>>> Feb 22 15:53:09 SC-2 osafimmnd[16359]: NO Implementer
>>>>> disconnected
>>>>>>> 17
>>>>>>>>>> <0, 2010f(down)> (@safAmfService2010f)
>>>>>>>>>> Feb 22 15:53:09 SC-2 osafimmnd[16359]: NO Implementer
>>>>> disconnected
>>>>>>> 18
>>>>>>>>>> <0, 2010f(down)> (@OpenSafImmReplicatorB)
>>>>>>>>>> Feb 22 15:53:09 SC-2 opensaf_reboot: Rebooting remote node in
>>> the
>>>>>>>>>> absence of PLM is outside the scope of OpenSAF
>>>>>>>>>>
>>>>>>>>>>
>>> ==============================================================
>>>>>>>>> ======================================================
>>>>>>>>>> -AVM
>>>>>>>>>>
>>>>>>>>>> On 2/22/2017 3:13 PM, Vu Minh Nguyen wrote:
>>>>>>>>>>> Hi Mahesh,
>>>>>>>>>>>
>>>>>>>>>>> I put all required patches into one. Try to use this and see if
>> you
>>>>>>>>>>> still
>>>>>>>>>>> have that problem or not.
>>>>>>>>>>>
>>>>>>>>>>> Regards, Vu
>>>>>>>>>>>
>>>>>>>>>>>> -----Original Message-----
>>>>>>>>>>>> From: A V Mahesh [mailto:mahesh.va...@oracle.com]
>>>>>>>>>>>> Sent: Wednesday, February 22, 2017 3:35 PM
>>>>>>>>>>>> To: Vu Minh Nguyen <vu.m.ngu...@dektech.com.au>;
>>>>>>>>>>>> lennart.l...@ericsson.com; canh.v.tru...@dektech.com.au
>>>>>>>>>>>> Cc: opensaf-devel@lists.sourceforge.net
>>>>>>>>>>>> Subject: Re: [devel] [PATCH 0 of 3] Review Request for log: 
>>>>>>>>>>>> add
>>>>>>>>>>> alternative
>>>>>>>>>>>> destinations of log records [#2258] V4
>>>>>>>>>>>>
>>>>>>>>>>>> Hi Vu,
>>>>>>>>>>>>
>>>>>>>>>>>> I used  new #3 and #4 patches ,  Can you please re-send All 
>>>>>>>>>>>> the
>>>>>> final
>>>>>>>>>>>> patch in go,
>>>>>>>>>>>>
>>>>>>>>>>>> which i need to apply on today`s  staging ( if possible 
>>>>>>>>>>>> publish
>>>> the
>>>>>>>>>>>> with
>>>>>>>>>>>> new version )
>>>>>>>>>>>>
>>>>>>>>>>>> -AVM
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> On 2/22/2017 1:52 PM, Vu Minh Nguyen wrote:
>>>>>>>>>>>>> Hi Mahesh,
>>>>>>>>>>>>>
>>>>>>>>>>>>>> Core was generated by `/usr/lib64/opensaf/osaflogd'.
>>>>>>>>>>>>>> Program terminated with signal 11, Segmentation fault.
>>>>>>>>>>>>>> #0  ckpt_proc_cfg_stream(lgs_cb*, void*) () at
>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc:2195
>>>>>>>>>>>>>> 2195    src/log/logd/lgs_mbcsv.cc: No such file or 
>>>>>>>>>>>>>> directory.
>>>>>>>>>>>>>>                in src/log/logd/lgs_mbcsv.cc
>>>>>>>>>>>>> Backtrace still points to old position (lgs_mbcsv:2195). I
>> guess
>>>>>> the
>>>>>>>>>>>>> osaflogd binary has not been updated with the fixed patch.
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>> Regards, Vu
>>>>>>>>>>>>>
>>>>>>>>>>>>>> -----Original Message-----
>>>>>>>>>>>>>> From: A V Mahesh [mailto:mahesh.va...@oracle.com]
>>>>>>>>>>>>>> Sent: Wednesday, February 22, 2017 3:18 PM
>>>>>>>>>>>>>> To: Vu Minh Nguyen <vu.m.ngu...@dektech.com.au>;
>>>>>>>>>>>>>> lennart.l...@ericsson.com; canh.v.tru...@dektech.com.au
>>>>>>>>>>>>>> Cc: opensaf-devel@lists.sourceforge.net
>>>>>>>>>>>>>> Subject: Re: [devel] [PATCH 0 of 3] Review Request for log:
>>> add
>>>>>>>>>>>>> alternative
>>>>>>>>>>>>>> destinations of log records [#2258] V4
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> Hi Vu,
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> SC-2 standby osaflogd core dumped still occurs ( Not 
>>>>>>>>>>>>>> resolved)
>>> ,
>>>>>> the
>>>>>>>>>>> new
>>>>>>>>>>>>>> patch only resolved the application (/usr/bin/logtest )
>>>>>> Segmentation
>>>>>>>>>>>>>> fault on SC-1 Active.
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>> ==============================================================
>>> ==============================================================
>>>>>>>>>>>>>> # gdb /usr/lib64/opensaf/osaflogd
>>>>>>> core_1487751055.osaflogd.4594
>>>>>>>>> GNU
>>>>>>>>>>>>>> gdb
>>>>>>>>>>>>>> (GDB) SUSE (7.3-0.6.1)
>>>>>>>>>>>>>> Copyright (C) 2011 Free Software Foundation, Inc.
>>>>>>>>>>>>>> .......
>>>>>>>>>>>>>> Core was generated by `/usr/lib64/opensaf/osaflogd'.
>>>>>>>>>>>>>> Program terminated with signal 11, Segmentation fault.
>>>>>>>>>>>>>> #0  ckpt_proc_cfg_stream(lgs_cb*, void*) () at
>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc:2195
>>>>>>>>>>>>>> 2195    src/log/logd/lgs_mbcsv.cc: No such file or 
>>>>>>>>>>>>>> directory.
>>>>>>>>>>>>>>                in src/log/logd/lgs_mbcsv.cc
>>>>>>>>>>>>>> (gdb) bt
>>>>>>>>>>>>>> #0  ckpt_proc_cfg_stream(lgs_cb*, void*) () at
>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc:2195
>>>>>>>>>>>>>> #1  0x00007f97b026f960 in ckpt_decode_log_struct(lgs_cb*,
>>>>>>>>>>>>>> ncs_mbcsv_cb_arg*, void*, void*, unsigned int
>>> (*)(edu_hdl_tag*,
>>>>>>>>>>>>>> edu_tkn_tag*, void*, unsigned int*, edu_buf_env_tag*,
>>>>>>>>> EDP_OP_TYPE,
>>>>>>>>>>>>>> EDU_ERR*)) () at src/log/logd/lgs_mbcsv.cc:950
>>>>>>>>>>>>>> #2  0x00007f97b02710dc in
>>> ckpt_decode_async_update(lgs_cb*,
>>>>>>>>>>>>>> ncs_mbcsv_cb_arg*) () at src/log/logd/lgs_mbcsv.cc:1086
>>>>>>>>>>>>>> #3  0x00007f97b0273941 in
>>> mbcsv_callback(ncs_mbcsv_cb_arg*)
>>>>> ()
>>>>>>> at
>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc:880
>>>>>>>>>>>>>> #4  0x00007f97af372596 in ncs_mbscv_rcv_decode () from
>>>>>>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0
>>>>>>>>>>>>>> #5  0x00007f97af372766 in ncs_mbcsv_rcv_async_update ()
>>> from
>>>>>>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0
>>>>>>>>>>>>>> #6  0x00007f97af379370 in mbcsv_process_events () from
>>>>>>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0
>>>>>>>>>>>>>> #7  0x00007f97af3794db in mbcsv_hdl_dispatch_all () from
>>>>>>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0
>>>>>>>>>>>>>> #8  0x00007f97af373ce2 in mbcsv_process_dispatch_request ()
>>>>> at
>>>>>>>>>>>>>> src/mbc/mbcsv_api.c:423
>>>>>>>>>>>>>> #9  0x00007f97b027096e in lgs_mbcsv_dispatch(unsigned int)
>>> ()
>>>>> at
>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc:327
>>>>>>>>>>>>>> #10 0x00007f97b024d9f2 in main () at
>>>>> src/log/logd/lgs_main.cc:583
>>>>>>>>>>>>>> (gdb) bt full
>>>>>>>>>>>>>> #0  ckpt_proc_cfg_stream(lgs_cb*, void*) () at
>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc:2195
>>>>>>>>>>>>>>                ckpt_data_handler = {0x7f97b0270300
>>>>>>>>>>>>>> <ckpt_proc_initialize_client(lgs_cb*, void*)>, 
>>>>>>>>>>>>>> 0x7f97b02701c0
>>>>>>>>>>>>>> <ckpt_proc_finalize_client(lgs_cb*, void*)>,
>>>>>>>>>>>>>>                  0x7f97b0270060 
>>>>>>>>>>>>>> <ckpt_proc_agent_down(lgs_cb*,
>>>>>> void*)>,
>>>>>>>>>>>>>> 0x7f97b02712f0 <ckpt_proc_log_write(lgs_cb*, void*)>,
>>>>>>>>> 0x7f97b0271ab0
>>>>>>>>>>>>>> <ckpt_proc_open_stream(lgs_cb*, void*)>,
>>>>>>>>>>>>>>                  0x7f97b026fe80
>> <ckpt_proc_close_stream(lgs_cb*,
>>>>>>>> void*)>,
>>>>>>>>>>>>>> 0x7f97b0272380 <ckpt_proc_cfg_stream(lgs_cb*, void*)>,
>>>>>>>>>>>> 0x7f97b0274800
>>>>>>>>>>>>>> <ckpt_proc_lgs_cfg_v2(lgs_cb*, void*)>,
>>>>>>>>>>>>>>                  0x7f97b0274e10 
>>>>>>>>>>>>>> <ckpt_proc_lgs_cfg_v3(lgs_cb*,
>>>>>> void*)>,
>>>>>>>>>>>>>> 0x7f97b02754f0 <ckpt_proc_lgs_cfg_v5(lgs_cb*, void*)>}
>>>>>>>>>>>>>> #1  0x00007f97b026f960 in ckpt_decode_log_struct(lgs_cb*,
>>>>>>>>>>>>>> ncs_mbcsv_cb_arg*, void*, void*, unsigned int
>>> (*)(edu_hdl_tag*,
>>>>>>>>>>>>>> edu_tkn_tag*, void*, unsigned int*, edu_buf_env_tag*,
>>>>>>>>> EDP_OP_TYPE,
>>>>>>>>>>>>>> EDU_ERR*)) () at src/log/logd/lgs_mbcsv.cc:950
>>>>>>>>>>>>>>                ckpt_data_handler = {0x7f97b0270300
>>>>>>>>>>>>>> <ckpt_proc_initialize_client(lgs_cb*, void*)>, 
>>>>>>>>>>>>>> 0x7f97b02701c0
>>>>>>>>>>>>>> <ckpt_proc_finalize_client(lgs_cb*, void*)>,
>>>>>>>>>>>>>>                  0x7f97b0270060 
>>>>>>>>>>>>>> <ckpt_proc_agent_down(lgs_cb*,
>>>>>> void*)>,
>>>>>>>>>>>>>> 0x7f97b02712f0 <ckpt_proc_log_write(lgs_cb*, void*)>,
>>>>>>>>> 0x7f97b0271ab0
>>>>>>>>>>>>>> <ckpt_proc_open_stream(lgs_cb*, void*)>,
>>>>>>>>>>>>>>                  0x7f97b026fe80
>> <ckpt_proc_close_stream(lgs_cb*,
>>>>>>>> void*)>,
>>>>>>>>>>>>>> 0x7f97b0272380 <ckpt_proc_cfg_stream(lgs_cb*, void*)>,
>>>>>>>>>>>> 0x7f97b0274800
>>>>>>>>>>>>>> <ckpt_proc_lgs_cfg_v2(lgs_cb*, void*)>,
>>>>>>>>>>>>>>                  0x7f97b0274e10 
>>>>>>>>>>>>>> <ckpt_proc_lgs_cfg_v3(lgs_cb*,
>>>>>> void*)>,
>>>>>>>>>>>>>> 0x7f97b02754f0 <ckpt_proc_lgs_cfg_v5(lgs_cb*, void*)>}
>>>>>>>>>>>>>> #2  0x00007f97b02710dc in
>>> ckpt_decode_async_update(lgs_cb*,
>>>>>>>>>>>>>> ncs_mbcsv_cb_arg*) () at src/log/logd/lgs_mbcsv.cc:1086
>>>>>>>>>>>>>>                ckpt_data_handler = {0x7f97b0270300
>>>>>>>>>>>>>> <ckpt_proc_initialize_client(lgs_cb*, void*)>, 
>>>>>>>>>>>>>> 0x7f97b02701c0
>>>>>>>>>>>>>> <ckpt_proc_finalize_client(lgs_cb*, void*)>,
>>>>>>>>>>>>>>                  0x7f97b0270060 
>>>>>>>>>>>>>> <ckpt_proc_agent_down(lgs_cb*,
>>>>>> void*)>,
>>>>>>>>>>>>>> 0x7f97b02712f0 <ckpt_proc_log_write(lgs_cb*, void*)>,
>>>>>>>>> 0x7f97b0271ab0
>>>>>>>>>>>>>> <ckpt_proc_open_stream(lgs_cb*, void*)>,
>>>>>>>>>>>>>>                  0x7f97b026fe80
>> <ckpt_proc_close_stream(lgs_cb*,
>>>>>>>> void*)>,
>>>>>>>>>>>>>> 0x7f97b0272380 <ckpt_proc_cfg_stream(lgs_cb*, void*)>,
>>>>>>>>>>>> 0x7f97b0274800
>>>>>>>>>>>>>> <ckpt_proc_lgs_cfg_v2(lgs_cb*, void*)>,
>>>>>>>>>>>>>>                  0x7f97b0274e10 
>>>>>>>>>>>>>> <ckpt_proc_lgs_cfg_v3(lgs_cb*,
>>>>>> void*)>,
>>>>>>>>>>>>>> 0x7f97b02754f0 <ckpt_proc_lgs_cfg_v5(lgs_cb*, void*)>}
>>>>>>>>>>>>>> #3  0x00007f97b0273941 in
>>> mbcsv_callback(ncs_mbcsv_cb_arg*)
>>>>> ()
>>>>>>> at
>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc:880
>>>>>>>>>>>>>>                ckpt_data_handler = {0x7f97b0270300
>>>>>>>>>>>>>> <ckpt_proc_initialize_client(lgs_cb*, void*)>, 
>>>>>>>>>>>>>> 0x7f97b02701c0
>>>>>>>>>>>>>> <ckpt_proc_finalize_client(lgs_cb*, void*)>,
>>>>>>>>>>>>>>                  0x7f97b0270060 
>>>>>>>>>>>>>> <ckpt_proc_agent_down(lgs_cb*,
>>>>>> void*)>,
>>>>>>>>>>>>>> 0x7f97b02712f0 <ckpt_proc_log_write(lgs_cb*, void*)>,
>>>>>>>>> 0x7f97b0271ab0
>>>>>>>>>>>>>> <ckpt_proc_open_stream(lgs_cb*, void*)>,
>>>>>>>>>>>>>>                  0x7f97b026fe80
>> <ckpt_proc_close_stream(lgs_cb*,
>>>>>>>> void*)>,
>>>>>>>>>>>>>> 0x7f97b0272380 <ckpt_proc_cfg_stream(lgs_cb*, void*)>,
>>>>>>>>>>>> 0x7f97b0274800
>>>>>>>>>>>>>> <ckpt_proc_lgs_cfg_v2(lgs_cb*, void*)>,
>>>>>>>>>>>>>>                  0x7f97b0274e10 
>>>>>>>>>>>>>> <ckpt_proc_lgs_cfg_v3(lgs_cb*,
>>>>>> void*)>,
>>>>>>>>>>>>>> 0x7f97b02754f0 <ckpt_proc_lgs_cfg_v5(lgs_cb*, void*)>}
>>>>>>>>>>>>>> #4  0x00007f97af372596 in ncs_mbscv_rcv_decode () from
>>>>>>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0
>>>>>>>>>>>>>> mbcsv_init_process_req_func = {0x7f97af373630
>>>>>>>>>>>>>> <mbcsv_process_initialize_request>, 0x7f97af374f10
>>>>>>>>>>>>>> <mbcsv_process_get_sel_obj_request>,
>>>>>>>>>>>>>>                  0x7f97af373be0
>> <mbcsv_process_dispatch_request>,
>>>>>>>>>>>>>> 0x7f97af373af0 <mbcsv_process_finalize_request>,
>>>>>>> 0x7f97af373cf0
>>>>>>>>>>>>>> <mbcsv_process_open_request>,
>>>>>>>>>>>>>>                  0x7f97af374050 
>>>>>>>>>>>>>> <mbcsv_process_close_request>,
>>>>>>>>>>> 0x7f97af3741e0
>>>>>>>>>>>>>> <mbcsv_process_chg_role_request>, 0x7f97af3744c0
>>>>>>>>>>>>>> <mbcsv_process_snd_ckpt_request>,
>>>>>>>>>>>>>>                  0x7f97af3747d0
>> <mbcsv_process_snd_ntfy_request>,
>>>>>>>>>>>>>> 0x7f97af374970 <mbcsv_process_snd_data_req>,
>>>>> 0x7f97af373930
>>>>>>>>>>>>>> <mbcsv_process_get_request>,
>>>>>>>>>>>>>>                  0x7f97af374bd0 <mbcsv_process_set_request>}
>>>>>>>>>>>>>> #5  0x00007f97af372766 in ncs_mbcsv_rcv_async_update ()
>>> from
>>>>>>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0
>>>>>>>>>>>>>> mbcsv_init_process_req_func = {0x7f97af373630
>>>>>>>>>>>>>> <mbcsv_process_initialize_request>, 0x7f97af374f10
>>>>>>>>>>>>>> <mbcsv_process_get_sel_obj_request>,
>>>>>>>>>>>>>>                  0x7f97af373be0
>> <mbcsv_process_dispatch_request>,
>>>>>>>>>>>>>> 0x7f97af373af0 <mbcsv_process_finalize_request>,
>>>>>>> 0x7f97af373cf0
>>>>>>>>>>>>>> <mbcsv_process_open_request>,
>>>>>>>>>>>>>>                  0x7f97af374050 
>>>>>>>>>>>>>> <mbcsv_process_close_request>,
>>>>>>>>>>> 0x7f97af3741e0
>>>>>>>>>>>>>> <mbcsv_process_chg_role_request>, 0x7f97af3744c0
>>>>>>>>>>>>>> <mbcsv_process_snd_ckpt_request>,
>>>>>>>>>>>>>>                  0x7f97af3747d0
>> <mbcsv_process_snd_ntfy_request>,
>>>>>>>>>>>>>> 0x7f97af374970 <mbcsv_process_snd_data_req>,
>>>>> 0x7f97af373930
>>>>>>>>>>>>>> <mbcsv_process_get_request>,
>>>>>>>>>>>>>>                  0x7f97af374bd0 <mbcsv_process_set_request>}
>>>>>>>>>>>>>> #6  0x00007f97af379370 in mbcsv_process_events () from
>>>>>>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0
>>>>>>>>>>>>>> No symbol table info available.
>>>>>>>>>>>>>> #7  0x00007f97af3794db in mbcsv_hdl_dispatch_all () from
>>>>>>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0
>>>>>>>>>>>>>> No symbol table info available.
>>>>>>>>>>>>>> #8  0x00007f97af373ce2 in mbcsv_process_dispatch_request ()
>>>>> at
>>>>>>>>>>>>>> src/mbc/mbcsv_api.c:423
>>>>>>>>>>>>>> mbcsv_init_process_req_func = {0x7f97af373630
>>>>>>>>>>>>>> <mbcsv_process_initialize_request>, 0x7f97af374f10
>>>>>>>>>>>>>> <mbcsv_process_get_sel_obj_request>,
>>>>>>>>>>>>>>                  0x7f97af373be0
>> <mbcsv_process_dispatch_request>,
>>>>>>>>>>>>>> 0x7f97af373af0 <mbcsv_process_finalize_request>,
>>>>>>> 0x7f97af373cf0
>>>>>>>>>>>>>> <mbcsv_process_open_request>,
>>>>>>>>>>>>>>                  0x7f97af374050 
>>>>>>>>>>>>>> <mbcsv_process_close_request>,
>>>>>>>>>>> 0x7f97af3741e0
>>>>>>>>>>>>>> <mbcsv_process_chg_role_request>, 0x7f97af3744c0
>>>>>>>>>>>>>> <mbcsv_process_snd_ckpt_request>,
>>>>>>>>>>>>>>                  0x7f97af3747d0
>> <mbcsv_process_snd_ntfy_request>,
>>>>>>>>>>>>>> 0x7f97af374970 <mbcsv_process_snd_data_req>,
>>>>> 0x7f97af373930
>>>>>>>>>>>>>> <mbcsv_process_get_request>,
>>>>>>>>>>>>>>                  0x7f97af374bd0 <mbcsv_process_set_request>}
>>>>>>>>>>>>>> #9  0x00007f97b027096e in lgs_mbcsv_dispatch(unsigned int)
>>> ()
>>>>> at
>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc:327
>>>>>>>>>>>>>>                ckpt_data_handler = {0x7f97b0270300
>>>>>>>>>>>>>> <ckpt_proc_initialize_client(lgs_cb*, void*)>, 
>>>>>>>>>>>>>> 0x7f97b02701c0
>>>>>>>>>>>>>> <ckpt_proc_finalize_client(lgs_cb*, void*)>,
>>>>>>>>>>>>>>                  0x7f97b0270060 
>>>>>>>>>>>>>> <ckpt_proc_agent_down(lgs_cb*,
>>>>>> void*)>,
>>>>>>>>>>>>>> 0x7f97b02712f0 <ckpt_proc_log_write(lgs_cb*, void*)>,
>>>>>>>>> 0x7f97b0271ab0
>>>>>>>>>>>>>> <ckpt_proc_open_stream(lgs_cb*, void*)>,
>>>>>>>>>>>>>>                  0x7f97b026fe80
>> <ckpt_proc_close_stream(lgs_cb*,
>>>>>>>> void*)>,
>>>>>>>>>>>>>> 0x7f97b0272380 <ckpt_proc_cfg_stream(lgs_cb*, void*)>,
>>>>>>>>>>>> 0x7f97b0274800
>>>>>>>>>>>>>> <ckpt_proc_lgs_cfg_v2(lgs_cb*, void*)>,
>>>>>>>>>>>>>>                  0x7f97b0274e10 
>>>>>>>>>>>>>> <ckpt_proc_lgs_cfg_v3(lgs_cb*,
>>>>>> void*)>,
>>>>>>>>>>>>>> 0x7f97b02754f0 <ckpt_proc_lgs_cfg_v5(lgs_cb*, void*)>}
>>>>>>>>>>>>>> ---Type <return> to continue, or q <return> to quit---
>>>>>>>>>>>>>> #10 0x00007f97b024d9f2 in main () at
>>>>> src/log/logd/lgs_main.cc:583
>>>>>>>>>>>>>> usr1_sel_obj = {raise_obj = -1, rmv_obj = -1}
>>>>>>>>>>>>>>                _lgs_cb = {mds_hdl = 65547, mds_role =
>>>>>>> V_DEST_RL_STANDBY,
>>>>>>>>>>>> vaddr
>>>>>>>>>>>>>> = 11, log_version = {releaseCode = 65 'A', majorVersion = 2
>>>>>> '\002',
>>>>>>>>>>>>>> minorVersion = 2 '\002'}, client_tree = {
>>>>>>>>>>>>>>                    root_node = {bit = -1, left =
>> 0x7f97b04cf1b0,
>>>>>> right =
>>>>>>>>>>>>>> 0x7f97b04a2418, key_info = 0x7f97b04b7bd0 ""}, params =
>>>>>>> {key_size
>>>>>>>>>>>>>> = 4},
>>>>>>>>>>>>>> n_nodes = 8}, comp_name = {_opaque = {46, 24947,
>>>>>>>>>>>>>>                      17254, 28015, 15728, 20300, 11335, 
>>>>>>>>>>>>>> 24947,
>>>> 21350,
>>>>>>>>>>>>>> 15733,
>>>>>>>>>>>>>> 17235, 12845, 29484, 26209, 26451, 12861, 11342, 24947,
>>>>> 16742,
>>>>>>>>> 28784,
>>>>>>>>>>>>>> 20285, 25968, 21358, 17985,
>>>>>>>>>>>>>>                      0 <repeats 105 times>}}, amf_hdl =
>>>> 4288675841,
>>>>>>>>>>>>>> amfSelectionObject = 15, amf_invocation_id = 0,
>>> is_quiesced_set
>>>>> =
>>>>>>>>>>> false,
>>>>>>>>>>>>>> immOiHandle = 554050912783, immSelectionObject = 21,
>>>>>>>>>>>>>>                  clmSelectionObject = 17, clm_hdl = 
>>>>>>>>>>>>>> 4279238657,
>>>>>>>>>>>>>> ha_state =
>>>>>>>>>>>>>> SA_AMF_HA_STANDBY, last_client_id = 208, async_upd_cnt =
>>>>> 743,
>>>>>>>>>>>>>> ckpt_state
>>>>>>>>>>>>>> = COLD_SYNC_IDLE, mbcsv_hdl = 4293918753,
>>>>>>>>>>>>>>                  mbcsv_sel_obj = 23, mbcsv_ckpt_hdl =
>> 4292870177,
>>>>>>>>>>>>>> mbcsv_peer_version = 7, edu_hdl = {is_inited = true, tree =
>>>>>>>>>>>>>> {root_node
>>>>>>>>>>> =
>>>>>>>>>>>>>> {bit = -1, left = 0x7f97b04cf2e0,
>>>>>>>>>>>>>>                        right = 0x7f97b04a25b8, key_info =
>>>>>> 0x7f97b04b7d40
>>>>>>>>>>> ""},
>>>>>>>>>>>>>> params = {key_size = 8}, n_nodes = 12}, to_version = 1},
>>>>>>>>>>>>>> fully_initialized = true, lga_down_list_head = 0x0,
>>>>>>>>>>>>>>                  lga_down_list_tail = 0x0, 
>>>>>>>>>>>>>> clm_init_sel_obj =
>>>>>>>>>>>>>> {raise_obj =
>>>>>>>>>>> -1,
>>>>>>>>>>>>>> rmv_obj = -1}, nid_started = true, scAbsenceAllowed = 900,
>>>>>>>>>>>>>> lgs_recovery_state = LGS_NORMAL}
>>>>>>>>>>>>>>                nfds = 7
>>>>>>>>>>>>>>                fds = {{fd = 19, events = 1, revents = 0}, 
>>>>>>>>>>>>>> {fd =
>>>> 15,
>>>>>>>>>>>>>> events =
>>>>>>>>>>>>>> 1, revents = 0}, {fd = 23, events = 1, revents = 1}, {fd 
>>>>>>>>>>>>>> = 13,
>>>>>>>>>>>>>> events =
>>>>>>>>>>>>>> 1, revents = 0}, {fd = -1, events = 1,
>>>>>>>>>>>>>>                    revents = 0}, {fd = 17, events = 1, 
>>>>>>>>>>>>>> revents
>> =
>>>> 0},
>>>>>>>>>>>>>> {fd =
>>>>>>>>>>> 21,
>>>>>>>>>>>>>> events = 1, revents = 0}}
>>>>>>>>>>>>>>                mbox_msgs = {0, 0, 0, 0, 0}
>>>>>>>>>>>>>>                lgs_cb = 0x7f97b04a2400
>>>>>>>>>>>>>>                mbox_low = {0, 0, 0, 0, 0}
>>>>>>>>>>>>>>                lgs_mbox_init_mutex = {__data = {__lock = 0,
>>>> __count =
>>>>>> 0,
>>>>>>>>>>>>>> __owner = 0, __nusers = 0, __kind = 0, __spins = 0, __list =
>>>>>>>>>>>>>> {__prev =
>>>>>>>>>>>>>> 0x0, __next = 0x0}},
>>>>>>>>>>>>>>                  __size = '\000' <repeats 39 times>, 
>>>>>>>>>>>>>> __align =
>> 0}
>>>>>>>>>>>>>> lgs_mbx = 4291821569
>>>>>>>>>>>>>>                mbox_high = {0, 0, 0, 0, 0}
>>>>>>>>>>>>>>                mbox_full = {false, false, false, false, 
>>>>>>>>>>>>>> false}
>>>>>>>>>>>>>> (gdb)
>>>>>>>>>>>>>> (gdb)
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> Feb 22 13:40:55 SC-2 osafimmnd[4584]: NO Ccb 131
>>>>> COMMITTED
>>>>>>>>>>>>>> (immcfg_SC-1_18714)
>>>>>>>>>>>>>> Feb 22 13:40:56 SC-2 osafamfnd[4634]: NO
>>>>>>>>>>>>>> 'safComp=LOG,safSu=SC-2,safSg=2N,safApp=OpenSAF' faulted
>>>>> due
>>>>>>> to
>>>>>>>>>>>>>> 'avaDown' : Recovery is 'nodeFailfast'
>>>>>>>>>>>>>> Feb 22 13:40:56 SC-2 osafamfnd[4634]: ER
>>>>>>>>>>>>>> safComp=LOG,safSu=SC-2,safSg=2N,safApp=OpenSAF Faulted
>>>>> due
>>>>>>>>>>>>>> to:avaDown
>>>>>>>>>>>>>> Recovery is:nodeFailfast
>>>>>>>>>>>>>> Feb 22 13:40:56 SC-2 osafamfnd[4634]: Rebooting OpenSAF
>>>>> NodeId
>>>>>>> =
>>>>>>>>>>>>>> 131599
>>>>>>>>>>>>>> EE Name = , Reason: Component faulted: recovery is node
>>>>> failfast,
>>>>>>>>>>>>>> OwnNodeId = 131599, SupervisionTime = 60
>>>>>>>>>>>>>> Feb 22 13:40:56 SC-2 opensaf_reboot: Rebooting local node;
>>>>>>>>> timeout=60
>>> ==============================================================
>>> ==============================================================
>>>>>>>>>>>>>> On 2/22/2017 12:23 PM, A V Mahesh wrote:
>>>>>>>>>>>>>>> Hi Vu,
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> On 2/22/2017 12:19 PM, Vu Minh Nguyen wrote:
>>>>>>>>>>>>>>>> [Vu] I has sent you 02 patches. There is code change in
>>>>> osaflogd
>>>>>>>>>>>>>>>> code
>>>>>>>>>>>>> that
>>>>>>>>>>>>>>>> fix the coredump you have observed.
>>>>>>>>>>>>>>>> The other one is test code that fix the logtest coredump.
>>>>>>>>>>>>>>> Ok I will re-test , and update you .
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> -AVM
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> On 2/22/2017 12:19 PM, Vu Minh Nguyen wrote:
>>>>>>>>>>>>>>>> Hi Mahehs,
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> See my reply inline, [Vu].
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> Regards, Vu
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> -----Original Message-----
>>>>>>>>>>>>>>>>> From: A V Mahesh [mailto:mahesh.va...@oracle.com]
>>>>>>>>>>>>>>>>> Sent: Wednesday, February 22, 2017 1:36 PM
>>>>>>>>>>>>>>>>> To: Vu Minh Nguyen <vu.m.ngu...@dektech.com.au>;
>>>>>>>>>>>>>>>>> lennart.l...@ericsson.com;
>>> canh.v.tru...@dektech.com.au
>>>>>>>>>>>>>>>>> Cc: opensaf-devel@lists.sourceforge.net
>>>>>>>>>>>>>>>>> Subject: Re: [PATCH 0 of 3] Review Request for log: add
>>>>>>>>>>>>>>>>> alternative
>>>>>>>>>>>>>>>>> destinations of log records [#2258] V4
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> Hi Vu,
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> On 2/22/2017 11:52 AM, Vu Minh Nguyen wrote:
>>>>>>>>>>>>>>>>>> Hi Mahesh,
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> Have a code fault in uml test, and other one in
>>> checkpoint.
>>>>>>>>>>>>>>>>> [AVM] This is Normal Suse 11 VM ( not  UML).
>>>>>>>>>>>>>>>>>> I have just updated the code. Please re-apply for #3 and
>>> #4
>>>>>>>>>>> patches.
>>>>>>>>>>>>>>>>> [AVM] is these new patch has function changes or only 
>>>>>>>>>>>>>>>>> test
>>>>>>> code
>>>>>>>>>>>>>> changes ?
>>>>>>>>>>>>>>>> [Vu] I has sent you 02 patches. There is code change in
>>>>> osaflogd
>>>>>>>>>>>>>>>> code
>>>>>>>>>>>>> that
>>>>>>>>>>>>>>>> fix the coredump you have observed.
>>>>>>>>>>>>>>>> The other one is test code that fix the logtest coredump.
>>>>>>>>>>>>>>>>>> Note that, test case #14 of suite 17 should be run on
>>> active
>>>>>>>>>>>>>>>>>> node,
>>>>>>>>>>>>>>>>> otherwise
>>>>>>>>>>>>>>>>>> getting failed.
>>>>>>>>>>>>>>>>> [AVM]  Segmentation fault of /usr/bin/logtest Not a big
>>>> issue
>>>>>> ,
>>>>>>>>>>>>>>>>>           we need to debug why  osaflogd core dumped 
>>>>>>>>>>>>>>>>> and it
>> is
>>>>>>>>>>>>>>>>> critical
>>>>>>>>>>>>>>>> [Vu] I found the problem. You can try with the new one to
>>> see
>>>>> if
>>>>>>>>>>>>>>>> the
>>>>>>>>>>>>>>>> coredump is still there or not.
>>>>>>>>>>>>>>>>>> I will put condition check to that test case later.
>>>>>>>>>>>>>>>>> -AVM
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> Regards, Vu
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> -----Original Message-----
>>>>>>>>>>>>>>>>>>> From: A V Mahesh [mailto:mahesh.va...@oracle.com]
>>>>>>>>>>>>>>>>>>> Sent: Wednesday, February 22, 2017 12:16 PM
>>>>>>>>>>>>>>>>>>> To: Vu Minh Nguyen <vu.m.ngu...@dektech.com.au>;
>>>>>>>>>>>>>>>>>>> lennart.l...@ericsson.com;
>>>>> canh.v.tru...@dektech.com.au
>>>>>>>>>>>>>>>>>>> Cc: opensaf-devel@lists.sourceforge.net
>>>>>>>>>>>>>>>>>>> Subject: Re: [PATCH 0 of 3] Review Request for log: add
>>>>>>>>>>> alternative
>>>>>>>>>>>>>>>>>>> destinations of log records [#2258] V4
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> Hi Vu,
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> Thanks ,
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> While testing /usr/bin/logtest ,  SC-2 standby osaflogd
>>>>> core
>>>>>>>>>>>> dumped
>>>>>>>>>>>>>>>> and
>>>>>>>>>>>>>>>>>>> /usr/bin/logtest on SC-1 Active
>>>>>>>>>>>>>>>>>>> got Segmentation fault , am I missing any other patch (
>> i
>>>>> am
>>>>>>>>>>> using
>>>>>>>>>>>>>>>>>>> devel published patch only )
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> Following patches i am using :
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>            1) #2293 (sent by Anders Widel, but not yet
>>>> pushed)
>>>>>>>>>>>>>>>>>>> 2) #2258 (v2, sent by Lennart, but not yet
>>>> pushed
>>>>>> yet)
>>>>>>>>>>>>>>>>>>> 3) #2258 (v4, sent by Vu, but not yet pushed
>>>> yet)
>>> ==============================================================
>>>>>>>>>>>>>>>>>>> ========================================
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> Core was generated by `/usr/lib64/opensaf/osaflogd'.
>>>>>>>>>>>>>>>>>>> Program terminated with signal 11, Segmentation fault.
>>>>>>>>>>>>>>>>>>> #0 ckpt_proc_cfg_stream(lgs_cb*, void*) () at
>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc:2195
>>>>>>>>>>>>>>>>>>> 2195 src/log/logd/lgs_mbcsv.cc: No such file or
>>>>> directory.
>>>>>>>>>>>>>>>>>>> in src/log/logd/lgs_mbcsv.cc
>>>>>>>>>>>>>>>>>>> (gdb) bt
>>>>>>>>>>>>>>>>>>> #0 ckpt_proc_cfg_stream(lgs_cb*, void*) () at
>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc:2195
>>>>>>>>>>>>>>>>>>> #1  0x00007f12c3e22960 in
>>>>> ckpt_decode_log_struct(lgs_cb*,
>>>>>>>>>>>>>>>>>>> ncs_mbcsv_cb_arg*, void*, void*, unsigned int
>>>>>>>>> (*)(edu_hdl_tag*,
>>>>>>>>>>>>>>>>>>> edu_tkn_tag*, void*, unsigned int*, edu_buf_env_tag*,
>>>>>>>>>>>>>> EDP_OP_TYPE,
>>>>>>>>>>>>>>>>>>> EDU_ERR*)) () at src/log/logd/lgs_mbcsv.cc:950
>>>>>>>>>>>>>>>>>>> #2  0x00007f12c3e240dc in
>>>>>>>>> ckpt_decode_async_update(lgs_cb*,
>>>>>>>>>>>>>>>>>>> ncs_mbcsv_cb_arg*) () at
>>> src/log/logd/lgs_mbcsv.cc:1086
>>>>>>>>>>>>>>>>>>> #3 0x00007f12c3e26941 in
>>>>>>>>> mbcsv_callback(ncs_mbcsv_cb_arg*) ()
>>>>>>>>>>>> at
>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc:880
>>>>>>>>>>>>>>>>>>> #4  0x00007f12c2f25596 in ncs_mbscv_rcv_decode ()
>>> from
>>>>>>>>>>>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0
>>>>>>>>>>>>>>>>>>> #5  0x00007f12c2f25766 in
>>> ncs_mbcsv_rcv_async_update
>>>>> ()
>>>>>>>>> from
>>>>>>>>>>>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0
>>>>>>>>>>>>>>>>>>> #6  0x00007f12c2f2c370 in mbcsv_process_events ()
>>> from
>>>>>>>>>>>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0
>>>>>>>>>>>>>>>>>>> #7  0x00007f12c2f2c4db in mbcsv_hdl_dispatch_all ()
>>> from
>>>>>>>>>>>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0
>>>>>>>>>>>>>>>>>>> #8  0x00007f12c2f26ce2 in
>>>>> mbcsv_process_dispatch_request
>>>>>>> ()
>>>>>>>>> at
>>>>>>>>>>>>>>>>>>> src/mbc/mbcsv_api.c:423
>>>>>>>>>>>>>>>>>>> #9  0x00007f12c3e2396e in
>>> lgs_mbcsv_dispatch(unsigned
>>>>> int)
>>>>>>> ()
>>>>>>>>> at
>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc:327
>>>>>>>>>>>>>>>>>>> #10 0x00007f12c3e009f2 in main () at
>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_main.cc:583
>>>>>>>>>>>>>>>>>>> (gdb)
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>
>>> ==============================================================
>>>>>>>>>>>>>>>>>>> ========================================
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> Feb 22 10:37:06 SC-1 osafimmnd[4020]: NO Invalid error
>>>>>>>>> reported
>>>>>>>>>>>>>>>>>>> implementer 'safLogService', Ccb 161 will be aborted
>>>>>>>>>>>>>>>>>>> Feb 22 10:37:06 SC-1 osafimmnd[4020]: NO Ccb 161
>>>>> aborted
>>>>>>> in
>>>>>>>>>>>>>>>>> COMPLETED
>>>>>>>>>>>>>>>>>>> processing (validation)
>>>>>>>>>>>>>>>>>>> Feb 22 10:37:06 SC-1 osafimmnd[4020]: NO Ccb 161
>>>>>>> ABORTED
>>>>>>>>>>>>>>>>> (immcfg_SC-
>>>>>>>>>>>>>>>>>>> 1_5394)
>>>>>>>>>>>>>>>>>>> Add values Fail
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> Program received signal SIGSEGV, Segmentation fault.
>>>>>>>>>>>>>>>>>>> 0x000055555556929a in read_and_compare.isra.7 () at
>>>>>>>>>>>>>>>>>>> src/log/apitest/tet_LogOiOps.c:1891
>>>>>>>>>>>>>>>>>>> 1891 src/log/apitest/tet_LogOiOps.c: No such file or
>>>>>>>>>>>>>>>>>>> directory.
>>>>>>>>>>>>>>>>>>>                   in src/log/apitest/tet_LogOiOps.c
>>>>>>>>>>>>>>>>>>> (gdb) Feb 22 10:37:07 SC-1 sshd[5298]: Accepted
>>> keyboard-
>>>>>>>>>>>>>>>>> interactive/pam
>>>>>>>>>>>>>>>>>>> for root from 10.176.178.22 port 51945 ssh2
>>>>>>>>>>>>>>>>>>> bt
>>>>>>>>>>>>>>>>>>> #0  0x000055555556929a in read_and_compare.isra.7 ()
>>> at
>>>>>>>>>>>>>>>>>>> src/log/apitest/tet_LogOiOps.c:1891
>>>>>>>>>>>>>>>>>>> #1  0x0000555555569bbb in
>>>>>>>>>>>>>>>>>>> check_logRecordDestinationConfigurationEmpty
>>>>>>>>>>>>>>>>>>> () at src/log/apitest/tet_LogOiOps.c:2179
>>>>>>>>>>>>>>>>>>> #2  0x0000555555573495 in run_test_case ()
>>>>>>>>>>>>>>>>>>> #3  0x0000555555573934 in test_run ()
>>>>>>>>>>>>>>>>>>> #4  0x000055555555c7cd in main () at
>>>>>>>>>>>>>>>>>>> src/log/apitest/logtest.c:569
>>>>>>>>>>>>>>>>>>> (gdb)
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>
>>> ==============================================================
>>>>>>>>>>>>>>>>>>> ========================================
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> -AVM
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> On 2/22/2017 9:48 AM, Vu Minh Nguyen wrote:
>>>>>>>>>>>>>>>>>>>> Hi Mahesh,
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>> I send them in attachment instead, and name them in
>>> the
>>>>>>>>> order.
>>>>>>>>>>>>>>>>>>>> I just pull the latest code, and apply them without
>>>> getting
>>>>>>>> any
>>>>>>>>>>>>> hunk
>>>>>>>>>>>>>>>>>> error.
>>>>>>>>>>>>>>>>>>>> Please try with them, and let me know if you see any
>>>>>>> problem.
>>>>>>>>>>>>>>>>>>>> Regards, Vu
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>> -----Original Message-----
>>>>>>>>>>>>>>>>>>>>> From: A V Mahesh [mailto:mahesh.va...@oracle.com]
>>>>>>>>>>>>>>>>>>>>> Sent: Wednesday, February 22, 2017 11:09 AM
>>>>>>>>>>>>>>>>>>>>> To: Vu Minh Nguyen
>>> <vu.m.ngu...@dektech.com.au>;
>>>>>>>>>>>>>>>>>>>>> lennart.l...@ericsson.com;
>>>>>>> canh.v.tru...@dektech.com.au
>>>>>>>>>>>>>>>>>>>>> Cc: opensaf-devel@lists.sourceforge.net
>>>>>>>>>>>>>>>>>>>>> Subject: Re: [PATCH 0 of 3] Review Request for log:
>>> add
>>>>>>>>>>>>> alternative
>>>>>>>>>>>>>>>>>>>>> destinations of log records [#2258] V4
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>> Hi Vu,
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>> I did follow that still i get Hunk #2 FAILED even on
>>>>>> today's
>>>>>>>>>>>>> staging
>>> ==============================================================
>>>>>>>>>>>>>>>>>>>>> ==================
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>> [root@dhcp-hyd-scp-5fl-10-176-177-96 staging]#
>>> patch
>>>>> -
>>>>>>> p1
>>>>>>>>>>>> <2293
>>>>>>>>>>>>>>>>>>>>> patching file src/base/Makefile.am
>>>>>>>>>>>>>>>>>>>>> Hunk #1 succeeded at 33 (offset 1 line).
>>>>>>>>>>>>>>>>>>>>> Hunk #3 succeeded at 183 (offset 1 line).
>>>>>>>>>>>>>>>>>>>>> patching file src/base/file_descriptor.cc
>>>>>>>>>>>>>>>>>>>>> patching file src/base/file_descriptor.h
>>>>>>>>>>>>>>>>>>>>> patching file src/base/tests/unix_socket_test.cc
>>>>>>>>>>>>>>>>>>>>> patching file src/base/unix_client_socket.cc
>>>>>>>>>>>>>>>>>>>>> patching file src/base/unix_server_socket.cc
>>>>>>>>>>>>>>>>>>>>> patching file src/base/unix_socket.cc
>>>>>>>>>>>>>>>>>>>>> patching file src/base/unix_socket.h
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>> [root@dhcp-hyd-scp-5fl-10-176-177-96 staging]#
>>> patch
>>>>> -
>>>>>>> p1
>>>>>>>>>>>> <2258-
>>>>>>>>>>>>>> 1
>>>>>>>>>>>>>>>>>>>>> patching file src/log/Makefile.am
>>>>>>>>>>>>>>>>>>>>> Hunk #1 succeeded at 71 (offset -1 lines).
>>>>>>>>>>>>>>>>>>>>> patching file src/log/config/logsv_classes.xml
>>>>>>>>>>>>>>>>>>>>> Hunk #1 FAILED at 147.
>>>>>>>>>>>>>>>>>>>>> 1 out of 1 hunk FAILED -- saving rejects to file
>>>>>>>>>>>>>>>>>>>>> src/log/config/logsv_classes.xml.rej
>>>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_config.cc
>>>>>>>>>>>>>>>>>>>>> Hunk #1 succeeded at 35 (offset -5 lines).
>>>>>>>>>>>>>>>>>>>>> Hunk #2 FAILED at 705.
>>>>>>>>>>>>>>>>>>>>> Hunk #3 FAILED at 971.
>>>>>>>>>>>>>>>>>>>>> 2 out of 3 hunks FAILED -- saving rejects to file
>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_config.cc.rej
>>>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_config.h
>>>>>>>>>>>>>>>>>>>>> Hunk #1 FAILED at 304.
>>>>>>>>>>>>>>>>>>>>> 1 out of 1 hunk FAILED -- saving rejects to file
>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_config.h.rej
>>>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_dest.cc
>>>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_dest.h
>>>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_evt.cc
>>>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_imm.cc
>>>>>>>>>>>>>>>>>>>>> Hunk #1 FAILED at 45.
>>>>>>>>>>>>>>>>>>>>> Hunk #2 succeeded at 235 (offset -1 lines).
>>>>>>>>>>>>>>>>>>>>> Hunk #3 FAILED at 877.
>>>>>>>>>>>>>>>>>>>>> Hunk #4 succeeded at 1273 (offset -20 lines).
>>>>>>>>>>>>>>>>>>>>> Hunk #5 succeeded at 1404 (offset -1 lines).
>>>>>>>>>>>>>>>>>>>>> Hunk #6 succeeded at 1449 (offset -20 lines).
>>>>>>>>>>>>>>>>>>>>> Hunk #7 succeeded at 2032 (offset -1 lines).
>>>>>>>>>>>>>>>>>>>>> Hunk #8 FAILED at 2181.
>>>>>>>>>>>>>>>>>>>>> Hunk #9 succeeded at 2271 (offset -54 lines).
>>>>>>>>>>>>>>>>>>>>> Hunk #10 succeeded at 2387 (offset -1 lines).
>>>>>>>>>>>>>>>>>>>>> Hunk #11 succeeded at 2377 (offset -54 lines).
>>>>>>>>>>>>>>>>>>>>> Hunk #12 succeeded at 2478 (offset -1 lines).
>>>>>>>>>>>>>>>>>>>>> Hunk #13 succeeded at 2684 (offset -54 lines).
>>>>>>>>>>>>>>>>>>>>> Hunk #14 succeeded at 2821 (offset -1 lines).
>>>>>>>>>>>>>>>>>>>>> 3 out of 14 hunks FAILED -- saving rejects to file
>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_imm.cc.rej
>>>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_main.cc
>>>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_mbcsv.cc
>>>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_mbcsv.h
>>>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_mbcsv_v5.cc
>>>>>>>>>>>>>>>>>>>>> Hunk #3 succeeded at 133 (offset -1 lines).
>>>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_mbcsv_v7.cc
>>>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_mbcsv_v7.h
>>>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_stream.cc
>>>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_stream.h
>>>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_util.cc
>>>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_util.h
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>> [root@dhcp-hyd-scp-5fl-10-176-177-96 staging]#
>>> patch
>>>>> -
>>>>>>> p1
>>>>>>>>>>>> <2258-
>>>>>>>>>>>>>> 2
>>>>>>>>>>>>>>>>>>>>> patching file src/log/Makefile.am
>>>>>>>>>>>>>>>>>>>>> Hunk #1 succeeded at 180 (offset -3 lines).
>>>>>>>>>>>>>>>>>>>>> patching file src/log/apitest/tet_LogOiOps.c
>>>>>>>>>>>>>>>>>>>>> Hunk #1 FAILED at 1923.
>>>>>>>>>>>>>>>>>>>>> Hunk #2 FAILED at 1979.
>>>>>>>>>>>>>>>>>>>>> Hunk #3 FAILED at 2067.
>>>>>>>>>>>>>>>>>>>>> Hunk #4 FAILED at 2094.
>>>>>>>>>>>>>>>>>>>>> 4 out of 4 hunks FAILED -- saving rejects to file
>>>>>>>>>>>>>>>>>>>>> src/log/apitest/tet_LogOiOps.c.rej
>>>>>>>>>>>>>>>>>>>>> patching file src/log/apitest/tet_cfg_destination.c
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>> [root@dhcp-hyd-scp-5fl-10-176-177-96 staging]#
>>> patch
>>>>> -
>>>>>>> p1
>>>>>>>>>>>> <2258-
>>>>>>>>>>>>>> 3
>>>>>>>>>>>>>>>>>>>>> patching file src/log/Makefile
>>>>>>>>>>>>>>>>>>>>> patching file src/log/Makefile.am
>>>>>>>>>>>>>>>>>>>>> Hunk #1 succeeded at 80 (offset -1 lines).
>>>>>>>>>>>>>>>>>>>>> Hunk #2 succeeded at 217 (offset -2 lines).
>>>>>>>>>>>>>>>>>>>>> patching file src/log/tests/Makefile
>>>>>>>>>>>>>>>>>>>>> patching file src/log/tests/lgs_dest_test.cc
>>>>>>>>>>>>>>>>>>>>> [root@dhcp-hyd-scp-5fl-10-176-177-96 staging]# vi
>>>>>>>>>>>>>>>>>>>>> src/log/apitest/tet_LogOiOps.c.rej
>>>>>>>>>>>>>>>>>>>>> [root@dhcp-hyd-scp-5fl-10-176-177-96 staging]#
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>
>>> ==============================================================
>>>>>>>>>>>>>>>>>>>>> ========================
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>> -AVM
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>> On 2/21/2017 3:53 PM, Vu Minh Nguyen wrote:
>>>>>>>>>>>>>>>>>>>>>> Hi Mahesh,
>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>> As I has mentioned in below:
>>>>>>>>>>>>>>>>>>>>>>>> To run the test, this patch has
>> dependent
>>>> on
>>>>>>>>>>> following
>>>>>>>>>>>>>>>> patches:
>>>>>>>>>>>>>>>>>>>>>>>> 1) #2293 (sent by Anders Widel, but not
>> yet
>>>>>>>>>>>>>>>>>>>>>>>> pushed)
>>>>>>>>>>>>>>>>>>>>>>>>              2) #2258 (v2, sent by Lennart, but 
>>>>>>>>>>>>>>>>>>>>>>>> not
>> yet
>>>>>> pushed
>>>>>>>>>>> yet)
>>>>>>>>>>>>>>>>>>>>>> So, you need to apply #2293 first, then #2258 which
>>>>> sent
>>>>>>> by
>>>>>>>>>>>>>> Lennart
>>>>>>>>>>>>>>>>>>>>>> yesterday, then mine.
>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>> Regards, Vu
>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> -----Original Message-----
>>>>>>>>>>>>>>>>>>>>>>> From: A V Mahesh
>>> [mailto:mahesh.va...@oracle.com]
>>>>>>>>>>>>>>>>>>>>>>> Sent: Tuesday, February 21, 2017 5:10 PM
>>>>>>>>>>>>>>>>>>>>>>> To: Vu Minh Nguyen
>>>>> <vu.m.ngu...@dektech.com.au>;
>>>>>>>>>>>>>>>>>>>>>>> lennart.l...@ericsson.com;
>>>>>>>>> canh.v.tru...@dektech.com.au
>>>>>>>>>>>>>>>>>>>>>>> Cc: opensaf-devel@lists.sourceforge.net
>>>>>>>>>>>>>>>>>>>>>>> Subject: Re: [PATCH 0 of 3] Review Request for log:
>>>>> add
>>>>>>>>>>>>>> alternative
>>>>>>>>>>>>>>>>>>>>>>> destinations of log records [#2258] V4
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> Hi Vu,
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> Is this applies on top of   log #2146  - V4 , I see
>>>>>> both
>>>>>>>>>>>>> #tickets
>>>>>>>>>>>>>>>>>>>> has
>>>>>>>>>>>>>>>>>>>>>>> version changes ?
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> in which order i need to apply  ( #2146 & #2258 )
>>> or
>>>>>>>>>>>>>>>>>>>>>>> (#2258
>>>>>>>>>>> &
>>>>>>>>>>>>>>>>>>>> #2146).
>>> =========================================================
>>>>>>>>>>>>>>>>>>>>>>> patching file src/log/Makefile.am
>>>>>>>>>>>>>>>>>>>>>>> Hunk #1 FAILED at 72.
>>>>>>>>>>>>>>>>>>>>>>> Hunk #2 FAILED at 120.
>>>>>>>>>>>>>>>>>>>>>>> 2 out of 2 hunks FAILED -- saving rejects to file
>>>>>>>>>>>>>>>>>>>> src/log/Makefile.am.rej
>>>>>>>>>>>>>>>>>>>>>>> patching file src/log/config/logsv_classes.xml
>>>>>>>>>>>>>>>>>>>>>>> Hunk #1 FAILED at 147.
>>>>>>>>>>>>>>>>>>>>>>> 1 out of 1 hunk FAILED -- saving rejects to file
>>>>>>>>>>>>>>>>>>>>>>> src/log/config/logsv_classes.xml.rej
>>>>>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_config.cc
>>>>>>>>>>>>>>>>>>>>>>> Hunk #1 succeeded at 35 (offset -5 lines).
>>>>>>>>>>>>>>>>>>>>>>> Hunk #2 FAILED at 705.
>>>>>>>>>>>>>>>>>>>>>>> Hunk #3 FAILED at 971.
>>>>>>>>>>>>>>>>>>>>>>> 2 out of 3 hunks FAILED -- saving rejects to file
>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_config.cc.rej
>>>>>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_config.h
>>>>>>>>>>>>>>>>>>>>>>> Hunk #1 FAILED at 304.
>>>>>>>>>>>>>>>>>>>>>>> 1 out of 1 hunk FAILED -- saving rejects to file
>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_config.h.rej
>>>>>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_dest.cc
>>>>>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_dest.h
>>>>>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_evt.cc
>>>>>>>>>>>>>>>>>>>>>>> Hunk #1 FAILED at 1.
>>>>>>>>>>>>>>>>>>>>>>> Hunk #2 succeeded at 30 with fuzz 2 (offset 2 
>>>>>>>>>>>>>>>>>>>>>>> lines).
>>>>>>>>>>>>>>>>>>>>>>> Hunk #3 succeeded at 1282 (offset 45 lines).
>>>>>>>>>>>>>>>>>>>>>>> Hunk #4 succeeded at 1300 (offset 2 lines).
>>>>>>>>>>>>>>>>>>>>>>> 1 out of 4 hunks FAILED -- saving rejects to file
>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_evt.cc.rej
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>
>>> ==============================================================
>>>>>>>>>>>>>>>>>>>>>>> ===
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> -AVM
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> On 2/21/2017 3:03 PM, Vu Minh Nguyen wrote:
>>>>>>>>>>>>>>>>>>>>>>>> Summary: log: add alternative destinations of log
>>>>>>> records
>>>>>>>>>>>>>> [#2258]
>>>>>>>>>>>>>>>>>>>>>>>> Review request for Trac Ticket(s): #2258
>>>>>>>>>>>>>>>>>>>>>>>> Peer Reviewer(s): Lennart, Canh, Mahesh
>>>>>>>>>>>>>>>>>>>>>>>> Pull request to: <<LIST THE PERSON WITH PUSH
>>>>>>> ACCESS
>>>>>>>>>>>> HERE>>
>>>>>>>>>>>>>>>>>>>>>>>> Affected branch(es): Default
>>>>>>>>>>>>>>>>>>>>>>>> Development branch: Default
>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>> --------------------------------
>>>>>>>>>>>>>>>>>>>>>>>> Impacted area Impact y/n
>>>>>>>>>>>>>>>>>>>>>>>> --------------------------------
>>>>>>>>>>>>>>>>>>>>>>>> Docs n
>>>>>>>>>>>>>>>>>>>>>>>>              Build system            n
>>>>>>>>>>>>>>>>>>>>>>>> RPM/packaging n
>>>>>>>>>>>>>>>>>>>>>>>> Configuration files     n
>>>>>>>>>>>>>>>>>>>>>>>>              Startup scripts         n
>>>>>>>>>>>>>>>>>>>>>>>>              SAF services n
>>>>>>>>>>>>>>>>>>>>>>>>              OpenSAF services        y
>>>>>>>>>>>>>>>>>>>>>>>>              Core libraries          n
>>>>>>>>>>>>>>>>>>>>>>>> Samples n
>>>>>>>>>>>>>>>>>>>>>>>> Tests y
>>>>>>>>>>>>>>>>>>>>>>>> Other n
>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>> Comments (indicate scope for each "y" above):
>>>>>>>>>>>>>>>>>>>>>>>> ---------------------------------------------
>>>>>>>>>>>>>>>>>>>>>>>>              To run the test, this patch has
>> dependent
>>>> on
>>>>>>>>>>> following
>>>>>>>>>>>>>>>> patches:
>>>>>>>>>>>>>>>>>>>>>>>> 1) #2293 (sent by Anders Widel, but not
>> yet
>>>>>>>>>>>>>>>>>>>>>>>> pushed)
>>>>>>>>>>>>>>>>>>>>>>>>              2) #2258 (v2, sent by Lennart, but 
>>>>>>>>>>>>>>>>>>>>>>>> not
>> yet
>>>>>> pushed
>>>>>>>>>>> yet)
>>>>>>>>>>>>>>>>>>>>>>>> changeset
>>>>>>>>> d74aaf3025c99cade3165a15831124548f4d85bd
>>>>>>>>>>>>>>>>>>>>>>>> Author: Vu Minh Nguyen
>>>>>>>>> <vu.m.ngu...@dektech.com.au>
>>>>>>>>>>>>>>>>>>>>>>>> Date: Wed, 15 Feb 2017 14:36:00 +0700
>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>         log: add alternative destinations of log
>> records
>>>>>>>>>>>>>>>>>>>>>>>> [#2258]
>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>         Here are major info, detailed info will be
>>> added
>>>> to
>>>>>>>>>>>>>>>>>>>>>>>> PR doc
>>>>>>>>>>>>>>>>>> soon. 1)
>>>>>>>>>>>>>>>>>>>>>>> Add
>>>>>>>>>>>>>>>>>>>>>>>> attribute "saLogRecordDestination" to log
>>>>> stream.
>>>>>> 2)
>>>>>>>>>>>>>>>>>>>>>>>> Add
>>>>>>>>>>>>>>>>>> Local
>>>>>>>>>>>>>>>>>>>>>>> socket
>>>>>>>>>>>>>>>>>>>>>>>> destintion handler 3) Integrate into first
>>>>>> increment
>>>>>>>>>>>>>>>>>>>>>>>> made by
>>>>>>>>>>>>>>>>>> Lennart
>>>>>>>>>>>>>>>>>>>>>>>> changeset
>>>>>>>>> 4bae27a478c235df3058f43c92d3a5483233b01d
>>>>>>>>>>>>>>>>>>>>>>>> Author: Vu Minh Nguyen
>>>>>>>>> <vu.m.ngu...@dektech.com.au>
>>>>>>>>>>>>>>>>>>>>>>>> Date: Wed, 15 Feb 2017 15:07:09 +0700
>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>         log: add UML test case to verify 
>>>>>>>>>>>>>>>>>>>>>>>> alternative
>>>>>>>>>>>>>>>>>>>>>>>> destination
>>>>>>>>>>>>>>>>>> [#2258]
>>>>>>>>>>>>>>>>>>>>>>>> Major changes: 1) Modify Lennart's test cases
>>>>>>> because
>>>>>>>>>>>>>>>>>> enhancing
>>>>>>>>>>>>>>>>>>>>>>> destination
>>>>>>>>>>>>>>>>>>>>>>>> configuration validation rules. 2) Add test
>>>> suite
>>>>>>>>>>>>>>>>>>>>>>>> #17 to
>>>>>>>>>>>>>>>>>> verify
>>>>>>>>>>>>>>>>>>>>>>> alternative
>>>>>>>>>>>>>>>>>>>>>>>> destination
>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>> changeset
>>>>>>> bc375725fed22bb4f8cb3ae3df5f96fb9d281efb
>>>>>>>>>>>>>>>>>>>>>>>> Author: Vu Minh Nguyen
>>>>>>>>> <vu.m.ngu...@dektech.com.au>
>>>>>>>>>>>>>>>>>>>>>>>> Date: Thu, 16 Feb 2017 17:22:13 +0700
>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>         log: add unit tests to verify interfaces
>>>> provided
>>>>>> by
>>>>>>>>>>>>>>>>>> destination
>>>>>>>>>>>>>>>>>>>>>>> handler
>>>>>>>>>>>>>>>>>>>>>>>> [#2258]
>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>         Unit tests to verify major interfaces: 1)
>>>>>>>>>>>>>>>>>>>>>>>> CfgDestination()
>>>>>>>>>>>>>>>>>> 2)
>>>>>>>>>>>>>>>>>>>>>>>> WriteToDestination()
>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>> Added Files:
>>>>>>>>>>>>>>>>>>>>>>>> ------------
>>>>>>>>>>>>>>>>>>>>>>>> src/log/apitest/tet_cfg_destination.c
>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_dest.cc
>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_dest.h
>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv_v7.cc
>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv_v7.h
>>>>>>>>>>>>>>>>>>>>>>>> src/log/tests/lgs_dest_test.cc
>>>>>>>>>>>>>>>>>>>>>>>> src/log/tests/Makefile
>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>> Complete diffstat:
>>>>>>>>>>>>>>>>>>>>>>>> ------------------
>>>>>>>>>>>>>>>>>>>>>>>> src/log/Makefile |    4 +
>>>>>>>>>>>>>>>>>>>>>>>> src/log/Makefile.am |   31 +++++-
>>>>>>>>>>>>>>>>>>>>>>>> src/log/apitest/tet_LogOiOps.c |    8 +-
>>>>>>>>>>>>>>>>>>>>>>>> src/log/apitest/tet_cfg_destination.c |  483
>>> ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++
>>>>>>>>>>>>>>>>>>>>>>> ++++++++++++++++++++++++++++++++++
>>>>>>>>>>>>>>>>>>>>>>>> src/log/config/logsv_classes.xml |    7 +-
>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_config.cc |  169
>>>>>>>>>>>>>>>>>>>>>>> +++++++++++++++++++++++++++++---
>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_config.h |    3 +-
>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_dest.cc |  707
>>> ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++
>>> ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++
>>>>>>>>>>>>>>>>>>>>>>> +++++++++++++++++
>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_dest.h |  576
>>> ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++
>>>>>>>>>>>> ++++++++++++++++++++++++++++++++++++++++++++++++++++
>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_evt.cc |   33 ++++++
>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_imm.cc |  202
>>>>>>>>>>>>>>>>>>>>>>> +++++++++++++++++++++++++++++++++------
>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_main.cc |    8 +
>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc |  103
>>>>>>>>>>>>>>>> ++++++++++++++++++-
>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv.h |    6 +-
>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv_v5.cc |   10 +
>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv_v7.cc |  177
>>>>>>>>>>>>>>>>>>>>>>> +++++++++++++++++++++++++++++++++++
>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv_v7.h |   67
>>>>>>>>>>>>> +++++++++++++
>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_stream.cc |   60
>>>>>>>>>>>>> +++++++++++-
>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_stream.h |   16 +++
>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_util.cc |   63
>>>>>>>>>>>>> ++++++++++++
>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_util.h |   11 +-
>>>>>>>>>>>>>>>>>>>>>>>> src/log/tests/Makefile |   20 +++
>>>>>>>>>>>>>>>>>>>>>>>> src/log/tests/lgs_dest_test.cc |  209
>>>>>>>>>>>>>>>>>>>>>>> +++++++++++++++++++++++++++++++++++++++++
>>>>>>>>>>>>>>>>>>>>>>>> 23 files changed, 2896 insertions(+), 77
>>>>>>>>>>>>>>>>>>>>>>>> deletions(-)
>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>> Testing Commands:
>>>>>>>>>>>>>>>>>>>>>>>> -----------------
>>>>>>>>>>>>>>>>>>>>>>>>              Run UML test suite #17
>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>> Testing, Expected Results:
>>>>>>>>>>>>>>>>>>>>>>>> --------------------------
>>>>>>>>>>>>>>>>>>>>>>>>              All test passed
>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>> Conditions of Submission:
>>>>>>>>>>>>>>>>>>>>>>>> -------------------------
>>>>>>>>>>>>>>>>>>>>>>>> <<HOW MANY DAYS BEFORE PUSHING,
>>>>>>> CONSENSUS
>>>>>>>>>>>> ETC>>
>>>>>>>>>>>>>>>>>>>>>>>> Arch Built     Started Linux distro
>>>>>>>>>>>>>>>>>>>>>>>> -------------------------------------------
>>>>>>>>>>>>>>>>>>>>>>>> mips n          n
>>>>>>>>>>>>>>>>>>>>>>>> mips64 n          n
>>>>>>>>>>>>>>>>>>>>>>>> x86 n          n
>>>>>>>>>>>>>>>>>>>>>>>> x86_64 n          n
>>>>>>>>>>>>>>>>>>>>>>>> powerpc n          n
>>>>>>>>>>>>>>>>>>>>>>>> powerpc64 n          n
>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>> Reviewer Checklist:
>>>>>>>>>>>>>>>>>>>>>>>> -------------------
>>>>>>>>>>>>>>>>>>>>>>>> [Submitters: make sure that your review doesn't
>>>>> trigger
>>>>>>>>> any
>>>>>>>>>>>>>>>>>>>>> checkmarks!]
>>>>>>>>>>>>>>>>>>>>>>>> Your checkin has not passed review because (see
>>>>>>> checked
>>>>>>>>>>>>>> entries):
>>>>>>>>>>>>>>>>>>>>>>>> ___ Your RR template is generally incomplete; it
>>> has
>>>>>>> too
>>>>>>>>>>>> many
>>>>>>>>>>>>>>>> blank
>>>>>>>>>>>>>>>>>>>>>> entries
>>>>>>>>>>>>>>>>>>>>>>>> that need proper data filled in.
>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>> ___ You have failed to nominate the proper
>>> persons
>>>>> for
>>>>>>>>>>>> review
>>>>>>>>>>>>>> and
>>>>>>>>>>>>>>>>>>>> push.
>>>>>>>>>>>>>>>>>>>>>>>> ___ Your patches do not have proper short+long
>>>>>>> header
>>>>>>>>>>>>>>>>>>>>>>>> ___ You have grammar/spelling in your header
>>> that
>>>>> is
>>>>>>>>>>>>>> unacceptable.
>>>>>>>>>>>>>>>>>>>>>>>> ___ You have exceeded a sensible line length in
>>> your
>>>>>>>>>>>>>>>>>>>>>>> headers/comments/text.
>>>>>>>>>>>>>>>>>>>>>>>> ___ You have failed to put in a proper Trac Ticket
>>> #
>>>>>> into
>>>>>>>>>>> your
>>>>>>>>>>>>>>>>>>>> commits.
>>>>>>>>>>>>>>>>>>>>>>>> ___ You have incorrectly put/left internal data in
>>>>> your
>>>>>>>>>>>>>>>>>> comments/files
>>>>>>>>>>>>>>>>>>>>>>>> (i.e. internal bug tracking tool IDs, product
>>>>>>>>>>> names
>>>>>>>>>>>>> etc)
>>>>>>>>>>>>>>>>>>>>>>>> ___ You have not given any evidence of testing
>>>>> beyond
>>>>>>>>> basic
>>>>>>>>>>>>>> build
>>>>>>>>>>>>>>>>>>>> tests.
>>>>>>>>>>>>>>>>>>>>>>>> Demonstrate some level of runtime or other sanity
>>>>>>>>>>>>> testing.
>>>>>>>>>>>>>>>>>>>>>>>> ___ You have ^M present in some of your files.
>>>>> These
>>>>>>>>>>>>>>>>>>>>>>>> have to
>>>>>>>>>>>>>> be
>>>>>>>>>>>>>>>>>>>>> removed.
>>>>>>>>>>>>>>>>>>>>>>>> ___ You have needlessly changed whitespace or
>>>>> added
>>>>>>>>>>>>>> whitespace
>>>>>>>>>>>>>>>>>>>>> crimes
>>>>>>>>>>>>>>>>>>>>>>>> like trailing spaces, or spaces
>> before
>>>>>> tabs.
>>>>>>>>>>>>>>>>>>>>>>>> ___ You have mixed real technical changes with
>>>>>>>>> whitespace
>>>>>>>>>>>> and
>>>>>>>>>>>>>>>>> other
>>>>>>>>>>>>>>>>>>>>>>>> cosmetic code cleanup changes. These have to be
>>>>>>>>>>>>>> separate
>>>>>>>>>>>>>>>>>>>> commits.
>>>>>>>>>>>>>>>>>>>>>>>> ___ You need to refactor your submission into
>>>>> logical
>>>>>>>>>>>>>>>>>>>>>>>> chunks;
>>>>>>>>>>>>>>>> there
>>>>>>>>>>>>>>>>>> is
>>>>>>>>>>>>>>>>>>>>>>>> too much content into a single
>> commit.
>>>>>>>>>>>>>>>>>>>>>>>> ___ You have extraneous garbage in your review
>>>>>>> (merge
>>>>>>>>>>>>>> commits
>>>>>>>>>>>>>>>>> etc)
>>>>>>>>>>>>>>>>>>>>>>>> ___ You have giant attachments which should
>>> never
>>>>>>> have
>>>>>>>>>>>> been
>>>>>>>>>>>>>>>>> sent;
>>>>>>>>>>>>>>>>>>>>>>>> Instead you should place your content in a public
>>>>>>>>>>>>> tree to
>>>>>>>>>>>>>>>> be
>>>>>>>>>>>>>>>>>>>>>> pulled.
>>>>>>>>>>>>>>>>>>>>>>>> ___ You have too many commits attached to an e-
>>>>> mail;
>>>>>>>>>>>> resend
>>>>>>>>>>>>>> as
>>>>>>>>>>>>>>>>>>>>> threaded
>>>>>>>>>>>>>>>>>>>>>>>> commits, or place in a public tree for a pull.
>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>> ___ You have resent this content multiple times
>>>>>>> without a
>>>>>>>>>>>> clear
>>>>>>>>>>>>>>>>>>>>>> indication
>>>>>>>>>>>>>>>>>>>>>>>> of what has changed between each
>>>> re-send.
>>>>>>>>>>>>>>>>>>>>>>>> ___ You have failed to adequately and individually
>>>>>>> address
>>>>>>>>>>> all
>>>>>>>>>>>>> of
>>>>>>>>>>>>>>>>>> the
>>>>>>>>>>>>>>>>>>>>>>>> comments and change requests that were
>>> proposed
>>>>> in
>>>>>>>>>>>>> the
>>>>>>>>>>>>>>>>>> initial
>>>>>>>>>>>>>>>>>>>>>> review.
>>>>>>>>>>>>>>>>>>>>>>>> ___ You have a misconfigured ~/.hgrc file (i.e.
>>>>>>> username,
>>>>>>>>>>>> email
>>>>>>>>>>>>>>>> etc)
>>>>>>>>>>>>>>>>>>>>>>>> ___ Your computer have a badly configured date
>>> and
>>>>>>>>> time;
>>>>>>>>>>>>>>>>> confusing
>>>>>>>>>>>>>>>>>>> the
>>>>>>>>>>>>>>>>>>>>>>>> the threaded patch review.
>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>> ___ Your changes affect IPC mechanism, and you
>>>>> don't
>>>>>>>>>>>> present
>>>>>>>>>>>>>> any
>>>>>>>>>>>>>>>>>>>>> results
>>>>>>>>>>>>>>>>>>>>>>>> for in-service upgradability test.
>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>> ___ Your changes affect user manual and
>>>>>>> documentation,
>>>>>>>>>>>> your
>>>>>>>>>>>>>>>>> patch
>>>>>>>>>>>>>>>>>>>>> series
>>>>>>>>>>>>>>>>>>>>>>>> do not contain the patch that updates
>>>> the
>>>>>>>>>>>>>>>>>>>>>>>> Doxygen
>>>>>>>>>>>>>> manual.
>> ---------------------------------------------------------------------------- 
>>
>>>>>>>>>>>>> -- 
>>>>>>>>>>>>>>> Check out the vibrant tech community on one of the world's
>>>>> most
>>>>>>>>>>>>>>> engaging tech sites, SlashDot.org! http://sdm.link/slashdot
>>>>>>>>>>>>>>> _______________________________________________
>>>>>>>>>>>>>>> Opensaf-devel mailing list
>>>>>>>>>>>>>>> Opensaf-devel@lists.sourceforge.net
>>>>>>>>>>>>>>> https://lists.sourceforge.net/lists/listinfo/opensaf-devel
>


------------------------------------------------------------------------------
Check out the vibrant tech community on one of the world's most
engaging tech sites, SlashDot.org! http://sdm.link/slashdot
_______________________________________________
Opensaf-devel mailing list
Opensaf-devel@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/opensaf-devel

Reply via email to