Hi Vu,

Now we are now able to proceed further with  V7 `2258_v7.patch` in 
service upgraded working fine,
because of Encode/decode changes done in V7 patch.

But we have another small test case issue (/usr/bin/logtest 5 17 
Segmentation fault),
once we resolve this also, we can conclude that all the basic 
functionality is working,
then you can re-publish the V7 patch  ( if change occurred in Lennart 
#2258 V2 please do publish that as well )
so that I can go for CODE review.

Steps to reproduce the test case issue :

1) Bring up old node as Active  ( with out `2258_v7.patch` )
2) Bring-up new node as Standby  ( with `2258_v7.patch` )
3)  Do `amf-adm si-swap safSi=SC-2N,safApp=OpenSAF`
4)  Run `/usr/bin/logtest 5 17 ` on  new Active (because of si-swap )

Note :  both nodes has the new XLM attributes populated .

===================================================================================

gdb /usr/bin/logtest
(gdb) r 5

    16  PASSED   CCB Object Modify, change root directory. Path exist. OK;
Detaching after fork from child process 13797.
Set values Fail
[New Thread 0x7ffff7ff7b00 (LWP 13801)]
[New Thread 0x7ffff7fc4b00 (LWP 13802)]

Program received signal SIGSEGV, Segmentation fault.
0x00005555555688ea in read_and_compare.isra.7 () at 
src/log/apitest/tet_LogOiOps.c:1891
1891    src/log/apitest/tet_LogOiOps.c: No such file or directory.
         in src/log/apitest/tet_LogOiOps.c
(gdb) bt
#0  0x00005555555688ea in read_and_compare.isra.7 () at 
src/log/apitest/tet_LogOiOps.c:1891
#1  0x0000555555568a4b in check_logRecordDestinationConfigurationAdd () 
at src/log/apitest/tet_LogOiOps.c:1941
#2  0x0000555555571b05 in run_test_case ()
#3  0x0000555555571feb in test_run ()
#4  0x000055555555bfad in main () at src/log/apitest/logtest.c:569
(gdb)

===================================================================================


-AVM

On 2/23/2017 11:44 AM, Vu Minh Nguyen wrote:
> Hi Mahesh,
>
> Maybe it was broken when transmitting. I zipped to a tar file. Please try it
> one more.
>
> Regards, Vu
>
>
>> -----Original Message-----
>> From: A V Mahesh [mailto:mahesh.va...@oracle.com]
>> Sent: Thursday, February 23, 2017 12:54 PM
>> To: Vu Minh Nguyen <vu.m.ngu...@dektech.com.au>;
>> lennart.l...@ericsson.com; canh.v.tru...@dektech.com.au
>> Cc: opensaf-devel@lists.sourceforge.net
>> Subject: Re: [devel] [PATCH 0 of 3] Review Request for log: add
> alternative
>> destinations of log records [#2258] V4
>>
>> Hi Vu,
>>
>> On 2/23/2017 10:20 AM, Vu Minh Nguyen wrote:
>>> Hi Mahesh,
>>>
>>> Can you try with 2258_v7.patch I just sent to you?
>> I striped    changeset:   8610 of today's latest staging  (  `hg strip
>> 8610` which  removed log: implement SaLogFilterSetCallbackT and version
>> handling [#2146])
>> and try to apply your `2258_v7.patch`, it says `malformed patch at line
>> 3324`.
>>
>> -AVM
>>> I have pulled the latest code on OpenSAF 5.1 branch, re-created the
> cluster.
>>> And it works with the case old active SC-1 (OpenSAF 5.1) and new standby
>>> SC-2 (with 2258_v7.patch included in).
>>>
>>> To apply 2258_v7.patch, please do remove the just pushed ticket "log:
>>> implement SaLogFilterSetCallbackT and version handling [#2146]" ,
>>> I have not rebased the code on that yet.
>>>
>>> Regards, Vu
>>>
>>>> -----Original Message-----
>>>> From: A V Mahesh [mailto:mahesh.va...@oracle.com]
>>>> Sent: Thursday, February 23, 2017 11:45 AM
>>>> To: Vu Minh Nguyen <vu.m.ngu...@dektech.com.au>;
>>>> lennart.l...@ericsson.com; canh.v.tru...@dektech.com.au
>>>> Cc: opensaf-devel@lists.sourceforge.net
>>>> Subject: Re: [devel] [PATCH 0 of 3] Review Request for log: add
>>> alternative
>>>> destinations of log records [#2258] V4
>>>>
>>>> Hi Vu/Lennart,
>>>>
>>>>
>>>> In broad WITHOUT the #2258 patch, the same code/setup working fine
>> with
>>>> 2 sc node (staging  changeset:   8609 ),
>>>> as soon as we apply `2258_v5.patch` V5 patch on staging (changeset:
>>>> 8609 ) that you have provided yesterday,
>>>> on one sc node and try to bring up that in to cluster (in-service test)
>>>> we are observing the issue of new node (with #2258 patch) not joining
>>>> cluster.
>>>>
>>>>
>> ==============================================================
>>>> ====================================================
>>>> eb 23 10:01:59 SC-1 osafimmnd[15279]: NO Implementer (applier)
>>>> connected: 15 (@safAmfService2010f) <127, 2010f>
>>>> Feb 23 10:01:59 SC-1 osaflogd[15289]: NO LOGSV_DATA_GROUPNAME
>> not
>>>> found
>>>> Feb 23 10:01:59 SC-1 osaflogd[15289]: NO LOG root directory is:
>>>> "/var/log/opensaf/saflog"
>>>> Feb 23 10:01:59 SC-1 osaflogd[15289]: NO LOG data group is: ""
>>>> Feb 23 10:01:59 SC-1 osaflogd[15289]: NO LGS_MBCSV_VERSION = 7
>>>> Feb 23 10:01:59 SC-1 osafamfnd[15329]: NO Assigned
>>>> 'safSi=SC-2N,safApp=OpenSAF' STANDBY to 'safSu=SC-
>>>> 1,safSg=2N,safApp=OpenSAF'
>>>> Feb 23 10:01:59 SC-1 opensafd: OpenSAF(5.1.M0 - ) services successfully
>>>> started
>>>> Feb 23 10:01:59 SC-1 osafamfnd[15329]: NO
>>>> 'safComp=LOG,safSu=SC-1,safSg=2N,safApp=OpenSAF' faulted due to
>>>> 'avaDown' : Recovery is 'nodeFailfast'
>>>> Feb 23 10:01:59 SC-1 osafamfnd[15329]: ER
>>>> safComp=LOG,safSu=SC-1,safSg=2N,safApp=OpenSAF Faulted due
>>>> to:avaDown
>>>> Recovery is:nodeFailfast
>>>> Feb 23 10:01:59 SC-1 osafamfnd[15329]: Rebooting OpenSAF NodeId =
>>>> 131343
>>>> EE Name = , Reason: Component faulted: recovery is node failfast,
>>>> OwnNodeId = 131343, SupervisionTime = 60
>>>> Feb 23 10:01:59 SC-1 opensaf_reboot: Rebooting local node; timeout=60
>>>> Feb 23 10:02:00 SC-1 osafimmnd[15279]: NO Implementer (applier)
>>>> connected: 16 (@OpenSafImmReplicatorB) <144, 2010f>
>>>> Feb 23 10:01:59 SC-1 opensaf_reboot: Rebooting local node; timeout=60
>>>>
>> ==============================================================
>>>> ====================================================
>>>>
>>>> So it is evident that in-service upgrade part code of this need to be
>>>> corrected.
>>>>
>>>> Please see my comments as [AVM] and let me know if you need some
>> traces
>>>> .
>>>>
>>>> If you're planing to prepare new V6 patch , please do prepare on top of
>>>> today's latest staging.
>>>>
>>>> On 2/23/2017 9:33 AM, Vu Minh Nguyen wrote:
>>>>> Hi Mahesh,
>>>>>
>>>>> I have done in-service upgrade/downgrade with following cases:
>>>>> 1) New Active SC-1 (OpenSAF 5.2 with the attached patch) + old standby
>>>> SC-2
>>>>> (OpenSAF 5.1)
>>>>> --> Work fine
>>>> [AVM] This is not a practical use cause of in-service upgrade , we can
>>>> ignore this test further
>>>>> 2) Old Active SC-1 (OpenSAF 5.1) + new standby SC-2 (with or without
>>>>> attached patch)
>>>>> --> SC-2 is restarted & not able to join the cluster.
>>>> [AVM] This use cause/flow is  we do get in in-service upgrade , so we
>>>> need to address this.
>>>>> I got following messages in syslog:
>>>>> Feb 23 09:32:42 SC-2 user.notice opensafd: OpenSAF(5.2.M0 -
>>>>> 8529:b5addd36e45d:default) services successfully started
>>>>> Feb 23 09:32:43 SC-2 local0.warn osafntfimcnd[701]: WA
>>>> ntfimcn_imm_init
>>>>> saImmOiImplementerSet() returned SA_AIS_ERR_TIMEOUT (5)
>>>>> Feb 23 09:32:45 SC-2 local0.warn osafntfimcnd[701]: WA
>>>> ntfimcn_imm_init
>>>>> saImmOiImplementerSet() returned SA_AIS_ERR_TIMEOUT (5)
>>>>> Feb 23 09:32:47 SC-2 local0.warn osafntfimcnd[701]: WA
>>>> ntfimcn_imm_init
>>>>> saImmOiImplementerSet() returned SA_AIS_ERR_TIMEOUT (5)
>>>>> Feb 23 09:32:49 SC-2 local0.warn osafntfimcnd[701]: WA
>>>> ntfimcn_imm_init
>>>>> saImmOiImplementerSet() returned SA_AIS_ERR_TIMEOUT (5)
>>>>> Feb 23 09:32:50 SC-2 local0.err osafmsgnd[592]: ER
>>>> saImmOiImplementerSet
>>>>> FAILED:5
>>>>> Feb 23 09:32:50 SC-2 local0.err osafmsgnd[592]: ER
>>>> saImmOiImplementerSet
>>>>> FAILED:5
>>>>> Feb 23 09:32:50 SC-2 local0.notice osafamfnd[496]: NO
>>>>> 'safSu=SC-2,safSg=NoRed,safApp=OpenSAF' component restart
>> probation
>>>> timer
>>>>> started (timeout: 60000000000 ns)
>>>>> Feb 23 09:32:50 SC-2 local0.notice osafamfnd[496]: NO Restarting a
>>>> component
>>>>> of 'safSu=SC-2,safSg=NoRed,safApp=OpenSAF' (comp restart count: 1)
>>>>> Feb 23 09:32:50 SC-2 local0.notice osafamfnd[496]: NO
>>>>> 'safComp=MQND,safSu=SC-2,safSg=NoRed,safApp=OpenSAF' faulted
>> due
>>>> to
>>>>> 'avaDown' : Recovery is 'componentRestart'
>>>>> Feb 23 09:32:50 SC-2 local0.info osafmsgnd[736]: mkfifo already
> exists:
>>>>> /var/lib/opensaf/osafmsgnd.fifo File exists
>>>>>
>>>>> And sometimes, on active SC-1 (OpenSAF 5.1), the node is not able to
> up
>>>>> because of following error:
>>>>>
>>>>> Feb 23 11:00:32 SC-1 local0.err osafclmna[406]: MDTM:TIPC Dsock
>> Socket
>>>>> creation failed in MDTM_INIT err :Address family not supported by
>>>> protocol
>>>>> Feb 23 11:00:32 SC-1 local0.err osafclmna[406]: ER ncs_agents_startup
>>>> FAILED
>>>> [AVM]  No such issues ( with both TCP & TIPC)  (staging changeset:
> 8609
>>> )
>>>>> Are you getting similar problem at your side?
>>>>> Please note that, the problem is existed WITH or WITHOUT the #2258
>>>> patch.
>>>> [AVM] No , problem only if we apply `2258_v5.patch` V5 patch on staging
>>>> (changeset:   8609 )
>>>>               try to bring up that node in to cluster.
>>>>
>>>>
>>>> -AVM
>>>>
>>>>> I have informed this to IMM to have a look, not sure any problem with
>>> MDS
>>>>> layer or any problem with my environment setup.
>>>>> In the meantime, please have a look at the updated patch,  I will
>>> continue
>>>>> checking the problem. Will keep you updated.
>>>> [AVM] I haven't seen any IMM problems
>>>>> Regards, Vu
>>>>>
>>>>>> -----Original Message-----
>>>>>> From: A V Mahesh [mailto:mahesh.va...@oracle.com]
>>>>>> Sent: Wednesday, February 22, 2017 5:36 PM
>>>>>> To: Vu Minh Nguyen <vu.m.ngu...@dektech.com.au>;
>>>>>> lennart.l...@ericsson.com; canh.v.tru...@dektech.com.au
>>>>>> Cc: opensaf-devel@lists.sourceforge.net
>>>>>> Subject: Re: [devel] [PATCH 0 of 3] Review Request for log: add
>>>>> alternative
>>>>>> destinations of log records [#2258] V4
>>>>>>
>>>>>> Please see correction New Standby SC-1 ( with patch )
>>>>>>
>>>>>> -AVM
>>>>>>
>>>>>> On 2/22/2017 4:02 PM, A V Mahesh wrote:
>>>>>>> Hi Vu,
>>>>>>>
>>>>>>> With this new patch , we have another issue :
>>>>>>>
>>>>>>> 1)  standby Core by `/usr/lib64/opensaf/osaflogd'  issue got
> resolved
>>> .
>>>>>>> 2) In-service upgrade is Not working , I have Old  Active SC-2 (
> with
>>>>>>> out patch )  and New Standby SC-1 ( with patch )
>>>>>>>
>>>>>>>        the new New Standby SC-1 not joining the cluster ( in-service
>>>>>>> upgrade  failed )
>>>>>>>
>>>>>>> New Standby SC-1
>>>>>>>
>>>>>>>
>> ==============================================================
>>>>>> ======================================================
>>>>>>> Feb 22 15:53:05 SC-1 osafamfnd[15752]: NO
>>>>>>> 'safSu=SC-1,safSg=NoRed,safApp=OpenSAF' Presence State
>>>>>> INSTANTIATING
>>>>>>> => INSTANTIATED
>>>>>>> Feb 22 15:53:05 SC-1 osafamfnd[15752]: NO Assigning
>>>>>>> 'safSi=NoRed4,safApp=OpenSAF' ACTIVE to
>>>>>>> 'safSu=SC-1,safSg=NoRed,safApp=OpenSAF'
>>>>>>> Feb 22 15:53:05 SC-1 osafamfnd[15752]: NO Assigned
>>>>>>> 'safSi=NoRed4,safApp=OpenSAF' ACTIVE to
>>>>>>> 'safSu=SC-1,safSg=NoRed,safApp=OpenSAF'
>>>>>>> Feb 22 15:53:05 SC-1 osafsmfd[15889]: Started
>>>>>>> Feb 22 15:53:05 SC-1 osafamfnd[15752]: NO
>>>>>>> 'safSu=SC-1,safSg=2N,safApp=OpenSAF' Presence State
>> INSTANTIATING
>>>> =>
>>>>>>> INSTANTIATED
>>>>>>> Feb 22 15:53:05 SC-1 osafamfnd[15752]: NO Assigning
>>>>>>> 'safSi=SC-2N,safApp=OpenSAF' STANDBY to
>>>>>>> 'safSu=SC-1,safSg=2N,safApp=OpenSAF'
>>>>>>> Feb 22 15:53:05 SC-1 osafrded[15672]: NO RDE role set to STANDBY
>>>>>>> Feb 22 15:53:05 SC-1 osafrded[15672]: NO Peer up on node 0x2020f
>>>>>>> Feb 22 15:53:05 SC-1 osafrded[15672]: NO Got peer info request from
>>>>>>> node 0x2020f with role ACTIVE
>>>>>>> Feb 22 15:53:05 SC-1 osafrded[15672]: NO Got peer info response
>> from
>>>>>>> node 0x2020f with role ACTIVE
>>>>>>> Feb 22 15:53:05 SC-1 osafimmd[15691]: NO MDS event from svc_id 24
>>>>>>> (change:5, dest:13)
>>>>>>> Feb 22 15:53:05 SC-1 osafimmd[15691]: NO MDS event from svc_id 24
>>>>>>> (change:3, dest:13)
>>>>>>> Feb 22 15:53:05 SC-1 osafimmd[15691]: NO MDS event from svc_id 24
>>>>>>> (change:5, dest:13)
>>>>>>> Feb 22 15:53:05 SC-1 osafimmd[15691]: NO MDS event from svc_id 25
>>>>>>> (change:3, dest:567412424453430)
>>>>>>> Feb 22 15:53:05 SC-1 osafimmd[15691]: NO MDS event from svc_id 25
>>>>>>> (change:3, dest:565213401202663)
>>>>>>> Feb 22 15:53:05 SC-1 osafimmd[15691]: NO MDS event from svc_id 25
>>>>>>> (change:3, dest:566312912825221)
>>>>>>> Feb 22 15:53:05 SC-1 osafimmd[15691]: NO MDS event from svc_id 25
>>>>>>> (change:3, dest:564113889574230)
>>>>>>> Feb 22 15:53:05 SC-1 osafimmnd[15702]: NO Implementer (applier)
>>>>>>> connected: 17 (@safAmfService2010f) <127, 2010f>
>>>>>>> Feb 22 15:53:05 SC-1 osaflogd[15712]: NO
>> LOGSV_DATA_GROUPNAME
>>>>>> not found
>>>>>>> Feb 22 15:53:05 SC-1 osaflogd[15712]: NO LOG root directory is:
>>>>>>> "/var/log/opensaf/saflog"
>>>>>>> Feb 22 15:53:05 SC-1 osaflogd[15712]: NO LOG data group is: ""
>>>>>>> Feb 22 15:53:05 SC-1 osaflogd[15712]: NO LGS_MBCSV_VERSION = 7
>>>>>>> Feb 22 15:53:05 SC-1 osafamfnd[15752]: NO Assigned
>>>>>>> 'safSi=SC-2N,safApp=OpenSAF' STANDBY to
>>>>>>> 'safSu=SC-1,safSg=2N,safApp=OpenSAF'
>>>>>>> Feb 22 15:53:05 SC-1 opensafd: OpenSAF(5.1.M0 - ) services
>>>>>>> successfully started
>>>>>>> Feb 22 15:53:05 SC-1 osafamfnd[15752]: NO
>>>>>>> 'safComp=LOG,safSu=SC-1,safSg=2N,safApp=OpenSAF' faulted due to
>>>>>>> 'avaDown' : Recovery is 'nodeFailfast'
>>>>>>> Feb 22 15:53:05 SC-1 osafamfnd[15752]: ER
>>>>>>> safComp=LOG,safSu=SC-1,safSg=2N,safApp=OpenSAF Faulted due
>>>>>> to:avaDown
>>>>>>> Recovery is:nodeFailfast
>>>>>>> Feb 22 15:53:05 SC-1 osafamfnd[15752]: Rebooting OpenSAF NodeId
>> =
>>>>>>> 131343 EE Name = , Reason: Component faulted: recovery is node
>>>>>>> failfast, OwnNodeId = 131343, SupervisionTime = 60
>>>>>>> Feb 22 15:53:05 SC-1 opensaf_reboot: Rebooting local node;
>>>> timeout=60
>>>>>>> Feb 22 15:53:43 SC-1 syslog-ng[1171]: syslog-ng starting up;
>>>>>>> version='2.0.9'
>>>>>>>
>>>>>>>
>> ==============================================================
>>>>>> ======================================================
>>>>>>> Old - Active - SC-2
>>>>>>>
>>>>>>>
>> ==============================================================
>>>>>> ======================================================
>>>>>>> Feb 22 15:53:02 SC-2 osafimmnd[16359]: NO NODE STATE->
>>>>>>> IMM_NODE_R_AVAILABLE
>>>>>>> Feb 22 15:53:02 SC-2 osafimmloadd: NO Sync starting
>>>>>>> Feb 22 15:53:02 SC-2 osafimmloadd: IN Synced 390 objects in total
>>>>>>> Feb 22 15:53:02 SC-2 osafimmnd[16359]: NO NODE STATE->
>>>>>>> IMM_NODE_FULLY_AVAILABLE 18511
>>>>>>> Feb 22 15:53:02 SC-2 osafimmnd[16359]: NO Epoch set to 3 in
>>>> ImmModel
>>>>>>> Feb 22 15:53:02 SC-2 osafimmd[16346]: NO ACT: New Epoch for
>>>> IMMND
>>>>>>> process at node 2020f old epoch: 2  new epoch:3
>>>>>>> Feb 22 15:53:02 SC-2 osafimmd[16346]: NO ACT: New Epoch for
>>>> IMMND
>>>>>>> process at node 2040f old epoch: 2  new epoch:3
>>>>>>> Feb 22 15:53:02 SC-2 osafimmd[16346]: NO ACT: New Epoch for
>>>> IMMND
>>>>>>> process at node 2030f old epoch: 2  new epoch:3
>>>>>>> Feb 22 15:53:02 SC-2 osafimmloadd: NO Sync ending normally
>>>>>>> Feb 22 15:53:02 SC-2 osafimmd[16346]: NO ACT: New Epoch for
>>>> IMMND
>>>>>>> process at node 2010f old epoch: 0  new epoch:3
>>>>>>> Feb 22 15:53:02 SC-2 osafimmnd[16359]: NO SERVER STATE:
>>>>>>> IMM_SERVER_SYNC_SERVER --> IMM_SERVER_READY
>>>>>>> Feb 22 15:53:03 SC-2 osafamfd[16408]: NO Received node_up from
>>>> 2010f:
>>>>>>> msg_id 1
>>>>>>> Feb 22 15:53:03 SC-2 osafamfd[16408]: NO Node 'SC-1' joined the
>>>> cluster
>>>>>>> Feb 22 15:53:03 SC-2 osafimmnd[16359]: NO Implementer connected:
>>>> 16
>>>>>>> (MsgQueueService131343) <0, 2010f>
>>>>>>> Feb 22 15:53:03 SC-2 osafrded[16327]: NO Peer up on node 0x2010f
>>>>>>> Feb 22 15:53:03 SC-2 osafrded[16327]: NO Got peer info request from
>>>>>>> node 0x2010f with role STANDBY
>>>>>>> Feb 22 15:53:03 SC-2 osafrded[16327]: NO Got peer info response
>> from
>>>>>>> node 0x2010f with role STANDBY
>>>>>>> Feb 22 15:53:03 SC-2 osafimmd[16346]: NO MDS event from svc_id 24
>>>>>>> (change:5, dest:13)
>>>>>>> Feb 22 15:53:03 SC-2 osafimmnd[16359]: NO Implementer (applier)
>>>>>>> connected: 17 (@safAmfService2010f) <0, 2010f>
>>>>>>> Feb 22 15:53:03 SC-2 osafamfd[16408]: NO Cluster startup is done
>>>>>>> Feb 22 15:53:04 SC-2 osafimmnd[16359]: NO Implementer (applier)
>>>>>>> connected: 18 (@OpenSafImmReplicatorB) <0, 2010f>
>>>>>>> Feb 22 15:53:09 SC-2 osafdtmd[16304]: NO Lost contact with 'SC-1'
>>>>>>> Feb 22 15:53:09 SC-2 osaffmd[16336]: NO Node Down event for node
>> id
>>>>>>> 2010f:
>>>>>>> Feb 22 15:53:09 SC-2 osafimmd[16346]: NO MDS event from svc_id 24
>>>>>>> (change:6, dest:13)
>>>>>>> Feb 22 15:53:09 SC-2 osafimmd[16346]: NO MDS event from svc_id 25
>>>>>>> (change:4, dest:564113889574230)
>>>>>>> Feb 22 15:53:09 SC-2 osaffmd[16336]: NO Current role: ACTIVE
>>>>>>> Feb 22 15:53:09 SC-2 osaffmd[16336]: Rebooting OpenSAF NodeId =
>>>>>> 131343
>>>>>>> EE Name = , Reason: Received Node Down for peer controller,
>>>> OwnNodeId
>>>>>>> = 131599, SupervisionTime = 60
>>>>>>> Feb 22 15:53:09 SC-2 osafrded[16327]: NO Peer down on node
>> 0x2010f
>>>>>>> Feb 22 15:53:09 SC-2 osafamfd[16408]: NO Node 'SC-1' left the
>> cluster
>>>>>>> Feb 22 15:53:09 SC-2 osaflogd[16372]: NO Failed (2) to send of WRITE
>>>>>>> ack to: 2010f00003d6a
>>>>>>> Feb 22 15:53:09 SC-2 osaflogd[16372]: NO Failed (2) to send of WRITE
>>>>>>> ack to: 2010f00003d6a
>>>>>>> Feb 22 15:53:09 SC-2 osaflogd[16372]: NO Failed (2) to send of WRITE
>>>>>>> ack to: 2010f00003d74
>>>>>>> Feb 22 15:53:09 SC-2 osafclmd[16398]: NO Node 131343 went down.
>>>> Not
>>>>>>> sending track callback for agents on that node
>>>>>>> Feb 22 15:53:09 SC-2 osafclmd[16398]: NO Node 131343 went down.
>>>> Not
>>>>>>> sending track callback for agents on that node
>>>>>>> Feb 22 15:53:09 SC-2 osafclmd[16398]: NO Node 131343 went down.
>>>> Not
>>>>>>> sending track callback for agents on that node
>>>>>>> Feb 22 15:53:09 SC-2 osafclmd[16398]: NO Node 131343 went down.
>>>> Not
>>>>>>> sending track callback for agents on that node
>>>>>>> Feb 22 15:53:09 SC-2 osafclmd[16398]: NO Node 131343 went down.
>>>> Not
>>>>>>> sending track callback for agents on that node
>>>>>>> Feb 22 15:53:09 SC-2 osafclmd[16398]: NO Node 131343 went down.
>>>> Not
>>>>>>> sending track callback for agents on that node
>>>>>>> Feb 22 15:53:09 SC-2 osafclmd[16398]: NO Node 131343 went down.
>>>> Not
>>>>>>> sending track callback for agents on that node
>>>>>>> Feb 22 15:53:09 SC-2 osafimmd[16346]: WA IMMD lost contact with
>>>> peer
>>>>>>> IMMD (NCSMDS_RED_DOWN)
>>>>>>> Feb 22 15:53:09 SC-2 osafimmnd[16359]: NO Global discard node
>>>> received
>>>>>>> for nodeId:2010f pid:15702
>>>>>>> Feb 22 15:53:09 SC-2 osafimmnd[16359]: NO Implementer
>> disconnected
>>>> 16
>>>>>>> <0, 2010f(down)> (MsgQueueService131343)
>>>>>>> Feb 22 15:53:09 SC-2 osafimmnd[16359]: NO Implementer
>> disconnected
>>>> 17
>>>>>>> <0, 2010f(down)> (@safAmfService2010f)
>>>>>>> Feb 22 15:53:09 SC-2 osafimmnd[16359]: NO Implementer
>> disconnected
>>>> 18
>>>>>>> <0, 2010f(down)> (@OpenSafImmReplicatorB)
>>>>>>> Feb 22 15:53:09 SC-2 opensaf_reboot: Rebooting remote node in the
>>>>>>> absence of PLM is outside the scope of OpenSAF
>>>>>>>
>>>>>>>
>> ==============================================================
>>>>>> ======================================================
>>>>>>> -AVM
>>>>>>>
>>>>>>> On 2/22/2017 3:13 PM, Vu Minh Nguyen wrote:
>>>>>>>> Hi Mahesh,
>>>>>>>>
>>>>>>>> I put all required patches into one. Try to use this and see if you
>>>>>>>> still
>>>>>>>> have that problem or not.
>>>>>>>>
>>>>>>>> Regards, Vu
>>>>>>>>
>>>>>>>>> -----Original Message-----
>>>>>>>>> From: A V Mahesh [mailto:mahesh.va...@oracle.com]
>>>>>>>>> Sent: Wednesday, February 22, 2017 3:35 PM
>>>>>>>>> To: Vu Minh Nguyen <vu.m.ngu...@dektech.com.au>;
>>>>>>>>> lennart.l...@ericsson.com; canh.v.tru...@dektech.com.au
>>>>>>>>> Cc: opensaf-devel@lists.sourceforge.net
>>>>>>>>> Subject: Re: [devel] [PATCH 0 of 3] Review Request for log: add
>>>>>>>> alternative
>>>>>>>>> destinations of log records [#2258] V4
>>>>>>>>>
>>>>>>>>> Hi Vu,
>>>>>>>>>
>>>>>>>>> I used  new #3 and #4 patches ,  Can you please re-send All the
>>> final
>>>>>>>>> patch in go,
>>>>>>>>>
>>>>>>>>> which i need to apply on today`s  staging ( if possible publish
> the
>>>>>>>>> with
>>>>>>>>> new version )
>>>>>>>>>
>>>>>>>>> -AVM
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> On 2/22/2017 1:52 PM, Vu Minh Nguyen wrote:
>>>>>>>>>> Hi Mahesh,
>>>>>>>>>>
>>>>>>>>>>> Core was generated by `/usr/lib64/opensaf/osaflogd'.
>>>>>>>>>>> Program terminated with signal 11, Segmentation fault.
>>>>>>>>>>> #0  ckpt_proc_cfg_stream(lgs_cb*, void*) () at
>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc:2195
>>>>>>>>>>> 2195    src/log/logd/lgs_mbcsv.cc: No such file or directory.
>>>>>>>>>>>               in src/log/logd/lgs_mbcsv.cc
>>>>>>>>>> Backtrace still points to old position (lgs_mbcsv:2195). I guess
>>> the
>>>>>>>>>> osaflogd binary has not been updated with the fixed patch.
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> Regards, Vu
>>>>>>>>>>
>>>>>>>>>>> -----Original Message-----
>>>>>>>>>>> From: A V Mahesh [mailto:mahesh.va...@oracle.com]
>>>>>>>>>>> Sent: Wednesday, February 22, 2017 3:18 PM
>>>>>>>>>>> To: Vu Minh Nguyen <vu.m.ngu...@dektech.com.au>;
>>>>>>>>>>> lennart.l...@ericsson.com; canh.v.tru...@dektech.com.au
>>>>>>>>>>> Cc: opensaf-devel@lists.sourceforge.net
>>>>>>>>>>> Subject: Re: [devel] [PATCH 0 of 3] Review Request for log: add
>>>>>>>>>> alternative
>>>>>>>>>>> destinations of log records [#2258] V4
>>>>>>>>>>>
>>>>>>>>>>> Hi Vu,
>>>>>>>>>>>
>>>>>>>>>>> SC-2 standby osaflogd core dumped still occurs ( Not resolved) ,
>>> the
>>>>>>>> new
>>>>>>>>>>> patch only resolved the application (/usr/bin/logtest )
>>> Segmentation
>>>>>>>>>>> fault on SC-1 Active.
>>>>>>>>>>>
>>>>>>>>>>>
>> ==============================================================
>> ==============================================================
>>>>>>>>>>> # gdb /usr/lib64/opensaf/osaflogd
>>>> core_1487751055.osaflogd.4594
>>>>>> GNU
>>>>>>>>>>> gdb
>>>>>>>>>>> (GDB) SUSE (7.3-0.6.1)
>>>>>>>>>>> Copyright (C) 2011 Free Software Foundation, Inc.
>>>>>>>>>>> .......
>>>>>>>>>>> Core was generated by `/usr/lib64/opensaf/osaflogd'.
>>>>>>>>>>> Program terminated with signal 11, Segmentation fault.
>>>>>>>>>>> #0  ckpt_proc_cfg_stream(lgs_cb*, void*) () at
>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc:2195
>>>>>>>>>>> 2195    src/log/logd/lgs_mbcsv.cc: No such file or directory.
>>>>>>>>>>>               in src/log/logd/lgs_mbcsv.cc
>>>>>>>>>>> (gdb) bt
>>>>>>>>>>> #0  ckpt_proc_cfg_stream(lgs_cb*, void*) () at
>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc:2195
>>>>>>>>>>> #1  0x00007f97b026f960 in ckpt_decode_log_struct(lgs_cb*,
>>>>>>>>>>> ncs_mbcsv_cb_arg*, void*, void*, unsigned int (*)(edu_hdl_tag*,
>>>>>>>>>>> edu_tkn_tag*, void*, unsigned int*, edu_buf_env_tag*,
>>>>>> EDP_OP_TYPE,
>>>>>>>>>>> EDU_ERR*)) () at src/log/logd/lgs_mbcsv.cc:950
>>>>>>>>>>> #2  0x00007f97b02710dc in ckpt_decode_async_update(lgs_cb*,
>>>>>>>>>>> ncs_mbcsv_cb_arg*) () at src/log/logd/lgs_mbcsv.cc:1086
>>>>>>>>>>> #3  0x00007f97b0273941 in mbcsv_callback(ncs_mbcsv_cb_arg*)
>> ()
>>>> at
>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc:880
>>>>>>>>>>> #4  0x00007f97af372596 in ncs_mbscv_rcv_decode () from
>>>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0
>>>>>>>>>>> #5  0x00007f97af372766 in ncs_mbcsv_rcv_async_update () from
>>>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0
>>>>>>>>>>> #6  0x00007f97af379370 in mbcsv_process_events () from
>>>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0
>>>>>>>>>>> #7  0x00007f97af3794db in mbcsv_hdl_dispatch_all () from
>>>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0
>>>>>>>>>>> #8  0x00007f97af373ce2 in mbcsv_process_dispatch_request ()
>> at
>>>>>>>>>>> src/mbc/mbcsv_api.c:423
>>>>>>>>>>> #9  0x00007f97b027096e in lgs_mbcsv_dispatch(unsigned int) ()
>> at
>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc:327
>>>>>>>>>>> #10 0x00007f97b024d9f2 in main () at
>> src/log/logd/lgs_main.cc:583
>>>>>>>>>>> (gdb) bt full
>>>>>>>>>>> #0  ckpt_proc_cfg_stream(lgs_cb*, void*) () at
>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc:2195
>>>>>>>>>>>               ckpt_data_handler = {0x7f97b0270300
>>>>>>>>>>> <ckpt_proc_initialize_client(lgs_cb*, void*)>, 0x7f97b02701c0
>>>>>>>>>>> <ckpt_proc_finalize_client(lgs_cb*, void*)>,
>>>>>>>>>>>                 0x7f97b0270060 <ckpt_proc_agent_down(lgs_cb*,
>>> void*)>,
>>>>>>>>>>> 0x7f97b02712f0 <ckpt_proc_log_write(lgs_cb*, void*)>,
>>>>>> 0x7f97b0271ab0
>>>>>>>>>>> <ckpt_proc_open_stream(lgs_cb*, void*)>,
>>>>>>>>>>>                 0x7f97b026fe80 <ckpt_proc_close_stream(lgs_cb*,
>>>>> void*)>,
>>>>>>>>>>> 0x7f97b0272380 <ckpt_proc_cfg_stream(lgs_cb*, void*)>,
>>>>>>>>> 0x7f97b0274800
>>>>>>>>>>> <ckpt_proc_lgs_cfg_v2(lgs_cb*, void*)>,
>>>>>>>>>>>                 0x7f97b0274e10 <ckpt_proc_lgs_cfg_v3(lgs_cb*,
>>> void*)>,
>>>>>>>>>>> 0x7f97b02754f0 <ckpt_proc_lgs_cfg_v5(lgs_cb*, void*)>}
>>>>>>>>>>> #1  0x00007f97b026f960 in ckpt_decode_log_struct(lgs_cb*,
>>>>>>>>>>> ncs_mbcsv_cb_arg*, void*, void*, unsigned int (*)(edu_hdl_tag*,
>>>>>>>>>>> edu_tkn_tag*, void*, unsigned int*, edu_buf_env_tag*,
>>>>>> EDP_OP_TYPE,
>>>>>>>>>>> EDU_ERR*)) () at src/log/logd/lgs_mbcsv.cc:950
>>>>>>>>>>>               ckpt_data_handler = {0x7f97b0270300
>>>>>>>>>>> <ckpt_proc_initialize_client(lgs_cb*, void*)>, 0x7f97b02701c0
>>>>>>>>>>> <ckpt_proc_finalize_client(lgs_cb*, void*)>,
>>>>>>>>>>>                 0x7f97b0270060 <ckpt_proc_agent_down(lgs_cb*,
>>> void*)>,
>>>>>>>>>>> 0x7f97b02712f0 <ckpt_proc_log_write(lgs_cb*, void*)>,
>>>>>> 0x7f97b0271ab0
>>>>>>>>>>> <ckpt_proc_open_stream(lgs_cb*, void*)>,
>>>>>>>>>>>                 0x7f97b026fe80 <ckpt_proc_close_stream(lgs_cb*,
>>>>> void*)>,
>>>>>>>>>>> 0x7f97b0272380 <ckpt_proc_cfg_stream(lgs_cb*, void*)>,
>>>>>>>>> 0x7f97b0274800
>>>>>>>>>>> <ckpt_proc_lgs_cfg_v2(lgs_cb*, void*)>,
>>>>>>>>>>>                 0x7f97b0274e10 <ckpt_proc_lgs_cfg_v3(lgs_cb*,
>>> void*)>,
>>>>>>>>>>> 0x7f97b02754f0 <ckpt_proc_lgs_cfg_v5(lgs_cb*, void*)>}
>>>>>>>>>>> #2  0x00007f97b02710dc in ckpt_decode_async_update(lgs_cb*,
>>>>>>>>>>> ncs_mbcsv_cb_arg*) () at src/log/logd/lgs_mbcsv.cc:1086
>>>>>>>>>>>               ckpt_data_handler = {0x7f97b0270300
>>>>>>>>>>> <ckpt_proc_initialize_client(lgs_cb*, void*)>, 0x7f97b02701c0
>>>>>>>>>>> <ckpt_proc_finalize_client(lgs_cb*, void*)>,
>>>>>>>>>>>                 0x7f97b0270060 <ckpt_proc_agent_down(lgs_cb*,
>>> void*)>,
>>>>>>>>>>> 0x7f97b02712f0 <ckpt_proc_log_write(lgs_cb*, void*)>,
>>>>>> 0x7f97b0271ab0
>>>>>>>>>>> <ckpt_proc_open_stream(lgs_cb*, void*)>,
>>>>>>>>>>>                 0x7f97b026fe80 <ckpt_proc_close_stream(lgs_cb*,
>>>>> void*)>,
>>>>>>>>>>> 0x7f97b0272380 <ckpt_proc_cfg_stream(lgs_cb*, void*)>,
>>>>>>>>> 0x7f97b0274800
>>>>>>>>>>> <ckpt_proc_lgs_cfg_v2(lgs_cb*, void*)>,
>>>>>>>>>>>                 0x7f97b0274e10 <ckpt_proc_lgs_cfg_v3(lgs_cb*,
>>> void*)>,
>>>>>>>>>>> 0x7f97b02754f0 <ckpt_proc_lgs_cfg_v5(lgs_cb*, void*)>}
>>>>>>>>>>> #3  0x00007f97b0273941 in mbcsv_callback(ncs_mbcsv_cb_arg*)
>> ()
>>>> at
>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc:880
>>>>>>>>>>>               ckpt_data_handler = {0x7f97b0270300
>>>>>>>>>>> <ckpt_proc_initialize_client(lgs_cb*, void*)>, 0x7f97b02701c0
>>>>>>>>>>> <ckpt_proc_finalize_client(lgs_cb*, void*)>,
>>>>>>>>>>>                 0x7f97b0270060 <ckpt_proc_agent_down(lgs_cb*,
>>> void*)>,
>>>>>>>>>>> 0x7f97b02712f0 <ckpt_proc_log_write(lgs_cb*, void*)>,
>>>>>> 0x7f97b0271ab0
>>>>>>>>>>> <ckpt_proc_open_stream(lgs_cb*, void*)>,
>>>>>>>>>>>                 0x7f97b026fe80 <ckpt_proc_close_stream(lgs_cb*,
>>>>> void*)>,
>>>>>>>>>>> 0x7f97b0272380 <ckpt_proc_cfg_stream(lgs_cb*, void*)>,
>>>>>>>>> 0x7f97b0274800
>>>>>>>>>>> <ckpt_proc_lgs_cfg_v2(lgs_cb*, void*)>,
>>>>>>>>>>>                 0x7f97b0274e10 <ckpt_proc_lgs_cfg_v3(lgs_cb*,
>>> void*)>,
>>>>>>>>>>> 0x7f97b02754f0 <ckpt_proc_lgs_cfg_v5(lgs_cb*, void*)>}
>>>>>>>>>>> #4  0x00007f97af372596 in ncs_mbscv_rcv_decode () from
>>>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0
>>>>>>>>>>>               mbcsv_init_process_req_func = {0x7f97af373630
>>>>>>>>>>> <mbcsv_process_initialize_request>, 0x7f97af374f10
>>>>>>>>>>> <mbcsv_process_get_sel_obj_request>,
>>>>>>>>>>>                 0x7f97af373be0 <mbcsv_process_dispatch_request>,
>>>>>>>>>>> 0x7f97af373af0 <mbcsv_process_finalize_request>,
>>>> 0x7f97af373cf0
>>>>>>>>>>> <mbcsv_process_open_request>,
>>>>>>>>>>>                 0x7f97af374050 <mbcsv_process_close_request>,
>>>>>>>> 0x7f97af3741e0
>>>>>>>>>>> <mbcsv_process_chg_role_request>, 0x7f97af3744c0
>>>>>>>>>>> <mbcsv_process_snd_ckpt_request>,
>>>>>>>>>>>                 0x7f97af3747d0 <mbcsv_process_snd_ntfy_request>,
>>>>>>>>>>> 0x7f97af374970 <mbcsv_process_snd_data_req>,
>> 0x7f97af373930
>>>>>>>>>>> <mbcsv_process_get_request>,
>>>>>>>>>>>                 0x7f97af374bd0 <mbcsv_process_set_request>}
>>>>>>>>>>> #5  0x00007f97af372766 in ncs_mbcsv_rcv_async_update () from
>>>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0
>>>>>>>>>>>               mbcsv_init_process_req_func = {0x7f97af373630
>>>>>>>>>>> <mbcsv_process_initialize_request>, 0x7f97af374f10
>>>>>>>>>>> <mbcsv_process_get_sel_obj_request>,
>>>>>>>>>>>                 0x7f97af373be0 <mbcsv_process_dispatch_request>,
>>>>>>>>>>> 0x7f97af373af0 <mbcsv_process_finalize_request>,
>>>> 0x7f97af373cf0
>>>>>>>>>>> <mbcsv_process_open_request>,
>>>>>>>>>>>                 0x7f97af374050 <mbcsv_process_close_request>,
>>>>>>>> 0x7f97af3741e0
>>>>>>>>>>> <mbcsv_process_chg_role_request>, 0x7f97af3744c0
>>>>>>>>>>> <mbcsv_process_snd_ckpt_request>,
>>>>>>>>>>>                 0x7f97af3747d0 <mbcsv_process_snd_ntfy_request>,
>>>>>>>>>>> 0x7f97af374970 <mbcsv_process_snd_data_req>,
>> 0x7f97af373930
>>>>>>>>>>> <mbcsv_process_get_request>,
>>>>>>>>>>>                 0x7f97af374bd0 <mbcsv_process_set_request>}
>>>>>>>>>>> #6  0x00007f97af379370 in mbcsv_process_events () from
>>>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0
>>>>>>>>>>> No symbol table info available.
>>>>>>>>>>> #7  0x00007f97af3794db in mbcsv_hdl_dispatch_all () from
>>>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0
>>>>>>>>>>> No symbol table info available.
>>>>>>>>>>> #8  0x00007f97af373ce2 in mbcsv_process_dispatch_request ()
>> at
>>>>>>>>>>> src/mbc/mbcsv_api.c:423
>>>>>>>>>>>               mbcsv_init_process_req_func = {0x7f97af373630
>>>>>>>>>>> <mbcsv_process_initialize_request>, 0x7f97af374f10
>>>>>>>>>>> <mbcsv_process_get_sel_obj_request>,
>>>>>>>>>>>                 0x7f97af373be0 <mbcsv_process_dispatch_request>,
>>>>>>>>>>> 0x7f97af373af0 <mbcsv_process_finalize_request>,
>>>> 0x7f97af373cf0
>>>>>>>>>>> <mbcsv_process_open_request>,
>>>>>>>>>>>                 0x7f97af374050 <mbcsv_process_close_request>,
>>>>>>>> 0x7f97af3741e0
>>>>>>>>>>> <mbcsv_process_chg_role_request>, 0x7f97af3744c0
>>>>>>>>>>> <mbcsv_process_snd_ckpt_request>,
>>>>>>>>>>>                 0x7f97af3747d0 <mbcsv_process_snd_ntfy_request>,
>>>>>>>>>>> 0x7f97af374970 <mbcsv_process_snd_data_req>,
>> 0x7f97af373930
>>>>>>>>>>> <mbcsv_process_get_request>,
>>>>>>>>>>>                 0x7f97af374bd0 <mbcsv_process_set_request>}
>>>>>>>>>>> #9  0x00007f97b027096e in lgs_mbcsv_dispatch(unsigned int) ()
>> at
>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc:327
>>>>>>>>>>>               ckpt_data_handler = {0x7f97b0270300
>>>>>>>>>>> <ckpt_proc_initialize_client(lgs_cb*, void*)>, 0x7f97b02701c0
>>>>>>>>>>> <ckpt_proc_finalize_client(lgs_cb*, void*)>,
>>>>>>>>>>>                 0x7f97b0270060 <ckpt_proc_agent_down(lgs_cb*,
>>> void*)>,
>>>>>>>>>>> 0x7f97b02712f0 <ckpt_proc_log_write(lgs_cb*, void*)>,
>>>>>> 0x7f97b0271ab0
>>>>>>>>>>> <ckpt_proc_open_stream(lgs_cb*, void*)>,
>>>>>>>>>>>                 0x7f97b026fe80 <ckpt_proc_close_stream(lgs_cb*,
>>>>> void*)>,
>>>>>>>>>>> 0x7f97b0272380 <ckpt_proc_cfg_stream(lgs_cb*, void*)>,
>>>>>>>>> 0x7f97b0274800
>>>>>>>>>>> <ckpt_proc_lgs_cfg_v2(lgs_cb*, void*)>,
>>>>>>>>>>>                 0x7f97b0274e10 <ckpt_proc_lgs_cfg_v3(lgs_cb*,
>>> void*)>,
>>>>>>>>>>> 0x7f97b02754f0 <ckpt_proc_lgs_cfg_v5(lgs_cb*, void*)>}
>>>>>>>>>>> ---Type <return> to continue, or q <return> to quit---
>>>>>>>>>>> #10 0x00007f97b024d9f2 in main () at
>> src/log/logd/lgs_main.cc:583
>>>>>>>>>>>               usr1_sel_obj = {raise_obj = -1, rmv_obj = -1}
>>>>>>>>>>>               _lgs_cb = {mds_hdl = 65547, mds_role =
>>>> V_DEST_RL_STANDBY,
>>>>>>>>> vaddr
>>>>>>>>>>> = 11, log_version = {releaseCode = 65 'A', majorVersion = 2
>>> '\002',
>>>>>>>>>>> minorVersion = 2 '\002'}, client_tree = {
>>>>>>>>>>>                   root_node = {bit = -1, left = 0x7f97b04cf1b0,
>>> right =
>>>>>>>>>>> 0x7f97b04a2418, key_info = 0x7f97b04b7bd0 ""}, params =
>>>> {key_size
>>>>>>>>>>> = 4},
>>>>>>>>>>> n_nodes = 8}, comp_name = {_opaque = {46, 24947,
>>>>>>>>>>>                     17254, 28015, 15728, 20300, 11335, 24947,
> 21350,
>>>>>>>>>>> 15733,
>>>>>>>>>>> 17235, 12845, 29484, 26209, 26451, 12861, 11342, 24947,
>> 16742,
>>>>>> 28784,
>>>>>>>>>>> 20285, 25968, 21358, 17985,
>>>>>>>>>>>                     0 <repeats 105 times>}}, amf_hdl =
> 4288675841,
>>>>>>>>>>> amfSelectionObject = 15, amf_invocation_id = 0, is_quiesced_set
>> =
>>>>>>>> false,
>>>>>>>>>>> immOiHandle = 554050912783, immSelectionObject = 21,
>>>>>>>>>>>                 clmSelectionObject = 17, clm_hdl = 4279238657,
>>>>>>>>>>> ha_state =
>>>>>>>>>>> SA_AMF_HA_STANDBY, last_client_id = 208, async_upd_cnt =
>> 743,
>>>>>>>>>>> ckpt_state
>>>>>>>>>>> = COLD_SYNC_IDLE, mbcsv_hdl = 4293918753,
>>>>>>>>>>>                 mbcsv_sel_obj = 23, mbcsv_ckpt_hdl = 4292870177,
>>>>>>>>>>> mbcsv_peer_version = 7, edu_hdl = {is_inited = true, tree =
>>>>>>>>>>> {root_node
>>>>>>>> =
>>>>>>>>>>> {bit = -1, left = 0x7f97b04cf2e0,
>>>>>>>>>>>                       right = 0x7f97b04a25b8, key_info =
>>> 0x7f97b04b7d40
>>>>>>>> ""},
>>>>>>>>>>> params = {key_size = 8}, n_nodes = 12}, to_version = 1},
>>>>>>>>>>> fully_initialized = true, lga_down_list_head = 0x0,
>>>>>>>>>>>                 lga_down_list_tail = 0x0, clm_init_sel_obj =
>>>>>>>>>>> {raise_obj =
>>>>>>>> -1,
>>>>>>>>>>> rmv_obj = -1}, nid_started = true, scAbsenceAllowed = 900,
>>>>>>>>>>> lgs_recovery_state = LGS_NORMAL}
>>>>>>>>>>>               nfds = 7
>>>>>>>>>>>               fds = {{fd = 19, events = 1, revents = 0}, {fd =
> 15,
>>>>>>>>>>> events =
>>>>>>>>>>> 1, revents = 0}, {fd = 23, events = 1, revents = 1}, {fd = 13,
>>>>>>>>>>> events =
>>>>>>>>>>> 1, revents = 0}, {fd = -1, events = 1,
>>>>>>>>>>>                   revents = 0}, {fd = 17, events = 1, revents =
> 0},
>>>>>>>>>>> {fd =
>>>>>>>> 21,
>>>>>>>>>>> events = 1, revents = 0}}
>>>>>>>>>>>               mbox_msgs = {0, 0, 0, 0, 0}
>>>>>>>>>>>               lgs_cb = 0x7f97b04a2400
>>>>>>>>>>>               mbox_low = {0, 0, 0, 0, 0}
>>>>>>>>>>>               lgs_mbox_init_mutex = {__data = {__lock = 0,
> __count =
>>> 0,
>>>>>>>>>>> __owner = 0, __nusers = 0, __kind = 0, __spins = 0, __list =
>>>>>>>>>>> {__prev =
>>>>>>>>>>> 0x0, __next = 0x0}},
>>>>>>>>>>>                 __size = '\000' <repeats 39 times>, __align = 0}
>>>>>>>>>>>               lgs_mbx = 4291821569
>>>>>>>>>>>               mbox_high = {0, 0, 0, 0, 0}
>>>>>>>>>>>               mbox_full = {false, false, false, false, false}
>>>>>>>>>>> (gdb)
>>>>>>>>>>> (gdb)
>>>>>>>>>>>
>>>>>>>>>>> Feb 22 13:40:55 SC-2 osafimmnd[4584]: NO Ccb 131
>> COMMITTED
>>>>>>>>>>> (immcfg_SC-1_18714)
>>>>>>>>>>> Feb 22 13:40:56 SC-2 osafamfnd[4634]: NO
>>>>>>>>>>> 'safComp=LOG,safSu=SC-2,safSg=2N,safApp=OpenSAF' faulted
>> due
>>>> to
>>>>>>>>>>> 'avaDown' : Recovery is 'nodeFailfast'
>>>>>>>>>>> Feb 22 13:40:56 SC-2 osafamfnd[4634]: ER
>>>>>>>>>>> safComp=LOG,safSu=SC-2,safSg=2N,safApp=OpenSAF Faulted
>> due
>>>>>>>>>>> to:avaDown
>>>>>>>>>>> Recovery is:nodeFailfast
>>>>>>>>>>> Feb 22 13:40:56 SC-2 osafamfnd[4634]: Rebooting OpenSAF
>> NodeId
>>>> =
>>>>>>>>>>> 131599
>>>>>>>>>>> EE Name = , Reason: Component faulted: recovery is node
>> failfast,
>>>>>>>>>>> OwnNodeId = 131599, SupervisionTime = 60
>>>>>>>>>>> Feb 22 13:40:56 SC-2 opensaf_reboot: Rebooting local node;
>>>>>> timeout=60
>> ==============================================================
>> ==============================================================
>>>>>>>>>>> On 2/22/2017 12:23 PM, A V Mahesh wrote:
>>>>>>>>>>>> Hi Vu,
>>>>>>>>>>>>
>>>>>>>>>>>> On 2/22/2017 12:19 PM, Vu Minh Nguyen wrote:
>>>>>>>>>>>>> [Vu] I has sent you 02 patches. There is code change in
>> osaflogd
>>>>>>>>>>>>> code
>>>>>>>>>> that
>>>>>>>>>>>>> fix the coredump you have observed.
>>>>>>>>>>>>> The other one is test code that fix the logtest coredump.
>>>>>>>>>>>> Ok I will re-test , and update you .
>>>>>>>>>>>>
>>>>>>>>>>>> -AVM
>>>>>>>>>>>>
>>>>>>>>>>>> On 2/22/2017 12:19 PM, Vu Minh Nguyen wrote:
>>>>>>>>>>>>> Hi Mahehs,
>>>>>>>>>>>>>
>>>>>>>>>>>>> See my reply inline, [Vu].
>>>>>>>>>>>>>
>>>>>>>>>>>>> Regards, Vu
>>>>>>>>>>>>>
>>>>>>>>>>>>>> -----Original Message-----
>>>>>>>>>>>>>> From: A V Mahesh [mailto:mahesh.va...@oracle.com]
>>>>>>>>>>>>>> Sent: Wednesday, February 22, 2017 1:36 PM
>>>>>>>>>>>>>> To: Vu Minh Nguyen <vu.m.ngu...@dektech.com.au>;
>>>>>>>>>>>>>> lennart.l...@ericsson.com; canh.v.tru...@dektech.com.au
>>>>>>>>>>>>>> Cc: opensaf-devel@lists.sourceforge.net
>>>>>>>>>>>>>> Subject: Re: [PATCH 0 of 3] Review Request for log: add
>>>>>>>>>>>>>> alternative
>>>>>>>>>>>>>> destinations of log records [#2258] V4
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> Hi Vu,
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> On 2/22/2017 11:52 AM, Vu Minh Nguyen wrote:
>>>>>>>>>>>>>>> Hi Mahesh,
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> Have a code fault in uml test, and other one in checkpoint.
>>>>>>>>>>>>>> [AVM] This is Normal Suse 11 VM ( not  UML).
>>>>>>>>>>>>>>> I have just updated the code. Please re-apply for #3 and #4
>>>>>>>> patches.
>>>>>>>>>>>>>> [AVM] is these new patch has function changes or only test
>>>> code
>>>>>>>>>>> changes ?
>>>>>>>>>>>>> [Vu] I has sent you 02 patches. There is code change in
>> osaflogd
>>>>>>>>>>>>> code
>>>>>>>>>> that
>>>>>>>>>>>>> fix the coredump you have observed.
>>>>>>>>>>>>> The other one is test code that fix the logtest coredump.
>>>>>>>>>>>>>>> Note that, test case #14 of suite 17 should be run on active
>>>>>>>>>>>>>>> node,
>>>>>>>>>>>>>> otherwise
>>>>>>>>>>>>>>> getting failed.
>>>>>>>>>>>>>> [AVM]  Segmentation fault of  /usr/bin/logtest Not a big
> issue
>>> ,
>>>>>>>>>>>>>>          we need to debug why  osaflogd core dumped and it is
>>>>>>>>>>>>>> critical
>>>>>>>>>>>>> [Vu] I found the problem. You can try with the new one to see
>> if
>>>>>>>>>>>>> the
>>>>>>>>>>>>> coredump is still there or not.
>>>>>>>>>>>>>>> I will put condition check to that test case later.
>>>>>>>>>>>>>> -AVM
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> Regards, Vu
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> -----Original Message-----
>>>>>>>>>>>>>>>> From: A V Mahesh [mailto:mahesh.va...@oracle.com]
>>>>>>>>>>>>>>>> Sent: Wednesday, February 22, 2017 12:16 PM
>>>>>>>>>>>>>>>> To: Vu Minh Nguyen <vu.m.ngu...@dektech.com.au>;
>>>>>>>>>>>>>>>> lennart.l...@ericsson.com;
>> canh.v.tru...@dektech.com.au
>>>>>>>>>>>>>>>> Cc: opensaf-devel@lists.sourceforge.net
>>>>>>>>>>>>>>>> Subject: Re: [PATCH 0 of 3] Review Request for log: add
>>>>>>>> alternative
>>>>>>>>>>>>>>>> destinations of log records [#2258] V4
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> Hi Vu,
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> Thanks ,
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> While testing  /usr/bin/logtest ,  SC-2 standby osaflogd
>> core
>>>>>>>>> dumped
>>>>>>>>>>>>> and
>>>>>>>>>>>>>>>> /usr/bin/logtest on SC-1 Active
>>>>>>>>>>>>>>>> got Segmentation fault , am I missing any other patch (  i
>> am
>>>>>>>> using
>>>>>>>>>>>>>>>> devel published patch only )
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> Following patches i am using :
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>           1) #2293 (sent by Anders Widel, but not yet
> pushed)
>>>>>>>>>>>>>>>>             2) #2258 (v2, sent by Lennart, but not yet
> pushed
>>> yet)
>>>>>>>>>>>>>>>>             3) #2258 (v4, sent by Vu, but not yet pushed
> yet)
>>>>>>>>>>>>>>>>
>> ==============================================================
>>>>>>>>>>>>>>>> ========================================
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> Core was generated by `/usr/lib64/opensaf/osaflogd'.
>>>>>>>>>>>>>>>> Program terminated with signal 11, Segmentation fault.
>>>>>>>>>>>>>>>> #0  ckpt_proc_cfg_stream(lgs_cb*, void*) () at
>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc:2195
>>>>>>>>>>>>>>>> 2195    src/log/logd/lgs_mbcsv.cc: No such file or
>> directory.
>>>>>>>>>>>>>>>>                  in src/log/logd/lgs_mbcsv.cc
>>>>>>>>>>>>>>>> (gdb) bt
>>>>>>>>>>>>>>>> #0  ckpt_proc_cfg_stream(lgs_cb*, void*) () at
>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc:2195
>>>>>>>>>>>>>>>> #1  0x00007f12c3e22960 in
>> ckpt_decode_log_struct(lgs_cb*,
>>>>>>>>>>>>>>>> ncs_mbcsv_cb_arg*, void*, void*, unsigned int
>>>>>> (*)(edu_hdl_tag*,
>>>>>>>>>>>>>>>> edu_tkn_tag*, void*, unsigned int*, edu_buf_env_tag*,
>>>>>>>>>>> EDP_OP_TYPE,
>>>>>>>>>>>>>>>> EDU_ERR*)) () at src/log/logd/lgs_mbcsv.cc:950
>>>>>>>>>>>>>>>> #2  0x00007f12c3e240dc in
>>>>>> ckpt_decode_async_update(lgs_cb*,
>>>>>>>>>>>>>>>> ncs_mbcsv_cb_arg*) () at src/log/logd/lgs_mbcsv.cc:1086
>>>>>>>>>>>>>>>> #3  0x00007f12c3e26941 in
>>>>>> mbcsv_callback(ncs_mbcsv_cb_arg*) ()
>>>>>>>>> at
>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc:880
>>>>>>>>>>>>>>>> #4  0x00007f12c2f25596 in ncs_mbscv_rcv_decode () from
>>>>>>>>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0
>>>>>>>>>>>>>>>> #5  0x00007f12c2f25766 in ncs_mbcsv_rcv_async_update
>> ()
>>>>>> from
>>>>>>>>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0
>>>>>>>>>>>>>>>> #6  0x00007f12c2f2c370 in mbcsv_process_events () from
>>>>>>>>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0
>>>>>>>>>>>>>>>> #7  0x00007f12c2f2c4db in mbcsv_hdl_dispatch_all () from
>>>>>>>>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0
>>>>>>>>>>>>>>>> #8  0x00007f12c2f26ce2 in
>> mbcsv_process_dispatch_request
>>>> ()
>>>>>> at
>>>>>>>>>>>>>>>> src/mbc/mbcsv_api.c:423
>>>>>>>>>>>>>>>> #9  0x00007f12c3e2396e in lgs_mbcsv_dispatch(unsigned
>> int)
>>>> ()
>>>>>> at
>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc:327
>>>>>>>>>>>>>>>> #10 0x00007f12c3e009f2 in main () at
>>>>>>>>>>>>>>>> src/log/logd/lgs_main.cc:583
>>>>>>>>>>>>>>>> (gdb)
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>
>> ==============================================================
>>>>>>>>>>>>>>>> ========================================
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> Feb 22 10:37:06 SC-1 osafimmnd[4020]: NO Invalid error
>>>>>> reported
>>>>>>>>>>>>>>>> implementer 'safLogService', Ccb 161 will be aborted
>>>>>>>>>>>>>>>> Feb 22 10:37:06 SC-1 osafimmnd[4020]: NO Ccb 161
>> aborted
>>>> in
>>>>>>>>>>>>>> COMPLETED
>>>>>>>>>>>>>>>> processing (validation)
>>>>>>>>>>>>>>>> Feb 22 10:37:06 SC-1 osafimmnd[4020]: NO Ccb 161
>>>> ABORTED
>>>>>>>>>>>>>> (immcfg_SC-
>>>>>>>>>>>>>>>> 1_5394)
>>>>>>>>>>>>>>>> Add values Fail
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> Program received signal SIGSEGV, Segmentation fault.
>>>>>>>>>>>>>>>> 0x000055555556929a in read_and_compare.isra.7 () at
>>>>>>>>>>>>>>>> src/log/apitest/tet_LogOiOps.c:1891
>>>>>>>>>>>>>>>> 1891    src/log/apitest/tet_LogOiOps.c: No such file or
>>>>>>>>>>>>>>>> directory.
>>>>>>>>>>>>>>>>                  in src/log/apitest/tet_LogOiOps.c
>>>>>>>>>>>>>>>> (gdb) Feb 22 10:37:07 SC-1 sshd[5298]: Accepted keyboard-
>>>>>>>>>>>>>> interactive/pam
>>>>>>>>>>>>>>>> for root from 10.176.178.22 port 51945 ssh2
>>>>>>>>>>>>>>>> bt
>>>>>>>>>>>>>>>> #0  0x000055555556929a in read_and_compare.isra.7 () at
>>>>>>>>>>>>>>>> src/log/apitest/tet_LogOiOps.c:1891
>>>>>>>>>>>>>>>> #1  0x0000555555569bbb in
>>>>>>>>>>>>>>>> check_logRecordDestinationConfigurationEmpty
>>>>>>>>>>>>>>>> () at src/log/apitest/tet_LogOiOps.c:2179
>>>>>>>>>>>>>>>> #2  0x0000555555573495 in run_test_case ()
>>>>>>>>>>>>>>>> #3  0x0000555555573934 in test_run ()
>>>>>>>>>>>>>>>> #4  0x000055555555c7cd in main () at
>>>>>>>>>>>>>>>> src/log/apitest/logtest.c:569
>>>>>>>>>>>>>>>> (gdb)
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>
>> ==============================================================
>>>>>>>>>>>>>>>> ========================================
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> -AVM
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> On 2/22/2017 9:48 AM, Vu Minh Nguyen wrote:
>>>>>>>>>>>>>>>>> Hi Mahesh,
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> I send them in attachment instead, and name them in the
>>>>>> order.
>>>>>>>>>>>>>>>>> I just pull the latest code, and apply them without
> getting
>>>>> any
>>>>>>>>>> hunk
>>>>>>>>>>>>>>> error.
>>>>>>>>>>>>>>>>> Please try with them, and let me know if you see any
>>>> problem.
>>>>>>>>>>>>>>>>> Regards, Vu
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> -----Original Message-----
>>>>>>>>>>>>>>>>>> From: A V Mahesh [mailto:mahesh.va...@oracle.com]
>>>>>>>>>>>>>>>>>> Sent: Wednesday, February 22, 2017 11:09 AM
>>>>>>>>>>>>>>>>>> To: Vu Minh Nguyen <vu.m.ngu...@dektech.com.au>;
>>>>>>>>>>>>>>>>>> lennart.l...@ericsson.com;
>>>> canh.v.tru...@dektech.com.au
>>>>>>>>>>>>>>>>>> Cc: opensaf-devel@lists.sourceforge.net
>>>>>>>>>>>>>>>>>> Subject: Re: [PATCH 0 of 3] Review Request for log: add
>>>>>>>>>> alternative
>>>>>>>>>>>>>>>>>> destinations of log records [#2258] V4
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> Hi Vu,
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> I did follow that still i get Hunk #2 FAILED even on
>>> today's
>>>>>>>>>> staging
>> ==============================================================
>>>>>>>>>>>>>>>>>> ==================
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> [root@dhcp-hyd-scp-5fl-10-176-177-96 staging]# patch
>> -
>>>> p1
>>>>>>>>> <2293
>>>>>>>>>>>>>>>>>> patching file src/base/Makefile.am
>>>>>>>>>>>>>>>>>> Hunk #1 succeeded at 33 (offset 1 line).
>>>>>>>>>>>>>>>>>> Hunk #3 succeeded at 183 (offset 1 line).
>>>>>>>>>>>>>>>>>> patching file src/base/file_descriptor.cc
>>>>>>>>>>>>>>>>>> patching file src/base/file_descriptor.h
>>>>>>>>>>>>>>>>>> patching file src/base/tests/unix_socket_test.cc
>>>>>>>>>>>>>>>>>> patching file src/base/unix_client_socket.cc
>>>>>>>>>>>>>>>>>> patching file src/base/unix_server_socket.cc
>>>>>>>>>>>>>>>>>> patching file src/base/unix_socket.cc
>>>>>>>>>>>>>>>>>> patching file src/base/unix_socket.h
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> [root@dhcp-hyd-scp-5fl-10-176-177-96 staging]# patch
>> -
>>>> p1
>>>>>>>>> <2258-
>>>>>>>>>>> 1
>>>>>>>>>>>>>>>>>> patching file src/log/Makefile.am
>>>>>>>>>>>>>>>>>> Hunk #1 succeeded at 71 (offset -1 lines).
>>>>>>>>>>>>>>>>>> patching file src/log/config/logsv_classes.xml
>>>>>>>>>>>>>>>>>> Hunk #1 FAILED at 147.
>>>>>>>>>>>>>>>>>> 1 out of 1 hunk FAILED -- saving rejects to file
>>>>>>>>>>>>>>>>>> src/log/config/logsv_classes.xml.rej
>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_config.cc
>>>>>>>>>>>>>>>>>> Hunk #1 succeeded at 35 (offset -5 lines).
>>>>>>>>>>>>>>>>>> Hunk #2 FAILED at 705.
>>>>>>>>>>>>>>>>>> Hunk #3 FAILED at 971.
>>>>>>>>>>>>>>>>>> 2 out of 3 hunks FAILED -- saving rejects to file
>>>>>>>>>>>>>>>>>> src/log/logd/lgs_config.cc.rej
>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_config.h
>>>>>>>>>>>>>>>>>> Hunk #1 FAILED at 304.
>>>>>>>>>>>>>>>>>> 1 out of 1 hunk FAILED -- saving rejects to file
>>>>>>>>>>>>>>>>>> src/log/logd/lgs_config.h.rej
>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_dest.cc
>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_dest.h
>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_evt.cc
>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_imm.cc
>>>>>>>>>>>>>>>>>> Hunk #1 FAILED at 45.
>>>>>>>>>>>>>>>>>> Hunk #2 succeeded at 235 (offset -1 lines).
>>>>>>>>>>>>>>>>>> Hunk #3 FAILED at 877.
>>>>>>>>>>>>>>>>>> Hunk #4 succeeded at 1273 (offset -20 lines).
>>>>>>>>>>>>>>>>>> Hunk #5 succeeded at 1404 (offset -1 lines).
>>>>>>>>>>>>>>>>>> Hunk #6 succeeded at 1449 (offset -20 lines).
>>>>>>>>>>>>>>>>>> Hunk #7 succeeded at 2032 (offset -1 lines).
>>>>>>>>>>>>>>>>>> Hunk #8 FAILED at 2181.
>>>>>>>>>>>>>>>>>> Hunk #9 succeeded at 2271 (offset -54 lines).
>>>>>>>>>>>>>>>>>> Hunk #10 succeeded at 2387 (offset -1 lines).
>>>>>>>>>>>>>>>>>> Hunk #11 succeeded at 2377 (offset -54 lines).
>>>>>>>>>>>>>>>>>> Hunk #12 succeeded at 2478 (offset -1 lines).
>>>>>>>>>>>>>>>>>> Hunk #13 succeeded at 2684 (offset -54 lines).
>>>>>>>>>>>>>>>>>> Hunk #14 succeeded at 2821 (offset -1 lines).
>>>>>>>>>>>>>>>>>> 3 out of 14 hunks FAILED -- saving rejects to file
>>>>>>>>>>>>>>>>>> src/log/logd/lgs_imm.cc.rej
>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_main.cc
>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_mbcsv.cc
>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_mbcsv.h
>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_mbcsv_v5.cc
>>>>>>>>>>>>>>>>>> Hunk #3 succeeded at 133 (offset -1 lines).
>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_mbcsv_v7.cc
>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_mbcsv_v7.h
>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_stream.cc
>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_stream.h
>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_util.cc
>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_util.h
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> [root@dhcp-hyd-scp-5fl-10-176-177-96 staging]# patch
>> -
>>>> p1
>>>>>>>>> <2258-
>>>>>>>>>>> 2
>>>>>>>>>>>>>>>>>> patching file src/log/Makefile.am
>>>>>>>>>>>>>>>>>> Hunk #1 succeeded at 180 (offset -3 lines).
>>>>>>>>>>>>>>>>>> patching file src/log/apitest/tet_LogOiOps.c
>>>>>>>>>>>>>>>>>> Hunk #1 FAILED at 1923.
>>>>>>>>>>>>>>>>>> Hunk #2 FAILED at 1979.
>>>>>>>>>>>>>>>>>> Hunk #3 FAILED at 2067.
>>>>>>>>>>>>>>>>>> Hunk #4 FAILED at 2094.
>>>>>>>>>>>>>>>>>> 4 out of 4 hunks FAILED -- saving rejects to file
>>>>>>>>>>>>>>>>>> src/log/apitest/tet_LogOiOps.c.rej
>>>>>>>>>>>>>>>>>> patching file src/log/apitest/tet_cfg_destination.c
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> [root@dhcp-hyd-scp-5fl-10-176-177-96 staging]# patch
>> -
>>>> p1
>>>>>>>>> <2258-
>>>>>>>>>>> 3
>>>>>>>>>>>>>>>>>> patching file src/log/Makefile
>>>>>>>>>>>>>>>>>> patching file src/log/Makefile.am
>>>>>>>>>>>>>>>>>> Hunk #1 succeeded at 80 (offset -1 lines).
>>>>>>>>>>>>>>>>>> Hunk #2 succeeded at 217 (offset -2 lines).
>>>>>>>>>>>>>>>>>> patching file src/log/tests/Makefile
>>>>>>>>>>>>>>>>>> patching file src/log/tests/lgs_dest_test.cc
>>>>>>>>>>>>>>>>>> [root@dhcp-hyd-scp-5fl-10-176-177-96 staging]# vi
>>>>>>>>>>>>>>>>>> src/log/apitest/tet_LogOiOps.c.rej
>>>>>>>>>>>>>>>>>> [root@dhcp-hyd-scp-5fl-10-176-177-96 staging]#
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>
>> ==============================================================
>>>>>>>>>>>>>>>>>> ========================
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> -AVM
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> On 2/21/2017 3:53 PM, Vu Minh Nguyen wrote:
>>>>>>>>>>>>>>>>>>> Hi Mahesh,
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> As I has mentioned in below:
>>>>>>>>>>>>>>>>>>>>>             To run the test, this patch has dependent
> on
>>>>>>>> following
>>>>>>>>>>>>> patches:
>>>>>>>>>>>>>>>>>>>>>             1) #2293 (sent by Anders Widel, but not yet
>>>>>>>>>>>>>>>>>>>>> pushed)
>>>>>>>>>>>>>>>>>>>>>             2) #2258 (v2, sent by Lennart, but not yet
>>> pushed
>>>>>>>> yet)
>>>>>>>>>>>>>>>>>>> So, you need to apply #2293 first, then #2258 which
>> sent
>>>> by
>>>>>>>>>>> Lennart
>>>>>>>>>>>>>>>>>>> yesterday, then mine.
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> Regards, Vu
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>> -----Original Message-----
>>>>>>>>>>>>>>>>>>>> From: A V Mahesh [mailto:mahesh.va...@oracle.com]
>>>>>>>>>>>>>>>>>>>> Sent: Tuesday, February 21, 2017 5:10 PM
>>>>>>>>>>>>>>>>>>>> To: Vu Minh Nguyen
>> <vu.m.ngu...@dektech.com.au>;
>>>>>>>>>>>>>>>>>>>> lennart.l...@ericsson.com;
>>>>>> canh.v.tru...@dektech.com.au
>>>>>>>>>>>>>>>>>>>> Cc: opensaf-devel@lists.sourceforge.net
>>>>>>>>>>>>>>>>>>>> Subject: Re: [PATCH 0 of 3] Review Request for log:
>> add
>>>>>>>>>>> alternative
>>>>>>>>>>>>>>>>>>>> destinations of log records [#2258] V4
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>> Hi Vu,
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>> Is this applies on top of   log #2146  - V4  , I see
>>> both
>>>>>>>>>> #tickets
>>>>>>>>>>>>>>>>> has
>>>>>>>>>>>>>>>>>>>> version changes ?
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>> in which order i need to apply  ( #2146 & #2258 )    or
>>>>>>>>>>>>>>>>>>>> (#2258
>>>>>>>> &
>>>>>>>>>>>>>>>>> #2146).
>>>>>> =========================================================
>>>>>>>>>>>>>>>>>>>> patching file src/log/Makefile.am
>>>>>>>>>>>>>>>>>>>> Hunk #1 FAILED at 72.
>>>>>>>>>>>>>>>>>>>> Hunk #2 FAILED at 120.
>>>>>>>>>>>>>>>>>>>> 2 out of 2 hunks FAILED -- saving rejects to file
>>>>>>>>>>>>>>>>> src/log/Makefile.am.rej
>>>>>>>>>>>>>>>>>>>> patching file src/log/config/logsv_classes.xml
>>>>>>>>>>>>>>>>>>>> Hunk #1 FAILED at 147.
>>>>>>>>>>>>>>>>>>>> 1 out of 1 hunk FAILED -- saving rejects to file
>>>>>>>>>>>>>>>>>>>> src/log/config/logsv_classes.xml.rej
>>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_config.cc
>>>>>>>>>>>>>>>>>>>> Hunk #1 succeeded at 35 (offset -5 lines).
>>>>>>>>>>>>>>>>>>>> Hunk #2 FAILED at 705.
>>>>>>>>>>>>>>>>>>>> Hunk #3 FAILED at 971.
>>>>>>>>>>>>>>>>>>>> 2 out of 3 hunks FAILED -- saving rejects to file
>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_config.cc.rej
>>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_config.h
>>>>>>>>>>>>>>>>>>>> Hunk #1 FAILED at 304.
>>>>>>>>>>>>>>>>>>>> 1 out of 1 hunk FAILED -- saving rejects to file
>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_config.h.rej
>>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_dest.cc
>>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_dest.h
>>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_evt.cc
>>>>>>>>>>>>>>>>>>>> Hunk #1 FAILED at 1.
>>>>>>>>>>>>>>>>>>>> Hunk #2 succeeded at 30 with fuzz 2 (offset 2 lines).
>>>>>>>>>>>>>>>>>>>> Hunk #3 succeeded at 1282 (offset 45 lines).
>>>>>>>>>>>>>>>>>>>> Hunk #4 succeeded at 1300 (offset 2 lines).
>>>>>>>>>>>>>>>>>>>> 1 out of 4 hunks FAILED -- saving rejects to file
>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_evt.cc.rej
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>
>> ==============================================================
>>>>>>>>>>>>>>>>>>>> ===
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>> -AVM
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>> On 2/21/2017 3:03 PM, Vu Minh Nguyen wrote:
>>>>>>>>>>>>>>>>>>>>> Summary: log: add alternative destinations of log
>>>> records
>>>>>>>>>>> [#2258]
>>>>>>>>>>>>>>>>>>>>> Review request for Trac Ticket(s): #2258
>>>>>>>>>>>>>>>>>>>>> Peer Reviewer(s): Lennart, Canh, Mahesh
>>>>>>>>>>>>>>>>>>>>> Pull request to: <<LIST THE PERSON WITH PUSH
>>>> ACCESS
>>>>>>>>> HERE>>
>>>>>>>>>>>>>>>>>>>>> Affected branch(es): Default
>>>>>>>>>>>>>>>>>>>>> Development branch: Default
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>> --------------------------------
>>>>>>>>>>>>>>>>>>>>> Impacted area       Impact y/n
>>>>>>>>>>>>>>>>>>>>> --------------------------------
>>>>>>>>>>>>>>>>>>>>>             Docs                    n
>>>>>>>>>>>>>>>>>>>>>             Build system            n
>>>>>>>>>>>>>>>>>>>>>             RPM/packaging           n
>>>>>>>>>>>>>>>>>>>>>             Configuration files     n
>>>>>>>>>>>>>>>>>>>>>             Startup scripts         n
>>>>>>>>>>>>>>>>>>>>>             SAF services            n
>>>>>>>>>>>>>>>>>>>>>             OpenSAF services        y
>>>>>>>>>>>>>>>>>>>>>             Core libraries          n
>>>>>>>>>>>>>>>>>>>>>             Samples                 n
>>>>>>>>>>>>>>>>>>>>>             Tests                   y
>>>>>>>>>>>>>>>>>>>>>             Other                   n
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>> Comments (indicate scope for each "y" above):
>>>>>>>>>>>>>>>>>>>>> ---------------------------------------------
>>>>>>>>>>>>>>>>>>>>>             To run the test, this patch has dependent
> on
>>>>>>>> following
>>>>>>>>>>>>> patches:
>>>>>>>>>>>>>>>>>>>>>             1) #2293 (sent by Anders Widel, but not yet
>>>>>>>>>>>>>>>>>>>>> pushed)
>>>>>>>>>>>>>>>>>>>>>             2) #2258 (v2, sent by Lennart, but not yet
>>> pushed
>>>>>>>> yet)
>>>>>>>>>>>>>>>>>>>>> changeset
>>>>>> d74aaf3025c99cade3165a15831124548f4d85bd
>>>>>>>>>>>>>>>>>>>>> Author:    Vu Minh Nguyen
>>>>>> <vu.m.ngu...@dektech.com.au>
>>>>>>>>>>>>>>>>>>>>> Date:    Wed, 15 Feb 2017 14:36:00 +0700
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>        log: add alternative destinations of log records
>>>>>>>>>>>>>>>>>>>>> [#2258]
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>        Here are major info, detailed info will be added
> to
>>>>>>>>>>>>>>>>>>>>> PR doc
>>>>>>>>>>>>>>> soon. 1)
>>>>>>>>>>>>>>>>>>>> Add
>>>>>>>>>>>>>>>>>>>>>        attribute "saLogRecordDestination" to log
>> stream.
>>> 2)
>>>>>>>>>>>>>>>>>>>>> Add
>>>>>>>>>>>>>>> Local
>>>>>>>>>>>>>>>>>>>> socket
>>>>>>>>>>>>>>>>>>>>>        destintion handler 3) Integrate into first
>>> increment
>>>>>>>>>>>>>>>>>>>>> made by
>>>>>>>>>>>>>>> Lennart
>>>>>>>>>>>>>>>>>>>>> changeset
>>>>>> 4bae27a478c235df3058f43c92d3a5483233b01d
>>>>>>>>>>>>>>>>>>>>> Author:    Vu Minh Nguyen
>>>>>> <vu.m.ngu...@dektech.com.au>
>>>>>>>>>>>>>>>>>>>>> Date:    Wed, 15 Feb 2017 15:07:09 +0700
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>        log: add UML test case to verify alternative
>>>>>>>>>>>>>>>>>>>>> destination
>>>>>>>>>>>>>>> [#2258]
>>>>>>>>>>>>>>>>>>>>>        Major changes: 1) Modify Lennart's test cases
>>>> because
>>>>>>>>>>>>>>> enhancing
>>>>>>>>>>>>>>>>>>>> destination
>>>>>>>>>>>>>>>>>>>>>        configuration validation rules. 2) Add test
> suite
>>>>>>>>>>>>>>>>>>>>> #17 to
>>>>>>>>>>>>>>> verify
>>>>>>>>>>>>>>>>>>>> alternative
>>>>>>>>>>>>>>>>>>>>>        destination
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>> changeset
>>>> bc375725fed22bb4f8cb3ae3df5f96fb9d281efb
>>>>>>>>>>>>>>>>>>>>> Author:    Vu Minh Nguyen
>>>>>> <vu.m.ngu...@dektech.com.au>
>>>>>>>>>>>>>>>>>>>>> Date:    Thu, 16 Feb 2017 17:22:13 +0700
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>        log: add unit tests to verify interfaces
> provided
>>> by
>>>>>>>>>>>>>>> destination
>>>>>>>>>>>>>>>>>>>> handler
>>>>>>>>>>>>>>>>>>>>>        [#2258]
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>        Unit tests to verify major interfaces: 1)
>>>>>>>>>>>>>>>>>>>>> CfgDestination()
>>>>>>>>>>>>>>> 2)
>>>>>>>>>>>>>>>>>>>>>        WriteToDestination()
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>> Added Files:
>>>>>>>>>>>>>>>>>>>>> ------------
>>>>>>>>>>>>>>>>>>>>> src/log/apitest/tet_cfg_destination.c
>>>>>>>>>>>>>>>>>>>>>             src/log/logd/lgs_dest.cc
>>>>>>>>>>>>>>>>>>>>>             src/log/logd/lgs_dest.h
>>>>>>>>>>>>>>>>>>>>>             src/log/logd/lgs_mbcsv_v7.cc
>>>>>>>>>>>>>>>>>>>>>             src/log/logd/lgs_mbcsv_v7.h
>>>>>>>>>>>>>>>>>>>>> src/log/tests/lgs_dest_test.cc
>>>>>>>>>>>>>>>>>>>>>             src/log/tests/Makefile
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>> Complete diffstat:
>>>>>>>>>>>>>>>>>>>>> ------------------
>>>>>>>>>>>>>>>>>>>>> src/log/Makefile |    4 +
>>>>>>>>>>>>>>>>>>>>> src/log/Makefile.am |   31 +++++-
>>>>>>>>>>>>>>>>>>>>> src/log/apitest/tet_LogOiOps.c |    8 +-
>>>>>>>>>>>>>>>>>>>>> src/log/apitest/tet_cfg_destination.c |  483
>> ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++
>>>>>>>>>>>>>>>>>>>> ++++++++++++++++++++++++++++++++++
>>>>>>>>>>>>>>>>>>>>> src/log/config/logsv_classes.xml |    7 +-
>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_config.cc |  169
>>>>>>>>>>>>>>>>>>>> +++++++++++++++++++++++++++++---
>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_config.h |    3 +-
>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_dest.cc |  707
>> ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++
>> ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++
>>>>>>>>>>>>>>>>>>>> +++++++++++++++++
>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_dest.h |  576
>> ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++
>>>>>>>>> ++++++++++++++++++++++++++++++++++++++++++++++++++++
>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_evt.cc |   33 ++++++
>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_imm.cc |  202
>>>>>>>>>>>>>>>>>>>> +++++++++++++++++++++++++++++++++------
>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_main.cc |    8 +
>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc |  103
>>>>>>>>>>>>> ++++++++++++++++++-
>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv.h |    6 +-
>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv_v5.cc |   10 +
>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv_v7.cc |  177
>>>>>>>>>>>>>>>>>>>> +++++++++++++++++++++++++++++++++++
>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv_v7.h |   67
>>>>>>>>>> +++++++++++++
>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_stream.cc |   60
>>>>>>>>>> +++++++++++-
>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_stream.h |   16 +++
>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_util.cc |   63
>>>>>>>>>> ++++++++++++
>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_util.h |   11 +-
>>>>>>>>>>>>>>>>>>>>> src/log/tests/Makefile |   20 +++
>>>>>>>>>>>>>>>>>>>>> src/log/tests/lgs_dest_test.cc |  209
>>>>>>>>>>>>>>>>>>>> +++++++++++++++++++++++++++++++++++++++++
>>>>>>>>>>>>>>>>>>>>>             23 files changed, 2896 insertions(+), 77
>>>>>>>>>>>>>>>>>>>>> deletions(-)
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>> Testing Commands:
>>>>>>>>>>>>>>>>>>>>> -----------------
>>>>>>>>>>>>>>>>>>>>>             Run UML test suite #17
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>> Testing, Expected Results:
>>>>>>>>>>>>>>>>>>>>> --------------------------
>>>>>>>>>>>>>>>>>>>>>             All test passed
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>> Conditions of Submission:
>>>>>>>>>>>>>>>>>>>>> -------------------------
>>>>>>>>>>>>>>>>>>>>>             <<HOW MANY DAYS BEFORE PUSHING,
>>>> CONSENSUS
>>>>>>>>> ETC>>
>>>>>>>>>>>>>>>>>>>>> Arch      Built     Started    Linux distro
>>>>>>>>>>>>>>>>>>>>> -------------------------------------------
>>>>>>>>>>>>>>>>>>>>> mips        n          n
>>>>>>>>>>>>>>>>>>>>> mips64      n          n
>>>>>>>>>>>>>>>>>>>>> x86         n          n
>>>>>>>>>>>>>>>>>>>>> x86_64      n          n
>>>>>>>>>>>>>>>>>>>>> powerpc     n          n
>>>>>>>>>>>>>>>>>>>>> powerpc64   n          n
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>> Reviewer Checklist:
>>>>>>>>>>>>>>>>>>>>> -------------------
>>>>>>>>>>>>>>>>>>>>> [Submitters: make sure that your review doesn't
>> trigger
>>>>>> any
>>>>>>>>>>>>>>>>>> checkmarks!]
>>>>>>>>>>>>>>>>>>>>> Your checkin has not passed review because (see
>>>> checked
>>>>>>>>>>> entries):
>>>>>>>>>>>>>>>>>>>>> ___ Your RR template is generally incomplete; it has
>>>> too
>>>>>>>>> many
>>>>>>>>>>>>> blank
>>>>>>>>>>>>>>>>>>> entries
>>>>>>>>>>>>>>>>>>>>>                that need proper data filled in.
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>> ___ You have failed to nominate the proper persons
>> for
>>>>>>>>> review
>>>>>>>>>>> and
>>>>>>>>>>>>>>>>> push.
>>>>>>>>>>>>>>>>>>>>> ___ Your patches do not have proper short+long
>>>> header
>>>>>>>>>>>>>>>>>>>>> ___ You have grammar/spelling in your header that
>> is
>>>>>>>>>>> unacceptable.
>>>>>>>>>>>>>>>>>>>>> ___ You have exceeded a sensible line length in your
>>>>>>>>>>>>>>>>>>>> headers/comments/text.
>>>>>>>>>>>>>>>>>>>>> ___ You have failed to put in a proper Trac Ticket #
>>> into
>>>>>>>> your
>>>>>>>>>>>>>>>>> commits.
>>>>>>>>>>>>>>>>>>>>> ___ You have incorrectly put/left internal data in
>> your
>>>>>>>>>>>>>>> comments/files
>>>>>>>>>>>>>>>>>>>>> (i.e. internal bug tracking tool IDs, product
>>>>>>>> names
>>>>>>>>>> etc)
>>>>>>>>>>>>>>>>>>>>> ___ You have not given any evidence of testing
>> beyond
>>>>>> basic
>>>>>>>>>>> build
>>>>>>>>>>>>>>>>> tests.
>>>>>>>>>>>>>>>>>>>>> Demonstrate some level of runtime or other sanity
>>>>>>>>>> testing.
>>>>>>>>>>>>>>>>>>>>> ___ You have ^M present in some of your files.
>> These
>>>>>>>>>>>>>>>>>>>>> have to
>>>>>>>>>>> be
>>>>>>>>>>>>>>>>>> removed.
>>>>>>>>>>>>>>>>>>>>> ___ You have needlessly changed whitespace or
>> added
>>>>>>>>>>> whitespace
>>>>>>>>>>>>>>>>>> crimes
>>>>>>>>>>>>>>>>>>>>>                like trailing spaces, or spaces before
>>> tabs.
>>>>>>>>>>>>>>>>>>>>> ___ You have mixed real technical changes with
>>>>>> whitespace
>>>>>>>>> and
>>>>>>>>>>>>>> other
>>>>>>>>>>>>>>>>>>>>> cosmetic code cleanup changes. These have to be
>>>>>>>>>>> separate
>>>>>>>>>>>>>>>>> commits.
>>>>>>>>>>>>>>>>>>>>> ___ You need to refactor your submission into
>> logical
>>>>>>>>>>>>>>>>>>>>> chunks;
>>>>>>>>>>>>> there
>>>>>>>>>>>>>>> is
>>>>>>>>>>>>>>>>>>>>>                too much content into a single commit.
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>> ___ You have extraneous garbage in your review
>>>> (merge
>>>>>>>>>>> commits
>>>>>>>>>>>>>> etc)
>>>>>>>>>>>>>>>>>>>>> ___ You have giant attachments which should never
>>>> have
>>>>>>>>> been
>>>>>>>>>>>>>> sent;
>>>>>>>>>>>>>>>>>>>>> Instead you should place your content in a public
>>>>>>>>>> tree to
>>>>>>>>>>>>> be
>>>>>>>>>>>>>>>>>>> pulled.
>>>>>>>>>>>>>>>>>>>>> ___ You have too many commits attached to an e-
>> mail;
>>>>>>>>> resend
>>>>>>>>>>> as
>>>>>>>>>>>>>>>>>> threaded
>>>>>>>>>>>>>>>>>>>>> commits, or place in a public tree for a pull.
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>> ___ You have resent this content multiple times
>>>> without a
>>>>>>>>> clear
>>>>>>>>>>>>>>>>>>> indication
>>>>>>>>>>>>>>>>>>>>>                of what has changed between each
> re-send.
>>>>>>>>>>>>>>>>>>>>> ___ You have failed to adequately and individually
>>>> address
>>>>>>>> all
>>>>>>>>>> of
>>>>>>>>>>>>>>> the
>>>>>>>>>>>>>>>>>>>>> comments and change requests that were proposed
>> in
>>>>>>>>>> the
>>>>>>>>>>>>>>> initial
>>>>>>>>>>>>>>>>>>> review.
>>>>>>>>>>>>>>>>>>>>> ___ You have a misconfigured ~/.hgrc file (i.e.
>>>> username,
>>>>>>>>> email
>>>>>>>>>>>>> etc)
>>>>>>>>>>>>>>>>>>>>> ___ Your computer have a badly configured date and
>>>>>> time;
>>>>>>>>>>>>>> confusing
>>>>>>>>>>>>>>>> the
>>>>>>>>>>>>>>>>>>>>>                the threaded patch review.
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>> ___ Your changes affect IPC mechanism, and you
>> don't
>>>>>>>>> present
>>>>>>>>>>> any
>>>>>>>>>>>>>>>>>> results
>>>>>>>>>>>>>>>>>>>>>                for in-service upgradability test.
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>> ___ Your changes affect user manual and
>>>> documentation,
>>>>>>>>> your
>>>>>>>>>>>>>> patch
>>>>>>>>>>>>>>>>>> series
>>>>>>>>>>>>>>>>>>>>>                do not contain the patch that updates
> the
>>>>>>>>>>>>>>>>>>>>> Doxygen
>>>>>>>>>>> manual.
> ----------------------------------------------------------------------------
>>>>>>>>>> --
>>>>>>>>>>>> Check out the vibrant tech community on one of the world's
>> most
>>>>>>>>>>>> engaging tech sites, SlashDot.org! http://sdm.link/slashdot
>>>>>>>>>>>> _______________________________________________
>>>>>>>>>>>> Opensaf-devel mailing list
>>>>>>>>>>>> Opensaf-devel@lists.sourceforge.net
>>>>>>>>>>>> https://lists.sourceforge.net/lists/listinfo/opensaf-devel


------------------------------------------------------------------------------
Check out the vibrant tech community on one of the world's most
engaging tech sites, SlashDot.org! http://sdm.link/slashdot
_______________________________________________
Opensaf-devel mailing list
Opensaf-devel@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/opensaf-devel

Reply via email to