Hi Vu, Issue still exist even with `2258_v10.patch`. fix all the issue and republish the patch.
========================================================================================================================================== Feb 23 17:09:06 SC-2 osafamfnd[3774]: NO 'safSu=SC-2,safSg=2N,safApp=OpenSAF' Presence State INSTANTIATING => INSTANTIATED Feb 23 17:09:06 SC-2 osafamfnd[3774]: NO Assigning 'safSi=SC-2N,safApp=OpenSAF' STANDBY to 'safSu=SC-2,safSg=2N,safApp=OpenSAF' Feb 23 17:09:06 SC-2 osafrded[3694]: NO RDE role set to STANDBY Feb 23 17:09:06 SC-2 osafrded[3694]: NO Peer up on node 0x2010f Feb 23 17:09:06 SC-2 osafrded[3694]: NO Got peer info request from node 0x2010f with role ACTIVE Feb 23 17:09:06 SC-2 osafrded[3694]: NO Got peer info response from node 0x2010f with role ACTIVE Feb 23 17:09:06 SC-2 osafimmd[3713]: NO MDS event from svc_id 24 (change:3, dest:13) Feb 23 17:09:06 SC-2 osafimmd[3713]: NO MDS event from svc_id 24 (change:5, dest:13) Feb 23 17:09:06 SC-2 osafimmd[3713]: NO MDS event from svc_id 24 (change:5, dest:13) Feb 23 17:09:06 SC-2 osafimmd[3713]: NO MDS event from svc_id 25 (change:3, dest:565213468688400) Feb 23 17:09:06 SC-2 osafimmd[3713]: NO MDS event from svc_id 25 (change:3, dest:564116543111184) Feb 23 17:09:06 SC-2 osafimmnd[3724]: NO Implementer (applier) connected: 15 (@safAmfService2020f) <127, 2020f> Feb 23 17:09:06 SC-2 osaflogd[3734]: NO LOGSV_DATA_GROUPNAME not found Feb 23 17:09:06 SC-2 osaflogd[3734]: NO LOG root directory is: "/var/log/opensaf/saflog" Feb 23 17:09:06 SC-2 osaflogd[3734]: NO LOG data group is: "" Feb 23 17:09:06 SC-2 osaflogd[3734]: NO LGS_MBCSV_VERSION = 7 Feb 23 17:09:06 SC-2 osafamfnd[3774]: NO Assigned 'safSi=SC-2N,safApp=OpenSAF' STANDBY to 'safSu=SC-2,safSg=2N,safApp=OpenSAF' Feb 23 17:09:06 SC-2 opensafd: OpenSAF(5.1.M0 - ) services successfully started done SC-2:~ # Feb 23 17:09:06 SC-2 osafimmnd[3724]: NO Implementer (applier) connected: 16 (@OpenSafImmReplicatorB) <150, 2020f> Feb 23 17:09:06 SC-2 osafntfimcnd[3931]: NO Started Feb 23 17:09:08 SC-2 osafamfd[3764]: NO Cold sync complete! Feb 23 17:09:08 SC-2 osaflogd[3734]: WA FAILED: ncs_patricia_tree_add, client_id 0 Feb 23 17:09:08 SC-2 osaflogd[3734]: ER Exiting with message: Could not create new client Feb 23 17:09:08 SC-2 osafamfnd[3774]: NO 'safSu=SC-2,safSg=2N,safApp=OpenSAF' component restart probation timer started (timeout: 60000000000 ns) Feb 23 17:09:08 SC-2 osafamfnd[3774]: NO Restarting a component of 'safSu=SC-2,safSg=2N,safApp=OpenSAF' (comp restart count: 1) Feb 23 17:09:08 SC-2 osafamfnd[3774]: NO 'safComp=LOG,safSu=SC-2,safSg=2N,safApp=OpenSAF' faulted due to 'errorReport' : Recovery is 'componentRestart' ========================================================================================================================================== -AVM On 2/23/2017 4:25 PM, Vu Minh Nguyen wrote: > Hi Mahesh, > > Sorry, I missed fixing other point. See 2258_additional_fix_err.patch. > > For your convenience, I pack them all to new version 2258_v10.patch. > > Regards, Vu > >> -----Original Message----- >> From: Vu Minh Nguyen [mailto:vu.m.ngu...@dektech.com.au] >> Sent: Thursday, February 23, 2017 5:39 PM >> To: 'A V Mahesh' <mahesh.va...@oracle.com>; 'lennart.l...@ericsson.com' >> <lennart.l...@ericsson.com>; 'canh.v.tru...@dektech.com.au' >> <canh.v.tru...@dektech.com.au> >> Cc: 'opensaf-devel@lists.sourceforge.net' <opensaf- >> de...@lists.sourceforge.net> >> Subject: RE: [devel] [PATCH 0 of 3] Review Request for log: add > alternative >> destinations of log records [#2258] V4 >> >> Hi Mahesh, >> >> I found the root cause. It is because in ` log: implement >> SaLogFilterSetCallbackT and version handling [#2146]`, >> Canh introduced MBCSV version #6, but I missed adding that info when >> rebasing. >> >> The attached patch contains the fix. Can you apply it to see if the > problem >> still occur? Thanks. >> >> Regards, Vu >> >>> -----Original Message----- >>> From: A V Mahesh [mailto:mahesh.va...@oracle.com] >>> Sent: Thursday, February 23, 2017 5:32 PM >>> To: Vu Minh Nguyen <vu.m.ngu...@dektech.com.au>; >>> lennart.l...@ericsson.com; canh.v.tru...@dektech.com.au >>> Cc: opensaf-devel@lists.sourceforge.net >>> Subject: Re: [devel] [PATCH 0 of 3] Review Request for log: add > alternative >>> destinations of log records [#2258] V4 >>> >>> Hi Vu, >>> >>> Please find attached , below is the time stamp of SC-2 >>> >>> >> ============================================================== >>> ============================================== >>> >>> Feb 23 15:55:30 SC-2 osafimmnd[6978]: NO Implementer (applier) >>> connected: 15 (@safAmfService2020f) <127, 2020f> >>> Feb 23 15:55:30 SC-2 osaflogd[6988]: NO LOGSV_DATA_GROUPNAME not >>> found >>> Feb 23 15:55:30 SC-2 osaflogd[6988]: NO LOG root directory is: >>> "/var/log/opensaf/saflog" >>> Feb 23 15:55:30 SC-2 osaflogd[6988]: NO LOG data group is: "" >>> Feb 23 15:55:30 SC-2 osaflogd[6988]: NO LGS_MBCSV_VERSION = 7 >>> Feb 23 15:55:30 SC-2 osafamfnd[7028]: NO Assigned >>> 'safSi=SC-2N,safApp=OpenSAF' STANDBY to 'safSu=SC- >>> 2,safSg=2N,safApp=OpenSAF' >>> Feb 23 15:55:30 SC-2 opensafd: OpenSAF(5.1.M0 - ) services successfully >>> started >>> done >>> SC-2:/var/log/opensaf # Feb 23 15:55:31 SC-2 osafimmnd[6978]: NO >>> Implementer (applier) connected: 16 (@OpenSafImmReplicatorB) <144, >>> 2020f> >>> Feb 23 15:55:31 SC-2 osafntfimcnd[7185]: NO Started >>> Feb 23 15:55:33 SC-2 osafamfd[7018]: NO Cold sync complete! >>> Feb 23 15:55:33 SC-2 osaflogd[6988]: WA FAILED: ncs_patricia_tree_add, >>> client_id 0 >>> Feb 23 15:55:33 SC-2 osaflogd[6988]: ER Exiting with message: Could not >>> create new client >>> Feb 23 15:55:33 SC-2 osafamfnd[7028]: NO >>> 'safSu=SC-2,safSg=2N,safApp=OpenSAF' component restart probation >> timer >>> started (timeout: 60000000000 ns) >>> Feb 23 15:55:33 SC-2 osafamfnd[7028]: NO Restarting a component of >>> 'safSu=SC-2,safSg=2N,safApp=OpenSAF' (comp restart count: 1) >>> Feb 23 15:55:33 SC-2 osafamfnd[7028]: NO >>> 'safComp=LOG,safSu=SC-2,safSg=2N,safApp=OpenSAF' faulted due to >>> 'errorReport' : Recovery is 'componentRestart' >>> >>> >> ============================================================== >>> ============================================== >>> >>> -AVM >>> >>> >>> On 2/23/2017 3:39 PM, Vu Minh Nguyen wrote: >>>> Hi Mahesh, >>>> >>>> No change in V7 vs V9. Just do rebase the code on latest changeset. >>>> >>>> I have tried to clean up all, and rebuild the cluster to see what you > are >>>> observing, >>>> and I am not able to reproduce the problem, I have tried several > times. >>>> Can you provide me the osaflogd trace on both SCs node? Thanks. >>>> >>>> Regards, Vu >>>> >>>>> -----Original Message----- >>>>> From: A V Mahesh [mailto:mahesh.va...@oracle.com] >>>>> Sent: Thursday, February 23, 2017 4:48 PM >>>>> To: Vu Minh Nguyen <vu.m.ngu...@dektech.com.au>; >>>>> lennart.l...@ericsson.com; canh.v.tru...@dektech.com.au >>>>> Cc: opensaf-devel@lists.sourceforge.net >>>>> Subject: Re: [devel] [PATCH 0 of 3] Review Request for log: add >>>> alternative >>>>> destinations of log records [#2258] V4 >>>>> >>>>> Hi Vu, >>>>> >>>>> On 2/23/2017 3:13 PM, A V Mahesh wrote: >>>>>> Not sure what are other change compare to V7 to V9 , New problems >>> got >>>>>> introduced >>>>>> >>>>>> Both nodes SC-1 & SC-2 ( with 2258_v9.patch ) , trying bring up > both >>>>>> SC`s simple node bringup , >>>>>> >>>>>> SC-2 going for reboot with following : >>>>>> >>>>>> >> ============================================================== >> ============================================================== >>>>> ============ >>>>>> Feb 23 15:05:32 SC-2 osaflogd[29988]: NO >> LOGSV_DATA_GROUPNAME >>>>> not found >>>>>> Feb 23 15:05:32 SC-2 osaflogd[29988]: NO LOG root directory is: >>>>>> "/var/log/opensaf/saflog" >>>>>> Feb 23 15:05:32 SC-2 osaflogd[29988]: NO LOG data group is: "" >>>>>> Feb 23 15:05:32 SC-2 osafimmnd[29978]: NO Implementer (applier) >>>>>> connected: 16 (@safAmfService2020f) <127, 2020f> >>>>>> Feb 23 15:05:32 SC-2 osaflogd[29988]: NO LGS_MBCSV_VERSION = 7 >>>>>> Feb 23 15:05:32 SC-2 osaflogd[29988]: WA FAILED: >>>>>> ncs_patricia_tree_add, client_id 0 >>>>>> Feb 23 15:05:32 SC-2 osafamfnd[30028]: NO Assigned >>>>>> 'safSi=SC-2N,safApp=OpenSAF' STANDBY to >>>>>> 'safSu=SC-2,safSg=2N,safApp=OpenSAF' >>>>>> Feb 23 15:05:32 SC-2 osaflogd[29988]: ER Exiting with message: Could >>>>>> not create new client >>>>>> Feb 23 15:05:32 SC-2 osafamfnd[30028]: NO >>>>>> 'safSu=SC-2,safSg=2N,safApp=OpenSAF' component restart probation >>>>> timer >>>>>> started (timeout: 60000000000 ns) >>>>>> Feb 23 15:05:32 SC-2 osafamfnd[30028]: NO Restarting a component >> of >>>>>> 'safSu=SC-2,safSg=2N,safApp=OpenSAF' (comp restart count: 1) >>>>>> Feb 23 15:05:32 SC-2 osafamfnd[30028]: NO >>>>>> 'safComp=LOG,safSu=SC-2,safSg=2N,safApp=OpenSAF' faulted due to >>>>>> 'errorReport' : Recovery is 'componentRestart' >>>>>> Feb 23 15:05:32 SC-2 opensafd[29908]: ER Service LOGD has >>> unexpectedly >>>>>> crashed. Unable to continue, exiting >>>>>> Feb 23 15:05:32 SC-2 osafamfd[30018]: exiting for shutdown >>>>>> Feb 23 15:05:32 SC-2 osafamfnd[30028]: ER AMFD has unexpectedly >>>>>> crashed. Rebooting node >>>>>> Feb 23 15:05:32 SC-2 osafamfnd[30028]: Rebooting OpenSAF NodeId = >>>>>> 131599 EE Name = , Reason: AMFD has unexpectedly crashed. >>> Rebooting >>>>>> node, OwnNodeId = 131599, SupervisionTime = 60 >>>>>> Feb 23 15:05:32 SC-2 opensaf_reboot: Rebooting local node; >>> timeout=60 >>>>>> Feb 23 15:06:04 SC-2 syslog-ng[1180]: syslog-ng starting up; >>>>>> version='2.0.9' >>>>>> >>>>>> >> ============================================================== >> ============================================================== >>>>> ============ >>>>> Some times : >>>>> >>>>> >> ============================================================== >> ============================================================== >>>>> ============ >>>>> >>>>> Feb 23 15:15:19 SC-2 osafrded[3858]: NO RDE role set to STANDBY >>>>> Feb 23 15:15:19 SC-2 osafrded[3858]: NO Peer up on node 0x2010f >>>>> Feb 23 15:15:19 SC-2 osafrded[3858]: NO Got peer info request from >>> node >>>>> 0x2010f with role ACTIVE >>>>> Feb 23 15:15:19 SC-2 osafrded[3858]: NO Got peer info response from >>> node >>>>> 0x2010f with role ACTIVE >>>>> Feb 23 15:15:19 SC-2 osafimmd[3877]: NO MDS event from svc_id 24 >>>>> (change:3, dest:13) >>>>> Feb 23 15:15:19 SC-2 osafimmd[3877]: NO MDS event from svc_id 24 >>>>> (change:5, dest:13) >>>>> Feb 23 15:15:19 SC-2 osafimmd[3877]: NO MDS event from svc_id 24 >>>>> (change:5, dest:13) >>>>> Feb 23 15:15:19 SC-2 osafimmd[3877]: NO MDS event from svc_id 25 >>>>> (change:3, dest:565217560625168) >>>>> Feb 23 15:15:19 SC-2 osafimmd[3877]: NO MDS event from svc_id 25 >>>>> (change:3, dest:564114674417680) >>>>> Feb 23 15:15:19 SC-2 osaflogd[3898]: NO LOGSV_DATA_GROUPNAME >>> not >>>>> found >>>>> Feb 23 15:15:19 SC-2 osaflogd[3898]: NO LOG root directory is: >>>>> "/var/log/opensaf/saflog" >>>>> Feb 23 15:15:19 SC-2 osaflogd[3898]: NO LOG data group is: "" >>>>> Feb 23 15:15:19 SC-2 osafimmnd[3888]: NO Implementer (applier) >>>>> connected: 15 (@safAmfService2020f) <127, 2020f> >>>>> Feb 23 15:15:19 SC-2 osaflogd[3898]: NO LGS_MBCSV_VERSION = 7 >>>>> Feb 23 15:15:19 SC-2 osaflogd[3898]: ER Exiting with message: Client >>>>> attributes differ >>>>> Feb 23 15:15:19 SC-2 osafamfnd[3938]: NO >>>>> 'safSu=SC-2,safSg=2N,safApp=OpenSAF' component restart probation >>> timer >>>>> started (timeout: 60000000000 ns) >>>>> Feb 23 15:15:19 SC-2 osafamfnd[3938]: NO Restarting a component of >>>>> 'safSu=SC-2,safSg=2N,safApp=OpenSAF' (comp restart count: 1) >>>>> Feb 23 15:15:19 SC-2 osafamfnd[3938]: NO >>>>> 'safComp=LOG,safSu=SC-2,safSg=2N,safApp=OpenSAF' faulted due to >>>>> 'errorReport' : Recovery is 'componentRestart' >>>>> Feb 23 15:15:19 SC-2 opensafd[3818]: ER Service LOGD has >> unexpectedly >>>>> crashed. Unable to continue, exiting >>>>> Feb 23 15:15:20 SC-2 osafamfd[3928]: exiting for shutdown >>>>> Feb 23 15:15:20 SC-2 osafamfnd[3938]: ER AMFD has unexpectedly >>> crashed. >>>>> Rebooting node >>>>> Feb 23 15:15:20 SC-2 osafamfnd[3938]: Rebooting OpenSAF NodeId = >>>>> 131599 >>>>> EE Name = , Reason: AMFD has unexpectedly crashed. Rebooting node, >>>>> OwnNodeId = 131599, SupervisionTime = 60 >>>>> Feb 23 15:15:20 SC-2 osafimmnd[3888]: NO Implementer locally >>>>> disconnected. Marking it as doomed 15 <127, 2020f> >>> (@safAmfService2020f) >>>>> Feb 23 15:15:20 SC-2 osafimmnd[3888]: NO Implementer disconnected >> 15 >>>>> <127, 2020f> (@safAmfService2020f) >>>>> Feb 23 15:15:20 SC-2 opensaf_reboot: Rebooting local node; >> timeout=60 >> ============================================================== >> ============================================================== >>>>> ============ >>>>> >>>>> >>>>>> -AVM >>>>>> >>>>>> >>>>>> On 2/23/2017 2:20 PM, Vu Minh Nguyen wrote: >>>>>>> Hi Mahesh, >>>>>>> >>>>>>> This is the latest code has been rebased on the latest changeset. >>>>>>> >>>>>>> Note that, in the attached patch, I have included one more >>> dependency, >>>>>>> that is on base::Hash() function, the patch sent by Anders [#2266] >>>>>>> >>>>>>> Please review the patch, then comment if any. Thanks. >>>>>>> >>>>>>> Regards, Vu >>>>>>> >>>>>>>> -----Original Message----- >>>>>>>> From: A V Mahesh [mailto:mahesh.va...@oracle.com] >>>>>>>> Sent: Thursday, February 23, 2017 2:03 PM >>>>>>>> To: Vu Minh Nguyen <vu.m.ngu...@dektech.com.au>; >>>>>>>> lennart.l...@ericsson.com; canh.v.tru...@dektech.com.au >>>>>>>> Cc: opensaf-devel@lists.sourceforge.net >>>>>>>> Subject: Re: [devel] [PATCH 0 of 3] Review Request for log: add >>>>>>> alternative >>>>>>>> destinations of log records [#2258] V4 >>>>>>>> >>>>>>>> Hi Vu, >>>>>>>> >>>>>>>> Now we are now able to proceed further with V7 `2258_v7.patch` in >>>>>>>> service upgraded working fine, >>>>>>>> because of Encode/decode changes done in V7 patch. >>>>>>>> >>>>>>>> But we have another small test case issue (/usr/bin/logtest 5 17 >>>>>>>> Segmentation fault), >>>>>>>> once we resolve this also, we can conclude that all the basic >>>>>>>> functionality is working, >>>>>>>> then you can re-publish the V7 patch ( if change occurred in > Lennart >>>>>>>> #2258 V2 please do publish that as well ) >>>>>>>> so that I can go for CODE review. >>>>>>>> >>>>>>>> Steps to reproduce the test case issue : >>>>>>>> >>>>>>>> 1) Bring up old node as Active ( with out `2258_v7.patch` ) >>>>>>>> 2) Bring-up new node as Standby ( with `2258_v7.patch` ) >>>>>>>> 3) Do `amf-adm si-swap safSi=SC-2N,safApp=OpenSAF` >>>>>>>> 4) Run `/usr/bin/logtest 5 17 ` on new Active (because of > si-swap ) >>>>>>>> Note : both nodes has the new XLM attributes populated . >>>>>>>> >>>>>>>> >> ============================================================== >>>>>>>> ===================== >>>>>>>> >>>>>>>> gdb /usr/bin/logtest >>>>>>>> (gdb) r 5 >>>>>>>> >>>>>>>> 16 PASSED CCB Object Modify, change root directory. Path >>>>>>>> exist. OK; >>>>>>>> Detaching after fork from child process 13797. >>>>>>>> Set values Fail >>>>>>>> [New Thread 0x7ffff7ff7b00 (LWP 13801)] >>>>>>>> [New Thread 0x7ffff7fc4b00 (LWP 13802)] >>>>>>>> >>>>>>>> Program received signal SIGSEGV, Segmentation fault. >>>>>>>> 0x00005555555688ea in read_and_compare.isra.7 () at >>>>>>>> src/log/apitest/tet_LogOiOps.c:1891 >>>>>>>> 1891 src/log/apitest/tet_LogOiOps.c: No such file or directory. >>>>>>>> in src/log/apitest/tet_LogOiOps.c >>>>>>>> (gdb) bt >>>>>>>> #0 0x00005555555688ea in read_and_compare.isra.7 () at >>>>>>>> src/log/apitest/tet_LogOiOps.c:1891 >>>>>>>> #1 0x0000555555568a4b in >>>>> check_logRecordDestinationConfigurationAdd () >>>>>>>> at src/log/apitest/tet_LogOiOps.c:1941 >>>>>>>> #2 0x0000555555571b05 in run_test_case () >>>>>>>> #3 0x0000555555571feb in test_run () >>>>>>>> #4 0x000055555555bfad in main () at src/log/apitest/logtest.c:569 >>>>>>>> (gdb) >>>>>>>> >>>>>>>> >> ============================================================== >>>>>>>> ===================== >>>>>>>> >>>>>>>> >>>>>>>> -AVM >>>>>>>> >>>>>>>> On 2/23/2017 11:44 AM, Vu Minh Nguyen wrote: >>>>>>>>> Hi Mahesh, >>>>>>>>> >>>>>>>>> Maybe it was broken when transmitting. I zipped to a tar file. >> Please >>>>>>> try it >>>>>>>>> one more. >>>>>>>>> >>>>>>>>> Regards, Vu >>>>>>>>> >>>>>>>>> >>>>>>>>>> -----Original Message----- >>>>>>>>>> From: A V Mahesh [mailto:mahesh.va...@oracle.com] >>>>>>>>>> Sent: Thursday, February 23, 2017 12:54 PM >>>>>>>>>> To: Vu Minh Nguyen <vu.m.ngu...@dektech.com.au>; >>>>>>>>>> lennart.l...@ericsson.com; canh.v.tru...@dektech.com.au >>>>>>>>>> Cc: opensaf-devel@lists.sourceforge.net >>>>>>>>>> Subject: Re: [devel] [PATCH 0 of 3] Review Request for log: add >>>>>>>>> alternative >>>>>>>>>> destinations of log records [#2258] V4 >>>>>>>>>> >>>>>>>>>> Hi Vu, >>>>>>>>>> >>>>>>>>>> On 2/23/2017 10:20 AM, Vu Minh Nguyen wrote: >>>>>>>>>>> Hi Mahesh, >>>>>>>>>>> >>>>>>>>>>> Can you try with 2258_v7.patch I just sent to you? >>>>>>>>>> I striped changeset: 8610 of today's latest staging ( `hg >>>> strip >>>>>>>>>> 8610` which removed log: implement SaLogFilterSetCallbackT >> and >>>>>>>>>> version >>>>>>>>>> handling [#2146]) >>>>>>>>>> and try to apply your `2258_v7.patch`, it says `malformed patch > at >>>>>>>>>> line >>>>>>>>>> 3324`. >>>>>>>>>> >>>>>>>>>> -AVM >>>>>>>>>>> I have pulled the latest code on OpenSAF 5.1 branch, re-created >>> the >>>>>>>>> cluster. >>>>>>>>>>> And it works with the case old active SC-1 (OpenSAF 5.1) and >> new >>>>>>>> standby >>>>>>>>>>> SC-2 (with 2258_v7.patch included in). >>>>>>>>>>> >>>>>>>>>>> To apply 2258_v7.patch, please do remove the just pushed ticket >>>>>>>>>>> "log: >>>>>>>>>>> implement SaLogFilterSetCallbackT and version handling >> [#2146]" , >>>>>>>>>>> I have not rebased the code on that yet. >>>>>>>>>>> >>>>>>>>>>> Regards, Vu >>>>>>>>>>> >>>>>>>>>>>> -----Original Message----- >>>>>>>>>>>> From: A V Mahesh [mailto:mahesh.va...@oracle.com] >>>>>>>>>>>> Sent: Thursday, February 23, 2017 11:45 AM >>>>>>>>>>>> To: Vu Minh Nguyen <vu.m.ngu...@dektech.com.au>; >>>>>>>>>>>> lennart.l...@ericsson.com; canh.v.tru...@dektech.com.au >>>>>>>>>>>> Cc: opensaf-devel@lists.sourceforge.net >>>>>>>>>>>> Subject: Re: [devel] [PATCH 0 of 3] Review Request for log: > add >>>>>>>>>>> alternative >>>>>>>>>>>> destinations of log records [#2258] V4 >>>>>>>>>>>> >>>>>>>>>>>> Hi Vu/Lennart, >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> In broad WITHOUT the #2258 patch, the same code/setup >>> working >>>>> fine >>>>>>>>>> with >>>>>>>>>>>> 2 sc node (staging changeset: 8609 ), >>>>>>>>>>>> as soon as we apply `2258_v5.patch` V5 patch on staging >>>>> (changeset: >>>>>>>>>>>> 8609 ) that you have provided yesterday, >>>>>>>>>>>> on one sc node and try to bring up that in to cluster > (in-service >>>>>>> test) >>>>>>>>>>>> we are observing the issue of new node (with #2258 patch) not >>>>>>>>>>>> joining >>>>>>>>>>>> cluster. >>>>>>>>>>>> >>>>>>>>>>>> >> ============================================================== >> ==================================================== >>>>>>>>>>>> eb 23 10:01:59 SC-1 osafimmnd[15279]: NO Implementer >>> (applier) >>>>>>>>>>>> connected: 15 (@safAmfService2010f) <127, 2010f> >>>>>>>>>>>> Feb 23 10:01:59 SC-1 osaflogd[15289]: NO >>>>> LOGSV_DATA_GROUPNAME >>>>>>>>>> not >>>>>>>>>>>> found >>>>>>>>>>>> Feb 23 10:01:59 SC-1 osaflogd[15289]: NO LOG root directory >> is: >>>>>>>>>>>> "/var/log/opensaf/saflog" >>>>>>>>>>>> Feb 23 10:01:59 SC-1 osaflogd[15289]: NO LOG data group is: >> "" >>>>>>>>>>>> Feb 23 10:01:59 SC-1 osaflogd[15289]: NO >>> LGS_MBCSV_VERSION = >>>>> 7 >>>>>>>>>>>> Feb 23 10:01:59 SC-1 osafamfnd[15329]: NO Assigned >>>>>>>>>>>> 'safSi=SC-2N,safApp=OpenSAF' STANDBY to 'safSu=SC- >>>>>>>>>>>> 1,safSg=2N,safApp=OpenSAF' >>>>>>>>>>>> Feb 23 10:01:59 SC-1 opensafd: OpenSAF(5.1.M0 - ) services >>>>>>>> successfully >>>>>>>>>>>> started >>>>>>>>>>>> Feb 23 10:01:59 SC-1 osafamfnd[15329]: NO >>>>>>>>>>>> 'safComp=LOG,safSu=SC-1,safSg=2N,safApp=OpenSAF' faulted >>> due >>>>> to >>>>>>>>>>>> 'avaDown' : Recovery is 'nodeFailfast' >>>>>>>>>>>> Feb 23 10:01:59 SC-1 osafamfnd[15329]: ER >>>>>>>>>>>> safComp=LOG,safSu=SC-1,safSg=2N,safApp=OpenSAF Faulted >>> due >>>>>>>>>>>> to:avaDown >>>>>>>>>>>> Recovery is:nodeFailfast >>>>>>>>>>>> Feb 23 10:01:59 SC-1 osafamfnd[15329]: Rebooting OpenSAF >>>>> NodeId = >>>>>>>>>>>> 131343 >>>>>>>>>>>> EE Name = , Reason: Component faulted: recovery is node >>> failfast, >>>>>>>>>>>> OwnNodeId = 131343, SupervisionTime = 60 >>>>>>>>>>>> Feb 23 10:01:59 SC-1 opensaf_reboot: Rebooting local node; >>>>>>>> timeout=60 >>>>>>>>>>>> Feb 23 10:02:00 SC-1 osafimmnd[15279]: NO Implementer >>> (applier) >>>>>>>>>>>> connected: 16 (@OpenSafImmReplicatorB) <144, 2010f> >>>>>>>>>>>> Feb 23 10:01:59 SC-1 opensaf_reboot: Rebooting local node; >>>>>>>> timeout=60 >>>>>>>> >> ============================================================== >> ==================================================== >>>>>>>>>>>> So it is evident that in-service upgrade part code of this > need >>>>>>>>>>>> to be >>>>>>>>>>>> corrected. >>>>>>>>>>>> >>>>>>>>>>>> Please see my comments as [AVM] and let me know if you need >>>>> some >>>>>>>>>> traces >>>>>>>>>>>> . >>>>>>>>>>>> >>>>>>>>>>>> If you're planing to prepare new V6 patch , please do prepare >> on >>>>>>>>>>>> top >>>>>>> of >>>>>>>>>>>> today's latest staging. >>>>>>>>>>>> >>>>>>>>>>>> On 2/23/2017 9:33 AM, Vu Minh Nguyen wrote: >>>>>>>>>>>>> Hi Mahesh, >>>>>>>>>>>>> >>>>>>>>>>>>> I have done in-service upgrade/downgrade with following >>> cases: >>>>>>>>>>>>> 1) New Active SC-1 (OpenSAF 5.2 with the attached patch) + >> old >>>>>>>> standby >>>>>>>>>>>> SC-2 >>>>>>>>>>>>> (OpenSAF 5.1) >>>>>>>>>>>>> --> Work fine >>>>>>>>>>>> [AVM] This is not a practical use cause of in-service upgrade > , >> we >>>>>>> can >>>>>>>>>>>> ignore this test further >>>>>>>>>>>>> 2) Old Active SC-1 (OpenSAF 5.1) + new standby SC-2 (with or >>>>>>>>>>>>> without >>>>>>>>>>>>> attached patch) >>>>>>>>>>>>> --> SC-2 is restarted & not able to join the cluster. >>>>>>>>>>>> [AVM] This use cause/flow is we do get in in-service upgrade > , >>>>>>>>>>>> so we >>>>>>>>>>>> need to address this. >>>>>>>>>>>>> I got following messages in syslog: >>>>>>>>>>>>> Feb 23 09:32:42 SC-2 user.notice opensafd: OpenSAF(5.2.M0 - >>>>>>>>>>>>> 8529:b5addd36e45d:default) services successfully started >>>>>>>>>>>>> Feb 23 09:32:43 SC-2 local0.warn osafntfimcnd[701]: WA >>>>>>>>>>>> ntfimcn_imm_init >>>>>>>>>>>>> saImmOiImplementerSet() returned SA_AIS_ERR_TIMEOUT (5) >>>>>>>>>>>>> Feb 23 09:32:45 SC-2 local0.warn osafntfimcnd[701]: WA >>>>>>>>>>>> ntfimcn_imm_init >>>>>>>>>>>>> saImmOiImplementerSet() returned SA_AIS_ERR_TIMEOUT (5) >>>>>>>>>>>>> Feb 23 09:32:47 SC-2 local0.warn osafntfimcnd[701]: WA >>>>>>>>>>>> ntfimcn_imm_init >>>>>>>>>>>>> saImmOiImplementerSet() returned SA_AIS_ERR_TIMEOUT (5) >>>>>>>>>>>>> Feb 23 09:32:49 SC-2 local0.warn osafntfimcnd[701]: WA >>>>>>>>>>>> ntfimcn_imm_init >>>>>>>>>>>>> saImmOiImplementerSet() returned SA_AIS_ERR_TIMEOUT (5) >>>>>>>>>>>>> Feb 23 09:32:50 SC-2 local0.err osafmsgnd[592]: ER >>>>>>>>>>>> saImmOiImplementerSet >>>>>>>>>>>>> FAILED:5 >>>>>>>>>>>>> Feb 23 09:32:50 SC-2 local0.err osafmsgnd[592]: ER >>>>>>>>>>>> saImmOiImplementerSet >>>>>>>>>>>>> FAILED:5 >>>>>>>>>>>>> Feb 23 09:32:50 SC-2 local0.notice osafamfnd[496]: NO >>>>>>>>>>>>> 'safSu=SC-2,safSg=NoRed,safApp=OpenSAF' component >> restart >>>>>>>>>> probation >>>>>>>>>>>> timer >>>>>>>>>>>>> started (timeout: 60000000000 ns) >>>>>>>>>>>>> Feb 23 09:32:50 SC-2 local0.notice osafamfnd[496]: NO >>> Restarting >>>>> a >>>>>>>>>>>> component >>>>>>>>>>>>> of 'safSu=SC-2,safSg=NoRed,safApp=OpenSAF' (comp restart >>>>> count: 1) >>>>>>>>>>>>> Feb 23 09:32:50 SC-2 local0.notice osafamfnd[496]: NO >>>>>>>>>>>>> 'safComp=MQND,safSu=SC-2,safSg=NoRed,safApp=OpenSAF' >>>>> faulted >>>>>>>>>> due >>>>>>>>>>>> to >>>>>>>>>>>>> 'avaDown' : Recovery is 'componentRestart' >>>>>>>>>>>>> Feb 23 09:32:50 SC-2 local0.info osafmsgnd[736]: mkfifo >>> already >>>>>>>>> exists: >>>>>>>>>>>>> /var/lib/opensaf/osafmsgnd.fifo File exists >>>>>>>>>>>>> >>>>>>>>>>>>> And sometimes, on active SC-1 (OpenSAF 5.1), the node is not >>>>>>>>>>>>> able to >>>>>>>>> up >>>>>>>>>>>>> because of following error: >>>>>>>>>>>>> >>>>>>>>>>>>> Feb 23 11:00:32 SC-1 local0.err osafclmna[406]: MDTM:TIPC >>>>> Dsock >>>>>>>>>> Socket >>>>>>>>>>>>> creation failed in MDTM_INIT err :Address family not >> supported >>> by >>>>>>>>>>>> protocol >>>>>>>>>>>>> Feb 23 11:00:32 SC-1 local0.err osafclmna[406]: ER >>>>>>>> ncs_agents_startup >>>>>>>>>>>> FAILED >>>>>>>>>>>> [AVM] No such issues ( with both TCP & TIPC) (staging >>> changeset: >>>>>>>>> 8609 >>>>>>>>>>> ) >>>>>>>>>>>>> Are you getting similar problem at your side? >>>>>>>>>>>>> Please note that, the problem is existed WITH or WITHOUT >> the >>>>> #2258 >>>>>>>>>>>> patch. >>>>>>>>>>>> [AVM] No , problem only if we apply `2258_v5.patch` V5 patch >> on >>>>>>>> staging >>>>>>>>>>>> (changeset: 8609 ) >>>>>>>>>>>> try to bring up that node in to cluster. >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> -AVM >>>>>>>>>>>> >>>>>>>>>>>>> I have informed this to IMM to have a look, not sure any >>> problem >>>>>>> with >>>>>>>>>>> MDS >>>>>>>>>>>>> layer or any problem with my environment setup. >>>>>>>>>>>>> In the meantime, please have a look at the updated patch, I >> will >>>>>>>>>>> continue >>>>>>>>>>>>> checking the problem. Will keep you updated. >>>>>>>>>>>> [AVM] I haven't seen any IMM problems >>>>>>>>>>>>> Regards, Vu >>>>>>>>>>>>> >>>>>>>>>>>>>> -----Original Message----- >>>>>>>>>>>>>> From: A V Mahesh [mailto:mahesh.va...@oracle.com] >>>>>>>>>>>>>> Sent: Wednesday, February 22, 2017 5:36 PM >>>>>>>>>>>>>> To: Vu Minh Nguyen <vu.m.ngu...@dektech.com.au>; >>>>>>>>>>>>>> lennart.l...@ericsson.com; canh.v.tru...@dektech.com.au >>>>>>>>>>>>>> Cc: opensaf-devel@lists.sourceforge.net >>>>>>>>>>>>>> Subject: Re: [devel] [PATCH 0 of 3] Review Request for log: >>> add >>>>>>>>>>>>> alternative >>>>>>>>>>>>>> destinations of log records [#2258] V4 >>>>>>>>>>>>>> >>>>>>>>>>>>>> Please see correction New Standby SC-1 ( with patch ) >>>>>>>>>>>>>> >>>>>>>>>>>>>> -AVM >>>>>>>>>>>>>> >>>>>>>>>>>>>> On 2/22/2017 4:02 PM, A V Mahesh wrote: >>>>>>>>>>>>>>> Hi Vu, >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> With this new patch , we have another issue : >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> 1) standby Core by `/usr/lib64/opensaf/osaflogd' issue > got >>>>>>>>> resolved >>>>>>>>>>> . >>>>>>>>>>>>>>> 2) In-service upgrade is Not working , I have Old Active > SC- >> 2 >>>> ( >>>>>>>>> with >>>>>>>>>>>>>>> out patch ) and New Standby SC-1 ( with patch ) >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> the new New Standby SC-1 not joining the cluster ( >>>>>>> in-service >>>>>>>>>>>>>>> upgrade failed ) >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> New Standby SC-1 >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >> ============================================================== >>>>> ====================================================== >>>>>>>>>>>>>>> Feb 22 15:53:05 SC-1 osafamfnd[15752]: NO >>>>>>>>>>>>>>> 'safSu=SC-1,safSg=NoRed,safApp=OpenSAF' Presence State >>>>>>>>>>>>>> INSTANTIATING >>>>>>>>>>>>>>> => INSTANTIATED >>>>>>>>>>>>>>> Feb 22 15:53:05 SC-1 osafamfnd[15752]: NO Assigning >>>>>>>>>>>>>>> 'safSi=NoRed4,safApp=OpenSAF' ACTIVE to >>>>>>>>>>>>>>> 'safSu=SC-1,safSg=NoRed,safApp=OpenSAF' >>>>>>>>>>>>>>> Feb 22 15:53:05 SC-1 osafamfnd[15752]: NO Assigned >>>>>>>>>>>>>>> 'safSi=NoRed4,safApp=OpenSAF' ACTIVE to >>>>>>>>>>>>>>> 'safSu=SC-1,safSg=NoRed,safApp=OpenSAF' >>>>>>>>>>>>>>> Feb 22 15:53:05 SC-1 osafsmfd[15889]: Started >>>>>>>>>>>>>>> Feb 22 15:53:05 SC-1 osafamfnd[15752]: NO >>>>>>>>>>>>>>> 'safSu=SC-1,safSg=2N,safApp=OpenSAF' Presence State >>>>>>>>>> INSTANTIATING >>>>>>>>>>>> => >>>>>>>>>>>>>>> INSTANTIATED >>>>>>>>>>>>>>> Feb 22 15:53:05 SC-1 osafamfnd[15752]: NO Assigning >>>>>>>>>>>>>>> 'safSi=SC-2N,safApp=OpenSAF' STANDBY to >>>>>>>>>>>>>>> 'safSu=SC-1,safSg=2N,safApp=OpenSAF' >>>>>>>>>>>>>>> Feb 22 15:53:05 SC-1 osafrded[15672]: NO RDE role set to >>>>>>>> STANDBY >>>>>>>>>>>>>>> Feb 22 15:53:05 SC-1 osafrded[15672]: NO Peer up on node >>>>>>>> 0x2020f >>>>>>>>>>>>>>> Feb 22 15:53:05 SC-1 osafrded[15672]: NO Got peer info >>>>> request >>>>>>>> from >>>>>>>>>>>>>>> node 0x2020f with role ACTIVE >>>>>>>>>>>>>>> Feb 22 15:53:05 SC-1 osafrded[15672]: NO Got peer info >>>>> response >>>>>>>>>> from >>>>>>>>>>>>>>> node 0x2020f with role ACTIVE >>>>>>>>>>>>>>> Feb 22 15:53:05 SC-1 osafimmd[15691]: NO MDS event >> from >>>>> svc_id >>>>>>>> 24 >>>>>>>>>>>>>>> (change:5, dest:13) >>>>>>>>>>>>>>> Feb 22 15:53:05 SC-1 osafimmd[15691]: NO MDS event >> from >>>>> svc_id >>>>>>>> 24 >>>>>>>>>>>>>>> (change:3, dest:13) >>>>>>>>>>>>>>> Feb 22 15:53:05 SC-1 osafimmd[15691]: NO MDS event >> from >>>>> svc_id >>>>>>>> 24 >>>>>>>>>>>>>>> (change:5, dest:13) >>>>>>>>>>>>>>> Feb 22 15:53:05 SC-1 osafimmd[15691]: NO MDS event >> from >>>>> svc_id >>>>>>>> 25 >>>>>>>>>>>>>>> (change:3, dest:567412424453430) >>>>>>>>>>>>>>> Feb 22 15:53:05 SC-1 osafimmd[15691]: NO MDS event >> from >>>>> svc_id >>>>>>>> 25 >>>>>>>>>>>>>>> (change:3, dest:565213401202663) >>>>>>>>>>>>>>> Feb 22 15:53:05 SC-1 osafimmd[15691]: NO MDS event >> from >>>>> svc_id >>>>>>>> 25 >>>>>>>>>>>>>>> (change:3, dest:566312912825221) >>>>>>>>>>>>>>> Feb 22 15:53:05 SC-1 osafimmd[15691]: NO MDS event >> from >>>>> svc_id >>>>>>>> 25 >>>>>>>>>>>>>>> (change:3, dest:564113889574230) >>>>>>>>>>>>>>> Feb 22 15:53:05 SC-1 osafimmnd[15702]: NO Implementer >>>>> (applier) >>>>>>>>>>>>>>> connected: 17 (@safAmfService2010f) <127, 2010f> >>>>>>>>>>>>>>> Feb 22 15:53:05 SC-1 osaflogd[15712]: NO >>>>>>>>>> LOGSV_DATA_GROUPNAME >>>>>>>>>>>>>> not found >>>>>>>>>>>>>>> Feb 22 15:53:05 SC-1 osaflogd[15712]: NO LOG root >>> directory >>>>> is: >>>>>>>>>>>>>>> "/var/log/opensaf/saflog" >>>>>>>>>>>>>>> Feb 22 15:53:05 SC-1 osaflogd[15712]: NO LOG data group >> is: >>> "" >>>>>>>>>>>>>>> Feb 22 15:53:05 SC-1 osaflogd[15712]: NO >>>>> LGS_MBCSV_VERSION = >>>>>>>> 7 >>>>>>>>>>>>>>> Feb 22 15:53:05 SC-1 osafamfnd[15752]: NO Assigned >>>>>>>>>>>>>>> 'safSi=SC-2N,safApp=OpenSAF' STANDBY to >>>>>>>>>>>>>>> 'safSu=SC-1,safSg=2N,safApp=OpenSAF' >>>>>>>>>>>>>>> Feb 22 15:53:05 SC-1 opensafd: OpenSAF(5.1.M0 - ) services >>>>>>>>>>>>>>> successfully started >>>>>>>>>>>>>>> Feb 22 15:53:05 SC-1 osafamfnd[15752]: NO >>>>>>>>>>>>>>> 'safComp=LOG,safSu=SC-1,safSg=2N,safApp=OpenSAF' >>> faulted >>>>> due >>>>>>>> to >>>>>>>>>>>>>>> 'avaDown' : Recovery is 'nodeFailfast' >>>>>>>>>>>>>>> Feb 22 15:53:05 SC-1 osafamfnd[15752]: ER >>>>>>>>>>>>>>> safComp=LOG,safSu=SC-1,safSg=2N,safApp=OpenSAF >>> Faulted >>>>> due >>>>>>>>>>>>>> to:avaDown >>>>>>>>>>>>>>> Recovery is:nodeFailfast >>>>>>>>>>>>>>> Feb 22 15:53:05 SC-1 osafamfnd[15752]: Rebooting >> OpenSAF >>>>>>>> NodeId >>>>>>>>>> = >>>>>>>>>>>>>>> 131343 EE Name = , Reason: Component faulted: recovery >> is >>>>> node >>>>>>>>>>>>>>> failfast, OwnNodeId = 131343, SupervisionTime = 60 >>>>>>>>>>>>>>> Feb 22 15:53:05 SC-1 opensaf_reboot: Rebooting local >> node; >>>>>>>>>>>> timeout=60 >>>>>>>>>>>>>>> Feb 22 15:53:43 SC-1 syslog-ng[1171]: syslog-ng starting >> up; >>>>>>>>>>>>>>> version='2.0.9' >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >> ============================================================== >>>>> ====================================================== >>>>>>>>>>>>>>> Old - Active - SC-2 >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >> ============================================================== >>>>> ====================================================== >>>>>>>>>>>>>>> Feb 22 15:53:02 SC-2 osafimmnd[16359]: NO NODE STATE- >>>>>>>>>>>>>>> IMM_NODE_R_AVAILABLE >>>>>>>>>>>>>>> Feb 22 15:53:02 SC-2 osafimmloadd: NO Sync starting >>>>>>>>>>>>>>> Feb 22 15:53:02 SC-2 osafimmloadd: IN Synced 390 objects >> in >>>>>>>>>>>>>>> total >>>>>>>>>>>>>>> Feb 22 15:53:02 SC-2 osafimmnd[16359]: NO NODE STATE- >>>>>>>>>>>>>>> IMM_NODE_FULLY_AVAILABLE 18511 >>>>>>>>>>>>>>> Feb 22 15:53:02 SC-2 osafimmnd[16359]: NO Epoch set to 3 >>> in >>>>>>>>>>>> ImmModel >>>>>>>>>>>>>>> Feb 22 15:53:02 SC-2 osafimmd[16346]: NO ACT: New >> Epoch >>>>> for >>>>>>>>>>>> IMMND >>>>>>>>>>>>>>> process at node 2020f old epoch: 2 new epoch:3 >>>>>>>>>>>>>>> Feb 22 15:53:02 SC-2 osafimmd[16346]: NO ACT: New >> Epoch >>>>> for >>>>>>>>>>>> IMMND >>>>>>>>>>>>>>> process at node 2040f old epoch: 2 new epoch:3 >>>>>>>>>>>>>>> Feb 22 15:53:02 SC-2 osafimmd[16346]: NO ACT: New >> Epoch >>>>> for >>>>>>>>>>>> IMMND >>>>>>>>>>>>>>> process at node 2030f old epoch: 2 new epoch:3 >>>>>>>>>>>>>>> Feb 22 15:53:02 SC-2 osafimmloadd: NO Sync ending >>> normally >>>>>>>>>>>>>>> Feb 22 15:53:02 SC-2 osafimmd[16346]: NO ACT: New >> Epoch >>>>> for >>>>>>>>>>>> IMMND >>>>>>>>>>>>>>> process at node 2010f old epoch: 0 new epoch:3 >>>>>>>>>>>>>>> Feb 22 15:53:02 SC-2 osafimmnd[16359]: NO SERVER >> STATE: >>>>>>>>>>>>>>> IMM_SERVER_SYNC_SERVER --> IMM_SERVER_READY >>>>>>>>>>>>>>> Feb 22 15:53:03 SC-2 osafamfd[16408]: NO Received >>> node_up >>>>> from >>>>>>>>>>>> 2010f: >>>>>>>>>>>>>>> msg_id 1 >>>>>>>>>>>>>>> Feb 22 15:53:03 SC-2 osafamfd[16408]: NO Node 'SC-1' >>> joined >>>>> the >>>>>>>>>>>> cluster >>>>>>>>>>>>>>> Feb 22 15:53:03 SC-2 osafimmnd[16359]: NO Implementer >>>>>>>> connected: >>>>>>>>>>>> 16 >>>>>>>>>>>>>>> (MsgQueueService131343) <0, 2010f> >>>>>>>>>>>>>>> Feb 22 15:53:03 SC-2 osafrded[16327]: NO Peer up on node >>>>>>>> 0x2010f >>>>>>>>>>>>>>> Feb 22 15:53:03 SC-2 osafrded[16327]: NO Got peer info >>>>> request >>>>>>>> from >>>>>>>>>>>>>>> node 0x2010f with role STANDBY >>>>>>>>>>>>>>> Feb 22 15:53:03 SC-2 osafrded[16327]: NO Got peer info >>>>> response >>>>>>>>>> from >>>>>>>>>>>>>>> node 0x2010f with role STANDBY >>>>>>>>>>>>>>> Feb 22 15:53:03 SC-2 osafimmd[16346]: NO MDS event >> from >>>>> svc_id >>>>>>>> 24 >>>>>>>>>>>>>>> (change:5, dest:13) >>>>>>>>>>>>>>> Feb 22 15:53:03 SC-2 osafimmnd[16359]: NO Implementer >>>>> (applier) >>>>>>>>>>>>>>> connected: 17 (@safAmfService2010f) <0, 2010f> >>>>>>>>>>>>>>> Feb 22 15:53:03 SC-2 osafamfd[16408]: NO Cluster startup >> is >>>>> done >>>>>>>>>>>>>>> Feb 22 15:53:04 SC-2 osafimmnd[16359]: NO Implementer >>>>> (applier) >>>>>>>>>>>>>>> connected: 18 (@OpenSafImmReplicatorB) <0, 2010f> >>>>>>>>>>>>>>> Feb 22 15:53:09 SC-2 osafdtmd[16304]: NO Lost contact >> with >>>>>>>>>>>>>>> 'SC-1' >>>>>>>>>>>>>>> Feb 22 15:53:09 SC-2 osaffmd[16336]: NO Node Down >> event >>>>> for >>>>>>>> node >>>>>>>>>> id >>>>>>>>>>>>>>> 2010f: >>>>>>>>>>>>>>> Feb 22 15:53:09 SC-2 osafimmd[16346]: NO MDS event >> from >>>>> svc_id >>>>>>>> 24 >>>>>>>>>>>>>>> (change:6, dest:13) >>>>>>>>>>>>>>> Feb 22 15:53:09 SC-2 osafimmd[16346]: NO MDS event >> from >>>>> svc_id >>>>>>>> 25 >>>>>>>>>>>>>>> (change:4, dest:564113889574230) >>>>>>>>>>>>>>> Feb 22 15:53:09 SC-2 osaffmd[16336]: NO Current role: >>> ACTIVE >>>>>>>>>>>>>>> Feb 22 15:53:09 SC-2 osaffmd[16336]: Rebooting OpenSAF >>>>> NodeId >>>>>>>> = >>>>>>>>>>>>>> 131343 >>>>>>>>>>>>>>> EE Name = , Reason: Received Node Down for peer >>> controller, >>>>>>>>>>>> OwnNodeId >>>>>>>>>>>>>>> = 131599, SupervisionTime = 60 >>>>>>>>>>>>>>> Feb 22 15:53:09 SC-2 osafrded[16327]: NO Peer down on >>> node >>>>>>>>>> 0x2010f >>>>>>>>>>>>>>> Feb 22 15:53:09 SC-2 osafamfd[16408]: NO Node 'SC-1' left >>> the >>>>>>>>>> cluster >>>>>>>>>>>>>>> Feb 22 15:53:09 SC-2 osaflogd[16372]: NO Failed (2) to >> send >>> of >>>>>>>> WRITE >>>>>>>>>>>>>>> ack to: 2010f00003d6a >>>>>>>>>>>>>>> Feb 22 15:53:09 SC-2 osaflogd[16372]: NO Failed (2) to >> send >>> of >>>>>>>> WRITE >>>>>>>>>>>>>>> ack to: 2010f00003d6a >>>>>>>>>>>>>>> Feb 22 15:53:09 SC-2 osaflogd[16372]: NO Failed (2) to >> send >>> of >>>>>>>> WRITE >>>>>>>>>>>>>>> ack to: 2010f00003d74 >>>>>>>>>>>>>>> Feb 22 15:53:09 SC-2 osafclmd[16398]: NO Node 131343 >>> went >>>>>>>> down. >>>>>>>>>>>> Not >>>>>>>>>>>>>>> sending track callback for agents on that node >>>>>>>>>>>>>>> Feb 22 15:53:09 SC-2 osafclmd[16398]: NO Node 131343 >>> went >>>>>>>> down. >>>>>>>>>>>> Not >>>>>>>>>>>>>>> sending track callback for agents on that node >>>>>>>>>>>>>>> Feb 22 15:53:09 SC-2 osafclmd[16398]: NO Node 131343 >>> went >>>>>>>> down. >>>>>>>>>>>> Not >>>>>>>>>>>>>>> sending track callback for agents on that node >>>>>>>>>>>>>>> Feb 22 15:53:09 SC-2 osafclmd[16398]: NO Node 131343 >>> went >>>>>>>> down. >>>>>>>>>>>> Not >>>>>>>>>>>>>>> sending track callback for agents on that node >>>>>>>>>>>>>>> Feb 22 15:53:09 SC-2 osafclmd[16398]: NO Node 131343 >>> went >>>>>>>> down. >>>>>>>>>>>> Not >>>>>>>>>>>>>>> sending track callback for agents on that node >>>>>>>>>>>>>>> Feb 22 15:53:09 SC-2 osafclmd[16398]: NO Node 131343 >>> went >>>>>>>> down. >>>>>>>>>>>> Not >>>>>>>>>>>>>>> sending track callback for agents on that node >>>>>>>>>>>>>>> Feb 22 15:53:09 SC-2 osafclmd[16398]: NO Node 131343 >>> went >>>>>>>> down. >>>>>>>>>>>> Not >>>>>>>>>>>>>>> sending track callback for agents on that node >>>>>>>>>>>>>>> Feb 22 15:53:09 SC-2 osafimmd[16346]: WA IMMD lost >>> contact >>>>>>>> with >>>>>>>>>>>> peer >>>>>>>>>>>>>>> IMMD (NCSMDS_RED_DOWN) >>>>>>>>>>>>>>> Feb 22 15:53:09 SC-2 osafimmnd[16359]: NO Global discard >>>>> node >>>>>>>>>>>> received >>>>>>>>>>>>>>> for nodeId:2010f pid:15702 >>>>>>>>>>>>>>> Feb 22 15:53:09 SC-2 osafimmnd[16359]: NO Implementer >>>>>>>>>> disconnected >>>>>>>>>>>> 16 >>>>>>>>>>>>>>> <0, 2010f(down)> (MsgQueueService131343) >>>>>>>>>>>>>>> Feb 22 15:53:09 SC-2 osafimmnd[16359]: NO Implementer >>>>>>>>>> disconnected >>>>>>>>>>>> 17 >>>>>>>>>>>>>>> <0, 2010f(down)> (@safAmfService2010f) >>>>>>>>>>>>>>> Feb 22 15:53:09 SC-2 osafimmnd[16359]: NO Implementer >>>>>>>>>> disconnected >>>>>>>>>>>> 18 >>>>>>>>>>>>>>> <0, 2010f(down)> (@OpenSafImmReplicatorB) >>>>>>>>>>>>>>> Feb 22 15:53:09 SC-2 opensaf_reboot: Rebooting remote >>> node >>>>> in >>>>>>>> the >>>>>>>>>>>>>>> absence of PLM is outside the scope of OpenSAF >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >> ============================================================== >>>>> ====================================================== >>>>>>>>>>>>>>> -AVM >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> On 2/22/2017 3:13 PM, Vu Minh Nguyen wrote: >>>>>>>>>>>>>>>> Hi Mahesh, >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> I put all required patches into one. Try to use this and > see >>>> if >>>>>>> you >>>>>>>>>>>>>>>> still >>>>>>>>>>>>>>>> have that problem or not. >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> Regards, Vu >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> -----Original Message----- >>>>>>>>>>>>>>>>> From: A V Mahesh [mailto:mahesh.va...@oracle.com] >>>>>>>>>>>>>>>>> Sent: Wednesday, February 22, 2017 3:35 PM >>>>>>>>>>>>>>>>> To: Vu Minh Nguyen <vu.m.ngu...@dektech.com.au>; >>>>>>>>>>>>>>>>> lennart.l...@ericsson.com; >>> canh.v.tru...@dektech.com.au >>>>>>>>>>>>>>>>> Cc: opensaf-devel@lists.sourceforge.net >>>>>>>>>>>>>>>>> Subject: Re: [devel] [PATCH 0 of 3] Review Request for >> log: >>>>>>>>>>>>>>>>> add >>>>>>>>>>>>>>>> alternative >>>>>>>>>>>>>>>>> destinations of log records [#2258] V4 >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> Hi Vu, >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> I used new #3 and #4 patches , Can you please re-send >> All >>>>>>>>>>>>>>>>> the >>>>>>>>>>> final >>>>>>>>>>>>>>>>> patch in go, >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> which i need to apply on today`s staging ( if possible >>>>>>>>>>>>>>>>> publish >>>>>>>>> the >>>>>>>>>>>>>>>>> with >>>>>>>>>>>>>>>>> new version ) >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> -AVM >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> On 2/22/2017 1:52 PM, Vu Minh Nguyen wrote: >>>>>>>>>>>>>>>>>> Hi Mahesh, >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> Core was generated by `/usr/lib64/opensaf/osaflogd'. >>>>>>>>>>>>>>>>>>> Program terminated with signal 11, Segmentation >> fault. >>>>>>>>>>>>>>>>>>> #0 ckpt_proc_cfg_stream(lgs_cb*, void*) () at >>>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc:2195 >>>>>>>>>>>>>>>>>>> 2195 src/log/logd/lgs_mbcsv.cc: No such file or >>>>>>>>>>>>>>>>>>> directory. >>>>>>>>>>>>>>>>>>> in src/log/logd/lgs_mbcsv.cc >>>>>>>>>>>>>>>>>> Backtrace still points to old position (lgs_mbcsv:2195). > I >>>>>>> guess >>>>>>>>>>> the >>>>>>>>>>>>>>>>>> osaflogd binary has not been updated with the fixed >>> patch. >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> Regards, Vu >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> -----Original Message----- >>>>>>>>>>>>>>>>>>> From: A V Mahesh [mailto:mahesh.va...@oracle.com] >>>>>>>>>>>>>>>>>>> Sent: Wednesday, February 22, 2017 3:18 PM >>>>>>>>>>>>>>>>>>> To: Vu Minh Nguyen >> <vu.m.ngu...@dektech.com.au>; >>>>>>>>>>>>>>>>>>> lennart.l...@ericsson.com; >>>>> canh.v.tru...@dektech.com.au >>>>>>>>>>>>>>>>>>> Cc: opensaf-devel@lists.sourceforge.net >>>>>>>>>>>>>>>>>>> Subject: Re: [devel] [PATCH 0 of 3] Review Request for >>> log: >>>>>>>> add >>>>>>>>>>>>>>>>>> alternative >>>>>>>>>>>>>>>>>>> destinations of log records [#2258] V4 >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> Hi Vu, >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> SC-2 standby osaflogd core dumped still occurs ( Not >>>>>>>>>>>>>>>>>>> resolved) >>>>>>>> , >>>>>>>>>>> the >>>>>>>>>>>>>>>> new >>>>>>>>>>>>>>>>>>> patch only resolved the application (/usr/bin/logtest ) >>>>>>>>>>> Segmentation >>>>>>>>>>>>>>>>>>> fault on SC-1 Active. >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> >> ============================================================== >> ============================================================== >>>>>>>>>>>>>>>>>>> # gdb /usr/lib64/opensaf/osaflogd >>>>>>>>>>>> core_1487751055.osaflogd.4594 >>>>>>>>>>>>>> GNU >>>>>>>>>>>>>>>>>>> gdb >>>>>>>>>>>>>>>>>>> (GDB) SUSE (7.3-0.6.1) >>>>>>>>>>>>>>>>>>> Copyright (C) 2011 Free Software Foundation, Inc. >>>>>>>>>>>>>>>>>>> ....... >>>>>>>>>>>>>>>>>>> Core was generated by `/usr/lib64/opensaf/osaflogd'. >>>>>>>>>>>>>>>>>>> Program terminated with signal 11, Segmentation >> fault. >>>>>>>>>>>>>>>>>>> #0 ckpt_proc_cfg_stream(lgs_cb*, void*) () at >>>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc:2195 >>>>>>>>>>>>>>>>>>> 2195 src/log/logd/lgs_mbcsv.cc: No such file or >>>>>>>>>>>>>>>>>>> directory. >>>>>>>>>>>>>>>>>>> in src/log/logd/lgs_mbcsv.cc >>>>>>>>>>>>>>>>>>> (gdb) bt >>>>>>>>>>>>>>>>>>> #0 ckpt_proc_cfg_stream(lgs_cb*, void*) () at >>>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc:2195 >>>>>>>>>>>>>>>>>>> #1 0x00007f97b026f960 in >>>>> ckpt_decode_log_struct(lgs_cb*, >>>>>>>>>>>>>>>>>>> ncs_mbcsv_cb_arg*, void*, void*, unsigned int >>>>>>>> (*)(edu_hdl_tag*, >>>>>>>>>>>>>>>>>>> edu_tkn_tag*, void*, unsigned int*, >> edu_buf_env_tag*, >>>>>>>>>>>>>> EDP_OP_TYPE, >>>>>>>>>>>>>>>>>>> EDU_ERR*)) () at src/log/logd/lgs_mbcsv.cc:950 >>>>>>>>>>>>>>>>>>> #2 0x00007f97b02710dc in >>>>>>>> ckpt_decode_async_update(lgs_cb*, >>>>>>>>>>>>>>>>>>> ncs_mbcsv_cb_arg*) () at >>> src/log/logd/lgs_mbcsv.cc:1086 >>>>>>>>>>>>>>>>>>> #3 0x00007f97b0273941 in >>>>>>>> mbcsv_callback(ncs_mbcsv_cb_arg*) >>>>>>>>>> () >>>>>>>>>>>> at >>>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc:880 >>>>>>>>>>>>>>>>>>> #4 0x00007f97af372596 in ncs_mbscv_rcv_decode () >>> from >>>>>>>>>>>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0 >>>>>>>>>>>>>>>>>>> #5 0x00007f97af372766 in >>> ncs_mbcsv_rcv_async_update >>>>> () >>>>>>>> from >>>>>>>>>>>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0 >>>>>>>>>>>>>>>>>>> #6 0x00007f97af379370 in mbcsv_process_events () >>> from >>>>>>>>>>>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0 >>>>>>>>>>>>>>>>>>> #7 0x00007f97af3794db in mbcsv_hdl_dispatch_all () >>> from >>>>>>>>>>>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0 >>>>>>>>>>>>>>>>>>> #8 0x00007f97af373ce2 in >>>>> mbcsv_process_dispatch_request () >>>>>>>>>> at >>>>>>>>>>>>>>>>>>> src/mbc/mbcsv_api.c:423 >>>>>>>>>>>>>>>>>>> #9 0x00007f97b027096e in >>> lgs_mbcsv_dispatch(unsigned >>>>> int) >>>>>>>> () >>>>>>>>>> at >>>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc:327 >>>>>>>>>>>>>>>>>>> #10 0x00007f97b024d9f2 in main () at >>>>>>>>>> src/log/logd/lgs_main.cc:583 >>>>>>>>>>>>>>>>>>> (gdb) bt full >>>>>>>>>>>>>>>>>>> #0 ckpt_proc_cfg_stream(lgs_cb*, void*) () at >>>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc:2195 >>>>>>>>>>>>>>>>>>> ckpt_data_handler = {0x7f97b0270300 >>>>>>>>>>>>>>>>>>> <ckpt_proc_initialize_client(lgs_cb*, void*)>, >>>>>>>>>>>>>>>>>>> 0x7f97b02701c0 >>>>>>>>>>>>>>>>>>> <ckpt_proc_finalize_client(lgs_cb*, void*)>, >>>>>>>>>>>>>>>>>>> 0x7f97b0270060 >>>>>>>>>>>>>>>>>>> <ckpt_proc_agent_down(lgs_cb*, >>>>>>>>>>> void*)>, >>>>>>>>>>>>>>>>>>> 0x7f97b02712f0 <ckpt_proc_log_write(lgs_cb*, >> void*)>, >>>>>>>>>>>>>> 0x7f97b0271ab0 >>>>>>>>>>>>>>>>>>> <ckpt_proc_open_stream(lgs_cb*, void*)>, >>>>>>>>>>>>>>>>>>> 0x7f97b026fe80 >>>>>>> <ckpt_proc_close_stream(lgs_cb*, >>>>>>>>>>>>> void*)>, >>>>>>>>>>>>>>>>>>> 0x7f97b0272380 <ckpt_proc_cfg_stream(lgs_cb*, >>> void*)>, >>>>>>>>>>>>>>>>> 0x7f97b0274800 >>>>>>>>>>>>>>>>>>> <ckpt_proc_lgs_cfg_v2(lgs_cb*, void*)>, >>>>>>>>>>>>>>>>>>> 0x7f97b0274e10 >>>>>>>>>>>>>>>>>>> <ckpt_proc_lgs_cfg_v3(lgs_cb*, >>>>>>>>>>> void*)>, >>>>>>>>>>>>>>>>>>> 0x7f97b02754f0 <ckpt_proc_lgs_cfg_v5(lgs_cb*, >>> void*)>} >>>>>>>>>>>>>>>>>>> #1 0x00007f97b026f960 in >>>>> ckpt_decode_log_struct(lgs_cb*, >>>>>>>>>>>>>>>>>>> ncs_mbcsv_cb_arg*, void*, void*, unsigned int >>>>>>>> (*)(edu_hdl_tag*, >>>>>>>>>>>>>>>>>>> edu_tkn_tag*, void*, unsigned int*, >> edu_buf_env_tag*, >>>>>>>>>>>>>> EDP_OP_TYPE, >>>>>>>>>>>>>>>>>>> EDU_ERR*)) () at src/log/logd/lgs_mbcsv.cc:950 >>>>>>>>>>>>>>>>>>> ckpt_data_handler = {0x7f97b0270300 >>>>>>>>>>>>>>>>>>> <ckpt_proc_initialize_client(lgs_cb*, void*)>, >>>>>>>>>>>>>>>>>>> 0x7f97b02701c0 >>>>>>>>>>>>>>>>>>> <ckpt_proc_finalize_client(lgs_cb*, void*)>, >>>>>>>>>>>>>>>>>>> 0x7f97b0270060 >>>>>>>>>>>>>>>>>>> <ckpt_proc_agent_down(lgs_cb*, >>>>>>>>>>> void*)>, >>>>>>>>>>>>>>>>>>> 0x7f97b02712f0 <ckpt_proc_log_write(lgs_cb*, >> void*)>, >>>>>>>>>>>>>> 0x7f97b0271ab0 >>>>>>>>>>>>>>>>>>> <ckpt_proc_open_stream(lgs_cb*, void*)>, >>>>>>>>>>>>>>>>>>> 0x7f97b026fe80 >>>>>>> <ckpt_proc_close_stream(lgs_cb*, >>>>>>>>>>>>> void*)>, >>>>>>>>>>>>>>>>>>> 0x7f97b0272380 <ckpt_proc_cfg_stream(lgs_cb*, >>> void*)>, >>>>>>>>>>>>>>>>> 0x7f97b0274800 >>>>>>>>>>>>>>>>>>> <ckpt_proc_lgs_cfg_v2(lgs_cb*, void*)>, >>>>>>>>>>>>>>>>>>> 0x7f97b0274e10 >>>>>>>>>>>>>>>>>>> <ckpt_proc_lgs_cfg_v3(lgs_cb*, >>>>>>>>>>> void*)>, >>>>>>>>>>>>>>>>>>> 0x7f97b02754f0 <ckpt_proc_lgs_cfg_v5(lgs_cb*, >>> void*)>} >>>>>>>>>>>>>>>>>>> #2 0x00007f97b02710dc in >>>>>>>> ckpt_decode_async_update(lgs_cb*, >>>>>>>>>>>>>>>>>>> ncs_mbcsv_cb_arg*) () at >>> src/log/logd/lgs_mbcsv.cc:1086 >>>>>>>>>>>>>>>>>>> ckpt_data_handler = {0x7f97b0270300 >>>>>>>>>>>>>>>>>>> <ckpt_proc_initialize_client(lgs_cb*, void*)>, >>>>>>>>>>>>>>>>>>> 0x7f97b02701c0 >>>>>>>>>>>>>>>>>>> <ckpt_proc_finalize_client(lgs_cb*, void*)>, >>>>>>>>>>>>>>>>>>> 0x7f97b0270060 >>>>>>>>>>>>>>>>>>> <ckpt_proc_agent_down(lgs_cb*, >>>>>>>>>>> void*)>, >>>>>>>>>>>>>>>>>>> 0x7f97b02712f0 <ckpt_proc_log_write(lgs_cb*, >> void*)>, >>>>>>>>>>>>>> 0x7f97b0271ab0 >>>>>>>>>>>>>>>>>>> <ckpt_proc_open_stream(lgs_cb*, void*)>, >>>>>>>>>>>>>>>>>>> 0x7f97b026fe80 >>>>>>> <ckpt_proc_close_stream(lgs_cb*, >>>>>>>>>>>>> void*)>, >>>>>>>>>>>>>>>>>>> 0x7f97b0272380 <ckpt_proc_cfg_stream(lgs_cb*, >>> void*)>, >>>>>>>>>>>>>>>>> 0x7f97b0274800 >>>>>>>>>>>>>>>>>>> <ckpt_proc_lgs_cfg_v2(lgs_cb*, void*)>, >>>>>>>>>>>>>>>>>>> 0x7f97b0274e10 >>>>>>>>>>>>>>>>>>> <ckpt_proc_lgs_cfg_v3(lgs_cb*, >>>>>>>>>>> void*)>, >>>>>>>>>>>>>>>>>>> 0x7f97b02754f0 <ckpt_proc_lgs_cfg_v5(lgs_cb*, >>> void*)>} >>>>>>>>>>>>>>>>>>> #3 0x00007f97b0273941 in >>>>>>>> mbcsv_callback(ncs_mbcsv_cb_arg*) >>>>>>>>>> () >>>>>>>>>>>> at >>>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc:880 >>>>>>>>>>>>>>>>>>> ckpt_data_handler = {0x7f97b0270300 >>>>>>>>>>>>>>>>>>> <ckpt_proc_initialize_client(lgs_cb*, void*)>, >>>>>>>>>>>>>>>>>>> 0x7f97b02701c0 >>>>>>>>>>>>>>>>>>> <ckpt_proc_finalize_client(lgs_cb*, void*)>, >>>>>>>>>>>>>>>>>>> 0x7f97b0270060 >>>>>>>>>>>>>>>>>>> <ckpt_proc_agent_down(lgs_cb*, >>>>>>>>>>> void*)>, >>>>>>>>>>>>>>>>>>> 0x7f97b02712f0 <ckpt_proc_log_write(lgs_cb*, >> void*)>, >>>>>>>>>>>>>> 0x7f97b0271ab0 >>>>>>>>>>>>>>>>>>> <ckpt_proc_open_stream(lgs_cb*, void*)>, >>>>>>>>>>>>>>>>>>> 0x7f97b026fe80 >>>>>>> <ckpt_proc_close_stream(lgs_cb*, >>>>>>>>>>>>> void*)>, >>>>>>>>>>>>>>>>>>> 0x7f97b0272380 <ckpt_proc_cfg_stream(lgs_cb*, >>> void*)>, >>>>>>>>>>>>>>>>> 0x7f97b0274800 >>>>>>>>>>>>>>>>>>> <ckpt_proc_lgs_cfg_v2(lgs_cb*, void*)>, >>>>>>>>>>>>>>>>>>> 0x7f97b0274e10 >>>>>>>>>>>>>>>>>>> <ckpt_proc_lgs_cfg_v3(lgs_cb*, >>>>>>>>>>> void*)>, >>>>>>>>>>>>>>>>>>> 0x7f97b02754f0 <ckpt_proc_lgs_cfg_v5(lgs_cb*, >>> void*)>} >>>>>>>>>>>>>>>>>>> #4 0x00007f97af372596 in ncs_mbscv_rcv_decode () >>> from >>>>>>>>>>>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0 >>>>>>>>>>>>>>>>>>> mbcsv_init_process_req_func = {0x7f97af373630 >>>>>>>>>>>>>>>>>>> <mbcsv_process_initialize_request>, 0x7f97af374f10 >>>>>>>>>>>>>>>>>>> <mbcsv_process_get_sel_obj_request>, >>>>>>>>>>>>>>>>>>> 0x7f97af373be0 >>>>>>> <mbcsv_process_dispatch_request>, >>>>>>>>>>>>>>>>>>> 0x7f97af373af0 <mbcsv_process_finalize_request>, >>>>>>>>>>>> 0x7f97af373cf0 >>>>>>>>>>>>>>>>>>> <mbcsv_process_open_request>, >>>>>>>>>>>>>>>>>>> 0x7f97af374050 >>>>>>>>>>>>>>>>>>> <mbcsv_process_close_request>, >>>>>>>>>>>>>>>> 0x7f97af3741e0 >>>>>>>>>>>>>>>>>>> <mbcsv_process_chg_role_request>, 0x7f97af3744c0 >>>>>>>>>>>>>>>>>>> <mbcsv_process_snd_ckpt_request>, >>>>>>>>>>>>>>>>>>> 0x7f97af3747d0 >>>>>>> <mbcsv_process_snd_ntfy_request>, >>>>>>>>>>>>>>>>>>> 0x7f97af374970 <mbcsv_process_snd_data_req>, >>>>>>>>>> 0x7f97af373930 >>>>>>>>>>>>>>>>>>> <mbcsv_process_get_request>, >>>>>>>>>>>>>>>>>>> 0x7f97af374bd0 >>>> <mbcsv_process_set_request>} >>>>>>>>>>>>>>>>>>> #5 0x00007f97af372766 in >>> ncs_mbcsv_rcv_async_update >>>>> () >>>>>>>> from >>>>>>>>>>>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0 >>>>>>>>>>>>>>>>>>> mbcsv_init_process_req_func = {0x7f97af373630 >>>>>>>>>>>>>>>>>>> <mbcsv_process_initialize_request>, 0x7f97af374f10 >>>>>>>>>>>>>>>>>>> <mbcsv_process_get_sel_obj_request>, >>>>>>>>>>>>>>>>>>> 0x7f97af373be0 >>>>>>> <mbcsv_process_dispatch_request>, >>>>>>>>>>>>>>>>>>> 0x7f97af373af0 <mbcsv_process_finalize_request>, >>>>>>>>>>>> 0x7f97af373cf0 >>>>>>>>>>>>>>>>>>> <mbcsv_process_open_request>, >>>>>>>>>>>>>>>>>>> 0x7f97af374050 >>>>>>>>>>>>>>>>>>> <mbcsv_process_close_request>, >>>>>>>>>>>>>>>> 0x7f97af3741e0 >>>>>>>>>>>>>>>>>>> <mbcsv_process_chg_role_request>, 0x7f97af3744c0 >>>>>>>>>>>>>>>>>>> <mbcsv_process_snd_ckpt_request>, >>>>>>>>>>>>>>>>>>> 0x7f97af3747d0 >>>>>>> <mbcsv_process_snd_ntfy_request>, >>>>>>>>>>>>>>>>>>> 0x7f97af374970 <mbcsv_process_snd_data_req>, >>>>>>>>>> 0x7f97af373930 >>>>>>>>>>>>>>>>>>> <mbcsv_process_get_request>, >>>>>>>>>>>>>>>>>>> 0x7f97af374bd0 >>>> <mbcsv_process_set_request>} >>>>>>>>>>>>>>>>>>> #6 0x00007f97af379370 in mbcsv_process_events () >>> from >>>>>>>>>>>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0 >>>>>>>>>>>>>>>>>>> No symbol table info available. >>>>>>>>>>>>>>>>>>> #7 0x00007f97af3794db in mbcsv_hdl_dispatch_all () >>> from >>>>>>>>>>>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0 >>>>>>>>>>>>>>>>>>> No symbol table info available. >>>>>>>>>>>>>>>>>>> #8 0x00007f97af373ce2 in >>>>> mbcsv_process_dispatch_request () >>>>>>>>>> at >>>>>>>>>>>>>>>>>>> src/mbc/mbcsv_api.c:423 >>>>>>>>>>>>>>>>>>> mbcsv_init_process_req_func = {0x7f97af373630 >>>>>>>>>>>>>>>>>>> <mbcsv_process_initialize_request>, 0x7f97af374f10 >>>>>>>>>>>>>>>>>>> <mbcsv_process_get_sel_obj_request>, >>>>>>>>>>>>>>>>>>> 0x7f97af373be0 >>>>>>> <mbcsv_process_dispatch_request>, >>>>>>>>>>>>>>>>>>> 0x7f97af373af0 <mbcsv_process_finalize_request>, >>>>>>>>>>>> 0x7f97af373cf0 >>>>>>>>>>>>>>>>>>> <mbcsv_process_open_request>, >>>>>>>>>>>>>>>>>>> 0x7f97af374050 >>>>>>>>>>>>>>>>>>> <mbcsv_process_close_request>, >>>>>>>>>>>>>>>> 0x7f97af3741e0 >>>>>>>>>>>>>>>>>>> <mbcsv_process_chg_role_request>, 0x7f97af3744c0 >>>>>>>>>>>>>>>>>>> <mbcsv_process_snd_ckpt_request>, >>>>>>>>>>>>>>>>>>> 0x7f97af3747d0 >>>>>>> <mbcsv_process_snd_ntfy_request>, >>>>>>>>>>>>>>>>>>> 0x7f97af374970 <mbcsv_process_snd_data_req>, >>>>>>>>>> 0x7f97af373930 >>>>>>>>>>>>>>>>>>> <mbcsv_process_get_request>, >>>>>>>>>>>>>>>>>>> 0x7f97af374bd0 >>>> <mbcsv_process_set_request>} >>>>>>>>>>>>>>>>>>> #9 0x00007f97b027096e in >>> lgs_mbcsv_dispatch(unsigned >>>>> int) >>>>>>>> () >>>>>>>>>> at >>>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc:327 >>>>>>>>>>>>>>>>>>> ckpt_data_handler = {0x7f97b0270300 >>>>>>>>>>>>>>>>>>> <ckpt_proc_initialize_client(lgs_cb*, void*)>, >>>>>>>>>>>>>>>>>>> 0x7f97b02701c0 >>>>>>>>>>>>>>>>>>> <ckpt_proc_finalize_client(lgs_cb*, void*)>, >>>>>>>>>>>>>>>>>>> 0x7f97b0270060 >>>>>>>>>>>>>>>>>>> <ckpt_proc_agent_down(lgs_cb*, >>>>>>>>>>> void*)>, >>>>>>>>>>>>>>>>>>> 0x7f97b02712f0 <ckpt_proc_log_write(lgs_cb*, >> void*)>, >>>>>>>>>>>>>> 0x7f97b0271ab0 >>>>>>>>>>>>>>>>>>> <ckpt_proc_open_stream(lgs_cb*, void*)>, >>>>>>>>>>>>>>>>>>> 0x7f97b026fe80 >>>>>>> <ckpt_proc_close_stream(lgs_cb*, >>>>>>>>>>>>> void*)>, >>>>>>>>>>>>>>>>>>> 0x7f97b0272380 <ckpt_proc_cfg_stream(lgs_cb*, >>> void*)>, >>>>>>>>>>>>>>>>> 0x7f97b0274800 >>>>>>>>>>>>>>>>>>> <ckpt_proc_lgs_cfg_v2(lgs_cb*, void*)>, >>>>>>>>>>>>>>>>>>> 0x7f97b0274e10 >>>>>>>>>>>>>>>>>>> <ckpt_proc_lgs_cfg_v3(lgs_cb*, >>>>>>>>>>> void*)>, >>>>>>>>>>>>>>>>>>> 0x7f97b02754f0 <ckpt_proc_lgs_cfg_v5(lgs_cb*, >>> void*)>} >>>>>>>>>>>>>>>>>>> ---Type <return> to continue, or q <return> to quit--- >>>>>>>>>>>>>>>>>>> #10 0x00007f97b024d9f2 in main () at >>>>>>>>>> src/log/logd/lgs_main.cc:583 >>>>>>>>>>>>>>>>>>> usr1_sel_obj = {raise_obj = -1, rmv_obj = -1} >>>>>>>>>>>>>>>>>>> _lgs_cb = {mds_hdl = 65547, mds_role = >>>>>>>>>>>> V_DEST_RL_STANDBY, >>>>>>>>>>>>>>>>> vaddr >>>>>>>>>>>>>>>>>>> = 11, log_version = {releaseCode = 65 'A', >> majorVersion >>> = 2 >>>>>>>>>>> '\002', >>>>>>>>>>>>>>>>>>> minorVersion = 2 '\002'}, client_tree = { >>>>>>>>>>>>>>>>>>> root_node = {bit = -1, left = >>>>>>> 0x7f97b04cf1b0, >>>>>>>>>>> right = >>>>>>>>>>>>>>>>>>> 0x7f97b04a2418, key_info = 0x7f97b04b7bd0 ""}, >>> params = >>>>>>>>>>>> {key_size >>>>>>>>>>>>>>>>>>> = 4}, >>>>>>>>>>>>>>>>>>> n_nodes = 8}, comp_name = {_opaque = {46, 24947, >>>>>>>>>>>>>>>>>>> 17254, 28015, 15728, 20300, > 11335, >>>>>>>>>>>>>>>>>>> 24947, >>>>>>>>> 21350, >>>>>>>>>>>>>>>>>>> 15733, >>>>>>>>>>>>>>>>>>> 17235, 12845, 29484, 26209, 26451, 12861, 11342, >>> 24947, >>>>>>>>>> 16742, >>>>>>>>>>>>>> 28784, >>>>>>>>>>>>>>>>>>> 20285, 25968, 21358, 17985, >>>>>>>>>>>>>>>>>>> 0 <repeats 105 times>}}, amf_hdl > = >>>>>>>>> 4288675841, >>>>>>>>>>>>>>>>>>> amfSelectionObject = 15, amf_invocation_id = 0, >>>>>>>> is_quiesced_set >>>>>>>>>> = >>>>>>>>>>>>>>>> false, >>>>>>>>>>>>>>>>>>> immOiHandle = 554050912783, immSelectionObject = >>> 21, >>>>>>>>>>>>>>>>>>> clmSelectionObject = 17, clm_hdl = >>>>>>>>>>>>>>>>>>> 4279238657, >>>>>>>>>>>>>>>>>>> ha_state = >>>>>>>>>>>>>>>>>>> SA_AMF_HA_STANDBY, last_client_id = 208, >>>>> async_upd_cnt = >>>>>>>>>> 743, >>>>>>>>>>>>>>>>>>> ckpt_state >>>>>>>>>>>>>>>>>>> = COLD_SYNC_IDLE, mbcsv_hdl = 4293918753, >>>>>>>>>>>>>>>>>>> mbcsv_sel_obj = 23, mbcsv_ckpt_hdl = >>>>>>> 4292870177, >>>>>>>>>>>>>>>>>>> mbcsv_peer_version = 7, edu_hdl = {is_inited = true, >> tree >>> = >>>>>>>>>>>>>>>>>>> {root_node >>>>>>>>>>>>>>>> = >>>>>>>>>>>>>>>>>>> {bit = -1, left = 0x7f97b04cf2e0, >>>>>>>>>>>>>>>>>>> right = 0x7f97b04a25b8, > key_info = >>>>>>>>>>> 0x7f97b04b7d40 >>>>>>>>>>>>>>>> ""}, >>>>>>>>>>>>>>>>>>> params = {key_size = 8}, n_nodes = 12}, to_version = >> 1}, >>>>>>>>>>>>>>>>>>> fully_initialized = true, lga_down_list_head = 0x0, >>>>>>>>>>>>>>>>>>> lga_down_list_tail = 0x0, >>>>>>>>>>>>>>>>>>> clm_init_sel_obj = >>>>>>>>>>>>>>>>>>> {raise_obj = >>>>>>>>>>>>>>>> -1, >>>>>>>>>>>>>>>>>>> rmv_obj = -1}, nid_started = true, scAbsenceAllowed = >>> 900, >>>>>>>>>>>>>>>>>>> lgs_recovery_state = LGS_NORMAL} >>>>>>>>>>>>>>>>>>> nfds = 7 >>>>>>>>>>>>>>>>>>> fds = {{fd = 19, events = 1, revents = > 0}, >>>>>>>>>>>>>>>>>>> {fd = >>>>>>>>> 15, >>>>>>>>>>>>>>>>>>> events = >>>>>>>>>>>>>>>>>>> 1, revents = 0}, {fd = 23, events = 1, revents = 1}, > {fd >>>>>>>>>>>>>>>>>>> = 13, >>>>>>>>>>>>>>>>>>> events = >>>>>>>>>>>>>>>>>>> 1, revents = 0}, {fd = -1, events = 1, >>>>>>>>>>>>>>>>>>> revents = 0}, {fd = 17, events = 1, >>>>>>>>>>>>>>>>>>> revents >>>>>>> = >>>>>>>>> 0}, >>>>>>>>>>>>>>>>>>> {fd = >>>>>>>>>>>>>>>> 21, >>>>>>>>>>>>>>>>>>> events = 1, revents = 0}} >>>>>>>>>>>>>>>>>>> mbox_msgs = {0, 0, 0, 0, 0} >>>>>>>>>>>>>>>>>>> lgs_cb = 0x7f97b04a2400 >>>>>>>>>>>>>>>>>>> mbox_low = {0, 0, 0, 0, 0} >>>>>>>>>>>>>>>>>>> lgs_mbox_init_mutex = {__data = {__lock > = 0, >>>>>>>>> __count = >>>>>>>>>>> 0, >>>>>>>>>>>>>>>>>>> __owner = 0, __nusers = 0, __kind = 0, __spins = 0, >> __list >>>> = >>>>>>>>>>>>>>>>>>> {__prev = >>>>>>>>>>>>>>>>>>> 0x0, __next = 0x0}}, >>>>>>>>>>>>>>>>>>> __size = '\000' <repeats 39 times>, >>>>>>>>>>>>>>>>>>> __align = >>>>>>> 0} >>>>>>>>>>>>>>>>>>> lgs_mbx = 4291821569 >>>>>>>>>>>>>>>>>>> mbox_high = {0, 0, 0, 0, 0} >>>>>>>>>>>>>>>>>>> mbox_full = {false, false, false, > false, >>>>>>>>>>>>>>>>>>> false} >>>>>>>>>>>>>>>>>>> (gdb) >>>>>>>>>>>>>>>>>>> (gdb) >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> Feb 22 13:40:55 SC-2 osafimmnd[4584]: NO Ccb 131 >>>>>>>>>> COMMITTED >>>>>>>>>>>>>>>>>>> (immcfg_SC-1_18714) >>>>>>>>>>>>>>>>>>> Feb 22 13:40:56 SC-2 osafamfnd[4634]: NO >>>>>>>>>>>>>>>>>>> 'safComp=LOG,safSu=SC-2,safSg=2N,safApp=OpenSAF' >>>>> faulted >>>>>>>>>> due >>>>>>>>>>>> to >>>>>>>>>>>>>>>>>>> 'avaDown' : Recovery is 'nodeFailfast' >>>>>>>>>>>>>>>>>>> Feb 22 13:40:56 SC-2 osafamfnd[4634]: ER >>>>>>>>>>>>>>>>>>> safComp=LOG,safSu=SC-2,safSg=2N,safApp=OpenSAF >>>>> Faulted >>>>>>>>>> due >>>>>>>>>>>>>>>>>>> to:avaDown >>>>>>>>>>>>>>>>>>> Recovery is:nodeFailfast >>>>>>>>>>>>>>>>>>> Feb 22 13:40:56 SC-2 osafamfnd[4634]: Rebooting >>>>> OpenSAF >>>>>>>>>> NodeId >>>>>>>>>>>> = >>>>>>>>>>>>>>>>>>> 131599 >>>>>>>>>>>>>>>>>>> EE Name = , Reason: Component faulted: recovery is >>> node >>>>>>>>>> failfast, >>>>>>>>>>>>>>>>>>> OwnNodeId = 131599, SupervisionTime = 60 >>>>>>>>>>>>>>>>>>> Feb 22 13:40:56 SC-2 opensaf_reboot: Rebooting local >>>>> node; >>>>>>>>>>>>>> timeout=60 >> ============================================================== >> ============================================================== >>>>>>>>>>>>>>>>>>> On 2/22/2017 12:23 PM, A V Mahesh wrote: >>>>>>>>>>>>>>>>>>>> Hi Vu, >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> On 2/22/2017 12:19 PM, Vu Minh Nguyen wrote: >>>>>>>>>>>>>>>>>>>>> [Vu] I has sent you 02 patches. There is code change >>> in >>>>>>>>>> osaflogd >>>>>>>>>>>>>>>>>>>>> code >>>>>>>>>>>>>>>>>> that >>>>>>>>>>>>>>>>>>>>> fix the coredump you have observed. >>>>>>>>>>>>>>>>>>>>> The other one is test code that fix the logtest >>> coredump. >>>>>>>>>>>>>>>>>>>> Ok I will re-test , and update you . >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> -AVM >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> On 2/22/2017 12:19 PM, Vu Minh Nguyen wrote: >>>>>>>>>>>>>>>>>>>>> Hi Mahehs, >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> See my reply inline, [Vu]. >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> Regards, Vu >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>> -----Original Message----- >>>>>>>>>>>>>>>>>>>>>> From: A V Mahesh >>> [mailto:mahesh.va...@oracle.com] >>>>>>>>>>>>>>>>>>>>>> Sent: Wednesday, February 22, 2017 1:36 PM >>>>>>>>>>>>>>>>>>>>>> To: Vu Minh Nguyen >>> <vu.m.ngu...@dektech.com.au>; >>>>>>>>>>>>>>>>>>>>>> lennart.l...@ericsson.com; >>>>>>>> canh.v.tru...@dektech.com.au >>>>>>>>>>>>>>>>>>>>>> Cc: opensaf-devel@lists.sourceforge.net >>>>>>>>>>>>>>>>>>>>>> Subject: Re: [PATCH 0 of 3] Review Request for log: >>> add >>>>>>>>>>>>>>>>>>>>>> alternative >>>>>>>>>>>>>>>>>>>>>> destinations of log records [#2258] V4 >>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>> Hi Vu, >>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>> On 2/22/2017 11:52 AM, Vu Minh Nguyen wrote: >>>>>>>>>>>>>>>>>>>>>>> Hi Mahesh, >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> Have a code fault in uml test, and other one in >>>>>>>> checkpoint. >>>>>>>>>>>>>>>>>>>>>> [AVM] This is Normal Suse 11 VM ( not UML). >>>>>>>>>>>>>>>>>>>>>>> I have just updated the code. Please re-apply for >> #3 >>>>> and >>>>>>>> #4 >>>>>>>>>>>>>>>> patches. >>>>>>>>>>>>>>>>>>>>>> [AVM] is these new patch has function changes or >>> only >>>>>>>>>>>>>>>>>>>>>> test >>>>>>>>>>>> code >>>>>>>>>>>>>>>>>>> changes ? >>>>>>>>>>>>>>>>>>>>> [Vu] I has sent you 02 patches. There is code change >>> in >>>>>>>>>> osaflogd >>>>>>>>>>>>>>>>>>>>> code >>>>>>>>>>>>>>>>>> that >>>>>>>>>>>>>>>>>>>>> fix the coredump you have observed. >>>>>>>>>>>>>>>>>>>>> The other one is test code that fix the logtest >>> coredump. >>>>>>>>>>>>>>>>>>>>>>> Note that, test case #14 of suite 17 should be run >>> on >>>>>>>> active >>>>>>>>>>>>>>>>>>>>>>> node, >>>>>>>>>>>>>>>>>>>>>> otherwise >>>>>>>>>>>>>>>>>>>>>>> getting failed. >>>>>>>>>>>>>>>>>>>>>> [AVM] Segmentation fault of /usr/bin/logtest Not >> a >>> big >>>>>>>>> issue >>>>>>>>>>> , >>>>>>>>>>>>>>>>>>>>>> we need to debug why osaflogd core >> dumped >>>>>>>>>>>>>>>>>>>>>> and it >>>>>>> is >>>>>>>>>>>>>>>>>>>>>> critical >>>>>>>>>>>>>>>>>>>>> [Vu] I found the problem. You can try with the new >>> one >>>>> to >>>>>>>> see >>>>>>>>>> if >>>>>>>>>>>>>>>>>>>>> the >>>>>>>>>>>>>>>>>>>>> coredump is still there or not. >>>>>>>>>>>>>>>>>>>>>>> I will put condition check to that test case later. >>>>>>>>>>>>>>>>>>>>>> -AVM >>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> Regards, Vu >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>> -----Original Message----- >>>>>>>>>>>>>>>>>>>>>>>> From: A V Mahesh >>>>> [mailto:mahesh.va...@oracle.com] >>>>>>>>>>>>>>>>>>>>>>>> Sent: Wednesday, February 22, 2017 12:16 PM >>>>>>>>>>>>>>>>>>>>>>>> To: Vu Minh Nguyen >>>>> <vu.m.ngu...@dektech.com.au>; >>>>>>>>>>>>>>>>>>>>>>>> lennart.l...@ericsson.com; >>>>>>>>>> canh.v.tru...@dektech.com.au >>>>>>>>>>>>>>>>>>>>>>>> Cc: opensaf-devel@lists.sourceforge.net >>>>>>>>>>>>>>>>>>>>>>>> Subject: Re: [PATCH 0 of 3] Review Request for >>> log: >>>>> add >>>>>>>>>>>>>>>> alternative >>>>>>>>>>>>>>>>>>>>>>>> destinations of log records [#2258] V4 >>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>> Hi Vu, >>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>> Thanks , >>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>> While testing /usr/bin/logtest , SC-2 standby >>>>> osaflogd >>>>>>>>>> core >>>>>>>>>>>>>>>>> dumped >>>>>>>>>>>>>>>>>>>>> and >>>>>>>>>>>>>>>>>>>>>>>> /usr/bin/logtest on SC-1 Active >>>>>>>>>>>>>>>>>>>>>>>> got Segmentation fault , am I missing any other >>>>> patch ( >>>>>>> i >>>>>>>>>> am >>>>>>>>>>>>>>>> using >>>>>>>>>>>>>>>>>>>>>>>> devel published patch only ) >>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>> Following patches i am using : >>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>> 1) #2293 (sent by Anders Widel, but > not >> yet >>>>>>>>> pushed) >>>>>>>>>>>>>>>>>>>>>>>> 2) #2258 (v2, sent by Lennart, but not yet >>>>>>>>> pushed >>>>>>>>>>> yet) >>>>>>>>>>>>>>>>>>>>>>>> 3) #2258 (v4, sent by Vu, but not yet pushed >>>>>>>>> yet) >> ============================================================== >> ======================================== >>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>> Core was generated by >>>>> `/usr/lib64/opensaf/osaflogd'. >>>>>>>>>>>>>>>>>>>>>>>> Program terminated with signal 11, >> Segmentation >>>>> fault. >>>>>>>>>>>>>>>>>>>>>>>> #0 ckpt_proc_cfg_stream(lgs_cb*, void*) () at >>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc:2195 >>>>>>>>>>>>>>>>>>>>>>>> 2195 src/log/logd/lgs_mbcsv.cc: No such file or >>>>>>>>>> directory. >>>>>>>>>>>>>>>>>>>>>>>> in src/log/logd/lgs_mbcsv.cc >>>>>>>>>>>>>>>>>>>>>>>> (gdb) bt >>>>>>>>>>>>>>>>>>>>>>>> #0 ckpt_proc_cfg_stream(lgs_cb*, void*) () at >>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc:2195 >>>>>>>>>>>>>>>>>>>>>>>> #1 0x00007f12c3e22960 in >>>>>>>>>> ckpt_decode_log_struct(lgs_cb*, >>>>>>>>>>>>>>>>>>>>>>>> ncs_mbcsv_cb_arg*, void*, void*, unsigned int >>>>>>>>>>>>>> (*)(edu_hdl_tag*, >>>>>>>>>>>>>>>>>>>>>>>> edu_tkn_tag*, void*, unsigned int*, >>>>> edu_buf_env_tag*, >>>>>>>>>>>>>>>>>>> EDP_OP_TYPE, >>>>>>>>>>>>>>>>>>>>>>>> EDU_ERR*)) () at src/log/logd/lgs_mbcsv.cc:950 >>>>>>>>>>>>>>>>>>>>>>>> #2 0x00007f12c3e240dc in >>>>>>>>>>>>>> ckpt_decode_async_update(lgs_cb*, >>>>>>>>>>>>>>>>>>>>>>>> ncs_mbcsv_cb_arg*) () at >>>>>>>> src/log/logd/lgs_mbcsv.cc:1086 >>>>>>>>>>>>>>>>>>>>>>>> #3 0x00007f12c3e26941 in >>>>>>>>>>>>>> mbcsv_callback(ncs_mbcsv_cb_arg*) () >>>>>>>>>>>>>>>>> at >>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc:880 >>>>>>>>>>>>>>>>>>>>>>>> #4 0x00007f12c2f25596 in >>> ncs_mbscv_rcv_decode () >>>>>>>> from >>>>>>>>>>>>>>>>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0 >>>>>>>>>>>>>>>>>>>>>>>> #5 0x00007f12c2f25766 in >>>>>>>> ncs_mbcsv_rcv_async_update >>>>>>>>>> () >>>>>>>>>>>>>> from >>>>>>>>>>>>>>>>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0 >>>>>>>>>>>>>>>>>>>>>>>> #6 0x00007f12c2f2c370 in >> mbcsv_process_events >>> () >>>>>>>> from >>>>>>>>>>>>>>>>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0 >>>>>>>>>>>>>>>>>>>>>>>> #7 0x00007f12c2f2c4db in >>> mbcsv_hdl_dispatch_all () >>>>>>>> from >>>>>>>>>>>>>>>>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0 >>>>>>>>>>>>>>>>>>>>>>>> #8 0x00007f12c2f26ce2 in >>>>>>>>>> mbcsv_process_dispatch_request >>>>>>>>>>>> () >>>>>>>>>>>>>> at >>>>>>>>>>>>>>>>>>>>>>>> src/mbc/mbcsv_api.c:423 >>>>>>>>>>>>>>>>>>>>>>>> #9 0x00007f12c3e2396e in >>>>>>>> lgs_mbcsv_dispatch(unsigned >>>>>>>>>> int) >>>>>>>>>>>> () >>>>>>>>>>>>>> at >>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc:327 >>>>>>>>>>>>>>>>>>>>>>>> #10 0x00007f12c3e009f2 in main () at >>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_main.cc:583 >>>>>>>>>>>>>>>>>>>>>>>> (gdb) >>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>> >> ============================================================== >> ======================================== >>>>>>>>>>>>>>>>>>>>>>>> Feb 22 10:37:06 SC-1 osafimmnd[4020]: NO >>> Invalid >>>>> error >>>>>>>>>>>>>> reported >>>>>>>>>>>>>>>>>>>>>>>> implementer 'safLogService', Ccb 161 will be >>> aborted >>>>>>>>>>>>>>>>>>>>>>>> Feb 22 10:37:06 SC-1 osafimmnd[4020]: NO Ccb >>> 161 >>>>>>>>>> aborted >>>>>>>>>>>> in >>>>>>>>>>>>>>>>>>>>>> COMPLETED >>>>>>>>>>>>>>>>>>>>>>>> processing (validation) >>>>>>>>>>>>>>>>>>>>>>>> Feb 22 10:37:06 SC-1 osafimmnd[4020]: NO Ccb >>> 161 >>>>>>>>>>>> ABORTED >>>>>>>>>>>>>>>>>>>>>> (immcfg_SC- >>>>>>>>>>>>>>>>>>>>>>>> 1_5394) >>>>>>>>>>>>>>>>>>>>>>>> Add values Fail >>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>> Program received signal SIGSEGV, Segmentation >>>>> fault. >>>>>>>>>>>>>>>>>>>>>>>> 0x000055555556929a in >> read_and_compare.isra.7 >>> () >>>>> at >>>>>>>>>>>>>>>>>>>>>>>> src/log/apitest/tet_LogOiOps.c:1891 >>>>>>>>>>>>>>>>>>>>>>>> 1891 src/log/apitest/tet_LogOiOps.c: No such >> file >>> or >>>>>>>>>>>>>>>>>>>>>>>> directory. >>>>>>>>>>>>>>>>>>>>>>>> in > src/log/apitest/tet_LogOiOps.c >>>>>>>>>>>>>>>>>>>>>>>> (gdb) Feb 22 10:37:07 SC-1 sshd[5298]: Accepted >>>>>>>> keyboard- >>>>>>>>>>>>>>>>>>>>>> interactive/pam >>>>>>>>>>>>>>>>>>>>>>>> for root from 10.176.178.22 port 51945 ssh2 >>>>>>>>>>>>>>>>>>>>>>>> bt >>>>>>>>>>>>>>>>>>>>>>>> #0 0x000055555556929a in >>>>> read_and_compare.isra.7 () >>>>>>>> at >>>>>>>>>>>>>>>>>>>>>>>> src/log/apitest/tet_LogOiOps.c:1891 >>>>>>>>>>>>>>>>>>>>>>>> #1 0x0000555555569bbb in >>>>>>>>>>>>>>>>>>>>>>>> check_logRecordDestinationConfigurationEmpty >>>>>>>>>>>>>>>>>>>>>>>> () at src/log/apitest/tet_LogOiOps.c:2179 >>>>>>>>>>>>>>>>>>>>>>>> #2 0x0000555555573495 in run_test_case () >>>>>>>>>>>>>>>>>>>>>>>> #3 0x0000555555573934 in test_run () >>>>>>>>>>>>>>>>>>>>>>>> #4 0x000055555555c7cd in main () at >>>>>>>>>>>>>>>>>>>>>>>> src/log/apitest/logtest.c:569 >>>>>>>>>>>>>>>>>>>>>>>> (gdb) >>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>> >> ============================================================== >> ======================================== >>>>>>>>>>>>>>>>>>>>>>>> -AVM >>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>> On 2/22/2017 9:48 AM, Vu Minh Nguyen wrote: >>>>>>>>>>>>>>>>>>>>>>>>> Hi Mahesh, >>>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>>> I send them in attachment instead, and name >>> them >>>>> in >>>>>>>> the >>>>>>>>>>>>>> order. >>>>>>>>>>>>>>>>>>>>>>>>> I just pull the latest code, and apply them >> without >>>>>>>>> getting >>>>>>>>>>>>> any >>>>>>>>>>>>>>>>>> hunk >>>>>>>>>>>>>>>>>>>>>>> error. >>>>>>>>>>>>>>>>>>>>>>>>> Please try with them, and let me know if you >> see >>>>> any >>>>>>>>>>>> problem. >>>>>>>>>>>>>>>>>>>>>>>>> Regards, Vu >>>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>>>> -----Original Message----- >>>>>>>>>>>>>>>>>>>>>>>>>> From: A V Mahesh >>>>> [mailto:mahesh.va...@oracle.com] >>>>>>>>>>>>>>>>>>>>>>>>>> Sent: Wednesday, February 22, 2017 11:09 >> AM >>>>>>>>>>>>>>>>>>>>>>>>>> To: Vu Minh Nguyen >>>>>>>> <vu.m.ngu...@dektech.com.au>; >>>>>>>>>>>>>>>>>>>>>>>>>> lennart.l...@ericsson.com; >>>>>>>>>>>> canh.v.tru...@dektech.com.au >>>>>>>>>>>>>>>>>>>>>>>>>> Cc: opensaf-devel@lists.sourceforge.net >>>>>>>>>>>>>>>>>>>>>>>>>> Subject: Re: [PATCH 0 of 3] Review Request >> for >>>>> log: >>>>>>>> add >>>>>>>>>>>>>>>>>> alternative >>>>>>>>>>>>>>>>>>>>>>>>>> destinations of log records [#2258] V4 >>>>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>>>> Hi Vu, >>>>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>>>> I did follow that still i get Hunk #2 FAILED > even >>> on >>>>>>>>>>> today's >>>>>>>>>>>>>>>>>> staging >> ============================================================== >>>>>>>>>>>>>>>>>>>>>>>>>> ================== >>>>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>>>> [root@dhcp-hyd-scp-5fl-10-176-177-96 >>> staging]# >>>>>>>> patch >>>>>>>>>> - >>>>>>>>>>>> p1 >>>>>>>>>>>>>>>>> <2293 >>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/base/Makefile.am >>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #1 succeeded at 33 (offset 1 line). >>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #3 succeeded at 183 (offset 1 line). >>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/base/file_descriptor.cc >>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/base/file_descriptor.h >>>>>>>>>>>>>>>>>>>>>>>>>> patching file >> src/base/tests/unix_socket_test.cc >>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/base/unix_client_socket.cc >>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/base/unix_server_socket.cc >>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/base/unix_socket.cc >>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/base/unix_socket.h >>>>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>>>> [root@dhcp-hyd-scp-5fl-10-176-177-96 >>> staging]# >>>>>>>> patch >>>>>>>>>> - >>>>>>>>>>>> p1 >>>>>>>>>>>>>>>>> <2258- >>>>>>>>>>>>>>>>>>> 1 >>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/log/Makefile.am >>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #1 succeeded at 71 (offset -1 lines). >>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/log/config/logsv_classes.xml >>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #1 FAILED at 147. >>>>>>>>>>>>>>>>>>>>>>>>>> 1 out of 1 hunk FAILED -- saving rejects to file >>>>>>>>>>>>>>>>>>>>>>>>>> src/log/config/logsv_classes.xml.rej >>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_config.cc >>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #1 succeeded at 35 (offset -5 lines). >>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #2 FAILED at 705. >>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #3 FAILED at 971. >>>>>>>>>>>>>>>>>>>>>>>>>> 2 out of 3 hunks FAILED -- saving rejects to > file >>>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_config.cc.rej >>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_config.h >>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #1 FAILED at 304. >>>>>>>>>>>>>>>>>>>>>>>>>> 1 out of 1 hunk FAILED -- saving rejects to file >>>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_config.h.rej >>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_dest.cc >>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_dest.h >>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_evt.cc >>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_imm.cc >>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #1 FAILED at 45. >>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #2 succeeded at 235 (offset -1 lines). >>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #3 FAILED at 877. >>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #4 succeeded at 1273 (offset -20 lines). >>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #5 succeeded at 1404 (offset -1 lines). >>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #6 succeeded at 1449 (offset -20 lines). >>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #7 succeeded at 2032 (offset -1 lines). >>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #8 FAILED at 2181. >>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #9 succeeded at 2271 (offset -54 lines). >>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #10 succeeded at 2387 (offset -1 lines). >>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #11 succeeded at 2377 (offset -54 lines). >>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #12 succeeded at 2478 (offset -1 lines). >>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #13 succeeded at 2684 (offset -54 lines). >>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #14 succeeded at 2821 (offset -1 lines). >>>>>>>>>>>>>>>>>>>>>>>>>> 3 out of 14 hunks FAILED -- saving rejects to >> file >>>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_imm.cc.rej >>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_main.cc >>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_mbcsv.cc >>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_mbcsv.h >>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_mbcsv_v5.cc >>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #3 succeeded at 133 (offset -1 lines). >>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_mbcsv_v7.cc >>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_mbcsv_v7.h >>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_stream.cc >>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_stream.h >>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_util.cc >>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_util.h >>>>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>>>> [root@dhcp-hyd-scp-5fl-10-176-177-96 >>> staging]# >>>>>>>> patch >>>>>>>>>> - >>>>>>>>>>>> p1 >>>>>>>>>>>>>>>>> <2258- >>>>>>>>>>>>>>>>>>> 2 >>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/log/Makefile.am >>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #1 succeeded at 180 (offset -3 lines). >>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/log/apitest/tet_LogOiOps.c >>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #1 FAILED at 1923. >>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #2 FAILED at 1979. >>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #3 FAILED at 2067. >>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #4 FAILED at 2094. >>>>>>>>>>>>>>>>>>>>>>>>>> 4 out of 4 hunks FAILED -- saving rejects to > file >>>>>>>>>>>>>>>>>>>>>>>>>> src/log/apitest/tet_LogOiOps.c.rej >>>>>>>>>>>>>>>>>>>>>>>>>> patching file >>> src/log/apitest/tet_cfg_destination.c >>>>>>>>>>>>>>>>>>>>>>>>>> [root@dhcp-hyd-scp-5fl-10-176-177-96 >>> staging]# >>>>>>>> patch >>>>>>>>>> - >>>>>>>>>>>> p1 >>>>>>>>>>>>>>>>> <2258- >>>>>>>>>>>>>>>>>>> 3 >>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/log/Makefile >>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/log/Makefile.am >>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #1 succeeded at 80 (offset -1 lines). >>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #2 succeeded at 217 (offset -2 lines). >>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/log/tests/Makefile >>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/log/tests/lgs_dest_test.cc >>>>>>>>>>>>>>>>>>>>>>>>>> [root@dhcp-hyd-scp-5fl-10-176-177-96 >>> staging]# >>>>> vi >>>>>>>>>>>>>>>>>>>>>>>>>> src/log/apitest/tet_LogOiOps.c.rej >>>>>>>>>>>>>>>>>>>>>>>>>> [root@dhcp-hyd-scp-5fl-10-176-177-96 >>> staging]# >>>>>>>>>>>>>>>>>>>>>>>>>> >> ============================================================== >>>>>>>>>>>>>>>>>>>>>>>>>> ======================== >>>>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>>>> -AVM >>>>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>>>> On 2/21/2017 3:53 PM, Vu Minh Nguyen >> wrote: >>>>>>>>>>>>>>>>>>>>>>>>>>> Hi Mahesh, >>>>>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>>>>> As I has mentioned in below: >>>>>>>>>>>>>>>>>>>>>>>>>>>>> To run the test, this patch has >>>>>>> dependent >>>>>>>>> on >>>>>>>>>>>>>>>> following >>>>>>>>>>>>>>>>>>>>> patches: >>>>>>>>>>>>>>>>>>>>>>>>>>>>> 1) #2293 (sent by Anders Widel, but not >>>>>>> yet >>>>>>>>>>>>>>>>>>>>>>>>>>>>> pushed) >>>>>>>>>>>>>>>>>>>>>>>>>>>>> 2) #2258 (v2, sent by Lennart, > but >>>>>>>>>>>>>>>>>>>>>>>>>>>>> not >>>>>>> yet >>>>>>>>>>> pushed >>>>>>>>>>>>>>>> yet) >>>>>>>>>>>>>>>>>>>>>>>>>>> So, you need to apply #2293 first, then #2258 >>>>> which >>>>>>>>>> sent >>>>>>>>>>>> by >>>>>>>>>>>>>>>>>>> Lennart >>>>>>>>>>>>>>>>>>>>>>>>>>> yesterday, then mine. >>>>>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>>>>> Regards, Vu >>>>>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>>>>>> -----Original Message----- >>>>>>>>>>>>>>>>>>>>>>>>>>>> From: A V Mahesh >>>>>>>> [mailto:mahesh.va...@oracle.com] >>>>>>>>>>>>>>>>>>>>>>>>>>>> Sent: Tuesday, February 21, 2017 5:10 PM >>>>>>>>>>>>>>>>>>>>>>>>>>>> To: Vu Minh Nguyen >>>>>>>>>> <vu.m.ngu...@dektech.com.au>; >>>>>>>>>>>>>>>>>>>>>>>>>>>> lennart.l...@ericsson.com; >>>>>>>>>>>>>> canh.v.tru...@dektech.com.au >>>>>>>>>>>>>>>>>>>>>>>>>>>> Cc: opensaf-devel@lists.sourceforge.net >>>>>>>>>>>>>>>>>>>>>>>>>>>> Subject: Re: [PATCH 0 of 3] Review Request >>> for >>>>> log: >>>>>>>>>> add >>>>>>>>>>>>>>>>>>> alternative >>>>>>>>>>>>>>>>>>>>>>>>>>>> destinations of log records [#2258] V4 >>>>>>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>>>>>> Hi Vu, >>>>>>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>>>>>> Is this applies on top of log #2146 - V4 , > I >>>> see >>>>>>>>>>> both >>>>>>>>>>>>>>>>>> #tickets >>>>>>>>>>>>>>>>>>>>>>>>> has >>>>>>>>>>>>>>>>>>>>>>>>>>>> version changes ? >>>>>>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>>>>>> in which order i need to apply ( #2146 & >>> #2258 ) >>>>>>>> or >>>>>>>>>>>>>>>>>>>>>>>>>>>> (#2258 >>>>>>>>>>>>>>>> & >>>>>>>>>>>>>>>>>>>>>>>>> #2146). >> ========================================================= >>>>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/log/Makefile.am >>>>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #1 FAILED at 72. >>>>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #2 FAILED at 120. >>>>>>>>>>>>>>>>>>>>>>>>>>>> 2 out of 2 hunks FAILED -- saving rejects to >>> file >>>>>>>>>>>>>>>>>>>>>>>>> src/log/Makefile.am.rej >>>>>>>>>>>>>>>>>>>>>>>>>>>> patching file >> src/log/config/logsv_classes.xml >>>>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #1 FAILED at 147. >>>>>>>>>>>>>>>>>>>>>>>>>>>> 1 out of 1 hunk FAILED -- saving rejects to >> file >>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/config/logsv_classes.xml.rej >>>>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_config.cc >>>>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #1 succeeded at 35 (offset -5 lines). >>>>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #2 FAILED at 705. >>>>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #3 FAILED at 971. >>>>>>>>>>>>>>>>>>>>>>>>>>>> 2 out of 3 hunks FAILED -- saving rejects to >>> file >>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_config.cc.rej >>>>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_config.h >>>>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #1 FAILED at 304. >>>>>>>>>>>>>>>>>>>>>>>>>>>> 1 out of 1 hunk FAILED -- saving rejects to >> file >>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_config.h.rej >>>>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_dest.cc >>>>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_dest.h >>>>>>>>>>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_evt.cc >>>>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #1 FAILED at 1. >>>>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #2 succeeded at 30 with fuzz 2 (offset >> 2 >>>>>>>>>>>>>>>>>>>>>>>>>>>> lines). >>>>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #3 succeeded at 1282 (offset 45 lines). >>>>>>>>>>>>>>>>>>>>>>>>>>>> Hunk #4 succeeded at 1300 (offset 2 lines). >>>>>>>>>>>>>>>>>>>>>>>>>>>> 1 out of 4 hunks FAILED -- saving rejects to >>> file >>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_evt.cc.rej >>>>>>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>>>>>> >> ============================================================== >>>>>>>>>>>>>>>>>>>>>>>>>>>> === >>>>>>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>>>>>> -AVM >>>>>>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>>>>>> On 2/21/2017 3:03 PM, Vu Minh Nguyen >>> wrote: >>>>>>>>>>>>>>>>>>>>>>>>>>>>> Summary: log: add alternative destinations >>> of >>>>> log >>>>>>>>>>>> records >>>>>>>>>>>>>>>>>>> [#2258] >>>>>>>>>>>>>>>>>>>>>>>>>>>>> Review request for Trac Ticket(s): #2258 >>>>>>>>>>>>>>>>>>>>>>>>>>>>> Peer Reviewer(s): Lennart, Canh, Mahesh >>>>>>>>>>>>>>>>>>>>>>>>>>>>> Pull request to: <<LIST THE PERSON WITH >>>>> PUSH >>>>>>>>>>>> ACCESS >>>>>>>>>>>>>>>>> HERE>> >>>>>>>>>>>>>>>>>>>>>>>>>>>>> Affected branch(es): Default >>>>>>>>>>>>>>>>>>>>>>>>>>>>> Development branch: Default >>>>>>>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>>>>>>> -------------------------------- >>>>>>>>>>>>>>>>>>>>>>>>>>>>> Impacted area Impact y/n >>>>>>>>>>>>>>>>>>>>>>>>>>>>> -------------------------------- >>>>>>>>>>>>>>>>>>>>>>>>>>>>> Docs n >>>>>>>>>>>>>>>>>>>>>>>>>>>>> Build system n >>>>>>>>>>>>>>>>>>>>>>>>>>>>> RPM/packaging n >>>>>>>>>>>>>>>>>>>>>>>>>>>>> Configuration files n >>>>>>>>>>>>>>>>>>>>>>>>>>>>> Startup scripts n >>>>>>>>>>>>>>>>>>>>>>>>>>>>> SAF services n >>>>>>>>>>>>>>>>>>>>>>>>>>>>> OpenSAF services y >>>>>>>>>>>>>>>>>>>>>>>>>>>>> Core libraries n >>>>>>>>>>>>>>>>>>>>>>>>>>>>> Samples n >>>>>>>>>>>>>>>>>>>>>>>>>>>>> Tests y >>>>>>>>>>>>>>>>>>>>>>>>>>>>> Other n >>>>>>>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>>>>>>> Comments (indicate scope for each "y" >>> above): >>>>>>>>>>>>>>>>>>>>>>>>>>>>> --------------------------------------------- >>>>>>>>>>>>>>>>>>>>>>>>>>>>> To run the test, this patch has >>>>>>> dependent >>>>>>>>> on >>>>>>>>>>>>>>>> following >>>>>>>>>>>>>>>>>>>>> patches: >>>>>>>>>>>>>>>>>>>>>>>>>>>>> 1) #2293 (sent by Anders Widel, but not >>>>>>> yet >>>>>>>>>>>>>>>>>>>>>>>>>>>>> pushed) >>>>>>>>>>>>>>>>>>>>>>>>>>>>> 2) #2258 (v2, sent by Lennart, > but >>>>>>>>>>>>>>>>>>>>>>>>>>>>> not >>>>>>> yet >>>>>>>>>>> pushed >>>>>>>>>>>>>>>> yet) >>>>>>>>>>>>>>>>>>>>>>>>>>>>> changeset >>>>>>>>>>>>>> d74aaf3025c99cade3165a15831124548f4d85bd >>>>>>>>>>>>>>>>>>>>>>>>>>>>> Author: Vu Minh Nguyen >>>>>>>>>>>>>> <vu.m.ngu...@dektech.com.au> >>>>>>>>>>>>>>>>>>>>>>>>>>>>> Date: Wed, 15 Feb 2017 14:36:00 +0700 >>>>>>>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>>>>>>> log: add alternative destinations of >> log >>>>>>> records >>>>>>>>>>>>>>>>>>>>>>>>>>>>> [#2258] >>>>>>>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>>>>>>> Here are major info, detailed info > will >>>> be >>>>>>>> added >>>>>>>>> to >>>>>>>>>>>>>>>>>>>>>>>>>>>>> PR doc >>>>>>>>>>>>>>>>>>>>>>> soon. 1) >>>>>>>>>>>>>>>>>>>>>>>>>>>> Add >>>>>>>>>>>>>>>>>>>>>>>>>>>>> attribute "saLogRecordDestination" to log >>>>>>>>>> stream. >>>>>>>>>>> 2) >>>>>>>>>>>>>>>>>>>>>>>>>>>>> Add >>>>>>>>>>>>>>>>>>>>>>> Local >>>>>>>>>>>>>>>>>>>>>>>>>>>> socket >>>>>>>>>>>>>>>>>>>>>>>>>>>>> destintion handler 3) Integrate into first >>>>>>>>>>> increment >>>>>>>>>>>>>>>>>>>>>>>>>>>>> made by >>>>>>>>>>>>>>>>>>>>>>> Lennart >>>>>>>>>>>>>>>>>>>>>>>>>>>>> changeset >>>>>>>>>>>>>> 4bae27a478c235df3058f43c92d3a5483233b01d >>>>>>>>>>>>>>>>>>>>>>>>>>>>> Author: Vu Minh Nguyen >>>>>>>>>>>>>> <vu.m.ngu...@dektech.com.au> >>>>>>>>>>>>>>>>>>>>>>>>>>>>> Date: Wed, 15 Feb 2017 15:07:09 +0700 >>>>>>>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>>>>>>> log: add UML test case to verify >>>>>>>>>>>>>>>>>>>>>>>>>>>>> alternative >>>>>>>>>>>>>>>>>>>>>>>>>>>>> destination >>>>>>>>>>>>>>>>>>>>>>> [#2258] >>>>>>>>>>>>>>>>>>>>>>>>>>>>> Major changes: 1) Modify Lennart's test >>> cases >>>>>>>>>>>> because >>>>>>>>>>>>>>>>>>>>>>> enhancing >>>>>>>>>>>>>>>>>>>>>>>>>>>> destination >>>>>>>>>>>>>>>>>>>>>>>>>>>>> configuration validation rules. 2) Add test >>>>>>>>> suite >>>>>>>>>>>>>>>>>>>>>>>>>>>>> #17 to >>>>>>>>>>>>>>>>>>>>>>> verify >>>>>>>>>>>>>>>>>>>>>>>>>>>> alternative >>>>>>>>>>>>>>>>>>>>>>>>>>>>> destination >>>>>>>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>>>>>>> changeset >>>>>>>>>>>> bc375725fed22bb4f8cb3ae3df5f96fb9d281efb >>>>>>>>>>>>>>>>>>>>>>>>>>>>> Author: Vu Minh Nguyen >>>>>>>>>>>>>> <vu.m.ngu...@dektech.com.au> >>>>>>>>>>>>>>>>>>>>>>>>>>>>> Date: Thu, 16 Feb 2017 17:22:13 +0700 >>>>>>>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>>>>>>> log: add unit tests to verify > interfaces >>>>>>>>> provided >>>>>>>>>>> by >>>>>>>>>>>>>>>>>>>>>>> destination >>>>>>>>>>>>>>>>>>>>>>>>>>>> handler >>>>>>>>>>>>>>>>>>>>>>>>>>>>> [#2258] >>>>>>>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>>>>>>> Unit tests to verify major > interfaces: >> 1) >>>>>>>>>>>>>>>>>>>>>>>>>>>>> CfgDestination() >>>>>>>>>>>>>>>>>>>>>>> 2) >>>>>>>>>>>>>>>>>>>>>>>>>>>>> WriteToDestination() >>>>>>>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>>>>>>> Added Files: >>>>>>>>>>>>>>>>>>>>>>>>>>>>> ------------ >>>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/apitest/tet_cfg_destination.c >>>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_dest.cc >>>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_dest.h >>>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv_v7.cc >>>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv_v7.h >>>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/tests/lgs_dest_test.cc >>>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/tests/Makefile >>>>>>>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>>>>>>> Complete diffstat: >>>>>>>>>>>>>>>>>>>>>>>>>>>>> ------------------ >>>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/Makefile | 4 + >>>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/Makefile.am | 31 +++++- >>>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/apitest/tet_LogOiOps.c | 8 +- >>>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/apitest/tet_cfg_destination.c | >> 483 >> ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++ >>>>>>>>>>>>>>>>>>>>>>>>>>>> ++++++++++++++++++++++++++++++++++ >>>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/config/logsv_classes.xml | 7 +- >>>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_config.cc | 169 >>>>>>>>>>>>>>>>>>>>>>>>>>>> +++++++++++++++++++++++++++++--- >>>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_config.h | 3 +- >>>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_dest.cc | 707 >> ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++ >> ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++ >>>>>>>>>>>>>>>>>>>>>>>>>>>> +++++++++++++++++ >>>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_dest.h | 576 >> ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++ >>>>> ++++++++++++++++++++++++++++++++++++++++++++++++++++ >>>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_evt.cc | 33 ++++++ >>>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_imm.cc | 202 >>>>>>>>>>>>>>>>>>>>>>>>>>>> +++++++++++++++++++++++++++++++++---- >> -- >>>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_main.cc | 8 + >>>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc | 103 >>>>>>>>>>>>>>>>>>>>> ++++++++++++++++++- >>>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv.h | 6 +- >>>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv_v5.cc | 10 + >>>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv_v7.cc | 177 >>>>>>>>>>>>>>>>>>>>>>>>>>>> +++++++++++++++++++++++++++++++++++ >>>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv_v7.h | 67 >>>>>>>>>>>>>>>>>> +++++++++++++ >>>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_stream.cc | 60 >>>>>>>>>>>>>>>>>> +++++++++++- >>>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_stream.h | 16 +++ >>>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_util.cc | 63 >>>>>>>>>>>>>>>>>> ++++++++++++ >>>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_util.h | 11 +- >>>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/tests/Makefile | 20 +++ >>>>>>>>>>>>>>>>>>>>>>>>>>>>> src/log/tests/lgs_dest_test.cc | 209 >>>>> +++++++++++++++++++++++++++++++++++++++++ >>>>>>>>>>>>>>>>>>>>>>>>>>>>> 23 files changed, 2896 insertions(+), 77 >>>>>>>>>>>>>>>>>>>>>>>>>>>>> deletions(-) >>>>>>>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>>>>>>> Testing Commands: >>>>>>>>>>>>>>>>>>>>>>>>>>>>> ----------------- >>>>>>>>>>>>>>>>>>>>>>>>>>>>> Run UML test suite #17 >>>>>>>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>>>>>>> Testing, Expected Results: >>>>>>>>>>>>>>>>>>>>>>>>>>>>> -------------------------- >>>>>>>>>>>>>>>>>>>>>>>>>>>>> All test passed >>>>>>>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>>>>>>> Conditions of Submission: >>>>>>>>>>>>>>>>>>>>>>>>>>>>> ------------------------- >>>>>>>>>>>>>>>>>>>>>>>>>>>>> <<HOW MANY DAYS BEFORE PUSHING, >>>>>>>>>>>> CONSENSUS >>>>>>>>>>>>>>>>> ETC>> >>>>>>>>>>>>>>>>>>>>>>>>>>>>> Arch Built Started Linux distro >>>>>>>>>>>>>>>>>>>>>>>>>>>>> ------------------------------------------- >>>>>>>>>>>>>>>>>>>>>>>>>>>>> mips n n >>>>>>>>>>>>>>>>>>>>>>>>>>>>> mips64 n n >>>>>>>>>>>>>>>>>>>>>>>>>>>>> x86 n n >>>>>>>>>>>>>>>>>>>>>>>>>>>>> x86_64 n n >>>>>>>>>>>>>>>>>>>>>>>>>>>>> powerpc n n >>>>>>>>>>>>>>>>>>>>>>>>>>>>> powerpc64 n n >>>>>>>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>>>>>>> Reviewer Checklist: >>>>>>>>>>>>>>>>>>>>>>>>>>>>> ------------------- >>>>>>>>>>>>>>>>>>>>>>>>>>>>> [Submitters: make sure that your review >>>>> doesn't >>>>>>>>>> trigger >>>>>>>>>>>>>> any >>>>>>>>>>>>>>>>>>>>>>>>>> checkmarks!] >>>>>>>>>>>>>>>>>>>>>>>>>>>>> Your checkin has not passed review >> because >>>>> (see >>>>>>>>>>>> checked >>>>>>>>>>>>>>>>>>> entries): >>>>>>>>>>>>>>>>>>>>>>>>>>>>> ___ Your RR template is generally >>> incomplete; >>>>> it >>>>>>>> has >>>>>>>>>>>> too >>>>>>>>>>>>>>>>> many >>>>>>>>>>>>>>>>>>>>> blank >>>>>>>>>>>>>>>>>>>>>>>>>>> entries >>>>>>>>>>>>>>>>>>>>>>>>>>>>> that need proper data filled in. >>>>>>>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>>>>>>> ___ You have failed to nominate the >> proper >>>>>>>> persons >>>>>>>>>> for >>>>>>>>>>>>>>>>> review >>>>>>>>>>>>>>>>>>> and >>>>>>>>>>>>>>>>>>>>>>>>> push. >>>>>>>>>>>>>>>>>>>>>>>>>>>>> ___ Your patches do not have proper >>>>> short+long >>>>>>>>>>>> header >>>>>>>>>>>>>>>>>>>>>>>>>>>>> ___ You have grammar/spelling in your >>> header >>>>>>>> that >>>>>>>>>> is >>>>>>>>>>>>>>>>>>> unacceptable. >>>>>>>>>>>>>>>>>>>>>>>>>>>>> ___ You have exceeded a sensible line >>> length >>>>> in >>>>>>>> your >>>>>>>>>>>>>>>>>>>>>>>>>>>> headers/comments/text. >>>>>>>>>>>>>>>>>>>>>>>>>>>>> ___ You have failed to put in a proper Trac >>>>> Ticket >>>>>>>> # >>>>>>>>>>> into >>>>>>>>>>>>>>>> your >>>>>>>>>>>>>>>>>>>>>>>>> commits. >>>>>>>>>>>>>>>>>>>>>>>>>>>>> ___ You have incorrectly put/left internal >>> data >>>>> in >>>>>>>>>> your >>>>>>>>>>>>>>>>>>>>>>> comments/files >>>>>>>>>>>>>>>>>>>>>>>>>>>>> (i.e. internal bug tracking tool IDs, product >>>>>>>>>>>>>>>> names >>>>>>>>>>>>>>>>>> etc) >>>>>>>>>>>>>>>>>>>>>>>>>>>>> ___ You have not given any evidence of >>> testing >>>>>>>>>> beyond >>>>>>>>>>>>>> basic >>>>>>>>>>>>>>>>>>> build >>>>>>>>>>>>>>>>>>>>>>>>> tests. >>>>>>>>>>>>>>>>>>>>>>>>>>>>> Demonstrate some level of runtime or >> other >>>>> sanity >>>>>>>>>>>>>>>>>> testing. >>>>>>>>>>>>>>>>>>>>>>>>>>>>> ___ You have ^M present in some of your >>> files. >>>>>>>>>> These >>>>>>>>>>>>>>>>>>>>>>>>>>>>> have to >>>>>>>>>>>>>>>>>>> be >>>>>>>>>>>>>>>>>>>>>>>>>> removed. >>>>>>>>>>>>>>>>>>>>>>>>>>>>> ___ You have needlessly changed >>> whitespace >>>>> or >>>>>>>>>> added >>>>>>>>>>>>>>>>>>> whitespace >>>>>>>>>>>>>>>>>>>>>>>>>> crimes >>>>>>>>>>>>>>>>>>>>>>>>>>>>> like trailing spaces, or spaces >>>>>>> before >>>>>>>>>>> tabs. >>>>>>>>>>>>>>>>>>>>>>>>>>>>> ___ You have mixed real technical changes >>>>> with >>>>>>>>>>>>>> whitespace >>>>>>>>>>>>>>>>> and >>>>>>>>>>>>>>>>>>>>>> other >>>>>>>>>>>>>>>>>>>>>>>>>>>>> cosmetic code cleanup changes. These >> have >>> to >>>>> be >>>>>>>>>>>>>>>>>>> separate >>>>>>>>>>>>>>>>>>>>>>>>> commits. >>>>>>>>>>>>>>>>>>>>>>>>>>>>> ___ You need to refactor your submission >>> into >>>>>>>>>> logical >>>>>>>>>>>>>>>>>>>>>>>>>>>>> chunks; >>>>>>>>>>>>>>>>>>>>> there >>>>>>>>>>>>>>>>>>>>>>> is >>>>>>>>>>>>>>>>>>>>>>>>>>>>> too much content into a single >>>>>>> commit. >>>>>>>>>>>>>>>>>>>>>>>>>>>>> ___ You have extraneous garbage in your >>>>> review >>>>>>>>>>>> (merge >>>>>>>>>>>>>>>>>>> commits >>>>>>>>>>>>>>>>>>>>>> etc) >>>>>>>>>>>>>>>>>>>>>>>>>>>>> ___ You have giant attachments which >>> should >>>>>>>> never >>>>>>>>>>>> have >>>>>>>>>>>>>>>>> been >>>>>>>>>>>>>>>>>>>>>> sent; >>>>>>>>>>>>>>>>>>>>>>>>>>>>> Instead you should place your content in a >>>>> public >>>>>>>>>>>>>>>>>> tree to >>>>>>>>>>>>>>>>>>>>> be >>>>>>>>>>>>>>>>>>>>>>>>>>> pulled. >>>>>>>>>>>>>>>>>>>>>>>>>>>>> ___ You have too many commits attached >>> to >>>>> an e- >>>>>>>>>> mail; >>>>>>>>>>>>>>>>> resend >>>>>>>>>>>>>>>>>>> as >>>>>>>>>>>>>>>>>>>>>>>>>> threaded >>>>>>>>>>>>>>>>>>>>>>>>>>>>> commits, or place in a public tree for a >> pull. >>>>>>>>>>>>>>>>>>>>>>>>>>>>> ___ You have resent this content multiple >>>>> times >>>>>>>>>>>> without a >>>>>>>>>>>>>>>>> clear >>>>>>>>>>>>>>>>>>>>>>>>>>> indication >>>>>>>>>>>>>>>>>>>>>>>>>>>>> of what has changed between each >>>>>>>>> re-send. >>>>>>>>>>>>>>>>>>>>>>>>>>>>> ___ You have failed to adequately and >>>>> individually >>>>>>>>>>>> address >>>>>>>>>>>>>>>> all >>>>>>>>>>>>>>>>>> of >>>>>>>>>>>>>>>>>>>>>>> the >>>>>>>>>>>>>>>>>>>>>>>>>>>>> comments and change requests that were >>>>>>>> proposed >>>>>>>>>> in >>>>>>>>>>>>>>>>>> the >>>>>>>>>>>>>>>>>>>>>>> initial >>>>>>>>>>>>>>>>>>>>>>>>>>> review. >>>>>>>>>>>>>>>>>>>>>>>>>>>>> ___ You have a misconfigured ~/.hgrc file >>> (i.e. >>>>>>>>>>>> username, >>>>>>>>>>>>>>>>> email >>>>>>>>>>>>>>>>>>>>> etc) >>>>>>>>>>>>>>>>>>>>>>>>>>>>> ___ Your computer have a badly >> configured >>>>> date >>>>>>>> and >>>>>>>>>>>>>> time; >>>>>>>>>>>>>>>>>>>>>> confusing >>>>>>>>>>>>>>>>>>>>>>>> the >>>>>>>>>>>>>>>>>>>>>>>>>>>>> the threaded patch review. >>>>>>>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>>>>>>> ___ Your changes affect IPC mechanism, >>> and >>>>> you >>>>>>>>>> don't >>>>>>>>>>>>>>>>> present >>>>>>>>>>>>>>>>>>> any >>>>>>>>>>>>>>>>>>>>>>>>>> results >>>>>>>>>>>>>>>>>>>>>>>>>>>>> for in-service upgradability test. >>>>>>>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>>>>>>> ___ Your changes affect user manual and >>>>>>>>>>>> documentation, >>>>>>>>>>>>>>>>> your >>>>>>>>>>>>>>>>>>>>>> patch >>>>>>>>>>>>>>>>>>>>>>>>>> series >>>>>>>>>>>>>>>>>>>>>>>>>>>>> do not contain the patch that updates >>>>>>>>> the >>>>>>>>>>>>>>>>>>>>>>>>>>>>> Doxygen >>>>>>>>>>>>>>>>>>> manual. > ---------------------------------------------------------------------------- >>>>>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>>>>>>> Check out the vibrant tech community on one of the >>>>> world's >>>>>>>>>> most >>>>>>>>>>>>>>>>>>>> engaging tech sites, SlashDot.org! >>>>> http://sdm.link/slashdot >>> _______________________________________________ >>>>>>>>>>>>>>>>>>>> Opensaf-devel mailing list >>>>>>>>>>>>>>>>>>>> Opensaf-devel@lists.sourceforge.net >>>>>>>>>>>>>>>>>>>> https://lists.sourceforge.net/lists/listinfo/opensaf- >>> devel ------------------------------------------------------------------------------ Check out the vibrant tech community on one of the world's most engaging tech sites, SlashDot.org! http://sdm.link/slashdot _______________________________________________ Opensaf-devel mailing list Opensaf-devel@lists.sourceforge.net https://lists.sourceforge.net/lists/listinfo/opensaf-devel