Hi Vu, Now we are now able to proceed further with V7 `2258_v7.patch` in service upgraded working fine, because of Encode/decode changes done in V7 patch.
But we have another small test case issue (/usr/bin/logtest 5 17 Segmentation fault), once we resolve this also, we can conclude that all the basic functionality is working, then you can re-publish the V7 patch ( if change occurred in Lennart #2258 V2 please do publish that as well ) so that I can go for CODE review. Steps to reproduce the test case issue : 1) Bring up old node as Active ( with out `2258_v7.patch` ) 2) Bring-up new node as Standby ( with `2258_v7.patch` ) 3) Do `amf-adm si-swap safSi=SC-2N,safApp=OpenSAF` 4) Run `/usr/bin/logtest 5 17 ` on new Active (because of si-swap ) Note : both nodes has the new XLM attributes populated . =================================================================================== gdb /usr/bin/logtest (gdb) r 5 16 PASSED CCB Object Modify, change root directory. Path exist. OK; Detaching after fork from child process 13797. Set values Fail [New Thread 0x7ffff7ff7b00 (LWP 13801)] [New Thread 0x7ffff7fc4b00 (LWP 13802)] Program received signal SIGSEGV, Segmentation fault. 0x00005555555688ea in read_and_compare.isra.7 () at src/log/apitest/tet_LogOiOps.c:1891 1891 src/log/apitest/tet_LogOiOps.c: No such file or directory. in src/log/apitest/tet_LogOiOps.c (gdb) bt #0 0x00005555555688ea in read_and_compare.isra.7 () at src/log/apitest/tet_LogOiOps.c:1891 #1 0x0000555555568a4b in check_logRecordDestinationConfigurationAdd () at src/log/apitest/tet_LogOiOps.c:1941 #2 0x0000555555571b05 in run_test_case () #3 0x0000555555571feb in test_run () #4 0x000055555555bfad in main () at src/log/apitest/logtest.c:569 (gdb) =================================================================================== -AVM On 2/23/2017 11:44 AM, Vu Minh Nguyen wrote: > Hi Mahesh, > > Maybe it was broken when transmitting. I zipped to a tar file. Please try it > one more. > > Regards, Vu > > >> -----Original Message----- >> From: A V Mahesh [mailto:mahesh.va...@oracle.com] >> Sent: Thursday, February 23, 2017 12:54 PM >> To: Vu Minh Nguyen <vu.m.ngu...@dektech.com.au>; >> lennart.l...@ericsson.com; canh.v.tru...@dektech.com.au >> Cc: opensaf-devel@lists.sourceforge.net >> Subject: Re: [devel] [PATCH 0 of 3] Review Request for log: add > alternative >> destinations of log records [#2258] V4 >> >> Hi Vu, >> >> On 2/23/2017 10:20 AM, Vu Minh Nguyen wrote: >>> Hi Mahesh, >>> >>> Can you try with 2258_v7.patch I just sent to you? >> I striped changeset: 8610 of today's latest staging ( `hg strip >> 8610` which removed log: implement SaLogFilterSetCallbackT and version >> handling [#2146]) >> and try to apply your `2258_v7.patch`, it says `malformed patch at line >> 3324`. >> >> -AVM >>> I have pulled the latest code on OpenSAF 5.1 branch, re-created the > cluster. >>> And it works with the case old active SC-1 (OpenSAF 5.1) and new standby >>> SC-2 (with 2258_v7.patch included in). >>> >>> To apply 2258_v7.patch, please do remove the just pushed ticket "log: >>> implement SaLogFilterSetCallbackT and version handling [#2146]" , >>> I have not rebased the code on that yet. >>> >>> Regards, Vu >>> >>>> -----Original Message----- >>>> From: A V Mahesh [mailto:mahesh.va...@oracle.com] >>>> Sent: Thursday, February 23, 2017 11:45 AM >>>> To: Vu Minh Nguyen <vu.m.ngu...@dektech.com.au>; >>>> lennart.l...@ericsson.com; canh.v.tru...@dektech.com.au >>>> Cc: opensaf-devel@lists.sourceforge.net >>>> Subject: Re: [devel] [PATCH 0 of 3] Review Request for log: add >>> alternative >>>> destinations of log records [#2258] V4 >>>> >>>> Hi Vu/Lennart, >>>> >>>> >>>> In broad WITHOUT the #2258 patch, the same code/setup working fine >> with >>>> 2 sc node (staging changeset: 8609 ), >>>> as soon as we apply `2258_v5.patch` V5 patch on staging (changeset: >>>> 8609 ) that you have provided yesterday, >>>> on one sc node and try to bring up that in to cluster (in-service test) >>>> we are observing the issue of new node (with #2258 patch) not joining >>>> cluster. >>>> >>>> >> ============================================================== >>>> ==================================================== >>>> eb 23 10:01:59 SC-1 osafimmnd[15279]: NO Implementer (applier) >>>> connected: 15 (@safAmfService2010f) <127, 2010f> >>>> Feb 23 10:01:59 SC-1 osaflogd[15289]: NO LOGSV_DATA_GROUPNAME >> not >>>> found >>>> Feb 23 10:01:59 SC-1 osaflogd[15289]: NO LOG root directory is: >>>> "/var/log/opensaf/saflog" >>>> Feb 23 10:01:59 SC-1 osaflogd[15289]: NO LOG data group is: "" >>>> Feb 23 10:01:59 SC-1 osaflogd[15289]: NO LGS_MBCSV_VERSION = 7 >>>> Feb 23 10:01:59 SC-1 osafamfnd[15329]: NO Assigned >>>> 'safSi=SC-2N,safApp=OpenSAF' STANDBY to 'safSu=SC- >>>> 1,safSg=2N,safApp=OpenSAF' >>>> Feb 23 10:01:59 SC-1 opensafd: OpenSAF(5.1.M0 - ) services successfully >>>> started >>>> Feb 23 10:01:59 SC-1 osafamfnd[15329]: NO >>>> 'safComp=LOG,safSu=SC-1,safSg=2N,safApp=OpenSAF' faulted due to >>>> 'avaDown' : Recovery is 'nodeFailfast' >>>> Feb 23 10:01:59 SC-1 osafamfnd[15329]: ER >>>> safComp=LOG,safSu=SC-1,safSg=2N,safApp=OpenSAF Faulted due >>>> to:avaDown >>>> Recovery is:nodeFailfast >>>> Feb 23 10:01:59 SC-1 osafamfnd[15329]: Rebooting OpenSAF NodeId = >>>> 131343 >>>> EE Name = , Reason: Component faulted: recovery is node failfast, >>>> OwnNodeId = 131343, SupervisionTime = 60 >>>> Feb 23 10:01:59 SC-1 opensaf_reboot: Rebooting local node; timeout=60 >>>> Feb 23 10:02:00 SC-1 osafimmnd[15279]: NO Implementer (applier) >>>> connected: 16 (@OpenSafImmReplicatorB) <144, 2010f> >>>> Feb 23 10:01:59 SC-1 opensaf_reboot: Rebooting local node; timeout=60 >>>> >> ============================================================== >>>> ==================================================== >>>> >>>> So it is evident that in-service upgrade part code of this need to be >>>> corrected. >>>> >>>> Please see my comments as [AVM] and let me know if you need some >> traces >>>> . >>>> >>>> If you're planing to prepare new V6 patch , please do prepare on top of >>>> today's latest staging. >>>> >>>> On 2/23/2017 9:33 AM, Vu Minh Nguyen wrote: >>>>> Hi Mahesh, >>>>> >>>>> I have done in-service upgrade/downgrade with following cases: >>>>> 1) New Active SC-1 (OpenSAF 5.2 with the attached patch) + old standby >>>> SC-2 >>>>> (OpenSAF 5.1) >>>>> --> Work fine >>>> [AVM] This is not a practical use cause of in-service upgrade , we can >>>> ignore this test further >>>>> 2) Old Active SC-1 (OpenSAF 5.1) + new standby SC-2 (with or without >>>>> attached patch) >>>>> --> SC-2 is restarted & not able to join the cluster. >>>> [AVM] This use cause/flow is we do get in in-service upgrade , so we >>>> need to address this. >>>>> I got following messages in syslog: >>>>> Feb 23 09:32:42 SC-2 user.notice opensafd: OpenSAF(5.2.M0 - >>>>> 8529:b5addd36e45d:default) services successfully started >>>>> Feb 23 09:32:43 SC-2 local0.warn osafntfimcnd[701]: WA >>>> ntfimcn_imm_init >>>>> saImmOiImplementerSet() returned SA_AIS_ERR_TIMEOUT (5) >>>>> Feb 23 09:32:45 SC-2 local0.warn osafntfimcnd[701]: WA >>>> ntfimcn_imm_init >>>>> saImmOiImplementerSet() returned SA_AIS_ERR_TIMEOUT (5) >>>>> Feb 23 09:32:47 SC-2 local0.warn osafntfimcnd[701]: WA >>>> ntfimcn_imm_init >>>>> saImmOiImplementerSet() returned SA_AIS_ERR_TIMEOUT (5) >>>>> Feb 23 09:32:49 SC-2 local0.warn osafntfimcnd[701]: WA >>>> ntfimcn_imm_init >>>>> saImmOiImplementerSet() returned SA_AIS_ERR_TIMEOUT (5) >>>>> Feb 23 09:32:50 SC-2 local0.err osafmsgnd[592]: ER >>>> saImmOiImplementerSet >>>>> FAILED:5 >>>>> Feb 23 09:32:50 SC-2 local0.err osafmsgnd[592]: ER >>>> saImmOiImplementerSet >>>>> FAILED:5 >>>>> Feb 23 09:32:50 SC-2 local0.notice osafamfnd[496]: NO >>>>> 'safSu=SC-2,safSg=NoRed,safApp=OpenSAF' component restart >> probation >>>> timer >>>>> started (timeout: 60000000000 ns) >>>>> Feb 23 09:32:50 SC-2 local0.notice osafamfnd[496]: NO Restarting a >>>> component >>>>> of 'safSu=SC-2,safSg=NoRed,safApp=OpenSAF' (comp restart count: 1) >>>>> Feb 23 09:32:50 SC-2 local0.notice osafamfnd[496]: NO >>>>> 'safComp=MQND,safSu=SC-2,safSg=NoRed,safApp=OpenSAF' faulted >> due >>>> to >>>>> 'avaDown' : Recovery is 'componentRestart' >>>>> Feb 23 09:32:50 SC-2 local0.info osafmsgnd[736]: mkfifo already > exists: >>>>> /var/lib/opensaf/osafmsgnd.fifo File exists >>>>> >>>>> And sometimes, on active SC-1 (OpenSAF 5.1), the node is not able to > up >>>>> because of following error: >>>>> >>>>> Feb 23 11:00:32 SC-1 local0.err osafclmna[406]: MDTM:TIPC Dsock >> Socket >>>>> creation failed in MDTM_INIT err :Address family not supported by >>>> protocol >>>>> Feb 23 11:00:32 SC-1 local0.err osafclmna[406]: ER ncs_agents_startup >>>> FAILED >>>> [AVM] No such issues ( with both TCP & TIPC) (staging changeset: > 8609 >>> ) >>>>> Are you getting similar problem at your side? >>>>> Please note that, the problem is existed WITH or WITHOUT the #2258 >>>> patch. >>>> [AVM] No , problem only if we apply `2258_v5.patch` V5 patch on staging >>>> (changeset: 8609 ) >>>> try to bring up that node in to cluster. >>>> >>>> >>>> -AVM >>>> >>>>> I have informed this to IMM to have a look, not sure any problem with >>> MDS >>>>> layer or any problem with my environment setup. >>>>> In the meantime, please have a look at the updated patch, I will >>> continue >>>>> checking the problem. Will keep you updated. >>>> [AVM] I haven't seen any IMM problems >>>>> Regards, Vu >>>>> >>>>>> -----Original Message----- >>>>>> From: A V Mahesh [mailto:mahesh.va...@oracle.com] >>>>>> Sent: Wednesday, February 22, 2017 5:36 PM >>>>>> To: Vu Minh Nguyen <vu.m.ngu...@dektech.com.au>; >>>>>> lennart.l...@ericsson.com; canh.v.tru...@dektech.com.au >>>>>> Cc: opensaf-devel@lists.sourceforge.net >>>>>> Subject: Re: [devel] [PATCH 0 of 3] Review Request for log: add >>>>> alternative >>>>>> destinations of log records [#2258] V4 >>>>>> >>>>>> Please see correction New Standby SC-1 ( with patch ) >>>>>> >>>>>> -AVM >>>>>> >>>>>> On 2/22/2017 4:02 PM, A V Mahesh wrote: >>>>>>> Hi Vu, >>>>>>> >>>>>>> With this new patch , we have another issue : >>>>>>> >>>>>>> 1) standby Core by `/usr/lib64/opensaf/osaflogd' issue got > resolved >>> . >>>>>>> 2) In-service upgrade is Not working , I have Old Active SC-2 ( > with >>>>>>> out patch ) and New Standby SC-1 ( with patch ) >>>>>>> >>>>>>> the new New Standby SC-1 not joining the cluster ( in-service >>>>>>> upgrade failed ) >>>>>>> >>>>>>> New Standby SC-1 >>>>>>> >>>>>>> >> ============================================================== >>>>>> ====================================================== >>>>>>> Feb 22 15:53:05 SC-1 osafamfnd[15752]: NO >>>>>>> 'safSu=SC-1,safSg=NoRed,safApp=OpenSAF' Presence State >>>>>> INSTANTIATING >>>>>>> => INSTANTIATED >>>>>>> Feb 22 15:53:05 SC-1 osafamfnd[15752]: NO Assigning >>>>>>> 'safSi=NoRed4,safApp=OpenSAF' ACTIVE to >>>>>>> 'safSu=SC-1,safSg=NoRed,safApp=OpenSAF' >>>>>>> Feb 22 15:53:05 SC-1 osafamfnd[15752]: NO Assigned >>>>>>> 'safSi=NoRed4,safApp=OpenSAF' ACTIVE to >>>>>>> 'safSu=SC-1,safSg=NoRed,safApp=OpenSAF' >>>>>>> Feb 22 15:53:05 SC-1 osafsmfd[15889]: Started >>>>>>> Feb 22 15:53:05 SC-1 osafamfnd[15752]: NO >>>>>>> 'safSu=SC-1,safSg=2N,safApp=OpenSAF' Presence State >> INSTANTIATING >>>> => >>>>>>> INSTANTIATED >>>>>>> Feb 22 15:53:05 SC-1 osafamfnd[15752]: NO Assigning >>>>>>> 'safSi=SC-2N,safApp=OpenSAF' STANDBY to >>>>>>> 'safSu=SC-1,safSg=2N,safApp=OpenSAF' >>>>>>> Feb 22 15:53:05 SC-1 osafrded[15672]: NO RDE role set to STANDBY >>>>>>> Feb 22 15:53:05 SC-1 osafrded[15672]: NO Peer up on node 0x2020f >>>>>>> Feb 22 15:53:05 SC-1 osafrded[15672]: NO Got peer info request from >>>>>>> node 0x2020f with role ACTIVE >>>>>>> Feb 22 15:53:05 SC-1 osafrded[15672]: NO Got peer info response >> from >>>>>>> node 0x2020f with role ACTIVE >>>>>>> Feb 22 15:53:05 SC-1 osafimmd[15691]: NO MDS event from svc_id 24 >>>>>>> (change:5, dest:13) >>>>>>> Feb 22 15:53:05 SC-1 osafimmd[15691]: NO MDS event from svc_id 24 >>>>>>> (change:3, dest:13) >>>>>>> Feb 22 15:53:05 SC-1 osafimmd[15691]: NO MDS event from svc_id 24 >>>>>>> (change:5, dest:13) >>>>>>> Feb 22 15:53:05 SC-1 osafimmd[15691]: NO MDS event from svc_id 25 >>>>>>> (change:3, dest:567412424453430) >>>>>>> Feb 22 15:53:05 SC-1 osafimmd[15691]: NO MDS event from svc_id 25 >>>>>>> (change:3, dest:565213401202663) >>>>>>> Feb 22 15:53:05 SC-1 osafimmd[15691]: NO MDS event from svc_id 25 >>>>>>> (change:3, dest:566312912825221) >>>>>>> Feb 22 15:53:05 SC-1 osafimmd[15691]: NO MDS event from svc_id 25 >>>>>>> (change:3, dest:564113889574230) >>>>>>> Feb 22 15:53:05 SC-1 osafimmnd[15702]: NO Implementer (applier) >>>>>>> connected: 17 (@safAmfService2010f) <127, 2010f> >>>>>>> Feb 22 15:53:05 SC-1 osaflogd[15712]: NO >> LOGSV_DATA_GROUPNAME >>>>>> not found >>>>>>> Feb 22 15:53:05 SC-1 osaflogd[15712]: NO LOG root directory is: >>>>>>> "/var/log/opensaf/saflog" >>>>>>> Feb 22 15:53:05 SC-1 osaflogd[15712]: NO LOG data group is: "" >>>>>>> Feb 22 15:53:05 SC-1 osaflogd[15712]: NO LGS_MBCSV_VERSION = 7 >>>>>>> Feb 22 15:53:05 SC-1 osafamfnd[15752]: NO Assigned >>>>>>> 'safSi=SC-2N,safApp=OpenSAF' STANDBY to >>>>>>> 'safSu=SC-1,safSg=2N,safApp=OpenSAF' >>>>>>> Feb 22 15:53:05 SC-1 opensafd: OpenSAF(5.1.M0 - ) services >>>>>>> successfully started >>>>>>> Feb 22 15:53:05 SC-1 osafamfnd[15752]: NO >>>>>>> 'safComp=LOG,safSu=SC-1,safSg=2N,safApp=OpenSAF' faulted due to >>>>>>> 'avaDown' : Recovery is 'nodeFailfast' >>>>>>> Feb 22 15:53:05 SC-1 osafamfnd[15752]: ER >>>>>>> safComp=LOG,safSu=SC-1,safSg=2N,safApp=OpenSAF Faulted due >>>>>> to:avaDown >>>>>>> Recovery is:nodeFailfast >>>>>>> Feb 22 15:53:05 SC-1 osafamfnd[15752]: Rebooting OpenSAF NodeId >> = >>>>>>> 131343 EE Name = , Reason: Component faulted: recovery is node >>>>>>> failfast, OwnNodeId = 131343, SupervisionTime = 60 >>>>>>> Feb 22 15:53:05 SC-1 opensaf_reboot: Rebooting local node; >>>> timeout=60 >>>>>>> Feb 22 15:53:43 SC-1 syslog-ng[1171]: syslog-ng starting up; >>>>>>> version='2.0.9' >>>>>>> >>>>>>> >> ============================================================== >>>>>> ====================================================== >>>>>>> Old - Active - SC-2 >>>>>>> >>>>>>> >> ============================================================== >>>>>> ====================================================== >>>>>>> Feb 22 15:53:02 SC-2 osafimmnd[16359]: NO NODE STATE-> >>>>>>> IMM_NODE_R_AVAILABLE >>>>>>> Feb 22 15:53:02 SC-2 osafimmloadd: NO Sync starting >>>>>>> Feb 22 15:53:02 SC-2 osafimmloadd: IN Synced 390 objects in total >>>>>>> Feb 22 15:53:02 SC-2 osafimmnd[16359]: NO NODE STATE-> >>>>>>> IMM_NODE_FULLY_AVAILABLE 18511 >>>>>>> Feb 22 15:53:02 SC-2 osafimmnd[16359]: NO Epoch set to 3 in >>>> ImmModel >>>>>>> Feb 22 15:53:02 SC-2 osafimmd[16346]: NO ACT: New Epoch for >>>> IMMND >>>>>>> process at node 2020f old epoch: 2 new epoch:3 >>>>>>> Feb 22 15:53:02 SC-2 osafimmd[16346]: NO ACT: New Epoch for >>>> IMMND >>>>>>> process at node 2040f old epoch: 2 new epoch:3 >>>>>>> Feb 22 15:53:02 SC-2 osafimmd[16346]: NO ACT: New Epoch for >>>> IMMND >>>>>>> process at node 2030f old epoch: 2 new epoch:3 >>>>>>> Feb 22 15:53:02 SC-2 osafimmloadd: NO Sync ending normally >>>>>>> Feb 22 15:53:02 SC-2 osafimmd[16346]: NO ACT: New Epoch for >>>> IMMND >>>>>>> process at node 2010f old epoch: 0 new epoch:3 >>>>>>> Feb 22 15:53:02 SC-2 osafimmnd[16359]: NO SERVER STATE: >>>>>>> IMM_SERVER_SYNC_SERVER --> IMM_SERVER_READY >>>>>>> Feb 22 15:53:03 SC-2 osafamfd[16408]: NO Received node_up from >>>> 2010f: >>>>>>> msg_id 1 >>>>>>> Feb 22 15:53:03 SC-2 osafamfd[16408]: NO Node 'SC-1' joined the >>>> cluster >>>>>>> Feb 22 15:53:03 SC-2 osafimmnd[16359]: NO Implementer connected: >>>> 16 >>>>>>> (MsgQueueService131343) <0, 2010f> >>>>>>> Feb 22 15:53:03 SC-2 osafrded[16327]: NO Peer up on node 0x2010f >>>>>>> Feb 22 15:53:03 SC-2 osafrded[16327]: NO Got peer info request from >>>>>>> node 0x2010f with role STANDBY >>>>>>> Feb 22 15:53:03 SC-2 osafrded[16327]: NO Got peer info response >> from >>>>>>> node 0x2010f with role STANDBY >>>>>>> Feb 22 15:53:03 SC-2 osafimmd[16346]: NO MDS event from svc_id 24 >>>>>>> (change:5, dest:13) >>>>>>> Feb 22 15:53:03 SC-2 osafimmnd[16359]: NO Implementer (applier) >>>>>>> connected: 17 (@safAmfService2010f) <0, 2010f> >>>>>>> Feb 22 15:53:03 SC-2 osafamfd[16408]: NO Cluster startup is done >>>>>>> Feb 22 15:53:04 SC-2 osafimmnd[16359]: NO Implementer (applier) >>>>>>> connected: 18 (@OpenSafImmReplicatorB) <0, 2010f> >>>>>>> Feb 22 15:53:09 SC-2 osafdtmd[16304]: NO Lost contact with 'SC-1' >>>>>>> Feb 22 15:53:09 SC-2 osaffmd[16336]: NO Node Down event for node >> id >>>>>>> 2010f: >>>>>>> Feb 22 15:53:09 SC-2 osafimmd[16346]: NO MDS event from svc_id 24 >>>>>>> (change:6, dest:13) >>>>>>> Feb 22 15:53:09 SC-2 osafimmd[16346]: NO MDS event from svc_id 25 >>>>>>> (change:4, dest:564113889574230) >>>>>>> Feb 22 15:53:09 SC-2 osaffmd[16336]: NO Current role: ACTIVE >>>>>>> Feb 22 15:53:09 SC-2 osaffmd[16336]: Rebooting OpenSAF NodeId = >>>>>> 131343 >>>>>>> EE Name = , Reason: Received Node Down for peer controller, >>>> OwnNodeId >>>>>>> = 131599, SupervisionTime = 60 >>>>>>> Feb 22 15:53:09 SC-2 osafrded[16327]: NO Peer down on node >> 0x2010f >>>>>>> Feb 22 15:53:09 SC-2 osafamfd[16408]: NO Node 'SC-1' left the >> cluster >>>>>>> Feb 22 15:53:09 SC-2 osaflogd[16372]: NO Failed (2) to send of WRITE >>>>>>> ack to: 2010f00003d6a >>>>>>> Feb 22 15:53:09 SC-2 osaflogd[16372]: NO Failed (2) to send of WRITE >>>>>>> ack to: 2010f00003d6a >>>>>>> Feb 22 15:53:09 SC-2 osaflogd[16372]: NO Failed (2) to send of WRITE >>>>>>> ack to: 2010f00003d74 >>>>>>> Feb 22 15:53:09 SC-2 osafclmd[16398]: NO Node 131343 went down. >>>> Not >>>>>>> sending track callback for agents on that node >>>>>>> Feb 22 15:53:09 SC-2 osafclmd[16398]: NO Node 131343 went down. >>>> Not >>>>>>> sending track callback for agents on that node >>>>>>> Feb 22 15:53:09 SC-2 osafclmd[16398]: NO Node 131343 went down. >>>> Not >>>>>>> sending track callback for agents on that node >>>>>>> Feb 22 15:53:09 SC-2 osafclmd[16398]: NO Node 131343 went down. >>>> Not >>>>>>> sending track callback for agents on that node >>>>>>> Feb 22 15:53:09 SC-2 osafclmd[16398]: NO Node 131343 went down. >>>> Not >>>>>>> sending track callback for agents on that node >>>>>>> Feb 22 15:53:09 SC-2 osafclmd[16398]: NO Node 131343 went down. >>>> Not >>>>>>> sending track callback for agents on that node >>>>>>> Feb 22 15:53:09 SC-2 osafclmd[16398]: NO Node 131343 went down. >>>> Not >>>>>>> sending track callback for agents on that node >>>>>>> Feb 22 15:53:09 SC-2 osafimmd[16346]: WA IMMD lost contact with >>>> peer >>>>>>> IMMD (NCSMDS_RED_DOWN) >>>>>>> Feb 22 15:53:09 SC-2 osafimmnd[16359]: NO Global discard node >>>> received >>>>>>> for nodeId:2010f pid:15702 >>>>>>> Feb 22 15:53:09 SC-2 osafimmnd[16359]: NO Implementer >> disconnected >>>> 16 >>>>>>> <0, 2010f(down)> (MsgQueueService131343) >>>>>>> Feb 22 15:53:09 SC-2 osafimmnd[16359]: NO Implementer >> disconnected >>>> 17 >>>>>>> <0, 2010f(down)> (@safAmfService2010f) >>>>>>> Feb 22 15:53:09 SC-2 osafimmnd[16359]: NO Implementer >> disconnected >>>> 18 >>>>>>> <0, 2010f(down)> (@OpenSafImmReplicatorB) >>>>>>> Feb 22 15:53:09 SC-2 opensaf_reboot: Rebooting remote node in the >>>>>>> absence of PLM is outside the scope of OpenSAF >>>>>>> >>>>>>> >> ============================================================== >>>>>> ====================================================== >>>>>>> -AVM >>>>>>> >>>>>>> On 2/22/2017 3:13 PM, Vu Minh Nguyen wrote: >>>>>>>> Hi Mahesh, >>>>>>>> >>>>>>>> I put all required patches into one. Try to use this and see if you >>>>>>>> still >>>>>>>> have that problem or not. >>>>>>>> >>>>>>>> Regards, Vu >>>>>>>> >>>>>>>>> -----Original Message----- >>>>>>>>> From: A V Mahesh [mailto:mahesh.va...@oracle.com] >>>>>>>>> Sent: Wednesday, February 22, 2017 3:35 PM >>>>>>>>> To: Vu Minh Nguyen <vu.m.ngu...@dektech.com.au>; >>>>>>>>> lennart.l...@ericsson.com; canh.v.tru...@dektech.com.au >>>>>>>>> Cc: opensaf-devel@lists.sourceforge.net >>>>>>>>> Subject: Re: [devel] [PATCH 0 of 3] Review Request for log: add >>>>>>>> alternative >>>>>>>>> destinations of log records [#2258] V4 >>>>>>>>> >>>>>>>>> Hi Vu, >>>>>>>>> >>>>>>>>> I used new #3 and #4 patches , Can you please re-send All the >>> final >>>>>>>>> patch in go, >>>>>>>>> >>>>>>>>> which i need to apply on today`s staging ( if possible publish > the >>>>>>>>> with >>>>>>>>> new version ) >>>>>>>>> >>>>>>>>> -AVM >>>>>>>>> >>>>>>>>> >>>>>>>>> On 2/22/2017 1:52 PM, Vu Minh Nguyen wrote: >>>>>>>>>> Hi Mahesh, >>>>>>>>>> >>>>>>>>>>> Core was generated by `/usr/lib64/opensaf/osaflogd'. >>>>>>>>>>> Program terminated with signal 11, Segmentation fault. >>>>>>>>>>> #0 ckpt_proc_cfg_stream(lgs_cb*, void*) () at >>>>>>>>>>> src/log/logd/lgs_mbcsv.cc:2195 >>>>>>>>>>> 2195 src/log/logd/lgs_mbcsv.cc: No such file or directory. >>>>>>>>>>> in src/log/logd/lgs_mbcsv.cc >>>>>>>>>> Backtrace still points to old position (lgs_mbcsv:2195). I guess >>> the >>>>>>>>>> osaflogd binary has not been updated with the fixed patch. >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> Regards, Vu >>>>>>>>>> >>>>>>>>>>> -----Original Message----- >>>>>>>>>>> From: A V Mahesh [mailto:mahesh.va...@oracle.com] >>>>>>>>>>> Sent: Wednesday, February 22, 2017 3:18 PM >>>>>>>>>>> To: Vu Minh Nguyen <vu.m.ngu...@dektech.com.au>; >>>>>>>>>>> lennart.l...@ericsson.com; canh.v.tru...@dektech.com.au >>>>>>>>>>> Cc: opensaf-devel@lists.sourceforge.net >>>>>>>>>>> Subject: Re: [devel] [PATCH 0 of 3] Review Request for log: add >>>>>>>>>> alternative >>>>>>>>>>> destinations of log records [#2258] V4 >>>>>>>>>>> >>>>>>>>>>> Hi Vu, >>>>>>>>>>> >>>>>>>>>>> SC-2 standby osaflogd core dumped still occurs ( Not resolved) , >>> the >>>>>>>> new >>>>>>>>>>> patch only resolved the application (/usr/bin/logtest ) >>> Segmentation >>>>>>>>>>> fault on SC-1 Active. >>>>>>>>>>> >>>>>>>>>>> >> ============================================================== >> ============================================================== >>>>>>>>>>> # gdb /usr/lib64/opensaf/osaflogd >>>> core_1487751055.osaflogd.4594 >>>>>> GNU >>>>>>>>>>> gdb >>>>>>>>>>> (GDB) SUSE (7.3-0.6.1) >>>>>>>>>>> Copyright (C) 2011 Free Software Foundation, Inc. >>>>>>>>>>> ....... >>>>>>>>>>> Core was generated by `/usr/lib64/opensaf/osaflogd'. >>>>>>>>>>> Program terminated with signal 11, Segmentation fault. >>>>>>>>>>> #0 ckpt_proc_cfg_stream(lgs_cb*, void*) () at >>>>>>>>>>> src/log/logd/lgs_mbcsv.cc:2195 >>>>>>>>>>> 2195 src/log/logd/lgs_mbcsv.cc: No such file or directory. >>>>>>>>>>> in src/log/logd/lgs_mbcsv.cc >>>>>>>>>>> (gdb) bt >>>>>>>>>>> #0 ckpt_proc_cfg_stream(lgs_cb*, void*) () at >>>>>>>>>>> src/log/logd/lgs_mbcsv.cc:2195 >>>>>>>>>>> #1 0x00007f97b026f960 in ckpt_decode_log_struct(lgs_cb*, >>>>>>>>>>> ncs_mbcsv_cb_arg*, void*, void*, unsigned int (*)(edu_hdl_tag*, >>>>>>>>>>> edu_tkn_tag*, void*, unsigned int*, edu_buf_env_tag*, >>>>>> EDP_OP_TYPE, >>>>>>>>>>> EDU_ERR*)) () at src/log/logd/lgs_mbcsv.cc:950 >>>>>>>>>>> #2 0x00007f97b02710dc in ckpt_decode_async_update(lgs_cb*, >>>>>>>>>>> ncs_mbcsv_cb_arg*) () at src/log/logd/lgs_mbcsv.cc:1086 >>>>>>>>>>> #3 0x00007f97b0273941 in mbcsv_callback(ncs_mbcsv_cb_arg*) >> () >>>> at >>>>>>>>>>> src/log/logd/lgs_mbcsv.cc:880 >>>>>>>>>>> #4 0x00007f97af372596 in ncs_mbscv_rcv_decode () from >>>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0 >>>>>>>>>>> #5 0x00007f97af372766 in ncs_mbcsv_rcv_async_update () from >>>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0 >>>>>>>>>>> #6 0x00007f97af379370 in mbcsv_process_events () from >>>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0 >>>>>>>>>>> #7 0x00007f97af3794db in mbcsv_hdl_dispatch_all () from >>>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0 >>>>>>>>>>> #8 0x00007f97af373ce2 in mbcsv_process_dispatch_request () >> at >>>>>>>>>>> src/mbc/mbcsv_api.c:423 >>>>>>>>>>> #9 0x00007f97b027096e in lgs_mbcsv_dispatch(unsigned int) () >> at >>>>>>>>>>> src/log/logd/lgs_mbcsv.cc:327 >>>>>>>>>>> #10 0x00007f97b024d9f2 in main () at >> src/log/logd/lgs_main.cc:583 >>>>>>>>>>> (gdb) bt full >>>>>>>>>>> #0 ckpt_proc_cfg_stream(lgs_cb*, void*) () at >>>>>>>>>>> src/log/logd/lgs_mbcsv.cc:2195 >>>>>>>>>>> ckpt_data_handler = {0x7f97b0270300 >>>>>>>>>>> <ckpt_proc_initialize_client(lgs_cb*, void*)>, 0x7f97b02701c0 >>>>>>>>>>> <ckpt_proc_finalize_client(lgs_cb*, void*)>, >>>>>>>>>>> 0x7f97b0270060 <ckpt_proc_agent_down(lgs_cb*, >>> void*)>, >>>>>>>>>>> 0x7f97b02712f0 <ckpt_proc_log_write(lgs_cb*, void*)>, >>>>>> 0x7f97b0271ab0 >>>>>>>>>>> <ckpt_proc_open_stream(lgs_cb*, void*)>, >>>>>>>>>>> 0x7f97b026fe80 <ckpt_proc_close_stream(lgs_cb*, >>>>> void*)>, >>>>>>>>>>> 0x7f97b0272380 <ckpt_proc_cfg_stream(lgs_cb*, void*)>, >>>>>>>>> 0x7f97b0274800 >>>>>>>>>>> <ckpt_proc_lgs_cfg_v2(lgs_cb*, void*)>, >>>>>>>>>>> 0x7f97b0274e10 <ckpt_proc_lgs_cfg_v3(lgs_cb*, >>> void*)>, >>>>>>>>>>> 0x7f97b02754f0 <ckpt_proc_lgs_cfg_v5(lgs_cb*, void*)>} >>>>>>>>>>> #1 0x00007f97b026f960 in ckpt_decode_log_struct(lgs_cb*, >>>>>>>>>>> ncs_mbcsv_cb_arg*, void*, void*, unsigned int (*)(edu_hdl_tag*, >>>>>>>>>>> edu_tkn_tag*, void*, unsigned int*, edu_buf_env_tag*, >>>>>> EDP_OP_TYPE, >>>>>>>>>>> EDU_ERR*)) () at src/log/logd/lgs_mbcsv.cc:950 >>>>>>>>>>> ckpt_data_handler = {0x7f97b0270300 >>>>>>>>>>> <ckpt_proc_initialize_client(lgs_cb*, void*)>, 0x7f97b02701c0 >>>>>>>>>>> <ckpt_proc_finalize_client(lgs_cb*, void*)>, >>>>>>>>>>> 0x7f97b0270060 <ckpt_proc_agent_down(lgs_cb*, >>> void*)>, >>>>>>>>>>> 0x7f97b02712f0 <ckpt_proc_log_write(lgs_cb*, void*)>, >>>>>> 0x7f97b0271ab0 >>>>>>>>>>> <ckpt_proc_open_stream(lgs_cb*, void*)>, >>>>>>>>>>> 0x7f97b026fe80 <ckpt_proc_close_stream(lgs_cb*, >>>>> void*)>, >>>>>>>>>>> 0x7f97b0272380 <ckpt_proc_cfg_stream(lgs_cb*, void*)>, >>>>>>>>> 0x7f97b0274800 >>>>>>>>>>> <ckpt_proc_lgs_cfg_v2(lgs_cb*, void*)>, >>>>>>>>>>> 0x7f97b0274e10 <ckpt_proc_lgs_cfg_v3(lgs_cb*, >>> void*)>, >>>>>>>>>>> 0x7f97b02754f0 <ckpt_proc_lgs_cfg_v5(lgs_cb*, void*)>} >>>>>>>>>>> #2 0x00007f97b02710dc in ckpt_decode_async_update(lgs_cb*, >>>>>>>>>>> ncs_mbcsv_cb_arg*) () at src/log/logd/lgs_mbcsv.cc:1086 >>>>>>>>>>> ckpt_data_handler = {0x7f97b0270300 >>>>>>>>>>> <ckpt_proc_initialize_client(lgs_cb*, void*)>, 0x7f97b02701c0 >>>>>>>>>>> <ckpt_proc_finalize_client(lgs_cb*, void*)>, >>>>>>>>>>> 0x7f97b0270060 <ckpt_proc_agent_down(lgs_cb*, >>> void*)>, >>>>>>>>>>> 0x7f97b02712f0 <ckpt_proc_log_write(lgs_cb*, void*)>, >>>>>> 0x7f97b0271ab0 >>>>>>>>>>> <ckpt_proc_open_stream(lgs_cb*, void*)>, >>>>>>>>>>> 0x7f97b026fe80 <ckpt_proc_close_stream(lgs_cb*, >>>>> void*)>, >>>>>>>>>>> 0x7f97b0272380 <ckpt_proc_cfg_stream(lgs_cb*, void*)>, >>>>>>>>> 0x7f97b0274800 >>>>>>>>>>> <ckpt_proc_lgs_cfg_v2(lgs_cb*, void*)>, >>>>>>>>>>> 0x7f97b0274e10 <ckpt_proc_lgs_cfg_v3(lgs_cb*, >>> void*)>, >>>>>>>>>>> 0x7f97b02754f0 <ckpt_proc_lgs_cfg_v5(lgs_cb*, void*)>} >>>>>>>>>>> #3 0x00007f97b0273941 in mbcsv_callback(ncs_mbcsv_cb_arg*) >> () >>>> at >>>>>>>>>>> src/log/logd/lgs_mbcsv.cc:880 >>>>>>>>>>> ckpt_data_handler = {0x7f97b0270300 >>>>>>>>>>> <ckpt_proc_initialize_client(lgs_cb*, void*)>, 0x7f97b02701c0 >>>>>>>>>>> <ckpt_proc_finalize_client(lgs_cb*, void*)>, >>>>>>>>>>> 0x7f97b0270060 <ckpt_proc_agent_down(lgs_cb*, >>> void*)>, >>>>>>>>>>> 0x7f97b02712f0 <ckpt_proc_log_write(lgs_cb*, void*)>, >>>>>> 0x7f97b0271ab0 >>>>>>>>>>> <ckpt_proc_open_stream(lgs_cb*, void*)>, >>>>>>>>>>> 0x7f97b026fe80 <ckpt_proc_close_stream(lgs_cb*, >>>>> void*)>, >>>>>>>>>>> 0x7f97b0272380 <ckpt_proc_cfg_stream(lgs_cb*, void*)>, >>>>>>>>> 0x7f97b0274800 >>>>>>>>>>> <ckpt_proc_lgs_cfg_v2(lgs_cb*, void*)>, >>>>>>>>>>> 0x7f97b0274e10 <ckpt_proc_lgs_cfg_v3(lgs_cb*, >>> void*)>, >>>>>>>>>>> 0x7f97b02754f0 <ckpt_proc_lgs_cfg_v5(lgs_cb*, void*)>} >>>>>>>>>>> #4 0x00007f97af372596 in ncs_mbscv_rcv_decode () from >>>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0 >>>>>>>>>>> mbcsv_init_process_req_func = {0x7f97af373630 >>>>>>>>>>> <mbcsv_process_initialize_request>, 0x7f97af374f10 >>>>>>>>>>> <mbcsv_process_get_sel_obj_request>, >>>>>>>>>>> 0x7f97af373be0 <mbcsv_process_dispatch_request>, >>>>>>>>>>> 0x7f97af373af0 <mbcsv_process_finalize_request>, >>>> 0x7f97af373cf0 >>>>>>>>>>> <mbcsv_process_open_request>, >>>>>>>>>>> 0x7f97af374050 <mbcsv_process_close_request>, >>>>>>>> 0x7f97af3741e0 >>>>>>>>>>> <mbcsv_process_chg_role_request>, 0x7f97af3744c0 >>>>>>>>>>> <mbcsv_process_snd_ckpt_request>, >>>>>>>>>>> 0x7f97af3747d0 <mbcsv_process_snd_ntfy_request>, >>>>>>>>>>> 0x7f97af374970 <mbcsv_process_snd_data_req>, >> 0x7f97af373930 >>>>>>>>>>> <mbcsv_process_get_request>, >>>>>>>>>>> 0x7f97af374bd0 <mbcsv_process_set_request>} >>>>>>>>>>> #5 0x00007f97af372766 in ncs_mbcsv_rcv_async_update () from >>>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0 >>>>>>>>>>> mbcsv_init_process_req_func = {0x7f97af373630 >>>>>>>>>>> <mbcsv_process_initialize_request>, 0x7f97af374f10 >>>>>>>>>>> <mbcsv_process_get_sel_obj_request>, >>>>>>>>>>> 0x7f97af373be0 <mbcsv_process_dispatch_request>, >>>>>>>>>>> 0x7f97af373af0 <mbcsv_process_finalize_request>, >>>> 0x7f97af373cf0 >>>>>>>>>>> <mbcsv_process_open_request>, >>>>>>>>>>> 0x7f97af374050 <mbcsv_process_close_request>, >>>>>>>> 0x7f97af3741e0 >>>>>>>>>>> <mbcsv_process_chg_role_request>, 0x7f97af3744c0 >>>>>>>>>>> <mbcsv_process_snd_ckpt_request>, >>>>>>>>>>> 0x7f97af3747d0 <mbcsv_process_snd_ntfy_request>, >>>>>>>>>>> 0x7f97af374970 <mbcsv_process_snd_data_req>, >> 0x7f97af373930 >>>>>>>>>>> <mbcsv_process_get_request>, >>>>>>>>>>> 0x7f97af374bd0 <mbcsv_process_set_request>} >>>>>>>>>>> #6 0x00007f97af379370 in mbcsv_process_events () from >>>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0 >>>>>>>>>>> No symbol table info available. >>>>>>>>>>> #7 0x00007f97af3794db in mbcsv_hdl_dispatch_all () from >>>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0 >>>>>>>>>>> No symbol table info available. >>>>>>>>>>> #8 0x00007f97af373ce2 in mbcsv_process_dispatch_request () >> at >>>>>>>>>>> src/mbc/mbcsv_api.c:423 >>>>>>>>>>> mbcsv_init_process_req_func = {0x7f97af373630 >>>>>>>>>>> <mbcsv_process_initialize_request>, 0x7f97af374f10 >>>>>>>>>>> <mbcsv_process_get_sel_obj_request>, >>>>>>>>>>> 0x7f97af373be0 <mbcsv_process_dispatch_request>, >>>>>>>>>>> 0x7f97af373af0 <mbcsv_process_finalize_request>, >>>> 0x7f97af373cf0 >>>>>>>>>>> <mbcsv_process_open_request>, >>>>>>>>>>> 0x7f97af374050 <mbcsv_process_close_request>, >>>>>>>> 0x7f97af3741e0 >>>>>>>>>>> <mbcsv_process_chg_role_request>, 0x7f97af3744c0 >>>>>>>>>>> <mbcsv_process_snd_ckpt_request>, >>>>>>>>>>> 0x7f97af3747d0 <mbcsv_process_snd_ntfy_request>, >>>>>>>>>>> 0x7f97af374970 <mbcsv_process_snd_data_req>, >> 0x7f97af373930 >>>>>>>>>>> <mbcsv_process_get_request>, >>>>>>>>>>> 0x7f97af374bd0 <mbcsv_process_set_request>} >>>>>>>>>>> #9 0x00007f97b027096e in lgs_mbcsv_dispatch(unsigned int) () >> at >>>>>>>>>>> src/log/logd/lgs_mbcsv.cc:327 >>>>>>>>>>> ckpt_data_handler = {0x7f97b0270300 >>>>>>>>>>> <ckpt_proc_initialize_client(lgs_cb*, void*)>, 0x7f97b02701c0 >>>>>>>>>>> <ckpt_proc_finalize_client(lgs_cb*, void*)>, >>>>>>>>>>> 0x7f97b0270060 <ckpt_proc_agent_down(lgs_cb*, >>> void*)>, >>>>>>>>>>> 0x7f97b02712f0 <ckpt_proc_log_write(lgs_cb*, void*)>, >>>>>> 0x7f97b0271ab0 >>>>>>>>>>> <ckpt_proc_open_stream(lgs_cb*, void*)>, >>>>>>>>>>> 0x7f97b026fe80 <ckpt_proc_close_stream(lgs_cb*, >>>>> void*)>, >>>>>>>>>>> 0x7f97b0272380 <ckpt_proc_cfg_stream(lgs_cb*, void*)>, >>>>>>>>> 0x7f97b0274800 >>>>>>>>>>> <ckpt_proc_lgs_cfg_v2(lgs_cb*, void*)>, >>>>>>>>>>> 0x7f97b0274e10 <ckpt_proc_lgs_cfg_v3(lgs_cb*, >>> void*)>, >>>>>>>>>>> 0x7f97b02754f0 <ckpt_proc_lgs_cfg_v5(lgs_cb*, void*)>} >>>>>>>>>>> ---Type <return> to continue, or q <return> to quit--- >>>>>>>>>>> #10 0x00007f97b024d9f2 in main () at >> src/log/logd/lgs_main.cc:583 >>>>>>>>>>> usr1_sel_obj = {raise_obj = -1, rmv_obj = -1} >>>>>>>>>>> _lgs_cb = {mds_hdl = 65547, mds_role = >>>> V_DEST_RL_STANDBY, >>>>>>>>> vaddr >>>>>>>>>>> = 11, log_version = {releaseCode = 65 'A', majorVersion = 2 >>> '\002', >>>>>>>>>>> minorVersion = 2 '\002'}, client_tree = { >>>>>>>>>>> root_node = {bit = -1, left = 0x7f97b04cf1b0, >>> right = >>>>>>>>>>> 0x7f97b04a2418, key_info = 0x7f97b04b7bd0 ""}, params = >>>> {key_size >>>>>>>>>>> = 4}, >>>>>>>>>>> n_nodes = 8}, comp_name = {_opaque = {46, 24947, >>>>>>>>>>> 17254, 28015, 15728, 20300, 11335, 24947, > 21350, >>>>>>>>>>> 15733, >>>>>>>>>>> 17235, 12845, 29484, 26209, 26451, 12861, 11342, 24947, >> 16742, >>>>>> 28784, >>>>>>>>>>> 20285, 25968, 21358, 17985, >>>>>>>>>>> 0 <repeats 105 times>}}, amf_hdl = > 4288675841, >>>>>>>>>>> amfSelectionObject = 15, amf_invocation_id = 0, is_quiesced_set >> = >>>>>>>> false, >>>>>>>>>>> immOiHandle = 554050912783, immSelectionObject = 21, >>>>>>>>>>> clmSelectionObject = 17, clm_hdl = 4279238657, >>>>>>>>>>> ha_state = >>>>>>>>>>> SA_AMF_HA_STANDBY, last_client_id = 208, async_upd_cnt = >> 743, >>>>>>>>>>> ckpt_state >>>>>>>>>>> = COLD_SYNC_IDLE, mbcsv_hdl = 4293918753, >>>>>>>>>>> mbcsv_sel_obj = 23, mbcsv_ckpt_hdl = 4292870177, >>>>>>>>>>> mbcsv_peer_version = 7, edu_hdl = {is_inited = true, tree = >>>>>>>>>>> {root_node >>>>>>>> = >>>>>>>>>>> {bit = -1, left = 0x7f97b04cf2e0, >>>>>>>>>>> right = 0x7f97b04a25b8, key_info = >>> 0x7f97b04b7d40 >>>>>>>> ""}, >>>>>>>>>>> params = {key_size = 8}, n_nodes = 12}, to_version = 1}, >>>>>>>>>>> fully_initialized = true, lga_down_list_head = 0x0, >>>>>>>>>>> lga_down_list_tail = 0x0, clm_init_sel_obj = >>>>>>>>>>> {raise_obj = >>>>>>>> -1, >>>>>>>>>>> rmv_obj = -1}, nid_started = true, scAbsenceAllowed = 900, >>>>>>>>>>> lgs_recovery_state = LGS_NORMAL} >>>>>>>>>>> nfds = 7 >>>>>>>>>>> fds = {{fd = 19, events = 1, revents = 0}, {fd = > 15, >>>>>>>>>>> events = >>>>>>>>>>> 1, revents = 0}, {fd = 23, events = 1, revents = 1}, {fd = 13, >>>>>>>>>>> events = >>>>>>>>>>> 1, revents = 0}, {fd = -1, events = 1, >>>>>>>>>>> revents = 0}, {fd = 17, events = 1, revents = > 0}, >>>>>>>>>>> {fd = >>>>>>>> 21, >>>>>>>>>>> events = 1, revents = 0}} >>>>>>>>>>> mbox_msgs = {0, 0, 0, 0, 0} >>>>>>>>>>> lgs_cb = 0x7f97b04a2400 >>>>>>>>>>> mbox_low = {0, 0, 0, 0, 0} >>>>>>>>>>> lgs_mbox_init_mutex = {__data = {__lock = 0, > __count = >>> 0, >>>>>>>>>>> __owner = 0, __nusers = 0, __kind = 0, __spins = 0, __list = >>>>>>>>>>> {__prev = >>>>>>>>>>> 0x0, __next = 0x0}}, >>>>>>>>>>> __size = '\000' <repeats 39 times>, __align = 0} >>>>>>>>>>> lgs_mbx = 4291821569 >>>>>>>>>>> mbox_high = {0, 0, 0, 0, 0} >>>>>>>>>>> mbox_full = {false, false, false, false, false} >>>>>>>>>>> (gdb) >>>>>>>>>>> (gdb) >>>>>>>>>>> >>>>>>>>>>> Feb 22 13:40:55 SC-2 osafimmnd[4584]: NO Ccb 131 >> COMMITTED >>>>>>>>>>> (immcfg_SC-1_18714) >>>>>>>>>>> Feb 22 13:40:56 SC-2 osafamfnd[4634]: NO >>>>>>>>>>> 'safComp=LOG,safSu=SC-2,safSg=2N,safApp=OpenSAF' faulted >> due >>>> to >>>>>>>>>>> 'avaDown' : Recovery is 'nodeFailfast' >>>>>>>>>>> Feb 22 13:40:56 SC-2 osafamfnd[4634]: ER >>>>>>>>>>> safComp=LOG,safSu=SC-2,safSg=2N,safApp=OpenSAF Faulted >> due >>>>>>>>>>> to:avaDown >>>>>>>>>>> Recovery is:nodeFailfast >>>>>>>>>>> Feb 22 13:40:56 SC-2 osafamfnd[4634]: Rebooting OpenSAF >> NodeId >>>> = >>>>>>>>>>> 131599 >>>>>>>>>>> EE Name = , Reason: Component faulted: recovery is node >> failfast, >>>>>>>>>>> OwnNodeId = 131599, SupervisionTime = 60 >>>>>>>>>>> Feb 22 13:40:56 SC-2 opensaf_reboot: Rebooting local node; >>>>>> timeout=60 >> ============================================================== >> ============================================================== >>>>>>>>>>> On 2/22/2017 12:23 PM, A V Mahesh wrote: >>>>>>>>>>>> Hi Vu, >>>>>>>>>>>> >>>>>>>>>>>> On 2/22/2017 12:19 PM, Vu Minh Nguyen wrote: >>>>>>>>>>>>> [Vu] I has sent you 02 patches. There is code change in >> osaflogd >>>>>>>>>>>>> code >>>>>>>>>> that >>>>>>>>>>>>> fix the coredump you have observed. >>>>>>>>>>>>> The other one is test code that fix the logtest coredump. >>>>>>>>>>>> Ok I will re-test , and update you . >>>>>>>>>>>> >>>>>>>>>>>> -AVM >>>>>>>>>>>> >>>>>>>>>>>> On 2/22/2017 12:19 PM, Vu Minh Nguyen wrote: >>>>>>>>>>>>> Hi Mahehs, >>>>>>>>>>>>> >>>>>>>>>>>>> See my reply inline, [Vu]. >>>>>>>>>>>>> >>>>>>>>>>>>> Regards, Vu >>>>>>>>>>>>> >>>>>>>>>>>>>> -----Original Message----- >>>>>>>>>>>>>> From: A V Mahesh [mailto:mahesh.va...@oracle.com] >>>>>>>>>>>>>> Sent: Wednesday, February 22, 2017 1:36 PM >>>>>>>>>>>>>> To: Vu Minh Nguyen <vu.m.ngu...@dektech.com.au>; >>>>>>>>>>>>>> lennart.l...@ericsson.com; canh.v.tru...@dektech.com.au >>>>>>>>>>>>>> Cc: opensaf-devel@lists.sourceforge.net >>>>>>>>>>>>>> Subject: Re: [PATCH 0 of 3] Review Request for log: add >>>>>>>>>>>>>> alternative >>>>>>>>>>>>>> destinations of log records [#2258] V4 >>>>>>>>>>>>>> >>>>>>>>>>>>>> Hi Vu, >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> On 2/22/2017 11:52 AM, Vu Minh Nguyen wrote: >>>>>>>>>>>>>>> Hi Mahesh, >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> Have a code fault in uml test, and other one in checkpoint. >>>>>>>>>>>>>> [AVM] This is Normal Suse 11 VM ( not UML). >>>>>>>>>>>>>>> I have just updated the code. Please re-apply for #3 and #4 >>>>>>>> patches. >>>>>>>>>>>>>> [AVM] is these new patch has function changes or only test >>>> code >>>>>>>>>>> changes ? >>>>>>>>>>>>> [Vu] I has sent you 02 patches. There is code change in >> osaflogd >>>>>>>>>>>>> code >>>>>>>>>> that >>>>>>>>>>>>> fix the coredump you have observed. >>>>>>>>>>>>> The other one is test code that fix the logtest coredump. >>>>>>>>>>>>>>> Note that, test case #14 of suite 17 should be run on active >>>>>>>>>>>>>>> node, >>>>>>>>>>>>>> otherwise >>>>>>>>>>>>>>> getting failed. >>>>>>>>>>>>>> [AVM] Segmentation fault of /usr/bin/logtest Not a big > issue >>> , >>>>>>>>>>>>>> we need to debug why osaflogd core dumped and it is >>>>>>>>>>>>>> critical >>>>>>>>>>>>> [Vu] I found the problem. You can try with the new one to see >> if >>>>>>>>>>>>> the >>>>>>>>>>>>> coredump is still there or not. >>>>>>>>>>>>>>> I will put condition check to that test case later. >>>>>>>>>>>>>> -AVM >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>>> Regards, Vu >>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> -----Original Message----- >>>>>>>>>>>>>>>> From: A V Mahesh [mailto:mahesh.va...@oracle.com] >>>>>>>>>>>>>>>> Sent: Wednesday, February 22, 2017 12:16 PM >>>>>>>>>>>>>>>> To: Vu Minh Nguyen <vu.m.ngu...@dektech.com.au>; >>>>>>>>>>>>>>>> lennart.l...@ericsson.com; >> canh.v.tru...@dektech.com.au >>>>>>>>>>>>>>>> Cc: opensaf-devel@lists.sourceforge.net >>>>>>>>>>>>>>>> Subject: Re: [PATCH 0 of 3] Review Request for log: add >>>>>>>> alternative >>>>>>>>>>>>>>>> destinations of log records [#2258] V4 >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> Hi Vu, >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> Thanks , >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> While testing /usr/bin/logtest , SC-2 standby osaflogd >> core >>>>>>>>> dumped >>>>>>>>>>>>> and >>>>>>>>>>>>>>>> /usr/bin/logtest on SC-1 Active >>>>>>>>>>>>>>>> got Segmentation fault , am I missing any other patch ( i >> am >>>>>>>> using >>>>>>>>>>>>>>>> devel published patch only ) >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> Following patches i am using : >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> 1) #2293 (sent by Anders Widel, but not yet > pushed) >>>>>>>>>>>>>>>> 2) #2258 (v2, sent by Lennart, but not yet > pushed >>> yet) >>>>>>>>>>>>>>>> 3) #2258 (v4, sent by Vu, but not yet pushed > yet) >>>>>>>>>>>>>>>> >> ============================================================== >>>>>>>>>>>>>>>> ======================================== >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> Core was generated by `/usr/lib64/opensaf/osaflogd'. >>>>>>>>>>>>>>>> Program terminated with signal 11, Segmentation fault. >>>>>>>>>>>>>>>> #0 ckpt_proc_cfg_stream(lgs_cb*, void*) () at >>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc:2195 >>>>>>>>>>>>>>>> 2195 src/log/logd/lgs_mbcsv.cc: No such file or >> directory. >>>>>>>>>>>>>>>> in src/log/logd/lgs_mbcsv.cc >>>>>>>>>>>>>>>> (gdb) bt >>>>>>>>>>>>>>>> #0 ckpt_proc_cfg_stream(lgs_cb*, void*) () at >>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc:2195 >>>>>>>>>>>>>>>> #1 0x00007f12c3e22960 in >> ckpt_decode_log_struct(lgs_cb*, >>>>>>>>>>>>>>>> ncs_mbcsv_cb_arg*, void*, void*, unsigned int >>>>>> (*)(edu_hdl_tag*, >>>>>>>>>>>>>>>> edu_tkn_tag*, void*, unsigned int*, edu_buf_env_tag*, >>>>>>>>>>> EDP_OP_TYPE, >>>>>>>>>>>>>>>> EDU_ERR*)) () at src/log/logd/lgs_mbcsv.cc:950 >>>>>>>>>>>>>>>> #2 0x00007f12c3e240dc in >>>>>> ckpt_decode_async_update(lgs_cb*, >>>>>>>>>>>>>>>> ncs_mbcsv_cb_arg*) () at src/log/logd/lgs_mbcsv.cc:1086 >>>>>>>>>>>>>>>> #3 0x00007f12c3e26941 in >>>>>> mbcsv_callback(ncs_mbcsv_cb_arg*) () >>>>>>>>> at >>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc:880 >>>>>>>>>>>>>>>> #4 0x00007f12c2f25596 in ncs_mbscv_rcv_decode () from >>>>>>>>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0 >>>>>>>>>>>>>>>> #5 0x00007f12c2f25766 in ncs_mbcsv_rcv_async_update >> () >>>>>> from >>>>>>>>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0 >>>>>>>>>>>>>>>> #6 0x00007f12c2f2c370 in mbcsv_process_events () from >>>>>>>>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0 >>>>>>>>>>>>>>>> #7 0x00007f12c2f2c4db in mbcsv_hdl_dispatch_all () from >>>>>>>>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0 >>>>>>>>>>>>>>>> #8 0x00007f12c2f26ce2 in >> mbcsv_process_dispatch_request >>>> () >>>>>> at >>>>>>>>>>>>>>>> src/mbc/mbcsv_api.c:423 >>>>>>>>>>>>>>>> #9 0x00007f12c3e2396e in lgs_mbcsv_dispatch(unsigned >> int) >>>> () >>>>>> at >>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc:327 >>>>>>>>>>>>>>>> #10 0x00007f12c3e009f2 in main () at >>>>>>>>>>>>>>>> src/log/logd/lgs_main.cc:583 >>>>>>>>>>>>>>>> (gdb) >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >> ============================================================== >>>>>>>>>>>>>>>> ======================================== >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> Feb 22 10:37:06 SC-1 osafimmnd[4020]: NO Invalid error >>>>>> reported >>>>>>>>>>>>>>>> implementer 'safLogService', Ccb 161 will be aborted >>>>>>>>>>>>>>>> Feb 22 10:37:06 SC-1 osafimmnd[4020]: NO Ccb 161 >> aborted >>>> in >>>>>>>>>>>>>> COMPLETED >>>>>>>>>>>>>>>> processing (validation) >>>>>>>>>>>>>>>> Feb 22 10:37:06 SC-1 osafimmnd[4020]: NO Ccb 161 >>>> ABORTED >>>>>>>>>>>>>> (immcfg_SC- >>>>>>>>>>>>>>>> 1_5394) >>>>>>>>>>>>>>>> Add values Fail >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> Program received signal SIGSEGV, Segmentation fault. >>>>>>>>>>>>>>>> 0x000055555556929a in read_and_compare.isra.7 () at >>>>>>>>>>>>>>>> src/log/apitest/tet_LogOiOps.c:1891 >>>>>>>>>>>>>>>> 1891 src/log/apitest/tet_LogOiOps.c: No such file or >>>>>>>>>>>>>>>> directory. >>>>>>>>>>>>>>>> in src/log/apitest/tet_LogOiOps.c >>>>>>>>>>>>>>>> (gdb) Feb 22 10:37:07 SC-1 sshd[5298]: Accepted keyboard- >>>>>>>>>>>>>> interactive/pam >>>>>>>>>>>>>>>> for root from 10.176.178.22 port 51945 ssh2 >>>>>>>>>>>>>>>> bt >>>>>>>>>>>>>>>> #0 0x000055555556929a in read_and_compare.isra.7 () at >>>>>>>>>>>>>>>> src/log/apitest/tet_LogOiOps.c:1891 >>>>>>>>>>>>>>>> #1 0x0000555555569bbb in >>>>>>>>>>>>>>>> check_logRecordDestinationConfigurationEmpty >>>>>>>>>>>>>>>> () at src/log/apitest/tet_LogOiOps.c:2179 >>>>>>>>>>>>>>>> #2 0x0000555555573495 in run_test_case () >>>>>>>>>>>>>>>> #3 0x0000555555573934 in test_run () >>>>>>>>>>>>>>>> #4 0x000055555555c7cd in main () at >>>>>>>>>>>>>>>> src/log/apitest/logtest.c:569 >>>>>>>>>>>>>>>> (gdb) >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >> ============================================================== >>>>>>>>>>>>>>>> ======================================== >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> -AVM >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> On 2/22/2017 9:48 AM, Vu Minh Nguyen wrote: >>>>>>>>>>>>>>>>> Hi Mahesh, >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> I send them in attachment instead, and name them in the >>>>>> order. >>>>>>>>>>>>>>>>> I just pull the latest code, and apply them without > getting >>>>> any >>>>>>>>>> hunk >>>>>>>>>>>>>>> error. >>>>>>>>>>>>>>>>> Please try with them, and let me know if you see any >>>> problem. >>>>>>>>>>>>>>>>> Regards, Vu >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> -----Original Message----- >>>>>>>>>>>>>>>>>> From: A V Mahesh [mailto:mahesh.va...@oracle.com] >>>>>>>>>>>>>>>>>> Sent: Wednesday, February 22, 2017 11:09 AM >>>>>>>>>>>>>>>>>> To: Vu Minh Nguyen <vu.m.ngu...@dektech.com.au>; >>>>>>>>>>>>>>>>>> lennart.l...@ericsson.com; >>>> canh.v.tru...@dektech.com.au >>>>>>>>>>>>>>>>>> Cc: opensaf-devel@lists.sourceforge.net >>>>>>>>>>>>>>>>>> Subject: Re: [PATCH 0 of 3] Review Request for log: add >>>>>>>>>> alternative >>>>>>>>>>>>>>>>>> destinations of log records [#2258] V4 >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> Hi Vu, >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> I did follow that still i get Hunk #2 FAILED even on >>> today's >>>>>>>>>> staging >> ============================================================== >>>>>>>>>>>>>>>>>> ================== >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> [root@dhcp-hyd-scp-5fl-10-176-177-96 staging]# patch >> - >>>> p1 >>>>>>>>> <2293 >>>>>>>>>>>>>>>>>> patching file src/base/Makefile.am >>>>>>>>>>>>>>>>>> Hunk #1 succeeded at 33 (offset 1 line). >>>>>>>>>>>>>>>>>> Hunk #3 succeeded at 183 (offset 1 line). >>>>>>>>>>>>>>>>>> patching file src/base/file_descriptor.cc >>>>>>>>>>>>>>>>>> patching file src/base/file_descriptor.h >>>>>>>>>>>>>>>>>> patching file src/base/tests/unix_socket_test.cc >>>>>>>>>>>>>>>>>> patching file src/base/unix_client_socket.cc >>>>>>>>>>>>>>>>>> patching file src/base/unix_server_socket.cc >>>>>>>>>>>>>>>>>> patching file src/base/unix_socket.cc >>>>>>>>>>>>>>>>>> patching file src/base/unix_socket.h >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> [root@dhcp-hyd-scp-5fl-10-176-177-96 staging]# patch >> - >>>> p1 >>>>>>>>> <2258- >>>>>>>>>>> 1 >>>>>>>>>>>>>>>>>> patching file src/log/Makefile.am >>>>>>>>>>>>>>>>>> Hunk #1 succeeded at 71 (offset -1 lines). >>>>>>>>>>>>>>>>>> patching file src/log/config/logsv_classes.xml >>>>>>>>>>>>>>>>>> Hunk #1 FAILED at 147. >>>>>>>>>>>>>>>>>> 1 out of 1 hunk FAILED -- saving rejects to file >>>>>>>>>>>>>>>>>> src/log/config/logsv_classes.xml.rej >>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_config.cc >>>>>>>>>>>>>>>>>> Hunk #1 succeeded at 35 (offset -5 lines). >>>>>>>>>>>>>>>>>> Hunk #2 FAILED at 705. >>>>>>>>>>>>>>>>>> Hunk #3 FAILED at 971. >>>>>>>>>>>>>>>>>> 2 out of 3 hunks FAILED -- saving rejects to file >>>>>>>>>>>>>>>>>> src/log/logd/lgs_config.cc.rej >>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_config.h >>>>>>>>>>>>>>>>>> Hunk #1 FAILED at 304. >>>>>>>>>>>>>>>>>> 1 out of 1 hunk FAILED -- saving rejects to file >>>>>>>>>>>>>>>>>> src/log/logd/lgs_config.h.rej >>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_dest.cc >>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_dest.h >>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_evt.cc >>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_imm.cc >>>>>>>>>>>>>>>>>> Hunk #1 FAILED at 45. >>>>>>>>>>>>>>>>>> Hunk #2 succeeded at 235 (offset -1 lines). >>>>>>>>>>>>>>>>>> Hunk #3 FAILED at 877. >>>>>>>>>>>>>>>>>> Hunk #4 succeeded at 1273 (offset -20 lines). >>>>>>>>>>>>>>>>>> Hunk #5 succeeded at 1404 (offset -1 lines). >>>>>>>>>>>>>>>>>> Hunk #6 succeeded at 1449 (offset -20 lines). >>>>>>>>>>>>>>>>>> Hunk #7 succeeded at 2032 (offset -1 lines). >>>>>>>>>>>>>>>>>> Hunk #8 FAILED at 2181. >>>>>>>>>>>>>>>>>> Hunk #9 succeeded at 2271 (offset -54 lines). >>>>>>>>>>>>>>>>>> Hunk #10 succeeded at 2387 (offset -1 lines). >>>>>>>>>>>>>>>>>> Hunk #11 succeeded at 2377 (offset -54 lines). >>>>>>>>>>>>>>>>>> Hunk #12 succeeded at 2478 (offset -1 lines). >>>>>>>>>>>>>>>>>> Hunk #13 succeeded at 2684 (offset -54 lines). >>>>>>>>>>>>>>>>>> Hunk #14 succeeded at 2821 (offset -1 lines). >>>>>>>>>>>>>>>>>> 3 out of 14 hunks FAILED -- saving rejects to file >>>>>>>>>>>>>>>>>> src/log/logd/lgs_imm.cc.rej >>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_main.cc >>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_mbcsv.cc >>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_mbcsv.h >>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_mbcsv_v5.cc >>>>>>>>>>>>>>>>>> Hunk #3 succeeded at 133 (offset -1 lines). >>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_mbcsv_v7.cc >>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_mbcsv_v7.h >>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_stream.cc >>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_stream.h >>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_util.cc >>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_util.h >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> [root@dhcp-hyd-scp-5fl-10-176-177-96 staging]# patch >> - >>>> p1 >>>>>>>>> <2258- >>>>>>>>>>> 2 >>>>>>>>>>>>>>>>>> patching file src/log/Makefile.am >>>>>>>>>>>>>>>>>> Hunk #1 succeeded at 180 (offset -3 lines). >>>>>>>>>>>>>>>>>> patching file src/log/apitest/tet_LogOiOps.c >>>>>>>>>>>>>>>>>> Hunk #1 FAILED at 1923. >>>>>>>>>>>>>>>>>> Hunk #2 FAILED at 1979. >>>>>>>>>>>>>>>>>> Hunk #3 FAILED at 2067. >>>>>>>>>>>>>>>>>> Hunk #4 FAILED at 2094. >>>>>>>>>>>>>>>>>> 4 out of 4 hunks FAILED -- saving rejects to file >>>>>>>>>>>>>>>>>> src/log/apitest/tet_LogOiOps.c.rej >>>>>>>>>>>>>>>>>> patching file src/log/apitest/tet_cfg_destination.c >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> [root@dhcp-hyd-scp-5fl-10-176-177-96 staging]# patch >> - >>>> p1 >>>>>>>>> <2258- >>>>>>>>>>> 3 >>>>>>>>>>>>>>>>>> patching file src/log/Makefile >>>>>>>>>>>>>>>>>> patching file src/log/Makefile.am >>>>>>>>>>>>>>>>>> Hunk #1 succeeded at 80 (offset -1 lines). >>>>>>>>>>>>>>>>>> Hunk #2 succeeded at 217 (offset -2 lines). >>>>>>>>>>>>>>>>>> patching file src/log/tests/Makefile >>>>>>>>>>>>>>>>>> patching file src/log/tests/lgs_dest_test.cc >>>>>>>>>>>>>>>>>> [root@dhcp-hyd-scp-5fl-10-176-177-96 staging]# vi >>>>>>>>>>>>>>>>>> src/log/apitest/tet_LogOiOps.c.rej >>>>>>>>>>>>>>>>>> [root@dhcp-hyd-scp-5fl-10-176-177-96 staging]# >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> >> ============================================================== >>>>>>>>>>>>>>>>>> ======================== >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> -AVM >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> On 2/21/2017 3:53 PM, Vu Minh Nguyen wrote: >>>>>>>>>>>>>>>>>>> Hi Mahesh, >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> As I has mentioned in below: >>>>>>>>>>>>>>>>>>>>> To run the test, this patch has dependent > on >>>>>>>> following >>>>>>>>>>>>> patches: >>>>>>>>>>>>>>>>>>>>> 1) #2293 (sent by Anders Widel, but not yet >>>>>>>>>>>>>>>>>>>>> pushed) >>>>>>>>>>>>>>>>>>>>> 2) #2258 (v2, sent by Lennart, but not yet >>> pushed >>>>>>>> yet) >>>>>>>>>>>>>>>>>>> So, you need to apply #2293 first, then #2258 which >> sent >>>> by >>>>>>>>>>> Lennart >>>>>>>>>>>>>>>>>>> yesterday, then mine. >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> Regards, Vu >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> -----Original Message----- >>>>>>>>>>>>>>>>>>>> From: A V Mahesh [mailto:mahesh.va...@oracle.com] >>>>>>>>>>>>>>>>>>>> Sent: Tuesday, February 21, 2017 5:10 PM >>>>>>>>>>>>>>>>>>>> To: Vu Minh Nguyen >> <vu.m.ngu...@dektech.com.au>; >>>>>>>>>>>>>>>>>>>> lennart.l...@ericsson.com; >>>>>> canh.v.tru...@dektech.com.au >>>>>>>>>>>>>>>>>>>> Cc: opensaf-devel@lists.sourceforge.net >>>>>>>>>>>>>>>>>>>> Subject: Re: [PATCH 0 of 3] Review Request for log: >> add >>>>>>>>>>> alternative >>>>>>>>>>>>>>>>>>>> destinations of log records [#2258] V4 >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> Hi Vu, >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> Is this applies on top of log #2146 - V4 , I see >>> both >>>>>>>>>> #tickets >>>>>>>>>>>>>>>>> has >>>>>>>>>>>>>>>>>>>> version changes ? >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> in which order i need to apply ( #2146 & #2258 ) or >>>>>>>>>>>>>>>>>>>> (#2258 >>>>>>>> & >>>>>>>>>>>>>>>>> #2146). >>>>>> ========================================================= >>>>>>>>>>>>>>>>>>>> patching file src/log/Makefile.am >>>>>>>>>>>>>>>>>>>> Hunk #1 FAILED at 72. >>>>>>>>>>>>>>>>>>>> Hunk #2 FAILED at 120. >>>>>>>>>>>>>>>>>>>> 2 out of 2 hunks FAILED -- saving rejects to file >>>>>>>>>>>>>>>>> src/log/Makefile.am.rej >>>>>>>>>>>>>>>>>>>> patching file src/log/config/logsv_classes.xml >>>>>>>>>>>>>>>>>>>> Hunk #1 FAILED at 147. >>>>>>>>>>>>>>>>>>>> 1 out of 1 hunk FAILED -- saving rejects to file >>>>>>>>>>>>>>>>>>>> src/log/config/logsv_classes.xml.rej >>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_config.cc >>>>>>>>>>>>>>>>>>>> Hunk #1 succeeded at 35 (offset -5 lines). >>>>>>>>>>>>>>>>>>>> Hunk #2 FAILED at 705. >>>>>>>>>>>>>>>>>>>> Hunk #3 FAILED at 971. >>>>>>>>>>>>>>>>>>>> 2 out of 3 hunks FAILED -- saving rejects to file >>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_config.cc.rej >>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_config.h >>>>>>>>>>>>>>>>>>>> Hunk #1 FAILED at 304. >>>>>>>>>>>>>>>>>>>> 1 out of 1 hunk FAILED -- saving rejects to file >>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_config.h.rej >>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_dest.cc >>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_dest.h >>>>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_evt.cc >>>>>>>>>>>>>>>>>>>> Hunk #1 FAILED at 1. >>>>>>>>>>>>>>>>>>>> Hunk #2 succeeded at 30 with fuzz 2 (offset 2 lines). >>>>>>>>>>>>>>>>>>>> Hunk #3 succeeded at 1282 (offset 45 lines). >>>>>>>>>>>>>>>>>>>> Hunk #4 succeeded at 1300 (offset 2 lines). >>>>>>>>>>>>>>>>>>>> 1 out of 4 hunks FAILED -- saving rejects to file >>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_evt.cc.rej >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> >> ============================================================== >>>>>>>>>>>>>>>>>>>> === >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> -AVM >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> On 2/21/2017 3:03 PM, Vu Minh Nguyen wrote: >>>>>>>>>>>>>>>>>>>>> Summary: log: add alternative destinations of log >>>> records >>>>>>>>>>> [#2258] >>>>>>>>>>>>>>>>>>>>> Review request for Trac Ticket(s): #2258 >>>>>>>>>>>>>>>>>>>>> Peer Reviewer(s): Lennart, Canh, Mahesh >>>>>>>>>>>>>>>>>>>>> Pull request to: <<LIST THE PERSON WITH PUSH >>>> ACCESS >>>>>>>>> HERE>> >>>>>>>>>>>>>>>>>>>>> Affected branch(es): Default >>>>>>>>>>>>>>>>>>>>> Development branch: Default >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> -------------------------------- >>>>>>>>>>>>>>>>>>>>> Impacted area Impact y/n >>>>>>>>>>>>>>>>>>>>> -------------------------------- >>>>>>>>>>>>>>>>>>>>> Docs n >>>>>>>>>>>>>>>>>>>>> Build system n >>>>>>>>>>>>>>>>>>>>> RPM/packaging n >>>>>>>>>>>>>>>>>>>>> Configuration files n >>>>>>>>>>>>>>>>>>>>> Startup scripts n >>>>>>>>>>>>>>>>>>>>> SAF services n >>>>>>>>>>>>>>>>>>>>> OpenSAF services y >>>>>>>>>>>>>>>>>>>>> Core libraries n >>>>>>>>>>>>>>>>>>>>> Samples n >>>>>>>>>>>>>>>>>>>>> Tests y >>>>>>>>>>>>>>>>>>>>> Other n >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> Comments (indicate scope for each "y" above): >>>>>>>>>>>>>>>>>>>>> --------------------------------------------- >>>>>>>>>>>>>>>>>>>>> To run the test, this patch has dependent > on >>>>>>>> following >>>>>>>>>>>>> patches: >>>>>>>>>>>>>>>>>>>>> 1) #2293 (sent by Anders Widel, but not yet >>>>>>>>>>>>>>>>>>>>> pushed) >>>>>>>>>>>>>>>>>>>>> 2) #2258 (v2, sent by Lennart, but not yet >>> pushed >>>>>>>> yet) >>>>>>>>>>>>>>>>>>>>> changeset >>>>>> d74aaf3025c99cade3165a15831124548f4d85bd >>>>>>>>>>>>>>>>>>>>> Author: Vu Minh Nguyen >>>>>> <vu.m.ngu...@dektech.com.au> >>>>>>>>>>>>>>>>>>>>> Date: Wed, 15 Feb 2017 14:36:00 +0700 >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> log: add alternative destinations of log records >>>>>>>>>>>>>>>>>>>>> [#2258] >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> Here are major info, detailed info will be added > to >>>>>>>>>>>>>>>>>>>>> PR doc >>>>>>>>>>>>>>> soon. 1) >>>>>>>>>>>>>>>>>>>> Add >>>>>>>>>>>>>>>>>>>>> attribute "saLogRecordDestination" to log >> stream. >>> 2) >>>>>>>>>>>>>>>>>>>>> Add >>>>>>>>>>>>>>> Local >>>>>>>>>>>>>>>>>>>> socket >>>>>>>>>>>>>>>>>>>>> destintion handler 3) Integrate into first >>> increment >>>>>>>>>>>>>>>>>>>>> made by >>>>>>>>>>>>>>> Lennart >>>>>>>>>>>>>>>>>>>>> changeset >>>>>> 4bae27a478c235df3058f43c92d3a5483233b01d >>>>>>>>>>>>>>>>>>>>> Author: Vu Minh Nguyen >>>>>> <vu.m.ngu...@dektech.com.au> >>>>>>>>>>>>>>>>>>>>> Date: Wed, 15 Feb 2017 15:07:09 +0700 >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> log: add UML test case to verify alternative >>>>>>>>>>>>>>>>>>>>> destination >>>>>>>>>>>>>>> [#2258] >>>>>>>>>>>>>>>>>>>>> Major changes: 1) Modify Lennart's test cases >>>> because >>>>>>>>>>>>>>> enhancing >>>>>>>>>>>>>>>>>>>> destination >>>>>>>>>>>>>>>>>>>>> configuration validation rules. 2) Add test > suite >>>>>>>>>>>>>>>>>>>>> #17 to >>>>>>>>>>>>>>> verify >>>>>>>>>>>>>>>>>>>> alternative >>>>>>>>>>>>>>>>>>>>> destination >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> changeset >>>> bc375725fed22bb4f8cb3ae3df5f96fb9d281efb >>>>>>>>>>>>>>>>>>>>> Author: Vu Minh Nguyen >>>>>> <vu.m.ngu...@dektech.com.au> >>>>>>>>>>>>>>>>>>>>> Date: Thu, 16 Feb 2017 17:22:13 +0700 >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> log: add unit tests to verify interfaces > provided >>> by >>>>>>>>>>>>>>> destination >>>>>>>>>>>>>>>>>>>> handler >>>>>>>>>>>>>>>>>>>>> [#2258] >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> Unit tests to verify major interfaces: 1) >>>>>>>>>>>>>>>>>>>>> CfgDestination() >>>>>>>>>>>>>>> 2) >>>>>>>>>>>>>>>>>>>>> WriteToDestination() >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> Added Files: >>>>>>>>>>>>>>>>>>>>> ------------ >>>>>>>>>>>>>>>>>>>>> src/log/apitest/tet_cfg_destination.c >>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_dest.cc >>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_dest.h >>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv_v7.cc >>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv_v7.h >>>>>>>>>>>>>>>>>>>>> src/log/tests/lgs_dest_test.cc >>>>>>>>>>>>>>>>>>>>> src/log/tests/Makefile >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> Complete diffstat: >>>>>>>>>>>>>>>>>>>>> ------------------ >>>>>>>>>>>>>>>>>>>>> src/log/Makefile | 4 + >>>>>>>>>>>>>>>>>>>>> src/log/Makefile.am | 31 +++++- >>>>>>>>>>>>>>>>>>>>> src/log/apitest/tet_LogOiOps.c | 8 +- >>>>>>>>>>>>>>>>>>>>> src/log/apitest/tet_cfg_destination.c | 483 >> ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++ >>>>>>>>>>>>>>>>>>>> ++++++++++++++++++++++++++++++++++ >>>>>>>>>>>>>>>>>>>>> src/log/config/logsv_classes.xml | 7 +- >>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_config.cc | 169 >>>>>>>>>>>>>>>>>>>> +++++++++++++++++++++++++++++--- >>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_config.h | 3 +- >>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_dest.cc | 707 >> ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++ >> ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++ >>>>>>>>>>>>>>>>>>>> +++++++++++++++++ >>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_dest.h | 576 >> ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++ >>>>>>>>> ++++++++++++++++++++++++++++++++++++++++++++++++++++ >>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_evt.cc | 33 ++++++ >>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_imm.cc | 202 >>>>>>>>>>>>>>>>>>>> +++++++++++++++++++++++++++++++++------ >>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_main.cc | 8 + >>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc | 103 >>>>>>>>>>>>> ++++++++++++++++++- >>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv.h | 6 +- >>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv_v5.cc | 10 + >>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv_v7.cc | 177 >>>>>>>>>>>>>>>>>>>> +++++++++++++++++++++++++++++++++++ >>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv_v7.h | 67 >>>>>>>>>> +++++++++++++ >>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_stream.cc | 60 >>>>>>>>>> +++++++++++- >>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_stream.h | 16 +++ >>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_util.cc | 63 >>>>>>>>>> ++++++++++++ >>>>>>>>>>>>>>>>>>>>> src/log/logd/lgs_util.h | 11 +- >>>>>>>>>>>>>>>>>>>>> src/log/tests/Makefile | 20 +++ >>>>>>>>>>>>>>>>>>>>> src/log/tests/lgs_dest_test.cc | 209 >>>>>>>>>>>>>>>>>>>> +++++++++++++++++++++++++++++++++++++++++ >>>>>>>>>>>>>>>>>>>>> 23 files changed, 2896 insertions(+), 77 >>>>>>>>>>>>>>>>>>>>> deletions(-) >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> Testing Commands: >>>>>>>>>>>>>>>>>>>>> ----------------- >>>>>>>>>>>>>>>>>>>>> Run UML test suite #17 >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> Testing, Expected Results: >>>>>>>>>>>>>>>>>>>>> -------------------------- >>>>>>>>>>>>>>>>>>>>> All test passed >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> Conditions of Submission: >>>>>>>>>>>>>>>>>>>>> ------------------------- >>>>>>>>>>>>>>>>>>>>> <<HOW MANY DAYS BEFORE PUSHING, >>>> CONSENSUS >>>>>>>>> ETC>> >>>>>>>>>>>>>>>>>>>>> Arch Built Started Linux distro >>>>>>>>>>>>>>>>>>>>> ------------------------------------------- >>>>>>>>>>>>>>>>>>>>> mips n n >>>>>>>>>>>>>>>>>>>>> mips64 n n >>>>>>>>>>>>>>>>>>>>> x86 n n >>>>>>>>>>>>>>>>>>>>> x86_64 n n >>>>>>>>>>>>>>>>>>>>> powerpc n n >>>>>>>>>>>>>>>>>>>>> powerpc64 n n >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> Reviewer Checklist: >>>>>>>>>>>>>>>>>>>>> ------------------- >>>>>>>>>>>>>>>>>>>>> [Submitters: make sure that your review doesn't >> trigger >>>>>> any >>>>>>>>>>>>>>>>>> checkmarks!] >>>>>>>>>>>>>>>>>>>>> Your checkin has not passed review because (see >>>> checked >>>>>>>>>>> entries): >>>>>>>>>>>>>>>>>>>>> ___ Your RR template is generally incomplete; it has >>>> too >>>>>>>>> many >>>>>>>>>>>>> blank >>>>>>>>>>>>>>>>>>> entries >>>>>>>>>>>>>>>>>>>>> that need proper data filled in. >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> ___ You have failed to nominate the proper persons >> for >>>>>>>>> review >>>>>>>>>>> and >>>>>>>>>>>>>>>>> push. >>>>>>>>>>>>>>>>>>>>> ___ Your patches do not have proper short+long >>>> header >>>>>>>>>>>>>>>>>>>>> ___ You have grammar/spelling in your header that >> is >>>>>>>>>>> unacceptable. >>>>>>>>>>>>>>>>>>>>> ___ You have exceeded a sensible line length in your >>>>>>>>>>>>>>>>>>>> headers/comments/text. >>>>>>>>>>>>>>>>>>>>> ___ You have failed to put in a proper Trac Ticket # >>> into >>>>>>>> your >>>>>>>>>>>>>>>>> commits. >>>>>>>>>>>>>>>>>>>>> ___ You have incorrectly put/left internal data in >> your >>>>>>>>>>>>>>> comments/files >>>>>>>>>>>>>>>>>>>>> (i.e. internal bug tracking tool IDs, product >>>>>>>> names >>>>>>>>>> etc) >>>>>>>>>>>>>>>>>>>>> ___ You have not given any evidence of testing >> beyond >>>>>> basic >>>>>>>>>>> build >>>>>>>>>>>>>>>>> tests. >>>>>>>>>>>>>>>>>>>>> Demonstrate some level of runtime or other sanity >>>>>>>>>> testing. >>>>>>>>>>>>>>>>>>>>> ___ You have ^M present in some of your files. >> These >>>>>>>>>>>>>>>>>>>>> have to >>>>>>>>>>> be >>>>>>>>>>>>>>>>>> removed. >>>>>>>>>>>>>>>>>>>>> ___ You have needlessly changed whitespace or >> added >>>>>>>>>>> whitespace >>>>>>>>>>>>>>>>>> crimes >>>>>>>>>>>>>>>>>>>>> like trailing spaces, or spaces before >>> tabs. >>>>>>>>>>>>>>>>>>>>> ___ You have mixed real technical changes with >>>>>> whitespace >>>>>>>>> and >>>>>>>>>>>>>> other >>>>>>>>>>>>>>>>>>>>> cosmetic code cleanup changes. These have to be >>>>>>>>>>> separate >>>>>>>>>>>>>>>>> commits. >>>>>>>>>>>>>>>>>>>>> ___ You need to refactor your submission into >> logical >>>>>>>>>>>>>>>>>>>>> chunks; >>>>>>>>>>>>> there >>>>>>>>>>>>>>> is >>>>>>>>>>>>>>>>>>>>> too much content into a single commit. >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> ___ You have extraneous garbage in your review >>>> (merge >>>>>>>>>>> commits >>>>>>>>>>>>>> etc) >>>>>>>>>>>>>>>>>>>>> ___ You have giant attachments which should never >>>> have >>>>>>>>> been >>>>>>>>>>>>>> sent; >>>>>>>>>>>>>>>>>>>>> Instead you should place your content in a public >>>>>>>>>> tree to >>>>>>>>>>>>> be >>>>>>>>>>>>>>>>>>> pulled. >>>>>>>>>>>>>>>>>>>>> ___ You have too many commits attached to an e- >> mail; >>>>>>>>> resend >>>>>>>>>>> as >>>>>>>>>>>>>>>>>> threaded >>>>>>>>>>>>>>>>>>>>> commits, or place in a public tree for a pull. >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> ___ You have resent this content multiple times >>>> without a >>>>>>>>> clear >>>>>>>>>>>>>>>>>>> indication >>>>>>>>>>>>>>>>>>>>> of what has changed between each > re-send. >>>>>>>>>>>>>>>>>>>>> ___ You have failed to adequately and individually >>>> address >>>>>>>> all >>>>>>>>>> of >>>>>>>>>>>>>>> the >>>>>>>>>>>>>>>>>>>>> comments and change requests that were proposed >> in >>>>>>>>>> the >>>>>>>>>>>>>>> initial >>>>>>>>>>>>>>>>>>> review. >>>>>>>>>>>>>>>>>>>>> ___ You have a misconfigured ~/.hgrc file (i.e. >>>> username, >>>>>>>>> email >>>>>>>>>>>>> etc) >>>>>>>>>>>>>>>>>>>>> ___ Your computer have a badly configured date and >>>>>> time; >>>>>>>>>>>>>> confusing >>>>>>>>>>>>>>>> the >>>>>>>>>>>>>>>>>>>>> the threaded patch review. >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> ___ Your changes affect IPC mechanism, and you >> don't >>>>>>>>> present >>>>>>>>>>> any >>>>>>>>>>>>>>>>>> results >>>>>>>>>>>>>>>>>>>>> for in-service upgradability test. >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> ___ Your changes affect user manual and >>>> documentation, >>>>>>>>> your >>>>>>>>>>>>>> patch >>>>>>>>>>>>>>>>>> series >>>>>>>>>>>>>>>>>>>>> do not contain the patch that updates > the >>>>>>>>>>>>>>>>>>>>> Doxygen >>>>>>>>>>> manual. > ---------------------------------------------------------------------------- >>>>>>>>>> -- >>>>>>>>>>>> Check out the vibrant tech community on one of the world's >> most >>>>>>>>>>>> engaging tech sites, SlashDot.org! http://sdm.link/slashdot >>>>>>>>>>>> _______________________________________________ >>>>>>>>>>>> Opensaf-devel mailing list >>>>>>>>>>>> Opensaf-devel@lists.sourceforge.net >>>>>>>>>>>> https://lists.sourceforge.net/lists/listinfo/opensaf-devel ------------------------------------------------------------------------------ Check out the vibrant tech community on one of the world's most engaging tech sites, SlashDot.org! http://sdm.link/slashdot _______________________________________________ Opensaf-devel mailing list Opensaf-devel@lists.sourceforge.net https://lists.sourceforge.net/lists/listinfo/opensaf-devel