Hi Vu, On 2/23/2017 10:20 AM, Vu Minh Nguyen wrote: > Hi Mahesh, > > Can you try with 2258_v7.patch I just sent to you? I striped changeset: 8610 of today's latest staging ( `hg strip 8610` which removed log: implement SaLogFilterSetCallbackT and version handling [#2146]) and try to apply your `2258_v7.patch`, it says `malformed patch at line 3324`.
-AVM > > I have pulled the latest code on OpenSAF 5.1 branch, re-created the cluster. > And it works with the case old active SC-1 (OpenSAF 5.1) and new standby > SC-2 (with 2258_v7.patch included in). > > To apply 2258_v7.patch, please do remove the just pushed ticket "log: > implement SaLogFilterSetCallbackT and version handling [#2146]" , > I have not rebased the code on that yet. > > Regards, Vu > >> -----Original Message----- >> From: A V Mahesh [mailto:mahesh.va...@oracle.com] >> Sent: Thursday, February 23, 2017 11:45 AM >> To: Vu Minh Nguyen <vu.m.ngu...@dektech.com.au>; >> lennart.l...@ericsson.com; canh.v.tru...@dektech.com.au >> Cc: opensaf-devel@lists.sourceforge.net >> Subject: Re: [devel] [PATCH 0 of 3] Review Request for log: add > alternative >> destinations of log records [#2258] V4 >> >> Hi Vu/Lennart, >> >> >> In broad WITHOUT the #2258 patch, the same code/setup working fine with >> 2 sc node (staging changeset: 8609 ), >> as soon as we apply `2258_v5.patch` V5 patch on staging (changeset: >> 8609 ) that you have provided yesterday, >> on one sc node and try to bring up that in to cluster (in-service test) >> we are observing the issue of new node (with #2258 patch) not joining >> cluster. >> >> ============================================================== >> ==================================================== >> eb 23 10:01:59 SC-1 osafimmnd[15279]: NO Implementer (applier) >> connected: 15 (@safAmfService2010f) <127, 2010f> >> Feb 23 10:01:59 SC-1 osaflogd[15289]: NO LOGSV_DATA_GROUPNAME not >> found >> Feb 23 10:01:59 SC-1 osaflogd[15289]: NO LOG root directory is: >> "/var/log/opensaf/saflog" >> Feb 23 10:01:59 SC-1 osaflogd[15289]: NO LOG data group is: "" >> Feb 23 10:01:59 SC-1 osaflogd[15289]: NO LGS_MBCSV_VERSION = 7 >> Feb 23 10:01:59 SC-1 osafamfnd[15329]: NO Assigned >> 'safSi=SC-2N,safApp=OpenSAF' STANDBY to 'safSu=SC- >> 1,safSg=2N,safApp=OpenSAF' >> Feb 23 10:01:59 SC-1 opensafd: OpenSAF(5.1.M0 - ) services successfully >> started >> Feb 23 10:01:59 SC-1 osafamfnd[15329]: NO >> 'safComp=LOG,safSu=SC-1,safSg=2N,safApp=OpenSAF' faulted due to >> 'avaDown' : Recovery is 'nodeFailfast' >> Feb 23 10:01:59 SC-1 osafamfnd[15329]: ER >> safComp=LOG,safSu=SC-1,safSg=2N,safApp=OpenSAF Faulted due >> to:avaDown >> Recovery is:nodeFailfast >> Feb 23 10:01:59 SC-1 osafamfnd[15329]: Rebooting OpenSAF NodeId = >> 131343 >> EE Name = , Reason: Component faulted: recovery is node failfast, >> OwnNodeId = 131343, SupervisionTime = 60 >> Feb 23 10:01:59 SC-1 opensaf_reboot: Rebooting local node; timeout=60 >> Feb 23 10:02:00 SC-1 osafimmnd[15279]: NO Implementer (applier) >> connected: 16 (@OpenSafImmReplicatorB) <144, 2010f> >> Feb 23 10:01:59 SC-1 opensaf_reboot: Rebooting local node; timeout=60 >> ============================================================== >> ==================================================== >> >> So it is evident that in-service upgrade part code of this need to be >> corrected. >> >> Please see my comments as [AVM] and let me know if you need some traces >> . >> >> If you're planing to prepare new V6 patch , please do prepare on top of >> today's latest staging. >> >> On 2/23/2017 9:33 AM, Vu Minh Nguyen wrote: >>> Hi Mahesh, >>> >>> I have done in-service upgrade/downgrade with following cases: >>> 1) New Active SC-1 (OpenSAF 5.2 with the attached patch) + old standby >> SC-2 >>> (OpenSAF 5.1) >>> --> Work fine >> [AVM] This is not a practical use cause of in-service upgrade , we can >> ignore this test further >>> 2) Old Active SC-1 (OpenSAF 5.1) + new standby SC-2 (with or without >>> attached patch) >>> --> SC-2 is restarted & not able to join the cluster. >> [AVM] This use cause/flow is we do get in in-service upgrade , so we >> need to address this. >>> I got following messages in syslog: >>> Feb 23 09:32:42 SC-2 user.notice opensafd: OpenSAF(5.2.M0 - >>> 8529:b5addd36e45d:default) services successfully started >>> Feb 23 09:32:43 SC-2 local0.warn osafntfimcnd[701]: WA >> ntfimcn_imm_init >>> saImmOiImplementerSet() returned SA_AIS_ERR_TIMEOUT (5) >>> Feb 23 09:32:45 SC-2 local0.warn osafntfimcnd[701]: WA >> ntfimcn_imm_init >>> saImmOiImplementerSet() returned SA_AIS_ERR_TIMEOUT (5) >>> Feb 23 09:32:47 SC-2 local0.warn osafntfimcnd[701]: WA >> ntfimcn_imm_init >>> saImmOiImplementerSet() returned SA_AIS_ERR_TIMEOUT (5) >>> Feb 23 09:32:49 SC-2 local0.warn osafntfimcnd[701]: WA >> ntfimcn_imm_init >>> saImmOiImplementerSet() returned SA_AIS_ERR_TIMEOUT (5) >>> Feb 23 09:32:50 SC-2 local0.err osafmsgnd[592]: ER >> saImmOiImplementerSet >>> FAILED:5 >>> Feb 23 09:32:50 SC-2 local0.err osafmsgnd[592]: ER >> saImmOiImplementerSet >>> FAILED:5 >>> Feb 23 09:32:50 SC-2 local0.notice osafamfnd[496]: NO >>> 'safSu=SC-2,safSg=NoRed,safApp=OpenSAF' component restart probation >> timer >>> started (timeout: 60000000000 ns) >>> Feb 23 09:32:50 SC-2 local0.notice osafamfnd[496]: NO Restarting a >> component >>> of 'safSu=SC-2,safSg=NoRed,safApp=OpenSAF' (comp restart count: 1) >>> Feb 23 09:32:50 SC-2 local0.notice osafamfnd[496]: NO >>> 'safComp=MQND,safSu=SC-2,safSg=NoRed,safApp=OpenSAF' faulted due >> to >>> 'avaDown' : Recovery is 'componentRestart' >>> Feb 23 09:32:50 SC-2 local0.info osafmsgnd[736]: mkfifo already exists: >>> /var/lib/opensaf/osafmsgnd.fifo File exists >>> >>> And sometimes, on active SC-1 (OpenSAF 5.1), the node is not able to up >>> because of following error: >>> >>> Feb 23 11:00:32 SC-1 local0.err osafclmna[406]: MDTM:TIPC Dsock Socket >>> creation failed in MDTM_INIT err :Address family not supported by >> protocol >>> Feb 23 11:00:32 SC-1 local0.err osafclmna[406]: ER ncs_agents_startup >> FAILED >> [AVM] No such issues ( with both TCP & TIPC) (staging changeset: 8609 > ) >>> Are you getting similar problem at your side? >>> Please note that, the problem is existed WITH or WITHOUT the #2258 >> patch. >> [AVM] No , problem only if we apply `2258_v5.patch` V5 patch on staging >> (changeset: 8609 ) >> try to bring up that node in to cluster. >> >> >> -AVM >> >>> I have informed this to IMM to have a look, not sure any problem with > MDS >>> layer or any problem with my environment setup. >>> In the meantime, please have a look at the updated patch, I will > continue >>> checking the problem. Will keep you updated. >> [AVM] I haven't seen any IMM problems >>> Regards, Vu >>> >>>> -----Original Message----- >>>> From: A V Mahesh [mailto:mahesh.va...@oracle.com] >>>> Sent: Wednesday, February 22, 2017 5:36 PM >>>> To: Vu Minh Nguyen <vu.m.ngu...@dektech.com.au>; >>>> lennart.l...@ericsson.com; canh.v.tru...@dektech.com.au >>>> Cc: opensaf-devel@lists.sourceforge.net >>>> Subject: Re: [devel] [PATCH 0 of 3] Review Request for log: add >>> alternative >>>> destinations of log records [#2258] V4 >>>> >>>> Please see correction New Standby SC-1 ( with patch ) >>>> >>>> -AVM >>>> >>>> On 2/22/2017 4:02 PM, A V Mahesh wrote: >>>>> Hi Vu, >>>>> >>>>> With this new patch , we have another issue : >>>>> >>>>> 1) standby Core by `/usr/lib64/opensaf/osaflogd' issue got resolved > . >>>>> 2) In-service upgrade is Not working , I have Old Active SC-2 ( with >>>>> out patch ) and New Standby SC-1 ( with patch ) >>>>> >>>>> the new New Standby SC-1 not joining the cluster ( in-service >>>>> upgrade failed ) >>>>> >>>>> New Standby SC-1 >>>>> >>>>> >> ============================================================== >>>> ====================================================== >>>>> Feb 22 15:53:05 SC-1 osafamfnd[15752]: NO >>>>> 'safSu=SC-1,safSg=NoRed,safApp=OpenSAF' Presence State >>>> INSTANTIATING >>>>> => INSTANTIATED >>>>> Feb 22 15:53:05 SC-1 osafamfnd[15752]: NO Assigning >>>>> 'safSi=NoRed4,safApp=OpenSAF' ACTIVE to >>>>> 'safSu=SC-1,safSg=NoRed,safApp=OpenSAF' >>>>> Feb 22 15:53:05 SC-1 osafamfnd[15752]: NO Assigned >>>>> 'safSi=NoRed4,safApp=OpenSAF' ACTIVE to >>>>> 'safSu=SC-1,safSg=NoRed,safApp=OpenSAF' >>>>> Feb 22 15:53:05 SC-1 osafsmfd[15889]: Started >>>>> Feb 22 15:53:05 SC-1 osafamfnd[15752]: NO >>>>> 'safSu=SC-1,safSg=2N,safApp=OpenSAF' Presence State INSTANTIATING >> => >>>>> INSTANTIATED >>>>> Feb 22 15:53:05 SC-1 osafamfnd[15752]: NO Assigning >>>>> 'safSi=SC-2N,safApp=OpenSAF' STANDBY to >>>>> 'safSu=SC-1,safSg=2N,safApp=OpenSAF' >>>>> Feb 22 15:53:05 SC-1 osafrded[15672]: NO RDE role set to STANDBY >>>>> Feb 22 15:53:05 SC-1 osafrded[15672]: NO Peer up on node 0x2020f >>>>> Feb 22 15:53:05 SC-1 osafrded[15672]: NO Got peer info request from >>>>> node 0x2020f with role ACTIVE >>>>> Feb 22 15:53:05 SC-1 osafrded[15672]: NO Got peer info response from >>>>> node 0x2020f with role ACTIVE >>>>> Feb 22 15:53:05 SC-1 osafimmd[15691]: NO MDS event from svc_id 24 >>>>> (change:5, dest:13) >>>>> Feb 22 15:53:05 SC-1 osafimmd[15691]: NO MDS event from svc_id 24 >>>>> (change:3, dest:13) >>>>> Feb 22 15:53:05 SC-1 osafimmd[15691]: NO MDS event from svc_id 24 >>>>> (change:5, dest:13) >>>>> Feb 22 15:53:05 SC-1 osafimmd[15691]: NO MDS event from svc_id 25 >>>>> (change:3, dest:567412424453430) >>>>> Feb 22 15:53:05 SC-1 osafimmd[15691]: NO MDS event from svc_id 25 >>>>> (change:3, dest:565213401202663) >>>>> Feb 22 15:53:05 SC-1 osafimmd[15691]: NO MDS event from svc_id 25 >>>>> (change:3, dest:566312912825221) >>>>> Feb 22 15:53:05 SC-1 osafimmd[15691]: NO MDS event from svc_id 25 >>>>> (change:3, dest:564113889574230) >>>>> Feb 22 15:53:05 SC-1 osafimmnd[15702]: NO Implementer (applier) >>>>> connected: 17 (@safAmfService2010f) <127, 2010f> >>>>> Feb 22 15:53:05 SC-1 osaflogd[15712]: NO LOGSV_DATA_GROUPNAME >>>> not found >>>>> Feb 22 15:53:05 SC-1 osaflogd[15712]: NO LOG root directory is: >>>>> "/var/log/opensaf/saflog" >>>>> Feb 22 15:53:05 SC-1 osaflogd[15712]: NO LOG data group is: "" >>>>> Feb 22 15:53:05 SC-1 osaflogd[15712]: NO LGS_MBCSV_VERSION = 7 >>>>> Feb 22 15:53:05 SC-1 osafamfnd[15752]: NO Assigned >>>>> 'safSi=SC-2N,safApp=OpenSAF' STANDBY to >>>>> 'safSu=SC-1,safSg=2N,safApp=OpenSAF' >>>>> Feb 22 15:53:05 SC-1 opensafd: OpenSAF(5.1.M0 - ) services >>>>> successfully started >>>>> Feb 22 15:53:05 SC-1 osafamfnd[15752]: NO >>>>> 'safComp=LOG,safSu=SC-1,safSg=2N,safApp=OpenSAF' faulted due to >>>>> 'avaDown' : Recovery is 'nodeFailfast' >>>>> Feb 22 15:53:05 SC-1 osafamfnd[15752]: ER >>>>> safComp=LOG,safSu=SC-1,safSg=2N,safApp=OpenSAF Faulted due >>>> to:avaDown >>>>> Recovery is:nodeFailfast >>>>> Feb 22 15:53:05 SC-1 osafamfnd[15752]: Rebooting OpenSAF NodeId = >>>>> 131343 EE Name = , Reason: Component faulted: recovery is node >>>>> failfast, OwnNodeId = 131343, SupervisionTime = 60 >>>>> Feb 22 15:53:05 SC-1 opensaf_reboot: Rebooting local node; >> timeout=60 >>>>> Feb 22 15:53:43 SC-1 syslog-ng[1171]: syslog-ng starting up; >>>>> version='2.0.9' >>>>> >>>>> >> ============================================================== >>>> ====================================================== >>>>> Old - Active - SC-2 >>>>> >>>>> >> ============================================================== >>>> ====================================================== >>>>> Feb 22 15:53:02 SC-2 osafimmnd[16359]: NO NODE STATE-> >>>>> IMM_NODE_R_AVAILABLE >>>>> Feb 22 15:53:02 SC-2 osafimmloadd: NO Sync starting >>>>> Feb 22 15:53:02 SC-2 osafimmloadd: IN Synced 390 objects in total >>>>> Feb 22 15:53:02 SC-2 osafimmnd[16359]: NO NODE STATE-> >>>>> IMM_NODE_FULLY_AVAILABLE 18511 >>>>> Feb 22 15:53:02 SC-2 osafimmnd[16359]: NO Epoch set to 3 in >> ImmModel >>>>> Feb 22 15:53:02 SC-2 osafimmd[16346]: NO ACT: New Epoch for >> IMMND >>>>> process at node 2020f old epoch: 2 new epoch:3 >>>>> Feb 22 15:53:02 SC-2 osafimmd[16346]: NO ACT: New Epoch for >> IMMND >>>>> process at node 2040f old epoch: 2 new epoch:3 >>>>> Feb 22 15:53:02 SC-2 osafimmd[16346]: NO ACT: New Epoch for >> IMMND >>>>> process at node 2030f old epoch: 2 new epoch:3 >>>>> Feb 22 15:53:02 SC-2 osafimmloadd: NO Sync ending normally >>>>> Feb 22 15:53:02 SC-2 osafimmd[16346]: NO ACT: New Epoch for >> IMMND >>>>> process at node 2010f old epoch: 0 new epoch:3 >>>>> Feb 22 15:53:02 SC-2 osafimmnd[16359]: NO SERVER STATE: >>>>> IMM_SERVER_SYNC_SERVER --> IMM_SERVER_READY >>>>> Feb 22 15:53:03 SC-2 osafamfd[16408]: NO Received node_up from >> 2010f: >>>>> msg_id 1 >>>>> Feb 22 15:53:03 SC-2 osafamfd[16408]: NO Node 'SC-1' joined the >> cluster >>>>> Feb 22 15:53:03 SC-2 osafimmnd[16359]: NO Implementer connected: >> 16 >>>>> (MsgQueueService131343) <0, 2010f> >>>>> Feb 22 15:53:03 SC-2 osafrded[16327]: NO Peer up on node 0x2010f >>>>> Feb 22 15:53:03 SC-2 osafrded[16327]: NO Got peer info request from >>>>> node 0x2010f with role STANDBY >>>>> Feb 22 15:53:03 SC-2 osafrded[16327]: NO Got peer info response from >>>>> node 0x2010f with role STANDBY >>>>> Feb 22 15:53:03 SC-2 osafimmd[16346]: NO MDS event from svc_id 24 >>>>> (change:5, dest:13) >>>>> Feb 22 15:53:03 SC-2 osafimmnd[16359]: NO Implementer (applier) >>>>> connected: 17 (@safAmfService2010f) <0, 2010f> >>>>> Feb 22 15:53:03 SC-2 osafamfd[16408]: NO Cluster startup is done >>>>> Feb 22 15:53:04 SC-2 osafimmnd[16359]: NO Implementer (applier) >>>>> connected: 18 (@OpenSafImmReplicatorB) <0, 2010f> >>>>> Feb 22 15:53:09 SC-2 osafdtmd[16304]: NO Lost contact with 'SC-1' >>>>> Feb 22 15:53:09 SC-2 osaffmd[16336]: NO Node Down event for node id >>>>> 2010f: >>>>> Feb 22 15:53:09 SC-2 osafimmd[16346]: NO MDS event from svc_id 24 >>>>> (change:6, dest:13) >>>>> Feb 22 15:53:09 SC-2 osafimmd[16346]: NO MDS event from svc_id 25 >>>>> (change:4, dest:564113889574230) >>>>> Feb 22 15:53:09 SC-2 osaffmd[16336]: NO Current role: ACTIVE >>>>> Feb 22 15:53:09 SC-2 osaffmd[16336]: Rebooting OpenSAF NodeId = >>>> 131343 >>>>> EE Name = , Reason: Received Node Down for peer controller, >> OwnNodeId >>>>> = 131599, SupervisionTime = 60 >>>>> Feb 22 15:53:09 SC-2 osafrded[16327]: NO Peer down on node 0x2010f >>>>> Feb 22 15:53:09 SC-2 osafamfd[16408]: NO Node 'SC-1' left the cluster >>>>> Feb 22 15:53:09 SC-2 osaflogd[16372]: NO Failed (2) to send of WRITE >>>>> ack to: 2010f00003d6a >>>>> Feb 22 15:53:09 SC-2 osaflogd[16372]: NO Failed (2) to send of WRITE >>>>> ack to: 2010f00003d6a >>>>> Feb 22 15:53:09 SC-2 osaflogd[16372]: NO Failed (2) to send of WRITE >>>>> ack to: 2010f00003d74 >>>>> Feb 22 15:53:09 SC-2 osafclmd[16398]: NO Node 131343 went down. >> Not >>>>> sending track callback for agents on that node >>>>> Feb 22 15:53:09 SC-2 osafclmd[16398]: NO Node 131343 went down. >> Not >>>>> sending track callback for agents on that node >>>>> Feb 22 15:53:09 SC-2 osafclmd[16398]: NO Node 131343 went down. >> Not >>>>> sending track callback for agents on that node >>>>> Feb 22 15:53:09 SC-2 osafclmd[16398]: NO Node 131343 went down. >> Not >>>>> sending track callback for agents on that node >>>>> Feb 22 15:53:09 SC-2 osafclmd[16398]: NO Node 131343 went down. >> Not >>>>> sending track callback for agents on that node >>>>> Feb 22 15:53:09 SC-2 osafclmd[16398]: NO Node 131343 went down. >> Not >>>>> sending track callback for agents on that node >>>>> Feb 22 15:53:09 SC-2 osafclmd[16398]: NO Node 131343 went down. >> Not >>>>> sending track callback for agents on that node >>>>> Feb 22 15:53:09 SC-2 osafimmd[16346]: WA IMMD lost contact with >> peer >>>>> IMMD (NCSMDS_RED_DOWN) >>>>> Feb 22 15:53:09 SC-2 osafimmnd[16359]: NO Global discard node >> received >>>>> for nodeId:2010f pid:15702 >>>>> Feb 22 15:53:09 SC-2 osafimmnd[16359]: NO Implementer disconnected >> 16 >>>>> <0, 2010f(down)> (MsgQueueService131343) >>>>> Feb 22 15:53:09 SC-2 osafimmnd[16359]: NO Implementer disconnected >> 17 >>>>> <0, 2010f(down)> (@safAmfService2010f) >>>>> Feb 22 15:53:09 SC-2 osafimmnd[16359]: NO Implementer disconnected >> 18 >>>>> <0, 2010f(down)> (@OpenSafImmReplicatorB) >>>>> Feb 22 15:53:09 SC-2 opensaf_reboot: Rebooting remote node in the >>>>> absence of PLM is outside the scope of OpenSAF >>>>> >>>>> >> ============================================================== >>>> ====================================================== >>>>> -AVM >>>>> >>>>> On 2/22/2017 3:13 PM, Vu Minh Nguyen wrote: >>>>>> Hi Mahesh, >>>>>> >>>>>> I put all required patches into one. Try to use this and see if you >>>>>> still >>>>>> have that problem or not. >>>>>> >>>>>> Regards, Vu >>>>>> >>>>>>> -----Original Message----- >>>>>>> From: A V Mahesh [mailto:mahesh.va...@oracle.com] >>>>>>> Sent: Wednesday, February 22, 2017 3:35 PM >>>>>>> To: Vu Minh Nguyen <vu.m.ngu...@dektech.com.au>; >>>>>>> lennart.l...@ericsson.com; canh.v.tru...@dektech.com.au >>>>>>> Cc: opensaf-devel@lists.sourceforge.net >>>>>>> Subject: Re: [devel] [PATCH 0 of 3] Review Request for log: add >>>>>> alternative >>>>>>> destinations of log records [#2258] V4 >>>>>>> >>>>>>> Hi Vu, >>>>>>> >>>>>>> I used new #3 and #4 patches , Can you please re-send All the > final >>>>>>> patch in go, >>>>>>> >>>>>>> which i need to apply on today`s staging ( if possible publish the >>>>>>> with >>>>>>> new version ) >>>>>>> >>>>>>> -AVM >>>>>>> >>>>>>> >>>>>>> On 2/22/2017 1:52 PM, Vu Minh Nguyen wrote: >>>>>>>> Hi Mahesh, >>>>>>>> >>>>>>>>> Core was generated by `/usr/lib64/opensaf/osaflogd'. >>>>>>>>> Program terminated with signal 11, Segmentation fault. >>>>>>>>> #0 ckpt_proc_cfg_stream(lgs_cb*, void*) () at >>>>>>>>> src/log/logd/lgs_mbcsv.cc:2195 >>>>>>>>> 2195 src/log/logd/lgs_mbcsv.cc: No such file or directory. >>>>>>>>> in src/log/logd/lgs_mbcsv.cc >>>>>>>> Backtrace still points to old position (lgs_mbcsv:2195). I guess > the >>>>>>>> osaflogd binary has not been updated with the fixed patch. >>>>>>>> >>>>>>>> >>>>>>>> Regards, Vu >>>>>>>> >>>>>>>>> -----Original Message----- >>>>>>>>> From: A V Mahesh [mailto:mahesh.va...@oracle.com] >>>>>>>>> Sent: Wednesday, February 22, 2017 3:18 PM >>>>>>>>> To: Vu Minh Nguyen <vu.m.ngu...@dektech.com.au>; >>>>>>>>> lennart.l...@ericsson.com; canh.v.tru...@dektech.com.au >>>>>>>>> Cc: opensaf-devel@lists.sourceforge.net >>>>>>>>> Subject: Re: [devel] [PATCH 0 of 3] Review Request for log: add >>>>>>>> alternative >>>>>>>>> destinations of log records [#2258] V4 >>>>>>>>> >>>>>>>>> Hi Vu, >>>>>>>>> >>>>>>>>> SC-2 standby osaflogd core dumped still occurs ( Not resolved) , > the >>>>>> new >>>>>>>>> patch only resolved the application (/usr/bin/logtest ) > Segmentation >>>>>>>>> fault on SC-1 Active. >>>>>>>>> >>>>>>>>> >> ============================================================== >> ============================================================== >>>>>>>>> # gdb /usr/lib64/opensaf/osaflogd >> core_1487751055.osaflogd.4594 >>>> GNU >>>>>>>>> gdb >>>>>>>>> (GDB) SUSE (7.3-0.6.1) >>>>>>>>> Copyright (C) 2011 Free Software Foundation, Inc. >>>>>>>>> ....... >>>>>>>>> Core was generated by `/usr/lib64/opensaf/osaflogd'. >>>>>>>>> Program terminated with signal 11, Segmentation fault. >>>>>>>>> #0 ckpt_proc_cfg_stream(lgs_cb*, void*) () at >>>>>>>>> src/log/logd/lgs_mbcsv.cc:2195 >>>>>>>>> 2195 src/log/logd/lgs_mbcsv.cc: No such file or directory. >>>>>>>>> in src/log/logd/lgs_mbcsv.cc >>>>>>>>> (gdb) bt >>>>>>>>> #0 ckpt_proc_cfg_stream(lgs_cb*, void*) () at >>>>>>>>> src/log/logd/lgs_mbcsv.cc:2195 >>>>>>>>> #1 0x00007f97b026f960 in ckpt_decode_log_struct(lgs_cb*, >>>>>>>>> ncs_mbcsv_cb_arg*, void*, void*, unsigned int (*)(edu_hdl_tag*, >>>>>>>>> edu_tkn_tag*, void*, unsigned int*, edu_buf_env_tag*, >>>> EDP_OP_TYPE, >>>>>>>>> EDU_ERR*)) () at src/log/logd/lgs_mbcsv.cc:950 >>>>>>>>> #2 0x00007f97b02710dc in ckpt_decode_async_update(lgs_cb*, >>>>>>>>> ncs_mbcsv_cb_arg*) () at src/log/logd/lgs_mbcsv.cc:1086 >>>>>>>>> #3 0x00007f97b0273941 in mbcsv_callback(ncs_mbcsv_cb_arg*) () >> at >>>>>>>>> src/log/logd/lgs_mbcsv.cc:880 >>>>>>>>> #4 0x00007f97af372596 in ncs_mbscv_rcv_decode () from >>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0 >>>>>>>>> #5 0x00007f97af372766 in ncs_mbcsv_rcv_async_update () from >>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0 >>>>>>>>> #6 0x00007f97af379370 in mbcsv_process_events () from >>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0 >>>>>>>>> #7 0x00007f97af3794db in mbcsv_hdl_dispatch_all () from >>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0 >>>>>>>>> #8 0x00007f97af373ce2 in mbcsv_process_dispatch_request () at >>>>>>>>> src/mbc/mbcsv_api.c:423 >>>>>>>>> #9 0x00007f97b027096e in lgs_mbcsv_dispatch(unsigned int) () at >>>>>>>>> src/log/logd/lgs_mbcsv.cc:327 >>>>>>>>> #10 0x00007f97b024d9f2 in main () at src/log/logd/lgs_main.cc:583 >>>>>>>>> (gdb) bt full >>>>>>>>> #0 ckpt_proc_cfg_stream(lgs_cb*, void*) () at >>>>>>>>> src/log/logd/lgs_mbcsv.cc:2195 >>>>>>>>> ckpt_data_handler = {0x7f97b0270300 >>>>>>>>> <ckpt_proc_initialize_client(lgs_cb*, void*)>, 0x7f97b02701c0 >>>>>>>>> <ckpt_proc_finalize_client(lgs_cb*, void*)>, >>>>>>>>> 0x7f97b0270060 <ckpt_proc_agent_down(lgs_cb*, > void*)>, >>>>>>>>> 0x7f97b02712f0 <ckpt_proc_log_write(lgs_cb*, void*)>, >>>> 0x7f97b0271ab0 >>>>>>>>> <ckpt_proc_open_stream(lgs_cb*, void*)>, >>>>>>>>> 0x7f97b026fe80 <ckpt_proc_close_stream(lgs_cb*, >>> void*)>, >>>>>>>>> 0x7f97b0272380 <ckpt_proc_cfg_stream(lgs_cb*, void*)>, >>>>>>> 0x7f97b0274800 >>>>>>>>> <ckpt_proc_lgs_cfg_v2(lgs_cb*, void*)>, >>>>>>>>> 0x7f97b0274e10 <ckpt_proc_lgs_cfg_v3(lgs_cb*, > void*)>, >>>>>>>>> 0x7f97b02754f0 <ckpt_proc_lgs_cfg_v5(lgs_cb*, void*)>} >>>>>>>>> #1 0x00007f97b026f960 in ckpt_decode_log_struct(lgs_cb*, >>>>>>>>> ncs_mbcsv_cb_arg*, void*, void*, unsigned int (*)(edu_hdl_tag*, >>>>>>>>> edu_tkn_tag*, void*, unsigned int*, edu_buf_env_tag*, >>>> EDP_OP_TYPE, >>>>>>>>> EDU_ERR*)) () at src/log/logd/lgs_mbcsv.cc:950 >>>>>>>>> ckpt_data_handler = {0x7f97b0270300 >>>>>>>>> <ckpt_proc_initialize_client(lgs_cb*, void*)>, 0x7f97b02701c0 >>>>>>>>> <ckpt_proc_finalize_client(lgs_cb*, void*)>, >>>>>>>>> 0x7f97b0270060 <ckpt_proc_agent_down(lgs_cb*, > void*)>, >>>>>>>>> 0x7f97b02712f0 <ckpt_proc_log_write(lgs_cb*, void*)>, >>>> 0x7f97b0271ab0 >>>>>>>>> <ckpt_proc_open_stream(lgs_cb*, void*)>, >>>>>>>>> 0x7f97b026fe80 <ckpt_proc_close_stream(lgs_cb*, >>> void*)>, >>>>>>>>> 0x7f97b0272380 <ckpt_proc_cfg_stream(lgs_cb*, void*)>, >>>>>>> 0x7f97b0274800 >>>>>>>>> <ckpt_proc_lgs_cfg_v2(lgs_cb*, void*)>, >>>>>>>>> 0x7f97b0274e10 <ckpt_proc_lgs_cfg_v3(lgs_cb*, > void*)>, >>>>>>>>> 0x7f97b02754f0 <ckpt_proc_lgs_cfg_v5(lgs_cb*, void*)>} >>>>>>>>> #2 0x00007f97b02710dc in ckpt_decode_async_update(lgs_cb*, >>>>>>>>> ncs_mbcsv_cb_arg*) () at src/log/logd/lgs_mbcsv.cc:1086 >>>>>>>>> ckpt_data_handler = {0x7f97b0270300 >>>>>>>>> <ckpt_proc_initialize_client(lgs_cb*, void*)>, 0x7f97b02701c0 >>>>>>>>> <ckpt_proc_finalize_client(lgs_cb*, void*)>, >>>>>>>>> 0x7f97b0270060 <ckpt_proc_agent_down(lgs_cb*, > void*)>, >>>>>>>>> 0x7f97b02712f0 <ckpt_proc_log_write(lgs_cb*, void*)>, >>>> 0x7f97b0271ab0 >>>>>>>>> <ckpt_proc_open_stream(lgs_cb*, void*)>, >>>>>>>>> 0x7f97b026fe80 <ckpt_proc_close_stream(lgs_cb*, >>> void*)>, >>>>>>>>> 0x7f97b0272380 <ckpt_proc_cfg_stream(lgs_cb*, void*)>, >>>>>>> 0x7f97b0274800 >>>>>>>>> <ckpt_proc_lgs_cfg_v2(lgs_cb*, void*)>, >>>>>>>>> 0x7f97b0274e10 <ckpt_proc_lgs_cfg_v3(lgs_cb*, > void*)>, >>>>>>>>> 0x7f97b02754f0 <ckpt_proc_lgs_cfg_v5(lgs_cb*, void*)>} >>>>>>>>> #3 0x00007f97b0273941 in mbcsv_callback(ncs_mbcsv_cb_arg*) () >> at >>>>>>>>> src/log/logd/lgs_mbcsv.cc:880 >>>>>>>>> ckpt_data_handler = {0x7f97b0270300 >>>>>>>>> <ckpt_proc_initialize_client(lgs_cb*, void*)>, 0x7f97b02701c0 >>>>>>>>> <ckpt_proc_finalize_client(lgs_cb*, void*)>, >>>>>>>>> 0x7f97b0270060 <ckpt_proc_agent_down(lgs_cb*, > void*)>, >>>>>>>>> 0x7f97b02712f0 <ckpt_proc_log_write(lgs_cb*, void*)>, >>>> 0x7f97b0271ab0 >>>>>>>>> <ckpt_proc_open_stream(lgs_cb*, void*)>, >>>>>>>>> 0x7f97b026fe80 <ckpt_proc_close_stream(lgs_cb*, >>> void*)>, >>>>>>>>> 0x7f97b0272380 <ckpt_proc_cfg_stream(lgs_cb*, void*)>, >>>>>>> 0x7f97b0274800 >>>>>>>>> <ckpt_proc_lgs_cfg_v2(lgs_cb*, void*)>, >>>>>>>>> 0x7f97b0274e10 <ckpt_proc_lgs_cfg_v3(lgs_cb*, > void*)>, >>>>>>>>> 0x7f97b02754f0 <ckpt_proc_lgs_cfg_v5(lgs_cb*, void*)>} >>>>>>>>> #4 0x00007f97af372596 in ncs_mbscv_rcv_decode () from >>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0 >>>>>>>>> mbcsv_init_process_req_func = {0x7f97af373630 >>>>>>>>> <mbcsv_process_initialize_request>, 0x7f97af374f10 >>>>>>>>> <mbcsv_process_get_sel_obj_request>, >>>>>>>>> 0x7f97af373be0 <mbcsv_process_dispatch_request>, >>>>>>>>> 0x7f97af373af0 <mbcsv_process_finalize_request>, >> 0x7f97af373cf0 >>>>>>>>> <mbcsv_process_open_request>, >>>>>>>>> 0x7f97af374050 <mbcsv_process_close_request>, >>>>>> 0x7f97af3741e0 >>>>>>>>> <mbcsv_process_chg_role_request>, 0x7f97af3744c0 >>>>>>>>> <mbcsv_process_snd_ckpt_request>, >>>>>>>>> 0x7f97af3747d0 <mbcsv_process_snd_ntfy_request>, >>>>>>>>> 0x7f97af374970 <mbcsv_process_snd_data_req>, 0x7f97af373930 >>>>>>>>> <mbcsv_process_get_request>, >>>>>>>>> 0x7f97af374bd0 <mbcsv_process_set_request>} >>>>>>>>> #5 0x00007f97af372766 in ncs_mbcsv_rcv_async_update () from >>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0 >>>>>>>>> mbcsv_init_process_req_func = {0x7f97af373630 >>>>>>>>> <mbcsv_process_initialize_request>, 0x7f97af374f10 >>>>>>>>> <mbcsv_process_get_sel_obj_request>, >>>>>>>>> 0x7f97af373be0 <mbcsv_process_dispatch_request>, >>>>>>>>> 0x7f97af373af0 <mbcsv_process_finalize_request>, >> 0x7f97af373cf0 >>>>>>>>> <mbcsv_process_open_request>, >>>>>>>>> 0x7f97af374050 <mbcsv_process_close_request>, >>>>>> 0x7f97af3741e0 >>>>>>>>> <mbcsv_process_chg_role_request>, 0x7f97af3744c0 >>>>>>>>> <mbcsv_process_snd_ckpt_request>, >>>>>>>>> 0x7f97af3747d0 <mbcsv_process_snd_ntfy_request>, >>>>>>>>> 0x7f97af374970 <mbcsv_process_snd_data_req>, 0x7f97af373930 >>>>>>>>> <mbcsv_process_get_request>, >>>>>>>>> 0x7f97af374bd0 <mbcsv_process_set_request>} >>>>>>>>> #6 0x00007f97af379370 in mbcsv_process_events () from >>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0 >>>>>>>>> No symbol table info available. >>>>>>>>> #7 0x00007f97af3794db in mbcsv_hdl_dispatch_all () from >>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0 >>>>>>>>> No symbol table info available. >>>>>>>>> #8 0x00007f97af373ce2 in mbcsv_process_dispatch_request () at >>>>>>>>> src/mbc/mbcsv_api.c:423 >>>>>>>>> mbcsv_init_process_req_func = {0x7f97af373630 >>>>>>>>> <mbcsv_process_initialize_request>, 0x7f97af374f10 >>>>>>>>> <mbcsv_process_get_sel_obj_request>, >>>>>>>>> 0x7f97af373be0 <mbcsv_process_dispatch_request>, >>>>>>>>> 0x7f97af373af0 <mbcsv_process_finalize_request>, >> 0x7f97af373cf0 >>>>>>>>> <mbcsv_process_open_request>, >>>>>>>>> 0x7f97af374050 <mbcsv_process_close_request>, >>>>>> 0x7f97af3741e0 >>>>>>>>> <mbcsv_process_chg_role_request>, 0x7f97af3744c0 >>>>>>>>> <mbcsv_process_snd_ckpt_request>, >>>>>>>>> 0x7f97af3747d0 <mbcsv_process_snd_ntfy_request>, >>>>>>>>> 0x7f97af374970 <mbcsv_process_snd_data_req>, 0x7f97af373930 >>>>>>>>> <mbcsv_process_get_request>, >>>>>>>>> 0x7f97af374bd0 <mbcsv_process_set_request>} >>>>>>>>> #9 0x00007f97b027096e in lgs_mbcsv_dispatch(unsigned int) () at >>>>>>>>> src/log/logd/lgs_mbcsv.cc:327 >>>>>>>>> ckpt_data_handler = {0x7f97b0270300 >>>>>>>>> <ckpt_proc_initialize_client(lgs_cb*, void*)>, 0x7f97b02701c0 >>>>>>>>> <ckpt_proc_finalize_client(lgs_cb*, void*)>, >>>>>>>>> 0x7f97b0270060 <ckpt_proc_agent_down(lgs_cb*, > void*)>, >>>>>>>>> 0x7f97b02712f0 <ckpt_proc_log_write(lgs_cb*, void*)>, >>>> 0x7f97b0271ab0 >>>>>>>>> <ckpt_proc_open_stream(lgs_cb*, void*)>, >>>>>>>>> 0x7f97b026fe80 <ckpt_proc_close_stream(lgs_cb*, >>> void*)>, >>>>>>>>> 0x7f97b0272380 <ckpt_proc_cfg_stream(lgs_cb*, void*)>, >>>>>>> 0x7f97b0274800 >>>>>>>>> <ckpt_proc_lgs_cfg_v2(lgs_cb*, void*)>, >>>>>>>>> 0x7f97b0274e10 <ckpt_proc_lgs_cfg_v3(lgs_cb*, > void*)>, >>>>>>>>> 0x7f97b02754f0 <ckpt_proc_lgs_cfg_v5(lgs_cb*, void*)>} >>>>>>>>> ---Type <return> to continue, or q <return> to quit--- >>>>>>>>> #10 0x00007f97b024d9f2 in main () at src/log/logd/lgs_main.cc:583 >>>>>>>>> usr1_sel_obj = {raise_obj = -1, rmv_obj = -1} >>>>>>>>> _lgs_cb = {mds_hdl = 65547, mds_role = >> V_DEST_RL_STANDBY, >>>>>>> vaddr >>>>>>>>> = 11, log_version = {releaseCode = 65 'A', majorVersion = 2 > '\002', >>>>>>>>> minorVersion = 2 '\002'}, client_tree = { >>>>>>>>> root_node = {bit = -1, left = 0x7f97b04cf1b0, > right = >>>>>>>>> 0x7f97b04a2418, key_info = 0x7f97b04b7bd0 ""}, params = >> {key_size >>>>>>>>> = 4}, >>>>>>>>> n_nodes = 8}, comp_name = {_opaque = {46, 24947, >>>>>>>>> 17254, 28015, 15728, 20300, 11335, 24947, 21350, >>>>>>>>> 15733, >>>>>>>>> 17235, 12845, 29484, 26209, 26451, 12861, 11342, 24947, 16742, >>>> 28784, >>>>>>>>> 20285, 25968, 21358, 17985, >>>>>>>>> 0 <repeats 105 times>}}, amf_hdl = 4288675841, >>>>>>>>> amfSelectionObject = 15, amf_invocation_id = 0, is_quiesced_set = >>>>>> false, >>>>>>>>> immOiHandle = 554050912783, immSelectionObject = 21, >>>>>>>>> clmSelectionObject = 17, clm_hdl = 4279238657, >>>>>>>>> ha_state = >>>>>>>>> SA_AMF_HA_STANDBY, last_client_id = 208, async_upd_cnt = 743, >>>>>>>>> ckpt_state >>>>>>>>> = COLD_SYNC_IDLE, mbcsv_hdl = 4293918753, >>>>>>>>> mbcsv_sel_obj = 23, mbcsv_ckpt_hdl = 4292870177, >>>>>>>>> mbcsv_peer_version = 7, edu_hdl = {is_inited = true, tree = >>>>>>>>> {root_node >>>>>> = >>>>>>>>> {bit = -1, left = 0x7f97b04cf2e0, >>>>>>>>> right = 0x7f97b04a25b8, key_info = > 0x7f97b04b7d40 >>>>>> ""}, >>>>>>>>> params = {key_size = 8}, n_nodes = 12}, to_version = 1}, >>>>>>>>> fully_initialized = true, lga_down_list_head = 0x0, >>>>>>>>> lga_down_list_tail = 0x0, clm_init_sel_obj = >>>>>>>>> {raise_obj = >>>>>> -1, >>>>>>>>> rmv_obj = -1}, nid_started = true, scAbsenceAllowed = 900, >>>>>>>>> lgs_recovery_state = LGS_NORMAL} >>>>>>>>> nfds = 7 >>>>>>>>> fds = {{fd = 19, events = 1, revents = 0}, {fd = 15, >>>>>>>>> events = >>>>>>>>> 1, revents = 0}, {fd = 23, events = 1, revents = 1}, {fd = 13, >>>>>>>>> events = >>>>>>>>> 1, revents = 0}, {fd = -1, events = 1, >>>>>>>>> revents = 0}, {fd = 17, events = 1, revents = 0}, >>>>>>>>> {fd = >>>>>> 21, >>>>>>>>> events = 1, revents = 0}} >>>>>>>>> mbox_msgs = {0, 0, 0, 0, 0} >>>>>>>>> lgs_cb = 0x7f97b04a2400 >>>>>>>>> mbox_low = {0, 0, 0, 0, 0} >>>>>>>>> lgs_mbox_init_mutex = {__data = {__lock = 0, __count = > 0, >>>>>>>>> __owner = 0, __nusers = 0, __kind = 0, __spins = 0, __list = >>>>>>>>> {__prev = >>>>>>>>> 0x0, __next = 0x0}}, >>>>>>>>> __size = '\000' <repeats 39 times>, __align = 0} >>>>>>>>> lgs_mbx = 4291821569 >>>>>>>>> mbox_high = {0, 0, 0, 0, 0} >>>>>>>>> mbox_full = {false, false, false, false, false} >>>>>>>>> (gdb) >>>>>>>>> (gdb) >>>>>>>>> >>>>>>>>> Feb 22 13:40:55 SC-2 osafimmnd[4584]: NO Ccb 131 COMMITTED >>>>>>>>> (immcfg_SC-1_18714) >>>>>>>>> Feb 22 13:40:56 SC-2 osafamfnd[4634]: NO >>>>>>>>> 'safComp=LOG,safSu=SC-2,safSg=2N,safApp=OpenSAF' faulted due >> to >>>>>>>>> 'avaDown' : Recovery is 'nodeFailfast' >>>>>>>>> Feb 22 13:40:56 SC-2 osafamfnd[4634]: ER >>>>>>>>> safComp=LOG,safSu=SC-2,safSg=2N,safApp=OpenSAF Faulted due >>>>>>>>> to:avaDown >>>>>>>>> Recovery is:nodeFailfast >>>>>>>>> Feb 22 13:40:56 SC-2 osafamfnd[4634]: Rebooting OpenSAF NodeId >> = >>>>>>>>> 131599 >>>>>>>>> EE Name = , Reason: Component faulted: recovery is node failfast, >>>>>>>>> OwnNodeId = 131599, SupervisionTime = 60 >>>>>>>>> Feb 22 13:40:56 SC-2 opensaf_reboot: Rebooting local node; >>>> timeout=60 >> ============================================================== >> ============================================================== >>>>>>>>> On 2/22/2017 12:23 PM, A V Mahesh wrote: >>>>>>>>>> Hi Vu, >>>>>>>>>> >>>>>>>>>> On 2/22/2017 12:19 PM, Vu Minh Nguyen wrote: >>>>>>>>>>> [Vu] I has sent you 02 patches. There is code change in osaflogd >>>>>>>>>>> code >>>>>>>> that >>>>>>>>>>> fix the coredump you have observed. >>>>>>>>>>> The other one is test code that fix the logtest coredump. >>>>>>>>>> Ok I will re-test , and update you . >>>>>>>>>> >>>>>>>>>> -AVM >>>>>>>>>> >>>>>>>>>> On 2/22/2017 12:19 PM, Vu Minh Nguyen wrote: >>>>>>>>>>> Hi Mahehs, >>>>>>>>>>> >>>>>>>>>>> See my reply inline, [Vu]. >>>>>>>>>>> >>>>>>>>>>> Regards, Vu >>>>>>>>>>> >>>>>>>>>>>> -----Original Message----- >>>>>>>>>>>> From: A V Mahesh [mailto:mahesh.va...@oracle.com] >>>>>>>>>>>> Sent: Wednesday, February 22, 2017 1:36 PM >>>>>>>>>>>> To: Vu Minh Nguyen <vu.m.ngu...@dektech.com.au>; >>>>>>>>>>>> lennart.l...@ericsson.com; canh.v.tru...@dektech.com.au >>>>>>>>>>>> Cc: opensaf-devel@lists.sourceforge.net >>>>>>>>>>>> Subject: Re: [PATCH 0 of 3] Review Request for log: add >>>>>>>>>>>> alternative >>>>>>>>>>>> destinations of log records [#2258] V4 >>>>>>>>>>>> >>>>>>>>>>>> Hi Vu, >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> On 2/22/2017 11:52 AM, Vu Minh Nguyen wrote: >>>>>>>>>>>>> Hi Mahesh, >>>>>>>>>>>>> >>>>>>>>>>>>> Have a code fault in uml test, and other one in checkpoint. >>>>>>>>>>>> [AVM] This is Normal Suse 11 VM ( not UML). >>>>>>>>>>>>> I have just updated the code. Please re-apply for #3 and #4 >>>>>> patches. >>>>>>>>>>>> [AVM] is these new patch has function changes or only test >> code >>>>>>>>> changes ? >>>>>>>>>>> [Vu] I has sent you 02 patches. There is code change in osaflogd >>>>>>>>>>> code >>>>>>>> that >>>>>>>>>>> fix the coredump you have observed. >>>>>>>>>>> The other one is test code that fix the logtest coredump. >>>>>>>>>>>>> Note that, test case #14 of suite 17 should be run on active >>>>>>>>>>>>> node, >>>>>>>>>>>> otherwise >>>>>>>>>>>>> getting failed. >>>>>>>>>>>> [AVM] Segmentation fault of /usr/bin/logtest Not a big issue > , >>>>>>>>>>>> we need to debug why osaflogd core dumped and it is >>>>>>>>>>>> critical >>>>>>>>>>> [Vu] I found the problem. You can try with the new one to see if >>>>>>>>>>> the >>>>>>>>>>> coredump is still there or not. >>>>>>>>>>>>> I will put condition check to that test case later. >>>>>>>>>>>> -AVM >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>>> Regards, Vu >>>>>>>>>>>>> >>>>>>>>>>>>>> -----Original Message----- >>>>>>>>>>>>>> From: A V Mahesh [mailto:mahesh.va...@oracle.com] >>>>>>>>>>>>>> Sent: Wednesday, February 22, 2017 12:16 PM >>>>>>>>>>>>>> To: Vu Minh Nguyen <vu.m.ngu...@dektech.com.au>; >>>>>>>>>>>>>> lennart.l...@ericsson.com; canh.v.tru...@dektech.com.au >>>>>>>>>>>>>> Cc: opensaf-devel@lists.sourceforge.net >>>>>>>>>>>>>> Subject: Re: [PATCH 0 of 3] Review Request for log: add >>>>>> alternative >>>>>>>>>>>>>> destinations of log records [#2258] V4 >>>>>>>>>>>>>> >>>>>>>>>>>>>> Hi Vu, >>>>>>>>>>>>>> >>>>>>>>>>>>>> Thanks , >>>>>>>>>>>>>> >>>>>>>>>>>>>> While testing /usr/bin/logtest , SC-2 standby osaflogd core >>>>>>> dumped >>>>>>>>>>> and >>>>>>>>>>>>>> /usr/bin/logtest on SC-1 Active >>>>>>>>>>>>>> got Segmentation fault , am I missing any other patch ( i am >>>>>> using >>>>>>>>>>>>>> devel published patch only ) >>>>>>>>>>>>>> >>>>>>>>>>>>>> Following patches i am using : >>>>>>>>>>>>>> >>>>>>>>>>>>>> 1) #2293 (sent by Anders Widel, but not yet pushed) >>>>>>>>>>>>>> 2) #2258 (v2, sent by Lennart, but not yet pushed > yet) >>>>>>>>>>>>>> 3) #2258 (v4, sent by Vu, but not yet pushed yet) >>>>>>>>>>>>>> >>>>>>>>>>>>>> >> ============================================================== >>>>>>>>>>>>>> ======================================== >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> Core was generated by `/usr/lib64/opensaf/osaflogd'. >>>>>>>>>>>>>> Program terminated with signal 11, Segmentation fault. >>>>>>>>>>>>>> #0 ckpt_proc_cfg_stream(lgs_cb*, void*) () at >>>>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc:2195 >>>>>>>>>>>>>> 2195 src/log/logd/lgs_mbcsv.cc: No such file or directory. >>>>>>>>>>>>>> in src/log/logd/lgs_mbcsv.cc >>>>>>>>>>>>>> (gdb) bt >>>>>>>>>>>>>> #0 ckpt_proc_cfg_stream(lgs_cb*, void*) () at >>>>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc:2195 >>>>>>>>>>>>>> #1 0x00007f12c3e22960 in ckpt_decode_log_struct(lgs_cb*, >>>>>>>>>>>>>> ncs_mbcsv_cb_arg*, void*, void*, unsigned int >>>> (*)(edu_hdl_tag*, >>>>>>>>>>>>>> edu_tkn_tag*, void*, unsigned int*, edu_buf_env_tag*, >>>>>>>>> EDP_OP_TYPE, >>>>>>>>>>>>>> EDU_ERR*)) () at src/log/logd/lgs_mbcsv.cc:950 >>>>>>>>>>>>>> #2 0x00007f12c3e240dc in >>>> ckpt_decode_async_update(lgs_cb*, >>>>>>>>>>>>>> ncs_mbcsv_cb_arg*) () at src/log/logd/lgs_mbcsv.cc:1086 >>>>>>>>>>>>>> #3 0x00007f12c3e26941 in >>>> mbcsv_callback(ncs_mbcsv_cb_arg*) () >>>>>>> at >>>>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc:880 >>>>>>>>>>>>>> #4 0x00007f12c2f25596 in ncs_mbscv_rcv_decode () from >>>>>>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0 >>>>>>>>>>>>>> #5 0x00007f12c2f25766 in ncs_mbcsv_rcv_async_update () >>>> from >>>>>>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0 >>>>>>>>>>>>>> #6 0x00007f12c2f2c370 in mbcsv_process_events () from >>>>>>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0 >>>>>>>>>>>>>> #7 0x00007f12c2f2c4db in mbcsv_hdl_dispatch_all () from >>>>>>>>>>>>>> /usr/lib/../lib64/libopensaf_core.so.0 >>>>>>>>>>>>>> #8 0x00007f12c2f26ce2 in mbcsv_process_dispatch_request >> () >>>> at >>>>>>>>>>>>>> src/mbc/mbcsv_api.c:423 >>>>>>>>>>>>>> #9 0x00007f12c3e2396e in lgs_mbcsv_dispatch(unsigned int) >> () >>>> at >>>>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc:327 >>>>>>>>>>>>>> #10 0x00007f12c3e009f2 in main () at >>>>>>>>>>>>>> src/log/logd/lgs_main.cc:583 >>>>>>>>>>>>>> (gdb) >>>>>>>>>>>>>> >>>>>>>>>>>>>> >> ============================================================== >>>>>>>>>>>>>> ======================================== >>>>>>>>>>>>>> >>>>>>>>>>>>>> Feb 22 10:37:06 SC-1 osafimmnd[4020]: NO Invalid error >>>> reported >>>>>>>>>>>>>> implementer 'safLogService', Ccb 161 will be aborted >>>>>>>>>>>>>> Feb 22 10:37:06 SC-1 osafimmnd[4020]: NO Ccb 161 aborted >> in >>>>>>>>>>>> COMPLETED >>>>>>>>>>>>>> processing (validation) >>>>>>>>>>>>>> Feb 22 10:37:06 SC-1 osafimmnd[4020]: NO Ccb 161 >> ABORTED >>>>>>>>>>>> (immcfg_SC- >>>>>>>>>>>>>> 1_5394) >>>>>>>>>>>>>> Add values Fail >>>>>>>>>>>>>> >>>>>>>>>>>>>> Program received signal SIGSEGV, Segmentation fault. >>>>>>>>>>>>>> 0x000055555556929a in read_and_compare.isra.7 () at >>>>>>>>>>>>>> src/log/apitest/tet_LogOiOps.c:1891 >>>>>>>>>>>>>> 1891 src/log/apitest/tet_LogOiOps.c: No such file or >>>>>>>>>>>>>> directory. >>>>>>>>>>>>>> in src/log/apitest/tet_LogOiOps.c >>>>>>>>>>>>>> (gdb) Feb 22 10:37:07 SC-1 sshd[5298]: Accepted keyboard- >>>>>>>>>>>> interactive/pam >>>>>>>>>>>>>> for root from 10.176.178.22 port 51945 ssh2 >>>>>>>>>>>>>> bt >>>>>>>>>>>>>> #0 0x000055555556929a in read_and_compare.isra.7 () at >>>>>>>>>>>>>> src/log/apitest/tet_LogOiOps.c:1891 >>>>>>>>>>>>>> #1 0x0000555555569bbb in >>>>>>>>>>>>>> check_logRecordDestinationConfigurationEmpty >>>>>>>>>>>>>> () at src/log/apitest/tet_LogOiOps.c:2179 >>>>>>>>>>>>>> #2 0x0000555555573495 in run_test_case () >>>>>>>>>>>>>> #3 0x0000555555573934 in test_run () >>>>>>>>>>>>>> #4 0x000055555555c7cd in main () at >>>>>>>>>>>>>> src/log/apitest/logtest.c:569 >>>>>>>>>>>>>> (gdb) >>>>>>>>>>>>>> >>>>>>>>>>>>>> >> ============================================================== >>>>>>>>>>>>>> ======================================== >>>>>>>>>>>>>> >>>>>>>>>>>>>> -AVM >>>>>>>>>>>>>> >>>>>>>>>>>>>> On 2/22/2017 9:48 AM, Vu Minh Nguyen wrote: >>>>>>>>>>>>>>> Hi Mahesh, >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> I send them in attachment instead, and name them in the >>>> order. >>>>>>>>>>>>>>> I just pull the latest code, and apply them without getting >>> any >>>>>>>> hunk >>>>>>>>>>>>> error. >>>>>>>>>>>>>>> Please try with them, and let me know if you see any >> problem. >>>>>>>>>>>>>>> Regards, Vu >>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> -----Original Message----- >>>>>>>>>>>>>>>> From: A V Mahesh [mailto:mahesh.va...@oracle.com] >>>>>>>>>>>>>>>> Sent: Wednesday, February 22, 2017 11:09 AM >>>>>>>>>>>>>>>> To: Vu Minh Nguyen <vu.m.ngu...@dektech.com.au>; >>>>>>>>>>>>>>>> lennart.l...@ericsson.com; >> canh.v.tru...@dektech.com.au >>>>>>>>>>>>>>>> Cc: opensaf-devel@lists.sourceforge.net >>>>>>>>>>>>>>>> Subject: Re: [PATCH 0 of 3] Review Request for log: add >>>>>>>> alternative >>>>>>>>>>>>>>>> destinations of log records [#2258] V4 >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> Hi Vu, >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> I did follow that still i get Hunk #2 FAILED even on > today's >>>>>>>> staging >> ============================================================== >>>>>>>>>>>>>>>> ================== >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> [root@dhcp-hyd-scp-5fl-10-176-177-96 staging]# patch - >> p1 >>>>>>> <2293 >>>>>>>>>>>>>>>> patching file src/base/Makefile.am >>>>>>>>>>>>>>>> Hunk #1 succeeded at 33 (offset 1 line). >>>>>>>>>>>>>>>> Hunk #3 succeeded at 183 (offset 1 line). >>>>>>>>>>>>>>>> patching file src/base/file_descriptor.cc >>>>>>>>>>>>>>>> patching file src/base/file_descriptor.h >>>>>>>>>>>>>>>> patching file src/base/tests/unix_socket_test.cc >>>>>>>>>>>>>>>> patching file src/base/unix_client_socket.cc >>>>>>>>>>>>>>>> patching file src/base/unix_server_socket.cc >>>>>>>>>>>>>>>> patching file src/base/unix_socket.cc >>>>>>>>>>>>>>>> patching file src/base/unix_socket.h >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> [root@dhcp-hyd-scp-5fl-10-176-177-96 staging]# patch - >> p1 >>>>>>> <2258- >>>>>>>>> 1 >>>>>>>>>>>>>>>> patching file src/log/Makefile.am >>>>>>>>>>>>>>>> Hunk #1 succeeded at 71 (offset -1 lines). >>>>>>>>>>>>>>>> patching file src/log/config/logsv_classes.xml >>>>>>>>>>>>>>>> Hunk #1 FAILED at 147. >>>>>>>>>>>>>>>> 1 out of 1 hunk FAILED -- saving rejects to file >>>>>>>>>>>>>>>> src/log/config/logsv_classes.xml.rej >>>>>>>>>>>>>>>> patching file src/log/logd/lgs_config.cc >>>>>>>>>>>>>>>> Hunk #1 succeeded at 35 (offset -5 lines). >>>>>>>>>>>>>>>> Hunk #2 FAILED at 705. >>>>>>>>>>>>>>>> Hunk #3 FAILED at 971. >>>>>>>>>>>>>>>> 2 out of 3 hunks FAILED -- saving rejects to file >>>>>>>>>>>>>>>> src/log/logd/lgs_config.cc.rej >>>>>>>>>>>>>>>> patching file src/log/logd/lgs_config.h >>>>>>>>>>>>>>>> Hunk #1 FAILED at 304. >>>>>>>>>>>>>>>> 1 out of 1 hunk FAILED -- saving rejects to file >>>>>>>>>>>>>>>> src/log/logd/lgs_config.h.rej >>>>>>>>>>>>>>>> patching file src/log/logd/lgs_dest.cc >>>>>>>>>>>>>>>> patching file src/log/logd/lgs_dest.h >>>>>>>>>>>>>>>> patching file src/log/logd/lgs_evt.cc >>>>>>>>>>>>>>>> patching file src/log/logd/lgs_imm.cc >>>>>>>>>>>>>>>> Hunk #1 FAILED at 45. >>>>>>>>>>>>>>>> Hunk #2 succeeded at 235 (offset -1 lines). >>>>>>>>>>>>>>>> Hunk #3 FAILED at 877. >>>>>>>>>>>>>>>> Hunk #4 succeeded at 1273 (offset -20 lines). >>>>>>>>>>>>>>>> Hunk #5 succeeded at 1404 (offset -1 lines). >>>>>>>>>>>>>>>> Hunk #6 succeeded at 1449 (offset -20 lines). >>>>>>>>>>>>>>>> Hunk #7 succeeded at 2032 (offset -1 lines). >>>>>>>>>>>>>>>> Hunk #8 FAILED at 2181. >>>>>>>>>>>>>>>> Hunk #9 succeeded at 2271 (offset -54 lines). >>>>>>>>>>>>>>>> Hunk #10 succeeded at 2387 (offset -1 lines). >>>>>>>>>>>>>>>> Hunk #11 succeeded at 2377 (offset -54 lines). >>>>>>>>>>>>>>>> Hunk #12 succeeded at 2478 (offset -1 lines). >>>>>>>>>>>>>>>> Hunk #13 succeeded at 2684 (offset -54 lines). >>>>>>>>>>>>>>>> Hunk #14 succeeded at 2821 (offset -1 lines). >>>>>>>>>>>>>>>> 3 out of 14 hunks FAILED -- saving rejects to file >>>>>>>>>>>>>>>> src/log/logd/lgs_imm.cc.rej >>>>>>>>>>>>>>>> patching file src/log/logd/lgs_main.cc >>>>>>>>>>>>>>>> patching file src/log/logd/lgs_mbcsv.cc >>>>>>>>>>>>>>>> patching file src/log/logd/lgs_mbcsv.h >>>>>>>>>>>>>>>> patching file src/log/logd/lgs_mbcsv_v5.cc >>>>>>>>>>>>>>>> Hunk #3 succeeded at 133 (offset -1 lines). >>>>>>>>>>>>>>>> patching file src/log/logd/lgs_mbcsv_v7.cc >>>>>>>>>>>>>>>> patching file src/log/logd/lgs_mbcsv_v7.h >>>>>>>>>>>>>>>> patching file src/log/logd/lgs_stream.cc >>>>>>>>>>>>>>>> patching file src/log/logd/lgs_stream.h >>>>>>>>>>>>>>>> patching file src/log/logd/lgs_util.cc >>>>>>>>>>>>>>>> patching file src/log/logd/lgs_util.h >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> [root@dhcp-hyd-scp-5fl-10-176-177-96 staging]# patch - >> p1 >>>>>>> <2258- >>>>>>>>> 2 >>>>>>>>>>>>>>>> patching file src/log/Makefile.am >>>>>>>>>>>>>>>> Hunk #1 succeeded at 180 (offset -3 lines). >>>>>>>>>>>>>>>> patching file src/log/apitest/tet_LogOiOps.c >>>>>>>>>>>>>>>> Hunk #1 FAILED at 1923. >>>>>>>>>>>>>>>> Hunk #2 FAILED at 1979. >>>>>>>>>>>>>>>> Hunk #3 FAILED at 2067. >>>>>>>>>>>>>>>> Hunk #4 FAILED at 2094. >>>>>>>>>>>>>>>> 4 out of 4 hunks FAILED -- saving rejects to file >>>>>>>>>>>>>>>> src/log/apitest/tet_LogOiOps.c.rej >>>>>>>>>>>>>>>> patching file src/log/apitest/tet_cfg_destination.c >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> [root@dhcp-hyd-scp-5fl-10-176-177-96 staging]# patch - >> p1 >>>>>>> <2258- >>>>>>>>> 3 >>>>>>>>>>>>>>>> patching file src/log/Makefile >>>>>>>>>>>>>>>> patching file src/log/Makefile.am >>>>>>>>>>>>>>>> Hunk #1 succeeded at 80 (offset -1 lines). >>>>>>>>>>>>>>>> Hunk #2 succeeded at 217 (offset -2 lines). >>>>>>>>>>>>>>>> patching file src/log/tests/Makefile >>>>>>>>>>>>>>>> patching file src/log/tests/lgs_dest_test.cc >>>>>>>>>>>>>>>> [root@dhcp-hyd-scp-5fl-10-176-177-96 staging]# vi >>>>>>>>>>>>>>>> src/log/apitest/tet_LogOiOps.c.rej >>>>>>>>>>>>>>>> [root@dhcp-hyd-scp-5fl-10-176-177-96 staging]# >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >> ============================================================== >>>>>>>>>>>>>>>> ======================== >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> -AVM >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> On 2/21/2017 3:53 PM, Vu Minh Nguyen wrote: >>>>>>>>>>>>>>>>> Hi Mahesh, >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> As I has mentioned in below: >>>>>>>>>>>>>>>>>>> To run the test, this patch has dependent on >>>>>> following >>>>>>>>>>> patches: >>>>>>>>>>>>>>>>>>> 1) #2293 (sent by Anders Widel, but not yet >>>>>>>>>>>>>>>>>>> pushed) >>>>>>>>>>>>>>>>>>> 2) #2258 (v2, sent by Lennart, but not yet > pushed >>>>>> yet) >>>>>>>>>>>>>>>>> So, you need to apply #2293 first, then #2258 which sent >> by >>>>>>>>> Lennart >>>>>>>>>>>>>>>>> yesterday, then mine. >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> Regards, Vu >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> -----Original Message----- >>>>>>>>>>>>>>>>>> From: A V Mahesh [mailto:mahesh.va...@oracle.com] >>>>>>>>>>>>>>>>>> Sent: Tuesday, February 21, 2017 5:10 PM >>>>>>>>>>>>>>>>>> To: Vu Minh Nguyen <vu.m.ngu...@dektech.com.au>; >>>>>>>>>>>>>>>>>> lennart.l...@ericsson.com; >>>> canh.v.tru...@dektech.com.au >>>>>>>>>>>>>>>>>> Cc: opensaf-devel@lists.sourceforge.net >>>>>>>>>>>>>>>>>> Subject: Re: [PATCH 0 of 3] Review Request for log: add >>>>>>>>> alternative >>>>>>>>>>>>>>>>>> destinations of log records [#2258] V4 >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> Hi Vu, >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> Is this applies on top of log #2146 - V4 , I see > both >>>>>>>> #tickets >>>>>>>>>>>>>>> has >>>>>>>>>>>>>>>>>> version changes ? >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> in which order i need to apply ( #2146 & #2258 ) or >>>>>>>>>>>>>>>>>> (#2258 >>>>>> & >>>>>>>>>>>>>>> #2146). >>>> ========================================================= >>>>>>>>>>>>>>>>>> patching file src/log/Makefile.am >>>>>>>>>>>>>>>>>> Hunk #1 FAILED at 72. >>>>>>>>>>>>>>>>>> Hunk #2 FAILED at 120. >>>>>>>>>>>>>>>>>> 2 out of 2 hunks FAILED -- saving rejects to file >>>>>>>>>>>>>>> src/log/Makefile.am.rej >>>>>>>>>>>>>>>>>> patching file src/log/config/logsv_classes.xml >>>>>>>>>>>>>>>>>> Hunk #1 FAILED at 147. >>>>>>>>>>>>>>>>>> 1 out of 1 hunk FAILED -- saving rejects to file >>>>>>>>>>>>>>>>>> src/log/config/logsv_classes.xml.rej >>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_config.cc >>>>>>>>>>>>>>>>>> Hunk #1 succeeded at 35 (offset -5 lines). >>>>>>>>>>>>>>>>>> Hunk #2 FAILED at 705. >>>>>>>>>>>>>>>>>> Hunk #3 FAILED at 971. >>>>>>>>>>>>>>>>>> 2 out of 3 hunks FAILED -- saving rejects to file >>>>>>>>>>>>>>>>>> src/log/logd/lgs_config.cc.rej >>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_config.h >>>>>>>>>>>>>>>>>> Hunk #1 FAILED at 304. >>>>>>>>>>>>>>>>>> 1 out of 1 hunk FAILED -- saving rejects to file >>>>>>>>>>>>>>>>>> src/log/logd/lgs_config.h.rej >>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_dest.cc >>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_dest.h >>>>>>>>>>>>>>>>>> patching file src/log/logd/lgs_evt.cc >>>>>>>>>>>>>>>>>> Hunk #1 FAILED at 1. >>>>>>>>>>>>>>>>>> Hunk #2 succeeded at 30 with fuzz 2 (offset 2 lines). >>>>>>>>>>>>>>>>>> Hunk #3 succeeded at 1282 (offset 45 lines). >>>>>>>>>>>>>>>>>> Hunk #4 succeeded at 1300 (offset 2 lines). >>>>>>>>>>>>>>>>>> 1 out of 4 hunks FAILED -- saving rejects to file >>>>>>>>>>>>>>>>>> src/log/logd/lgs_evt.cc.rej >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> >> ============================================================== >>>>>>>>>>>>>>>>>> === >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> -AVM >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> On 2/21/2017 3:03 PM, Vu Minh Nguyen wrote: >>>>>>>>>>>>>>>>>>> Summary: log: add alternative destinations of log >> records >>>>>>>>> [#2258] >>>>>>>>>>>>>>>>>>> Review request for Trac Ticket(s): #2258 >>>>>>>>>>>>>>>>>>> Peer Reviewer(s): Lennart, Canh, Mahesh >>>>>>>>>>>>>>>>>>> Pull request to: <<LIST THE PERSON WITH PUSH >> ACCESS >>>>>>> HERE>> >>>>>>>>>>>>>>>>>>> Affected branch(es): Default >>>>>>>>>>>>>>>>>>> Development branch: Default >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> -------------------------------- >>>>>>>>>>>>>>>>>>> Impacted area Impact y/n >>>>>>>>>>>>>>>>>>> -------------------------------- >>>>>>>>>>>>>>>>>>> Docs n >>>>>>>>>>>>>>>>>>> Build system n >>>>>>>>>>>>>>>>>>> RPM/packaging n >>>>>>>>>>>>>>>>>>> Configuration files n >>>>>>>>>>>>>>>>>>> Startup scripts n >>>>>>>>>>>>>>>>>>> SAF services n >>>>>>>>>>>>>>>>>>> OpenSAF services y >>>>>>>>>>>>>>>>>>> Core libraries n >>>>>>>>>>>>>>>>>>> Samples n >>>>>>>>>>>>>>>>>>> Tests y >>>>>>>>>>>>>>>>>>> Other n >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> Comments (indicate scope for each "y" above): >>>>>>>>>>>>>>>>>>> --------------------------------------------- >>>>>>>>>>>>>>>>>>> To run the test, this patch has dependent on >>>>>> following >>>>>>>>>>> patches: >>>>>>>>>>>>>>>>>>> 1) #2293 (sent by Anders Widel, but not yet >>>>>>>>>>>>>>>>>>> pushed) >>>>>>>>>>>>>>>>>>> 2) #2258 (v2, sent by Lennart, but not yet > pushed >>>>>> yet) >>>>>>>>>>>>>>>>>>> changeset >>>> d74aaf3025c99cade3165a15831124548f4d85bd >>>>>>>>>>>>>>>>>>> Author: Vu Minh Nguyen >>>> <vu.m.ngu...@dektech.com.au> >>>>>>>>>>>>>>>>>>> Date: Wed, 15 Feb 2017 14:36:00 +0700 >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> log: add alternative destinations of log records >>>>>>>>>>>>>>>>>>> [#2258] >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> Here are major info, detailed info will be added to >>>>>>>>>>>>>>>>>>> PR doc >>>>>>>>>>>>> soon. 1) >>>>>>>>>>>>>>>>>> Add >>>>>>>>>>>>>>>>>>> attribute "saLogRecordDestination" to log stream. > 2) >>>>>>>>>>>>>>>>>>> Add >>>>>>>>>>>>> Local >>>>>>>>>>>>>>>>>> socket >>>>>>>>>>>>>>>>>>> destintion handler 3) Integrate into first > increment >>>>>>>>>>>>>>>>>>> made by >>>>>>>>>>>>> Lennart >>>>>>>>>>>>>>>>>>> changeset >>>> 4bae27a478c235df3058f43c92d3a5483233b01d >>>>>>>>>>>>>>>>>>> Author: Vu Minh Nguyen >>>> <vu.m.ngu...@dektech.com.au> >>>>>>>>>>>>>>>>>>> Date: Wed, 15 Feb 2017 15:07:09 +0700 >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> log: add UML test case to verify alternative >>>>>>>>>>>>>>>>>>> destination >>>>>>>>>>>>> [#2258] >>>>>>>>>>>>>>>>>>> Major changes: 1) Modify Lennart's test cases >> because >>>>>>>>>>>>> enhancing >>>>>>>>>>>>>>>>>> destination >>>>>>>>>>>>>>>>>>> configuration validation rules. 2) Add test suite >>>>>>>>>>>>>>>>>>> #17 to >>>>>>>>>>>>> verify >>>>>>>>>>>>>>>>>> alternative >>>>>>>>>>>>>>>>>>> destination >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> changeset >> bc375725fed22bb4f8cb3ae3df5f96fb9d281efb >>>>>>>>>>>>>>>>>>> Author: Vu Minh Nguyen >>>> <vu.m.ngu...@dektech.com.au> >>>>>>>>>>>>>>>>>>> Date: Thu, 16 Feb 2017 17:22:13 +0700 >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> log: add unit tests to verify interfaces provided > by >>>>>>>>>>>>> destination >>>>>>>>>>>>>>>>>> handler >>>>>>>>>>>>>>>>>>> [#2258] >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> Unit tests to verify major interfaces: 1) >>>>>>>>>>>>>>>>>>> CfgDestination() >>>>>>>>>>>>> 2) >>>>>>>>>>>>>>>>>>> WriteToDestination() >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> Added Files: >>>>>>>>>>>>>>>>>>> ------------ >>>>>>>>>>>>>>>>>>> src/log/apitest/tet_cfg_destination.c >>>>>>>>>>>>>>>>>>> src/log/logd/lgs_dest.cc >>>>>>>>>>>>>>>>>>> src/log/logd/lgs_dest.h >>>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv_v7.cc >>>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv_v7.h >>>>>>>>>>>>>>>>>>> src/log/tests/lgs_dest_test.cc >>>>>>>>>>>>>>>>>>> src/log/tests/Makefile >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> Complete diffstat: >>>>>>>>>>>>>>>>>>> ------------------ >>>>>>>>>>>>>>>>>>> src/log/Makefile | 4 + >>>>>>>>>>>>>>>>>>> src/log/Makefile.am | 31 +++++- >>>>>>>>>>>>>>>>>>> src/log/apitest/tet_LogOiOps.c | 8 +- >>>>>>>>>>>>>>>>>>> src/log/apitest/tet_cfg_destination.c | 483 >> ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++ >>>>>>>>>>>>>>>>>> ++++++++++++++++++++++++++++++++++ >>>>>>>>>>>>>>>>>>> src/log/config/logsv_classes.xml | 7 +- >>>>>>>>>>>>>>>>>>> src/log/logd/lgs_config.cc | 169 >>>>>>>>>>>>>>>>>> +++++++++++++++++++++++++++++--- >>>>>>>>>>>>>>>>>>> src/log/logd/lgs_config.h | 3 +- >>>>>>>>>>>>>>>>>>> src/log/logd/lgs_dest.cc | 707 >> ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++ >> ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++ >>>>>>>>>>>>>>>>>> +++++++++++++++++ >>>>>>>>>>>>>>>>>>> src/log/logd/lgs_dest.h | 576 >> ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++ >>>>>>> ++++++++++++++++++++++++++++++++++++++++++++++++++++ >>>>>>>>>>>>>>>>>>> src/log/logd/lgs_evt.cc | 33 ++++++ >>>>>>>>>>>>>>>>>>> src/log/logd/lgs_imm.cc | 202 >>>>>>>>>>>>>>>>>> +++++++++++++++++++++++++++++++++------ >>>>>>>>>>>>>>>>>>> src/log/logd/lgs_main.cc | 8 + >>>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv.cc | 103 >>>>>>>>>>> ++++++++++++++++++- >>>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv.h | 6 +- >>>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv_v5.cc | 10 + >>>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv_v7.cc | 177 >>>>>>>>>>>>>>>>>> +++++++++++++++++++++++++++++++++++ >>>>>>>>>>>>>>>>>>> src/log/logd/lgs_mbcsv_v7.h | 67 >>>>>>>> +++++++++++++ >>>>>>>>>>>>>>>>>>> src/log/logd/lgs_stream.cc | 60 >>>>>>>> +++++++++++- >>>>>>>>>>>>>>>>>>> src/log/logd/lgs_stream.h | 16 +++ >>>>>>>>>>>>>>>>>>> src/log/logd/lgs_util.cc | 63 >>>>>>>> ++++++++++++ >>>>>>>>>>>>>>>>>>> src/log/logd/lgs_util.h | 11 +- >>>>>>>>>>>>>>>>>>> src/log/tests/Makefile | 20 +++ >>>>>>>>>>>>>>>>>>> src/log/tests/lgs_dest_test.cc | 209 >>>>>>>>>>>>>>>>>> +++++++++++++++++++++++++++++++++++++++++ >>>>>>>>>>>>>>>>>>> 23 files changed, 2896 insertions(+), 77 >>>>>>>>>>>>>>>>>>> deletions(-) >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> Testing Commands: >>>>>>>>>>>>>>>>>>> ----------------- >>>>>>>>>>>>>>>>>>> Run UML test suite #17 >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> Testing, Expected Results: >>>>>>>>>>>>>>>>>>> -------------------------- >>>>>>>>>>>>>>>>>>> All test passed >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> Conditions of Submission: >>>>>>>>>>>>>>>>>>> ------------------------- >>>>>>>>>>>>>>>>>>> <<HOW MANY DAYS BEFORE PUSHING, >> CONSENSUS >>>>>>> ETC>> >>>>>>>>>>>>>>>>>>> Arch Built Started Linux distro >>>>>>>>>>>>>>>>>>> ------------------------------------------- >>>>>>>>>>>>>>>>>>> mips n n >>>>>>>>>>>>>>>>>>> mips64 n n >>>>>>>>>>>>>>>>>>> x86 n n >>>>>>>>>>>>>>>>>>> x86_64 n n >>>>>>>>>>>>>>>>>>> powerpc n n >>>>>>>>>>>>>>>>>>> powerpc64 n n >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> Reviewer Checklist: >>>>>>>>>>>>>>>>>>> ------------------- >>>>>>>>>>>>>>>>>>> [Submitters: make sure that your review doesn't trigger >>>> any >>>>>>>>>>>>>>>> checkmarks!] >>>>>>>>>>>>>>>>>>> Your checkin has not passed review because (see >> checked >>>>>>>>> entries): >>>>>>>>>>>>>>>>>>> ___ Your RR template is generally incomplete; it has >> too >>>>>>> many >>>>>>>>>>> blank >>>>>>>>>>>>>>>>> entries >>>>>>>>>>>>>>>>>>> that need proper data filled in. >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> ___ You have failed to nominate the proper persons for >>>>>>> review >>>>>>>>> and >>>>>>>>>>>>>>> push. >>>>>>>>>>>>>>>>>>> ___ Your patches do not have proper short+long >> header >>>>>>>>>>>>>>>>>>> ___ You have grammar/spelling in your header that is >>>>>>>>> unacceptable. >>>>>>>>>>>>>>>>>>> ___ You have exceeded a sensible line length in your >>>>>>>>>>>>>>>>>> headers/comments/text. >>>>>>>>>>>>>>>>>>> ___ You have failed to put in a proper Trac Ticket # > into >>>>>> your >>>>>>>>>>>>>>> commits. >>>>>>>>>>>>>>>>>>> ___ You have incorrectly put/left internal data in your >>>>>>>>>>>>> comments/files >>>>>>>>>>>>>>>>>>> (i.e. internal bug tracking tool IDs, product >>>>>> names >>>>>>>> etc) >>>>>>>>>>>>>>>>>>> ___ You have not given any evidence of testing beyond >>>> basic >>>>>>>>> build >>>>>>>>>>>>>>> tests. >>>>>>>>>>>>>>>>>>> Demonstrate some level of runtime or other sanity >>>>>>>> testing. >>>>>>>>>>>>>>>>>>> ___ You have ^M present in some of your files. These >>>>>>>>>>>>>>>>>>> have to >>>>>>>>> be >>>>>>>>>>>>>>>> removed. >>>>>>>>>>>>>>>>>>> ___ You have needlessly changed whitespace or added >>>>>>>>> whitespace >>>>>>>>>>>>>>>> crimes >>>>>>>>>>>>>>>>>>> like trailing spaces, or spaces before > tabs. >>>>>>>>>>>>>>>>>>> ___ You have mixed real technical changes with >>>> whitespace >>>>>>> and >>>>>>>>>>>> other >>>>>>>>>>>>>>>>>>> cosmetic code cleanup changes. These have to be >>>>>>>>> separate >>>>>>>>>>>>>>> commits. >>>>>>>>>>>>>>>>>>> ___ You need to refactor your submission into logical >>>>>>>>>>>>>>>>>>> chunks; >>>>>>>>>>> there >>>>>>>>>>>>> is >>>>>>>>>>>>>>>>>>> too much content into a single commit. >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> ___ You have extraneous garbage in your review >> (merge >>>>>>>>> commits >>>>>>>>>>>> etc) >>>>>>>>>>>>>>>>>>> ___ You have giant attachments which should never >> have >>>>>>> been >>>>>>>>>>>> sent; >>>>>>>>>>>>>>>>>>> Instead you should place your content in a public >>>>>>>> tree to >>>>>>>>>>> be >>>>>>>>>>>>>>>>> pulled. >>>>>>>>>>>>>>>>>>> ___ You have too many commits attached to an e-mail; >>>>>>> resend >>>>>>>>> as >>>>>>>>>>>>>>>> threaded >>>>>>>>>>>>>>>>>>> commits, or place in a public tree for a pull. >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> ___ You have resent this content multiple times >> without a >>>>>>> clear >>>>>>>>>>>>>>>>> indication >>>>>>>>>>>>>>>>>>> of what has changed between each re-send. >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> ___ You have failed to adequately and individually >> address >>>>>> all >>>>>>>> of >>>>>>>>>>>>> the >>>>>>>>>>>>>>>>>>> comments and change requests that were proposed in >>>>>>>> the >>>>>>>>>>>>> initial >>>>>>>>>>>>>>>>> review. >>>>>>>>>>>>>>>>>>> ___ You have a misconfigured ~/.hgrc file (i.e. >> username, >>>>>>> email >>>>>>>>>>> etc) >>>>>>>>>>>>>>>>>>> ___ Your computer have a badly configured date and >>>> time; >>>>>>>>>>>> confusing >>>>>>>>>>>>>> the >>>>>>>>>>>>>>>>>>> the threaded patch review. >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> ___ Your changes affect IPC mechanism, and you don't >>>>>>> present >>>>>>>>> any >>>>>>>>>>>>>>>> results >>>>>>>>>>>>>>>>>>> for in-service upgradability test. >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> ___ Your changes affect user manual and >> documentation, >>>>>>> your >>>>>>>>>>>> patch >>>>>>>>>>>>>>>> series >>>>>>>>>>>>>>>>>>> do not contain the patch that updates the >>>>>>>>>>>>>>>>>>> Doxygen >>>>>>>>> manual. > ---------------------------------------------------------------------------- >>>>>>>> -- >>>>>>>>>> Check out the vibrant tech community on one of the world's most >>>>>>>>>> engaging tech sites, SlashDot.org! http://sdm.link/slashdot >>>>>>>>>> _______________________________________________ >>>>>>>>>> Opensaf-devel mailing list >>>>>>>>>> Opensaf-devel@lists.sourceforge.net >>>>>>>>>> https://lists.sourceforge.net/lists/listinfo/opensaf-devel > ------------------------------------------------------------------------------ Check out the vibrant tech community on one of the world's most engaging tech sites, SlashDot.org! http://sdm.link/slashdot _______________________________________________ Opensaf-devel mailing list Opensaf-devel@lists.sourceforge.net https://lists.sourceforge.net/lists/listinfo/opensaf-devel