Hi All, the following logs from corosync.log that might help.
Apr 25 10:29:32 [15334] gmlcdbw02 pengine: info: native_print: ipmi-fence-db01 (stonith:fence_ipmilan): Started gmlcdbw01 Apr 25 10:29:32 [15334] gmlcdbw02 pengine: info: native_print: ipmi-fence-db02 (stonith:fence_ipmilan): Started gmlcdbw02 Apr 25 10:29:32 [15334] gmlcdbw02 pengine: info: RecurringOp: Start recurring monitor (60s) for ipmi-fence-db01 on gmlcdbw02 Apr 25 10:29:32 [15334] gmlcdbw02 pengine: notice: LogActions: Move ipmi-fence-db01 (Started gmlcdbw01 -> gmlcdbw02) Apr 25 10:29:32 [15334] gmlcdbw02 pengine: info: LogActions: Leave ipmi-fence-db02 (Started gmlcdbw02) Apr 25 10:29:32 [15335] gmlcdbw02 crmd: notice: te_rsc_command: Initiating action 11: stop ipmi-fence-db01_stop_0 on gmlcdbw01 Apr 25 10:29:32 [15330] gmlcdbw02 cib: info: cib_perform_op: + /cib/status/node_state[@id='1']/lrm[@id='1']/lrm_resources/lrm_resource[@id='ipmi-fence-db01']/lrm_rsc_op[@id='ipmi-fence-db01_last_0']: @operation_key=ipmi-fence-db01_stop_0, @operation=stop, @transition-key=11:2485:0:27a91aab-060a-4de9-80b1-18abeb7bd850, @transition-magic=0:0;11:2485:0:27a91aab-060a-4de9-80b1-18abeb7bd850, @call-id=75, @last-run=1493087372, @last-rc-change=1493087372, @exec-time=0 Apr 25 10:29:32 [15330] gmlcdbw02 cib: info: cib_perform_op: + /cib/status/node_state[@id='1']/lrm[@id='1']/lrm_resources/lrm_resource[@id='ipmi-fence-db01']/lrm_rsc_op[@id='ipmi-fence-db01_last_0']: @operation_key=ipmi-fence-db01_stop_0, @operation=stop, @transition-key=11:2485:0:27a91aab-060a-4de9-80b1-18abeb7bd850, @transition-magic=0:0;11:2485:0:27a91aab-060a-4de9-80b1-18abeb7bd850, @call-id=75, @last-run=1493087372, @last-rc-change=1493087372, @exec-time=0 Apr 25 10:29:32 [15330] gmlcdbw02 cib: info: cib_perform_op: + /cib/status/node_state[@id='1']/lrm[@id='1']/lrm_resources/lrm_resource[@id='ipmi-fence-db01']/lrm_rsc_op[@id='ipmi-fence-db01_last_0']: @operation_key=ipmi-fence-db01_stop_0, @operation=stop, @transition-key=11:2485:0:27a91aab-060a-4de9-80b1-18abeb7bd850, @transition-magic=0:0;11:2485:0:27a91aab-060a-4de9-80b1-18abeb7bd850, @call-id=75, @last-run=1493087372, @last-rc-change=1493087372, @exec-time=0 Apr 25 10:29:32 [15335] gmlcdbw02 crmd: info: match_graph_event: Action ipmi-fence-db01_stop_0 (11) confirmed on gmlcdbw01 (rc=0) Apr 25 10:29:32 [15335] gmlcdbw02 crmd: notice: te_rsc_command: Initiating action 12: start ipmi-fence-db01_start_0 on gmlcdbw02 (local) Apr 25 10:29:32 [15335] gmlcdbw02 crmd: info: do_lrm_rsc_op: Performing key=12:2485:0:27a91aab-060a-4de9-80b1-18abeb7bd850 op=ipmi-fence-db01_start_0 Apr 25 10:29:32 [15332] gmlcdbw02 lrmd: info: log_execute: executing - rsc:ipmi-fence-db01 action:start call_id:65 Apr 25 10:29:32 [15332] gmlcdbw02 lrmd: info: log_finished: finished - rsc:ipmi-fence-db01 action:start call_id:65 exit-code:0 exec-time:45ms queue-time:0ms Apr 25 10:29:33 [15335] gmlcdbw02 crmd: notice: process_lrm_event: Operation ipmi-fence-db01_start_0: ok (node=gmlcdbw02, call=65, rc=0, cib-update=2571, confirmed=true) Apr 25 10:29:33 [15330] gmlcdbw02 cib: info: cib_perform_op: + /cib/status/node_state[@id='2']/lrm[@id='2']/lrm_resources/lrm_resource[@id='ipmi-fence-db01']/lrm_rsc_op[@id='ipmi-fence-db01_last_0']: @operation_key=ipmi-fence-db01_start_0, @operation=start, @transition-key=12:2485:0:27a91aab-060a-4de9-80b1-18abeb7bd850, @transition-magic=0:0;12:2485:0:27a91aab-060a-4de9-80b1-18abeb7bd850, @call-id=65, @last-run=1493087372, @last-rc-change=1493087372, @exec-time=45 Apr 25 10:29:33 [15330] gmlcdbw02 cib: info: cib_perform_op: + /cib/status/node_state[@id='2']/lrm[@id='2']/lrm_resources/lrm_resource[@id='ipmi-fence-db01']/lrm_rsc_op[@id='ipmi-fence-db01_last_0']: @operation_key=ipmi-fence-db01_start_0, @operation=start, @transition-key=12:2485:0:27a91aab-060a-4de9-80b1-18abeb7bd850, @transition-magic=0:0;12:2485:0:27a91aab-060a-4de9-80b1-18abeb7bd850, @call-id=65, @last-run=1493087372, @last-rc-change=1493087372, @exec-time=45 Apr 25 10:29:33 [15330] gmlcdbw02 cib: info: cib_perform_op: + /cib/status/node_state[@id='2']/lrm[@id='2']/lrm_resources/lrm_resource[@id='ipmi-fence-db01']/lrm_rsc_op[@id='ipmi-fence-db01_last_0']: @operation_key=ipmi-fence-db01_start_0, @operation=start, @transition-key=12:2485:0:27a91aab-060a-4de9-80b1-18abeb7bd850, @transition-magic=0:0;12:2485:0:27a91aab-060a-4de9-80b1-18abeb7bd850, @call-id=65, @last-run=1493087372, @last-rc-change=1493087372, @exec-time=45 Apr 25 10:29:33 [15335] gmlcdbw02 crmd: info: match_graph_event: Action ipmi-fence-db01_start_0 (12) confirmed on gmlcdbw02 (rc=0) Apr 25 10:29:33 [15335] gmlcdbw02 crmd: notice: te_rsc_command: Initiating action 13: monitor ipmi-fence-db01_monitor_60000 on gmlcdbw02 (local) Apr 25 10:29:33 [15335] gmlcdbw02 crmd: info: do_lrm_rsc_op: Performing key=13:2485:0:27a91aab-060a-4de9-80b1-18abeb7bd850 op=ipmi-fence-db01_monitor_60000 Apr 25 10:29:33 [15335] gmlcdbw02 crmd: info: process_lrm_event: Operation ipmi-fence-db01_monitor_60000: ok (node=gmlcdbw02, call=66, rc=0, cib-update=2577, confirmed=false) Apr 25 10:29:33 [15330] gmlcdbw02 cib: info: cib_perform_op: + /cib/status/node_state[@id='2']/lrm[@id='2']/lrm_resources/lrm_resource[@id='ipmi-fence-db01']/lrm_rsc_op[@id='ipmi-fence-db01_monitor_60000']: @transition-key=13:2485:0:27a91aab-060a-4de9-80b1-18abeb7bd850, @transition-magic=0:0;13:2485:0:27a91aab-060a-4de9-80b1-18abeb7bd850, @call-id=66, @last-rc-change=1493087373, @exec-time=39 Apr 25 10:29:33 [15330] gmlcdbw02 cib: info: cib_perform_op: + /cib/status/node_state[@id='2']/lrm[@id='2']/lrm_resources/lrm_resource[@id='ipmi-fence-db01']/lrm_rsc_op[@id='ipmi-fence-db01_monitor_60000']: @transition-key=13:2485:0:27a91aab-060a-4de9-80b1-18abeb7bd850, @transition-magic=0:0;13:2485:0:27a91aab-060a-4de9-80b1-18abeb7bd850, @call-id=66, @last-rc-change=1493087373, @exec-time=39 Apr 25 10:29:33 [15335] gmlcdbw02 crmd: info: match_graph_event: Action ipmi-fence-db01_monitor_60000 (13) confirmed on gmlcdbw02 (rc=0) Apr 25 10:35:37 [15333] gmlcdbw02 attrd: info: write_attribute: Sent update 6 with 1 changes for fail-count-ipmi-fence-db02, id=<n/a>, set=(null) Apr 25 10:35:37 [15333] gmlcdbw02 attrd: info: attrd_cib_callback: Update 6 for fail-count-ipmi-fence-db02: OK (0) Apr 25 10:35:37 [15333] gmlcdbw02 attrd: info: attrd_cib_callback: Update 6 for fail-count-ipmi-fence-db02[gmlcdbw02]=(null): OK (0) Apr 25 10:35:37 [15334] gmlcdbw02 pengine: info: native_print: ipmi-fence-db01 (stonith:fence_ipmilan): Started gmlcdbw02 Apr 25 10:35:37 [15334] gmlcdbw02 pengine: info: native_print: ipmi-fence-db02 (stonith:fence_ipmilan): Started gmlcdbw02 Apr 25 10:35:37 [15334] gmlcdbw02 pengine: info: native_color: Resource ipmi-fence-db01 cannot run anywhere Apr 25 10:35:37 [15334] gmlcdbw02 pengine: info: native_color: Resource ipmi-fence-db02 cannot run anywhere Apr 25 10:35:37 [15334] gmlcdbw02 pengine: notice: LogActions: Stop ipmi-fence-db01 (gmlcdbw02) Apr 25 10:35:37 [15334] gmlcdbw02 pengine: notice: LogActions: Stop ipmi-fence-db02 (gmlcdbw02) Apr 25 10:35:37 [15335] gmlcdbw02 crmd: notice: te_rsc_command: Initiating action 10: stop ipmi-fence-db01_stop_0 on gmlcdbw02 (local) if i create location constraint, can i force ipmi-fence-db01 keep on the gmlcdbw01? # pcs constraint location ipmi-fence-db01 prefers gmlcdbw01 # pcs constraint location ipmi-fence-db02 prefers gmlcdbw02 Thanks. On Tue, May 2, 2017 at 9:39 AM, Albert Weng <weng.alb...@gmail.com> wrote: > Hi All, > > I have created active/passive pacemaker cluster on RHEL 7. > > here is my environment: > clustera : 192.168.11.1 > clusterb : 192.168.11.2 > clustera-ilo4 : 192.168.11.10 > clusterb-ilo4 : 192.168.11.11 > > both nodes are connected SAN storage for shared storage. > > i used the following cmd to create my stonith devices on each node : > # pcs -f stonith_cfg stonith create ipmi-fence-node1 fence_ipmilan parms > lanplus="ture" pcmk_host_list="clustera" pcmk_host_check="static-list" > action="reboot" ipaddr="192.168.11.10" login=adminsitrator passwd=1234322 > op monitor interval=60s > > # pcs -f stonith_cfg stonith create ipmi-fence-node02 fence_ipmilan parms > lanplus="true" pcmk_host_list="clusterb" pcmk_host_check="static-list" > action="reboot" ipaddr="192.168.11.11" login=USERID passwd=password op > monitor interval=60s > > # pcs status > ipmi-fence-node1 clustera > ipmi-fence-node2 clusterb > > but when i failover to passive node, then i ran > # pcs status > > ipmi-fence-node1 clusterb > ipmi-fence-node2 clusterb > > why both fence device locate on the same node ? > > > -- > Kind regards, > Albert Weng > > > <https://www.avast.com/sig-email?utm_medium=email&utm_source=link&utm_campaign=sig-email&utm_content=webmail> > 不含病毒。www.avast.com > <https://www.avast.com/sig-email?utm_medium=email&utm_source=link&utm_campaign=sig-email&utm_content=webmail> > <#m_-4327774609408162231_DAB4FAD8-2DD7-40BB-A1B8-4E2AA1F9FDF2> > -- Kind regards, Albert Weng
_______________________________________________ Users mailing list: Users@clusterlabs.org http://lists.clusterlabs.org/mailman/listinfo/users Project Home: http://www.clusterlabs.org Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf Bugs: http://bugs.clusterlabs.org