I'm trying to understand what and how crm-fence-peer.sh does what it does. I'm using DRBD 8.4 with Pacemaker in a two node cluster, with a single primary.
I'm doing fabric fencing. And I believe I have my stonith setup configured correctly. I have the following in my drbd config: handlers { fence-peer "/usr/lib/drbd/crm-fence-peer.sh"; after-resync-target "/usr/lib/drbd/crm-unfence-peer.sh"; ... } disk { fencing resource-and-stonith; ... } I can run "stonith_admin -F <node>" and the switch ports for the other node will get disabled. Similarly, I can run "pcs stonith fence <node> --off" and I will get the same result. Now, I'm doing more fence testing, like running "init 6" on the primary node that is also the DC. When I do this, I see the following in logs of the surviving node: Jan 11 08:49:53 storage2 kernel: drbd r0: helper command: /sbin/drbdadm fence-peer r0 Jan 11 08:49:53 storage2 crm-fence-peer.sh[15594]: DRBD_CONF=/etc/drbd.conf DRBD_DONT_WARN_ON_VERSION_MISMATCH=1 DRBD_MINOR=1 DRBD_PEER=storage1 DRBD_PEERS=storage1 DRBD_PEER_ADDRESS=192.168.0.2 DRBD_PEER_AF=ipv4 DRBD_RESOURCE=r0 UP_TO_DATE_NODES='' /usr/lib/drbd/crm-fence-peer.sh Jan 11 08:49:53 storage2 crm-fence-peer.sh[15594]: INFO peer is reachable, my disk is UpToDate: placed constraint 'drbd-fence-by-handler-r0-StorageClusterClone' Jan 11 08:49:53 storage2 kernel: drbd r0: helper command: /sbin/drbdadm fence-peer r0 exit code 4 (0x400) Jan 11 08:49:53 storage2 kernel: drbd r0: fence-peer helper returned 4 (peer was fenced) But the switch ports connected to the fenced node are still enabled. What am I missing here? Thanks! Bryan Walton -- Bryan K. Walton 319-337-3877 Linux Systems Administrator Leepfrog Technologies, Inc _______________________________________________ drbd-user mailing list drbd-user@lists.linbit.com http://lists.linbit.com/mailman/listinfo/drbd-user