Hello Pascal

For disable startup fencing you need clean_start=1 in the fence_daemon tag,
i saw in your previous mail you are using expected_votes="1", with this
setting every cluster node will be partitioned into two clusters and
operate independently, i recommended using a quorim disk with master_wins
parameter


2013/9/11 Pascal Ehlert <pas...@hacksrus.net>

> Hi,
>
> I have recently setup an HA cluster with two nodes, IPMI based fencing
> and no quorum disk. Things worked nicely during the first tests, but to my
> very annoyance it blew up last night when I did another test of shutting
> down the network interface on my secondary node (node 2).
>
> The node was fenced as expected and came back online. This however
> resulted in an immediate fencing of the other node.
> Fencing went back and forth until I manually powered of node 2 and let
> node 1 a few minutes to settle down.
>
> Now when I switch node 2 back on, it looks like it joins the cluster and
> is kicked out immediately again, which again results in fencing of node
> 2. I have purposely set the post_join_delay to a high value, but it
> didn't help.
>
> Below are my cluster.conf and log files. My own guess would be that the
> problem is associated with the fact that the node tries to do a stateful
> merge, when it really should be joining without state after a clean
> reboot. (see fence_tool dump line 9).
>
> --------------
> root@rmg-de-1:~# cat /etc/pve/cluster.conf
> <?xml version="1.0"?>
> <cluster config_version="14" name="rmg-de-cl1">
>   <cman expected_votes="1" keyfile="/var/lib/pve-cluster/corosync.authkey"
> two_node="1"/>
>   <fencedevices>
>     <fencedevice agent="fence_ipmilan" ipaddr="10.xx.xx.11"
> login="FENCING" name="fenceNode1" passwd="abc"/>
>     <fencedevice agent="fence_ipmilan" ipaddr="10.xx.xx.12"
> login="FENCING" name="fenceNode2" passwd="abc"/>
>   </fencedevices>
>   <clusternodes>
>     <clusternode name="rmg-de-1" nodeid="1" votes="1">
>       <fence>
>         <method name="1">
>           <device action="reboot" name="fenceNode1"/>
>         </method>
>       </fence>
>     </clusternode>
>     <clusternode name="rmg-de-2" nodeid="2" votes="1">
>       <fence>
>         <method name="1">
>           <device action="reboot" name="fenceNode2"/>
>         </method>
>       </fence>
>     </clusternode>
>   </clusternodes>
>   <fence_daemon post_join_delay="360" />
>   <rm>
>     <pvevm autostart="1" vmid="101"/>
>     <pvevm autostart="1" vmid="100"/>
>     <pvevm autostart="1" vmid="104"/>
>     <pvevm autostart="1" vmid="103"/>
>     <pvevm autostart="1" vmid="102"/>
>   </rm>
> </cluster>
> --------------
>
> --------------
> root@rmg-de-1:~# fence_tool dump | tail -n 40
> 1378890849 daemon node 1 max 1.1.1.0 run 1.1.1.1
> 1378890849 daemon node 1 join 1378855487 left 0 local quorum 1378855487
> 1378890849 receive_start 1:12 len 152
> 1378890849 match_change 1:12 matches cg 12
> 1378890849 wait_messages cg 12 need 1 of 2
> 1378890850 receive_protocol from 2 max 1.1.1.0 run 1.1.1.1
> 1378890850 daemon node 2 max 0.0.0.0 run 0.0.0.0
> 1378890850 daemon node 2 join 1378890849 left 1378859110 local quorum
> 1378855487
> 1378890850 daemon node 2 stateful merge
> 1378890850 daemon node 2 kill due to stateful merge
> 1378890850 telling cman to remove nodeid 2 from cluster
> 1378890862 cluster node 2 removed seq 832
> 1378890862 fenced:daemon conf 1 0 1 memb 1 join left 2
> 1378890862 fenced:daemon ring 1:832 1 memb 1
> 1378890862 fenced:default conf 1 0 1 memb 1 join left 2
> 1378890862 add_change cg 13 remove nodeid 2 reason 3
> 1378890862 add_change cg 13 m 1 j 0 r 1 f 1
> 1378890862 add_victims node 2
> 1378890862 check_ringid cluster 832 cpg 1:828
> 1378890862 fenced:default ring 1:832 1 memb 1
> 1378890862 check_ringid done cluster 832 cpg 1:832
> 1378890862 check_quorum done
> 1378890862 send_start 1:13 flags 2 started 6 m 1 j 0 r 1 f 1
> 1378890862 cpg_mcast_joined retried 1 start
> 1378890862 receive_start 1:13 len 152
> 1378890862 match_change 1:13 skip cg 12 already start
> 1378890862 match_change 1:13 matches cg 13
> 1378890862 wait_messages cg 13 got all 1
> 1378890862 set_master from 1 to complete node 1
> 1378890862 delay post_join_delay 360 quorate_from_last_update 0
> 1378891222 delay of 360s leaves 1 victims
> 1378891222 rmg-de-2 not a cluster member after 360 sec post_join_delay
> 1378891222 fencing node rmg-de-2
> 1378891236 fence rmg-de-2 dev 0.0 agent fence_ipmilan result: success
> 1378891236 fence rmg-de-2 success
> 1378891236 send_victim_done cg 13 flags 2 victim nodeid 2
> 1378891236 send_complete 1:13 flags 2 started 6 m 1 j 0 r 1 f 1
> 1378891236 receive_victim_done 1:13 flags 2 len 80
> 1378891236 receive_victim_done 1:13 remove victim 2 time 1378891236 how 1
> 1378891236 receive_complete 1:13 len 152:
> --------------
>
> --------------
> root@rmg-de-1:~# tail -n 100 /var/log/cluster/corosync.log
> Sep 11 11:14:09 corosync [CLM   ] CLM CONFIGURATION CHANGE
> Sep 11 11:14:09 corosync [CLM   ] New Configuration:
> Sep 11 11:14:09 corosync [CLM   ]     r(0) ip(10.xx.xx.1)
> Sep 11 11:14:09 corosync [CLM   ] Members Left:
> Sep 11 11:14:09 corosync [CLM   ] Members Joined:
> Sep 11 11:14:09 corosync [CLM   ] CLM CONFIGURATION CHANGE
> Sep 11 11:14:09 corosync [CLM   ] New Configuration:
> Sep 11 11:14:09 corosync [CLM   ]     r(0) ip(10.xx.xx.1)
> Sep 11 11:14:09 corosync [CLM   ]     r(0) ip(10.xx.xx.2)
> Sep 11 11:14:09 corosync [CLM   ] Members Left:
> Sep 11 11:14:09 corosync [CLM   ] Members Joined:
> Sep 11 11:14:09 corosync [CLM   ]     r(0) ip(10.xx.xx.2)
> Sep 11 11:14:09 corosync [TOTEM ] A processor joined or left the
> membership and a new membership was formed.
> Sep 11 11:14:09 corosync [QUORUM] Members[2]: 1 2
> Sep 11 11:14:09 corosync [QUORUM] Members[2]: 1 2
> Sep 11 11:14:09 corosync [CPG   ] chosen downlist: sender r(0)
> ip(10.xx.xx.1) ; members(old:1 left:0)
> Sep 11 11:14:09 corosync [MAIN  ] Completed service synchronization, ready
> to provide service.
> Sep 11 11:14:20 corosync [TOTEM ] A processor failed, forming new
> configuration.
> Sep 11 11:14:22 corosync [CLM   ] CLM CONFIGURATION CHANGE
> Sep 11 11:14:22 corosync [CLM   ] New Configuration:
> Sep 11 11:14:22 corosync [CLM   ]     r(0) ip(10.xx.xx.1)
> Sep 11 11:14:22 corosync [CLM   ] Members Left:
> Sep 11 11:14:22 corosync [CLM   ]     r(0) ip(10.xx.xx.2)
> Sep 11 11:14:22 corosync [CLM   ] Members Joined:
> Sep 11 11:14:22 corosync [QUORUM] Members[1]: 1
> Sep 11 11:14:22 corosync [CLM   ] CLM CONFIGURATION CHANGE
> Sep 11 11:14:22 corosync [CLM   ] New Configuration:
> Sep 11 11:14:22 corosync [CLM   ]     r(0) ip(10.xx.xx.1)
> Sep 11 11:14:22 corosync [CLM   ] Members Left:
> Sep 11 11:14:22 corosync [CLM   ] Members Joined:
> Sep 11 11:14:22 corosync [TOTEM ] A processor joined or left the
> membership and a new membership was formed.
> Sep 11 11:14:22 corosync [CPG   ] chosen downlist: sender r(0)
> ip(10.xx.xx.1) ; members(old:2 left:1)
> Sep 11 11:14:22 corosync [MAIN  ] Completed service synchronization, ready
> to provide service.
> --------------
>
> --------------
> root@rmg-de-1:~# dlm_tool ls
> dlm lockspaces
> name          rgmanager
> id            0x5231f3eb
> flags         0x00000000
> change        member 1 joined 0 remove 1 failed 1 seq 12,13
> members       1
> --------------
>
> Unfortunately I only have the output of the currently operational node,
> as the other one is fenced very quickly and the logs are hard to
> retrieve. If someone has an idea however, I'll do my best to provide
> these as well.
>
> Thanks,
>
> Pascal
>
> --
> Linux-cluster mailing list
> Linux-cluster@redhat.com
> https://www.redhat.com/mailman/listinfo/linux-cluster
>



-- 
esta es mi vida e me la vivo hasta que dios quiera
-- 
Linux-cluster mailing list
Linux-cluster@redhat.com
https://www.redhat.com/mailman/listinfo/linux-cluster

Reply via email to