Hi,

On 3/22/2011 11:12 AM, Gianluca Cecchi wrote:
> Hello,
> I'm using latest updates on a 2 nodes rhel 6 based cluster.
> At the moment no quorum disk defined, so this line inside cluster.conf
> <cman expected_votes="1" two_node="1"/>
> 
> # rpm -q cman rgmanager fence-agents ricci corosync
> cman-3.0.12-23.el6_0.6.x86_64
> rgmanager-3.0.12-10.el6.x86_64
> fence-agents-3.0.12-8.el6_0.3.x86_64
> ricci-0.16.2-13.el6.x86_64
> corosync-1.2.3-21.el6_0.1.x86_64
> 
> # uname -r
> 2.6.32-71.18.2.el6.x86_64

For RHEL related questions you should always file a ticket with GSS.

> 
> If the initial situation is both nodes down and I start one of them, I
> get it powering on the other, that is not my intentional target...
> Is this an expected default behaviour in rh el 6 with two nodes
> without quorum disk? Or in general no matter if a quorum disk is
> defined?
> If so, how to change it if possible?

This is expected behavior.

The node that is booting/powering on, will gain quorum by itself and
since it does not detect the other node for N amount of seconds, it will
perform a fencing action to make sure the node is not accessing any
shared resource.

I am not sure why you want a one node cluster, but one easy workaround
is to start both of them at the same time, and then shutdown one of
them. At that point they have both seen each other and the one going
down will tell the other "I am going offline, no worries, it´s all good".

Fabio

--
Linux-cluster mailing list
Linux-cluster@redhat.com
https://www.redhat.com/mailman/listinfo/linux-cluster

Reply via email to