Hello Digimer,
Yes, I would agrre with you RHEL6.4 is old. We patched monthly, but I'm not
sure why these servers are still at 6.4. Most of our system are 6.6.
Here is my cluster config. All I want is using cluster to have BGFS2 mount via
/etc/fstab.
root@ustlvcmsp1955 ~]# cat /etc/cluster/cluster.conf
<?xml version="1.0"?>
<cluster config_version="15" name="p1954_to_p1958">
<clusternodes>
<clusternode name="ustlvcmsp1954" nodeid="1"/>
<clusternode name="ustlvcmsp1955" nodeid="2"/>
<clusternode name="ustlvcmsp1956" nodeid="3"/>
<clusternode name="ustlvcmsp1957" nodeid="4"/>
<clusternode name="ustlvcmsp1958" nodeid="5"/>
</clusternodes>
<fencedevices>
<fencedevice agent="fence_vmware_soap" ipaddr="10.30.197.108"
login="rhfence" name="p1954" passwd="xxxxxxxx"/>
<fencedevice agent="fence_vmware_soap" ipaddr="10.30.197.109"
login="rhfence" name="p1955" passwd=" xxxxxxxx "/>
<fencedevice agent="fence_vmware_soap" ipaddr="10.30.197.110"
login="rhfence" name="p1956" passwd=" xxxxxxxx "/>
<fencedevice agent="fence_vmware_soap" ipaddr="10.30.197.111"
login="rhfence" name="p1957" passwd=" xxxxxxxx "/>
<fencedevice agent="fence_vmware_soap" ipaddr="10.30.197.112"
login="rhfence" name="p1958" passwd=" xxxxxxxx "/>
</fencedevices>
</cluster>
clustat show:
Cluster Status for p1954_to_p1958 @ Wed Jan 7 15:38:00 2015
Member Status: Quorate
Member Name ID Status
------ ---- ---- ------
ustlvcmsp1954 1 Offline
ustlvcmsp1955 2 Online,
Local
ustlvcmsp1956 3 Online
ustlvcmsp1957 4 Offline
ustlvcmsp1958 5 Online
I need to make them all online, so I can use fencing for mounting shared disk.
Thanks,
Vinh
-----Original Message-----
From: [email protected]
[mailto:[email protected]] On Behalf Of Digimer
Sent: Wednesday, January 07, 2015 3:16 PM
To: linux clustering
Subject: Re: [Linux-cluster] needs helps GFS2 on 5 nodes cluster
My first though would be to set <fence_daemon post_join_delay="30" /> in
cluster.conf.
If that doesn't work, please share your configuration file. Then, with all
nodes offline, open a terminal to each node and run 'tail -f -n 0
/var/log/messages'. With that running, start all the nodes and wait for things
to settle down, then paste the five nodes' output as well.
Also, 6.4 is pretty old, why not upgrade to 6.6?
digimer
On 07/01/15 03:10 PM, Cao, Vinh wrote:
> Hello Cluster guru,
>
> I'm trying to setup Redhat 6.4 OS cluster with 5 nodes. With two nodes
> I don't have any issue.
>
> But with 5 nodes, when I ran clustat I got 3 nodes online and the
> other two off line.
>
> When I start the one that are off line. Service cman start. I got:
>
> [root@ustlvcmspxxx ~]# service cman status
>
> corosync is stopped
>
> [root@ustlvcmsp1954 ~]# service cman start
>
> Starting cluster:
>
> Checking if cluster has been disabled at boot... [ OK ]
>
> Checking Network Manager... [ OK ]
>
> Global setup... [ OK ]
>
> Loading kernel modules... [ OK ]
>
> Mounting configfs... [ OK ]
>
> Starting cman... [ OK ]
>
> Waiting for quorum... Timed-out waiting for cluster
>
> [FAILED]
>
> Stopping cluster:
>
> Leaving fence domain... [ OK ]
>
> Stopping gfs_controld... [ OK ]
>
> Stopping dlm_controld... [ OK ]
>
> Stopping fenced... [ OK ]
>
> Stopping cman... [ OK ]
>
> Waiting for corosync to shutdown: [ OK ]
>
> Unloading kernel modules... [ OK ]
>
> Unmounting configfs... [ OK ]
>
> Can you help?
>
> Thank you,
>
> Vinh
>
>
>
--
Digimer
Papers and Projects: https://alteeve.ca/w/ What if the cure for cancer is
trapped in the mind of a person without access to education?
--
Linux-cluster mailing list
[email protected]
https://www.redhat.com/mailman/listinfo/linux-cluster
--
Linux-cluster mailing list
[email protected]
https://www.redhat.com/mailman/listinfo/linux-cluster