Hi
I had/have similar problems

I have two nics:
eth0 direct connection without a switch between the two nodes (10.0.0.x) , 
hostnames: node1-direct, node2-direct
eth1 with a switch 192.168.x.x (hostnames: node1, node2)

using the node[12]-direct hostname in a simple cluster.conf a ran in the
same problem like ITec.

replacing the hostnames (-> node[12]) it works. adding altname
to the corresponding clusternode it works:
<altname name="node1-direct" port="5406" mcast="239.192.122.46" />


running: corosync-cfgtool -s # it show the two rings
Printing ring status.
Local node ID 2
RING ID 0
        id      = 192.168.25.52
        status  = ring 0 active with no faults
RING ID 1
        id      = 10.0.0.52
        status  = ring 1 active with no faults

maybe the helps..

KLor

-- 
You received this bug notification because you are a member of Ubuntu
Bugs, which is subscribed to Ubuntu.
https://bugs.launchpad.net/bugs/956383

Title:
  ccs_config_validate exits with 191 and broken link for cluster.rng

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/redhat-cluster/+bug/956383/+subscriptions

-- 
ubuntu-bugs mailing list
ubuntu-bugs@lists.ubuntu.com
https://lists.ubuntu.com/mailman/listinfo/ubuntu-bugs

Reply via email to