I'm having a bit of an issue under centos 6.4 x64. I have two duplcate hardware systems (raid arrays, 10G nics' etc) configured identically and drbd replication is working fine in the cluster between the two. When I started doing fail over testing, via power/communcations interruptions I found that I could not reliably shift the resources from cluster1 to cluster2 even though they are identical in every aspect. I AM able (by starting pacemaker first on one or the other) to get the cluster up on either node. I was told that this is a problem for a non-stonith 2 node cluster and to add a third server to provide the quorum vote to tell the survivor to host the cluster resources. Ok, thats the background, I've been trying to use, with limited success, the location resource rules in crmsh to say node1 and node2 can run whatever-they have the hardware and installed resources. but node3 cannot run any of the resources something akin to: location NeverRunDRBD ms_drbd_r0 \ rule $id="NeverRunDRBD-rule" $role="Master" -inf: #uname eq node3 and since all other resources are order and collocation constrained, it follows suit that they have to stay on node1 or node2, thats fine and appears to work, but on node3 (no drbd, no raid array not even close in hardware) I still get messages to the effect that the p_drbd_r0 monitor wont load becasue its not installed....well duh, its not installed but its not suppose to be running on node3 anyway why is trying to monitor on a node its not installed or permitted to run on ?
Incidentally I see the same behaviour sometimes when a ocf:heartbeat:exportfs is on node1 that the monitor for the same resource is trying to run on node2 where where the drbd volume is in secondary mode and not locally accessible nor should it be. I'm not asking for a fix, though I'd take one, I'm asking (no /var/log/messages is pretty useless on this issue), what other logs or debug flags can I toggle to help point out the way or does PCS / crm_xxxxx provide a better interface for configuring/debugging this ? Lastly, during my failover testing and configuration testing, I found the only surefire way to apply a new cluster config, is to cibadmin -f -E and cut and paste in a new one followed by a reboot....what a pain. You can sometimes get away with restarting pacemaker on all nodes, bringing up your intended primary first then the others later -- View this message in context: http://linux-ha.996297.n3.nabble.com/2-node-clustes-with-seperate-quorum-server-tp15159.html Sent from the Linux-HA mailing list archive at Nabble.com. _______________________________________________ Linux-HA mailing list Linux-HA@lists.linux-ha.org http://lists.linux-ha.org/mailman/listinfo/linux-ha See also: http://linux-ha.org/ReportingProblems