Re: [ClusterLabs] Pacemaker failover failure

2015-07-14 Thread Digimer
As said before, fencing. On 01/07/15 06:54 AM, alex austin wrote: > so did another test: > > two nodes: node1 and node2 > > Case: node1 is the active node > node2: is pasive > > if I killall -9 pacemakerd corosync on node 1 the services do not fail > over to node2, but if I start corosync and p

Re: [ClusterLabs] Pacemaker failover failure

2015-07-14 Thread Andrei Borzenkov
В Wed, 15 Jul 2015 00:31:45 +0100 alex austin пишет: > Unfortunately I have nothing yet ... > > There's something I don't quite understand though. What's the role of > stonith if the other machine crashes unexpectedly and totally unclean? Is > it to reboot the machine and recreate the cluster, t

Re: [ClusterLabs] Pacemaker failover failure

2015-07-14 Thread alex austin
Unfortunately I have nothing yet ... There's something I don't quite understand though. What's the role of stonith if the other machine crashes unexpectedly and totally unclean? Is it to reboot the machine and recreate the cluster, thus making the drbd volume available again? or is it other? The

[ClusterLabs] [Announce] clufter-0.50.1 released

2015-07-14 Thread Jan Pokorný
I am happy to announce that clufter-0.50.1, a tool/library for transforming/analyzing cluster configuration formats, has been released and published (incl. signature using my 60BCBB4F5CD7F9EF key):

Re: [ClusterLabs] Forcing rgmanager to re-read the config

2015-07-14 Thread Digimer
On 14/07/15 10:34 AM, Jan Pokorný wrote: > On 08/07/15 17:41 -0700, Digimer wrote: >> Second time in a couple of years I hit a weird bug... I have no idea how >> to reproduce it, sadly. >> >> I added a VM to cluster.conf via ccs and it showed up on node 1 but not >> node 2. I confirmed that node 2

Re: [ClusterLabs] Forcing rgmanager to re-read the config

2015-07-14 Thread Jan Pokorný
On 08/07/15 17:41 -0700, Digimer wrote: > Second time in a couple of years I hit a weird bug... I have no idea how > to reproduce it, sadly. > > I added a VM to cluster.conf via ccs and it showed up on node 1 but not > node 2. I confirmed that node 2 had the updated cluster.conf, but the > running