Re: [Pacemaker] Pacemaker and OpenStack RA

2013-09-10 Thread Andrew Beekhof
I replied to your entry here:
   
http://openstack.redhat.com/forum/discussion/538/pacemaker-and-openstack-ra-issues#Item_3

But for completeness:

> Any ideas?
> 

This:

> p_glance-api_start_0 (node=main-controller, call=71, rc=1, status=Timed Out): 
> unknown error
> 

suggests that there was not enough time for glance to start before the default 
20s timeout.

Without knowing anything about glance, I can't say if that's unusual and 
without logs it's impossible to know what the agent was doing for those 20s.


On 10/09/2013, at 1:24 AM, Mael Lavault  wrote:

> 
> Hi,
> 
> I'm setting up an Active/Passive configuration for OpenStack using 2
> nodes (2 controller nodes).
> 
> I'm having trouble to make some OpenStack ressources agent work.
> 
> pcs status command give me this :
> 
> Last updated: Mon Sep  9 11:05:13 2013
> Last change: Fri Sep  6 17:15:09 2013 via cibadmin on main-controller
> Stack: classic openais (with plugin)
> Current DC: main-controller - partition with quorum
> Version: 1.1.8-7.el6-394e906
> 2 Nodes configured, 2 expected votes
> 11 Resources configured.
> 
> 
> Online: [ main-controller secondary-controller ]
> 
> Full list of resources:
> 
> Resource Group: g_mysql
> p_ip_mysql(ocf::heartbeat:IPaddr2): Started main-controller
> p_fs_mysql(ocf::heartbeat:Filesystem): Started main-controller
> p_mysql   (ocf::heartbeat:mysql): Started main-controller
> Master/Slave Set: ms_drbd_mysql [p_drbd_mysql]
> Masters: [ main-controller ]
> Slaves: [ secondary-controller ]
> Resource Group: g_services_api
> p_api-ip(ocf::heartbeat:IPaddr2): Started secondary-controller
> p_keystone(ocf::openstack:keystone): Started secondary-controller
>p_glance-api(ocf::openstack:glance-api): Stopped 
> p_cinder-api(ocf::openstack:cinder-api): Stopped 
> p_neutron-server(ocf::openstack:neutron-server): Stopped 
> p_glance-registry(ocf::openstack:glance-registry): Stopped 
> 
> Failed actions:
>p_glance-api_start_0 (node=main-controller, call=71, rc=1,
> status=Timed Out): unknown error
>p_glance-api_start_0 (node=secondary-controller, call=71, rc=1,
> status=Timed Out): unknown error
> 
> Here is my conf: 
> 
> Corosync Nodes:
> 
> Pacemaker Nodes:
> main-controller secondary-controller 
> 
> Resources: 
> Group: g_mysql
>  Resource: p_ip_mysql (provider=heartbeat type=IPaddr2 class=ocf)
>   Attributes: ip=192.168.16.49 cidr_netmask=24 
>  Resource: p_fs_mysql (provider=heartbeat type=Filesystem class=ocf)
>   Attributes: device=/dev/drbd0 directory=/var/lib/mysql fstype=xfs 
>  Resource: p_mysql (provider=heartbeat type=mysql class=ocf)
>   Attributes: config=/etc/my.cnf pid=/var/run/mysqld/mysqld.pid 
> socket=/var/run/mysqld/mysqld.sock log=/var/log/mysqld.log 
> Master: ms_drbd_mysql
>  Resource: p_drbd_mysql (provider=linbit type=drbd class=ocf)
>   Attributes: drbd_resource=mysql 
>   Operations: start interval=29s role=Master timeout=180s
> Group: g_services_api
>  Resource: p_api-ip (provider=heartbeat type=IPaddr2 class=ocf)
>   Attributes: ip=192.168.16.48 cidr_netmask=24 
>   Operations: monitor interval=30s
>  Resource: p_keystone (provider=openstack type=keystone class=ocf)
>   Attributes: config=/etc/keystone/keystone.conf os_password=8d99add1e17545f0 
> os_username=admin os_tenant_name=admin 
> os_auth_url=http://192.168.16.48:5000/v2.0/ user=keystone 
>   Operations: monitor interval=30s timeout=30s
>  Resource: p_glance-api (provider=openstack type=glance-api class=ocf)
>   Attributes: config=/etc/glance/glance-api.conf os_password=8d99add1e17545f0 
> os_username=admin os_tenant_name=admin 
> os_auth_url=http://192.168.16.48:5000/v2.0/ 
>   Operations: monitor interval=30s timeout=30s
>  Resource: p_cinder-api (provider=openstack type=cinder-api class=ocf)
>   Attributes: config=/etc/cinder/cinder.conf os_password=8d99add1e17545f0 
> os_username=admin os_tenant_name=admin 
> os_auth_url=http://192.168.16.48:5000/v2.0/ 
>   Operations: monitor interval=30s timeout=30s
>  Resource: p_neutron-server (provider=openstack type=neutron-server class=ocf)
>   Attributes: os_password=8d99add1e17545f0 os_username=admin 
> os_tenant_name=admin keystone_get_token_url=http://192.168.16.48:5000/v2.0/ 
>   Operations: monitor interval=30s timeout=30s
>  Resource: p_glance-registry (provider=openstack type=glance-registry 
> class=ocf)
>   Attributes: config=/etc/glance/glance-registry.conf 
> os_password=8d99add1e17545f0 os_username=admin os_tenant_name=admin 
> os_auth_url=http://192.168.16.48:5000/v2.0/ 
>   Operations: monitor interval=30s timeout=30s
> 
> Location Constraints:
> Ordering Constraints:
>  promote ms_drbd_mysql then start g_mysql
> Colocation Constraints:
>  g_mysql with ms_drbd_mysql (with-rsc-role:Master)
> 
> Cluster Properties:
> dc-version: 1.1.8-7.el6-394e906
> cluster-infrastructure: classic openais (with plugin)
> expected-quorum-votes: 2
> no-quorum-policy: ignore
> pe-warn-series-max: 1000
> pe-input-series-max: 1000
> pe-erro

Re: [Pacemaker] different behavior cibadmin -Ql with cman and corosync2

2013-09-10 Thread Andrey Groshev
Hello Christine, Andrew and all.

I'm sorry - a little was unwell, so did not answer.
What we end this stream of messages?
Who will change? corosync or pacemaker?


05.09.2013, 15:49, "Christine Caulfield" :
> On 05/09/13 11:33, Andrew Beekhof wrote:
>
>>  On 05/09/2013, at 6:37 PM, Christine Caulfield  wrote:
>>>  On 03/09/13 22:03, Andrew Beekhof wrote:
  On 03/09/2013, at 11:49 PM, Christine Caulfield  
 wrote:
>  On 03/09/13 05:20, Andrew Beekhof wrote:
>>  On 02/09/2013, at 5:27 PM, Andrey Groshev  wrote:
>>>  30.08.2013, 07:18, "Andrew Beekhof" :
  On 29/08/2013, at 7:31 PM, Andrey Groshev  wrote:
>    29.08.2013, 12:25, "Andrey Groshev" :
>>    29.08.2013, 02:55, "Andrew Beekhof" :
>>> On 28/08/2013, at 5:38 PM, Andrey Groshev  
>>> wrote:
  28.08.2013, 04:06, "Andrew Beekhof" :
>  On 27/08/2013, at 1:13 PM, Andrey Groshev  
> wrote:
>>   27.08.2013, 05:39, "Andrew Beekhof" :
>>>   On 26/08/2013, at 3:09 PM, Andrey Groshev 
>>>  wrote:
    26.08.2013, 03:34, "Andrew Beekhof" 
 :
>    On 23/08/2013, at 9:39 PM, Andrey Groshev 
>  wrote:
>> Hello,
>>
>> Today I try remake my test cluster from cman to 
>> corosync2.
>> I drew attention to the following:
>> If I reset cluster with cman through cibadmin 
>> --erase --force
>> In cib is still there exist names of nodes.
>    Yes, the cluster puts back entries for all the nodes 
> it know about automagically.
>> cibadmin -Ql
>> .
>>    
>>  > uname="dev-cluster2-node2"/>
>>  > uname="dev-cluster2-node4"/>
>>  > uname="dev-cluster2-node3"/>
>>    
>> 
>>
>> Even if cman and pacemaker running only one node.
>    I'm assuming all three are configured in cluster.conf?
    Yes, there exist list nodes.
>> And if I do too on cluster with corosync2
>> I see only names of nodes which run corosync and 
>> pacemaker.
>    Since you're not included your config, I can only 
> guess that your corosync.conf does not have a nodelist.
>    If it did, you should get the same behaviour.
    I try and expected_node and nodelist.
>>>   And it didn't work? What version of pacemaker?
>>   It does not work as I expected.
>  Thats because you've used IP addresses in the node list.
>  ie.
>
>  node {
>    ring0_addr: 10.76.157.17
>  }
>
>  try including the node name as well, eg.
>
>  node {
>    name: dev-cluster2-node2
>    ring0_addr: 10.76.157.17
>  }
  The same thing.
>>> I don't know what to say.  I tested it here yesterday and it 
>>> worked as expected.
>>    I found that the reason that You and I have different results - I 
>> did not have reverse DNS zone for these nodes.
>>    I know what it should be, but (PACEMAKER + CMAN) worked without a 
>> reverse area!
>    Hasty. Deleted all. Reinstalled. Configured. Not working again. 
> Damn!
  It would have surprised me... pacemaker 1.1.11 doesn't do any dns 
 lookups - reverse or otherwise.
  Can you set

    PCMK_trace_files=corosync.c

  in your environment and retest?

  On RHEL6 that means putting the following in /etc/sysconfig/pacemaker
 export PCMK_trace_files=corosync.c

  It should produce additional logging[1] that will help diagnose the 
 issue.

  [1] http://blog.clusterlabs.org/blog/2013/pacemaker-logging/
>>>  Hello, Andrew.
>>>
>>>  You are a little misunderstood me.
>>  No, I understood you fine.
>>>  I wrote that I rushed to judgment.
>>>  After I did the reverse DNS zone, the cluster behaved correctly.
>>>  BUT after I took apart the cluster dropped configs and restarted on 
>>> the new cluster,
>>>  cluster again don't showed all the nodes in the nodes (only node with 
>>> running pacemaker).
>>>
>>>  A small portion of the log. Full log

Re: [Pacemaker] Moving cloned resources

2013-09-10 Thread Andrew Beekhof

On 11/09/2013, at 12:58 AM, Matias R. Cuenca del Rey  wrote:

> Thank you Andrew,
> 
> If I can't move instances of a clone. How can I do to rebalance my IP 
> resource trough the three nodes?

Set its resource-stickiness to 0 until everything moves back

> 
> Thanks,
> 
> Matías R. Cuenca del Rey
> 
> 
> On Fri, Aug 9, 2013 at 12:15 PM, Matias R. Cuenca del Rey  
> wrote:
> [root@www-proxylb01 ~]# rpm -qi pcs
> Name: pcs  Relocations: (not relocatable)
> Version : 0.9.26Vendor: CentOS
> Release : 10.el6_4.1Build Date: Mon 18 Mar 2013 
> 11:39:25 AM ART
> Install Date: Tue 04 Jun 2013 05:19:49 PM ART  Build Host: 
> c6b10.bsys.dev.centos.org
> Group   : System Environment/Base   Source RPM: 
> pcs-0.9.26-10.el6_4.1.src.rpm
> Size: 254791   License: GPLv2
> Signature   : RSA/SHA1, Mon 18 Mar 2013 12:01:56 PM ART, Key ID 
> 0946fca2c105b9de
> Packager: CentOS BuildSystem 
> URL : http://github.com/feist/pcs
> Summary : Pacemaker Configuration System
> Description :
> pcs is a corosync and pacemaker configuration tool.  It permits users to
> easily view, modify and created pacemaker based clusters.
> 
> 
> [root@www-proxylb01 ~]# rpm -qi pacemaker
> Name: pacemakerRelocations: (not relocatable)
> Version : 1.1.8 Vendor: CentOS
> Release : 7.el6 Build Date: Fri 22 Feb 2013 
> 02:07:28 AM ART
> Install Date: Tue 04 Jun 2013 04:32:24 PM ART  Build Host: 
> c6b9.bsys.dev.centos.org
> Group   : System Environment/DaemonsSource RPM: 
> pacemaker-1.1.8-7.el6.src.rpm
> Size: 1269655  License: GPLv2+ and LGPLv2+
> Signature   : RSA/SHA1, Sat 23 Feb 2013 02:41:32 PM ART, Key ID 
> 0946fca2c105b9de
> Packager: CentOS BuildSystem 
> URL : http://www.clusterlabs.org
> Summary : Scalable High-Availability cluster resource manager
> Description :
> Pacemaker is an advanced, scalable High-Availability cluster resource
> manager for Linux-HA (Heartbeat) and/or Corosync.
> 
> It supports "n-node" clusters with significant capabilities for
> managing resources and dependencies.
> 
> It will run scripts at initialization, when machines go up or down,
> when related resources fail and can be configured to periodically check
> resource health.
> 
> Available rpmbuild rebuild options:
>   --with(out) : heartbeat cman corosync doc publican snmp esmtp pre_release
> 
> 
> [root@www-proxylb01 ~]# rpm -qi corosync
> Name: corosync Relocations: (not relocatable)
> Version : 1.4.1 Vendor: CentOS
> Release : 15.el6_4.1Build Date: Tue 14 May 2013 
> 06:09:27 PM ART
> Install Date: Tue 04 Jun 2013 04:32:24 PM ART  Build Host: 
> c6b7.bsys.dev.centos.org
> Group   : System Environment/Base   Source RPM: 
> corosync-1.4.1-15.el6_4.1.src.rpm
> Size: 438998   License: BSD
> Signature   : RSA/SHA1, Tue 14 May 2013 08:03:55 PM ART, Key ID 
> 0946fca2c105b9de
> Packager: CentOS BuildSystem 
> URL : http://ftp.corosync.org
> Summary : The Corosync Cluster Engine and Application Programming 
> Interfaces
> Description :
> This package contains the Corosync Cluster Engine Executive, several default
> APIs and libraries, default configuration files, and an init script.
> 
> 
> Thanks,
> 
> 
> Matías R. Cuenca del Rey
> 
> 
> On Thu, Aug 8, 2013 at 7:44 PM, Chris Feist  wrote:
> On 08/08/2013 01:25 PM, Matias R. Cuenca del Rey wrote:
> Hi,
> 
> This is my first mail. I'm playing with active/active cluster with 
> cman+pacemaker
> I have 3 nodes working great. When I reboot one node, my IP resource move to
> another node, but when the rebooted node comes back, my IP resource doesn't 
> move
> in again. I tried to move mannualy with pcs but I get the following error:
> 
> [root@www-proxylb01 ~]# pcs config
> Corosync Nodes:
> 
> Pacemaker Nodes:
>   www-proxylb01 www-proxylb02 www-proxylb03
> 
> Resources:
>   Clone: ip-xxx.xxx.xxx.xxx-clone
>Resource: ip-xxx.xxx.xxx.xxx (provider=heartbeat type=IPaddr2 class=ocf)
> Attributes: ip=xxx.xxx.xxx.xxx cidr_netmask=32
> clusterip_hash=sourceip-sourceport
> Operations: monitor interval=30s
>   Clone: fs-usr.share.haproxy-clone
>Resource: fs-usr.share.haproxy (provider=heartbeat type=Filesystem 
> class=ocf)
> Attributes: device=/dev/xvdc directory=/usr/share/haproxy/ fstype=gfs2
>   Clone: haproxy-xxx.xxx.xxx.xxx-clone
>Resource: haproxy-xxx.xxx.xxx.xxx (provider=heartbeat type=haproxy 
> class=ocf)
> Attributes: conffile=/etc/haproxy/haproxy.cfg
> Operations: monitor interval=30s
> 
> Location Constraints:
> Ordering Constraints:
>ip-xxx.xxx.xxx.xxx-clone then haproxy-xxx.

Re: [Pacemaker] Postgresql Replication

2013-09-10 Thread Takatoshi MATSUO
Hi Eloy

Please see http://clusterlabs.org/wiki/PgSQL_Replicated_Cluster .
In the document, it uses virtual IP to receive connection,
so it doesn't need to change recovery.conf.

Thanks,
Takatoshi MATSUO


2013/9/11 Eloy Coto Pereiro :
> Hi,
>
> In Postgresql if you use wal replication
>  when the master
> servers fails need to change the recovery.conf on the slave server.
>
> In this case any tool, when the master is down, execute a command and get
> this info?
> Is this the right tool for postgresql's replication?
>
> Cheers
>
>
>
>
> ___
> Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>
> Project Home: http://www.clusterlabs.org
> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
> Bugs: http://bugs.clusterlabs.org
>

___
Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
http://oss.clusterlabs.org/mailman/listinfo/pacemaker

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org


[Pacemaker] Postgresql Replication

2013-09-10 Thread Eloy Coto Pereiro
Hi,

In Postgresql if you use wal replication <
http://wiki.postgresql.org/wiki/Streaming_Replication> when the master
servers fails need to change the recovery.conf on the slave server.

In this case any tool, when the master is down, execute a command and get
this info?
Is this the right tool for postgresql's replication?

Cheers
___
Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
http://oss.clusterlabs.org/mailman/listinfo/pacemaker

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org


Re: [Pacemaker] Moving cloned resources

2013-09-10 Thread Matias R. Cuenca del Rey
Thank you Andrew,

If I can't move instances of a clone. How can I do to rebalance my IP
resource trough the three nodes?

Thanks,

Matías R. Cuenca del Rey


On Fri, Aug 9, 2013 at 12:15 PM, Matias R. Cuenca del Rey  wrote:

> [root@www-proxylb01 ~]# rpm -qi pcs
> Name: pcs  Relocations: (not relocatable)
> Version : 0.9.26Vendor: CentOS
> Release : 10.el6_4.1Build Date: Mon 18 Mar 2013
> 11:39:25 AM ART
> Install Date: Tue 04 Jun 2013 05:19:49 PM ART  Build Host:
> c6b10.bsys.dev.centos.org
> Group   : System Environment/Base   Source RPM:
> pcs-0.9.26-10.el6_4.1.src.rpm
> Size: 254791   License: GPLv2
> Signature   : RSA/SHA1, Mon 18 Mar 2013 12:01:56 PM ART, Key ID
> 0946fca2c105b9de
> Packager: CentOS BuildSystem 
> URL : http://github.com/feist/pcs
> Summary : Pacemaker Configuration System
> Description :
> pcs is a corosync and pacemaker configuration tool.  It permits users to
> easily view, modify and created pacemaker based clusters.
>
>
> [root@www-proxylb01 ~]# rpm -qi pacemaker
> Name: pacemakerRelocations: (not relocatable)
> Version : 1.1.8 Vendor: CentOS
> Release : 7.el6 Build Date: Fri 22 Feb 2013
> 02:07:28 AM ART
> Install Date: Tue 04 Jun 2013 04:32:24 PM ART  Build Host:
> c6b9.bsys.dev.centos.org
> Group   : System Environment/DaemonsSource RPM:
> pacemaker-1.1.8-7.el6.src.rpm
> Size: 1269655  License: GPLv2+ and LGPLv2+
> Signature   : RSA/SHA1, Sat 23 Feb 2013 02:41:32 PM ART, Key ID
> 0946fca2c105b9de
> Packager: CentOS BuildSystem 
> URL : http://www.clusterlabs.org
> Summary : Scalable High-Availability cluster resource manager
> Description :
> Pacemaker is an advanced, scalable High-Availability cluster resource
> manager for Linux-HA (Heartbeat) and/or Corosync.
>
> It supports "n-node" clusters with significant capabilities for
> managing resources and dependencies.
>
> It will run scripts at initialization, when machines go up or down,
> when related resources fail and can be configured to periodically check
> resource health.
>
> Available rpmbuild rebuild options:
>   --with(out) : heartbeat cman corosync doc publican snmp esmtp pre_release
>
>
> [root@www-proxylb01 ~]# rpm -qi corosync
> Name: corosync Relocations: (not relocatable)
> Version : 1.4.1 Vendor: CentOS
> Release : 15.el6_4.1Build Date: Tue 14 May 2013
> 06:09:27 PM ART
> Install Date: Tue 04 Jun 2013 04:32:24 PM ART  Build Host:
> c6b7.bsys.dev.centos.org
> Group   : System Environment/Base   Source RPM:
> corosync-1.4.1-15.el6_4.1.src.rpm
> Size: 438998   License: BSD
> Signature   : RSA/SHA1, Tue 14 May 2013 08:03:55 PM ART, Key ID
> 0946fca2c105b9de
> Packager: CentOS BuildSystem 
> URL : http://ftp.corosync.org
> Summary : The Corosync Cluster Engine and Application Programming
> Interfaces
> Description :
> This package contains the Corosync Cluster Engine Executive, several
> default
> APIs and libraries, default configuration files, and an init script.
>
>
> Thanks,
>
>
> Matías R. Cuenca del Rey
>
>
> On Thu, Aug 8, 2013 at 7:44 PM, Chris Feist  wrote:
>
>> On 08/08/2013 01:25 PM, Matias R. Cuenca del Rey wrote:
>>
>>> Hi,
>>>
>>> This is my first mail. I'm playing with active/active cluster with
>>> cman+pacemaker
>>> I have 3 nodes working great. When I reboot one node, my IP resource
>>> move to
>>> another node, but when the rebooted node comes back, my IP resource
>>> doesn't move
>>> in again. I tried to move mannualy with pcs but I get the following
>>> error:
>>>
>>> [root@www-proxylb01 ~]# pcs config
>>> Corosync Nodes:
>>>
>>> Pacemaker Nodes:
>>>   www-proxylb01 www-proxylb02 www-proxylb03
>>>
>>> Resources:
>>>   Clone: ip-xxx.xxx.xxx.xxx-clone
>>>Resource: ip-xxx.xxx.xxx.xxx (provider=heartbeat type=IPaddr2
>>> class=ocf)
>>> Attributes: ip=xxx.xxx.xxx.xxx cidr_netmask=32
>>> clusterip_hash=sourceip-**sourceport
>>> Operations: monitor interval=30s
>>>   Clone: fs-usr.share.haproxy-clone
>>>Resource: fs-usr.share.haproxy (provider=heartbeat type=Filesystem
>>> class=ocf)
>>> Attributes: device=/dev/xvdc directory=/usr/share/haproxy/
>>> fstype=gfs2
>>>   Clone: haproxy-xxx.xxx.xxx.xxx-clone
>>>Resource: haproxy-xxx.xxx.xxx.xxx (provider=heartbeat type=haproxy
>>> class=ocf)
>>> Attributes: conffile=/etc/haproxy/haproxy.**cfg
>>> Operations: monitor interval=30s
>>>
>>> Location Constraints:
>>> Ordering Constraints:
>>>ip-xxx.xxx.xxx.xxx-clone then haproxy-xxx.xxx.xxx.xxx-clone
>>>fs-usr.share.haproxy-clone then haproxy-xxx.xxx.xxx.xxx-clone
>>>

Re: [Pacemaker] Resource ordering/colocating question (DRBD + LVM + FS)

2013-09-10 Thread Heikki Manninen
Not sure whether I'm doing this the right way but here goes..

With resources started on node #1:

# crm_simulate -L -s -d pgdbsrv01.cl1.local

Current cluster status:
Online: [ pgdbsrv01.cl1.local pgdbsrv02.cl1.local ]

 Master/Slave Set: DRBD_ms_data01 [DRBD_data01]
 Masters: [ pgdbsrv01.cl1.local ]
 Slaves: [ pgdbsrv02.cl1.local ]
 Master/Slave Set: DRBD_ms_data02 [DRBD_data02]
 Masters: [ pgdbsrv01.cl1.local ]
 Slaves: [ pgdbsrv02.cl1.local ]
 Resource Group: GRP_data01
 LVM_vgdata01   (ocf::heartbeat:LVM):   Started pgdbsrv01.cl1.local
 FS_data01  (ocf::heartbeat:Filesystem):Started pgdbsrv01.cl1.local
 Resource Group: GRP_data02
 LVM_vgdata02   (ocf::heartbeat:LVM):   Started pgdbsrv01.cl1.local
 FS_data02  (ocf::heartbeat:Filesystem):Started pgdbsrv01.cl1.local
 fusion-fencing (stonith:fence_fusion): Started pgdbsrv02.cl1.local

Performing requested modifications
 + Taking node pgdbsrv01.cl1.local offline
Allocation scores:
clone_color: DRBD_ms_data01 allocation score on pgdbsrv01.cl1.local: 0
clone_color: DRBD_ms_data01 allocation score on pgdbsrv02.cl1.local: 0
clone_color: DRBD_data01:0 allocation score on pgdbsrv01.cl1.local: 0
clone_color: DRBD_data01:0 allocation score on pgdbsrv02.cl1.local: 1
clone_color: DRBD_data01:1 allocation score on pgdbsrv01.cl1.local: 0
clone_color: DRBD_data01:1 allocation score on pgdbsrv02.cl1.local: 1
native_color: DRBD_data01:0 allocation score on pgdbsrv01.cl1.local: -INFINITY
native_color: DRBD_data01:0 allocation score on pgdbsrv02.cl1.local: 1
native_color: DRBD_data01:1 allocation score on pgdbsrv01.cl1.local: -INFINITY
native_color: DRBD_data01:1 allocation score on pgdbsrv02.cl1.local: -INFINITY
DRBD_data01:0 promotion score on pgdbsrv02.cl1.local: 1
DRBD_data01:1 promotion score on none: 0
clone_color: DRBD_ms_data02 allocation score on pgdbsrv01.cl1.local: 0
clone_color: DRBD_ms_data02 allocation score on pgdbsrv02.cl1.local: 0
clone_color: DRBD_data02:0 allocation score on pgdbsrv01.cl1.local: 0
clone_color: DRBD_data02:0 allocation score on pgdbsrv02.cl1.local: 1
clone_color: DRBD_data02:1 allocation score on pgdbsrv01.cl1.local: 0
clone_color: DRBD_data02:1 allocation score on pgdbsrv02.cl1.local: 1
native_color: DRBD_data02:0 allocation score on pgdbsrv01.cl1.local: -INFINITY
native_color: DRBD_data02:0 allocation score on pgdbsrv02.cl1.local: 1
native_color: DRBD_data02:1 allocation score on pgdbsrv01.cl1.local: -INFINITY
native_color: DRBD_data02:1 allocation score on pgdbsrv02.cl1.local: -INFINITY
DRBD_data02:0 promotion score on pgdbsrv02.cl1.local: 1
DRBD_data02:1 promotion score on none: 0
group_color: GRP_data01 allocation score on pgdbsrv01.cl1.local: 0
group_color: GRP_data01 allocation score on pgdbsrv02.cl1.local: 0
group_color: LVM_vgdata01 allocation score on pgdbsrv01.cl1.local: 0
group_color: LVM_vgdata01 allocation score on pgdbsrv02.cl1.local: 0
group_color: FS_data01 allocation score on pgdbsrv01.cl1.local: 0
group_color: FS_data01 allocation score on pgdbsrv02.cl1.local: 0
native_color: LVM_vgdata01 allocation score on pgdbsrv01.cl1.local: -INFINITY
native_color: LVM_vgdata01 allocation score on pgdbsrv02.cl1.local: 1
native_color: FS_data01 allocation score on pgdbsrv01.cl1.local: -INFINITY
native_color: FS_data01 allocation score on pgdbsrv02.cl1.local: 0
group_color: GRP_data02 allocation score on pgdbsrv01.cl1.local: 0
group_color: GRP_data02 allocation score on pgdbsrv02.cl1.local: 0
group_color: LVM_vgdata02 allocation score on pgdbsrv01.cl1.local: 0
group_color: LVM_vgdata02 allocation score on pgdbsrv02.cl1.local: 0
group_color: FS_data02 allocation score on pgdbsrv01.cl1.local: 0
group_color: FS_data02 allocation score on pgdbsrv02.cl1.local: 0
native_color: LVM_vgdata02 allocation score on pgdbsrv01.cl1.local: -INFINITY
native_color: LVM_vgdata02 allocation score on pgdbsrv02.cl1.local: 1
native_color: FS_data02 allocation score on pgdbsrv01.cl1.local: -INFINITY
native_color: FS_data02 allocation score on pgdbsrv02.cl1.local: 0
native_color: fusion-fencing allocation score on pgdbsrv01.cl1.local: 0
native_color: fusion-fencing allocation score on pgdbsrv02.cl1.local: 0

Transition Summary:
 * Promote DRBD_data01:0(Slave -> Master pgdbsrv02.cl1.local)
 * Demote  DRBD_data01:1(Master -> Stopped pgdbsrv01.cl1.local)
 * Promote DRBD_data02:0(Slave -> Master pgdbsrv02.cl1.local)
 * Demote  DRBD_data02:1(Master -> Stopped pgdbsrv01.cl1.local)
 * MoveLVM_vgdata01 (Started pgdbsrv01.cl1.local -> pgdbsrv02.cl1.local)
 * MoveFS_data01(Started pgdbsrv01.cl1.local -> pgdbsrv02.cl1.local)
 * MoveLVM_vgdata02 (Started pgdbsrv01.cl1.local -> pgdbsrv02.cl1.local)
 * MoveFS_data02(Started pgdbsrv01.cl1.local -> pgdbsrv02.cl1.local)


..taking node #1 offline (standby) for real, resources running on node #2, then:

# crm_simulate -L -s -u pgdbsrv01.cl1.local

Current cluster status:
Node pgdbsrv01

Re: [Pacemaker] [Linux-HA] Probably a regression of the linbit drbd agent between pacemaker 1.1.8 and 1.1.10

2013-09-10 Thread Lars Ellenberg
On Mon, Sep 09, 2013 at 01:41:17PM +0200, Andreas Mock wrote:
> Hi Lars,
> 
> here also my official "Thank you very much" looking
> at the problem.

> I've been looking forward to the official release
> of drbd 8.4.4.
> 
> Or do you need disoriented rc testers like me? ;-)

Why not?
That's what release candidates are intended for.
You'd only have to confirm that it works for you now.

Respectively, that it still does not,
in which case you better report that now
than after the release, right?


-- 
: Lars Ellenberg
: LINBIT | Your Way to High Availability
: DRBD/HA support and consulting http://www.linbit.com

DRBD® and LINBIT® are registered trademarks of LINBIT, Austria.

___
Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
http://oss.clusterlabs.org/mailman/listinfo/pacemaker

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org