Hi,

I have configured 8 master/slave resources on two virtual machines named 
fc12-64-1 and fc12-64-2.

They are running fedora core 12 64 bit with the version of pacemaker from a 
"yum install pacemaker" or:
   Name       : heartbeat
   Version    : 3.0.0

   Name       : pacemaker
   Version    : 1.0.5

  Name       : pacemaker-libs
  Version    : 1.0.5

  Name       : cluster-glue
  Version    : 1.0

  Name       : cluster-glue-libs
  Version    : 1.0

My requirements are:

1.Four resources start as master on each node and the slave for each resource 
starts
on the other node - i.e.

SS0 (master) on fc12-64-1
SS0 (slave) on fc12-64-2
2.After a failover (power off one node) I have all 8 resources running as 
master on one node.

3.I do not want automatic failback when a node comes back.  I only want it to 
occur under operator control.

4.If the process associated with each resource dies, it will be restarted as a 
slave and the other node
will convert the process to master.

My problem:

I am able to start both nodes and have four masters on each node with a slave 
on the second node.  Also, failover
works as expected - all 8 resources are master on the remaining node if one 
node dies and if I have process death I only
failover the one resource.

However, I am not sure how I can cause a fail back to occur from the command 
line (requirement #3)

I start with this to figure out how to do a failback:

# crm_mon -n

============
Last updated: Tue Feb 16 19:44:12 2010
Stack: Heartbeat
Current DC: fc12-64-1 (d7b30d08-d835-4014-b9c6-ebf53099cbe3) - partition with 
quorum
Version: 1.0.5-ee19d8e83c2a5d45988f1cee36d334a631d84fc7
2 Nodes configured, unknown expected votes
8 Resources configured.
============

Node fc12-64-1 (d7b30d08-d835-4014-b9c6-ebf53099cbe3): online
        SS6:0   (ocf::omneon:ss) Master
       SS3:0   (ocf::omneon:ss) Master
        SS7:0   (ocf::omneon:ss) Master
        SS0:0   (ocf::omneon:ss) Master
        SS4:0   (ocf::omneon:ss) Master
        SS1:0   (ocf::omneon:ss) Master
        SS5:0   (ocf::omneon:ss) Master
        SS2:0   (ocf::omneon:ss) Master
Node fc12-64-2 (b69df3a6-a630-4edb-adf4-28727f8c1222): online
        SS0:1   (ocf::omneon:ss) Slave
        SS2:1   (ocf::omneon:ss) Slave
        SS1:1   (ocf::omneon:ss) Slave
        SS3:1   (ocf::omneon:ss) Slave
        SS5:1   (ocf::omneon:ss) Slave
        SS4:1   (ocf::omneon:ss) Slave
        SS7:1   (ocf::omneon:ss) Slave
        SS6:1   (ocf::omneon:ss) Slave

And tried these steps to do a failback:

# crm resource migrate ms-SS0 fc12-64-2
Error performing operation: ms-SS0 is already active on fc12-64-2

====> which makes sense since it is a "slave" on fc12-64-2

# crm resource
crm(live)resource# promote ms-SS0:1
ERROR: ms-SS0:1 is not a master-slave resource crm(live)resource# promote SS0:1 
ERROR: SS0:1 is not a master-slave resource crm(live)resource# promote SS0 
ERROR: SS0 is not a master-slave resource crm(live)resource# promote ms-SS0 
Multiple attributes match name=target-role Value: Started        
(id=ms-SS0-meta_attributes-target-role) Value: Started        
(id=SS0-meta_attributes-target-role) Error performing operation: Required data 
for this CIB API call not found crm(live)resource# demote ms-SS0 Multiple 
attributes match name=target-role Value: Started        
(id=ms-SS0-meta_attributes-target-role) Value: Started        
(id=SS0-meta_attributes-target-role) Error performing operation: Required data 
for this CIB API call not found crm(live)resource#

Since I have a location constraint for ms-SS0 to fc12-64-1, I tried the same 
operation using ms-SS1 which has a location constraint for fc12-64-2.

This gave me the same messages or:

crm(live)resource# promote ms-SS1
Multiple attributes match name=target-role
  Value: Started        (id=ms-SS1-meta_attributes-target-role)
   Value: Started        (id=SS1-meta_attributes-target-role)
Error performing operation: Required data for this CIB API call not found
crm(live)resource# demote ms-SS1
Multiple attributes match name=target-role
  Value: Started        (id=ms-SS1-meta_attributes-target-role)
  Value: Started        (id=SS1-meta_attributes-target-role)
Error performing operation: Required data for this CIB API call not found

Could you tell me what I am doing wrong?

My configuration file is attached below.


Thanks,

Bob

--------------------------- my configuration 
-----------------------------------------------------------
node $id="b69df3a6-a630-4edb-adf4-28727f8c1222" fc12-64-2
node $id="d7b30d08-d835-4014-b9c6-ebf53099cbe3" fc12-64-1
primitive SS0 ocf:omneon:ss \
        params ss_resource="SS0" \
        params ssconf="/tmp/config.0" \
        op monitor interval="59s" role="Master" timeout="30s" \
        op monitor interval="60s" role="Slave" timeout="28" \
        meta target-role="Started"
primitive SS1 ocf:omneon:ss \
        params ss_resource="SS1" \
        params ssconf="/tmp/config.1" \
        op monitor interval="59s" role="Master" timeout="30s" \
        op monitor interval="60s" role="Slave" timeout="28" \
        meta target-role="Started"
primitive SS2 ocf:omneon:ss \
        params ss_resource="SS2" \
        params ssconf="/tmp/config.2" \
        op monitor interval="59s" role="Master" timeout="30s" \
        op monitor interval="60s" role="Slave" timeout="28" \
        meta target-role="Started"
primitive SS3 ocf:omneon:ss \
        params ss_resource="SS3" \
        params ssconf="/tmp/config.3" \
        op monitor interval="59s" role="Master" timeout="30s" \
        op monitor interval="60s" role="Slave" timeout="28" \
        meta target-role="Started"
primitive SS4 ocf:omneon:ss \
        params ss_resource="SS4" \
        params ssconf="/tmp/config.4" \
        op monitor interval="59s" role="Master" timeout="30s" \
        op monitor interval="60s" role="Slave" timeout="28" \
        meta target-role="Started"
primitive SS5 ocf:omneon:ss \
        params ss_resource="SS5" \
        params ssconf="/tmp/config.5" \
        op monitor interval="59s" role="Master" timeout="30s" \
        op monitor interval="60s" role="Slave" timeout="28" \
        meta target-role="Started"
primitive SS6 ocf:omneon:ss \
        params ss_resource="SS6" \
        params ssconf="/tmp/config.6" \
        op monitor interval="59s" role="Master" timeout="30s" \
        op monitor interval="60s" role="Slave" timeout="28" \
        meta target-role="Started"
primitive SS7 ocf:omneon:ss \
        params ss_resource="SS7" \
        params ssconf="/tmp/config.7" \
        op monitor interval="59s" role="Master" timeout="30s" \
        op monitor interval="60s" role="Slave" timeout="28" \
        meta target-role="Started"
ms ms-SS0 SS0 \
        meta clone-max="2" notify="true" globaally-unique="false" 
target-role="Started"
ms ms-SS1 SS1 \
        meta clone-max="2" notify="true" globaally-unique="false" 
target-role="Started"
ms ms-SS2 SS2 \
        meta clone-max="2" notify="true" globaally-unique="false" 
target-role="Started"
ms ms-SS3 SS3 \
        meta clone-max="2" notify="true" globaally-unique="false" 
target-role="Started"
ms ms-SS4 SS4 \
        meta clone-max="2" notify="true" globaally-unique="false" 
target-role="Started"
ms ms-SS5 SS5 \
        meta clone-max="2" notify="true" globaally-unique="false" 
target-role="Started"
ms ms-SS6 SS6 \
        meta clone-max="2" notify="true" globaally-unique="false" 
target-role="Started"
ms ms-SS7 SS7 \
        meta clone-max="2" notify="true" globaally-unique="false" 
target-role="Started"
location ms-SS0-master-w1 ms-SS0 \
        rule $id="ms-SS0-master-w1-rule" $role="master" 100: #uname eq fc12-64-1
location ms-SS1-master-w1 ms-SS1 \
        rule $id="ms-SS1-master-w1-rule" $role="master" 100: #uname eq fc12-64-2
location ms-SS2-master-w1 ms-SS2 \
        rule $id="ms-SS2-master-w1-rule" $role="master" 100: #uname eq fc12-64-1
location ms-SS3-master-w1 ms-SS3 \
        rule $id="ms-SS3-master-w1-rule" $role="master" 100: #uname eq fc12-64-2
location ms-SS4-master-w1 ms-SS4 \
        rule $id="ms-SS4-master-w1-rule" $role="master" 100: #uname eq fc12-64-1
location ms-SS5-master-w1 ms-SS5 \
        rule $id="ms-SS5-master-w1-rule" $role="master" 100: #uname eq fc12-64-2
location ms-SS6-master-w1 ms-SS6 \
        rule $id="ms-SS6-master-w1-rule" $role="master" 100: #uname eq fc12-64-1
location ms-SS7-master-w1 ms-SS7 \
        rule $id="ms-SS7-master-w1-rule" $role="master" 100: #uname eq fc12-64-2
property $id="cib-bootstrap-options" \
        dc-version="1.0.5-ee19d8e83c2a5d45988f1cee36d334a631d84fc7" \
        cluster-infrastructure="Heartbeat" \
        stonith-enabled="false" \
        symmetric-cluster="true"


      
_______________________________________________
Linux-HA mailing list
[email protected]
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems

Reply via email to