Hi, On Tue, Feb 16, 2010 at 08:01:10PM -0800, Bob Schatz wrote: > Hi, > > I have configured 8 master/slave resources on two virtual machines named > fc12-64-1 and fc12-64-2. > > They are running fedora core 12 64 bit with the version of pacemaker from a > "yum install pacemaker" or: > Name : heartbeat > Version : 3.0.0 > > Name : pacemaker > Version : 1.0.5 > > Name : pacemaker-libs > Version : 1.0.5 > > Name : cluster-glue > Version : 1.0 > > Name : cluster-glue-libs > Version : 1.0 > > My requirements are: > > 1.Four resources start as master on each node and the slave for each resource > starts > on the other node - i.e. > > SS0 (master) on fc12-64-1 > SS0 (slave) on fc12-64-2 > 2.After a failover (power off one node) I have all 8 resources running as > master on one node. > > 3.I do not want automatic failback when a node comes back. I only want it to > occur under operator control. > > 4.If the process associated with each resource dies, it will be restarted as > a slave and the other node > will convert the process to master. > > My problem: > > I am able to start both nodes and have four masters on each node with a slave > on the second node. Also, failover > works as expected - all 8 resources are master on the remaining node if one > node dies and if I have process death I only > failover the one resource. > > However, I am not sure how I can cause a fail back to occur from the command > line (requirement #3) > > I start with this to figure out how to do a failback: > > # crm_mon -n > > ============ > Last updated: Tue Feb 16 19:44:12 2010 > Stack: Heartbeat > Current DC: fc12-64-1 (d7b30d08-d835-4014-b9c6-ebf53099cbe3) - partition with > quorum > Version: 1.0.5-ee19d8e83c2a5d45988f1cee36d334a631d84fc7 > 2 Nodes configured, unknown expected votes > 8 Resources configured. > ============ > > Node fc12-64-1 (d7b30d08-d835-4014-b9c6-ebf53099cbe3): online > SS6:0 (ocf::omneon:ss) Master > SS3:0 (ocf::omneon:ss) Master > SS7:0 (ocf::omneon:ss) Master > SS0:0 (ocf::omneon:ss) Master > SS4:0 (ocf::omneon:ss) Master > SS1:0 (ocf::omneon:ss) Master > SS5:0 (ocf::omneon:ss) Master > SS2:0 (ocf::omneon:ss) Master > Node fc12-64-2 (b69df3a6-a630-4edb-adf4-28727f8c1222): online > SS0:1 (ocf::omneon:ss) Slave > SS2:1 (ocf::omneon:ss) Slave > SS1:1 (ocf::omneon:ss) Slave > SS3:1 (ocf::omneon:ss) Slave > SS5:1 (ocf::omneon:ss) Slave > SS4:1 (ocf::omneon:ss) Slave > SS7:1 (ocf::omneon:ss) Slave > SS6:1 (ocf::omneon:ss) Slave > > And tried these steps to do a failback: > > # crm resource migrate ms-SS0 fc12-64-2 > Error performing operation: ms-SS0 is already active on fc12-64-2 > > ====> which makes sense since it is a "slave" on fc12-64-2 > > # crm resource > crm(live)resource# promote ms-SS0:1 > ERROR: ms-SS0:1 is not a master-slave resource crm(live)resource# promote > SS0:1 ERROR: SS0:1 is not a master-slave resource crm(live)resource# promote > SS0 ERROR: SS0 is not a master-slave resource crm(live)resource# promote > ms-SS0 Multiple attributes match name=target-role Value: Started > (id=ms-SS0-meta_attributes-target-role) Value: Started > (id=SS0-meta_attributes-target-role) Error performing operation: Required > data for this CIB API call not found crm(live)resource# demote ms-SS0 > Multiple attributes match name=target-role Value: Started > (id=ms-SS0-meta_attributes-target-role) Value: Started > (id=SS0-meta_attributes-target-role) Error performing operation: Required > data for this CIB API call not found crm(live)resource# > > Since I have a location constraint for ms-SS0 to fc12-64-1, I tried the same > operation using ms-SS1 which has a location constraint for fc12-64-2. > > This gave me the same messages or: > > crm(live)resource# promote ms-SS1 > Multiple attributes match name=target-role > Value: Started (id=ms-SS1-meta_attributes-target-role) > Value: Started (id=SS1-meta_attributes-target-role) > Error performing operation: Required data for this CIB API call not found > crm(live)resource# demote ms-SS1 > Multiple attributes match name=target-role > Value: Started (id=ms-SS1-meta_attributes-target-role) > Value: Started (id=SS1-meta_attributes-target-role) > Error performing operation: Required data for this CIB API call not found > > Could you tell me what I am doing wrong?
There are multiple target-roles, i.e. for both the clone and the resource which is cloned. The tools can't decide which one to change. The crm shell will try to deal with that with the next Pacemaker release (I think that there's already an enhancement bugzilla open). In the meantime, you can edit your configuration and remove the meta attributes from primitives. Thanks, Dejan > My configuration file is attached below. > > > Thanks, > > Bob > > --------------------------- my configuration > ----------------------------------------------------------- > node $id="b69df3a6-a630-4edb-adf4-28727f8c1222" fc12-64-2 > node $id="d7b30d08-d835-4014-b9c6-ebf53099cbe3" fc12-64-1 > primitive SS0 ocf:omneon:ss \ > params ss_resource="SS0" \ > params ssconf="/tmp/config.0" \ > op monitor interval="59s" role="Master" timeout="30s" \ > op monitor interval="60s" role="Slave" timeout="28" \ > meta target-role="Started" > primitive SS1 ocf:omneon:ss \ > params ss_resource="SS1" \ > params ssconf="/tmp/config.1" \ > op monitor interval="59s" role="Master" timeout="30s" \ > op monitor interval="60s" role="Slave" timeout="28" \ > meta target-role="Started" > primitive SS2 ocf:omneon:ss \ > params ss_resource="SS2" \ > params ssconf="/tmp/config.2" \ > op monitor interval="59s" role="Master" timeout="30s" \ > op monitor interval="60s" role="Slave" timeout="28" \ > meta target-role="Started" > primitive SS3 ocf:omneon:ss \ > params ss_resource="SS3" \ > params ssconf="/tmp/config.3" \ > op monitor interval="59s" role="Master" timeout="30s" \ > op monitor interval="60s" role="Slave" timeout="28" \ > meta target-role="Started" > primitive SS4 ocf:omneon:ss \ > params ss_resource="SS4" \ > params ssconf="/tmp/config.4" \ > op monitor interval="59s" role="Master" timeout="30s" \ > op monitor interval="60s" role="Slave" timeout="28" \ > meta target-role="Started" > primitive SS5 ocf:omneon:ss \ > params ss_resource="SS5" \ > params ssconf="/tmp/config.5" \ > op monitor interval="59s" role="Master" timeout="30s" \ > op monitor interval="60s" role="Slave" timeout="28" \ > meta target-role="Started" > primitive SS6 ocf:omneon:ss \ > params ss_resource="SS6" \ > params ssconf="/tmp/config.6" \ > op monitor interval="59s" role="Master" timeout="30s" \ > op monitor interval="60s" role="Slave" timeout="28" \ > meta target-role="Started" > primitive SS7 ocf:omneon:ss \ > params ss_resource="SS7" \ > params ssconf="/tmp/config.7" \ > op monitor interval="59s" role="Master" timeout="30s" \ > op monitor interval="60s" role="Slave" timeout="28" \ > meta target-role="Started" > ms ms-SS0 SS0 \ > meta clone-max="2" notify="true" globaally-unique="false" > target-role="Started" > ms ms-SS1 SS1 \ > meta clone-max="2" notify="true" globaally-unique="false" > target-role="Started" > ms ms-SS2 SS2 \ > meta clone-max="2" notify="true" globaally-unique="false" > target-role="Started" > ms ms-SS3 SS3 \ > meta clone-max="2" notify="true" globaally-unique="false" > target-role="Started" > ms ms-SS4 SS4 \ > meta clone-max="2" notify="true" globaally-unique="false" > target-role="Started" > ms ms-SS5 SS5 \ > meta clone-max="2" notify="true" globaally-unique="false" > target-role="Started" > ms ms-SS6 SS6 \ > meta clone-max="2" notify="true" globaally-unique="false" > target-role="Started" > ms ms-SS7 SS7 \ > meta clone-max="2" notify="true" globaally-unique="false" > target-role="Started" > location ms-SS0-master-w1 ms-SS0 \ > rule $id="ms-SS0-master-w1-rule" $role="master" 100: #uname eq > fc12-64-1 > location ms-SS1-master-w1 ms-SS1 \ > rule $id="ms-SS1-master-w1-rule" $role="master" 100: #uname eq > fc12-64-2 > location ms-SS2-master-w1 ms-SS2 \ > rule $id="ms-SS2-master-w1-rule" $role="master" 100: #uname eq > fc12-64-1 > location ms-SS3-master-w1 ms-SS3 \ > rule $id="ms-SS3-master-w1-rule" $role="master" 100: #uname eq > fc12-64-2 > location ms-SS4-master-w1 ms-SS4 \ > rule $id="ms-SS4-master-w1-rule" $role="master" 100: #uname eq > fc12-64-1 > location ms-SS5-master-w1 ms-SS5 \ > rule $id="ms-SS5-master-w1-rule" $role="master" 100: #uname eq > fc12-64-2 > location ms-SS6-master-w1 ms-SS6 \ > rule $id="ms-SS6-master-w1-rule" $role="master" 100: #uname eq > fc12-64-1 > location ms-SS7-master-w1 ms-SS7 \ > rule $id="ms-SS7-master-w1-rule" $role="master" 100: #uname eq > fc12-64-2 > property $id="cib-bootstrap-options" \ > dc-version="1.0.5-ee19d8e83c2a5d45988f1cee36d334a631d84fc7" \ > cluster-infrastructure="Heartbeat" \ > stonith-enabled="false" \ > symmetric-cluster="true" > > > > _______________________________________________ > Linux-HA mailing list > [email protected] > http://lists.linux-ha.org/mailman/listinfo/linux-ha > See also: http://linux-ha.org/ReportingProblems _______________________________________________ Linux-HA mailing list [email protected] http://lists.linux-ha.org/mailman/listinfo/linux-ha See also: http://linux-ha.org/ReportingProblems
