Hi, On Thu, Feb 18, 2010 at 03:14:05PM +0200, Marian Marinov wrote: > I had almost identical problem as you. > > I'm currently working on a solution for this problem. I hope next week I'll > have that part finished and I'll file the enhancement bugzilla for that > feature.
You mean you're preparing a patch? > For now, what you can do is remove the target-role meta attribute from the > resources you want to promote using: > crm> resource meta RESOUCE_PRIMITIVE delete target-role Oh, completely forgot that it can be done this way too. Thanks for mentioning it. Dejan > You have to remove the target role only from the local primitive not from the > clone. > > This is how we deal with the problem. > > Regards, > Marian > > On Thursday 18 February 2010 14:56:03 Dejan Muhamedagic wrote: > > Hi, > > > > On Tue, Feb 16, 2010 at 08:01:10PM -0800, Bob Schatz wrote: > > > Hi, > > > > > > I have configured 8 master/slave resources on two virtual machines named > > > fc12-64-1 and fc12-64-2. > > > > > > They are running fedora core 12 64 bit with the version of pacemaker from > > > a "yum install pacemaker" or: Name : heartbeat > > > Version : 3.0.0 > > > > > > Name : pacemaker > > > Version : 1.0.5 > > > > > > Name : pacemaker-libs > > > Version : 1.0.5 > > > > > > Name : cluster-glue > > > Version : 1.0 > > > > > > Name : cluster-glue-libs > > > Version : 1.0 > > > > > > My requirements are: > > > > > > 1.Four resources start as master on each node and the slave for each > > > resource starts on the other node - i.e. > > > > > > SS0 (master) on fc12-64-1 > > > SS0 (slave) on fc12-64-2 > > > 2.After a failover (power off one node) I have all 8 resources running as > > > master on one node. > > > > > > 3.I do not want automatic failback when a node comes back. I only want > > > it to occur under operator control. > > > > > > 4.If the process associated with each resource dies, it will be restarted > > > as a slave and the other node will convert the process to master. > > > > > > My problem: > > > > > > I am able to start both nodes and have four masters on each node with a > > > slave on the second node. Also, failover works as expected - all 8 > > > resources are master on the remaining node if one node dies and if I have > > > process death I only failover the one resource. > > > > > > However, I am not sure how I can cause a fail back to occur from the > > > command line (requirement #3) > > > > > > I start with this to figure out how to do a failback: > > > > > > # crm_mon -n > > > > > > ============ > > > Last updated: Tue Feb 16 19:44:12 2010 > > > Stack: Heartbeat > > > Current DC: fc12-64-1 (d7b30d08-d835-4014-b9c6-ebf53099cbe3) - partition > > > with quorum Version: 1.0.5-ee19d8e83c2a5d45988f1cee36d334a631d84fc7 > > > 2 Nodes configured, unknown expected votes > > > 8 Resources configured. > > > ============ > > > > > > Node fc12-64-1 (d7b30d08-d835-4014-b9c6-ebf53099cbe3): online > > > SS6:0 (ocf::omneon:ss) Master > > > SS3:0 (ocf::omneon:ss) Master > > > SS7:0 (ocf::omneon:ss) Master > > > SS0:0 (ocf::omneon:ss) Master > > > SS4:0 (ocf::omneon:ss) Master > > > SS1:0 (ocf::omneon:ss) Master > > > SS5:0 (ocf::omneon:ss) Master > > > SS2:0 (ocf::omneon:ss) Master > > > Node fc12-64-2 (b69df3a6-a630-4edb-adf4-28727f8c1222): online > > > SS0:1 (ocf::omneon:ss) Slave > > > SS2:1 (ocf::omneon:ss) Slave > > > SS1:1 (ocf::omneon:ss) Slave > > > SS3:1 (ocf::omneon:ss) Slave > > > SS5:1 (ocf::omneon:ss) Slave > > > SS4:1 (ocf::omneon:ss) Slave > > > SS7:1 (ocf::omneon:ss) Slave > > > SS6:1 (ocf::omneon:ss) Slave > > > > > > And tried these steps to do a failback: > > > > > > # crm resource migrate ms-SS0 fc12-64-2 > > > Error performing operation: ms-SS0 is already active on fc12-64-2 > > > > > > ====> which makes sense since it is a "slave" on fc12-64-2 > > > > > > # crm resource > > > crm(live)resource# promote ms-SS0:1 > > > ERROR: ms-SS0:1 is not a master-slave resource crm(live)resource# promote > > > SS0:1 ERROR: SS0:1 is not a master-slave resource crm(live)resource# > > > promote SS0 ERROR: SS0 is not a master-slave resource crm(live)resource# > > > promote ms-SS0 Multiple attributes match name=target-role Value: Started > > > (id=ms-SS0-meta_attributes-target-role) Value: Started > > > (id=SS0-meta_attributes-target-role) Error performing operation: Required > > > data for this CIB API call not found crm(live)resource# demote ms-SS0 > > > Multiple attributes match name=target-role Value: Started > > > (id=ms-SS0-meta_attributes-target-role) Value: Started > > > (id=SS0-meta_attributes-target-role) Error performing operation: Required > > > data for this CIB API call not found crm(live)resource# > > > > > > Since I have a location constraint for ms-SS0 to fc12-64-1, I tried the > > > same operation using ms-SS1 which has a location constraint for > > > fc12-64-2. > > > > > > This gave me the same messages or: > > > > > > crm(live)resource# promote ms-SS1 > > > Multiple attributes match name=target-role > > > Value: Started (id=ms-SS1-meta_attributes-target-role) > > > Value: Started (id=SS1-meta_attributes-target-role) > > > Error performing operation: Required data for this CIB API call not found > > > crm(live)resource# demote ms-SS1 > > > Multiple attributes match name=target-role > > > Value: Started (id=ms-SS1-meta_attributes-target-role) > > > Value: Started (id=SS1-meta_attributes-target-role) > > > Error performing operation: Required data for this CIB API call not found > > > > > > Could you tell me what I am doing wrong? > > > > There are multiple target-roles, i.e. for both the clone and the > > resource which is cloned. The tools can't decide which one to > > change. The crm shell will try to deal with that with the next > > Pacemaker release (I think that there's already an enhancement > > bugzilla open). In the meantime, you can edit your configuration > > and remove the meta attributes from primitives. > > > > Thanks, > > > > Dejan > > > > > My configuration file is attached below. > > > > > > > > > Thanks, > > > > > > Bob > > > > > > --------------------------- my configuration > > > ----------------------------------------------------------- node > > > $id="b69df3a6-a630-4edb-adf4-28727f8c1222" fc12-64-2 > > > node $id="d7b30d08-d835-4014-b9c6-ebf53099cbe3" fc12-64-1 > > > primitive SS0 ocf:omneon:ss \ > > > params ss_resource="SS0" \ > > > params ssconf="/tmp/config.0" \ > > > op monitor interval="59s" role="Master" timeout="30s" \ > > > op monitor interval="60s" role="Slave" timeout="28" \ > > > meta target-role="Started" > > > primitive SS1 ocf:omneon:ss \ > > > params ss_resource="SS1" \ > > > params ssconf="/tmp/config.1" \ > > > op monitor interval="59s" role="Master" timeout="30s" \ > > > op monitor interval="60s" role="Slave" timeout="28" \ > > > meta target-role="Started" > > > primitive SS2 ocf:omneon:ss \ > > > params ss_resource="SS2" \ > > > params ssconf="/tmp/config.2" \ > > > op monitor interval="59s" role="Master" timeout="30s" \ > > > op monitor interval="60s" role="Slave" timeout="28" \ > > > meta target-role="Started" > > > primitive SS3 ocf:omneon:ss \ > > > params ss_resource="SS3" \ > > > params ssconf="/tmp/config.3" \ > > > op monitor interval="59s" role="Master" timeout="30s" \ > > > op monitor interval="60s" role="Slave" timeout="28" \ > > > meta target-role="Started" > > > primitive SS4 ocf:omneon:ss \ > > > params ss_resource="SS4" \ > > > params ssconf="/tmp/config.4" \ > > > op monitor interval="59s" role="Master" timeout="30s" \ > > > op monitor interval="60s" role="Slave" timeout="28" \ > > > meta target-role="Started" > > > primitive SS5 ocf:omneon:ss \ > > > params ss_resource="SS5" \ > > > params ssconf="/tmp/config.5" \ > > > op monitor interval="59s" role="Master" timeout="30s" \ > > > op monitor interval="60s" role="Slave" timeout="28" \ > > > meta target-role="Started" > > > primitive SS6 ocf:omneon:ss \ > > > params ss_resource="SS6" \ > > > params ssconf="/tmp/config.6" \ > > > op monitor interval="59s" role="Master" timeout="30s" \ > > > op monitor interval="60s" role="Slave" timeout="28" \ > > > meta target-role="Started" > > > primitive SS7 ocf:omneon:ss \ > > > params ss_resource="SS7" \ > > > params ssconf="/tmp/config.7" \ > > > op monitor interval="59s" role="Master" timeout="30s" \ > > > op monitor interval="60s" role="Slave" timeout="28" \ > > > meta target-role="Started" > > > ms ms-SS0 SS0 \ > > > meta clone-max="2" notify="true" globaally-unique="false" > > > target-role="Started" ms ms-SS1 SS1 \ > > > meta clone-max="2" notify="true" globaally-unique="false" > > > target-role="Started" ms ms-SS2 SS2 \ > > > meta clone-max="2" notify="true" globaally-unique="false" > > > target-role="Started" ms ms-SS3 SS3 \ > > > meta clone-max="2" notify="true" globaally-unique="false" > > > target-role="Started" ms ms-SS4 SS4 \ > > > meta clone-max="2" notify="true" globaally-unique="false" > > > target-role="Started" ms ms-SS5 SS5 \ > > > meta clone-max="2" notify="true" globaally-unique="false" > > > target-role="Started" ms ms-SS6 SS6 \ > > > meta clone-max="2" notify="true" globaally-unique="false" > > > target-role="Started" ms ms-SS7 SS7 \ > > > meta clone-max="2" notify="true" globaally-unique="false" > > > target-role="Started" location ms-SS0-master-w1 ms-SS0 \ > > > rule $id="ms-SS0-master-w1-rule" $role="master" 100: #uname eq > > > fc12-64-1 location ms-SS1-master-w1 ms-SS1 \ > > > rule $id="ms-SS1-master-w1-rule" $role="master" 100: #uname eq > > > fc12-64-2 location ms-SS2-master-w1 ms-SS2 \ > > > rule $id="ms-SS2-master-w1-rule" $role="master" 100: #uname eq > > > fc12-64-1 location ms-SS3-master-w1 ms-SS3 \ > > > rule $id="ms-SS3-master-w1-rule" $role="master" 100: #uname eq > > > fc12-64-2 location ms-SS4-master-w1 ms-SS4 \ > > > rule $id="ms-SS4-master-w1-rule" $role="master" 100: #uname eq > > > fc12-64-1 location ms-SS5-master-w1 ms-SS5 \ > > > rule $id="ms-SS5-master-w1-rule" $role="master" 100: #uname eq > > > fc12-64-2 location ms-SS6-master-w1 ms-SS6 \ > > > rule $id="ms-SS6-master-w1-rule" $role="master" 100: #uname eq > > > fc12-64-1 location ms-SS7-master-w1 ms-SS7 \ > > > rule $id="ms-SS7-master-w1-rule" $role="master" 100: #uname eq > > > fc12-64-2 property $id="cib-bootstrap-options" \ > > > dc-version="1.0.5-ee19d8e83c2a5d45988f1cee36d334a631d84fc7" \ > > > cluster-infrastructure="Heartbeat" \ > > > stonith-enabled="false" \ > > > symmetric-cluster="true" > > > > > > > > -- > Best regards, > Marian Marinov > _______________________________________________ > Linux-HA mailing list > [email protected] > http://lists.linux-ha.org/mailman/listinfo/linux-ha > See also: http://linux-ha.org/ReportingProblems _______________________________________________ Linux-HA mailing list [email protected] http://lists.linux-ha.org/mailman/listinfo/linux-ha See also: http://linux-ha.org/ReportingProblems
