No reason. Need to know which combination is expected to work.
using 0.9.26 with new 1.1.10 pacemaker
and 0.9.90 with old pacemaker has issues in resource deletion.

Below is the commands executed on 0.9.26 with 1.1.10 pacemaker

[root@vsanqa11 tmp]# rpm -qa | grep pcs; rpm -qa | grep pacemaker; rpm -qa
| grep corosync; rpm -qa | grep libqb
pcs-0.9.26-10.el6.noarch
pacemaker-cli-1.1.10-14.el6_5.3.x86_64
pacemaker-libs-1.1.10-14.el6_5.3.x86_64
pacemaker-1.1.10-14.el6_5.3.x86_64
pacemaker-cluster-libs-1.1.10-14.el6_5.3.x86_64
corosynclib-1.4.1-17.el6_5.1.x86_64
corosync-1.4.1-17.el6_5.1.x86_64
libqb-devel-0.16.0-2.el6.x86_64
libqb-0.16.0-2.el6.x86_64


[root@vsanqa11 tmp]# pcs status
Last updated: Mon May 26 22:39:42 2014
Last change: Mon May 26 22:39:01 2014 via cibadmin on vsanqa11
Stack: cman
Current DC: vsanqa11 - partition with quorum
Version: 1.1.10-14.el6_5.3-368c726
2 Nodes configured
2 Resources configured


Online: [ vsanqa11 vsanqa12 ]

Full list of resources:

 Master/Slave Set: ms-9fc36888-cf2a-417a-907c-db3f5e9b7a8a
[vha-9fc36888-cf2a-417a-907c-db3f5e9b7a8a]
     Masters: [ vsanqa12 ]
     Slaves: [ vsanqa11 ]



[root@vsanqa11 tmp]# pcs config
Corosync Nodes:

Pacemaker Nodes:
 vsanqa11 vsanqa12

Resources:

Location Constraints:
  Resource: ms-9fc36888-cf2a-417a-907c-db3f5e9b7a8a
    Enabled on: vsanqa11
    Enabled on: vsanqa12
  Resource: vha-9fc36888-cf2a-417a-907c-db3f5e9b7a8a
    Enabled on: vsanqa11
    Enabled on: vsanqa12
Ordering Constraints:
Colocation Constraints:

Cluster Properties:
 dc-version: 1.1.10-14.el6_5.3-368c726
 cluster-infrastructure: cman
 last-lrm-refresh: 1401098102
 expected-quorum-votes: 2
 stonith-enabled: false
 no-quorum-policy: ignore



[root@vsanqa11 tmp]# pcs resource show
ms-9fc36888-cf2a-417a-907c-db3f5e9b7a8a
Resource: ms-9fc36888-cf2a-417a-907c-db3f5e9b7a8a
  cluster_uuid: 9fc36888-cf2a-417a-907c-db3f5e9b7a8a
  clone-max: 2
  globally-unique: false
  target-role: started
  op monitor interval=31s role=Slave timeout=100s
[root@vsanqa11 tmp]# pcs resource show
vha-9fc36888-cf2a-417a-907c-db3f5e9b7a8a
Resource: vha-9fc36888-cf2a-417a-907c-db3f5e9b7a8a
  cluster_uuid: 9fc36888-cf2a-417a-907c-db3f5e9b7a8a
  op monitor interval=31s role=Slave timeout=100s


[root@vsanqa11 tmp]# pcs resource delete
ms-9fc36888-cf2a-417a-907c-db3f5e9b7a8a
ERROR: Unable to update cib
Call cib_replace failed (-203): Update does not conform to the configured
schema
<cib admin_epoch="0" cib-last-written="Mon May 26 22:40:45 2014"
crm_feature_set="3.0.7" dc-uuid="vsanqa11" epoch="10550" have-quorum="1"
num_updates="1" update-client="cibadmin" update-origin="vsanqa11"
validate-with="pacemaker-1.2">
  <configuration>
    <crm_config>
      <cluster_property_set id="cib-bootstrap-options">
        <nvpair id="cib-bootstrap-options-dc-version" name="dc-version"
value="1.1.10-14.el6_5.3-368c726"/>
        <nvpair id="cib-bootstrap-options-cluster-infrastructure"
name="cluster-infrastructure" value="cman"/>
        <nvpair id="cib-bootstrap-options-last-lrm-refresh"
name="last-lrm-refresh" value="1401098102"/>
        <nvpair id="cib-bootstrap-options-expected-quorum-votes"
name="expected-quorum-votes" value="2"/>
        <nvpair id="cib-bootstrap-options-stonith-enabled"
name="stonith-enabled" value="false"/>
        <nvpair id="cib-bootstrap-options-no-quorum-policy"
name="no-quorum-policy" value="ignore"/>
      </cluster_property_set>
    </crm_config>
    <nodes>
      <node id="vsanqa11" uname="vsanqa11"/>
      <node id="vsanqa12" uname="vsanqa12"/>
    </nodes>
    <resources/>
    <constraints>
      <rsc_location
id="location-vha-9fc36888-cf2a-417a-907c-db3f5e9b7a8a-vsanqa11-INFINITY"
node="vsanqa11" rsc="vha-9fc36888-cf2a-417a-907c-db3f5e9b7a8a"
score="INFINITY"/>
      <rsc_location
id="location-vha-9fc36888-cf2a-417a-907c-db3f5e9b7a8a-vsanqa12-INFINITY"
node="vsanqa12" rsc="vha-9fc36888-cf2a-417a-907c-db3f5e9b7a8a"
score="INFINITY"/>
      <rsc_location
id="location-ms-9fc36888-cf2a-417a-907c-db3f5e9b7a8a-vsanqa11-INFINITY"
node="vsanqa11" rsc="ms-9fc36888-cf2a-417a-907c-db3f5e9b7a8a"
score="INFINITY"/>
      <rsc_location
id="location-ms-9fc36888-cf2a-417a-907c-db3f5e9b7a8a-vsanqa12-INFINITY"
node="vsanqa12" rsc="ms-9fc36888-cf2a-417a-907c-db3f5e9b7a8a"
score="INFINITY"/>
    </constraints>
    <rsc_defaults>
      <meta_attributes id="rsc_defaults-options">
        <nvpair id="rsc_defaults-options-resource-stickiness"
name="resource-stickiness" value="100"/>
        <nvpair id="rsc_defaults-options-timeout" name="timeout"
value="100s"/>
      </meta_attributes>
    </rsc_defaults>
  </configuration>
  <status>
    <node_state crm-debug-origin="do_update_resource" crmd="online"
expected="member" id="vsanqa11" in_ccm="true" join="member"
uname="vsanqa11">
      <transient_attributes id="vsanqa11">
        <instance_attributes id="status-vsanqa11">
          <nvpair id="status-vsanqa11-probe_complete" name="probe_complete"
value="true"/>
          <nvpair
id="status-vsanqa11-master-vha-9fc36888-cf2a-417a-907c-db3f5e9b7a8a"
name="master-vha-9fc36888-cf2a-417a-907c-db3f5e9b7a8a" value="4"/>
        </instance_attributes>
      </transient_attributes>
      <lrm id="vsanqa11">
        <lrm_resources>
          <lrm_resource id="vha-9fc36888-cf2a-417a-907c-db3f5e9b7a8a"
type="vgc-cm-agent.ocf" class="ocf" provider="heartbeat">
            <lrm_rsc_op
id="vha-9fc36888-cf2a-417a-907c-db3f5e9b7a8a_last_0"
operation_key="vha-9fc36888-cf2a-417a-907c-db3f5e9b7a8a_start_0"
operation="start" crm-debug-origin="do_update_resource"
crm_feature_set="3.0.7"
transition-key="7:195:0:79ecdaeb-e637-4fdf-b8e8-ebfc7e2eca39"
transition-magic="0:0;7:195:0:79ecdaeb-e637-4fdf-b8e8-ebfc7e2eca39"
call-id="290" rc-code="0" op-status="0" interval="0" last-run="1401169141"
last-rc-change="1401169141" exec-time="1103" queue-time="0"
op-digest="494c7d757cceae4f35487404ebc12a10"/>
            <lrm_rsc_op
id="vha-9fc36888-cf2a-417a-907c-db3f5e9b7a8a_monitor_31000"
operation_key="vha-9fc36888-cf2a-417a-907c-db3f5e9b7a8a_monitor_31000"
operation="monitor" crm-debug-origin="do_update_resource"
crm_feature_set="3.0.7"
transition-key="8:195:0:79ecdaeb-e637-4fdf-b8e8-ebfc7e2eca39"
transition-magic="0:0;8:195:0:79ecdaeb-e637-4fdf-b8e8-ebfc7e2eca39"
call-id="293" rc-code="0" op-status="0" interval="31000"
last-rc-change="1401169142" exec-time="76" queue-time="0"
op-digest="f81bd1d7870f6bb69f88312740132d65"/>
          </lrm_resource>
        </lrm_resources>
      </lrm>
    </node_state>
    <node_state crm-debug-origin="do_update_resource" crmd="online"
expected="member" id="vsanqa12" in_ccm="true" join="member"
uname="vsanqa12">
      <transient_attributes id="vsanqa12">
        <instance_attributes id="status-vsanqa12">
          <nvpair id="status-vsanqa12-probe_complete" name="probe_complete"
value="true"/>
          <nvpair
id="status-vsanqa12-master-vha-9fc36888-cf2a-417a-907c-db3f5e9b7a8a"
name="master-vha-9fc36888-cf2a-417a-907c-db3f5e9b7a8a" value="5"/>
        </instance_attributes>
      </transient_attributes>
      <lrm id="vsanqa12">
        <lrm_resources>
          <lrm_resource id="vha-9fc36888-cf2a-417a-907c-db3f5e9b7a8a"
type="vgc-cm-agent.ocf" class="ocf" provider="heartbeat">
            <lrm_rsc_op
id="vha-9fc36888-cf2a-417a-907c-db3f5e9b7a8a_last_0"
operation_key="vha-9fc36888-cf2a-417a-907c-db3f5e9b7a8a_promote_0"
operation="promote" crm-debug-origin="do_update_resource"
crm_feature_set="3.0.7"
transition-key="11:196:0:79ecdaeb-e637-4fdf-b8e8-ebfc7e2eca39"
transition-magic="0:0;11:196:0:79ecdaeb-e637-4fdf-b8e8-ebfc7e2eca39"
call-id="275" rc-code="0" op-status="0" interval="0" last-run="1401169174"
last-rc-change="1401169174" exec-time="196" queue-time="0"
op-digest="494c7d757cceae4f35487404ebc12a10"/>
          </lrm_resource>
        </lrm_resources>
      </lrm>
    </node_state>
  </status>
</cib>


****deleted on base (non clone) resource works however it stays in orphaned
state for few seconds

[root@vsanqa11 tmp]# pcs resource delete
vha-9fc36888-cf2a-417a-907c-db3f5e9b7a8a; pcs status
Removing Constraint -
location-ms-9fc36888-cf2a-417a-907c-db3f5e9b7a8a-vsanqa11-INFINITY
Removing Constraint -
location-ms-9fc36888-cf2a-417a-907c-db3f5e9b7a8a-vsanqa12-INFINITY
Removing Constraint -
location-vha-9fc36888-cf2a-417a-907c-db3f5e9b7a8a-vsanqa11-INFINITY
Removing Constraint -
location-vha-9fc36888-cf2a-417a-907c-db3f5e9b7a8a-vsanqa12-INFINITY
Deleting Resource - vha-9fc36888-cf2a-417a-907c-db3f5e9b7a8a
Last updated: Mon May 26 22:41:08 2014
Last change: Mon May 26 22:41:07 2014 via cibadmin on vsanqa11
Stack: cman
Current DC: vsanqa11 - partition with quorum
Version: 1.1.10-14.el6_5.3-368c726
2 Nodes configured
0 Resources configured


Online: [ vsanqa11 vsanqa12 ]

Full list of resources:

 vha-9fc36888-cf2a-417a-907c-db3f5e9b7a8a
(ocf::heartbeat:vgc-cm-agent.ocf):       ORPHANED Master [ vsanqa11
vsanqa12 ]

[root@vsanqa11 tmp]# pcs status
Last updated: Mon May 26 22:41:18 2014
Last change: Mon May 26 22:41:07 2014 via cibadmin on vsanqa11
Stack: cman
Current DC: vsanqa11 - partition with quorum
Version: 1.1.10-14.el6_5.3-368c726
2 Nodes configured
0 Resources configured


Online: [ vsanqa11 vsanqa12 ]

Full list of resources:



On Tue, May 27, 2014 at 11:01 AM, Andrew Beekhof <and...@beekhof.net> wrote:

>
> On 27 May 2014, at 2:34 pm, K Mehta <kiranmehta1...@gmail.com> wrote:
>
> > I have seen that 0.9.26 works with 1.1.8 pacemaker and 0.9.90 works with
> 1.1.10 pacemaker.
> > However, with 0.9.90 pcs and 1.1.8 pacemaker, pcs delete resource
> <multistate resource name> fails with error CIB update failed because of
> schema error
>
> Any specific reason to stay on 1.1.8?
>
> >
> >
> > On Tue, May 27, 2014 at 5:28 AM, Andrew Beekhof <and...@beekhof.net>
> wrote:
> >
> > On 26 May 2014, at 5:15 pm, K Mehta <kiranmehta1...@gmail.com> wrote:
> >
> > > pcs versions 0.9.26 and 0.9.90
> > > pacemaker versions 1.1.8 and 1.1.10
> > >
> > > Which pcs versions are expected to work with which pacemaker versions ?
> >
> > I think for the most part, all versions will work together.
> > There may be the odd command that requires a flag exposed by newer
> versions of pacemaker, but that should be minimal.
> >
> > >
> > >
> > > Regards,
> > >  Kiran
> > > _______________________________________________
> > > Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
> > > http://oss.clusterlabs.org/mailman/listinfo/pacemaker
> > >
> > > Project Home: http://www.clusterlabs.org
> > > Getting started:
> http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
> > > Bugs: http://bugs.clusterlabs.org
> >
> >
> > _______________________________________________
> > Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
> > http://oss.clusterlabs.org/mailman/listinfo/pacemaker
> >
> > Project Home: http://www.clusterlabs.org
> > Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
> > Bugs: http://bugs.clusterlabs.org
> >
> >
> > _______________________________________________
> > Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
> > http://oss.clusterlabs.org/mailman/listinfo/pacemaker
> >
> > Project Home: http://www.clusterlabs.org
> > Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
> > Bugs: http://bugs.clusterlabs.org
>
>
> _______________________________________________
> Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>
> Project Home: http://www.clusterlabs.org
> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
> Bugs: http://bugs.clusterlabs.org
>
>
_______________________________________________
Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
http://oss.clusterlabs.org/mailman/listinfo/pacemaker

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org

Reply via email to