Hi,
I have a cluster with 2 nodes (with SUSE SLES 10 SP2 OS).
I'm using Heartbeat v2, from SUSE:
heartbeat-pils-2.1.3-0.9
heartbeat-2.1.3-0.9
sles-heartbeat_en-10.1-0.20
heartbeat-cmpi-2.1.3-0.9
yast2-heartbeat-2.13.13-0.3
heartbeat-stonith-2.1.3-0.9
I have one resource group (RG) including an Oracle DB on the first node
(maedoc1) and another RG including a tomcat6 on the second one (maedoc2):
maedoc2:~ # crm_mon -1
============
Last updated: Mon Apr 26 09:09:56 2010
Current DC: maedoc2 (39858c1a-cc4a-47a5-8e27-6d610ad614ca)
2 Nodes configured.
4 Resources configured.
============
Node: maedoc2 (39858c1a-cc4a-47a5-8e27-6d610ad614ca): online
Node: maedoc1 (a2efb703-e840-4476-a005-581446294efc): online
Resource Group: group_oracle
resource_oracle_IP (ocf::heartbeat:IPaddr2): Started maedoc1
resource_oracle_LUN (ocf::heartbeat:Filesystem): Started maedoc1
resource_oracle_DB (lsb:oracle-MAE): Started maedoc1
Resource Group: group_tomcat
resource_tomcat_IP (ocf::heartbeat:IPaddr2): Started maedoc2
resource_tomcat_LUN (ocf::heartbeat:Filesystem): Started maedoc2
resource_tomcat_TOMCAT (lsb:tomcat6): Started maedoc2
resource_stonith_maedoc1 (stonith:external/riloe): Started maedoc1
resource_stonith_maedoc2 (stonith:external/riloe): Started maedoc2
When I try to migrate the resource with, for example, this command (to
migrate the tomcat6 RG):
crm_resource -f -M -r group_tomcat -H maedoc1
These resources:
resource_tomcat_IP
resource_tomcat_LUN
migrate without problem.
On the other hand the la resource:
resource_tomcat_TOMCAT
does not start on the new node (it reamains inactive).
Only when I make (via GUI) a resource cleanup, the resource starts
corretly on the new node.
I have this error (also present into log file):
maedoc1:~ # crm_verify -L -V
crm_verify[29480]: 2010/04/23_16:36:11 ERROR: unpack_rsc_op: Hard error:
resource_tomcat_TOMCAT_monitor_0 failed with rc=5.
crm_verify[29480]: 2010/04/23_16:36:11 ERROR: unpack_rsc_op:
Preventing resource_tomcat_TOMCAT from re-starting on maedoc2
crm_verify[29970]: 2010/04/23_16:36:45 WARN: native_color: Resource
resource_tomcat_TOMCAT cannot run anywhere
Warnings found during check: config may not be valid
But the log information are not so exhaustive to known what the problem.
I found that the cleanup action does this:
Delete a resource from the LRM.
Why my resource goes up only after this operation?
I attach my cib.xml.
Thank you
Alessandra
--
Alessandra Giovanardi
e-mail a.giovana...@cineca.it
Cineca - DSET http://www.cineca.it
Phone +39 051 6171939
<cib generated="true" admin_epoch="0" have_quorum="true" ignore_dtd="false" num_peers="2" cib_feature_revision="2.0" crm_feature_set="2.0" ccm_transition="2" dc_uuid="39858c1a-cc4a-47a5-8e27-6d610ad614ca" num_updates="1" epoch="750" cib-last-written="Fri Apr 23 18:20:04 2010">
<configuration>
<crm_config>
<cluster_property_set id="cib-bootstrap-options">
<attributes>
<nvpair id="cib-bootstrap-options-dc-version" name="dc-version" value="2.1.3-node: a3184d5240c6e7032aef9cce6e5b7752ded544b3"/>
<nvpair id="cib-bootstrap-options-stonith-enabled" name="stonith-enabled" value="false"/>
<nvpair id="cib-bootstrap-options-stonith-action" name="stonith-action" value="poweroff"/>
<nvpair id="cib-bootstrap-options-default-resource-stickiness" name="default-resource-stickiness" value="101"/>
<nvpair name="last-lrm-refresh" id="cib-bootstrap-options-last-lrm-refresh" value="1272039583"/>
<nvpair id="cib-bootstrap-options-default-action-timeout" name="default-action-timeout" value="90s"/>
</attributes>
</cluster_property_set>
</crm_config>
<nodes>
<node uname="maedoc2" type="normal" id="39858c1a-cc4a-47a5-8e27-6d610ad614ca">
<instance_attributes id="nodes-39858c1a-cc4a-47a5-8e27-6d610ad614ca">
<attributes>
<nvpair name="standby" id="standby-39858c1a-cc4a-47a5-8e27-6d610ad614ca" value="off"/>
</attributes>
</instance_attributes>
</node>
<node uname="maedoc1" type="normal" id="a2efb703-e840-4476-a005-581446294efc">
<instance_attributes id="nodes-a2efb703-e840-4476-a005-581446294efc">
<attributes>
<nvpair name="standby" id="standby-a2efb703-e840-4476-a005-581446294efc" value="off"/>
</attributes>
</instance_attributes>
</node>
</nodes>
<resources>
<group id="group_oracle">
<meta_attributes id="group_oracle_meta_attrs">
<attributes>
<nvpair id="group_oracle_metaattr_ordered" name="ordered" value="true"/>
<nvpair id="group_oracle_metaattr_collocated" name="collocated" value="true"/>
<nvpair id="group_oracle_metaattr_target_role" name="target_role" value="started"/>
</attributes>
</meta_attributes>
<primitive id="resource_oracle_IP" class="ocf" type="IPaddr2" provider="heartbeat">
<instance_attributes id="resource_oracle_IP_instance_attrs">
<attributes>
<nvpair id="5a34fbf1-2c91-4ed0-8c12-15301f4843e8" name="ip" value="192.168.1.131"/>
</attributes>
</instance_attributes>
<meta_attributes id="resource_oracle_IP_meta_attrs">
<attributes>
<nvpair id="resource_oracle_IP_metaattr_target_role" name="target_role" value="started"/>
</attributes>
</meta_attributes>
</primitive>
<primitive id="resource_oracle_LUN" class="ocf" type="Filesystem" provider="heartbeat">
<instance_attributes id="resource_oracle_LUN_instance_attrs">
<attributes>
<nvpair id="73aae2cb-6097-4a5d-8f2f-7a8ebdd2e297" name="device" value="/dev/mapper/3600508b4000ad32f00008000001f0000-part1"/>
<nvpair id="c5630cac-34fb-4e0f-b2bb-c4ceaf3dfd3a" name="directory" value="/SAN/ORADATA_LUN"/>
<nvpair id="15ed23b8-c31e-433d-adc6-762bf6b8bf4d" name="fstype" value="ext3"/>
</attributes>
</instance_attributes>
<meta_attributes id="resource_oracle_LUN_meta_attrs">
<attributes>
<nvpair id="resource_oracle_LUN_metaattr_target_role" name="target_role" value="started"/>
</attributes>
</meta_attributes>
</primitive>
<primitive id="resource_oracle_DB" class="lsb" type="oracle-MAE" provider="heartbeat">
<meta_attributes id="resource_oracle_DB_meta_attrs">
<attributes>
<nvpair id="resource_oracle_DB_metaattr_target_role" name="target_role" value="started"/>
</attributes>
</meta_attributes>
</primitive>
</group>
<group id="group_tomcat">
<meta_attributes id="group_tomcat_meta_attrs">
<attributes>
<nvpair id="group_tomcat_metaattr_ordered" name="ordered" value="true"/>
<nvpair id="group_tomcat_metaattr_collocated" name="collocated" value="true"/>
<nvpair id="group_tomcat_metaattr_target_role" name="target_role" value="started"/>
</attributes>
</meta_attributes>
<primitive id="resource_tomcat_IP" class="ocf" type="IPaddr2" provider="heartbeat">
<instance_attributes id="resource_tomcat_IP_instance_attrs">
<attributes>
<nvpair id="fd4228cb-af3a-4e58-84ce-e0492cd8a06a" name="ip" value="192.168.1.132"/>
</attributes>
</instance_attributes>
<operations/>
<meta_attributes id="resource_tomcat_IP_meta_attrs">
<attributes>
<nvpair id="resource_tomcat_IP_metaattr_target_role" name="target_role" value="started"/>
</attributes>
</meta_attributes>
</primitive>
<primitive id="resource_tomcat_LUN" class="ocf" type="Filesystem" provider="heartbeat">
<instance_attributes id="resource_tomcat_LUN_instance_attrs">
<attributes>
<nvpair id="097b485a-8e75-472a-9924-161917d01477" name="device" value="/dev/mapper/3600508b4000ad32f0000800000220000"/>
<nvpair id="0f5ee72b-3b7b-455c-9676-c1892f8a6fd9" name="directory" value="/SAN/ATDOC_LUN"/>
<nvpair id="baa501de-d72c-4c87-b2e3-d8fd35608197" name="fstype" value="ext3"/>
</attributes>
</instance_attributes>
<operations/>
<meta_attributes id="resource_tomcat_LUN_meta_attrs">
<attributes>
<nvpair id="resource_tomcat_LUN_metaattr_target_role" name="target_role" value="started"/>
</attributes>
</meta_attributes>
</primitive>
<primitive id="resource_tomcat_TOMCAT" class="lsb" type="tomcat6" provider="heartbeat">
<meta_attributes id="resource_tomcat_TOMCAT_meta_attrs">
<attributes>
<nvpair id="resource_tomcat_TOMCAT_metaattr_target_role" name="target_role" value="started"/>
</attributes>
</meta_attributes>
</primitive>
</group>
<primitive id="resource_stonith_maedoc1" class="stonith" type="external/riloe" provider="heartbeat">
<meta_attributes id="resource_stonith_maedoc1_meta_attrs">
<attributes/>
</meta_attributes>
<instance_attributes id="resource_stonith_maedoc1_instance_attrs">
<attributes>
<nvpair id="fbc7c10c-9cbe-4a46-90cc-599ce95b3344" name="hostlist" value="10.0.0.20"/>
<nvpair id="773c3e2d-87e8-4f9b-b213-454677d6457c" name="ilo_hostname" value="10.0.0.20"/>
<nvpair id="f9062414-a113-4d69-920e-6a8a283fc452" name="ilo_user" value="Administrator"/>
<nvpair id="193e4b72-dd86-49c6-94c9-d7d04c840653" name="ilo_password" value="i.l.o.v.2"/>
<nvpair id="fe62fa88-07cd-450d-84d1-7cc8b7c7ad57" name="ilo_can_reset" value="1"/>
<nvpair id="ddbea44c-e480-4393-b29a-1bdb79616240" name="ilo_protocol" value="2.0"/>
<nvpair id="d8338c4a-72e2-4ba0-92ce-d1aeff677831" name="ilo_powerdown_method" value="power"/>
</attributes>
</instance_attributes>
</primitive>
<primitive id="resource_stonith_maedoc2" class="stonith" type="external/riloe" provider="heartbeat">
<meta_attributes id="resource_stonith_maedoc2_meta_attrs">
<attributes/>
</meta_attributes>
<instance_attributes id="resource_stonith_maedoc2_instance_attrs">
<attributes>
<nvpair id="0f68a8b7-b2a7-4128-b1ca-1a7924b8e1bc" name="hostlist" value="10.0.0.10"/>
<nvpair id="287e625d-5f23-4a14-925d-c6234d54ad66" name="ilo_hostname" value="10.0.0.10"/>
<nvpair id="96eac4bc-4883-4cc0-b799-4ccf90416643" name="ilo_user" value="Administrator"/>
<nvpair id="e87714b5-39fe-4535-82d0-e93132219ce9" name="ilo_password" value="i.l.o.v.2"/>
<nvpair id="9074b348-9e16-4873-b53f-2502fadf9a43" name="ilo_can_reset" value="1"/>
<nvpair id="60664375-1ece-4d5d-accf-057bf94921bd" name="ilo_protocol" value="2.0"/>
<nvpair id="eb9e3eff-4839-421a-a7c6-83893747eade" name="ilo_powerdown_method" value="power"/>
</attributes>
</instance_attributes>
</primitive>
</resources>
<constraints>
<rsc_location id="location_oracle" rsc="group_oracle">
<rule id="prefered_location_oracle" score="100">
<expression attribute="#uname" id="db81f525-6d4e-45dd-97dd-9c9b0fc223da" operation="eq" value="maedoc1"/>
</rule>
</rsc_location>
<rsc_location id="location_tomcat" rsc="group_tomcat">
<rule id="prefered_location_tomcat" score="100">
<expression attribute="#uname" id="7b6c859a-356f-4868-b645-0280dcace5da" operation="eq" value="maedoc2"/>
</rule>
</rsc_location>
<rsc_location id="location_stonith_maedoc1" rsc="resource_stonith_maedoc1">
<rule id="prefered_location_stonith_maedoc1" score="-INFINITY">
<expression attribute="#uname" id="ef1a56e3-8cfe-43f7-be96-3e607b7afe92" operation="eq" value="maedoc2"/>
</rule>
</rsc_location>
<rsc_location id="location_stonith_maedoc2" rsc="resource_stonith_maedoc2">
<rule id="prefered_location_stonith_maedoc2" score="-INFINITY">
<expression attribute="#uname" id="f2219467-2399-426a-8a83-8607602cc9c2" operation="eq" value="maedoc1"/>
</rule>
</rsc_location>
</constraints>
</configuration>
</cib>
_______________________________________________
Linux-HA mailing list
Linux-HA@lists.linux-ha.org
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems