Hi,

i have a Cluster running on SLES 11 SP3. A resource group is defined. So every 
entry in that resource group is a hard dependency for the folliowing resource 
in that group. If just only one resource fails on both nodes, the resource 
res_MyApplication won't be started. I want to change that behavior. I want the 
cluster to start the resources on just one host. The local resources should be 
started on just the same host, regardless of the status of the CIFS Shares. If 
the CIFS Shares are available they have to be mounted on the same node running 
the application. I would try to accomplish this by deleting the resource group 
grp_application, creating colocation constrains for local fs, service ip and 
application. Additionally I would create co-location constrains for each CIFS 
Share to mount it on the node running the application. Any hints/thoughts on 
that? Is that the "right way" to achieve what I want?

Here's the currently running config:

node mynodea\
        attributes standby="off"
node mynodeb \
        attributes standby="off"
primitive res_Service-IP ocf:heartbeat:IPaddr2 \
        params ip="192.168.10.120" cidr_netmask="24" \
        op monitor interval="10s" timeout="20s" depth="0" \
        meta target-role="started"
primitive res_MyApplication lsb:myapp \
        operations $id="res_MyApplication-operations" \
        op start interval="0" timeout="180" \
        op stop interval="0" timeout="180" \
        op status interval="0" timeout="600" \
        op monitor interval="15" timeout="15" start-delay="120" \
        op meta-data interval="0" timeout="600" \
        meta target-role="Started"
primitive res_mount-CIFSshareData ocf:heartbeat:Filesystem \
        params device="//CIFS/Share1" directory="/Datafstype="cifs" 
options="uid=30666,gid=30666" \
        op monitor interval="20" timeout="40" depth="0" \
        meta target-role="started"
primitive res_mount-CIFSshareData2 ocf:heartbeat:Filesystem \
        params device="//CIFS/Share2" directory="/Data2" fstype="cifs" 
options="uid=30666,gid=30666" \
        op monitor interval="20" timeout="40" depth="0" \
        meta target-role="started"
primitive res_mount-CIFSshareData3 ocf:heartbeat:Filesystem \
        params device="//CIFS/Share3" directory="/Data3" fstype="cifs" 
options="uid=30666,gid=30666" \
        op monitor interval="20" timeout="40" depth="0" \
        meta target-role="started"
primitive res_mount-application ocf:heartbeat:Filesystem \
        params 
device="/dev/disk/by-id/scsi-36000d56e0000324561d9dcae19034e90-part1" 
directory="/MyApp" fstype="ext3" \
        op monitor interval="20" timeout="40" depth="0" \
        meta target-role="started"
primitive stonith-sbd stonith:external/sbd \
        params 
sbd_device="/dev/disk/by-id/scsi-36000d56e0000a31a36d8dcaebaf5a439" \
        meta target-role="Started"
group grp_application res_mount-application res_Service-IP 
res_mount-CIFSshareData res_mount-CIFSshareData2  res_mount-CIFSshareData3 
res_MyApplication
location prefer-application grp_application 50: nodea
property $id="cib-bootstrap-options" \
        stonith-enabled="true" \
        no-quorum-policy="ignore" \
        placement-strategy="balanced" \
        dc-version="1.1.9-2db99f1" \
        cluster-infrastructure="classic openais (with plugin)" \
        expected-quorum-votes="2" \
        last-lrm-refresh="1412758622" \
        stonith-action="poweroff" \
        stonith-timeout="216s" \
        maintenance-mode="true"
rsc_defaults $id="rsc-options" \
        resource-stickiness="100" \
        migration-threshold="3"
op_defaults $id="op-options" \
        timeout="600" \
        record-pending="true"

Thanks and kind regards

Sven

_______________________________________________
Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
http://oss.clusterlabs.org/mailman/listinfo/pacemaker

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org

Reply via email to