On Tue, Dec 9, 2008 at 19:40, guillermo <[EMAIL PROTECTED]> wrote:
> Andrew;
>
> sorry for having been so unclear, basically the cluster is working in
> the right way, when the services start this is made in the right
> moment , even the files's assembly make it after the dbrd starts and put
> the right node as master, this is:
>
> the cluster starts:
>
> - the ip virtual sets.
> - the drbd starts.
> - mysql starts.
> - asterisk starts.
> - the principal node designed to master states.
> then the system's files mount in the node which was designed as a
> master.
>
> so far it's ok , if for example , then I turn off the cluster,

using "/etc/init.d/heartbeat stop" or something else?

> which is
> the primary , the services turn off in time and they do it ok  and then
> they  set in the other node in time and ok, as I' ve already said
> everything is ok now, what is my problem?, if for example a service
> falls, as mysql, which is happening is that only a piece of the services
> which are found in the group_1 they pass to the other node ,
> the
> services that pass to the other node are:
> ipaddr
> mysql
> and asterick,
> the drbd and the file's system continue in the secondary node, which I
> need is before the first failure of any of the services , they pass, all
> of them, to the other node.

So when a member of the group fails, you want _everything_ to move to
the other node.
Is that correct?

If so, then you need to colocate the group with drbd.

> In the contraints, resources location I have declared the pbx-2 node as
> prefered, when the services start, in the resources order I declared the
> drbd started before the filesystem mount in the pbx-2 node and the
> colocation resource that the file's system mount in the same node where
> the drbd master is.
>
> This is what the crm_resource -L give me back:
>
>  Master/Slave Set: ms-r0
>    r0:0        (ocf::heartbeat:drbd)
>    r0:1        (ocf::heartbeat:drbd)
>  Resource Group: group_1
>    IPaddr_192_168_123_205      (ocf::heartbeat:IPaddr)
>    mysql_2     (ocf::heartbeat:mysql)
>    apache2_2   (lsb:apache2)
>    asterisk_3  (lsb:asterisk)
>  fs0     (ocf::heartbeat:Filesystem)
>
>
> I wish having been clear, any information you can provide me is
> welcomed.
> regards,
>
> Guillermo
>
>
> El mar, 09-12-2008 a las 11:40 +0100, Andrew Beekhof escribió:
>> I'm sorry, but I'm having trouble parsing this...
>>
>> On Fri, Dec 5, 2008 at 19:25, guillermo <[EMAIL PROTECTED]> wrote:
>> > I have the following problem and I am not able to solve it; I have
>> > configured Heartbeat v2 plus DRBD in two nodes, one of then is
>> > configured to act as DRBD master, so it has preference in relation to
>> > the other node when it has to start,
>>
>> are you talking about rsc_location or rsc_order constraints here?
>>
>> > because in the master assembles the
>> > file system, together with the DRBD I have configured others services,
>> > these services carry out in the master node, however, the problem is
>> > when I want to before the first failure in some of the resources, all of
>> > the servcices change to the other node, as I have configured it to the
>> > scores, for example if I kill the mysql process the group_1 passes to
>> > the other node, but the DRBD's and the file system assembled remain in
>> > the same node, what I need is, before the failure of any of the other
>> > resources all the servcies pass to the other nodes.
>>
>> you lost me here... but looking at your configuration, I think you
>> need some rsc_colocation and rsc_order constraints so that the group
>> runs on the same machine as the filesystem (and starts after it too).
>>
>> > I am using Debian etch, Heartbeat 2.1.4 (stable) and DRBD 0.7
>> >
>> > This is what the crm_resource -L give me back:
>> >
>> > Master/Slave Set: ms-r0
>> >    r0:0        (ocf::heartbeat:drbd)
>> >    r0:1        (ocf::heartbeat:drbd)
>> > Resource Group: group_1
>> >    IPaddr_192_168_123_205      (ocf::heartbeat:IPaddr)
>> >    mysql_2     (ocf::heartbeat:mysql)
>> >    apache2_2   (lsb:apache2)
>> >    asterisk_3  (lsb:asterisk)
>> > fs0     (ocf::heartbeat:Filesystem)
>> >
>> >
>> > This is my cib.xml's file:
>> >
>> >  <cib generated="true" admin_epoch="0" have_quorum="true"
>> > ignore_dtd="false" num_peers="2" cib_feature_revision="2.0"
>> > crm_feature_set="2.0" epoch="52" num_updates="3" cib-last-written="Fri
>> > Dec  5 15:15:13 2008" ccm_transition="2"
>> > dc_uuid="813044e0-db95-40b2-9e48-67ec2e3e6584">
>> >   <configuration>
>> >     <crm_config>
>> >       <cluster_property_set id="cib-bootstrap-options">
>> >         <attributes>
>> >           <nvpair id="cib-bootstrap-options-dc-version"
>> > name="dc-version" value="2.1.4-node:
>> > aa909246edb386137b986c5773344b98c6969999"/>
>> >           <nvpair id="cib-bootstrap-options-symmetric-cluster"
>> > name="symmetric-cluster" value="true"/>
>> >           <nvpair id="cib-bootstrap-options-no-quorum-policy"
>> > name="no-quorum-policy" value="stop"/>
>> >           <nvpair
>> > id="cib-bootstrap-options-default-resource-stickiness"
>> > name="default-resource-stickiness" value="100"/>
>> >           <nvpair
>> > id="cib-bootstrap-options-default-resource-failure-stickiness"
>> > name="default-resource-failure-stickiness" value="-400"/>
>> >           <nvpair id="cib-bootstrap-options-stonith-enabled"
>> > name="stonith-enabled" value="false"/>
>> >           <nvpair id="cib-bootstrap-options-stonith-action"
>> > name="stonith-action" value="reboot"/>
>> >           <nvpair id="cib-bootstrap-options-startup-fencing"
>> > name="startup-fencing" value="true"/>
>> >           <nvpair id="cib-bootstrap-options-stop-orphan-resources"
>> > name="stop-orphan-resources" value="true"/>
>> >           <nvpair id="cib-bootstrap-options-stop-orphan-actions"
>> > name="stop-orphan-actions" value="true"/>
>> >           <nvpair id="cib-bootstrap-options-remove-after-stop"
>> > name="remove-after-stop" value="false"/>
>> >           <nvpair id="cib-bootstrap-options-short-resource-names"
>> > name="short-resource-names" value="true"/>
>> >           <nvpair id="cib-bootstrap-options-transition-idle-timeout"
>> > name="transition-idle-timeout" value="5min"/>
>> >           <nvpair id="cib-bootstrap-options-default-action-timeout"
>> > name="default-action-timeout" value="20s"/>
>> >           <nvpair id="cib-bootstrap-options-is-managed-default"
>> > name="is-managed-default" value="true"/>
>> >           <nvpair id="cib-bootstrap-options-cluster-delay"
>> > name="cluster-delay" value="60s"/>
>> >           <nvpair id="cib-bootstrap-options-pe-error-series-max"
>> > name="pe-error-series-max" value="-1"/>
>> >           <nvpair id="cib-bootstrap-options-pe-warn-series-max"
>> > name="pe-warn-series-max" value="-1"/>
>> >           <nvpair id="cib-bootstrap-options-pe-input-series-max"
>> > name="pe-input-series-max" value="-1"/>
>> >         </attributes>
>> >       </cluster_property_set>
>> >     </crm_config>
>> >     <nodes>
>> >       <node id="813044e0-db95-40b2-9e48-67ec2e3e6584" uname="pbx-2"
>> > type="normal"/>
>> >       <node id="91471b98-85a3-4f9c-a414-1f889143c8be" uname="pbx-1"
>> > type="normal"/>
>> >     </nodes>
>> >     <resources>
>> >       <master_slave id="ms-r0">
>> >         <meta_attributes id="ma-ms-r0">
>> >           <attributes>
>> >             <nvpair id="ma-ms-r0-1" name="clone_max" value="2"/>
>> >             <nvpair id="ma-ms-r0-2" name="clone_node_max" value="1"/>
>> >             <nvpair id="ma-ms-r0-3" name="master_max" value="1"/>
>> >             <nvpair id="ma-ms-r0-4" name="master_node_max" value="1"/>
>> >             <nvpair id="ma-ms-r0-5" name="notify" value="yes"/>
>> >             <nvpair id="ma-ms-r0-6" name="globally_unique"
>> > value="false"/>
>> >             <nvpair id="ma-ms-r0-7" name="target_role"
>> > value="started"/>
>> >           </attributes>
>> >         </meta_attributes>
>> >         <primitive id="r0" class="ocf" provider="heartbeat"
>> > type="drbd">
>> >           <instance_attributes id="ia-r0">
>> >             <attributes>
>> >               <nvpair id="ia-r0-1" name="drbd_resource" value="r0"/>
>> >             </attributes>
>> >           </instance_attributes>
>> >           <operations>
>> >             <op id="op-r0-1" name="monitor" interval="59s"
>> > timeout="10s" role="Master"/>
>> >             <op id="op-r0-2" name="monitor" interval="60s"
>> > timeout="10s" role="Slave"/>
>> >           </operations>
>> >         </primitive>
>> >       </master_slave>
>> >       <group id="group_1">
>> >         <primitive class="ocf" id="IPaddr_192_168_123_205"
>> > provider="heartbeat" type="IPaddr">
>> >           <operations>
>> >             <op id="IPaddr_192_168_123_205_mon" interval="5s"
>> > name="monitor" timeout="5s"/>
>> >           </operations>
>> >           <instance_attributes id="IPaddr_192_168_123_205_inst_attr">
>> >             <attributes>
>> >               <nvpair id="IPaddr_192_168_123_205_attr_0" name="ip"
>> > value="192.168.123.205"/>
>> >               <nvpair id="IPaddr_192_168_123_205_attr_1" name="netmask"
>> > value="24"/>
>> >               <nvpair id="IPaddr_192_168_123_205_attr_2" name="nic"
>> > value="eth1"/>
>> >             </attributes>
>> >           </instance_attributes>
>> >         </primitive>
>> >         <primitive class="ocf" id="mysql_2" provider="heartbeat"
>> > type="mysql">
>> >           <operations>
>> >             <op id="mysql_2_mon" interval="120s" name="monitor"
>> > timeout="60s"/>
>> >           </operations>
>> >         </primitive>
>> >         <primitive class="lsb" id="apache2_2" provider="heartbeat"
>> > type="apache2">
>> >           <operations>
>> >             <op id="apache2_2_mon" interval="120s" name="monitor"
>> > timeout="60s"/>
>> >           </operations>
>> >         </primitive>
>> >         <primitive class="lsb" id="asterisk_3" provider="heartbeat"
>> > type="asterisk">
>> >           <operations>
>> >             <op id="asterisk_3_mon" interval="120s" name="monitor"
>> > timeout="60s"/>
>> >           </operations>
>> >         </primitive>
>> >       </group>
>> >       <primitive class="ocf" provider="heartbeat" type="Filesystem"
>> > id="fs0">
>> >         <instance_attributes id="ia-group_1">
>> >           <attributes>
>> >             <nvpair id="ia-fs0-1" name="fstype" value="ext3"/>
>> >             <nvpair id="ia-fs0-2" name="directory"
>> > value="/replicated"/>
>> >             <nvpair id="ia-fs0-3" name="device" value="/dev/drbd0"/>
>> >           </attributes>
>> >         </instance_attributes>
>> >       </primitive>
>> >     </resources>
>> >     <constraints>
>> >       <rsc_location id="rsc_location_group_1" rsc="group_1">
>> >         <rule id="prefered_location_group_1" score="100">
>> >           <expression attribute="#uname"
>> > id="prefered_location_group_1_expr" operation="eq" value="pbx-2"/>
>> >         </rule>
>> >       </rsc_location>
>> >       <rsc_location id="r0_master_on_pbx-2" rsc="ms-r0">
>> >         <rule id="r0_master_on_pbx-2_rule1" role="master" score="100">
>> >           <expression id="r0_master_on_pbx-2_expression1"
>> > attribute="#uname" operation="eq" value="pbx-2"/>
>> >         </rule>
>> >       </rsc_location>
>> >       <rsc_order id="r0_before_fs0" to_action="start" to="fs0"
>> > type="before" action="promote" from="ms-r0"/>
>> >       <rsc_colocation id="fs0_on_r0" to="ms-r0" to_role="master"
>> > from="fs0" score="100"/>
>> >     </constraints>
>> >   </configuration>
>> >  </cib>
>> >
>> >
>> >
>> > I will appreciate if anyone can help me to solve it.
>> >
>> > Regards
>> >
>> > Guillermo
>> >
>> > _______________________________________________
>> > Linux-HA mailing list
>> > Linux-HA@lists.linux-ha.org
>> > http://lists.linux-ha.org/mailman/listinfo/linux-ha
>> > See also: http://linux-ha.org/ReportingProblems
>> >
>> _______________________________________________
>> Linux-HA mailing list
>> Linux-HA@lists.linux-ha.org
>> http://lists.linux-ha.org/mailman/listinfo/linux-ha
>> See also: http://linux-ha.org/ReportingProblems
>
> _______________________________________________
> Linux-HA mailing list
> Linux-HA@lists.linux-ha.org
> http://lists.linux-ha.org/mailman/listinfo/linux-ha
> See also: http://linux-ha.org/ReportingProblems
>
_______________________________________________
Linux-HA mailing list
Linux-HA@lists.linux-ha.org
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems

Reply via email to