I have disabled the services and run
drbdadm secondary all
drbdadm detach all
drbdadm down all
service drbd stop

before testing as far as I can see (cat /proc/drbd on both nodes) drbd is
shutdown

cat: /proc/drbd: No such file or directory


I have taken the command that heartbeat is running (drbdsetup /dev/drbd0
disk /dev/sdb /dev/sdb internal --set-defaults --create-device
--on-io-error=pass_on') and run it against the nodes when heartbeat is not
in control and this command will bring the resources online, but re-running
this command will generate the error, so I am kind of leaning twords the
command being run twice?

Thanks

Jason

2009/2/11 Dominik Klein <d...@in-telegence.net>

> Hi Jason
>
> any chance you started drbd at boot or the drbd device was active at the
> time you started the cluster resource? If so, read the introduction of
> the howto again and correct your setup.
>
> Jason Fitzpatrick wrote:
> > Hi Dominik
> >
> > I have upgraded to HB 2.9xx and have been following the instructions that
> > you provided (thanks for those) and have added a resource as follows
> >
> > crm
> > configure
> > primitive Storage1 ocf:heartbeat:drbd \
> > params drbd_resource=Storage1 \
> > op monitor role=Master interval=59s timeout=30s \
> > op monitor role=Slave interval=60s timeout=30s
> > ms DRBD_Storage Storage1 \
> > meta clone-max=2 notify=true globally-unique=false target-role=stopped
> > commit
> > exit
> >
> > no errors are reported and the resource is visable from within the hb_gui
> >
> > when I try to bring the resource online with
> >
> > crm resource start DRBD_Storage
> >
> > I see the resource attempt to come online and then fail, it seems to be
> > starting the services, changing the status of the devices to attached
> (from
> > detached) but not setting any device to master
> >
> > the following is from the ha-log
> >
> > crmd[8020]: 2009/02/10_17:22:32 info: do_lrm_rsc_op: Performing
> > key=7:166:0:b57f7f7c-4e2d-4134-9c14-b1a2b7db11a7 op=Storage1:1_start_0 )
> > lrmd[8016]: 2009/02/10_17:22:32 info: rsc:Storage1:1: start
> > lrmd[8016]: 2009/02/10_17:22:32 info: RA output:
> (Storage1:1:start:stdout)
> > /dev/drbd0: Failure: (124) Device is attached to a disk (use detach
> first)
> > Command
> >  'drbdsetup /dev/drbd0 disk /dev/sdb /dev/sdb internal --set-defaults
> > --create-device --on-io-error=pass_on' terminated with exit code 10
>
> This looks like "drbdadm up" is failing because the device is already
> attached to the lower level storage device.
>
> Regards
> Dominik
>
> > drbd[22270]:    2009/02/10_17:22:32 ERROR: Storage1 start: not in
> Secondary
> > mode after start.
> > crmd[8020]: 2009/02/10_17:22:32 info: process_lrm_event: LRM operation
> > Storage1:1_start_0 (call=189, rc=1, cib-update=380, confirmed=true)
> complete
> > unknown e
> > rror
> > .
> >
> > I have checked the DRBD device Storage1 and it is in secondary mode after
> > the start, and should I choose I can make it primary on either node
> >
> > Thanks
> >
> > Jason
> >
> > 2009/2/10 Jason Fitzpatrick <jayfitzpatr...@gmail.com>
> >
> >> Thanks,
> >>
> >> This was the latest version in the Fedora Repos, I will upgrade and see
> >> what happens
> >>
> >> Jason
> >>
> >> 2009/2/10 Dominik Klein <d...@in-telegence.net>
> >>
> >> Jason Fitzpatrick wrote:
> >>>>> Hi All
> >>>>>
> >>>>> I am having a hell of a time trying to get heartbeat to fail over my
> >>> DRBD
> >>>>> harddisk and am hoping for some help.
> >>>>>
> >>>>> I have a 2 node cluster, heartbeat is working as I am able to fail
> over
> >>> IP
> >>>>> Addresses and services successfully, but when I try to fail over my
> >>> DRBD
> >>>>> resource from secondary to primary I am hitting a brick wall, I can
> >>> fail
> >>>>> over the DRBD resource manually so I know that it does work at some
> >>> level
> >>>>> DRBD version 8.3 Heartbeat version heartbeat-2.1.3-1.fc9.i386
> >>> Please upgrade. Thats too old for reliable master/slave behaviour.
> >>> Preferrably upgrade to pacemaker and ais or heartbeat 2.99. Read
> >>> http://www.clusterlabs.org/wiki/Install for install notes.
> >>>
> >>>>> and using
> >>>>> heartbeat-gui to configure
> >>> Don't use the gui to configure complex (ie clone or master/slave)
> >>> resources.
> >>>
> >>> Once you upgraded to the latest pacemaker, please refer to
> >>> http://www.clusterlabs.org/wiki/DRBD_HowTo_1.0 for drbd's cluster
> >>> configuration.
> >>>
> >>> Regards
> >>> Dominik
> >>>
> >>>>> DRBD Resource is called Storage1, the 2 nodes are connected via 2
> >>> x-over
> >>>>> cables (1 heartbeat 1 Replication)
> >>>>>
> >>>>> I have stripped down my config to the bare bones and tried every
> option
> >>>>> that I can think off but know that I am missing something simple,
> >>>>>
> >>>>> I have attached my cib.xml but have removed domain names from the
> >>> systems
> >>>>> for privacy reasons
> >>>>>
> >>>>> Thanks in advance
> >>>>>
> >>>>> Jason
> >>>>>
> >>>>>  <cib admin_epoch="0" have_quorum="true" ignore_dtd="false"
> >>>>> cib_feature_revision="2.0" num_peers="2" generated="true"
> >>>>> ccm_transition="22" dc_uuid="9d8abc28-4fa3-408a-a695-fb36b0d67a48"
> >>>>> epoch="733" num_updates="1" cib-last-written="Mon Feb  9 18:31:19
> >>> 2009">
> >>>>>    <configuration>
> >>>>>      <crm_config>
> >>>>>        <cluster_property_set id="cib-bootstrap-options">
> >>>>>          <attributes>
> >>>>>            <nvpair id="cib-bootstrap-options-dc-version"
> >>> name="dc-version"
> >>>>> value="2.1.3-node: 552305612591183b1628baa5bc6e903e0f1e26a3"/>
> >>>>>            <nvpair name="last-lrm-refresh"
> >>>>> id="cib-bootstrap-options-last-lrm-refresh" value="1234204278"/>
> >>>>>          </attributes>
> >>>>>        </cluster_property_set>
> >>>>>      </crm_config>
> >>>>>      <nodes>
> >>>>>        <node id="df707752-d5fb-405a-8ca7-049e25a227b7"
> >>> uname="lpissan1001"
> >>>>> type="normal">
> >>>>>          <instance_attributes
> >>>>> id="nodes-df707752-d5fb-405a-8ca7-049e25a227b7">
> >>>>>            <attributes>
> >>>>>              <nvpair
> id="standby-df707752-d5fb-405a-8ca7-049e25a227b7"
> >>>>> name="standby" value="off"/>
> >>>>>            </attributes>
> >>>>>          </instance_attributes>
> >>>>>        </node>
> >>>>>        <node id="9d8abc28-4fa3-408a-a695-fb36b0d67a48"
> >>> uname="lpissan1002"
> >>>>> type="normal">
> >>>>>          <instance_attributes
> >>>>> id="nodes-9d8abc28-4fa3-408a-a695-fb36b0d67a48">
> >>>>>            <attributes>
> >>>>>              <nvpair
> id="standby-9d8abc28-4fa3-408a-a695-fb36b0d67a48"
> >>>>> name="standby" value="off"/>
> >>>>>            </attributes>
> >>>>>          </instance_attributes>
> >>>>>        </node>
> >>>>>      </nodes>
> >>>>>      <resources>
> >>>>>        <master_slave id="Storage1">
> >>>>>          <meta_attributes id="Storage1_meta_attrs">
> >>>>>            <attributes>
> >>>>>              <nvpair id="Storage1_metaattr_target_role"
> >>> name="target_role"
> >>>>> value="started"/>
> >>>>>              <nvpair id="Storage1_metaattr_clone_max"
> name="clone_max"
> >>>>> value="2"/>
> >>>>>              <nvpair id="Storage1_metaattr_clone_node_max"
> >>>>> name="clone_node_max" value="1"/>
> >>>>>              <nvpair id="Storage1_metaattr_master_max"
> >>> name="master_max"
> >>>>> value="1"/>
> >>>>>              <nvpair id="Storage1_metaattr_master_node_max"
> >>>>> name="master_node_max" value="1"/>
> >>>>>              <nvpair id="Storage1_metaattr_notify" name="notify"
> >>>>> value="true"/>
> >>>>>              <nvpair id="Storage1_metaattr_globally_unique"
> >>>>> name="globally_unique" value="false"/>
> >>>>>            </attributes>
> >>>>>          </meta_attributes>
> >>>>>          <primitive id="Storage1" class="ocf" type="drbd"
> >>>>> provider="heartbeat">
> >>>>>            <instance_attributes id="Storage1_instance_attrs">
> >>>>>              <attributes>
> >>>>>                <nvpair id="273a1bb2-4867-42dd-a9e5-7cebbf48ef3b"
> >>>>> name="drbd_resource" value="Storage1"/>
> >>>>>              </attributes>
> >>>>>            </instance_attributes>
> >>>>>            <operations>
> >>>>>              <op id="9ddc0ce9-4090-4546-a7d5-787fe47de872"
> >>> name="monitor"
> >>>>> description="master" interval="29" timeout="10" start_delay="1m"
> >>>>> role="Master"/>
> >>>>>              <op id="56a7508f-fa42-46f8-9924-3b284cdb97f0"
> >>> name="monitor"
> >>>>> description="slave" interval="29" timeout="10" start_delay="1m"
> >>>>> role="Slave"/>
> >>>>>            </operations>
> >>>>>          </primitive>
> >>>>>        </master_slave>
> >>>>>      </resources>
> >>>>>      <constraints/>
> >>>>>    </configuration>
> >>>>>  </cib>
> >>>>>
> >>>>>
> >>>> _______________________________________________
> >>>> Linux-HA mailing list
> >>>> Linux-HA@lists.linux-ha.org
> >>>> http://lists.linux-ha.org/mailman/listinfo/linux-ha
> >>>> See also: http://linux-ha.org/ReportingProblems
> >>>>
> >>> _______________________________________________
> >>> Linux-HA mailing list
> >>> Linux-HA@lists.linux-ha.org
> >>> http://lists.linux-ha.org/mailman/listinfo/linux-ha
> >>> See also: http://linux-ha.org/ReportingProblems
> >>>
> >>
> > _______________________________________________
> > Linux-HA mailing list
> > Linux-HA@lists.linux-ha.org
> > http://lists.linux-ha.org/mailman/listinfo/linux-ha
> > See also: http://linux-ha.org/ReportingProblems
> >
>
> _______________________________________________
> Linux-HA mailing list
> Linux-HA@lists.linux-ha.org
> http://lists.linux-ha.org/mailman/listinfo/linux-ha
> See also: http://linux-ha.org/ReportingProblems
>
_______________________________________________
Linux-HA mailing list
Linux-HA@lists.linux-ha.org
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems

Reply via email to