sorry i'm just curious, why you are using drbd if you have a share storage?
2013/8/28 Goutam Baul <goutam.b...@rp-sg.in> > ** ** > > ** ** > > >On 2013/8/27 wrote**** > > > > >Are you using san with emc powerpath?**** > > ** ** > > Yes**** > > ** ** > > >drbd service start on boot?**** > > >start drbd by hand and give the command ls -l *dev/drbd/by-res/r0/1 and > ls -l * > *>dev/drbd/by-res/r0/0***** > > ** ** > > It works fine if we do things manually. The only issue comes when we try > to run things under rgmanager.**** > > >>On 2013/8/27 Goutam Baul wrote**** > > >>Dear List,**** > > >>We have configured DRBD 8.4.3 (community version) in the sense that we > are able to manually start DRBD at both the nodes, make nodes primary or > secondary and >>data is getting replicated between the primary and the > secondary node. We are doing this for our postfix mail server. The drbd > configuration file is**** > > >>resource r0 {**** > > >>startup {**** > > >> wfc-timeout 1;**** > > >> degr-wfc-timeout 2;**** > > >> }**** > > >> volume 0 {**** > > >> device /dev/drbd1;**** > > >> disk /dev/emcpowera2;**** > > >> meta-disk /dev/emcpowera5;**** > > >> }**** > > >> volume 1 {**** > > >> device /dev/drbd2;**** > > >> disk /dev/emcpowera3;**** > > >> meta-disk /dev/emcpowera6;**** > > >> }**** > > >> on drmail1.cesc.co.in {**** > > >> address 10.50.4.14:7789;**** > > >> }**** > > >> on drbdtest2.cesc.co.in {**** > > >> address 10.50.81.253:7789;**** > > >> }**** > > >>}**** > > **** > > >>We want to run postfix mailing system on a RHCS cluster suite for high > availability. The mail-store and the mail-queue of postfix will be on the > DRBD setup so that >>they are replicated from our data center to our > disaster recovery site. The cluster.conf is as follows:**** > > **** > > >><?xml version="1.0"?>**** > > >><cluster config_version="27" name="dr-mail">**** > > >> <fence_daemon clean_start="1" post_join_delay="30"/>**** > > >> <clusternodes>**** > > >> <clusternode name="drmail1.cesc.co.in" nodeid="1">**** > > >> <fence>**** > > >> <method name="Method1_drac">**** > > >> <device name="fence_drac1"/>**** > > >> </method>**** > > >> </fence>**** > > >> </clusternode>**** > > >> <clusternode name="drmail2.cesc.co.in" nodeid="2">**** > > >> <fence>**** > > >> <method name="Method1_drac">**** > > >> <device name="fence_drac2"/>**** > > >> </method>**** > > >> </fence>**** > > >> </clusternode>**** > > >> </clusternodes>**** > > >> <cman expected_votes="1" two_node="1"/>**** > > >> <fencedevices>**** > > >> <fencedevice agent="fence_drac5" > cmd_prompt="admin1->" ipaddr="10.50.4.22" login="root" > module_name="fence_drac1" name="fence_drac1" passwd="calvin" secure="on"/> > **** > > >> <fencedevice agent="fence_drac5" > cmd_prompt="admin1->" ipaddr="10.50.4.23" login="root" > module_name="fence_drac2" name="fence_drac2" passwd="calvin" secure="on"/> > **** > > >> </fencedevices>**** > > >> <rm>**** > > >> <failoverdomains>**** > > >> <failoverdomain name="dr_mail" nofailback="0" > ordered="1" restricted="1">**** > > >> <failoverdomainnode name=" > drmail1.cesc.co.in" priority="1"/>**** > > >> <failoverdomainnode name=" > drmail2.cesc.co.in" priority="2"/>**** > > </failoverdomain>**** > > >> </failoverdomains>**** > > >> <resources>**** > > >> <ip address="10.50.4.20" monitor_link="on" > sleeptime="10"/>**** > > >> <drbd name="drbd-mail" resource="r0">**** > > >> <fs device="/dev/drbd/by-res/r0/0" > force_unmount="1" fsid="61850" mountpoint="/home" fstype="ext3" > name="drbd-mailstore" options="noatime"/>**** > > >> <fs device="/dev/drbd/by-res/r0/1" > force_unmount="1" fsid="61855" mountpoint="/var/spool/postfix" > fstype="ext3" name="drbd-spool" options="noatime"/>**** > > >> </drbd>**** > > >> <script file="/etc/init.d/postfix" > name="postfix" />**** > > >> </resources>**** > > >> <service domain="dr_mail" name="mail" > recovery="relocate">**** > > >> <ip ref="10.50.4.20"/>**** > > >> <drbd ref="drbd-mail" />**** > > <script ref="postfix" />**** > > >> </service>**** > > >> </rm>**** > > >> </cluster>**** > > >> If we run this RHCS cluster, we get the error in the log as **** > > >> *Aug 27 17:38:19 drmail1 rgmanager[4695]: drbd not found in > /proc/modules. Do you need to modprobe?***** > > * >> *If we run the RHCS cluster after issuing the modprobe drbd command > we get the following message**** > > >> *1: Failure: (127) Device minor not allocated***** > > *>> additional info from kernel:***** > > *>> unknown minor***** > > *>> Command 'drbdsetup role 1' terminated with exit code 10***** > > *>> <debug> DRBD resource r0 already configured***** > > *>> DRBD resource r0 already configured***** > > *>> Start of mail complete***** > > * >> *But the DRBD service is not starting up and therefore the > corresponding devices are not being created. Thus the log file returns the > error**** > > *>> Aug 27 17:38:33 drmail1 rgmanager[5147]: stop: Could not match > /dev/drbd/by-res/r0/1 with a real device***** > > *>> Aug 27 17:38:33 drmail1 rgmanager[5166]: stop: Could not match > /dev/drbd/by-res/r0/0 with a real device***** > > >> We have built the DRBD rpm files from the source using rpmbuild and > copied the drbd.sh.rhcs and drbd.metadata.rhcs files from the script folder > of the source to the >> /usr/share/cluster folder as drbd.sh and > drbd.metadata respectively. We are using RHEL 6.1 (64 bit).**** > > **** > > -- > Linux-cluster mailing list > Linux-cluster@redhat.com > https://www.redhat.com/mailman/listinfo/linux-cluster > -- esta es mi vida e me la vivo hasta que dios quiera
-- Linux-cluster mailing list Linux-cluster@redhat.com https://www.redhat.com/mailman/listinfo/linux-cluster