sorry i'm just curious, why you are using drbd if you have a share storage?


2013/8/28 Goutam Baul <goutam.b...@rp-sg.in>

> ** **
>
> ** **
>
> >On 2013/8/27 wrote****
>
>
>
> >Are you using san with emc powerpath?****
>
> ** **
>
> Yes****
>
> ** **
>
> >drbd service start on boot?****
>
> >start drbd by hand and give the command ls -l *dev/drbd/by-res/r0/1 and
> ls -l *
> *>dev/drbd/by-res/r0/0*****
>
> ** **
>
> It works fine if we do things manually. The only issue comes when we try
> to run things under rgmanager.****
>
> >>On 2013/8/27 Goutam Baul wrote****
>
> >>Dear List,****
>
>  >>We have configured DRBD 8.4.3 (community version) in the sense that we
> are able to manually  start DRBD at both the nodes, make nodes primary or
> secondary and >>data is getting replicated between the primary and the
> secondary node. We are doing this for our postfix mail server. The drbd
> configuration file is****
>
>  >>resource r0 {****
>
> >>startup {****
>
> >>    wfc-timeout 1;****
>
> >>    degr-wfc-timeout 2;****
>
> >>  }****
>
> >>  volume 0 {****
>
> >>    device    /dev/drbd1;****
>
> >>    disk      /dev/emcpowera2;****
>
> >>    meta-disk /dev/emcpowera5;****
>
> >>  }****
>
> >>  volume 1 {****
>
> >>    device    /dev/drbd2;****
>
> >>    disk      /dev/emcpowera3;****
>
> >>    meta-disk /dev/emcpowera6;****
>
> >>  }****
>
> >>  on drmail1.cesc.co.in {****
>
> >>    address   10.50.4.14:7789;****
>
> >>  }****
>
> >>  on drbdtest2.cesc.co.in {****
>
> >>    address   10.50.81.253:7789;****
>
> >>  }****
>
> >>}****
>
>  ****
>
> >>We want to run postfix mailing system on a RHCS cluster suite for high
> availability. The mail-store and the mail-queue of postfix will be on the
> DRBD setup so that >>they are replicated from our data center to our
> disaster recovery site. The cluster.conf is as follows:****
>
>  ****
>
> >><?xml version="1.0"?>****
>
> >><cluster config_version="27" name="dr-mail">****
>
>  >>       <fence_daemon clean_start="1" post_join_delay="30"/>****
>
>  >>       <clusternodes>****
>
> >>                <clusternode name="drmail1.cesc.co.in" nodeid="1">****
>
> >>                        <fence>****
>
> >>                                <method name="Method1_drac">****
>
> >>                                        <device name="fence_drac1"/>****
>
>  >>                               </method>****
>
>  >>                        </fence>****
>
>  >>               </clusternode>****
>
> >>                <clusternode name="drmail2.cesc.co.in" nodeid="2">****
>
> >>                        <fence>****
>
> >>                                <method name="Method1_drac">****
>
> >>                                        <device name="fence_drac2"/>****
>
> >>                                </method>****
>
>  >>                        </fence>****
>
> >>                </clusternode>****
>
> >>        </clusternodes>****
>
> >>        <cman expected_votes="1" two_node="1"/>****
>
> >>        <fencedevices>****
>
> >>                <fencedevice agent="fence_drac5"
> cmd_prompt="admin1-&gt;" ipaddr="10.50.4.22" login="root"
> module_name="fence_drac1" name="fence_drac1" passwd="calvin" secure="on"/>
> ****
>
>  >>               <fencedevice agent="fence_drac5"
> cmd_prompt="admin1-&gt;" ipaddr="10.50.4.23" login="root"
> module_name="fence_drac2" name="fence_drac2" passwd="calvin" secure="on"/>
> ****
>
>  >>        </fencedevices>****
>
>   >>      <rm>****
>
>    >>             <failoverdomains>****
>
>  >>                       <failoverdomain name="dr_mail" nofailback="0"
> ordered="1" restricted="1">****
>
>  >>                               <failoverdomainnode name="
> drmail1.cesc.co.in" priority="1"/>****
>
> >>                                <failoverdomainnode name="
> drmail2.cesc.co.in" priority="2"/>****
>
>                         </failoverdomain>****
>
>  >>               </failoverdomains>****
>
>  >>               <resources>****
>
>   >>                     <ip address="10.50.4.20" monitor_link="on"
> sleeptime="10"/>****
>
>  >>                       <drbd name="drbd-mail" resource="r0">****
>
>   >>                            <fs device="/dev/drbd/by-res/r0/0"
> force_unmount="1" fsid="61850" mountpoint="/home" fstype="ext3"
> name="drbd-mailstore" options="noatime"/>****
>
>     >>                          <fs device="/dev/drbd/by-res/r0/1"
> force_unmount="1" fsid="61855" mountpoint="/var/spool/postfix"
> fstype="ext3" name="drbd-spool" options="noatime"/>****
>
>   >>                      </drbd>****
>
>   >>                      <script file="/etc/init.d/postfix"
> name="postfix" />****
>
>   >>              </resources>****
>
>  >>               <service domain="dr_mail" name="mail"
> recovery="relocate">****
>
>  >>                       <ip ref="10.50.4.20"/>****
>
>   >>                      <drbd ref="drbd-mail" />****
>
>                         <script ref="postfix" />****
>
>   >>              </service>****
>
>   >>      </rm>****
>
> >> </cluster>****
>
> >> If we run this RHCS cluster, we get the error in the log as ****
>
>  >> *Aug 27 17:38:19 drmail1 rgmanager[4695]: drbd not found in
> /proc/modules. Do you need to modprobe?*****
>
> * >> *If we run the RHCS cluster after issuing the modprobe drbd command
> we get the following message****
>
>  >> *1: Failure: (127) Device minor not allocated*****
>
> *>> additional info from kernel:*****
>
> *>> unknown minor*****
>
> *>> Command 'drbdsetup role 1' terminated with exit code 10*****
>
> *>> <debug>  DRBD resource r0 already configured*****
>
> *>> DRBD resource r0 already configured*****
>
> *>> Start of mail complete*****
>
> * >> *But the DRBD service is not starting up and therefore the
> corresponding devices are not being created. Thus the log file returns the
> error****
>
> *>> Aug 27 17:38:33 drmail1 rgmanager[5147]: stop: Could not match
> /dev/drbd/by-res/r0/1 with a real device*****
>
> *>> Aug 27 17:38:33 drmail1 rgmanager[5166]: stop: Could not match
> /dev/drbd/by-res/r0/0 with a real device*****
>
>  >> We have built the DRBD rpm files from the source using rpmbuild and
> copied the drbd.sh.rhcs and drbd.metadata.rhcs files from the script folder
> of the source to the >> /usr/share/cluster folder as drbd.sh and
> drbd.metadata respectively. We are using RHEL 6.1 (64 bit).****
>
>   ****
>
> --
> Linux-cluster mailing list
> Linux-cluster@redhat.com
> https://www.redhat.com/mailman/listinfo/linux-cluster
>



-- 
esta es mi vida e me la vivo hasta que dios quiera
-- 
Linux-cluster mailing list
Linux-cluster@redhat.com
https://www.redhat.com/mailman/listinfo/linux-cluster

Reply via email to