I remake my CRM config and reboot the virtual machines, and not have successful. the DRBD not start the primary, and not mount the /dev/drbd0 :-(
ha-slave: version: 8.3.13 (api:88/proto:86-96) srcversion: 697DE8B1973B1D8914F04DB 0: cs:Connected ro:Secondary/Secondary ds:UpToDate/UpToDate C r----- ns:0 nr:28672 dw:28672 dr:0 al:0 bm:5 lo:0 pe:0 ua:0 ap:0 ep:1 wo:n oos:0 ha-master: version: 8.3.13 (api:88/proto:86-96) srcversion: 697DE8B1973B1D8914F04DB 0: cs:Connected ro:Secondary/Secondary ds:UpToDate/UpToDate C r----- ns:28672 nr:0 dw:0 dr:28672 al:0 bm:5 lo:0 pe:0 ua:0 ap:0 ep:1 wo:n oos:0 crm(live)# configure crm(live)configure# show node ha-master node ha-slave primitive drbd_postgresql ocf:heartbeat:drbd \ params drbd_resource="postgresql" primitive fs_postgresql ocf:heartbeat:Filesystem \ params device="/dev/drbd/by-res/postgresql" directory="/mnt" fstype="ext4" primitive postgresqld lsb:postgresql primitive vip_cluster ocf:heartbeat:IPaddr2 \ params ip="172.70.65.200" nic="eth0:1" group postgresql fs_postgresql vip_cluster postgresqld \ meta target-role="Started" ms ms_drbd_postgresql drbd_postgresql \ meta master-max="1" master-node-max="1" clone-max="2" clone-node-max="1" notify="true" colocation postgresql_on_drbd inf: postgresql ms_drbd_postgresql:Master order postgresql_after_drbd inf: ms_drbd_postgresql:promote postgresql:start property $id="cib-bootstrap-options" \ dc-version="1.1.6-9971ebba4494012a93c03b40a2c58ec0eb60f50c" \ cluster-infrastructure="openais" \ expected-quorum-votes="2" \ stonith-enabled="false" \ no-quorum-policy="ignore" rsc_defaults $id="rsc-options" \ resource-stickiness="100" crm(live)# resource crm(live)resource# list Master/Slave Set: ms_drbd_postgresql [drbd_postgresql] Stopped: [ drbd_postgresql:0 drbd_postgresql:1 ] Resource Group: postgresql fs_postgresql (ocf::heartbeat:Filesystem) Stopped vip_cluster (ocf::heartbeat:IPaddr2) Stopped postgresqld (lsb:postgresql) Stopped ============ Last updated: Fri Oct 11 14:22:50 2013 Last change: Fri Oct 11 14:11:06 2013 via cibadmin on ha-slave Stack: openais Current DC: ha-slave - partition with quorum Version: 1.1.6-9971ebba4494012a93c03b40a2c58ec0eb60f50c 2 Nodes configured, 2 expected votes 5 Resources configured. ============ Online: [ ha-slave ha-master ] Failed actions: drbd_postgresql:0_start_0 (node=ha-slave, call=14, rc=1, status=complete): unknown error drbd_postgresql:0_start_0 (node=ha-master, call=18, rc=1, status=complete): unknown error On Fri, Oct 11, 2013 at 1:23 PM, emmanuel segura <emi2f...@gmail.com> wrote: > try with this constrains > > colocation col_postgresql inf: postgresql_cluster ms_drbd_postgresql:Master > order or_postgresql inf: ms_drbd_postgresql:promote > postgresql_cluster:start > > > > 2013/10/11 Thomaz Luiz Santos <thomaz.san...@gmail.com> > >> Dear all! >> >> I'm trying to make a sample cluster, in virtual machine, and after >> migrate to a physical machine, however i have problems to configure the >> pacemaker ( crm ), to startup the resources and failover. >> >> I cant mount the device /dev/drbd0 in the primary node and start >> postgresql manually, but use in crm resource, dont can mount the device, >> and start de postgresql. >> >> >> >> DRBD Version: 8.3.11 (api:88) >> Corosync Cluster Engine, version '1.4.2' >> Pacemaker 1.1.6 >> >> >> >> **** that is my configure on crm **** >> >> crm(live)# configure >> crm(live)configure# show >> node ha-master >> node ha-slave >> primitive drbd_postgresql ocf:heartbeat:drbd \ >> params drbd_resource="postgresql" >> primitive fs_postgresql ocf:heartbeat:Filesystem \ >> params device="/dev/drbd0" directory="/mnt" fstype="ext4" \ >> meta target-role="Started" >> primitive postgresql lsb:postgresql \ >> meta target-role="Started" >> primitive vip_cluster ocf:heartbeat:IPaddr2 \ >> params ip="172.70.65.200" nic="eth0:1" \ >> meta target-role="Started" >> group postgresql_cluster fs_postgresql vip_cluster postgresql >> ms ms_drbd_postgresql drbd_postgresql \ >> meta master-max="1" master-node-max="1" clone-max="2" >> clone-node-max="1" notify="true" target-role="Stopped" >> colocation col_postgresql inf: postgresql_cluster ms_drbd_postgresql >> order or_postgresql inf: ms_drbd_postgresql postgresql_cluster >> property $id="cib-bootstrap-options" \ >> dc-version="1.1.6-9971ebba4494012a93c03b40a2c58ec0eb60f50c" \ >> cluster-infrastructure="openais" \ >> expected-quorum-votes="2" \ >> stonith-enabled="false" \ >> no-quorum-policy="ignore" >> rsc_defaults $id="rsc-options" \ >> resource-stickiness="100" >> >> >> **** that is my global_common on drbd **** >> >> global { >> usage-count yes; >> # minor-count dialog-refresh disable-ip-verification >> } >> >> common { >> protocol C; >> >> handlers { >> pri-on-incon-degr >> "/usr/lib/drbd/notify-pri-on-incon-degr.sh; /usr/lib/drbd/not >> >> ify-emergency-reboot.sh; echo b > /proc/sysrq-trigger ; reboot >> -f"; >> pri-lost-after-sb >> "/usr/lib/drbd/notify-pri-lost-after-sb.sh; /usr/lib/drbd/not >> >> ify-emergency-reboot.sh; echo b > /proc/sysrq-trigger ; reboot >> -f"; >> local-io-error "/usr/lib/drbd/notify-io-error.sh; >> /usr/lib/drbd/notify-emergenc >> y-shutdown.sh; echo >> o > /proc/sysrq-trigger ; halt -f"; >> fence-peer "/usr/lib/drbd/crm-fence-peer.sh"; >> after-resync-target "/usr/lib/drbd/crm-unfence-peer.sh"; >> # split-brain "/usr/lib/drbd/notify-split-brain.sh root"; >> # out-of-sync "/usr/lib/drbd/notify-out-of-sync.sh root"; >> # before-resync-target >> "/usr/lib/drbd/snapshot-resync-target-lvm.sh -p 15 -- -c >> >> 16k"; >> # after-resync-target >> /usr/lib/drbd/unsnapshot-resync-target-lvm.sh; >> } >> >> startup { >> # wfc-timeout 15; >> # degr-wfc-timeout 60; >> # outdated-wfc-timeout wait-after-sb >> } >> >> disk { >> # on-io-error fencing use-bmbv no-disk-barrier >> no-disk-flushes >> # no-disk-drain no-md-flushes max-bio-bvecs >> } >> >> net { >> # cram-hmac-alg sha1; >> # shared-secret "secret"; >> # sndbuf-size rcvbuf-size timeout connect-int ping-int >> ping-timeout max-buffers >> # max-epoch-size ko-count allow-two-primaries >> cram-hmac-alg shared-secret >> # after-sb-0pri after-sb-1pri after-sb-2pri >> data-integrity-alg no-tcp-cork >> } >> >> syncer { >> # rate 150M; >> # rate after al-extents use-rle cpu-mask verify-alg >> csums-alg >> } >> } >> >> >> **** that is my postgresql.res **** >> >> resource postgresql { >> startup { >> wfc-timeout 15; >> degr-wfc-timeout 60; >> } >> >> syncer { >> rate 150M; >> verify-alg md5; >> } >> >> disk { >> on-io-error detach; >> no-disk-barrier; >> no-disk-flushes; >> no-disk-drain; >> fencing resource-only; >> } >> >> on ha-master { >> device /dev/drbd0; >> disk /dev/sdb1; >> address 172.70.65.210:7788; >> meta-disk internal; >> } >> >> on ha-slave { >> device /dev/drbd0; >> disk /dev/sdb1; >> address 172.70.65.220:7788; >> meta-disk internal; >> } >> >> >> } >> >> >> **** that is my corosync.conf **** >> >> >> compatibility: whitetank >> >> totem { >> version: 2 >> secauth: off >> threads: 0 >> interface { >> ringnumber: 0 >> bindnetaddr: 172.70.65.200 >> mcastaddr: 226.94.1.1 >> mcastport: 5405 >> ttl: 1 >> } >> } >> >> logging { >> fileline: off >> to_stderr: yes >> to_logfile: yes >> to_syslog: yes >> logfile: /var/log/cluster/corosync.log >> debug: on >> timestamp: on >> logger_subsys { >> subsys: AMF >> debug: off >> } >> } >> >> amf { >> mode: disabled >> } >> >> aisexec{ >> user : root >> group : root >> } >> >> service{ >> # Load the Pacemaker Cluster Resource Manager >> name : pacemaker >> ver : 0 >> } >> >> >> >> DRBD, postgresql, manually start : >> >> >> version: 8.3.13 (api:88/proto:86-96) >> srcversion: 697DE8B1973B1D8914F04DB >> 0: cs:Connected ro:Primary/Secondary ds:UpToDate/UpToDate C r----- >> ns:0 nr:0 dw:0 dr:664 al:0 bm:0 lo:0 pe:0 ua:0 ap:0 ep:1 wo:n oos:0 >> >> >> version: 8.3.13 (api:88/proto:86-96) >> srcversion: 697DE8B1973B1D8914F04DB >> 0: cs:Connected ro:Secondary/Primary ds:UpToDate/UpToDate C r----- >> ns:0 nr:0 dw:0 dr:0 al:0 bm:0 lo:0 pe:0 ua:0 ap:0 ep:1 wo:n oos:0 >> >> >> >> root@ha-master:/mnt# df -hT >> Sist. Arq. Tipo Tam. Usado Disp. Uso% Montado em >> /dev/sda1 ext4 4,0G 1,8G 2,1G 47% / >> udev devtmpfs 473M 4,0K 473M 1% /dev >> tmpfs tmpfs 193M 264K 193M 1% /run >> none tmpfs 5,0M 4,0K 5,0M 1% /run/lock >> none tmpfs 482M 17M 466M 4% /run/shm >> /dev/drbd0 ext4 2,0G 69M 1,9G 4% /mnt >> >> >> root@ha-master:/mnt# service postgresql status >> Running clusters: 9.1/main >> >> >> >> >> -- >> ------------------------------ >> Thomaz Luiz Santos >> Linux User: #359356 >> >> >> _______________________________________________ >> Pacemaker mailing list: Pacemaker@oss.clusterlabs.org >> http://oss.clusterlabs.org/mailman/listinfo/pacemaker >> >> Project Home: http://www.clusterlabs.org >> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf >> Bugs: http://bugs.clusterlabs.org >> >> > > > -- > esta es mi vida e me la vivo hasta que dios quiera > > _______________________________________________ > Pacemaker mailing list: Pacemaker@oss.clusterlabs.org > http://oss.clusterlabs.org/mailman/listinfo/pacemaker > > Project Home: http://www.clusterlabs.org > Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf > Bugs: http://bugs.clusterlabs.org > > -- ------------------------------ Thomaz Luiz Santos Linux User: #359356 http://thomaz.santos.googlepages.com/
_______________________________________________ Pacemaker mailing list: Pacemaker@oss.clusterlabs.org http://oss.clusterlabs.org/mailman/listinfo/pacemaker Project Home: http://www.clusterlabs.org Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf Bugs: http://bugs.clusterlabs.org