I remove the postgres start in the boot time, and work fine. now i make tests of failover, i reboot the master node and the slave node not assume the services in the log i have it
Oct 16 09:20:23 ha-slave drbd[7125]: ERROR: postgresql: Exit code 11 Oct 16 09:20:23 ha-slave drbd[7125]: ERROR: postgresql: Command output: Oct 16 09:20:24 ha-slave crmd: [1033]: ERROR: process_lrm_event: LRM operation drbd_postgresql:1_promote_0 (89) Timed Out (timeout=20000ms) Oct 16 09:20:29 ha-slave drbd[7585]: ERROR: postgresql: Called drbdadm -c /etc/drbd.conf primary postgresql I belive the slave node, not assume the primary on drbd On Tue, Oct 15, 2013 at 7:12 PM, emmanuel segura <emi2f...@gmail.com> wrote: > Check if your postgres is stopped or not start in the boot time. > > > 2013/10/15 Thomaz Luiz Santos <thomaz.san...@gmail.com> > >> dear all :-D >> >> I remake my crm config >> >> node ha-master >> node ha-slave >> primitive drbd_postgresql ocf:linbit:drbd \ >> params drbd_resource="postgresql" \ >> op monitor interval="30" role="Master" \ >> op monitor interval="33" role="Slave" >> primitive fs_postgresql ocf:heartbeat:Filesystem \ >> params device="/dev/drbd0" directory="/mnt" fstype="ext4" \ >> meta target-role="Started" >> primitive postgresqld lsb:postgresql \ >> meta target-role="Started" >> primitive vip_cluster ocf:heartbeat:IPaddr2 \ >> params ip="172.70.65.200" nic="eth0:1" \ >> meta target-role="Started" >> group postgresql fs_postgresql vip_cluster postgresqld \ >> meta is-managed="true" >> ms ms_drbd_postgresql drbd_postgresql \ >> meta master-max="1" master-node-max="1" clone-max="2" >> clone-node-max="1" notify="true" target-role="Started" is-managed="true" >> location cli-prefer-postgresql postgresql \ >> rule $id="cli-prefer-rule-postgresql" inf: #uname eq ha-master >> colocation postgresql_on_drbd inf: postgresql ms_drbd_postgresql:Master >> order postgresql_after_drbd inf: ms_drbd_postgresql:promote >> postgresql:start >> >> >> and fix some problems but have one error new rsrsrs. >> >> how fix it, its is the main problem ?? >> >> on ha-master syslog: >> pengine: [1100]: ERROR: native_create_actions: Resource postgresqld >> (lsb::postgresql) is active on 2 nodes attempting recovery >> >> >> >> On Fri, Oct 11, 2013 at 2:23 PM, Thomaz Luiz Santos < >> thomaz.san...@gmail.com> wrote: >> >>> I remake my CRM config and reboot the virtual machines, and not have >>> successful. >>> the DRBD not start the primary, and not mount the /dev/drbd0 :-( >>> >>> >>> ha-slave: >>> >>> version: 8.3.13 (api:88/proto:86-96) >>> srcversion: 697DE8B1973B1D8914F04DB >>> 0: cs:Connected ro:Secondary/Secondary ds:UpToDate/UpToDate C r----- >>> ns:0 nr:28672 dw:28672 dr:0 al:0 bm:5 lo:0 pe:0 ua:0 ap:0 ep:1 wo:n >>> oos:0 >>> >>> >>> ha-master: >>> version: 8.3.13 (api:88/proto:86-96) >>> srcversion: 697DE8B1973B1D8914F04DB >>> 0: cs:Connected ro:Secondary/Secondary ds:UpToDate/UpToDate C r----- >>> ns:28672 nr:0 dw:0 dr:28672 al:0 bm:5 lo:0 pe:0 ua:0 ap:0 ep:1 wo:n >>> oos:0 >>> >>> >>> >>> >>> >>> crm(live)# configure >>> crm(live)configure# show >>> node ha-master >>> node ha-slave >>> primitive drbd_postgresql ocf:heartbeat:drbd \ >>> params drbd_resource="postgresql" >>> primitive fs_postgresql ocf:heartbeat:Filesystem \ >>> params device="/dev/drbd/by-res/postgresql" directory="/mnt" >>> fstype="ext4" >>> primitive postgresqld lsb:postgresql >>> primitive vip_cluster ocf:heartbeat:IPaddr2 \ >>> params ip="172.70.65.200" nic="eth0:1" >>> group postgresql fs_postgresql vip_cluster postgresqld \ >>> meta target-role="Started" >>> ms ms_drbd_postgresql drbd_postgresql \ >>> meta master-max="1" master-node-max="1" clone-max="2" >>> clone-node-max="1" notify="true" >>> colocation postgresql_on_drbd inf: postgresql ms_drbd_postgresql:Master >>> order postgresql_after_drbd inf: ms_drbd_postgresql:promote >>> postgresql:start >>> property $id="cib-bootstrap-options" \ >>> dc-version="1.1.6-9971ebba4494012a93c03b40a2c58ec0eb60f50c" \ >>> cluster-infrastructure="openais" \ >>> expected-quorum-votes="2" \ >>> stonith-enabled="false" \ >>> no-quorum-policy="ignore" >>> rsc_defaults $id="rsc-options" \ >>> resource-stickiness="100" >>> >>> >>> >>> crm(live)# resource >>> crm(live)resource# list >>> Master/Slave Set: ms_drbd_postgresql [drbd_postgresql] >>> Stopped: [ drbd_postgresql:0 drbd_postgresql:1 ] >>> Resource Group: postgresql >>> fs_postgresql (ocf::heartbeat:Filesystem) Stopped >>> vip_cluster (ocf::heartbeat:IPaddr2) Stopped >>> postgresqld (lsb:postgresql) Stopped >>> >>> >>> >>> >>> ============ >>> Last updated: Fri Oct 11 14:22:50 2013 >>> Last change: Fri Oct 11 14:11:06 2013 via cibadmin on ha-slave >>> Stack: openais >>> Current DC: ha-slave - partition with quorum >>> Version: 1.1.6-9971ebba4494012a93c03b40a2c58ec0eb60f50c >>> 2 Nodes configured, 2 expected votes >>> 5 Resources configured. >>> ============ >>> >>> Online: [ ha-slave ha-master ] >>> >>> >>> Failed actions: >>> drbd_postgresql:0_start_0 (node=ha-slave, call=14, rc=1, >>> status=complete): unknown error >>> drbd_postgresql:0_start_0 (node=ha-master, call=18, rc=1, >>> status=complete): unknown error >>> >>> >>> >>> >>> On Fri, Oct 11, 2013 at 1:23 PM, emmanuel segura <emi2f...@gmail.com>wrote: >>> >>>> try with this constrains >>>> >>>> colocation col_postgresql inf: postgresql_cluster >>>> ms_drbd_postgresql:Master >>>> order or_postgresql inf: ms_drbd_postgresql:promote >>>> postgresql_cluster:start >>>> >>>> >>>> >>>> 2013/10/11 Thomaz Luiz Santos <thomaz.san...@gmail.com> >>>> >>>>> Dear all! >>>>> >>>>> I'm trying to make a sample cluster, in virtual machine, and after >>>>> migrate to a physical machine, however i have problems to configure the >>>>> pacemaker ( crm ), to startup the resources and failover. >>>>> >>>>> I cant mount the device /dev/drbd0 in the primary node and start >>>>> postgresql manually, but use in crm resource, dont can mount the device, >>>>> and start de postgresql. >>>>> >>>>> >>>>> >>>>> DRBD Version: 8.3.11 (api:88) >>>>> Corosync Cluster Engine, version '1.4.2' >>>>> Pacemaker 1.1.6 >>>>> >>>>> >>>>> >>>>> **** that is my configure on crm **** >>>>> >>>>> crm(live)# configure >>>>> crm(live)configure# show >>>>> node ha-master >>>>> node ha-slave >>>>> primitive drbd_postgresql ocf:heartbeat:drbd \ >>>>> params drbd_resource="postgresql" >>>>> primitive fs_postgresql ocf:heartbeat:Filesystem \ >>>>> params device="/dev/drbd0" directory="/mnt" fstype="ext4" \ >>>>> meta target-role="Started" >>>>> primitive postgresql lsb:postgresql \ >>>>> meta target-role="Started" >>>>> primitive vip_cluster ocf:heartbeat:IPaddr2 \ >>>>> params ip="172.70.65.200" nic="eth0:1" \ >>>>> meta target-role="Started" >>>>> group postgresql_cluster fs_postgresql vip_cluster postgresql >>>>> ms ms_drbd_postgresql drbd_postgresql \ >>>>> meta master-max="1" master-node-max="1" clone-max="2" >>>>> clone-node-max="1" notify="true" target-role="Stopped" >>>>> colocation col_postgresql inf: postgresql_cluster ms_drbd_postgresql >>>>> order or_postgresql inf: ms_drbd_postgresql postgresql_cluster >>>>> property $id="cib-bootstrap-options" \ >>>>> dc-version="1.1.6-9971ebba4494012a93c03b40a2c58ec0eb60f50c" \ >>>>> cluster-infrastructure="openais" \ >>>>> expected-quorum-votes="2" \ >>>>> stonith-enabled="false" \ >>>>> no-quorum-policy="ignore" >>>>> rsc_defaults $id="rsc-options" \ >>>>> resource-stickiness="100" >>>>> >>>>> >>>>> **** that is my global_common on drbd **** >>>>> >>>>> global { >>>>> usage-count yes; >>>>> # minor-count dialog-refresh disable-ip-verification >>>>> } >>>>> >>>>> common { >>>>> protocol C; >>>>> >>>>> handlers { >>>>> pri-on-incon-degr >>>>> "/usr/lib/drbd/notify-pri-on-incon-degr.sh; /usr/lib/drbd/not >>>>> >>>>> ify-emergency-reboot.sh; echo b > /proc/sysrq-trigger ; reboot >>>>> -f"; >>>>> pri-lost-after-sb >>>>> "/usr/lib/drbd/notify-pri-lost-after-sb.sh; /usr/lib/drbd/not >>>>> >>>>> ify-emergency-reboot.sh; echo b > /proc/sysrq-trigger ; reboot >>>>> -f"; >>>>> local-io-error "/usr/lib/drbd/notify-io-error.sh; >>>>> /usr/lib/drbd/notify-emergenc >>>>> y-shutdown.sh; echo >>>>> o > /proc/sysrq-trigger ; halt -f"; >>>>> fence-peer "/usr/lib/drbd/crm-fence-peer.sh"; >>>>> after-resync-target >>>>> "/usr/lib/drbd/crm-unfence-peer.sh"; >>>>> # split-brain "/usr/lib/drbd/notify-split-brain.sh >>>>> root"; >>>>> # out-of-sync "/usr/lib/drbd/notify-out-of-sync.sh >>>>> root"; >>>>> # before-resync-target >>>>> "/usr/lib/drbd/snapshot-resync-target-lvm.sh -p 15 -- -c >>>>> >>>>> 16k"; >>>>> # after-resync-target >>>>> /usr/lib/drbd/unsnapshot-resync-target-lvm.sh; >>>>> } >>>>> >>>>> startup { >>>>> # wfc-timeout 15; >>>>> # degr-wfc-timeout 60; >>>>> # outdated-wfc-timeout wait-after-sb >>>>> } >>>>> >>>>> disk { >>>>> # on-io-error fencing use-bmbv no-disk-barrier >>>>> no-disk-flushes >>>>> # no-disk-drain no-md-flushes max-bio-bvecs >>>>> } >>>>> >>>>> net { >>>>> # cram-hmac-alg sha1; >>>>> # shared-secret "secret"; >>>>> # sndbuf-size rcvbuf-size timeout connect-int ping-int >>>>> ping-timeout max-buffers >>>>> # max-epoch-size ko-count allow-two-primaries >>>>> cram-hmac-alg shared-secret >>>>> # after-sb-0pri after-sb-1pri after-sb-2pri >>>>> data-integrity-alg no-tcp-cork >>>>> } >>>>> >>>>> syncer { >>>>> # rate 150M; >>>>> # rate after al-extents use-rle cpu-mask verify-alg >>>>> csums-alg >>>>> } >>>>> } >>>>> >>>>> >>>>> **** that is my postgresql.res **** >>>>> >>>>> resource postgresql { >>>>> startup { >>>>> wfc-timeout 15; >>>>> degr-wfc-timeout 60; >>>>> } >>>>> >>>>> syncer { >>>>> rate 150M; >>>>> verify-alg md5; >>>>> } >>>>> >>>>> disk { >>>>> on-io-error detach; >>>>> no-disk-barrier; >>>>> no-disk-flushes; >>>>> no-disk-drain; >>>>> fencing resource-only; >>>>> } >>>>> >>>>> on ha-master { >>>>> device /dev/drbd0; >>>>> disk /dev/sdb1; >>>>> address 172.70.65.210:7788; >>>>> meta-disk internal; >>>>> } >>>>> >>>>> on ha-slave { >>>>> device /dev/drbd0; >>>>> disk /dev/sdb1; >>>>> address 172.70.65.220:7788; >>>>> meta-disk internal; >>>>> } >>>>> >>>>> >>>>> } >>>>> >>>>> >>>>> **** that is my corosync.conf **** >>>>> >>>>> >>>>> compatibility: whitetank >>>>> >>>>> totem { >>>>> version: 2 >>>>> secauth: off >>>>> threads: 0 >>>>> interface { >>>>> ringnumber: 0 >>>>> bindnetaddr: 172.70.65.200 >>>>> mcastaddr: 226.94.1.1 >>>>> mcastport: 5405 >>>>> ttl: 1 >>>>> } >>>>> } >>>>> >>>>> logging { >>>>> fileline: off >>>>> to_stderr: yes >>>>> to_logfile: yes >>>>> to_syslog: yes >>>>> logfile: /var/log/cluster/corosync.log >>>>> debug: on >>>>> timestamp: on >>>>> logger_subsys { >>>>> subsys: AMF >>>>> debug: off >>>>> } >>>>> } >>>>> >>>>> amf { >>>>> mode: disabled >>>>> } >>>>> >>>>> aisexec{ >>>>> user : root >>>>> group : root >>>>> } >>>>> >>>>> service{ >>>>> # Load the Pacemaker Cluster Resource Manager >>>>> name : pacemaker >>>>> ver : 0 >>>>> } >>>>> >>>>> >>>>> >>>>> DRBD, postgresql, manually start : >>>>> >>>>> >>>>> version: 8.3.13 (api:88/proto:86-96) >>>>> srcversion: 697DE8B1973B1D8914F04DB >>>>> 0: cs:Connected ro:Primary/Secondary ds:UpToDate/UpToDate C r----- >>>>> ns:0 nr:0 dw:0 dr:664 al:0 bm:0 lo:0 pe:0 ua:0 ap:0 ep:1 wo:n oos:0 >>>>> >>>>> >>>>> version: 8.3.13 (api:88/proto:86-96) >>>>> srcversion: 697DE8B1973B1D8914F04DB >>>>> 0: cs:Connected ro:Secondary/Primary ds:UpToDate/UpToDate C r----- >>>>> ns:0 nr:0 dw:0 dr:0 al:0 bm:0 lo:0 pe:0 ua:0 ap:0 ep:1 wo:n oos:0 >>>>> >>>>> >>>>> >>>>> root@ha-master:/mnt# df -hT >>>>> Sist. Arq. Tipo Tam. Usado Disp. Uso% Montado em >>>>> /dev/sda1 ext4 4,0G 1,8G 2,1G 47% / >>>>> udev devtmpfs 473M 4,0K 473M 1% /dev >>>>> tmpfs tmpfs 193M 264K 193M 1% /run >>>>> none tmpfs 5,0M 4,0K 5,0M 1% /run/lock >>>>> none tmpfs 482M 17M 466M 4% /run/shm >>>>> /dev/drbd0 ext4 2,0G 69M 1,9G 4% /mnt >>>>> >>>>> >>>>> root@ha-master:/mnt# service postgresql status >>>>> Running clusters: 9.1/main >>>>> >>>>> >>>>> >>>>> >>>>> -- >>>>> ------------------------------ >>>>> Thomaz Luiz Santos >>>>> Linux User: #359356 >>>>> >>>>> >>>>> _______________________________________________ >>>>> Pacemaker mailing list: Pacemaker@oss.clusterlabs.org >>>>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker >>>>> >>>>> Project Home: http://www.clusterlabs.org >>>>> Getting started: >>>>> http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf >>>>> Bugs: http://bugs.clusterlabs.org >>>>> >>>>> >>>> >>>> >>>> -- >>>> esta es mi vida e me la vivo hasta que dios quiera >>>> >>>> _______________________________________________ >>>> Pacemaker mailing list: Pacemaker@oss.clusterlabs.org >>>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker >>>> >>>> Project Home: http://www.clusterlabs.org >>>> Getting started: >>>> http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf >>>> Bugs: http://bugs.clusterlabs.org >>>> >>>> >>> >>> >>> -- >>> ------------------------------ >>> Thomaz Luiz Santos >>> Linux User: #359356 >>> http://thomaz.santos.googlepages.com/ >>> >> >> >> >> -- >> ------------------------------ >> Thomaz Luiz Santos >> Linux User: #359356 >> http://thomaz.santos.googlepages.com/ >> >> _______________________________________________ >> Pacemaker mailing list: Pacemaker@oss.clusterlabs.org >> http://oss.clusterlabs.org/mailman/listinfo/pacemaker >> >> Project Home: http://www.clusterlabs.org >> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf >> Bugs: http://bugs.clusterlabs.org >> >> > > > -- > esta es mi vida e me la vivo hasta que dios quiera > > _______________________________________________ > Pacemaker mailing list: Pacemaker@oss.clusterlabs.org > http://oss.clusterlabs.org/mailman/listinfo/pacemaker > > Project Home: http://www.clusterlabs.org > Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf > Bugs: http://bugs.clusterlabs.org > > -- ------------------------------ Thomaz Luiz Santos Linux User: #359356 http://thomaz.santos.googlepages.com/
_______________________________________________ Pacemaker mailing list: Pacemaker@oss.clusterlabs.org http://oss.clusterlabs.org/mailman/listinfo/pacemaker Project Home: http://www.clusterlabs.org Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf Bugs: http://bugs.clusterlabs.org