Hello.

I have got STOP cluster status when add\del new cluster node <pi05> after run <update pgsql> command:

How to add a node without STOP cluster?

I am doing command step's:

# pcs cluster auth pi01 pi02 pi03 pi05 -u hacluster -p hacluster

pi01: Authorized
pi02: Authorized
pi03: Authorized
pi05: Authorized

# pcs cluster node add pi05 --start

pi01: Corosync updated
pi02: Corosync updated
pi03: Corosync updated
pi05: Succeeded
pi05: Starting Cluster...

# pcs resource show --full

 Group: master-group
  Resource: vip-master (class=ocf provider=heartbeat type=IPaddr2)
   Attributes: ip=192.168.242.100 nic=eth0 cidr_netmask=24
Operations: start interval=0s timeout=60s on-fail=restart (vip-master-start-interval-0s) monitor interval=10s timeout=60s on-fail=restart (vip-master-monitor-interval-10s) stop interval=0s timeout=60s on-fail=block (vip-master-stop-interval-0s)
  Resource: vip-rep (class=ocf provider=heartbeat type=IPaddr2)
   Attributes: ip=192.168.242.101 nic=eth0 cidr_netmask=24
   Meta Attrs: migration-threshold=0
Operations: start interval=0s timeout=60s on-fail=stop (vip-rep-start-interval-0s) monitor interval=10s timeout=60s on-fail=restart (vip-rep-monitor-interval-10s) stop interval=0s timeout=60s on-fail=ignore (vip-rep-stop-interval-0s)
 Master: msPostgresql
Meta Attrs: master-max=1 master-node-max=1 clone-max=3 clone-node-max=1 notify=true
  Resource: pgsql (class=ocf provider=heartbeat type=pgsql)
Attributes: pgctl=/usr/pgsql-9.5/bin/pg_ctl psql=/usr/pgsql-9.5/bin/psql pgdata=/var/lib/pgsql/9.5/data/ rep_mode=sync node_list="pi01 pi02 pi03" restore_command="cp /var/lib/pgsql/9.5/data/wal_archive/%f %p" primary_conninfo_opt="user=repl password=super-pass-for-repl keepalives_idle=60 keepalives_interval=5 keepalives_count=5" master_ip=192.168.242.100 restart_on_promote=true check_wal_receiver=true Operations: start interval=0s timeout=60s on-fail=restart (pgsql-start-interval-0s) monitor interval=4s timeout=60s on-fail=restart (pgsql-monitor-interval-4s) monitor role=Master timeout=60s on-fail=restart interval=3s (pgsql-monitor-interval-3s-role-Master) promote interval=0s timeout=60s on-fail=restart (pgsql-promote-interval-0s) demote interval=0s timeout=60s on-fail=stop (pgsql-demote-interval-0s) stop interval=0s timeout=60s on-fail=block (pgsql-stop-interval-0s)
               notify interval=0s timeout=60s (pgsql-notify-interval-0s)


# pcs resource update msPostgresql pgsql master-max=1 master-node-max=1 clone-max=4 clone-node-max=1 notify=true

# pcs resource update pgsql pgsql node_list="pi01 pi02 pi03 pi05"

# crm_mon -Afr1

Last updated: Fri Oct 2 17:07:05 2015 Last change: Fri Oct 2 17:06:37 2015
 by root via cibadmin on pi01
Stack: corosync
Current DC: pi02 (version 1.1.13-a14efad) - partition with quorum
4 nodes and 9 resources configured

Online: [ pi01 pi02 pi03 pi05 ]

Full list of resources:

 Resource Group: master-group
     vip-master (ocf::heartbeat:IPaddr2):       Stopped
     vip-rep    (ocf::heartbeat:IPaddr2):       Stopped
 Master/Slave Set: msPostgresql [pgsql]
     Slaves: [ pi02 ]
     Stopped: [ pi01 pi03 pi05 ]
 fence-pi01     (stonith:fence_ssh):    Started pi02
 fence-pi02     (stonith:fence_ssh):    Started pi01
 fence-pi03     (stonith:fence_ssh):    Started pi01

Node Attributes:
* Node pi01:
    + master-pgsql                      : -INFINITY
    + pgsql-data-status                 : STREAMING|SYNC
    + pgsql-status                      : STOP
* Node pi02:
    + master-pgsql                      : -INFINITY
    + pgsql-data-status                 : LATEST
    + pgsql-status                      : STOP
* Node pi03:
    + master-pgsql                      : -INFINITY
    + pgsql-data-status                 : STREAMING|POTENTIAL
    + pgsql-status                      : STOP
* Node pi05:
    + master-pgsql                      : -INFINITY
    + pgsql-status                      : STOP

Migration Summary:
* Node pi01:
* Node pi03:
* Node pi02:
* Node pi05:

After some time is worked:

Every 2.0s: crm_mon -Afr1 Fri Oct 2 17:04:36 2015

Last updated: Fri Oct 2 17:04:36 2015 Last change: Fri Oct 2 17:04:07 2015 by root via
 cibadmin on pi01
Stack: corosync
Current DC: pi02 (version 1.1.13-a14efad) - partition with quorum
4 nodes and 9 resources configured

Online: [ pi01 pi02 pi03 pi05 ]

Full list of resources:

 Resource Group: master-group
     vip-master (ocf::heartbeat:IPaddr2):       Started pi02
     vip-rep    (ocf::heartbeat:IPaddr2):       Started pi02
 Master/Slave Set: msPostgresql [pgsql]
     Masters: [ pi02 ]
     Slaves: [ pi01 pi03 pi05 ]

 fence-pi01     (stonith:fence_ssh):    Started pi02
 fence-pi02     (stonith:fence_ssh):    Started pi01
 fence-pi03     (stonith:fence_ssh):    Started pi01

Node Attributes:
* Node pi01:
    + master-pgsql                      : 100
    + pgsql-data-status                 : STREAMING|SYNC
    + pgsql-receiver-status             : normal
    + pgsql-status                      : HS:sync
* Node pi02:
    + master-pgsql                      : 1000
    + pgsql-data-status                 : LATEST
    + pgsql-master-baseline             : 0000000008000098
    + pgsql-receiver-status             : ERROR
    + pgsql-status                      : PRI
* Node pi03:
    + master-pgsql                      : -INFINITY
    + pgsql-data-status                 : STREAMING|POTENTIAL
    + pgsql-receiver-status             : normal
    + pgsql-status                      : HS:potential
* Node pi05:
    + master-pgsql      : -INFINITY
    + pgsql-data-status                      : STREAMING|POTENTIAL
    + pgsql-receiver-status                  : normal
    + pgsql-status                           : HS:potential

Migration Summary:
* Node pi01:
* Node pi03:
* Node pi02:
* Node pi05:


--
Nikolay Popov

_______________________________________________
Users mailing list: Users@clusterlabs.org
http://clusterlabs.org/mailman/listinfo/users

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org

Reply via email to