On 12/10/2015 12:45 PM, Louis Munro wrote:
> Hello all,
> 
> I am trying to get a Corosync 2 cluster going on CentOS 6.7 but I am running 
> in a bit of a problem with either Corosync or Pacemaker.
> crm reports that all my nodes are offline and the stack is unknown (I am not 
> sure if that is relevant).
> 
> I believe both nodes are actually present and seen in corosync, but they may 
> not be considered as such by pacemaker.
> I have messages in the logs saying that the processes cannot get the node 
> name and default to uname -n: 
> 
> Dec 10 13:38:53 [2236] hack1.example.com       crmd:     info: 
> corosync_node_name:    Unable to get node name for nodeid 739513528
> Dec 10 13:38:53 [2236] hack1.example.com       crmd:   notice: get_node_name: 
> Defaulting to uname -n for the local corosync node name
> Dec 10 13:38:53 [2236] hack1.example.com       crmd:     info: crm_get_peer:  
> Node 739513528 is now known as hack1.example.com
> 
> The uname -n is correct as far that is concerned.
> 
> 
> Does this mean anything to anyone here? 
> 
> 
> [Lots of details to follow]...
> 
> I compiled my own versions of Corosync, Pacemaker, crm and the 
> resource-agents seemingly without problems.
> 
> Here is what I currently have installed:
> 
> # corosync -v
> Corosync Cluster Engine, version '2.3.5'
> Copyright (c) 2006-2009 Red Hat, Inc.
> 
> # pacemakerd -F
> Pacemaker 1.1.13 (Build: 5b41ae1)
>  Supporting v3.0.10:  generated-manpages agent-manpages ascii-docs ncurses 
> libqb-logging libqb-ipc lha-fencing upstart nagios  corosync-native 
> atomic-attrd libesmtp acls
> 
> # crm --version
> crm 2.2.0-rc3
> 
> 
> 
> Here is the output of crm status:
> 
> # crm status
> Last updated: Thu Dec 10 12:47:50 2015                Last change: Thu Dec 10 
> 12:02:33 2015 by root via cibadmin on hack1.example.com
> Stack: unknown
> Current DC: NONE
> 2 nodes and 0 resources configured
> 
> OFFLINE: [ hack1.example.com hack2.example.com ]
> 
> Full list of resources:
> 
> {nothing to see here}
> 
> 
> 
> # corosync-cmapctl | grep members
> runtime.totem.pg.mrp.srp.members.739513528.config_version (u64) = 0
> runtime.totem.pg.mrp.srp.members.739513528.ip (str) = r(0) ip(172.20.20.184)
> runtime.totem.pg.mrp.srp.members.739513528.join_count (u32) = 1
> runtime.totem.pg.mrp.srp.members.739513528.status (str) = joined
> runtime.totem.pg.mrp.srp.members.739513590.config_version (u64) = 0
> runtime.totem.pg.mrp.srp.members.739513590.ip (str) = r(0) ip(172.20.20.246)
> runtime.totem.pg.mrp.srp.members.739513590.join_count (u32) = 1
> runtime.totem.pg.mrp.srp.members.739513590.status (str) = joined
> 
> 
> # uname -n
> hack1.example.com
> 
> # corosync-cfgtool -s
> Printing ring status.
> Local node ID 739513528
> RING ID 0
>       id      = 172.20.20.184
>       status  = ring 0 active with no faults
> 
> 
> # uname -n
> hack2.example.com
> 
> 
> # corosync-cfgtool -s
> Printing ring status.
> Local node ID 739513590
> RING ID 0
>       id      = 172.20.20.246
>       status  = ring 0 active with no faults
> 
> 
> 
> 
> Shouldn’t I see both nodes in the same ring?

They are in the same ring, but the cfgtool will only print the local id.

> My corosync config is currently defined as:
> 
> # egrep -v '#' /etc/corosync/corosync.conf
> totem {
>       version: 2
> 
>       crypto_cipher: none
>       crypto_hash: none
>       clear_node_high_bit: yes
>       cluster_name: hack_cluster
>       interface {
>               ringnumber: 0
>               bindnetaddr: 172.20.0.0
>               mcastaddr: 239.255.1.1
>               mcastport: 5405
>               ttl: 1
>       }
> 
> }
> 
> logging {
>       fileline: on
>       to_stderr: no
>       to_logfile: yes
>       logfile: /var/log/cluster/corosync.log
>       to_syslog: yes
>       debug: off
>       timestamp: on
>       logger_subsys {
>               subsys: QUORUM
>               debug: off
>       }
> }
> 
> # cat /etc/corosync/service.d/pacemaker
> service {
>     name: pacemaker
>     ver: 1
> }

You don't want this section if you're using corosync 2. That's the old
"plugin" used with corosync 1.

> 
> 
> And here is my pacemaker configuration:
> 
> # crm config show xml
> <?xml version="1.0" ?>
> <cib num_updates="0" update-origin="hack1.example.com" 
> crm_feature_set="3.0.10" validate-with="pacemaker-2.4" 
> update-client="cibadmin" epoch="13" admin_epoch="0" update-user="root" 
> cib-last-written="Thu Dec 10 13:35:06 2015">
>   <configuration>
>     <crm_config>
>       <cluster_property_set id="cib-bootstrap-options">
>         <nvpair name="stonith-enabled" value="false" 
> id="cib-bootstrap-options-stonith-enabled"/>
>         <nvpair name="no-quorum-policy" value="ignore" 
> id="cib-bootstrap-options-no-quorum-policy"/>
>       </cluster_property_set>
>     </crm_config>
>     <nodes>
>       <node uname="hack1.example.com" id="hack1.example.com">
>         <instance_attributes id="hack1.example.com-instance_attributes">
>           <nvpair name="standby" value="off" 
> id="hack1.example.com-instance_attributes-standby"/>
>         </instance_attributes>
>       </node>
>       <node uname="hack2.example.com" id="hack2.example.com">
>         <instance_attributes id="hack2.example.com-instance_attributes">
>           <nvpair name="standby" value="off" 
> id="hack2.example.com-instance_attributes-standby"/>
>         </instance_attributes>
>       </node>
>     </nodes>
>     <resources/>
>     <constraints/>
>   </configuration>
> </cib>
> 
> 
> 
> 
> 
> 
> And finally some logs that might be relevant: 
> 
> Dec 10 13:38:50 [2227] hack1.example.com corosync notice  [MAIN  ] 
> main.c:1227 Corosync Cluster Engine ('2.3.5'): started and ready to provide 
> service.
> Dec 10 13:38:50 [2227] hack1.example.com corosync info    [MAIN  ] 
> main.c:1228 Corosync built-in features: pie relro bindnow
> Dec 10 13:38:50 [2227] hack1.example.com corosync notice  [TOTEM ] 
> totemnet.c:248 Initializing transport (UDP/IP Multicast).
> Dec 10 13:38:50 [2227] hack1.example.com corosync notice  [TOTEM ] 
> totemcrypto.c:579 Initializing transmit/receive security (NSS) crypto: none 
> hash: none
> Dec 10 13:38:50 [2227] hack1.example.com corosync notice  [TOTEM ] 
> totemudp.c:671 The network interface [172.20.20.184] is now up.
> Dec 10 13:38:50 [2227] hack1.example.com corosync notice  [SERV  ] 
> service.c:174 Service engine loaded: corosync configuration map access [0]
> Dec 10 13:38:50 [2227] hack1.example.com corosync info    [QB    ] 
> ipc_setup.c:377 server name: cmap
> Dec 10 13:38:50 [2227] hack1.example.com corosync notice  [SERV  ] 
> service.c:174 Service engine loaded: corosync configuration service [1]
> Dec 10 13:38:50 [2227] hack1.example.com corosync info    [QB    ] 
> ipc_setup.c:377 server name: cfg
> Dec 10 13:38:50 [2227] hack1.example.com corosync notice  [SERV  ] 
> service.c:174 Service engine loaded: corosync cluster closed process group 
> service v1.01 [2]
> Dec 10 13:38:50 [2227] hack1.example.com corosync info    [QB    ] 
> ipc_setup.c:377 server name: cpg
> Dec 10 13:38:50 [2227] hack1.example.com corosync notice  [SERV  ] 
> service.c:174 Service engine loaded: corosync profile loading service [4]
> Dec 10 13:38:50 [2227] hack1.example.com corosync notice  [SERV  ] 
> service.c:174 Service engine loaded: corosync cluster quorum service v0.1 [3]
> Dec 10 13:38:50 [2227] hack1.example.com corosync info    [QB    ] 
> ipc_setup.c:377 server name: quorum
> Dec 10 13:38:50 [2227] hack1.example.com corosync notice  [TOTEM ] 
> totemsrp.c:2095 A new membership (172.20.20.184:300) was formed. Members 
> joined: 739513528
> Dec 10 13:38:50 [2227] hack1.example.com corosync notice  [MAIN  ] main.c:305 
> Completed service synchronization, ready to provide service.
> Dec 10 13:38:50 [2227] hack1.example.com corosync notice  [TOTEM ] 
> totemsrp.c:2095 A new membership (172.20.20.184:304) was formed. Members 
> joined: 739513590
> Dec 10 13:38:50 [2227] hack1.example.com corosync notice  [MAIN  ] main.c:305 
> Completed service synchronization, ready to provide service.
> Dec 10 13:38:52 [2230] hack1.example.com pacemakerd:   notice: 
> mcp_read_config:       Configured corosync to accept connections from group 
> 500: OK (1)
> Dec 10 13:38:52 [2230] hack1.example.com pacemakerd:   notice: main:  
> Starting Pacemaker 1.1.13 (Build: 5b41ae1):  generated-manpages 
> agent-manpages ascii-docs ncurses libqb-logging libqb-ipc lha-fencing upstart 
> nagios  corosync-native atomic-attrd libesmtp acls
> Dec 10 13:38:52 [2230] hack1.example.com pacemakerd:     info: main:  Maximum 
> core file size is: 18446744073709551615
> Dec 10 13:38:52 [2230] hack1.example.com pacemakerd:     info: 
> qb_ipcs_us_publish:    server name: pacemakerd
> Dec 10 13:38:52 [2230] hack1.example.com pacemakerd:     info: 
> corosync_node_name:    Unable to get node name for nodeid 739513528
> Dec 10 13:38:52 [2230] hack1.example.com pacemakerd:   notice: get_node_name: 
> Could not obtain a node name for corosync nodeid 739513528
> Dec 10 13:38:52 [2230] hack1.example.com pacemakerd:     info: crm_get_peer:  
> Created entry 212e3751-e79f-4a72-927b-6e0176a9b35c/0x1b68e50 for node 
> (null)/739513528 (1 total)
> Dec 10 13:38:52 [2230] hack1.example.com pacemakerd:     info: crm_get_peer:  
> Node 739513528 has uuid 739513528
> Dec 10 13:38:52 [2230] hack1.example.com pacemakerd:     info: 
> crm_update_peer_proc:  cluster_connect_cpg: Node (null)[739513528] - 
> corosync-cpg is now online
> Dec 10 13:38:52 [2230] hack1.example.com pacemakerd:    error: 
> cluster_connect_quorum:        Corosync quorum is not configured
> Dec 10 13:38:52 [2230] hack1.example.com pacemakerd:     info: 
> corosync_node_name:    Unable to get node name for nodeid 739513528
> Dec 10 13:38:52 [2230] hack1.example.com pacemakerd:   notice: get_node_name: 
> Defaulting to uname -n for the local corosync node name
> Dec 10 13:38:52 [2230] hack1.example.com pacemakerd:     info: crm_get_peer:  
> Node 739513528 is now known as hack1.example.com
> Dec 10 13:38:52 [2230] hack1.example.com pacemakerd:     info: start_child:   
> Using uid=500 and group=500 for process cib
> Dec 10 13:38:52 [2230] hack1.example.com pacemakerd:     info: start_child:   
> Forked child 2231 for process cib
> Dec 10 13:38:52 [2230] hack1.example.com pacemakerd:     info: start_child:   
> Forked child 2232 for process stonith-ng
> Dec 10 13:38:52 [2230] hack1.example.com pacemakerd:     info: start_child:   
> Forked child 2233 for process lrmd
> Dec 10 13:38:52 [2230] hack1.example.com pacemakerd:     info: start_child:   
> Using uid=500 and group=500 for process attrd
> Dec 10 13:38:52 [2230] hack1.example.com pacemakerd:     info: start_child:   
> Forked child 2234 for process attrd
> Dec 10 13:38:52 [2230] hack1.example.com pacemakerd:     info: start_child:   
> Using uid=500 and group=500 for process pengine
> Dec 10 13:38:52 [2230] hack1.example.com pacemakerd:     info: start_child:   
> Forked child 2235 for process pengine
> Dec 10 13:38:52 [2230] hack1.example.com pacemakerd:     info: start_child:   
> Using uid=500 and group=500 for process crmd
> Dec 10 13:38:52 [2230] hack1.example.com pacemakerd:     info: start_child:   
> Forked child 2236 for process crmd
> Dec 10 13:38:52 [2230] hack1.example.com pacemakerd:     info: main:  
> Starting mainloop
> Dec 10 13:38:52 [2230] hack1.example.com pacemakerd:     info: 
> pcmk_cpg_membership:   Node 739513528 joined group pacemakerd (counter=0.0)
> Dec 10 13:38:52 [2230] hack1.example.com pacemakerd:     info: 
> pcmk_cpg_membership:   Node 739513528 still member of group pacemakerd 
> (peer=hack1.example.com, counter=0.0)
> Dec 10 13:38:52 [2231] hack1.example.com        cib:     info: crm_log_init:  
> Changed active directory to /var/lib/heartbeat/cores/hacluster
> Dec 10 13:38:52 [2231] hack1.example.com        cib:  warning: 
> crm_is_writable:       /var/lib/pacemaker/cib should be owned and r/w by 
> group haclient
> Dec 10 13:38:52 [2231] hack1.example.com        cib:     info: 
> get_cluster_type:      Verifying cluster type: 'corosync'
> Dec 10 13:38:52 [2231] hack1.example.com        cib:     info: 
> get_cluster_type:      Assuming an active 'corosync' cluster
> Dec 10 13:38:52 [2231] hack1.example.com        cib:     info: retrieveCib:   
> Reading cluster configuration file /var/lib/pacemaker/cib/cib.xml (digest: 
> /var/lib/pacemaker/cib/cib.xml.sig)
> Dec 10 13:38:52 [2231] hack1.example.com        cib:     info: 
> validate_with_relaxng: Creating RNG parser context
> Dec 10 13:38:52 [2234] hack1.example.com      attrd:     info: crm_log_init:  
> Changed active directory to /var/lib/heartbeat/cores/hacluster
> Dec 10 13:38:52 [2234] hack1.example.com      attrd:     info: main:  
> Starting up
> Dec 10 13:38:52 [2234] hack1.example.com      attrd:     info: 
> get_cluster_type:      Verifying cluster type: 'corosync'
> Dec 10 13:38:52 [2234] hack1.example.com      attrd:     info: 
> get_cluster_type:      Assuming an active 'corosync' cluster
> Dec 10 13:38:52 [2234] hack1.example.com      attrd:   notice: 
> crm_cluster_connect:   Connecting to cluster infrastructure: corosync
> Dec 10 13:38:52 [2236] hack1.example.com       crmd:     info: crm_log_init:  
> Changed active directory to /var/lib/heartbeat/cores/hacluster
> Dec 10 13:38:52 [2236] hack1.example.com       crmd:   notice: main:  CRM Git 
> Version: 1.1.13 (5b41ae1)
> Dec 10 13:38:52 [2232] hack1.example.com stonith-ng:     info: crm_log_init:  
> Changed active directory to /var/lib/heartbeat/cores/root
> Dec 10 13:38:52 [2232] hack1.example.com stonith-ng:     info: 
> get_cluster_type:      Verifying cluster type: 'corosync'
> Dec 10 13:38:52 [2232] hack1.example.com stonith-ng:     info: 
> get_cluster_type:      Assuming an active 'corosync' cluster
> Dec 10 13:38:52 [2232] hack1.example.com stonith-ng:   notice: 
> crm_cluster_connect:   Connecting to cluster infrastructure: corosync
> Dec 10 13:38:52 [2236] hack1.example.com       crmd:  warning: 
> crm_is_writable:       /var/lib/pacemaker/pengine should be owned and r/w by 
> group haclient
> Dec 10 13:38:52 [2236] hack1.example.com       crmd:  warning: 
> crm_is_writable:       /var/lib/pacemaker/cib should be owned and r/w by 
> group haclient

You definitely want to take care of these permissions issues.

> Dec 10 13:38:52 [2236] hack1.example.com       crmd:     info: do_log:        
> FSA: Input I_STARTUP from crmd_init() received in state S_STARTING
> Dec 10 13:38:52 [2236] hack1.example.com       crmd:     info: 
> get_cluster_type:      Verifying cluster type: 'corosync'
> Dec 10 13:38:52 [2236] hack1.example.com       crmd:     info: 
> get_cluster_type:      Assuming an active 'corosync' cluster
> Dec 10 13:38:52 [2231] hack1.example.com        cib:     info: startCib:      
> CIB Initialization completed successfully
> Dec 10 13:38:52 [2231] hack1.example.com        cib:   notice: 
> crm_cluster_connect:   Connecting to cluster infrastructure: corosync
> Dec 10 13:38:52 [2235] hack1.example.com    pengine:     info: crm_log_init:  
> Changed active directory to /var/lib/heartbeat/cores/hacluster
> Dec 10 13:38:52 [2235] hack1.example.com    pengine:  warning: 
> crm_is_writable:       /var/lib/pacemaker/pengine should be owned and r/w by 
> group haclient
> Dec 10 13:38:52 [2235] hack1.example.com    pengine:     info: 
> qb_ipcs_us_publish:    server name: pengine
> Dec 10 13:38:52 [2235] hack1.example.com    pengine:     info: main:  
> Starting pengine
> Dec 10 13:38:52 [2233] hack1.example.com       lrmd:     info: crm_log_init:  
> Changed active directory to /var/lib/heartbeat/cores/root
> Dec 10 13:38:52 [2233] hack1.example.com       lrmd:     info: 
> qb_ipcs_us_publish:    server name: lrmd
> Dec 10 13:38:52 [2233] hack1.example.com       lrmd:     info: main:  Starting
> Dec 10 13:38:52 [2230] hack1.example.com pacemakerd:     info: 
> corosync_node_name:    Unable to get node name for nodeid 739513590
> Dec 10 13:38:52 [2230] hack1.example.com pacemakerd:   notice: get_node_name: 
> Could not obtain a node name for corosync nodeid 739513590
> Dec 10 13:38:52 [2230] hack1.example.com pacemakerd:     info: crm_get_peer:  
> Created entry f34a4388-23fc-4f86-8932-43c814eb5ad7/0x1b6a920 for node 
> (null)/739513590 (2 total)
> Dec 10 13:38:52 [2230] hack1.example.com pacemakerd:     info: crm_get_peer:  
> Node 739513590 has uuid 739513590
> Dec 10 13:38:52 [2230] hack1.example.com pacemakerd:     info: 
> pcmk_cpg_membership:   Node 739513590 still member of group pacemakerd 
> (peer=(null), counter=0.1)
> Dec 10 13:38:52 [2230] hack1.example.com pacemakerd:     info: 
> crm_update_peer_proc:  pcmk_cpg_membership: Node (null)[739513590] - 
> corosync-cpg is now online
> Dec 10 13:38:52 [2230] hack1.example.com pacemakerd:     info: crm_get_peer:  
> Node 739513590 is now known as hack2.example.com
> Dec 10 13:38:52 [2230] hack1.example.com pacemakerd:     info: 
> mcp_cpg_deliver:       Ignoring process list sent by peer for local node
> Dec 10 13:38:52 [2230] hack1.example.com pacemakerd:     info: 
> mcp_cpg_deliver:       Ignoring process list sent by peer for local node
> Dec 10 13:38:52 [2230] hack1.example.com pacemakerd:     info: 
> mcp_cpg_deliver:       Ignoring process list sent by peer for local node
> Dec 10 13:38:52 [2230] hack1.example.com pacemakerd:     info: 
> mcp_cpg_deliver:       Ignoring process list sent by peer for local node
> Dec 10 13:38:52 [2230] hack1.example.com pacemakerd:     info: 
> mcp_cpg_deliver:       Ignoring process list sent by peer for local node
> Dec 10 13:38:52 [2230] hack1.example.com pacemakerd:     info: 
> mcp_cpg_deliver:       Ignoring process list sent by peer for local node
> Dec 10 13:38:52 [2230] hack1.example.com pacemakerd:     info: 
> mcp_cpg_deliver:       Ignoring process list sent by peer for local node
> Dec 10 13:38:53 [2234] hack1.example.com      attrd:     info: 
> corosync_node_name:    Unable to get node name for nodeid 739513528
> Dec 10 13:38:53 [2234] hack1.example.com      attrd:   notice: get_node_name: 
> Could not obtain a node name for corosync nodeid 739513528
> Dec 10 13:38:53 [2234] hack1.example.com      attrd:     info: crm_get_peer:  
> Created entry 5fbd08d1-fdbf-4f8c-85d0-bbbed1240969/0x180eb70 for node 
> (null)/739513528 (1 total)
> Dec 10 13:38:53 [2234] hack1.example.com      attrd:     info: crm_get_peer:  
> Node 739513528 has uuid 739513528
> Dec 10 13:38:53 [2234] hack1.example.com      attrd:     info: 
> crm_update_peer_proc:  cluster_connect_cpg: Node (null)[739513528] - 
> corosync-cpg is now online
> Dec 10 13:38:53 [2234] hack1.example.com      attrd:   notice: 
> crm_update_peer_state_iter:    crm_update_peer_proc: Node (null)[739513528] - 
> state is now member (was (null))
> Dec 10 13:38:53 [2234] hack1.example.com      attrd:     info: 
> init_cs_connection_once:       Connection to 'corosync': established
> Dec 10 13:38:53 [2232] hack1.example.com stonith-ng:     info: 
> corosync_node_name:    Unable to get node name for nodeid 739513528
> Dec 10 13:38:53 [2232] hack1.example.com stonith-ng:   notice: get_node_name: 
> Could not obtain a node name for corosync nodeid 739513528
> Dec 10 13:38:53 [2232] hack1.example.com stonith-ng:     info: crm_get_peer:  
> Created entry c2d2ccef-c87c-4b3c-8004-659f09113048/0x11fa8e0 for node 
> (null)/739513528 (1 total)
> Dec 10 13:38:53 [2232] hack1.example.com stonith-ng:     info: crm_get_peer:  
> Node 739513528 has uuid 739513528
> Dec 10 13:38:53 [2232] hack1.example.com stonith-ng:     info: 
> crm_update_peer_proc:  cluster_connect_cpg: Node (null)[739513528] - 
> corosync-cpg is now online
> Dec 10 13:38:53 [2232] hack1.example.com stonith-ng:   notice: 
> crm_update_peer_state_iter:    crm_update_peer_proc: Node (null)[739513528] - 
> state is now member (was (null))
> Dec 10 13:38:53 [2232] hack1.example.com stonith-ng:     info: 
> init_cs_connection_once:       Connection to 'corosync': established
> Dec 10 13:38:53 [2234] hack1.example.com      attrd:     info: 
> corosync_node_name:    Unable to get node name for nodeid 739513528
> Dec 10 13:38:53 [2234] hack1.example.com      attrd:   notice: get_node_name: 
> Defaulting to uname -n for the local corosync node name
> Dec 10 13:38:53 [2234] hack1.example.com      attrd:     info: crm_get_peer:  
> Node 739513528 is now known as hack1.example.com
> Dec 10 13:38:53 [2232] hack1.example.com stonith-ng:     info: 
> corosync_node_name:    Unable to get node name for nodeid 739513528
> Dec 10 13:38:53 [2232] hack1.example.com stonith-ng:   notice: get_node_name: 
> Defaulting to uname -n for the local corosync node name
> Dec 10 13:38:53 [2232] hack1.example.com stonith-ng:     info: crm_get_peer:  
> Node 739513528 is now known as hack1.example.com
> Dec 10 13:38:53 [2231] hack1.example.com        cib:     info: 
> corosync_node_name:    Unable to get node name for nodeid 739513528
> Dec 10 13:38:53 [2231] hack1.example.com        cib:   notice: get_node_name: 
> Could not obtain a node name for corosync nodeid 739513528
> Dec 10 13:38:53 [2231] hack1.example.com        cib:     info: crm_get_peer:  
> Created entry 97e43f18-db5c-4496-b8fc-bd7b4a6ea711/0x1deba80 for node 
> (null)/739513528 (1 total)
> Dec 10 13:38:53 [2231] hack1.example.com        cib:     info: crm_get_peer:  
> Node 739513528 has uuid 739513528
> Dec 10 13:38:53 [2231] hack1.example.com        cib:     info: 
> crm_update_peer_proc:  cluster_connect_cpg: Node (null)[739513528] - 
> corosync-cpg is now online
> Dec 10 13:38:53 [2231] hack1.example.com        cib:   notice: 
> crm_update_peer_state_iter:    crm_update_peer_proc: Node (null)[739513528] - 
> state is now member (was (null))
> Dec 10 13:38:53 [2231] hack1.example.com        cib:     info: 
> init_cs_connection_once:       Connection to 'corosync': established
> Dec 10 13:38:53 [2234] hack1.example.com      attrd:     info: main:  Cluster 
> connection active
> Dec 10 13:38:53 [2234] hack1.example.com      attrd:     info: 
> qb_ipcs_us_publish:    server name: attrd
> Dec 10 13:38:53 [2234] hack1.example.com      attrd:     info: main:  
> Accepting attribute updates
> Dec 10 13:38:53 [2231] hack1.example.com        cib:     info: 
> corosync_node_name:    Unable to get node name for nodeid 739513528
> Dec 10 13:38:53 [2231] hack1.example.com        cib:   notice: get_node_name: 
> Defaulting to uname -n for the local corosync node name
> Dec 10 13:38:53 [2231] hack1.example.com        cib:     info: crm_get_peer:  
> Node 739513528 is now known as hack1.example.com
> Dec 10 13:38:53 [2231] hack1.example.com        cib:     info: 
> qb_ipcs_us_publish:    server name: cib_ro
> Dec 10 13:38:53 [2231] hack1.example.com        cib:     info: 
> qb_ipcs_us_publish:    server name: cib_rw
> Dec 10 13:38:53 [2231] hack1.example.com        cib:     info: 
> qb_ipcs_us_publish:    server name: cib_shm
> Dec 10 13:38:53 [2231] hack1.example.com        cib:     info: cib_init:      
> Starting cib mainloop
> Dec 10 13:38:53 [2231] hack1.example.com        cib:     info: 
> cib_file_backup:       Archived previous version as 
> /var/lib/pacemaker/cib/cib-26.raw
> Dec 10 13:38:53 [2231] hack1.example.com        cib:     info: 
> cib_file_write_with_digest:    Wrote version 0.13.0 of the CIB to disk 
> (digest: d37e6c1873e854e52dd9b2caaa5e4beb)
> Dec 10 13:38:53 [2231] hack1.example.com        cib:     info: 
> cib_file_write_with_digest:    Reading cluster configuration file 
> /var/lib/pacemaker/cib/cib.dqtKlA (digest: /var/lib/pacemaker/cib/cib.BkV3eX)
> Dec 10 13:38:53 [2231] hack1.example.com        cib:     info: 
> pcmk_cpg_membership:   Node 739513528 joined group cib (counter=0.0)
> Dec 10 13:38:53 [2231] hack1.example.com        cib:     info: 
> pcmk_cpg_membership:   Node 739513528 still member of group cib 
> (peer=hack1.example.com, counter=0.0)
> Dec 10 13:38:53 [2231] hack1.example.com        cib:     info: 
> corosync_node_name:    Unable to get node name for nodeid 739513590
> Dec 10 13:38:53 [2231] hack1.example.com        cib:   notice: get_node_name: 
> Could not obtain a node name for corosync nodeid 739513590
> Dec 10 13:38:53 [2231] hack1.example.com        cib:     info: crm_get_peer:  
> Created entry 87459e91-7cff-4781-9c9a-5f78385e4b7c/0x1dec1d0 for node 
> (null)/739513590 (2 total)
> Dec 10 13:38:53 [2231] hack1.example.com        cib:     info: crm_get_peer:  
> Node 739513590 has uuid 739513590
> Dec 10 13:38:53 [2231] hack1.example.com        cib:     info: 
> pcmk_cpg_membership:   Node 739513590 still member of group cib (peer=(null), 
> counter=0.1)
> Dec 10 13:38:53 [2231] hack1.example.com        cib:     info: 
> crm_update_peer_proc:  pcmk_cpg_membership: Node (null)[739513590] - 
> corosync-cpg is now online
> Dec 10 13:38:53 [2231] hack1.example.com        cib:   notice: 
> crm_update_peer_state_iter:    crm_update_peer_proc: Node (null)[739513590] - 
> state is now member (was (null))
> Dec 10 13:38:53 [2236] hack1.example.com       crmd:     info: 
> do_cib_control:        CIB connection established
> Dec 10 13:38:53 [2236] hack1.example.com       crmd:   notice: 
> crm_cluster_connect:   Connecting to cluster infrastructure: corosync
> Dec 10 13:38:53 [2236] hack1.example.com       crmd:     info: 
> corosync_node_name:    Unable to get node name for nodeid 739513528
> Dec 10 13:38:53 [2236] hack1.example.com       crmd:   notice: get_node_name: 
> Could not obtain a node name for corosync nodeid 739513528
> Dec 10 13:38:53 [2236] hack1.example.com       crmd:     info: crm_get_peer:  
> Created entry 09d2eabe-fe50-4b40-b614-4f550dcbc677/0x156d0c0 for node 
> (null)/739513528 (1 total)
> Dec 10 13:38:53 [2236] hack1.example.com       crmd:     info: crm_get_peer:  
> Node 739513528 has uuid 739513528
> Dec 10 13:38:53 [2236] hack1.example.com       crmd:     info: 
> crm_update_peer_proc:  cluster_connect_cpg: Node (null)[739513528] - 
> corosync-cpg is now online
> Dec 10 13:38:53 [2236] hack1.example.com       crmd:     info: 
> init_cs_connection_once:       Connection to 'corosync': established
> Dec 10 13:38:53 [2236] hack1.example.com       crmd:     info: 
> corosync_node_name:    Unable to get node name for nodeid 739513528
> Dec 10 13:38:53 [2236] hack1.example.com       crmd:   notice: get_node_name: 
> Defaulting to uname -n for the local corosync node name
> Dec 10 13:38:53 [2236] hack1.example.com       crmd:     info: crm_get_peer:  
> Node 739513528 is now known as hack1.example.com
> Dec 10 13:38:53 [2236] hack1.example.com       crmd:     info: 
> peer_update_callback:  hack1.example.com is now in unknown state
> Dec 10 13:38:54 [2236] hack1.example.com       crmd:    error: 
> cluster_connect_quorum:        Corosync quorum is not configured
> Dec 10 13:38:54 [2234] hack1.example.com      attrd:     info: 
> attrd_cib_connect:     Connected to the CIB after 2 attempts
> Dec 10 13:38:54 [2234] hack1.example.com      attrd:     info: main:  CIB 
> connection active
> Dec 10 13:38:54 [2234] hack1.example.com      attrd:     info: 
> pcmk_cpg_membership:   Node 739513528 joined group attrd (counter=0.0)
> Dec 10 13:38:54 [2234] hack1.example.com      attrd:     info: 
> pcmk_cpg_membership:   Node 739513528 still member of group attrd 
> (peer=hack1.example.com, counter=0.0)
> Dec 10 13:38:54 [2232] hack1.example.com stonith-ng:   notice: setup_cib:     
> Watching for stonith topology changes
> Dec 10 13:38:54 [2232] hack1.example.com stonith-ng:     info: 
> qb_ipcs_us_publish:    server name: stonith-ng
> Dec 10 13:38:54 [2232] hack1.example.com stonith-ng:     info: main:  
> Starting stonith-ng mainloop
> Dec 10 13:38:54 [2232] hack1.example.com stonith-ng:     info: 
> pcmk_cpg_membership:   Node 739513528 joined group stonith-ng (counter=0.0)
> Dec 10 13:38:54 [2232] hack1.example.com stonith-ng:     info: 
> pcmk_cpg_membership:   Node 739513528 still member of group stonith-ng 
> (peer=hack1.example.com, counter=0.0)
> Dec 10 13:38:54 [2236] hack1.example.com       crmd:     info: do_ha_control: 
> Connected to the cluster
> Dec 10 13:38:54 [2236] hack1.example.com       crmd:     info: 
> lrmd_ipc_connect:      Connecting to lrmd
> Dec 10 13:38:54 [2234] hack1.example.com      attrd:     info: 
> corosync_node_name:    Unable to get node name for nodeid 739513590
> Dec 10 13:38:54 [2234] hack1.example.com      attrd:   notice: get_node_name: 
> Could not obtain a node name for corosync nodeid 739513590
> Dec 10 13:38:54 [2234] hack1.example.com      attrd:     info: crm_get_peer:  
> Created entry 1edc2c5f-1438-4f5c-b5aa-e95861c4205e/0x1814190 for node 
> (null)/739513590 (2 total)
> Dec 10 13:38:54 [2234] hack1.example.com      attrd:     info: crm_get_peer:  
> Node 739513590 has uuid 739513590
> Dec 10 13:38:54 [2234] hack1.example.com      attrd:     info: 
> pcmk_cpg_membership:   Node 739513590 still member of group attrd 
> (peer=(null), counter=0.1)
> Dec 10 13:38:54 [2234] hack1.example.com      attrd:     info: 
> crm_update_peer_proc:  pcmk_cpg_membership: Node (null)[739513590] - 
> corosync-cpg is now online
> Dec 10 13:38:54 [2234] hack1.example.com      attrd:   notice: 
> crm_update_peer_state_iter:    crm_update_peer_proc: Node (null)[739513590] - 
> state is now member (was (null))
> Dec 10 13:38:54 [2232] hack1.example.com stonith-ng:     info: 
> corosync_node_name:    Unable to get node name for nodeid 739513590
> Dec 10 13:38:54 [2232] hack1.example.com stonith-ng:   notice: get_node_name: 
> Could not obtain a node name for corosync nodeid 739513590
> Dec 10 13:38:54 [2232] hack1.example.com stonith-ng:     info: crm_get_peer:  
> Created entry 81f08c2d-6517-4b38-8fdc-51a4ada76af2/0x11fc5c0 for node 
> (null)/739513590 (2 total)
> Dec 10 13:38:54 [2232] hack1.example.com stonith-ng:     info: crm_get_peer:  
> Node 739513590 has uuid 739513590
> Dec 10 13:38:54 [2232] hack1.example.com stonith-ng:     info: 
> pcmk_cpg_membership:   Node 739513590 still member of group stonith-ng 
> (peer=(null), counter=0.1)
> Dec 10 13:38:54 [2232] hack1.example.com stonith-ng:     info: 
> crm_update_peer_proc:  pcmk_cpg_membership: Node (null)[739513590] - 
> corosync-cpg is now online
> Dec 10 13:38:54 [2232] hack1.example.com stonith-ng:   notice: 
> crm_update_peer_state_iter:    crm_update_peer_proc: Node (null)[739513590] - 
> state is now member (was (null))
> Dec 10 13:38:54 [2231] hack1.example.com        cib:     info: 
> cib_process_request:   Forwarding cib_modify operation for section nodes to 
> master (origin=local/crmd/3)
> Dec 10 13:38:54 [2236] hack1.example.com       crmd:     info: 
> do_lrm_control:        LRM connection established
> Dec 10 13:38:54 [2236] hack1.example.com       crmd:     info: do_started:    
> Delaying start, no membership data (0000000000100000)
> Dec 10 13:38:54 [2236] hack1.example.com       crmd:     info: 
> pcmk_cpg_membership:   Node 739513528 joined group crmd (counter=0.0)
> Dec 10 13:38:54 [2236] hack1.example.com       crmd:     info: 
> pcmk_cpg_membership:   Node 739513528 still member of group crmd 
> (peer=hack1.example.com, counter=0.0)
> Dec 10 13:38:54 [2232] hack1.example.com stonith-ng:     info: 
> corosync_node_name:    Unable to get node name for nodeid 739513528
> Dec 10 13:38:54 [2232] hack1.example.com stonith-ng:   notice: get_node_name: 
> Defaulting to uname -n for the local corosync node name
> Dec 10 13:38:54 [2232] hack1.example.com stonith-ng:     info: 
> init_cib_cache_cb:     Updating device list from the cib: init
> Dec 10 13:38:54 [2232] hack1.example.com stonith-ng:     info: 
> cib_devices_update:    Updating devices to version 0.13.0
> Dec 10 13:38:54 [2232] hack1.example.com stonith-ng:   notice: unpack_config: 
> On loss of CCM Quorum: Ignore
> Dec 10 13:38:54 [2232] hack1.example.com stonith-ng:     info: crm_get_peer:  
> Node 739513590 is now known as hack2.example.com
> Dec 10 13:38:54 [2231] hack1.example.com        cib:     info: 
> corosync_node_name:    Unable to get node name for nodeid 739513528
> Dec 10 13:38:54 [2231] hack1.example.com        cib:   notice: get_node_name: 
> Defaulting to uname -n for the local corosync node name
> Dec 10 13:38:54 [2236] hack1.example.com       crmd:     info: 
> corosync_node_name:    Unable to get node name for nodeid 739513590
> Dec 10 13:38:54 [2236] hack1.example.com       crmd:   notice: get_node_name: 
> Could not obtain a node name for corosync nodeid 739513590
> Dec 10 13:38:54 [2236] hack1.example.com       crmd:     info: crm_get_peer:  
> Created entry b34ca282-9c15-47de-8c98-cb59d1d8bed4/0x16b5130 for node 
> (null)/739513590 (2 total)
> Dec 10 13:38:54 [2236] hack1.example.com       crmd:     info: crm_get_peer:  
> Node 739513590 has uuid 739513590
> Dec 10 13:38:54 [2236] hack1.example.com       crmd:     info: 
> pcmk_cpg_membership:   Node 739513590 still member of group crmd 
> (peer=(null), counter=0.1)
> Dec 10 13:38:54 [2236] hack1.example.com       crmd:     info: 
> crm_update_peer_proc:  pcmk_cpg_membership: Node (null)[739513590] - 
> corosync-cpg is now online
> Dec 10 13:38:54 [2236] hack1.example.com       crmd:   notice: 
> crmd_enable_notifications:     Notifications disabled
> Dec 10 13:38:54 [2236] hack1.example.com       crmd:     info: do_started:    
> Delaying start, no membership data (0000000000100000)
> Dec 10 13:38:54 [2231] hack1.example.com        cib:     info: 
> cib_process_request:   Completed cib_modify operation for section nodes: OK 
> (rc=0, origin=hack1.example.com/crmd/3, version=0.13.0)
> 
> 
> Any help would be much appreciated.
> 
> Best  regards,
> --
> Louis Munro
> 
> 
> 
> 
> 
> _______________________________________________
> Users mailing list: Users@clusterlabs.org
> http://clusterlabs.org/mailman/listinfo/users
> 
> Project Home: http://www.clusterlabs.org
> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
> Bugs: http://bugs.clusterlabs.org
> 


_______________________________________________
Users mailing list: Users@clusterlabs.org
http://clusterlabs.org/mailman/listinfo/users

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org

Reply via email to