This doesn't look promising:

lrmd: [4939]: info: G_main_add_SignalHandler: Added signal handler for signal 15
lrmd: [4946]: info: Signal sent to pid=4939, waiting for process to exit
lrmd: [4939]: info: G_main_add_SignalHandler: Added signal handler for signal 17
lrmd: [4939]: info: enabling coredumps
lrmd: [4939]: info: G_main_add_SignalHandler: Added signal handler for signal 10
lrmd: [4939]: info: G_main_add_SignalHandler: Added signal handler for signal 12
lrmd: [4939]: info: Started.
lrmd: [4939]: info: lrmd is shutting down

The lrmd comes up but then immediately shuts down.
Perhaps try enabling debug to see if that sheds any light.

On 06/04/2013, at 4:58 AM, Jimmy Magee <jimmy.ma...@vennetics.com> wrote:

> Hi guys,
> 
> Apologies for reposting this query, it inadvertently got added to an existing 
> topic!
> 
> 
> We have a three node cluster deployed in a customer's network:
> - 2 nodes are on the same switch
> - 3rd node on the same subnet but there's a router in between.
> - IP Multicast is enabled and has been tested using omping as follows..
> 
> On each node ran..
> 
> omping node01 node02 node3
> 
> 
> ON node 3
> 
> Node01 :   unicast, xmt/rcv/%loss = 23/23/0%, min/avg/max/std-dev = 
> 0.128/0.181/0.255/0.025
> Node01 : multicast, xmt/rcv/%loss = 23/23/0%, min/avg/max/std-dev = 
> 0.140/0.187/0.219/0.021
> Node02 :   unicast, xmt/rcv/%loss = 8/8/0%, min/avg/max/std-dev = 
> 0.115/0.150/0.168/0.021
> Node02 : multicast, xmt/rcv/%loss = 8/8/0%, min/avg/max/std-dev = 
> 0.134/0.162/0.177/0.014
> 
> 
> On node 2
> 
> 
> Node01 :   unicast, xmt/rcv/%loss = 9/9/0%, min/avg/max/std-dev = 
> 0.168/0.191/0.205/0.014
> Node01 : multicast, xmt/rcv/%loss = 9/8/11% (seq>=2 0%), min/avg/max/std-dev 
> = 0.138/0.179/0.206/0.028
> Node03 :   unicast, xmt/rcv/%loss = 9/9/0%, min/avg/max/std-dev = 
> 0.112/0.149/0.175/0.022
> Node03 : multicast, xmt/rcv/%loss = 9/8/11% (seq>=2 0%), min/avg/max/std-dev 
> = 0.124/0.167/0.178/0.018
> 
> 
> 
> On node 1
> 
> Node02 :   unicast, xmt/rcv/%loss = 8/8/0%, min/avg/max/std-dev = 
> 0.154/0.185/0.208/0.019
> Node02 : multicast, xmt/rcv/%loss = 8/8/0%, min/avg/max/std-dev = 
> 0.175/0.198/0.214/0.015
> Node03 :   unicast, xmt/rcv/%loss = 23/23/0%, min/avg/max/std-dev = 
> 0.114/0.160/0.185/0.019
> Node03 : multicast, xmt/rcv/%loss = 23/22/4% (seq>=2 0%), min/avg/max/std-dev 
> = 0.124/0.172/0.197/0.019
> 
> 
> - Problem is intermittent but frequent. Occasionally starts fine when started 
> from scratch.
> 
> We suspect the problem is related to node 3 as we can see lrmd failures as 
> per the attached log. We've checked permissions are ok as per 
> https://bugs.launchpad.net/ubuntu/+source/cluster-glue/+bug/676391
> 
> 
> 
> stonith-ng[1437]:    error: ais_dispatch: AIS connection failed
> stonith-ng[1437]:    error: stonith_peer_ais_destroy: AIS connection 
> terminated
> corosync[1430]:   [SERV  ] Service engine unloaded: Pacemaker Cluster Manager 
> 1.1.6
> corosync[1430]:   [SERV  ] Service engine unloaded: corosync extended virtual 
> synchrony service
> corosync[1430]:   [SERV  ] Service engine unloaded: corosync configuration 
> service
> corosync[1430]:   [SERV  ] Service engine unloaded: corosync cluster closed 
> process group service v1.01
> corosync[1430]:   [SERV  ] Service engine unloaded: corosync cluster config 
> database access v1.01
> corosync[1430]:   [SERV  ] Service engine unloaded: corosync profile loading 
> service
> corosync[1430]:   [SERV  ] Service engine unloaded: corosync cluster quorum 
> service v0.1
> corosync[1430]:   [MAIN  ] Corosync Cluster Engine exiting with status 0 at 
> main.c:1894.
> 
> corosync[4931]:   [MAIN  ] Corosync built-in features: nss dbus rdma snmp
> corosync[4931]:   [MAIN  ] Successfully read main configuration file 
> '/etc/corosync/corosync.conf'.
> corosync[4931]:   [TOTEM ] Initializing transport (UDP/IP Multicast).
> corosync[4931]:   [TOTEM ] Initializing transmit/receive security: 
> libtomcrypt SOBER128/SHA1HMAC (mode 0).
> corosync[4931]:   [TOTEM ] The network interface [10.87.79.59] is now up.
> corosync[4931]:   [pcmk  ] Logging: Initialized pcmk_startup
> corosync[4931]:   [SERV  ] Service engine loaded: Pacemaker Cluster Manager 
> 1.1.6
> corosync[4931]:   [pcmk  ] Logging: Initialized pcmk_startup
> corosync[4931]:   [SERV  ] Service engine loaded: Pacemaker Cluster Manager 
> 1.1.6
> corosync[4931]:   [SERV  ] Service engine loaded: corosync extended virtual 
> synchrony service
> corosync[4931]:   [SERV  ] Service engine loaded: corosync configuration 
> service
> orosync[4931]:   [SERV  ] Service engine loaded: corosync cluster closed 
> process group service v1.01
> corosync[4931]:   [SERV  ] Service engine loaded: corosync cluster config 
> database access v1.01
> corosync[4931]:   [SERV  ] Service engine loaded: corosync profile loading 
> service
> corosync[4931]:   [SERV  ] Service engine loaded: corosync cluster quorum 
> service v0.1
> corosync[4931]:   [MAIN  ] Compatibility mode set to whitetank.  Using V1 and 
> V2 of the synchronization engine.
> corosync[4931]:   [TOTEM ] A processor joined or left the membership and a 
> new membership was formed.
> corosync[4931]:   [CPG   ] chosen downlist: sender r(0) ip(10.87.79.59) ; 
> members(old:0 left:0)
> corosync[4931]:   [MAIN  ] Completed service synchronization, ready to 
> provide service.
> cib[4937]:     info: crm_log_init_worker: Changed active directory to 
> /var/lib/heartbeat/cores/hacluster
> cib[4937]:     info: retrieveCib: Reading cluster configuration from: 
> /var/lib/heartbeat/crm/cib.xml (digest: /var/lib/heartbeat/crm/cib.xml.sig)
> cib[4937]:     info: validate_with_relaxng: Creating RNG parser context
> stonith-ng[4945]:     info: crm_log_init_worker: Changed active directory to 
> /var/lib/heartbeat/cores/root
> stonith-ng[4945]:     info: get_cluster_type: Cluster type is: 'openais'
> stonith-ng[4945]:   notice: crm_cluster_connect: Connecting to cluster 
> infrastructure: classic openais (with plugin)
> stonith-ng[4945]:     info: init_ais_connection_classic: Creating connection 
> to our Corosync plugin
> cib[4944]:     info: crm_log_init_worker: Changed active directory to 
> /var/lib/heartbeat/cores/hacluster
> cib[4944]:     info: retrieveCib: Reading cluster configuration from: 
> /var/lib/heartbeat/crm/cib.xml (digest: /var/lib/heartbeat/crm/cib.xml.sig)
> stonith-ng[4945]:     info: init_ais_connection_classic: AIS connection 
> established
> stonith-ng[4945]:     info: get_ais_nodeid: Server details: id=1003428268 
> uname=w0110Danmtapp03 cname=pcmk
> stonith-ng[4945]:     info: init_ais_connection_once: Connection to 'classic 
> openais (with plugin)': established
> stonith-ng[4945]:     info: crm_new_peer: Node node03 now has id: 1003428268
> stonith-ng[4945]:     info: crm_new_peer: Node 1003428268 is now known as 
> node03
> cib[4944]:     info: validate_with_relaxng: Creating RNG parser context
> lrmd: [4939]: info: G_main_add_SignalHandler: Added signal handler for signal 
> 15
> lrmd: [4946]: info: Signal sent to pid=4939, waiting for process to exit
> lrmd: [4939]: info: G_main_add_SignalHandler: Added signal handler for signal 
> 17
> lrmd: [4939]: info: enabling coredumps
> stonith-ng[4938]:     info: crm_log_init_worker: Changed active directory to 
> /var/lib/heartbeat/cores/root
> lrmd: [4939]: info: G_main_add_SignalHandler: Added signal handler for signal 
> 10
> lrmd: [4939]: info: G_main_add_SignalHandler: Added signal handler for signal 
> 12
> lrmd: [4939]: info: Started.
> stonith-ng[4938]:     info: get_cluster_type: Cluster type is: 'openais'
> lrmd: [4939]: info: lrmd is shutting down
> stonith-ng[4938]:   notice: crm_cluster_connect: Connecting to cluster 
> infrastructure: classic openais (with plugin)
> stonith-ng[4938]:     info: init_ais_connection_classic: Creating connection 
> to our Corosync plugin
> attrd[4940]:     info: crm_log_init_worker: Changed active directory to 
> /var/lib/heartbeat/cores/hacluster
> pengine[4941]:     info: crm_log_init_worker: Changed active directory to 
> /var/lib/heartbeat/cores/hacluster
> attrd[4940]:     info: main: Starting up
> attrd[4940]:     info: get_cluster_type: Cluster type is: 'openais'
> attrd[4940]:   notice: crm_cluster_connect: Connecting to cluster 
> infrastructure: classic openais (with plugin)
> attrd[4940]:     info: init_ais_connection_classic: Creating connection to 
> our Corosync plugin
> crmd[4942]:     info: crm_log_init_worker: Changed active directory to 
> /var/lib/heartbeat/cores/hacluster
> pengine[4941]:     info: main: Starting pengine
> crmd[4942]:   notice: main: CRM Hg Version: 
> 148fccfd5985c5590cc601123c6c16e966b85d14
> pengine[4948]:     info: crm_log_init_worker: Changed active directory to 
> /var/lib/heartbeat/cores/hacluster
> pengine[4948]:  warning: main: Terminating previous PE instance
> attrd[4947]:     info: crm_log_init_worker: Changed active directory to 
> /var/lib/heartbeat/cores/hacluster
> pengine[4941]:  warning: process_pe_message: Received quit message, 
> terminating
> attrd[4947]:     info: main: Starting up
> attrd[4947]:     info: get_cluster_type: Cluster type is: 'openais'
> attrd[4947]:   notice: crm_cluster_connect: Connecting to cluster 
> infrastructure: classic openais (with plugin)
> attrd[4947]:     info: init_ais_connection_classic: Creating connection to 
> our Corosync plugin
> crmd[4949]:     info: crm_log_init_worker: Changed active directory to 
> /var/lib/heartbeat/cores/hacluster
> crmd[4949]:   notice: main: CRM Hg Version: 
> 148fccfd5985c5590cc601123c6c16e966b85d14
> stonith-ng[4938]:     info: init_ais_connection_classic: AIS connection 
> established
> stonith-ng[4938]:     info: get_ais_nodeid: Server details: id=1003428268 
> uname=node03 cname=pcmk
> stonith-ng[4938]:     info: init_ais_connection_once: Connection to 'classic 
> openais (with plugin)': established
> stonith-ng[4938]:     info: crm_new_peer: Node node03 now has id: 1003428268
> stonith-ng[4938]:     info: crm_new_peer: Node 1003428268 is now known as 
> node03
> attrd[4940]:     info: init_ais_connection_classic: AIS connection established
> attrd[4940]:     info: get_ais_nodeid: Server details: id=1003428268 
> uname=node03 cname=pcmk
> attrd[4940]:     info: init_ais_connection_once: Connection to 'classic 
> openais (with plugin)': established
> attrd[4940]:     info: crm_new_peer: Node node03 now has id: 1003428268
> attrd[4940]:     info: crm_new_peer: Node 1003428268 is now known as node03
> attrd[4940]:     info: main: Cluster connection active
> attrd[4940]:     info: main: Accepting attribute updates
> attrd[4940]:   notice: main: Starting mainloop...
> attrd[4947]:     info: init_ais_connection_classic: AIS connection established
> attrd[4947]:     info: get_ais_nodeid: Server details: id=1003428268 
> uname=node03 cname=pcmk
> attrd[4947]:     info: init_ais_connection_once: Connection to 'classic 
> openais (with plugin)': established
> attrd[4947]:     info: crm_new_peer: Node node03 now has id: 1003428268
> attrd[4947]:     info: crm_new_peer: Node 1003428268 is now known as node03
> attrd[4947]:     info: main: Cluster connection active
> attrd[4947]:     info: main: Accepting attribute updates
> attrd[4947]:   notice: main: Starting mainloop...
> cib[4937]:     info: startCib: CIB Initialization completed successfully
> cib[4937]:     info: get_cluster_type: Cluster type is: 'openais'
> cib[4937]:   notice: crm_cluster_connect: Connecting to cluster 
> infrastructure: classic openais (with plugin)
> cib[4937]:     info: init_ais_connection_classic: Creating connection to our 
> Corosync plugin
> cib[4944]:     info: startCib: CIB Initialization completed successfully
> cib[4944]:     info: get_cluster_type: Cluster type is: 'openais'
> cib[4944]:   notice: crm_cluster_connect: Connecting to cluster 
> infrastructure: classic openais (with plugin)
> cib[4944]:     info: init_ais_connection_classic: Creating connection to our 
> Corosync plugin
> cib[4937]:     info: init_ais_connection_classic: AIS connection established
> cib[4937]:     info: get_ais_nodeid: Server details: id=1003428268 
> uname=node03 cname=pcmk
> cib[4937]:     info: init_ais_connection_once: Connection to 'classic openais 
> (with plugin)': established
> cib[4937]:     info: crm_new_peer: Node node03 now has id: 1003428268
> cib[4937]:     info: crm_new_peer: Node 1003428268 is now known as node03
> cib[4937]:     info: cib_init: Starting cib mainloop
> cib[4937]:     info: ais_dispatch_message: Membership 6892: quorum still lost
> cib[4937]:     info: crm_update_peer: Node node03: id=1003428268 state=member 
> (new) addr=r(0) ip(10.87.79.59)  (new) votes=1 (new) born=0 seen=6892 
> proc=00000000000000000000000000111312 (new)
> cib[4944]:     info: init_ais_connection_classic: AIS connection established
> cib[4944]:     info: get_ais_nodeid: Server details: id=1003428268 
> uname=node03 cname=pcmk
> cib[4944]:     info: init_ais_connection_once: Connection to 'classic openais 
> (with plugin)': established
> cib[4944]:     info: crm_new_peer: Node node03 now has id: 1003428268
> cib[4944]:     info: crm_new_peer: Node 1003428268 is now known as node03
> cib[4944]:     info: cib_init: Starting cib mainloop
> stonith-ng[4945]:   notice: setup_cib: Watching for stonith topology changes
> stonith-ng[4945]:     info: main: Starting stonith-ng mainloop
> cib[4937]:     info: ais_dispatch_message: Membership 6896: quorum still lost
> corosync[4931]:   [TOTEM ] A processor joined or left the membership and a 
> new membership was formed.
> cib[4937]:     info: crm_new_peer: Node <null> now has id: 969873836
> cib[4937]:     info: crm_update_peer: Node (null): id=969873836 state=member 
> (new) addr=r(0) ip(172.25.207.57)  votes=0 born=0 seen=6896 
> proc=00000000000000000000000000000000
> cib[4937]:     info: crm_new_peer: Node <null> now has id: 986651052
> cib[4937]:     info: crm_update_peer: Node (null): id=986651052 state=member 
> (new) addr=r(0) ip(172.25.207.58)  votes=0 born=0 seen=6896 
> proc=00000000000000000000000000000000
> cib[4937]:   notice: ais_dispatch_message: Membership 6896: quorum acquired
> cib[4937]:     info: crm_get_peer: Node 986651052 is now known as node02
> cib[4937]:     info: crm_update_peer: Node node02: id=986651052 state=member 
> addr=r(0) ip(172.25.207.58)  votes=1 (new) born=6812 seen=6896 
> proc=00000000000000000000000000111312 (new)
> cib[4937]:     info: ais_dispatch_message: Membership 6896: quorum retained
> cib[4937]:     info: crm_get_peer: Node 969873836 is now known as node01
> cib[4937]:     info: crm_update_peer: Node node01: id=969873836 state=member 
> addr=r(0) ip(172.25.207.57)  votes=1 (new) born=6848 seen=6896 
> proc=00000000000000000000000000111312 (new)
> rsyslogd-2177: imuxsock begins to drop messages from pid 4931 due to 
> rate-limiting
> crmd[4942]:     info: do_cib_control: CIB connection established
> crmd[4942]:     info: get_cluster_type: Cluster type is: 'openais'
> crmd[4942]:   notice: crm_cluster_connect: Connecting to cluster 
> infrastructure: classic openais (with plugin)
> crmd[4942]:     info: init_ais_connection_classic: Creating connection to our 
> Corosync plugin
> cib[4937]:     info: cib_process_diff: Diff 1.249.28 -> 1.249.29 not applied 
> to 1.249.0: current "num_updates" is less than required
> cib[4937]:     info: cib_server_process_diff: Requesting re-sync from peer
> crmd[4949]:     info: do_cib_control: CIB connection established
> crmd[4949]:     info: get_cluster_type: Cluster type is: 'openais'
> crmd[4949]:   notice: crm_cluster_connect: Connecting to cluster 
> infrastructure: classic openais (with plugin)
> crmd[4949]:     info: init_ais_connection_classic: Creating connection to our 
> Corosync plugin
> stonith-ng[4938]:   notice: setup_cib: Watching for stonith topology changes
> stonith-ng[4938]:     info: main: Starting stonith-ng mainloop
> cib[4937]:   notice: cib_server_process_diff: Not applying diff 1.249.29 -> 
> 1.249.30 (sync in progress)
> crmd[4942]:     info: init_ais_connection_classic: AIS connection established
> crmd[4942]:     info: get_ais_nodeid: Server details: id=1003428268 
> uname=node03 cname=pcmk
> crmd[4942]:     info: init_ais_connection_once: Connection to 'classic 
> openais (with plugin)': established
> crmd[4942]:     info: crm_new_peer: Node node03 now has id: 1003428268
> crmd[4942]:     info: crm_new_peer: Node 1003428268 is now known as node03
> crmd[4942]:     info: ais_status_callback: status: node03 is now unknown
> crmd[4942]:     info: do_ha_control: Connected to the cluster
> crmd[4942]:  warning: do_lrm_control: Failed to sign on to the LRM 1 (30 max) 
> times
> crmd[4949]:     info: init_ais_connection_classic: AIS connection established
> crmd[4949]:     info: get_ais_nodeid: Server details: id=1003428268 
> uname=node03 cname=pcmk
> crmd[4949]:     info: init_ais_connection_once: Connection to 'classic 
> openais (with plugin)': established
> crmd[4942]:   notice: ais_dispatch_message: Membership 6896: quorum acquired
> crmd[4949]:     info: crm_new_peer: Node node03 now has id: 1003428268
> crmd[4949]:     info: crm_new_peer: Node 1003428268 is now known as node03
> crmd[4942]:     info: crm_new_peer: Node node01 now has id: 969873836
> crmd[4949]:     info: ais_status_callback: status: node03 is now unknown
> crmd[4942]:     info: crm_new_peer: Node 969873836 is now known as node01
> crmd[4949]:     info: do_ha_control: Connected to the cluster
> crmd[4942]:     info: ais_status_callback: status: node01 is now unknown
> crmd[4942]:     info: ais_status_callback: status: node01 is now member (was 
> unknown)
> crmd[4942]:     info: crm_update_peer: Node node01: id=969873836 state=member 
> (new) addr=r(0) ip(172.25.207.57)  votes=1 born=6848 seen=6896 
> proc=00000000000000000000000000111312
> crmd[4942]:     info: crm_new_peer: Node node02 now has id: 986651052
> crmd[4942]:     info: crm_new_peer: Node 986651052 is now known as node02
> crmd[4942]:     info: ais_status_callback: status: node02 is now unknown
> crmd[4949]:  warning: do_lrm_control: Failed to sign on to the LRM 1 (30 max) 
> times
> crmd[4942]:     info: ais_status_callback: status: node02 is now member (was 
> unknown)
> crmd[4942]:     info: crm_update_peer: Node node02: id=986651052 state=member 
> (new) addr=r(0) ip(172.25.207.58)  votes=1 born=6812 seen=6896 
> proc=00000000000000000000000000111312
> crmd[4942]:   notice: crmd_peer_update: Status update: Client node03/crmd now 
> has status [online] (DC=<null>)
> crmd[4942]:     info: ais_status_callback: status: node03 is now member (was 
> unknown)
> crmd[4942]:     info: crm_update_peer: Node node03: id=1003428268 
> state=member (new) addr=r(0) ip(10.87.79.59)  (new) votes=1 (new) born=6896 
> seen=6896 proc=00000000000000000000000000111312 (new)
> crmd[4942]:     info: ais_dispatch_message: Membership 6896: quorum retained
> cib[4937]:   notice: cib_server_process_diff: Not applying diff 1.249.30 -> 
> 1.249.31 (sync in progress)
> crmd[4942]:  warning: do_lrm_control: Failed to sign on to the LRM 2 (30 max) 
> times
> crmd[4942]:  warning: do_lrm_control: Failed to sign on to the LRM 3 (30 max) 
> times
> crmd[4949]:  warning: do_lrm_control: Failed to sign on to the LRM 2 (30 max) 
> times
> crmd[4949]:   notice: ais_dispatch_message: Membership 6896: quorum acquired
> rsyslogd-2177: imuxsock begins to drop messages from pid 4937 due to 
> rate-limiting
> crmd[4942]:  warning: do_lrm_control: Failed to sign on to the LRM 4 (30 max) 
> times
> crmd[4942]:  warning: do_lrm_control: Failed to sign on to the LRM 5 (30 max) 
> times
> pengine[4948]:     info: main: Starting pengine
> crmd[4942]:     info: crm_timer_popped: Wait Timer (I_NULL) just popped 
> (2000ms)
> warning: do_lrm_control: Failed to sign on to the LRM 6 (30 max) times
> crmd[4949]:     info: crm_timer_popped: Wait Timer (I_NULL) just popped 
> (2000ms)
> crmd[4949]:  warning: do_lrm_control: Failed to sign on to the LRM 3 (30 max) 
> times
> attrd[4940]:     info: cib_connect: Connected to the CIB after 1 signon 
> attempts
> attrd[4940]:     info: cib_connect: Sending full refresh
> crmd[4942]:     info: crm_timer_popped: Wait Timer (I_NULL) just popped 
> (2000ms)
> crmd[4942]:  warning: do_lrm_control: Failed to sign on to the LRM 7 (30 max) 
> times
> attrd[4947]:     info: cib_connect: Connected to the CIB after 1 signon 
> attempts
> attrd[4947]:     info: cib_connect: Sending full refresh
> crmd[4949]:     info: crm_timer_popped: Wait Timer (I_NULL) just popped 
> (2000ms)
> crmd[4949]:  warning: do_lrm_control: Failed to sign on to the LRM 4 (30 max) 
> times
> crmd[4942]:     info: crm_timer_popped: Wait Timer (I_NULL) just popped 
> (2000ms)
> crmd[4942]:  warning: do_lrm_control: Failed to sign on to the LRM 8 (30 max) 
> times
> crmd[4949]:     info: crm_timer_popped: Wait Timer (I_NULL) just popped 
> (2000ms)
> crmd[4949]:  warning: do_lrm_control: Failed to sign on to the LRM 5 (30 max) 
> times
> crmd[4942]:     info: crm_timer_popped: Wait Timer (I_NULL) just popped 
> (2000ms)
> crmd[4942]:  warning: do_lrm_control: Failed to sign on to the LRM 9 (30 max) 
> times
> crmd[4949]:     info: crm_timer_popped: Wait Timer (I_NULL) just popped 
> (2000ms)
> crmd[4949]:  warning: do_lrm_control: Failed to sign on to the LRM 6 (30 max) 
> times
> crmd[4942]:     info: crm_timer_popped: Wait Timer (I_NULL) just popped 
> (2000ms)
> crmd[4942]:  warning: do_lrm_control: Failed to sign on to the LRM 10 (30 
> max) times
> crmd[4949]:     info: crm_timer_popped: Wait Timer (I_NULL) just popped 
> (2000ms)
> crmd[4949]:  warning: do_lrm_control: Failed to sign on to the LRM 7 (30 max) 
> times
> crmd[4942]:     info: crm_timer_popped: Wait Timer (I_NULL) just popped 
> (2000ms)
> crmd[4949]:     info: crm_timer_popped: Wait Timer (I_NULL) just popped 
> (2000ms)
> crmd[4942]:  warning: do_lrm_control: Failed to sign on to the LRM 11 (30 
> max) times
> crmd[4949]:  warning: do_lrm_control: Failed to sign on to the LRM 8 (30 max) 
> times
> crmd[4942]:     info: crm_timer_popped: Wait Timer (I_NULL) just popped 
> (2000ms)
> crmd[4949]:     info: crm_timer_popped: Wait Timer (I_NULL) just popped 
> (2000ms)
> crmd[4942]:  warning: do_lrm_control: Failed to sign on to the LRM 12 (30 
> max) times
> crmd[4949]:  warning: do_lrm_control: Failed to sign on to the LRM 9 (30 max) 
> times
> crmd[4949]:     info: crm_timer_popped: Wait Timer (I_NULL) just popped 
> (2000ms)
> crmd[4942]:     info: crm_timer_popped: Wait Timer (I_NULL) just popped 
> (2000ms)
> crmd[4942]:  warning: do_lrm_control: Failed to sign on to the LRM 13 (30 
> max) times
> crmd[4949]:  warning: do_lrm_control: Failed to sign on to the LRM 10 (30 
> max) times
> crmd[4949]:     info: crm_timer_popped: Wait Timer (I_NULL) just popped 
> (2000ms)
> crmd[4942]:     info: crm_timer_popped: Wait Timer (I_NULL) just popped 
> (2000ms)
> crmd[4942]:  warning: do_lrm_control: Failed to sign on to the LRM 14 (30 
> max) times
> crmd[4949]:  warning: do_lrm_control: Failed to sign on to the LRM 11 (30 
> max) times
> crmd[4949]:     info: crm_timer_popped: Wait Timer (I_NULL) just popped 
> (2000ms)
> crmd[4949]:  warning: do_lrm_control: Failed to sign on to the LRM 12 (30 
> max) times
> crmd[4942]:     info: crm_timer_popped: Wait Timer (I_NULL) just popped 
> (2000ms)
> crmd[4942]:  warning: do_lrm_control: Failed to sign on to the LRM 15 (30 
> max) times
> crmd[4949]:     info: crm_timer_popped: Wait Timer (I_NULL) just popped 
> (2000ms)
> crmd[4949]:  warning: do_lrm_control: Failed to sign on to the LRM 13 (30 
> max) times
> crmd[4942]:     info: crm_timer_popped: Wait Timer (I_NULL) just popped 
> (2000ms)
> crmd[4942]:  warning: do_lrm_control: Failed to sign on to the LRM 16 (30 
> max) times
> crmd[4949]:     info: crm_timer_popped: Wait Timer (I_NULL) just popped 
> (2000ms)
> crmd[4949]:  warning: do_lrm_control: Failed to sign on to the LRM 14 (30 
> max) times
> crmd[4942]:     info: crm_timer_popped: Wait Timer (I_NULL) just popped 
> (2000ms)
> crmd[4942]:  warning: do_lrm_control: Failed to sign on to the LRM 17 (30 
> max) times
> crmd[4949]:     info: crm_timer_popped: Wait Timer (I_NULL) just popped 
> (2000ms)
> crmd[4949]:  warning: do_lrm_control: Failed to sign on to the LRM 15 (30 
> max) times
> crmd[4942]:     info: crm_timer_popped: Wait Timer (I_NULL) just popped 
> (2000ms)
> crmd[4942]:  warning: do_lrm_control: Failed to sign on to the LRM 18 (30 
> max) times
> 
> 
> We have the following components installed..
> 
> 
> corosynclib-1.4.1-15.el6.x86_64
> corosync-1.4.1-15.el6.x86_64
> cluster-glue-libs-1.0.5-6.el6.x86_64
> clusterlib-3.0.12.1-49.el6.x86_64
> pacemaker-cluster-libs-1.1.7-6.el6.x86_64
> cluster-glue-1.0.5-6.el6.x86_64
> resource-agents-3.9.2-12.el6.x86_64
> 
> 
> 
> We'd appreciate assistance on how to debug what the issue may be and some 
> possible causes.
> 
> Cheers,
> Jimmy
> _______________________________________________
> Linux-HA mailing list
> Linux-HA@lists.linux-ha.org
> http://lists.linux-ha.org/mailman/listinfo/linux-ha
> See also: http://linux-ha.org/ReportingProblems

_______________________________________________
Linux-HA mailing list
Linux-HA@lists.linux-ha.org
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems

Reply via email to