Hello Keith,
you are right.
I don't have got an idea why there are two firewalls installed, beside iptables
(that I checked) there is a
second firewalld on our systems and this one was still active.
I noticed that there is something wrong when I changed the configuration back
to unicast and
saw ICMP drop packages.
Thanks,
Hans
________________________________
Von: Keith Ouellette <keith.ouelle...@airgas.com>
Hans,
Is the multicast port 5405 "opened" in the firewall? That has bitten me
before.
Thanks,
Keith
________________________________
From: Hans Bert
Sent: Wednesday, January 30, 2013 8:22 AM
To: and k; The Pacemaker cluster resource manager
Subject: Re: [Pacemaker] No communication between nodes (setup problem)
Hi,
in the meantime I modified the configuration to check if it works with multicast
totem {
version: 2
secauth: off
cluster_name: mcscluster
interface {
ringnumber: 0
bindnetaddr: 192.168.100.0
mcastaddr: 239.255.1.12
mcastport: 5405
ttl: 1
}
}
but unfortunately it is still not working.
I started wireshark and I can see on both hosts MC packages from both hosts.
YES selinux is disabled on both nodes
[root@server1 corosync]# selinuxenabled
[root@server1 corosync]# echo $?
1
Something else I found out is:
[root@server1 corosync]# pcs status nodes both
Error mapping 192.168.100.111
Error mapping 192.168.100.112
Corosync Nodes:
Online:
Offline: 192.168.100.111 192.168.100.112
Pacemaker Nodes:
Online: server1
Standby:
Offline:
[root@server2 corosync]# pcs status nodes both
Error mapping 192.168.100.111
Error mapping 192.168.100.112
Corosync Nodes:
Online:
Offline: 192.168.100.111 192.168.100.112
Pacemaker Nodes:
Online: server2
Standby:
Offline:
any further hints?
Best regards,
Hans
________________________________
Hi,
It seem to be problem with network traffic.
Have you tried to sniff network traffic to be sure that udp traffic reaches
from one node to another ??
Try on server1:
tcpdump -i interface -p udp -s 192.168.100.112
on server2:
tcpdump -i interface -p udp -s 192.168.100.111
if there will be no packet traffic, that means you have some network issue.
BTW: Is SELinux enabled on nodes ??
--
Regards
Andrew
2013/1/30 Hans Bert
Hello,
>
>we had to move from Fedora 16 to Fedora 18 and wanted to set up Corosync with
>Pacemaker and PCS as management tool.
>With F16 our cluster was running pretty good, but with F18 after 5 days we are
>reaching the point were we don't have
>got ideas what might be the problem(s).
>
>
>The cluster is build of two servers (server1=192.168.100.111;
>server2=192.168.100.112)
>
>Based on the Howto for F18 with pcs we created the following corosync.conf:
>
>totem {
> version: 2
> secauth: off
> cluster_name: mcscluster
> transport: udpu
>}
>
>nodelist {
> node {
> ring0_addr: 192.168.100.111
> }
> node {
> ring0_addr: 192.168.100.112
> }
>}
>
>quorum {
> provider: corosync_votequorum
>}
>
>logging {
> fileline: off
> to_stderr: no
> to_logfile: yes
> to_syslog: yes
> logfile: /var/log/cluster/corosync.log
> debug: on
> timestamp: on
>}
>
>
>
>After we started the server a status check shows us:
>
>
>[root@server1 corosync]#pcs status corosync
>
>Membership information
>----------------------
> Nodeid Votes Name
>1868867776 1 server1 (local)
>
>[root@server1 ~]# pcs status
>Last updated: Wed Jan 30 10:45:17 2013
>Last change: Wed Jan 30 10:18:56 2013 via cibadmin on server1
>Stack: corosync
>Current DC: server1 (1868867776) - partition WITHOUT quorum
>Version: 1.1.8-3.fc18-394e906
>1 Nodes configured, unknown expected votes
>0 Resources configured.
>
>
>Online: [ server1 ]
>
>Full list of resources:
>
>
>
>And on the other server:
>
>
>[root@server2 corosync]# pcs status corosync
>
>Membership information
>----------------------
> Nodeid Votes Name
>1885644992 1 server2 (local)
>
>[root@server2 corosync]# pcs status
>Last updated: Wed Jan 30 10:44:40 2013
>Last change: Wed Jan 30 10:19:36 2013 via cibadmin on server2
>Stack: corosync
>Current DC: server2 (1885644992) - partition WITHOUT quorum
>Version: 1.1.8-3.fc18-394e906
>1 Nodes configured, unknown expected votes
>0 Resources configured.
>
>
>Online: [ server2 ]
>
>
>
>
>
>
>The only warnings and errors in the logfile are:
>
>[root@server1 ~]# cat /var/log/cluster/corosync.log | egrep "warning|error"
>Jan 30 10:25:59 [1608] server1 crmd: warning: do_log: FSA: Input
>I_DC_TIMEOUT from crm_timer_popped() received in state S_PENDING
>Jan 30 10:25:59 [1607] server1 pengine: warning: cluster_status: We do
>not have quorum - fencing and resource management disabled
>Jan 30 10:28:25 [1525] server1 corosync debug [QUORUM] getinfo response
>error: 1
>Jan 30 10:40:59 [1607] server1 pengine: warning: cluster_status: We do
>not have quorum - fencing and resource management disabled
>
>
>root@server2 corosync]# cat /var/log/cluster/corosync.log | egrep
>"warning|error"
>Jan 30 10:27:18 [1458] server2 crmd: warning: do_log: FSA: Input
>I_DC_TIMEOUT from crm_timer_popped() received in state S_PENDING
>Jan 30 10:27:18 [1457] server2 pengine: warning: cluster_status: We do
>not have quorum - fencing and resource management disabled
>Jan 30 10:29:19 [1349] server2 corosync debug [QUORUM] getinfo response
>error: 1
>Jan 30 10:42:18 [1457] server2 pengine: warning: cluster_status: We do
>not have quorum - fencing and resource management disabled
>Jan 30 10:44:36 [1349] server2 corosync debug [QUORUM] getinfo response
>error: 1
>
>
>
>
>We have installed the following packages:
>
>corosync-2.2.0-1.fc18.i686
>corosynclib-2.2.0-1.fc18.i686
>drbd-bash-completion-8.3.13-1.fc18.i686
>drbd-pacemaker-8.3.13-1.fc18.i686
>drbd-utils-8.3.13-1.fc18.i686
>pacemaker-1.1.8-3.fc18.i686
>pacemaker-cli-1.1.8-3.fc18.i686
>pacemaker-cluster-libs-1.1.8-3.fc18.i686
>pacemaker-libs-1.1.8-3.fc18.i686
>pcs-0.9.27-3.fc18.i686
>
>
>
>Firewalls are disabled, Pinging and SSH communication is working without any
>problems.
>
>With best regards
>
>_______________________________________________
>Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
>http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>
>Project Home: http://www.clusterlabs.org
>Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>Bugs: http://bugs.clusterlabs.org
>
_______________________________________________
Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
http://oss.clusterlabs.org/mailman/listinfo/pacemaker
Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org