hi stefan,
it seems that's more stable but after 2 minute the issue is back again.
hopefully isn't a bug because it can reproduce it
node2 sents only unicast at sequenz 256...
node1
omping 10.0.0.22 10.0.0.21
10.0.0.22 : unicast, seq=257, size=69 bytes, dist=0, time=0.666ms
10.0.0.22 : multicast, seq=257, size=69 bytes, dist=0, time=0.677ms
10.0.0.22 : unicast, seq=258, size=69 bytes, dist=0, time=0.600ms
10.0.0.22 : multicast, seq=258, size=69 bytes, dist=0, time=0.610ms
10.0.0.22 : unicast, seq=259, size=69 bytes, dist=0, time=0.693ms
10.0.0.22 : multicast, seq=259, size=69 bytes, dist=0, time=0.702ms
10.0.0.22 : unicast, seq=260, size=69 bytes, dist=0, time=0.674ms
10.0.0.22 : multicast, seq=260, size=69 bytes, dist=0, time=0.685ms
10.0.0.22 : unicast, seq=261, size=69 bytes, dist=0, time=0.658ms
10.0.0.22 : multicast, seq=261, size=69 bytes, dist=0, time=0.669ms
10.0.0.22 : unicast, seq=262, size=69 bytes, dist=0, time=0.834ms
10.0.0.22 : multicast, seq=262, size=69 bytes, dist=0, time=0.845ms
10.0.0.22 : unicast, seq=263, size=69 bytes, dist=0, time=0.666ms
10.0.0.22 : multicast, seq=263, size=69 bytes, dist=0, time=0.677ms
10.0.0.22 : unicast, seq=264, size=69 bytes, dist=0, time=0.675ms
10.0.0.22 : multicast, seq=264, size=69 bytes, dist=0, time=0.687ms
10.0.0.22 : waiting for response msg
10.0.0.22 : server told us to stop
^C
10.0.0.22 : unicast, xmt/rcv/%loss = 264/264/0%, min/avg/max/std-dev =
0.542/0.663/0.860/0.035
10.0.0.22 : multicast, xmt/rcv/%loss = 264/264/0%, min/avg/max/std-dev =
0.553/0.675/0.876/0.035
node2:
10.0.0.21 : multicast, seq=251, size=69 bytes, dist=0, time=0.703ms
10.0.0.21 : unicast, seq=252, size=69 bytes, dist=0, time=0.714ms
10.0.0.21 : multicast, seq=252, size=69 bytes, dist=0, time=0.725ms
10.0.0.21 : unicast, seq=253, size=69 bytes, dist=0, time=0.662ms
10.0.0.21 : multicast, seq=253, size=69 bytes, dist=0, time=0.672ms
10.0.0.21 : unicast, seq=254, size=69 bytes, dist=0, time=0.662ms
10.0.0.21 : multicast, seq=254, size=69 bytes, dist=0, time=0.673ms
10.0.0.21 : unicast, seq=255, size=69 bytes, dist=0, time=0.668ms
10.0.0.21 : multicast, seq=255, size=69 bytes, dist=0, time=0.679ms
10.0.0.21 : unicast, seq=256, size=69 bytes, dist=0, time=0.674ms
10.0.0.21 : multicast, seq=256, size=69 bytes, dist=0, time=0.687ms
10.0.0.21 : unicast, seq=257, size=69 bytes, dist=0, time=0.618ms
10.0.0.21 : unicast, seq=258, size=69 bytes, dist=0, time=0.659ms
10.0.0.21 : unicast, seq=259, size=69 bytes, dist=0, time=0.705ms
10.0.0.21 : unicast, seq=260, size=69 bytes, dist=0, time=0.682ms
10.0.0.21 : unicast, seq=261, size=69 bytes, dist=0, time=0.760ms
10.0.0.21 : unicast, seq=262, size=69 bytes, dist=0, time=0.665ms
10.0.0.21 : unicast, seq=263, size=69 bytes, dist=0, time=0.711ms
^C
10.0.0.21 : unicast, xmt/rcv/%loss = 263/263/0%, min/avg/max/std-dev =
0.539/0.661/0.772/0.037
10.0.0.21 : multicast, xmt/rcv/%loss = 263/256/2%, min/avg/max/std-dev =
0.583/0.674/0.786/0.033
2014-02-14 9:59 GMT+01:00 Stefan Bauer <stefan.ba...@cubewerk.de>:
you have to disable all offloading features (rx, tx, tso...)
Mit freundlichen Grüßen
Stefan Bauer
--
Cubewerk GmbH
Herzog-Otto-Straße 32
83308 Trostberg
08621 - 99 60 237
HRB 22195 AG Traunstein
GF Stefan Bauer
Am 14.02.2014 um 09:40 schrieb "Beo Banks" <beo.ba...@googlemail.com>:
ethtool -K eth0 tx off
ethtool -K eth1 tx off
same result...retransmit issue
2014-02-14 9:31 GMT+01:00 Beo Banks <beo.ba...@googlemail.com>:
i have also try
"No more delay when you disable multicast snooping on the host:"
echo 0 > /sys/devices/virtual/net/br1/bridge/multicast_router
echo 0 > /sys/devices/virtual/net/br1/bridge/multicast_snooping
2014-02-14 9:28 GMT+01:00 Beo Banks <beo.ba...@googlemail.com>:
@jan and stefan
must i set it for both bridges
eth1 (br1) eth0 (br0) on the host or guest ?
2014-02-14 9:06 GMT+01:00 Jan Friesse <jfrie...@redhat.com>:
Beo,
do you experiencing cluster split? If answer is no, then you don't need
to do anything. Maybe network buffer is just filled. But, if answer is yes,
try reduce mtu size (netmtu in configuration) to value like 1000.
Regards,
Honza
Beo Banks napsal(a):
Hi,
i have a fresh 2 node cluster (kvm host1 -> guest = nodeA | kvm host2
->
guest = NodeB) and it seems to work but from time to time i have a lot
of
errors like
Feb 13 13:41:04 corosync [TOTEM ] Retransmit List: 196 198 184 185 186
187
188 189 18a 18b 18c 18d 18e 18f 190 191 192 193 194 195 197 199
Feb 13 13:41:04 corosync [TOTEM ] Retransmit List: 197 199 184 185 186
187
188 189 18a 18b 18c 18d 18e 18f 190 191 192 193 194 195 196 198
Feb 13 13:41:04 corosync [TOTEM ] Retransmit List: 196 198 184 185 186
187
188 189 18a 18b 18c 18d 18e 18f 190 191 192 193 194 195 197 199
Feb 13 13:41:04 corosync [TOTEM ] Retransmit List: 197 199 184 185 186
187
188 189 18a 18b 18c 18d 18e 18f 190 191 192 193 194 195 196 198
Feb 13 13:41:04 corosync [TOTEM ] Retransmit List: 196 198 184 185 186
187
188 189 18a 18b 18c 18d 18e 18f 190 191 192 193 194 195 197 199
Feb 13 13:41:04 corosync [TOTEM ] Retransmit List: 197 199 184 185 186
187
188 189 18a 18b 18c 18d 18e 18f 190 191 192 193 194 195 196 198
i used the newest rhel 6.5 version.
i have also already try solve the issue with
echo 1 > /sys/class/net/virbr0/bridge/multicast_querier (host system)
but no chance...
i have disable iptables,selinux..same issue
how can solve it?
thanks beo
_______________________________________________
Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
http://oss.clusterlabs.org/mailman/listinfo/pacemaker
Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/
doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org
_______________________________________________
Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
http://oss.clusterlabs.org/mailman/listinfo/pacemaker
Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/
doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org
_______________________________________________
Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
http://oss.clusterlabs.org/mailman/listinfo/pacemaker
Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org
_______________________________________________
Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
http://oss.clusterlabs.org/mailman/listinfo/pacemaker
Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org
_______________________________________________
Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
http://oss.clusterlabs.org/mailman/listinfo/pacemaker
Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org