Re: [PVE-User] SPAN port on bridge - not working in 3.x
Leftovers. Previous network card required that tuning, otherwise mirrored frames would be duplicated. Thanks for notifying. On Mon, Jul 8, 2013, at 09:10 AM, Alexandre Kouznetsov wrote: > Hello. > > El 08/07/13 08:44, Lex Rivera escribió: > > Well silly me. brctl location changed from /usr/sbin in squeeze to just > > /sbin in wheezy. Sorry for bothering, it's working now :) > > > > On Mon, Jul 8, 2013, at 06:32 AM, Lex Rivera wrote: > >> Hello. > >> I have machine which have span port from physical switch for VMs. > >> > >> Interface config: > >> > >> auto eth3 > >> allow-hotplug eth3 > >> iface eth3 inet manual > >> up /sbin/ifconfig $IFACE up > >> downp /sbin/ifconfig $IFACE down > >> > >> post-up ethtool -K eth4 sg off > >> post-up ethtool -K eth4 tso off > >> post-up ethtool -K eth4 gso off > >> post-up ethtool -K eth4 gro off > >> post-up ethtool -K eth4 lro off > >> post-up ethtool -K eth4 tco off > > Good it was solved. > Why are you running ethtool on eth4, while configuring eth3 ? > > -- > Alexandre Kouznetsov > > ___ > pve-user mailing list > pve-user@pve.proxmox.com > http://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-user ___ pve-user mailing list pve-user@pve.proxmox.com http://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-user
Re: [PVE-User] SPAN port on bridge - not working in 3.x
Hello. El 08/07/13 08:44, Lex Rivera escribió: Well silly me. brctl location changed from /usr/sbin in squeeze to just /sbin in wheezy. Sorry for bothering, it's working now :) On Mon, Jul 8, 2013, at 06:32 AM, Lex Rivera wrote: Hello. I have machine which have span port from physical switch for VMs. Interface config: auto eth3 allow-hotplug eth3 iface eth3 inet manual up /sbin/ifconfig $IFACE up downp /sbin/ifconfig $IFACE down post-up ethtool -K eth4 sg off post-up ethtool -K eth4 tso off post-up ethtool -K eth4 gso off post-up ethtool -K eth4 gro off post-up ethtool -K eth4 lro off post-up ethtool -K eth4 tco off Good it was solved. Why are you running ethtool on eth4, while configuring eth3 ? -- Alexandre Kouznetsov ___ pve-user mailing list pve-user@pve.proxmox.com http://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-user
Re: [PVE-User] SPAN port on bridge - not working in 3.x
Well silly me. brctl location changed from /usr/sbin in squeeze to just /sbin in wheezy. Sorry for bothering, it's working now :) On Mon, Jul 8, 2013, at 06:32 AM, Lex Rivera wrote: > Hello. > I have machine which have span port from physical switch for VMs. > > Interface config: > > auto eth3 > allow-hotplug eth3 > iface eth3 inet manual > up /sbin/ifconfig $IFACE up > downp /sbin/ifconfig $IFACE down > > post-up ethtool -K eth4 sg off > post-up ethtool -K eth4 tso off > post-up ethtool -K eth4 gso off > post-up ethtool -K eth4 gro off > post-up ethtool -K eth4 lro off > post-up ethtool -K eth4 tco off > > auto vmbrspan > iface vmbrspan inet manual > bridge_ports eth3 > bridge_stp off > bridge_fd 0 > post-up /usr/sbin/brctl setageing $IFACE 0 > > pveversion: > > lex@galleon > pveversion -v > pve-manager: 3.0-23 (pve-manager/3.0/957f0862) > running kernel: 2.6.32-20-pve > proxmox-ve-2.6.32: 3.0-100 > pve-kernel-2.6.32-20-pve: 2.6.32-100 > pve-kernel-2.6.32-19-pve: 2.6.32-96 > pve-kernel-2.6.32-18-pve: 2.6.32-88 > lvm2: 2.02.95-pve3 > clvm: 2.02.95-pve3 > corosync-pve: 1.4.5-1 > openais-pve: 1.1.4-3 > libqb0: 0.11.1-2 > redhat-cluster-pve: 3.2.0-2 > resource-agents-pve: 3.9.2-4 > fence-agents-pve: 4.0.0-1 > pve-cluster: 3.0-4 > qemu-server: 3.0-20 > pve-firmware: 1.0-22 > libpve-common-perl: 3.0-4 > libpve-access-control: 3.0-4 > libpve-storage-perl: 3.0-8 > vncterm: 1.1-4 > vzctl: 4.0-1pve3 > vzprocps: 2.0.11-2 > vzquota: 3.1-2 > pve-qemu-kvm: 1.4-13 > ksm-control-daemon: 1.1-1 > > It worked in proxmox 2.3.x, but after upgrade to 3.0-23 i can't see span > traffic from inside VMs. > I can see mirrored data on host node, both on eth3 and on vmbrspan, but > not on tapXXXiY interface. > Any ideas? ___ pve-user mailing list pve-user@pve.proxmox.com http://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-user
[PVE-User] SPAN port on bridge - not working in 3.x
Hello. I have machine which have span port from physical switch for VMs. Interface config: auto eth3 allow-hotplug eth3 iface eth3 inet manual up /sbin/ifconfig $IFACE up downp /sbin/ifconfig $IFACE down post-up ethtool -K eth4 sg off post-up ethtool -K eth4 tso off post-up ethtool -K eth4 gso off post-up ethtool -K eth4 gro off post-up ethtool -K eth4 lro off post-up ethtool -K eth4 tco off auto vmbrspan iface vmbrspan inet manual bridge_ports eth3 bridge_stp off bridge_fd 0 post-up /usr/sbin/brctl setageing $IFACE 0 pveversion: lex@galleon > pveversion -v pve-manager: 3.0-23 (pve-manager/3.0/957f0862) running kernel: 2.6.32-20-pve proxmox-ve-2.6.32: 3.0-100 pve-kernel-2.6.32-20-pve: 2.6.32-100 pve-kernel-2.6.32-19-pve: 2.6.32-96 pve-kernel-2.6.32-18-pve: 2.6.32-88 lvm2: 2.02.95-pve3 clvm: 2.02.95-pve3 corosync-pve: 1.4.5-1 openais-pve: 1.1.4-3 libqb0: 0.11.1-2 redhat-cluster-pve: 3.2.0-2 resource-agents-pve: 3.9.2-4 fence-agents-pve: 4.0.0-1 pve-cluster: 3.0-4 qemu-server: 3.0-20 pve-firmware: 1.0-22 libpve-common-perl: 3.0-4 libpve-access-control: 3.0-4 libpve-storage-perl: 3.0-8 vncterm: 1.1-4 vzctl: 4.0-1pve3 vzprocps: 2.0.11-2 vzquota: 3.1-2 pve-qemu-kvm: 1.4-13 ksm-control-daemon: 1.1-1 It worked in proxmox 2.3.x, but after upgrade to 3.0-23 i can't see span traffic from inside VMs. I can see mirrored data on host node, both on eth3 and on vmbrspan, but not on tapXXXiY interface. Any ideas? ___ pve-user mailing list pve-user@pve.proxmox.com http://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-user
Re: [PVE-User] Cman crash problem
>>It appeared after the network team changed network active equipments in >>the building (but this might not be the origin of the problem). Hi, what is the previous and new equipments ? (I have had some cisco problem in the past). - Mail original - De: "Jonathan Schaeffer" À: pve-user@pve.proxmox.com Envoyé: Lundi 8 Juillet 2013 12:56:48 Objet: [PVE-User] Cman crash problem Hi all, I'm experiencing a serious problem on our 4 nodes cluster (PVE 3.0). It appeared after the network team changed network active equipments in the building (but this might not be the origin of the problem). The symptoms are : - The nodes appear in red on the web gui, except the one hosting the web service IP - The VM, while still running correctly, do not show any information (running, rrd graphs, etc) - clustat shows nodes as "online" - some nodes seems to have been fenced (while not restarted) (see log extracts : barbossa_fenced.log and jim_fenced.log) - /var/log/cluster/corosync.log shows LOT of messages : Jul 08 07:06:49 corosync [TOTEM ] Retransmit List: 13f54a 13f54b 13f54c 13f54d 13f54e 13f54f 13f550 13f551 13f552 13f553 13f554 13f555 13f556 13f557 13f558 13f559 13f55a 13f55b 13f55c 13f55d 13f55e If I restart one node, the fencing is going to happen, other nodes will reboot and all the VMs hosted allong with them. I don't want this to happen. I can provide more logs if necessary. Do you have an idea to help me understand what is going on here ? Thanks, Jonathan barbossa_fenced.log : Jul 03 12:07:21 fenced fencing deferred to jim Jul 03 13:45:40 fenced receive_start 1:15 add node with started_count 8 Jul 03 13:45:40 fenced receive_start 2:11 add node with started_count 4 Jul 03 13:45:40 fenced receive_start 3:7 add node with started_count 1 Jul 04 00:29:35 fenced receive_start 1:16 add node with started_count 8 Jul 04 00:29:35 fenced receive_start 3:8 add node with started_count 1 Jul 04 00:38:31 fenced receive_start 2:17 add node with started_count 4 Jul 04 00:38:31 fenced receive_start 3:13 add node with started_count 1 Jul 04 00:38:31 fenced receive_start 1:21 add node with started_count 8 Jul 04 10:44:12 fenced receive_start 1:22 add node with started_count 8 Jul 04 10:44:12 fenced receive_start 3:14 add node with started_count 1 Jul 04 10:44:24 fenced receive_start 1:23 add node with started_count 8 Jul 04 10:44:24 fenced telling cman to remove nodeid 2 from cluster jim_fenced.log : Jul 03 12:07:21 fenced fencing node longjohn Jul 03 12:07:32 fenced fence longjohn success Jul 03 13:45:40 fenced receive_start 5:13 add node with started_count 6 Jul 03 13:45:40 fenced receive_start 2:11 add node with started_count 4 Jul 03 13:45:40 fenced receive_start 3:7 add node with started_count 1 Jul 04 00:29:35 fenced receive_start 3:8 add node with started_count 1 Jul 04 00:29:35 fenced receive_start 5:14 add node with started_count 6 Jul 04 00:38:31 fenced receive_start 2:17 add node with started_count 4 Jul 04 00:38:31 fenced receive_start 3:13 add node with started_count 1 Jul 04 00:38:31 fenced receive_start 5:19 add node with started_count 6 Jul 04 10:44:12 fenced receive_start 5:20 add node with started_count 6 Jul 04 10:44:12 fenced receive_start 3:14 add node with started_count 1 Jul 04 10:44:24 fenced telling cman to remove nodeid 2 from cluster Jul 04 10:44:24 fenced receive_start 2:23 add node with started_count 4 Jul 04 10:44:24 fenced receive_start 3:15 add node with started_count 1 Jul 04 10:44:24 fenced receive_start 5:21 add node with started_count 6 Jul 04 10:44:46 fenced receive_start 5:22 add node with started_count 6 Jul 04 10:44:46 fenced receive_start 3:16 add node with started_count 1 longjohn_fenced.log : Jul 03 09:47:12 fenced fenced 1352871249 started Jul 03 11:28:46 fenced cluster is down, exiting Jul 03 11:28:46 fenced daemon cpg_dispatch error 2 Jul 03 12:11:43 fenced fenced 1364188437 started Jul 03 13:45:40 fenced receive_start 5:13 add node with started_count 6 Jul 03 13:45:40 fenced receive_start 1:15 add node with started_count 8 Jul 03 13:45:40 fenced receive_start 2:11 add node with started_count 4 Jul 04 00:29:35 fenced receive_start 1:16 add node with started_count 8 Jul 04 00:29:35 fenced receive_start 5:14 add node with started_count 6 Jul 04 00:38:31 fenced receive_start 2:17 add node with started_count 4 Jul 04 00:38:31 fenced receive_start 1:21 add node with started_count 8 Jul 04 00:38:31 fenced receive_start 5:19 add node with started_count 6 Jul 04 10:44:12 fenced receive_start 1:22 add node with started_count 8 Jul 04 10:44:12 fenced receive_start 5:20 add node with started_count 6 Jul 04 10:44:24 fenced receive_start 1:23 add node with started_count 8 Jul 04 10:44:24 fenced telling cman to remove nodeid 2 from cluster Jul 04 10:44:24 fenced receive_start 2:23 add node with started_count 4 Jul 04 10:44:24 fenced receive_start 5:21 add node with started_count 6 Jul 04 10:44:46 f
[PVE-User] Cman crash problem
Hi all, I'm experiencing a serious problem on our 4 nodes cluster (PVE 3.0). It appeared after the network team changed network active equipments in the building (but this might not be the origin of the problem). The symptoms are : - The nodes appear in red on the web gui, except the one hosting the web service IP - The VM, while still running correctly, do not show any information (running, rrd graphs, etc) - clustat shows nodes as "online" - some nodes seems to have been fenced (while not restarted) (see log extracts : barbossa_fenced.log and jim_fenced.log) - /var/log/cluster/corosync.log shows LOT of messages : Jul 08 07:06:49 corosync [TOTEM ] Retransmit List: 13f54a 13f54b 13f54c 13f54d 13f54e 13f54f 13f550 13f551 13f552 13f553 13f554 13f555 13f556 13f557 13f558 13f559 13f55a 13f55b 13f55c 13f55d 13f55e If I restart one node, the fencing is going to happen, other nodes will reboot and all the VMs hosted allong with them. I don't want this to happen. I can provide more logs if necessary. Do you have an idea to help me understand what is going on here ? Thanks, Jonathan barbossa_fenced.log : Jul 03 12:07:21 fenced fencing deferred to jim Jul 03 13:45:40 fenced receive_start 1:15 add node with started_count 8 Jul 03 13:45:40 fenced receive_start 2:11 add node with started_count 4 Jul 03 13:45:40 fenced receive_start 3:7 add node with started_count 1 Jul 04 00:29:35 fenced receive_start 1:16 add node with started_count 8 Jul 04 00:29:35 fenced receive_start 3:8 add node with started_count 1 Jul 04 00:38:31 fenced receive_start 2:17 add node with started_count 4 Jul 04 00:38:31 fenced receive_start 3:13 add node with started_count 1 Jul 04 00:38:31 fenced receive_start 1:21 add node with started_count 8 Jul 04 10:44:12 fenced receive_start 1:22 add node with started_count 8 Jul 04 10:44:12 fenced receive_start 3:14 add node with started_count 1 Jul 04 10:44:24 fenced receive_start 1:23 add node with started_count 8 Jul 04 10:44:24 fenced telling cman to remove nodeid 2 from cluster jim_fenced.log : Jul 03 12:07:21 fenced fencing node longjohn Jul 03 12:07:32 fenced fence longjohn success Jul 03 13:45:40 fenced receive_start 5:13 add node with started_count 6 Jul 03 13:45:40 fenced receive_start 2:11 add node with started_count 4 Jul 03 13:45:40 fenced receive_start 3:7 add node with started_count 1 Jul 04 00:29:35 fenced receive_start 3:8 add node with started_count 1 Jul 04 00:29:35 fenced receive_start 5:14 add node with started_count 6 Jul 04 00:38:31 fenced receive_start 2:17 add node with started_count 4 Jul 04 00:38:31 fenced receive_start 3:13 add node with started_count 1 Jul 04 00:38:31 fenced receive_start 5:19 add node with started_count 6 Jul 04 10:44:12 fenced receive_start 5:20 add node with started_count 6 Jul 04 10:44:12 fenced receive_start 3:14 add node with started_count 1 Jul 04 10:44:24 fenced telling cman to remove nodeid 2 from cluster Jul 04 10:44:24 fenced receive_start 2:23 add node with started_count 4 Jul 04 10:44:24 fenced receive_start 3:15 add node with started_count 1 Jul 04 10:44:24 fenced receive_start 5:21 add node with started_count 6 Jul 04 10:44:46 fenced receive_start 5:22 add node with started_count 6 Jul 04 10:44:46 fenced receive_start 3:16 add node with started_count 1 longjohn_fenced.log : Jul 03 09:47:12 fenced fenced 1352871249 started Jul 03 11:28:46 fenced cluster is down, exiting Jul 03 11:28:46 fenced daemon cpg_dispatch error 2 Jul 03 12:11:43 fenced fenced 1364188437 started Jul 03 13:45:40 fenced receive_start 5:13 add node with started_count 6 Jul 03 13:45:40 fenced receive_start 1:15 add node with started_count 8 Jul 03 13:45:40 fenced receive_start 2:11 add node with started_count 4 Jul 04 00:29:35 fenced receive_start 1:16 add node with started_count 8 Jul 04 00:29:35 fenced receive_start 5:14 add node with started_count 6 Jul 04 00:38:31 fenced receive_start 2:17 add node with started_count 4 Jul 04 00:38:31 fenced receive_start 1:21 add node with started_count 8 Jul 04 00:38:31 fenced receive_start 5:19 add node with started_count 6 Jul 04 10:44:12 fenced receive_start 1:22 add node with started_count 8 Jul 04 10:44:12 fenced receive_start 5:20 add node with started_count 6 Jul 04 10:44:24 fenced receive_start 1:23 add node with started_count 8 Jul 04 10:44:24 fenced telling cman to remove nodeid 2 from cluster Jul 04 10:44:24 fenced receive_start 2:23 add node with started_count 4 Jul 04 10:44:24 fenced receive_start 5:21 add node with started_count 6 Jul 04 10:44:46 fenced receive_start 5:22 add node with started_count 6 Jul 04 10:44:46 fenced receive_start 1:24 add node with started_count 8 flint_fenced.log : Jul 03 11:18:30 fenced fenced 1364188437 started Jul 03 12:07:21 fenced fencing deferred to jim Jul 03 13:45:40 fenced receive_start 5:13 add node with started_count 6 Jul 03 13:45:40 fenced receive_start 1:15 add node with started_count 8 Jul 03 13:45:40 fenced receive_start 3:7 add node with started_count 1 J