Re: [PVE-User] SPAN port on bridge - not working in 3.x

2013-07-08 Thread Lex Rivera
Leftovers. Previous network card required that tuning, otherwise
mirrored frames would be duplicated.
Thanks for notifying.

On Mon, Jul 8, 2013, at 09:10 AM, Alexandre Kouznetsov wrote:
> Hello.
> 
> El 08/07/13 08:44, Lex Rivera escribió:
> > Well silly me. brctl location changed from /usr/sbin in squeeze to just
> > /sbin in wheezy. Sorry for bothering, it's working now :)
> >
> > On Mon, Jul 8, 2013, at 06:32 AM, Lex Rivera wrote:
> >> Hello.
> >> I have machine which have span port from physical switch for VMs.
> >>
> >> Interface config:
> >> 
> >> auto eth3
> >> allow-hotplug eth3
> >> iface eth3 inet manual
> >>  up /sbin/ifconfig $IFACE up
> >>  downp /sbin/ifconfig $IFACE down
> >>
> >>  post-up ethtool -K eth4 sg off
> >>  post-up ethtool -K eth4 tso off
> >>  post-up ethtool -K eth4 gso off
> >>  post-up ethtool -K eth4 gro off
> >>  post-up ethtool -K eth4 lro off
> >>  post-up ethtool -K eth4 tco off
> 
> Good it was solved.
> Why are you running ethtool on eth4, while configuring eth3 ?
> 
> -- 
> Alexandre Kouznetsov
> 
> ___
> pve-user mailing list
> pve-user@pve.proxmox.com
> http://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-user
___
pve-user mailing list
pve-user@pve.proxmox.com
http://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-user


Re: [PVE-User] SPAN port on bridge - not working in 3.x

2013-07-08 Thread Alexandre Kouznetsov

Hello.

El 08/07/13 08:44, Lex Rivera escribió:

Well silly me. brctl location changed from /usr/sbin in squeeze to just
/sbin in wheezy. Sorry for bothering, it's working now :)

On Mon, Jul 8, 2013, at 06:32 AM, Lex Rivera wrote:

Hello.
I have machine which have span port from physical switch for VMs.

Interface config:

auto eth3
allow-hotplug eth3
iface eth3 inet manual
 up /sbin/ifconfig $IFACE up
 downp /sbin/ifconfig $IFACE down

 post-up ethtool -K eth4 sg off
 post-up ethtool -K eth4 tso off
 post-up ethtool -K eth4 gso off
 post-up ethtool -K eth4 gro off
 post-up ethtool -K eth4 lro off
 post-up ethtool -K eth4 tco off


Good it was solved.
Why are you running ethtool on eth4, while configuring eth3 ?

--
Alexandre Kouznetsov

___
pve-user mailing list
pve-user@pve.proxmox.com
http://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-user


Re: [PVE-User] SPAN port on bridge - not working in 3.x

2013-07-08 Thread Lex Rivera
Well silly me. brctl location changed from /usr/sbin in squeeze to just
/sbin in wheezy. Sorry for bothering, it's working now :)

On Mon, Jul 8, 2013, at 06:32 AM, Lex Rivera wrote:
> Hello. 
> I have machine which have span port from physical switch for VMs.
> 
> Interface config:
> 
> auto eth3
> allow-hotplug eth3
> iface eth3 inet manual
> up /sbin/ifconfig $IFACE up
> downp /sbin/ifconfig $IFACE down
> 
> post-up ethtool -K eth4 sg off
> post-up ethtool -K eth4 tso off
> post-up ethtool -K eth4 gso off
> post-up ethtool -K eth4 gro off
> post-up ethtool -K eth4 lro off
> post-up ethtool -K eth4 tco off
> 
> auto vmbrspan
> iface vmbrspan inet manual
> bridge_ports eth3
> bridge_stp off
> bridge_fd 0
> post-up /usr/sbin/brctl setageing $IFACE 0
> 
> pveversion:
> 
> lex@galleon > pveversion -v
> pve-manager: 3.0-23 (pve-manager/3.0/957f0862)
> running kernel: 2.6.32-20-pve
> proxmox-ve-2.6.32: 3.0-100
> pve-kernel-2.6.32-20-pve: 2.6.32-100
> pve-kernel-2.6.32-19-pve: 2.6.32-96
> pve-kernel-2.6.32-18-pve: 2.6.32-88
> lvm2: 2.02.95-pve3
> clvm: 2.02.95-pve3
> corosync-pve: 1.4.5-1
> openais-pve: 1.1.4-3
> libqb0: 0.11.1-2
> redhat-cluster-pve: 3.2.0-2
> resource-agents-pve: 3.9.2-4
> fence-agents-pve: 4.0.0-1
> pve-cluster: 3.0-4
> qemu-server: 3.0-20
> pve-firmware: 1.0-22
> libpve-common-perl: 3.0-4
> libpve-access-control: 3.0-4
> libpve-storage-perl: 3.0-8
> vncterm: 1.1-4
> vzctl: 4.0-1pve3
> vzprocps: 2.0.11-2
> vzquota: 3.1-2
> pve-qemu-kvm: 1.4-13
> ksm-control-daemon: 1.1-1
> 
> It worked in proxmox 2.3.x, but after upgrade to 3.0-23 i can't see span
> traffic from inside VMs.
> I can see mirrored data on host node, both on eth3 and on vmbrspan, but
> not on tapXXXiY interface.
> Any ideas?
___
pve-user mailing list
pve-user@pve.proxmox.com
http://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-user


[PVE-User] SPAN port on bridge - not working in 3.x

2013-07-08 Thread Lex Rivera
Hello. 
I have machine which have span port from physical switch for VMs.

Interface config:

auto eth3
allow-hotplug eth3
iface eth3 inet manual
up /sbin/ifconfig $IFACE up
downp /sbin/ifconfig $IFACE down

post-up ethtool -K eth4 sg off
post-up ethtool -K eth4 tso off
post-up ethtool -K eth4 gso off
post-up ethtool -K eth4 gro off
post-up ethtool -K eth4 lro off
post-up ethtool -K eth4 tco off

auto vmbrspan
iface vmbrspan inet manual
bridge_ports eth3
bridge_stp off
bridge_fd 0
post-up /usr/sbin/brctl setageing $IFACE 0

pveversion:

lex@galleon > pveversion -v
pve-manager: 3.0-23 (pve-manager/3.0/957f0862)
running kernel: 2.6.32-20-pve
proxmox-ve-2.6.32: 3.0-100
pve-kernel-2.6.32-20-pve: 2.6.32-100
pve-kernel-2.6.32-19-pve: 2.6.32-96
pve-kernel-2.6.32-18-pve: 2.6.32-88
lvm2: 2.02.95-pve3
clvm: 2.02.95-pve3
corosync-pve: 1.4.5-1
openais-pve: 1.1.4-3
libqb0: 0.11.1-2
redhat-cluster-pve: 3.2.0-2
resource-agents-pve: 3.9.2-4
fence-agents-pve: 4.0.0-1
pve-cluster: 3.0-4
qemu-server: 3.0-20
pve-firmware: 1.0-22
libpve-common-perl: 3.0-4
libpve-access-control: 3.0-4
libpve-storage-perl: 3.0-8
vncterm: 1.1-4
vzctl: 4.0-1pve3
vzprocps: 2.0.11-2
vzquota: 3.1-2
pve-qemu-kvm: 1.4-13
ksm-control-daemon: 1.1-1

It worked in proxmox 2.3.x, but after upgrade to 3.0-23 i can't see span
traffic from inside VMs.
I can see mirrored data on host node, both on eth3 and on vmbrspan, but
not on tapXXXiY interface.
Any ideas?
___
pve-user mailing list
pve-user@pve.proxmox.com
http://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-user


Re: [PVE-User] Cman crash problem

2013-07-08 Thread Alexandre DERUMIER
>>It appeared after the network team changed network active equipments in 
>>the building (but this might not be the origin of the problem).

Hi, what is the previous and new equipments ? (I have had some cisco problem in 
the past).


- Mail original - 

De: "Jonathan Schaeffer"  
À: pve-user@pve.proxmox.com 
Envoyé: Lundi 8 Juillet 2013 12:56:48 
Objet: [PVE-User] Cman crash problem 

Hi all, 

I'm experiencing a serious problem on our 4 nodes cluster (PVE 3.0). 

It appeared after the network team changed network active equipments in 
the building (but this might not be the origin of the problem). 

The symptoms are : 

- The nodes appear in red on the web gui, except the one hosting the web 
service IP 
- The VM, while still running correctly, do not show any information 
(running, rrd graphs, etc) 

- clustat shows nodes as "online" 
- some nodes seems to have been fenced (while not restarted) 
(see log extracts : barbossa_fenced.log and jim_fenced.log) 

- /var/log/cluster/corosync.log shows LOT of messages : 
Jul 08 07:06:49 corosync [TOTEM ] Retransmit List: 13f54a 13f54b 13f54c 
13f54d 13f54e 13f54f 13f550 13f551 13f552 13f553 13f554 13f555 13f556 
13f557 13f558 13f559 13f55a 13f55b 13f55c 13f55d 13f55e 

If I restart one node, the fencing is going to happen, other nodes will 
reboot and all the VMs hosted allong with them. I don't want this to happen. 

I can provide more logs if necessary. Do you have an idea to help me 
understand what is going on here ? 

Thanks, 

Jonathan 


barbossa_fenced.log : 
Jul 03 12:07:21 fenced fencing deferred to jim 
Jul 03 13:45:40 fenced receive_start 1:15 add node with started_count 8 
Jul 03 13:45:40 fenced receive_start 2:11 add node with started_count 4 
Jul 03 13:45:40 fenced receive_start 3:7 add node with started_count 1 
Jul 04 00:29:35 fenced receive_start 1:16 add node with started_count 8 
Jul 04 00:29:35 fenced receive_start 3:8 add node with started_count 1 
Jul 04 00:38:31 fenced receive_start 2:17 add node with started_count 4 
Jul 04 00:38:31 fenced receive_start 3:13 add node with started_count 1 
Jul 04 00:38:31 fenced receive_start 1:21 add node with started_count 8 
Jul 04 10:44:12 fenced receive_start 1:22 add node with started_count 8 
Jul 04 10:44:12 fenced receive_start 3:14 add node with started_count 1 
Jul 04 10:44:24 fenced receive_start 1:23 add node with started_count 8 
Jul 04 10:44:24 fenced telling cman to remove nodeid 2 from cluster 


jim_fenced.log : 
Jul 03 12:07:21 fenced fencing node longjohn 
Jul 03 12:07:32 fenced fence longjohn success 
Jul 03 13:45:40 fenced receive_start 5:13 add node with started_count 6 
Jul 03 13:45:40 fenced receive_start 2:11 add node with started_count 4 
Jul 03 13:45:40 fenced receive_start 3:7 add node with started_count 1 
Jul 04 00:29:35 fenced receive_start 3:8 add node with started_count 1 
Jul 04 00:29:35 fenced receive_start 5:14 add node with started_count 6 
Jul 04 00:38:31 fenced receive_start 2:17 add node with started_count 4 
Jul 04 00:38:31 fenced receive_start 3:13 add node with started_count 1 
Jul 04 00:38:31 fenced receive_start 5:19 add node with started_count 6 
Jul 04 10:44:12 fenced receive_start 5:20 add node with started_count 6 
Jul 04 10:44:12 fenced receive_start 3:14 add node with started_count 1 
Jul 04 10:44:24 fenced telling cman to remove nodeid 2 from cluster 
Jul 04 10:44:24 fenced receive_start 2:23 add node with started_count 4 
Jul 04 10:44:24 fenced receive_start 3:15 add node with started_count 1 
Jul 04 10:44:24 fenced receive_start 5:21 add node with started_count 6 
Jul 04 10:44:46 fenced receive_start 5:22 add node with started_count 6 
Jul 04 10:44:46 fenced receive_start 3:16 add node with started_count 1 

longjohn_fenced.log : 
Jul 03 09:47:12 fenced fenced 1352871249 started 
Jul 03 11:28:46 fenced cluster is down, exiting 
Jul 03 11:28:46 fenced daemon cpg_dispatch error 2 
Jul 03 12:11:43 fenced fenced 1364188437 started 
Jul 03 13:45:40 fenced receive_start 5:13 add node with started_count 6 
Jul 03 13:45:40 fenced receive_start 1:15 add node with started_count 8 
Jul 03 13:45:40 fenced receive_start 2:11 add node with started_count 4 
Jul 04 00:29:35 fenced receive_start 1:16 add node with started_count 8 
Jul 04 00:29:35 fenced receive_start 5:14 add node with started_count 6 
Jul 04 00:38:31 fenced receive_start 2:17 add node with started_count 4 
Jul 04 00:38:31 fenced receive_start 1:21 add node with started_count 8 
Jul 04 00:38:31 fenced receive_start 5:19 add node with started_count 6 
Jul 04 10:44:12 fenced receive_start 1:22 add node with started_count 8 
Jul 04 10:44:12 fenced receive_start 5:20 add node with started_count 6 
Jul 04 10:44:24 fenced receive_start 1:23 add node with started_count 8 
Jul 04 10:44:24 fenced telling cman to remove nodeid 2 from cluster 
Jul 04 10:44:24 fenced receive_start 2:23 add node with started_count 4 
Jul 04 10:44:24 fenced receive_start 5:21 add node with started_count 6 
Jul 04 10:44:46 f

[PVE-User] Cman crash problem

2013-07-08 Thread Jonathan Schaeffer

Hi all,

I'm experiencing a serious problem on our 4 nodes cluster (PVE 3.0).

It appeared after the network team changed network active equipments in 
the building (but this might not be the origin of the problem).


The symptoms are :

- The nodes appear in red on the web gui, except the one hosting the web 
service IP
- The VM, while still running correctly, do not show any information 
(running, rrd graphs, etc)


- clustat shows nodes as "online"
- some nodes seems to have been fenced (while not restarted)
(see log extracts : barbossa_fenced.log and jim_fenced.log)

- /var/log/cluster/corosync.log shows LOT of messages :
Jul 08 07:06:49 corosync [TOTEM ] Retransmit List: 13f54a 13f54b 13f54c 
13f54d 13f54e 13f54f 13f550 13f551 13f552 13f553 13f554 13f555 13f556 
13f557 13f558 13f559 13f55a 13f55b 13f55c 13f55d 13f55e


If I restart one node, the fencing is going to happen, other nodes will 
reboot and all the VMs hosted allong with them. I don't want this to happen.


I can provide more logs if necessary. Do you have an idea to help me 
understand what is going on here ?


Thanks,

Jonathan


barbossa_fenced.log :
Jul 03 12:07:21 fenced fencing deferred to jim
Jul 03 13:45:40 fenced receive_start 1:15 add node with started_count 8
Jul 03 13:45:40 fenced receive_start 2:11 add node with started_count 4
Jul 03 13:45:40 fenced receive_start 3:7 add node with started_count 1
Jul 04 00:29:35 fenced receive_start 1:16 add node with started_count 8
Jul 04 00:29:35 fenced receive_start 3:8 add node with started_count 1
Jul 04 00:38:31 fenced receive_start 2:17 add node with started_count 4
Jul 04 00:38:31 fenced receive_start 3:13 add node with started_count 1
Jul 04 00:38:31 fenced receive_start 1:21 add node with started_count 8
Jul 04 10:44:12 fenced receive_start 1:22 add node with started_count 8
Jul 04 10:44:12 fenced receive_start 3:14 add node with started_count 1
Jul 04 10:44:24 fenced receive_start 1:23 add node with started_count 8
Jul 04 10:44:24 fenced telling cman to remove nodeid 2 from cluster


jim_fenced.log :
Jul 03 12:07:21 fenced fencing node longjohn
Jul 03 12:07:32 fenced fence longjohn success
Jul 03 13:45:40 fenced receive_start 5:13 add node with started_count 6
Jul 03 13:45:40 fenced receive_start 2:11 add node with started_count 4
Jul 03 13:45:40 fenced receive_start 3:7 add node with started_count 1
Jul 04 00:29:35 fenced receive_start 3:8 add node with started_count 1
Jul 04 00:29:35 fenced receive_start 5:14 add node with started_count 6
Jul 04 00:38:31 fenced receive_start 2:17 add node with started_count 4
Jul 04 00:38:31 fenced receive_start 3:13 add node with started_count 1
Jul 04 00:38:31 fenced receive_start 5:19 add node with started_count 6
Jul 04 10:44:12 fenced receive_start 5:20 add node with started_count 6
Jul 04 10:44:12 fenced receive_start 3:14 add node with started_count 1
Jul 04 10:44:24 fenced telling cman to remove nodeid 2 from cluster
Jul 04 10:44:24 fenced receive_start 2:23 add node with started_count 4
Jul 04 10:44:24 fenced receive_start 3:15 add node with started_count 1
Jul 04 10:44:24 fenced receive_start 5:21 add node with started_count 6
Jul 04 10:44:46 fenced receive_start 5:22 add node with started_count 6
Jul 04 10:44:46 fenced receive_start 3:16 add node with started_count 1

longjohn_fenced.log :
Jul 03 09:47:12 fenced fenced 1352871249 started
Jul 03 11:28:46 fenced cluster is down, exiting
Jul 03 11:28:46 fenced daemon cpg_dispatch error 2
Jul 03 12:11:43 fenced fenced 1364188437 started
Jul 03 13:45:40 fenced receive_start 5:13 add node with started_count 6
Jul 03 13:45:40 fenced receive_start 1:15 add node with started_count 8
Jul 03 13:45:40 fenced receive_start 2:11 add node with started_count 4
Jul 04 00:29:35 fenced receive_start 1:16 add node with started_count 8
Jul 04 00:29:35 fenced receive_start 5:14 add node with started_count 6
Jul 04 00:38:31 fenced receive_start 2:17 add node with started_count 4
Jul 04 00:38:31 fenced receive_start 1:21 add node with started_count 8
Jul 04 00:38:31 fenced receive_start 5:19 add node with started_count 6
Jul 04 10:44:12 fenced receive_start 1:22 add node with started_count 8
Jul 04 10:44:12 fenced receive_start 5:20 add node with started_count 6
Jul 04 10:44:24 fenced receive_start 1:23 add node with started_count 8
Jul 04 10:44:24 fenced telling cman to remove nodeid 2 from cluster
Jul 04 10:44:24 fenced receive_start 2:23 add node with started_count 4
Jul 04 10:44:24 fenced receive_start 5:21 add node with started_count 6
Jul 04 10:44:46 fenced receive_start 5:22 add node with started_count 6
Jul 04 10:44:46 fenced receive_start 1:24 add node with started_count 8

flint_fenced.log :
Jul 03 11:18:30 fenced fenced 1364188437 started
Jul 03 12:07:21 fenced fencing deferred to jim
Jul 03 13:45:40 fenced receive_start 5:13 add node with started_count 6
Jul 03 13:45:40 fenced receive_start 1:15 add node with started_count 8
Jul 03 13:45:40 fenced receive_start 3:7 add node with started_count 1
J