Re: [Xen-devel] kernel BUG at net/core/dev.c:1133!

2006-07-07 Thread Herbert Xu
On Fri, Jul 07, 2006 at 10:06:08PM +0200, Patrick McHardy wrote:
> 
> Are you refering to this code in ip_nat_fn()?
> 
> /* If we had a hardware checksum before, it's now invalid */
> if ((*pskb)->ip_summed == CHECKSUM_HW)
> if (skb_checksum_help(*pskb, (out == NULL)))
> return NF_DROP;

Yep that's the one.

> Doing incremental updates should work fine. This is something
> I wanted to take care of at some point, but didn't get to it
> yet.

No worries.  I'm going to do a workaround to fix the checksums in
GSO for now.

Cheers,
-- 
Visit Openswan at http://www.openswan.org/
Email: Herbert Xu ~{PmV>HI~} <[EMAIL PROTECTED]>
Home Page: http://gondor.apana.org.au/~herbert/
PGP Key: http://gondor.apana.org.au/~herbert/pubkey.txt
-
To unsubscribe from this list: send the line "unsubscribe netdev" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html


Re: [Xen-devel] kernel BUG at net/core/dev.c:1133!

2006-07-07 Thread Herbert Xu
On Fri, Jul 07, 2006 at 05:03:36PM +0200, Petersson, Mats wrote:
> 
> So, nothing going on there... I certainly haven't got NAT on my machine,
> as my machine is within the AMD network, and doesn't need NAT. AMD
> probably uses NAT as part of it's external communications, but I doubt
> it's used at all internally. 

Actually, just having it loaded is enough to break TSO.  So for all this
time anyone who had ip_nat loaded were silently corrupting all their TSO
checksums!

I'll send a patch soon once I've tested it.

Cheers,
-- 
Visit Openswan at http://www.openswan.org/
Email: Herbert Xu ~{PmV>HI~} <[EMAIL PROTECTED]>
Home Page: http://gondor.apana.org.au/~herbert/
PGP Key: http://gondor.apana.org.au/~herbert/pubkey.txt
-
To unsubscribe from this list: send the line "unsubscribe netdev" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html


Re: [Xen-devel] kernel BUG at net/core/dev.c:1133!

2006-07-07 Thread Patrick McHardy
Herbert Xu wrote:
> Petersson, Mats <[EMAIL PROTECTED]> wrote:
> 
>>Looks like the GSO is involved?
> 
> 
> It's certainly what crashed your machine :) It's probably not the
> guilty party though.  Someone is passing through a TSO packet with
> checksum set to something other than CHECKSUM_HW.
> 
> I bet it's netfilter and we just never noticed before because real
> NICS would simply corrupt the checksum silently.
> 
> Could you confirm that you have netfilter rules (in particular NAT
> rules) and that this goes away if you flush all your netfilter tables?
> 
> Patrick, do we really have to zap the checksum on outbound NAT? Could
> we update it instead?

Are you refering to this code in ip_nat_fn()?

/* If we had a hardware checksum before, it's now invalid */
if ((*pskb)->ip_summed == CHECKSUM_HW)
if (skb_checksum_help(*pskb, (out == NULL)))
return NF_DROP;

Doing incremental updates should work fine. This is something
I wanted to take care of at some point, but didn't get to it
yet.

-
To unsubscribe from this list: send the line "unsubscribe netdev" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html


RE: [Xen-devel] kernel BUG at net/core/dev.c:1133!

2006-07-07 Thread Petersson, Mats
> -Original Message-
> From: Herbert Xu [mailto:[EMAIL PROTECTED] 
> Sent: 07 July 2006 15:40
> To: Petersson, Mats
> Cc: [EMAIL PROTECTED]; netdev@vger.kernel.org; 
> [EMAIL PROTECTED]; [EMAIL PROTECTED]
> Subject: Re: [Xen-devel] kernel BUG at net/core/dev.c:1133!
> 
> Petersson, Mats <[EMAIL PROTECTED]> wrote:
> > Looks like the GSO is involved?
> 
> It's certainly what crashed your machine :) It's probably not the
> guilty party though.  Someone is passing through a TSO packet with
> checksum set to something other than CHECKSUM_HW.
> 
> I bet it's netfilter and we just never noticed before because real
> NICS would simply corrupt the checksum silently.
> 
> Could you confirm that you have netfilter rules (in particular NAT
> rules) and that this goes away if you flush all your netfilter tables?

If by netfilter, you mean "iptables", it says:
[EMAIL PROTECTED] ~]# iptables --list
Chain FORWARD (policy ACCEPT)
target prot opt source   destination

Chain INPUT (policy ACCEPT)
target prot opt source   destination

Chain OUTPUT (policy ACCEPT)
target prot opt source   destination

So, nothing going on there... I certainly haven't got NAT on my machine,
as my machine is within the AMD network, and doesn't need NAT. AMD
probably uses NAT as part of it's external communications, but I doubt
it's used at all internally. 

I also have noticed the crash happens when I try to access another
machine within my local switch - if that makes any difference... But not
instantly. I can do some communication with the machine next to it [like
I did "ssh cheetah" from my machine "quad" to get the iptables above,
and it works just fine - but when I did "xm dmesg" from "cheetah"
through ssh on "quad", it didn't work - presumably because it's a bit
more data being pushed - but I can't say for sure, as I have made no
attempt to really debug it]. 

I hope this info is of help to analyze the situation, and please feel
free to ask for further info.

--
Mats
> 
> Patrick, do we really have to zap the checksum on outbound NAT? Could
> we update it instead?
> 
> > I got this while running Dom0 only (no guests), with a
> > BOINC/[EMAIL PROTECTED] application running on all 4 cores. 
> > 
> > changeset:   10649:8e55c5c11475
> > 
> > Build: x86_32p (pae). 
> > 
> > [ cut here ]
> > kernel BUG at net/core/dev.c:1133!
> > invalid opcode:  [#1]
> > SMP 
> > CPU:0
> > EIP:0061:[]Not tainted VLI
> > EFLAGS: 00210297   (2.6.16.13-xen #12) 
> > EIP is at skb_gso_segment+0xf0/0x110
> > eax:    ebx: 0003   ecx: 0002   edx: c06e2e00
> > esi: 0008   edi: cd9e32e0   ebp: c63a7900   esp: c0de5ad0
> > ds: 007b   es: 007b   ss: 0069
> > Process rosetta_5.25_i6 (pid: 8826, threadinfo=c0de4000 
> task=cb019560)
> > Stack: <0>c8f69060  ffa3 0003 cd9e32e0 
> 0002 c63a7900
> > c04dcfb0 
> >   cd9e32e0 0003  cd9e32e0 cf8e3000 cf8e3140 c04dd07e
> > cd9e32e0 
> >   cf8e3000  cd9e32e0 cf8e3000 c04ec07e cd9e32e0 cf8e3000
> > c0895140 
> > Call Trace:
> > [] dev_gso_segment+0x30/0xb0
> > [] dev_hard_start_xmit+0x4e/0x110
> > [] __qdisc_run+0xbe/0x280
> > [] dev_queue_xmit+0x379/0x380
> > [] br_dev_queue_push_xmit+0xa4/0x140
> > [] br_nf_post_routing+0x102/0x1d0
> > [] br_nf_dev_queue_xmit+0x0/0x50
> > [] br_dev_queue_push_xmit+0x0/0x140
> > [] nf_iterate+0x6b/0xa0
> > [] br_dev_queue_push_xmit+0x0/0x140
> > [] br_dev_queue_push_xmit+0x0/0x140
> > [] nf_hook_slow+0x6e/0x120
> > [] br_dev_queue_push_xmit+0x0/0x140
> > [] br_forward_finish+0x60/0x70
> > [] br_dev_queue_push_xmit+0x0/0x140
> > [] br_nf_forward_finish+0x71/0x130
> > [] br_forward_finish+0x0/0x70
> > [] br_nf_forward_ip+0xf0/0x1a0
> > [] br_nf_forward_finish+0x0/0x130
> > [] br_forward_finish+0x0/0x70
> > [] nf_iterate+0x6b/0xa0
> > [] br_forward_finish+0x0/0x70
> > [] br_forward_finish+0x0/0x70
> > [] nf_hook_slow+0x6e/0x120
> > [] br_forward_finish+0x0/0x70
> > [] __br_forward+0x74/0x80
> > [] br_forward_finish+0x0/0x70
> > [] br_handle_frame_finish+0xd1/0x160
> > [] br_handle_frame_finish+0x0/0x160
> > [] br_nf_pre_routing_finish+0xfb/0x480
> > [] br_handle_frame_finish+0x0/0x160
> > [] br_nf_pre_routing_finish+0x0/0x480
> > [] ip_nat_in+0x43/0xc0
> > [] br_nf_pre_routing_finish+0x0/0x480
> > [] nf_iterate+0x6b/0xa0
> > [] br_nf_pre_routing_finish+0x0/0x480
> &g

RE: [Xen-devel] kernel BUG at net/core/dev.c:1133!

2006-07-07 Thread Petersson, Mats
> -Original Message-
> From: Tim Post [mailto:[EMAIL PROTECTED] 
> Sent: 07 July 2006 16:06
> To: Herbert Xu
> Cc: Petersson, Mats; netdev@vger.kernel.org; 
> [EMAIL PROTECTED]; [EMAIL PROTECTED]; [EMAIL PROTECTED]
> Subject: Re: [Xen-devel] kernel BUG at net/core/dev.c:1133!
> 
> I got the exact same thing when attempting to use BOINC on a 
> single node
> supporting a 5 node open SSI cluster, (5 guests) and yes the problem
> went away when I flushed the rules. 
> 
> I attributed this to a quirk with the cluster CVIP, because I had also
> assigned each node its own outbound IP in addition to the 
> incoming CVIP.
> 
> Since I felt it was due to my tendency to over-tinker, I 
> didn't mention
> it on the lists, was a few months ago. 
> 
> Thought I would chime in as it sounds like the same experience, up to
> and including BOINC.

I haven't been tinkering with anything [on purpose, at least] - the
system is a default installation of FC4, with the latest Xen-unstable
[bar the last dozen or so changesets - I don't pull the latest every
half-hour]. 

--
Mats
> 
> HTH
> 
> --Tim
> 
> On Sat, 2006-07-08 at 00:39 +1000, Herbert Xu wrote:
> > Petersson, Mats <[EMAIL PROTECTED]> wrote:
> > > Looks like the GSO is involved?
> > 
> > It's certainly what crashed your machine :) It's probably not the
> > guilty party though.  Someone is passing through a TSO packet with
> > checksum set to something other than CHECKSUM_HW.
> > 
> > I bet it's netfilter and we just never noticed before because real
> > NICS would simply corrupt the checksum silently.
> > 
> > Could you confirm that you have netfilter rules (in particular NAT
> > rules) and that this goes away if you flush all your 
> netfilter tables?
> > 
> > Patrick, do we really have to zap the checksum on outbound 
> NAT? Could
> > we update it instead?
> > 
> > > I got this while running Dom0 only (no guests), with a
> > > BOINC/[EMAIL PROTECTED] application running on all 4 cores. 
> > > 
> > > changeset:   10649:8e55c5c11475
> > > 
> > > Build: x86_32p (pae). 
> > > 
> > > [ cut here ]
> > > kernel BUG at net/core/dev.c:1133!
> > > invalid opcode:  [#1]
> > > SMP 
> > > CPU:0
> > > EIP:0061:[]Not tainted VLI
> > > EFLAGS: 00210297   (2.6.16.13-xen #12) 
> > > EIP is at skb_gso_segment+0xf0/0x110
> > > eax:    ebx: 0003   ecx: 0002   edx: c06e2e00
> > > esi: 0008   edi: cd9e32e0   ebp: c63a7900   esp: c0de5ad0
> > > ds: 007b   es: 007b   ss: 0069
> > > Process rosetta_5.25_i6 (pid: 8826, threadinfo=c0de4000 
> task=cb019560)
> > > Stack: <0>c8f69060  ffa3 0003 cd9e32e0 
> 0002 c63a7900
> > > c04dcfb0 
> > >   cd9e32e0 0003  cd9e32e0 cf8e3000 
> cf8e3140 c04dd07e
> > > cd9e32e0 
> > >   cf8e3000  cd9e32e0 cf8e3000 c04ec07e 
> cd9e32e0 cf8e3000
> > > c0895140 
> > > Call Trace:
> > > [] dev_gso_segment+0x30/0xb0
> > > [] dev_hard_start_xmit+0x4e/0x110
> > > [] __qdisc_run+0xbe/0x280
> > > [] dev_queue_xmit+0x379/0x380
> > > [] br_dev_queue_push_xmit+0xa4/0x140
> > > [] br_nf_post_routing+0x102/0x1d0
> > > [] br_nf_dev_queue_xmit+0x0/0x50
> > > [] br_dev_queue_push_xmit+0x0/0x140
> > > [] nf_iterate+0x6b/0xa0
> > > [] br_dev_queue_push_xmit+0x0/0x140
> > > [] br_dev_queue_push_xmit+0x0/0x140
> > > [] nf_hook_slow+0x6e/0x120
> > > [] br_dev_queue_push_xmit+0x0/0x140
> > > [] br_forward_finish+0x60/0x70
> > > [] br_dev_queue_push_xmit+0x0/0x140
> > > [] br_nf_forward_finish+0x71/0x130
> > > [] br_forward_finish+0x0/0x70
> > > [] br_nf_forward_ip+0xf0/0x1a0
> > > [] br_nf_forward_finish+0x0/0x130
> > > [] br_forward_finish+0x0/0x70
> > > [] nf_iterate+0x6b/0xa0
> > > [] br_forward_finish+0x0/0x70
> > > [] br_forward_finish+0x0/0x70
> > > [] nf_hook_slow+0x6e/0x120
> > > [] br_forward_finish+0x0/0x70
> > > [] __br_forward+0x74/0x80
> > > [] br_forward_finish+0x0/0x70
> > > [] br_handle_frame_finish+0xd1/0x160
> > > [] br_handle_frame_finish+0x0/0x160
> > > [] br_nf_pre_routing_finish+0xfb/0x480
> > > [] br_handle_frame_finish+0x0/0x160
> > > [] br_nf_pre_routing_finish+0x0/0x480
> > > [] ip_nat_in+0x43/0xc0
> > > [] br_nf_pre_routing_finish+0x0/0x480
> > > [] nf_iterate+0x6b/

Re: [Xen-devel] kernel BUG at net/core/dev.c:1133!

2006-07-07 Thread Tim Post
I got the exact same thing when attempting to use BOINC on a single node
supporting a 5 node open SSI cluster, (5 guests) and yes the problem
went away when I flushed the rules. 

I attributed this to a quirk with the cluster CVIP, because I had also
assigned each node its own outbound IP in addition to the incoming CVIP.

Since I felt it was due to my tendency to over-tinker, I didn't mention
it on the lists, was a few months ago. 

Thought I would chime in as it sounds like the same experience, up to
and including BOINC.

HTH

--Tim

On Sat, 2006-07-08 at 00:39 +1000, Herbert Xu wrote:
> Petersson, Mats <[EMAIL PROTECTED]> wrote:
> > Looks like the GSO is involved?
> 
> It's certainly what crashed your machine :) It's probably not the
> guilty party though.  Someone is passing through a TSO packet with
> checksum set to something other than CHECKSUM_HW.
> 
> I bet it's netfilter and we just never noticed before because real
> NICS would simply corrupt the checksum silently.
> 
> Could you confirm that you have netfilter rules (in particular NAT
> rules) and that this goes away if you flush all your netfilter tables?
> 
> Patrick, do we really have to zap the checksum on outbound NAT? Could
> we update it instead?
> 
> > I got this while running Dom0 only (no guests), with a
> > BOINC/[EMAIL PROTECTED] application running on all 4 cores. 
> > 
> > changeset:   10649:8e55c5c11475
> > 
> > Build: x86_32p (pae). 
> > 
> > [ cut here ]
> > kernel BUG at net/core/dev.c:1133!
> > invalid opcode:  [#1]
> > SMP 
> > CPU:0
> > EIP:0061:[]Not tainted VLI
> > EFLAGS: 00210297   (2.6.16.13-xen #12) 
> > EIP is at skb_gso_segment+0xf0/0x110
> > eax:    ebx: 0003   ecx: 0002   edx: c06e2e00
> > esi: 0008   edi: cd9e32e0   ebp: c63a7900   esp: c0de5ad0
> > ds: 007b   es: 007b   ss: 0069
> > Process rosetta_5.25_i6 (pid: 8826, threadinfo=c0de4000 task=cb019560)
> > Stack: <0>c8f69060  ffa3 0003 cd9e32e0 0002 c63a7900
> > c04dcfb0 
> >   cd9e32e0 0003  cd9e32e0 cf8e3000 cf8e3140 c04dd07e
> > cd9e32e0 
> >   cf8e3000  cd9e32e0 cf8e3000 c04ec07e cd9e32e0 cf8e3000
> > c0895140 
> > Call Trace:
> > [] dev_gso_segment+0x30/0xb0
> > [] dev_hard_start_xmit+0x4e/0x110
> > [] __qdisc_run+0xbe/0x280
> > [] dev_queue_xmit+0x379/0x380
> > [] br_dev_queue_push_xmit+0xa4/0x140
> > [] br_nf_post_routing+0x102/0x1d0
> > [] br_nf_dev_queue_xmit+0x0/0x50
> > [] br_dev_queue_push_xmit+0x0/0x140
> > [] nf_iterate+0x6b/0xa0
> > [] br_dev_queue_push_xmit+0x0/0x140
> > [] br_dev_queue_push_xmit+0x0/0x140
> > [] nf_hook_slow+0x6e/0x120
> > [] br_dev_queue_push_xmit+0x0/0x140
> > [] br_forward_finish+0x60/0x70
> > [] br_dev_queue_push_xmit+0x0/0x140
> > [] br_nf_forward_finish+0x71/0x130
> > [] br_forward_finish+0x0/0x70
> > [] br_nf_forward_ip+0xf0/0x1a0
> > [] br_nf_forward_finish+0x0/0x130
> > [] br_forward_finish+0x0/0x70
> > [] nf_iterate+0x6b/0xa0
> > [] br_forward_finish+0x0/0x70
> > [] br_forward_finish+0x0/0x70
> > [] nf_hook_slow+0x6e/0x120
> > [] br_forward_finish+0x0/0x70
> > [] __br_forward+0x74/0x80
> > [] br_forward_finish+0x0/0x70
> > [] br_handle_frame_finish+0xd1/0x160
> > [] br_handle_frame_finish+0x0/0x160
> > [] br_nf_pre_routing_finish+0xfb/0x480
> > [] br_handle_frame_finish+0x0/0x160
> > [] br_nf_pre_routing_finish+0x0/0x480
> > [] ip_nat_in+0x43/0xc0
> > [] br_nf_pre_routing_finish+0x0/0x480
> > [] nf_iterate+0x6b/0xa0
> > [] br_nf_pre_routing_finish+0x0/0x480
> > [] br_nf_pre_routing_finish+0x0/0x480
> > [] nf_hook_slow+0x6e/0x120
> > [] br_nf_pre_routing_finish+0x0/0x480
> > [] br_nf_pre_routing+0x404/0x580
> > [] br_nf_pre_routing_finish+0x0/0x480
> > [] nf_iterate+0x6b/0xa0
> > [] br_handle_frame_finish+0x0/0x160
> > [] br_handle_frame_finish+0x0/0x160
> > [] nf_hook_slow+0x6e/0x120
> > [] br_handle_frame_finish+0x0/0x160
> > [] br_handle_frame+0x1e4/0x250
> > [] br_handle_frame_finish+0x0/0x160
> > [] netif_receive_skb+0x165/0x2a0
> > [] process_backlog+0xbf/0x180
> > [] net_rx_action+0x11f/0x1d0
> > [] __do_softirq+0x86/0x120
> > [] do_softirq+0x75/0x90
> > [] do_IRQ+0x1f/0x30
> > [] evtchn_do_upcall+0x90/0x100
> > [] hypervisor_callback+0x3d/0x48
> > Code: c2 2b 57 24 29 d0 8d 14 2a 89 87 94 00 00 00 89 57 60 8b 44 24 08
> > 83 c4 0c 5b 5e 5f 5d c3 0f 0
> > b 69 03 fe 8c 66 c0 e9 69 ff ff ff <0f> 0b 6d 04 e8 ab 6c c0 e9 3a ff ff
> > ff 0f 0b 6c 04 e8 ab 6c c0 
> > <0>Kernel panic - not syncing: Fatal exception in interrupt
> 
> Cheers,

-
To unsubscribe from this list: send the line "unsubscribe netdev" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html


Re: [Xen-devel] kernel BUG at net/core/dev.c:1133!

2006-07-07 Thread Herbert Xu
Petersson, Mats <[EMAIL PROTECTED]> wrote:
> Looks like the GSO is involved?

It's certainly what crashed your machine :) It's probably not the
guilty party though.  Someone is passing through a TSO packet with
checksum set to something other than CHECKSUM_HW.

I bet it's netfilter and we just never noticed before because real
NICS would simply corrupt the checksum silently.

Could you confirm that you have netfilter rules (in particular NAT
rules) and that this goes away if you flush all your netfilter tables?

Patrick, do we really have to zap the checksum on outbound NAT? Could
we update it instead?

> I got this while running Dom0 only (no guests), with a
> BOINC/[EMAIL PROTECTED] application running on all 4 cores. 
> 
> changeset:   10649:8e55c5c11475
> 
> Build: x86_32p (pae). 
> 
> [ cut here ]
> kernel BUG at net/core/dev.c:1133!
> invalid opcode:  [#1]
> SMP 
> CPU:0
> EIP:0061:[]Not tainted VLI
> EFLAGS: 00210297   (2.6.16.13-xen #12) 
> EIP is at skb_gso_segment+0xf0/0x110
> eax:    ebx: 0003   ecx: 0002   edx: c06e2e00
> esi: 0008   edi: cd9e32e0   ebp: c63a7900   esp: c0de5ad0
> ds: 007b   es: 007b   ss: 0069
> Process rosetta_5.25_i6 (pid: 8826, threadinfo=c0de4000 task=cb019560)
> Stack: <0>c8f69060  ffa3 0003 cd9e32e0 0002 c63a7900
> c04dcfb0 
>   cd9e32e0 0003  cd9e32e0 cf8e3000 cf8e3140 c04dd07e
> cd9e32e0 
>   cf8e3000  cd9e32e0 cf8e3000 c04ec07e cd9e32e0 cf8e3000
> c0895140 
> Call Trace:
> [] dev_gso_segment+0x30/0xb0
> [] dev_hard_start_xmit+0x4e/0x110
> [] __qdisc_run+0xbe/0x280
> [] dev_queue_xmit+0x379/0x380
> [] br_dev_queue_push_xmit+0xa4/0x140
> [] br_nf_post_routing+0x102/0x1d0
> [] br_nf_dev_queue_xmit+0x0/0x50
> [] br_dev_queue_push_xmit+0x0/0x140
> [] nf_iterate+0x6b/0xa0
> [] br_dev_queue_push_xmit+0x0/0x140
> [] br_dev_queue_push_xmit+0x0/0x140
> [] nf_hook_slow+0x6e/0x120
> [] br_dev_queue_push_xmit+0x0/0x140
> [] br_forward_finish+0x60/0x70
> [] br_dev_queue_push_xmit+0x0/0x140
> [] br_nf_forward_finish+0x71/0x130
> [] br_forward_finish+0x0/0x70
> [] br_nf_forward_ip+0xf0/0x1a0
> [] br_nf_forward_finish+0x0/0x130
> [] br_forward_finish+0x0/0x70
> [] nf_iterate+0x6b/0xa0
> [] br_forward_finish+0x0/0x70
> [] br_forward_finish+0x0/0x70
> [] nf_hook_slow+0x6e/0x120
> [] br_forward_finish+0x0/0x70
> [] __br_forward+0x74/0x80
> [] br_forward_finish+0x0/0x70
> [] br_handle_frame_finish+0xd1/0x160
> [] br_handle_frame_finish+0x0/0x160
> [] br_nf_pre_routing_finish+0xfb/0x480
> [] br_handle_frame_finish+0x0/0x160
> [] br_nf_pre_routing_finish+0x0/0x480
> [] ip_nat_in+0x43/0xc0
> [] br_nf_pre_routing_finish+0x0/0x480
> [] nf_iterate+0x6b/0xa0
> [] br_nf_pre_routing_finish+0x0/0x480
> [] br_nf_pre_routing_finish+0x0/0x480
> [] nf_hook_slow+0x6e/0x120
> [] br_nf_pre_routing_finish+0x0/0x480
> [] br_nf_pre_routing+0x404/0x580
> [] br_nf_pre_routing_finish+0x0/0x480
> [] nf_iterate+0x6b/0xa0
> [] br_handle_frame_finish+0x0/0x160
> [] br_handle_frame_finish+0x0/0x160
> [] nf_hook_slow+0x6e/0x120
> [] br_handle_frame_finish+0x0/0x160
> [] br_handle_frame+0x1e4/0x250
> [] br_handle_frame_finish+0x0/0x160
> [] netif_receive_skb+0x165/0x2a0
> [] process_backlog+0xbf/0x180
> [] net_rx_action+0x11f/0x1d0
> [] __do_softirq+0x86/0x120
> [] do_softirq+0x75/0x90
> [] do_IRQ+0x1f/0x30
> [] evtchn_do_upcall+0x90/0x100
> [] hypervisor_callback+0x3d/0x48
> Code: c2 2b 57 24 29 d0 8d 14 2a 89 87 94 00 00 00 89 57 60 8b 44 24 08
> 83 c4 0c 5b 5e 5f 5d c3 0f 0
> b 69 03 fe 8c 66 c0 e9 69 ff ff ff <0f> 0b 6d 04 e8 ab 6c c0 e9 3a ff ff
> ff 0f 0b 6c 04 e8 ab 6c c0 
> <0>Kernel panic - not syncing: Fatal exception in interrupt

Cheers,
-- 
Visit Openswan at http://www.openswan.org/
Email: Herbert Xu ~{PmV>HI~} <[EMAIL PROTECTED]>
Home Page: http://gondor.apana.org.au/~herbert/
PGP Key: http://gondor.apana.org.au/~herbert/pubkey.txt
-
To unsubscribe from this list: send the line "unsubscribe netdev" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html