Re: [Xen-devel] kernel BUG at net/core/dev.c:1133!
On Fri, Jul 07, 2006 at 10:06:08PM +0200, Patrick McHardy wrote: > > Are you refering to this code in ip_nat_fn()? > > /* If we had a hardware checksum before, it's now invalid */ > if ((*pskb)->ip_summed == CHECKSUM_HW) > if (skb_checksum_help(*pskb, (out == NULL))) > return NF_DROP; Yep that's the one. > Doing incremental updates should work fine. This is something > I wanted to take care of at some point, but didn't get to it > yet. No worries. I'm going to do a workaround to fix the checksums in GSO for now. Cheers, -- Visit Openswan at http://www.openswan.org/ Email: Herbert Xu ~{PmV>HI~} <[EMAIL PROTECTED]> Home Page: http://gondor.apana.org.au/~herbert/ PGP Key: http://gondor.apana.org.au/~herbert/pubkey.txt - To unsubscribe from this list: send the line "unsubscribe netdev" in the body of a message to [EMAIL PROTECTED] More majordomo info at http://vger.kernel.org/majordomo-info.html
Re: [Xen-devel] kernel BUG at net/core/dev.c:1133!
On Fri, Jul 07, 2006 at 05:03:36PM +0200, Petersson, Mats wrote: > > So, nothing going on there... I certainly haven't got NAT on my machine, > as my machine is within the AMD network, and doesn't need NAT. AMD > probably uses NAT as part of it's external communications, but I doubt > it's used at all internally. Actually, just having it loaded is enough to break TSO. So for all this time anyone who had ip_nat loaded were silently corrupting all their TSO checksums! I'll send a patch soon once I've tested it. Cheers, -- Visit Openswan at http://www.openswan.org/ Email: Herbert Xu ~{PmV>HI~} <[EMAIL PROTECTED]> Home Page: http://gondor.apana.org.au/~herbert/ PGP Key: http://gondor.apana.org.au/~herbert/pubkey.txt - To unsubscribe from this list: send the line "unsubscribe netdev" in the body of a message to [EMAIL PROTECTED] More majordomo info at http://vger.kernel.org/majordomo-info.html
Re: [Xen-devel] kernel BUG at net/core/dev.c:1133!
Herbert Xu wrote: > Petersson, Mats <[EMAIL PROTECTED]> wrote: > >>Looks like the GSO is involved? > > > It's certainly what crashed your machine :) It's probably not the > guilty party though. Someone is passing through a TSO packet with > checksum set to something other than CHECKSUM_HW. > > I bet it's netfilter and we just never noticed before because real > NICS would simply corrupt the checksum silently. > > Could you confirm that you have netfilter rules (in particular NAT > rules) and that this goes away if you flush all your netfilter tables? > > Patrick, do we really have to zap the checksum on outbound NAT? Could > we update it instead? Are you refering to this code in ip_nat_fn()? /* If we had a hardware checksum before, it's now invalid */ if ((*pskb)->ip_summed == CHECKSUM_HW) if (skb_checksum_help(*pskb, (out == NULL))) return NF_DROP; Doing incremental updates should work fine. This is something I wanted to take care of at some point, but didn't get to it yet. - To unsubscribe from this list: send the line "unsubscribe netdev" in the body of a message to [EMAIL PROTECTED] More majordomo info at http://vger.kernel.org/majordomo-info.html
RE: [Xen-devel] kernel BUG at net/core/dev.c:1133!
> -Original Message- > From: Herbert Xu [mailto:[EMAIL PROTECTED] > Sent: 07 July 2006 15:40 > To: Petersson, Mats > Cc: [EMAIL PROTECTED]; netdev@vger.kernel.org; > [EMAIL PROTECTED]; [EMAIL PROTECTED] > Subject: Re: [Xen-devel] kernel BUG at net/core/dev.c:1133! > > Petersson, Mats <[EMAIL PROTECTED]> wrote: > > Looks like the GSO is involved? > > It's certainly what crashed your machine :) It's probably not the > guilty party though. Someone is passing through a TSO packet with > checksum set to something other than CHECKSUM_HW. > > I bet it's netfilter and we just never noticed before because real > NICS would simply corrupt the checksum silently. > > Could you confirm that you have netfilter rules (in particular NAT > rules) and that this goes away if you flush all your netfilter tables? If by netfilter, you mean "iptables", it says: [EMAIL PROTECTED] ~]# iptables --list Chain FORWARD (policy ACCEPT) target prot opt source destination Chain INPUT (policy ACCEPT) target prot opt source destination Chain OUTPUT (policy ACCEPT) target prot opt source destination So, nothing going on there... I certainly haven't got NAT on my machine, as my machine is within the AMD network, and doesn't need NAT. AMD probably uses NAT as part of it's external communications, but I doubt it's used at all internally. I also have noticed the crash happens when I try to access another machine within my local switch - if that makes any difference... But not instantly. I can do some communication with the machine next to it [like I did "ssh cheetah" from my machine "quad" to get the iptables above, and it works just fine - but when I did "xm dmesg" from "cheetah" through ssh on "quad", it didn't work - presumably because it's a bit more data being pushed - but I can't say for sure, as I have made no attempt to really debug it]. I hope this info is of help to analyze the situation, and please feel free to ask for further info. -- Mats > > Patrick, do we really have to zap the checksum on outbound NAT? Could > we update it instead? > > > I got this while running Dom0 only (no guests), with a > > BOINC/[EMAIL PROTECTED] application running on all 4 cores. > > > > changeset: 10649:8e55c5c11475 > > > > Build: x86_32p (pae). > > > > [ cut here ] > > kernel BUG at net/core/dev.c:1133! > > invalid opcode: [#1] > > SMP > > CPU:0 > > EIP:0061:[]Not tainted VLI > > EFLAGS: 00210297 (2.6.16.13-xen #12) > > EIP is at skb_gso_segment+0xf0/0x110 > > eax: ebx: 0003 ecx: 0002 edx: c06e2e00 > > esi: 0008 edi: cd9e32e0 ebp: c63a7900 esp: c0de5ad0 > > ds: 007b es: 007b ss: 0069 > > Process rosetta_5.25_i6 (pid: 8826, threadinfo=c0de4000 > task=cb019560) > > Stack: <0>c8f69060 ffa3 0003 cd9e32e0 > 0002 c63a7900 > > c04dcfb0 > > cd9e32e0 0003 cd9e32e0 cf8e3000 cf8e3140 c04dd07e > > cd9e32e0 > > cf8e3000 cd9e32e0 cf8e3000 c04ec07e cd9e32e0 cf8e3000 > > c0895140 > > Call Trace: > > [] dev_gso_segment+0x30/0xb0 > > [] dev_hard_start_xmit+0x4e/0x110 > > [] __qdisc_run+0xbe/0x280 > > [] dev_queue_xmit+0x379/0x380 > > [] br_dev_queue_push_xmit+0xa4/0x140 > > [] br_nf_post_routing+0x102/0x1d0 > > [] br_nf_dev_queue_xmit+0x0/0x50 > > [] br_dev_queue_push_xmit+0x0/0x140 > > [] nf_iterate+0x6b/0xa0 > > [] br_dev_queue_push_xmit+0x0/0x140 > > [] br_dev_queue_push_xmit+0x0/0x140 > > [] nf_hook_slow+0x6e/0x120 > > [] br_dev_queue_push_xmit+0x0/0x140 > > [] br_forward_finish+0x60/0x70 > > [] br_dev_queue_push_xmit+0x0/0x140 > > [] br_nf_forward_finish+0x71/0x130 > > [] br_forward_finish+0x0/0x70 > > [] br_nf_forward_ip+0xf0/0x1a0 > > [] br_nf_forward_finish+0x0/0x130 > > [] br_forward_finish+0x0/0x70 > > [] nf_iterate+0x6b/0xa0 > > [] br_forward_finish+0x0/0x70 > > [] br_forward_finish+0x0/0x70 > > [] nf_hook_slow+0x6e/0x120 > > [] br_forward_finish+0x0/0x70 > > [] __br_forward+0x74/0x80 > > [] br_forward_finish+0x0/0x70 > > [] br_handle_frame_finish+0xd1/0x160 > > [] br_handle_frame_finish+0x0/0x160 > > [] br_nf_pre_routing_finish+0xfb/0x480 > > [] br_handle_frame_finish+0x0/0x160 > > [] br_nf_pre_routing_finish+0x0/0x480 > > [] ip_nat_in+0x43/0xc0 > > [] br_nf_pre_routing_finish+0x0/0x480 > > [] nf_iterate+0x6b/0xa0 > > [] br_nf_pre_routing_finish+0x0/0x480 > &g
RE: [Xen-devel] kernel BUG at net/core/dev.c:1133!
> -Original Message- > From: Tim Post [mailto:[EMAIL PROTECTED] > Sent: 07 July 2006 16:06 > To: Herbert Xu > Cc: Petersson, Mats; netdev@vger.kernel.org; > [EMAIL PROTECTED]; [EMAIL PROTECTED]; [EMAIL PROTECTED] > Subject: Re: [Xen-devel] kernel BUG at net/core/dev.c:1133! > > I got the exact same thing when attempting to use BOINC on a > single node > supporting a 5 node open SSI cluster, (5 guests) and yes the problem > went away when I flushed the rules. > > I attributed this to a quirk with the cluster CVIP, because I had also > assigned each node its own outbound IP in addition to the > incoming CVIP. > > Since I felt it was due to my tendency to over-tinker, I > didn't mention > it on the lists, was a few months ago. > > Thought I would chime in as it sounds like the same experience, up to > and including BOINC. I haven't been tinkering with anything [on purpose, at least] - the system is a default installation of FC4, with the latest Xen-unstable [bar the last dozen or so changesets - I don't pull the latest every half-hour]. -- Mats > > HTH > > --Tim > > On Sat, 2006-07-08 at 00:39 +1000, Herbert Xu wrote: > > Petersson, Mats <[EMAIL PROTECTED]> wrote: > > > Looks like the GSO is involved? > > > > It's certainly what crashed your machine :) It's probably not the > > guilty party though. Someone is passing through a TSO packet with > > checksum set to something other than CHECKSUM_HW. > > > > I bet it's netfilter and we just never noticed before because real > > NICS would simply corrupt the checksum silently. > > > > Could you confirm that you have netfilter rules (in particular NAT > > rules) and that this goes away if you flush all your > netfilter tables? > > > > Patrick, do we really have to zap the checksum on outbound > NAT? Could > > we update it instead? > > > > > I got this while running Dom0 only (no guests), with a > > > BOINC/[EMAIL PROTECTED] application running on all 4 cores. > > > > > > changeset: 10649:8e55c5c11475 > > > > > > Build: x86_32p (pae). > > > > > > [ cut here ] > > > kernel BUG at net/core/dev.c:1133! > > > invalid opcode: [#1] > > > SMP > > > CPU:0 > > > EIP:0061:[]Not tainted VLI > > > EFLAGS: 00210297 (2.6.16.13-xen #12) > > > EIP is at skb_gso_segment+0xf0/0x110 > > > eax: ebx: 0003 ecx: 0002 edx: c06e2e00 > > > esi: 0008 edi: cd9e32e0 ebp: c63a7900 esp: c0de5ad0 > > > ds: 007b es: 007b ss: 0069 > > > Process rosetta_5.25_i6 (pid: 8826, threadinfo=c0de4000 > task=cb019560) > > > Stack: <0>c8f69060 ffa3 0003 cd9e32e0 > 0002 c63a7900 > > > c04dcfb0 > > > cd9e32e0 0003 cd9e32e0 cf8e3000 > cf8e3140 c04dd07e > > > cd9e32e0 > > > cf8e3000 cd9e32e0 cf8e3000 c04ec07e > cd9e32e0 cf8e3000 > > > c0895140 > > > Call Trace: > > > [] dev_gso_segment+0x30/0xb0 > > > [] dev_hard_start_xmit+0x4e/0x110 > > > [] __qdisc_run+0xbe/0x280 > > > [] dev_queue_xmit+0x379/0x380 > > > [] br_dev_queue_push_xmit+0xa4/0x140 > > > [] br_nf_post_routing+0x102/0x1d0 > > > [] br_nf_dev_queue_xmit+0x0/0x50 > > > [] br_dev_queue_push_xmit+0x0/0x140 > > > [] nf_iterate+0x6b/0xa0 > > > [] br_dev_queue_push_xmit+0x0/0x140 > > > [] br_dev_queue_push_xmit+0x0/0x140 > > > [] nf_hook_slow+0x6e/0x120 > > > [] br_dev_queue_push_xmit+0x0/0x140 > > > [] br_forward_finish+0x60/0x70 > > > [] br_dev_queue_push_xmit+0x0/0x140 > > > [] br_nf_forward_finish+0x71/0x130 > > > [] br_forward_finish+0x0/0x70 > > > [] br_nf_forward_ip+0xf0/0x1a0 > > > [] br_nf_forward_finish+0x0/0x130 > > > [] br_forward_finish+0x0/0x70 > > > [] nf_iterate+0x6b/0xa0 > > > [] br_forward_finish+0x0/0x70 > > > [] br_forward_finish+0x0/0x70 > > > [] nf_hook_slow+0x6e/0x120 > > > [] br_forward_finish+0x0/0x70 > > > [] __br_forward+0x74/0x80 > > > [] br_forward_finish+0x0/0x70 > > > [] br_handle_frame_finish+0xd1/0x160 > > > [] br_handle_frame_finish+0x0/0x160 > > > [] br_nf_pre_routing_finish+0xfb/0x480 > > > [] br_handle_frame_finish+0x0/0x160 > > > [] br_nf_pre_routing_finish+0x0/0x480 > > > [] ip_nat_in+0x43/0xc0 > > > [] br_nf_pre_routing_finish+0x0/0x480 > > > [] nf_iterate+0x6b/
Re: [Xen-devel] kernel BUG at net/core/dev.c:1133!
I got the exact same thing when attempting to use BOINC on a single node supporting a 5 node open SSI cluster, (5 guests) and yes the problem went away when I flushed the rules. I attributed this to a quirk with the cluster CVIP, because I had also assigned each node its own outbound IP in addition to the incoming CVIP. Since I felt it was due to my tendency to over-tinker, I didn't mention it on the lists, was a few months ago. Thought I would chime in as it sounds like the same experience, up to and including BOINC. HTH --Tim On Sat, 2006-07-08 at 00:39 +1000, Herbert Xu wrote: > Petersson, Mats <[EMAIL PROTECTED]> wrote: > > Looks like the GSO is involved? > > It's certainly what crashed your machine :) It's probably not the > guilty party though. Someone is passing through a TSO packet with > checksum set to something other than CHECKSUM_HW. > > I bet it's netfilter and we just never noticed before because real > NICS would simply corrupt the checksum silently. > > Could you confirm that you have netfilter rules (in particular NAT > rules) and that this goes away if you flush all your netfilter tables? > > Patrick, do we really have to zap the checksum on outbound NAT? Could > we update it instead? > > > I got this while running Dom0 only (no guests), with a > > BOINC/[EMAIL PROTECTED] application running on all 4 cores. > > > > changeset: 10649:8e55c5c11475 > > > > Build: x86_32p (pae). > > > > [ cut here ] > > kernel BUG at net/core/dev.c:1133! > > invalid opcode: [#1] > > SMP > > CPU:0 > > EIP:0061:[]Not tainted VLI > > EFLAGS: 00210297 (2.6.16.13-xen #12) > > EIP is at skb_gso_segment+0xf0/0x110 > > eax: ebx: 0003 ecx: 0002 edx: c06e2e00 > > esi: 0008 edi: cd9e32e0 ebp: c63a7900 esp: c0de5ad0 > > ds: 007b es: 007b ss: 0069 > > Process rosetta_5.25_i6 (pid: 8826, threadinfo=c0de4000 task=cb019560) > > Stack: <0>c8f69060 ffa3 0003 cd9e32e0 0002 c63a7900 > > c04dcfb0 > > cd9e32e0 0003 cd9e32e0 cf8e3000 cf8e3140 c04dd07e > > cd9e32e0 > > cf8e3000 cd9e32e0 cf8e3000 c04ec07e cd9e32e0 cf8e3000 > > c0895140 > > Call Trace: > > [] dev_gso_segment+0x30/0xb0 > > [] dev_hard_start_xmit+0x4e/0x110 > > [] __qdisc_run+0xbe/0x280 > > [] dev_queue_xmit+0x379/0x380 > > [] br_dev_queue_push_xmit+0xa4/0x140 > > [] br_nf_post_routing+0x102/0x1d0 > > [] br_nf_dev_queue_xmit+0x0/0x50 > > [] br_dev_queue_push_xmit+0x0/0x140 > > [] nf_iterate+0x6b/0xa0 > > [] br_dev_queue_push_xmit+0x0/0x140 > > [] br_dev_queue_push_xmit+0x0/0x140 > > [] nf_hook_slow+0x6e/0x120 > > [] br_dev_queue_push_xmit+0x0/0x140 > > [] br_forward_finish+0x60/0x70 > > [] br_dev_queue_push_xmit+0x0/0x140 > > [] br_nf_forward_finish+0x71/0x130 > > [] br_forward_finish+0x0/0x70 > > [] br_nf_forward_ip+0xf0/0x1a0 > > [] br_nf_forward_finish+0x0/0x130 > > [] br_forward_finish+0x0/0x70 > > [] nf_iterate+0x6b/0xa0 > > [] br_forward_finish+0x0/0x70 > > [] br_forward_finish+0x0/0x70 > > [] nf_hook_slow+0x6e/0x120 > > [] br_forward_finish+0x0/0x70 > > [] __br_forward+0x74/0x80 > > [] br_forward_finish+0x0/0x70 > > [] br_handle_frame_finish+0xd1/0x160 > > [] br_handle_frame_finish+0x0/0x160 > > [] br_nf_pre_routing_finish+0xfb/0x480 > > [] br_handle_frame_finish+0x0/0x160 > > [] br_nf_pre_routing_finish+0x0/0x480 > > [] ip_nat_in+0x43/0xc0 > > [] br_nf_pre_routing_finish+0x0/0x480 > > [] nf_iterate+0x6b/0xa0 > > [] br_nf_pre_routing_finish+0x0/0x480 > > [] br_nf_pre_routing_finish+0x0/0x480 > > [] nf_hook_slow+0x6e/0x120 > > [] br_nf_pre_routing_finish+0x0/0x480 > > [] br_nf_pre_routing+0x404/0x580 > > [] br_nf_pre_routing_finish+0x0/0x480 > > [] nf_iterate+0x6b/0xa0 > > [] br_handle_frame_finish+0x0/0x160 > > [] br_handle_frame_finish+0x0/0x160 > > [] nf_hook_slow+0x6e/0x120 > > [] br_handle_frame_finish+0x0/0x160 > > [] br_handle_frame+0x1e4/0x250 > > [] br_handle_frame_finish+0x0/0x160 > > [] netif_receive_skb+0x165/0x2a0 > > [] process_backlog+0xbf/0x180 > > [] net_rx_action+0x11f/0x1d0 > > [] __do_softirq+0x86/0x120 > > [] do_softirq+0x75/0x90 > > [] do_IRQ+0x1f/0x30 > > [] evtchn_do_upcall+0x90/0x100 > > [] hypervisor_callback+0x3d/0x48 > > Code: c2 2b 57 24 29 d0 8d 14 2a 89 87 94 00 00 00 89 57 60 8b 44 24 08 > > 83 c4 0c 5b 5e 5f 5d c3 0f 0 > > b 69 03 fe 8c 66 c0 e9 69 ff ff ff <0f> 0b 6d 04 e8 ab 6c c0 e9 3a ff ff > > ff 0f 0b 6c 04 e8 ab 6c c0 > > <0>Kernel panic - not syncing: Fatal exception in interrupt > > Cheers, - To unsubscribe from this list: send the line "unsubscribe netdev" in the body of a message to [EMAIL PROTECTED] More majordomo info at http://vger.kernel.org/majordomo-info.html
Re: [Xen-devel] kernel BUG at net/core/dev.c:1133!
Petersson, Mats <[EMAIL PROTECTED]> wrote: > Looks like the GSO is involved? It's certainly what crashed your machine :) It's probably not the guilty party though. Someone is passing through a TSO packet with checksum set to something other than CHECKSUM_HW. I bet it's netfilter and we just never noticed before because real NICS would simply corrupt the checksum silently. Could you confirm that you have netfilter rules (in particular NAT rules) and that this goes away if you flush all your netfilter tables? Patrick, do we really have to zap the checksum on outbound NAT? Could we update it instead? > I got this while running Dom0 only (no guests), with a > BOINC/[EMAIL PROTECTED] application running on all 4 cores. > > changeset: 10649:8e55c5c11475 > > Build: x86_32p (pae). > > [ cut here ] > kernel BUG at net/core/dev.c:1133! > invalid opcode: [#1] > SMP > CPU:0 > EIP:0061:[]Not tainted VLI > EFLAGS: 00210297 (2.6.16.13-xen #12) > EIP is at skb_gso_segment+0xf0/0x110 > eax: ebx: 0003 ecx: 0002 edx: c06e2e00 > esi: 0008 edi: cd9e32e0 ebp: c63a7900 esp: c0de5ad0 > ds: 007b es: 007b ss: 0069 > Process rosetta_5.25_i6 (pid: 8826, threadinfo=c0de4000 task=cb019560) > Stack: <0>c8f69060 ffa3 0003 cd9e32e0 0002 c63a7900 > c04dcfb0 > cd9e32e0 0003 cd9e32e0 cf8e3000 cf8e3140 c04dd07e > cd9e32e0 > cf8e3000 cd9e32e0 cf8e3000 c04ec07e cd9e32e0 cf8e3000 > c0895140 > Call Trace: > [] dev_gso_segment+0x30/0xb0 > [] dev_hard_start_xmit+0x4e/0x110 > [] __qdisc_run+0xbe/0x280 > [] dev_queue_xmit+0x379/0x380 > [] br_dev_queue_push_xmit+0xa4/0x140 > [] br_nf_post_routing+0x102/0x1d0 > [] br_nf_dev_queue_xmit+0x0/0x50 > [] br_dev_queue_push_xmit+0x0/0x140 > [] nf_iterate+0x6b/0xa0 > [] br_dev_queue_push_xmit+0x0/0x140 > [] br_dev_queue_push_xmit+0x0/0x140 > [] nf_hook_slow+0x6e/0x120 > [] br_dev_queue_push_xmit+0x0/0x140 > [] br_forward_finish+0x60/0x70 > [] br_dev_queue_push_xmit+0x0/0x140 > [] br_nf_forward_finish+0x71/0x130 > [] br_forward_finish+0x0/0x70 > [] br_nf_forward_ip+0xf0/0x1a0 > [] br_nf_forward_finish+0x0/0x130 > [] br_forward_finish+0x0/0x70 > [] nf_iterate+0x6b/0xa0 > [] br_forward_finish+0x0/0x70 > [] br_forward_finish+0x0/0x70 > [] nf_hook_slow+0x6e/0x120 > [] br_forward_finish+0x0/0x70 > [] __br_forward+0x74/0x80 > [] br_forward_finish+0x0/0x70 > [] br_handle_frame_finish+0xd1/0x160 > [] br_handle_frame_finish+0x0/0x160 > [] br_nf_pre_routing_finish+0xfb/0x480 > [] br_handle_frame_finish+0x0/0x160 > [] br_nf_pre_routing_finish+0x0/0x480 > [] ip_nat_in+0x43/0xc0 > [] br_nf_pre_routing_finish+0x0/0x480 > [] nf_iterate+0x6b/0xa0 > [] br_nf_pre_routing_finish+0x0/0x480 > [] br_nf_pre_routing_finish+0x0/0x480 > [] nf_hook_slow+0x6e/0x120 > [] br_nf_pre_routing_finish+0x0/0x480 > [] br_nf_pre_routing+0x404/0x580 > [] br_nf_pre_routing_finish+0x0/0x480 > [] nf_iterate+0x6b/0xa0 > [] br_handle_frame_finish+0x0/0x160 > [] br_handle_frame_finish+0x0/0x160 > [] nf_hook_slow+0x6e/0x120 > [] br_handle_frame_finish+0x0/0x160 > [] br_handle_frame+0x1e4/0x250 > [] br_handle_frame_finish+0x0/0x160 > [] netif_receive_skb+0x165/0x2a0 > [] process_backlog+0xbf/0x180 > [] net_rx_action+0x11f/0x1d0 > [] __do_softirq+0x86/0x120 > [] do_softirq+0x75/0x90 > [] do_IRQ+0x1f/0x30 > [] evtchn_do_upcall+0x90/0x100 > [] hypervisor_callback+0x3d/0x48 > Code: c2 2b 57 24 29 d0 8d 14 2a 89 87 94 00 00 00 89 57 60 8b 44 24 08 > 83 c4 0c 5b 5e 5f 5d c3 0f 0 > b 69 03 fe 8c 66 c0 e9 69 ff ff ff <0f> 0b 6d 04 e8 ab 6c c0 e9 3a ff ff > ff 0f 0b 6c 04 e8 ab 6c c0 > <0>Kernel panic - not syncing: Fatal exception in interrupt Cheers, -- Visit Openswan at http://www.openswan.org/ Email: Herbert Xu ~{PmV>HI~} <[EMAIL PROTECTED]> Home Page: http://gondor.apana.org.au/~herbert/ PGP Key: http://gondor.apana.org.au/~herbert/pubkey.txt - To unsubscribe from this list: send the line "unsubscribe netdev" in the body of a message to [EMAIL PROTECTED] More majordomo info at http://vger.kernel.org/majordomo-info.html