Vincent ETIENNE wrote:
Hi,
Summary :
        Got this trace when one network interface come down or up in a 2
        interfaces bonding.     So far, system seems to survive to this problem
        and works fine.

I'm investigating a similar/possibly identical bug. Do you experience packet loss or throughput stalls, beyond just the loss of the interface that went down, when this happens?

        -- Chris

 Full description
During testing of bonding of 2 interfaces, i have seen this from
        time to time in my log file ( the problem doesn't arrive each
time but one in 3 or 4 try ). SYSTEM : 2 NIC card bond on interface bond0 : intel PRO/1000 (e1000 )
                                Broadcomm ( tg3 )
I have also try a 2.6.20 and 2.6.19 vanilla kernel ( identical problem but in onecase the system doesn't survive : that the reason the problem catch my attention ) Keywords ; network, bonding Version : Linux version 2.6.21-rc6-mm1 ([EMAIL PROTECTED]) (gcc version 4.1.1
        (Gentoo 4.1.1-r3)) #3 SMP Thu Apr 26 08:45:06 CEST 2007
Output of /var/log/messages Apr 26 11:09:34 jupiter2 e1000: eth0: e1000_watchdog_task: NIC Link
                is Down Apr 26 11:09:34 jupiter2 bonding: bond0: link status
                definitely down for interface eth0, disabling it
                Apr 26 11:09:34 jupiter2 bonding: bond0: making interface eth1 
the new
                active one.
                Apr 26 11:09:34 jupiter2 RTNL: assertion failed at 
net/ipv4/devinet.c
                (1055) Apr 26 11:09:34 jupiter2
                Apr 26 11:09:34 jupiter2 Call Trace:
Apr 26 11:09:34 jupiter2 <IRQ> [<ffffffff8049b49e>] inetdev_event+0x48/0x283
                Apr 26 11:09:34 jupiter2 [<ffffffff804c8731>] 
_spin_lock_bh+0x9/0x19
                Apr 26 11:09:34 jupiter2 [<ffffffff80473df1>] 
rt_run_flush+0x7e/0xaf
Apr 26 11:09:34 jupiter2 [<ffffffff8022bdd0>] notifier_call_chain+0x29/0x56 Apr 26 11:09:34 jupiter2 [<ffffffff804560cc>] dev_set_mac_address+0x53/0x59 Apr 26 11:09:34 jupiter2 [<ffffffff88006d8d>] bonding:alb_set_slave_mac_addr+0x41/0x6c Apr 26 11:09:34 jupiter2 [<ffffffff88007215>] bonding:alb_swap_mac_addr+0x91/0x165 Apr 26 11:09:34 jupiter2 [<ffffffff88002029>] bonding:bond_change_active_slave+0x227/0x382
                Apr 26 11:09:34 jupiter2 [<ffffffff880024c9>]
                bonding:bond_select_active_slave+0xb7/0xe5
Apr 26 11:09:34 jupiter2 [<ffffffff88004182>] bonding:bond_mii_monitor+0x3cd/0x41e Apr 26 11:09:34 jupiter2 [<ffffffff88003db5>] bonding:bond_mii_monitor+0x0/0x41e
                Apr 26 11:09:34 jupiter2 [<ffffffff80228714>]
 run_timer_softirq+0x130/0x19f
Apr 26 11:09:34 jupiter2[<ffffffff8022618a>] __do_softirq+0x55/0xc4 Apr 26 11:09:34 jupiter2 [<ffffffff8020a5ac>] call_softirq+0x1c/0x28
                Apr 26 11:09:34 jupiter2 [<ffffffff8020bead>] 
do_softirq+0x2c/0x7d
Apr 26 11:09:34 jupiter2 [<ffffffff80213b2a>] smp_apic_timer_interrupt+0x49/0x5f
                 Apr 26 11:09:34 jupiter2 [<ffffffff802088a3>] 
mwait_idle+0x0/0x45
                Apr 26 11:09:34 jupiter2 [<ffffffff8020a056>] 
apic_timer_interrupt+0x66/0x70
Apr 26 11:09:34 jupiter2 <EOI> [<ffffffff802088e5>] mwait_idle+0x42/0x45 Apr 26 11:09:34 jupiter2 [<ffffffff8020883f>] cpu_idle+0x51/0x70 Apr 26 11:09:34 jupiter2 [<ffffffff806369bd>] start_kernel+0x242/0x24e
                Apr 26 11:09:34 jupiter2 [<ffffffff80636146>] 
_sinittext+0x146/0x14a
other informations (ver_linux, lspci, ... ) available at http://mail1.vetienne.net/linux

I'm a bit worried by the message so any help will be greatly appreciated

        Vincent

-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Reply via email to