Hi Jesse, thanks for your reply.
On Sat, May 13, 2023 at 2:41 AM Jesse Brandeburg
<[email protected]> wrote:
>
> On 5/11/2023 9:54 PM, Igor Cicimov wrote:
> > Hi,
> >
> > I have a problem with my 8086:1010 Intel Corporation 82546EB Gigabit
> > Ethernet Controller (Copper) dual port ethernet card and Ubuntu 22.04.2 LTS
> > using e1000 driver:
>
> This card is from 2003! :-) Nice that it's still running!
>
Time flies :-)
> ....
>
> Did you file a bug with Canonical against ubuntu or ask for help over
> there yet?
>
No, not yet. My assumption to be honest is that this card support has
maybe been removed in the driver or the kernel?
> > that I have configured in LACP bond0:
> >
> > # cat /proc/net/bonding/bond0
> > Ethernet Channel Bonding Driver: v5.15.0-69-generic
> >
> > Bonding Mode: IEEE 802.3ad Dynamic link aggregation
> > Transmit Hash Policy: layer2+3 (2)
> > MII Status: down
> > MII Polling Interval (ms): 100
> > Up Delay (ms): 100
> > Down Delay (ms): 100
> > Peer Notification Delay (ms): 0
> >
> > 802.3ad info
> > LACP active: on
> > LACP rate: fast
> > Min links: 0
> > Aggregator selection policy (ad_select): stable
> > System priority: 65535
> > System MAC address: MAC_BOND0
> > bond bond0 has no active aggregator
>
> Did you try bonding without MII link monitoring? I'm wondering if you're
> getting caught up in the ethtool transition to netlink for some reason.
>
No but that's a good idea.
>
> >
> > Slave Interface: eth1
> > MII Status: down
> > Speed: 1000 Mbps
> > Duplex: full
> > Link Failure Count: 0
> > Permanent HW addr: MAC_ETH1
> > Slave queue ID: 0
> > Aggregator ID: 1
> > Actor Churn State: churned
> > Partner Churn State: churned
> > Actor Churned Count: 1
> > Partner Churned Count: 1
> > details actor lacp pdu:
> > system priority: 65535
> > system mac address: MAC_BOND0
> > port key: 0
> > port priority: 255
> > port number: 1
> > port state: 71
> > details partner lacp pdu:
> > system priority: 65535
> > system mac address: 00:00:00:00:00:00
> > oper key: 1
> > port priority: 255
> > port number: 1
> > port state: 1
> >
> > Slave Interface: eth2
> > MII Status: down
> > Speed: 1000 Mbps
> > Duplex: full
> > Link Failure Count: 0
> > Permanent HW addr: MAC_ETH2
> > Slave queue ID: 0
> > Aggregator ID: 2
> > Actor Churn State: churned
> > Partner Churn State: churned
> > Actor Churned Count: 1
> > Partner Churned Count: 1
> > details actor lacp pdu:
> > system priority: 65535
> > system mac address: MAC_BOND0
> > port key: 0
> > port priority: 255
> > port number: 2
> > port state: 71
> > details partner lacp pdu:
> > system priority: 65535
> > system mac address: 00:00:00:00:00:00
> > oper key: 1
> > port priority: 255
> > port number: 1
> > port state: 1
> >
> > that is in state down of course since both interfaces have MII Status:
> > down. The dmesg shows:
> >
> > # dmesg | grep -E "bond0|eth[1|2]"
> > [ 42.999281] e1000 0000:01:0a.0 eth1: (PCI:33MHz:32-bit) MAC_ETH1
> > [ 42.999292] e1000 0000:01:0a.0 eth1: Intel(R) PRO/1000 Network Connection
> > [ 43.323358] e1000 0000:01:0a.1 eth2: (PCI:33MHz:32-bit) MAC_ETH2
> > [ 43.323366] e1000 0000:01:0a.1 eth2: Intel(R) PRO/1000 Network Connection
> > [ 65.617020] bonding: bond0 is being created...
> > [ 65.787883] 8021q: adding VLAN 0 to HW filter on device eth1
> > [ 67.790638] 8021q: adding VLAN 0 to HW filter on device eth2
> > [ 70.094511] 8021q: adding VLAN 0 to HW filter on device bond0
> > [ 70.558364] 8021q: adding VLAN 0 to HW filter on device eth1
> > [ 70.558675] bond0: (slave eth1): Enslaving as a backup interface with a
> > down link
> > [ 70.560050] 8021q: adding VLAN 0 to HW filter on device eth2
> > [ 70.560354] bond0: (slave eth2): Enslaving as a backup interface with a
> > down link
> >
> > So both eth1 and eth2 are UP and recognised, ethtool says "Link detected:
> > yes" but their links are DOWN. I have a confusing port type of FIBRE
> > reported by ethtool (capabilities reported by lshw are capabilities: pm
> > pcix msi cap_list rom ethernet physical fibre 1000bt-fd autonegotiation).
> > It is weird and I suspect some hardware or firmware issue. Any ideas are
> > welcome.
>
> You didn't post your bonding options enabled or bonding config file:
>
Sure, this is the content of /etc/network/interfaces that I've been
dragging since 12.04 or even maybe earlier, so many times I planned to
migrate to systemd or even better netplan but never got the time:
auto eth1
allow-hotplug eth1
allow-bond0 eth1
iface eth1 inet manual
bond-master bond0
auto eth2
allow-hotplug eth2
allow-bond0 eth2
iface eth2 inet manual
bond-master bond0
auto bond0
iface bond0 inet static
pre-up /sbin/ifconfig eth1 0.0.0.0 up || /bin/true && \
sleep 2 && /sbin/ifconfig eth2 0.0.0.0 up || /bin/true && sleep 2
post-up ifenslave bond0 eth1 eth2
pre-down ifenslave -d bond0 eth1 eth2
#there are several modes, this is also known as mode 4
bond-mode 802.3ad
bond-miimon 100
bond-lacp_rate fast
bond-xmit_hash_policy layer2+3
bond-downdelay 100
bond-updelay 100
bond-slaves none
address 172.128.1.129
netmask 255.255.255.0
Might be outdated though since there are some pre and post commands I
had to use due to some bug in previous ubuntu versions, not sure if
that hurts or not.
> did you try the use_carrier=1 option, it's the default but you're not
> setting it to zero are you??
>
Nope I haven't. As you say that's the default though and I'm not
setting it to zero.
> >
> > P.S: It is not the switch or the switch ports and it is not the cables
> > already tested that. The same setup, switch+cables+card was working fine up
> > to Ubuntu 18.04
>
> The
> Supported ports: [ FIBRE ]
>
> thing is strange, but it really shouldn't matter.
>
>
>
Cheers,
Igor
_______________________________________________
E1000-devel mailing list
[email protected]
https://lists.sourceforge.net/lists/listinfo/e1000-devel
To learn more about Intel Ethernet, visit
https://community.intel.com/t5/Ethernet-Products/bd-p/ethernet-products