Hi Jesse, thanks for your reply.

On Sat, May 13, 2023 at 2:41 AM Jesse Brandeburg
<jesse.brandeb...@intel.com> wrote:
>
> On 5/11/2023 9:54 PM, Igor Cicimov wrote:
> > Hi,
> >
> > I have a problem with my 8086:1010 Intel Corporation 82546EB Gigabit
> > Ethernet Controller (Copper) dual port ethernet card and Ubuntu 22.04.2 LTS
> > using e1000 driver:
>
> This card is from 2003! :-) Nice that it's still running!
>

Time flies :-)

> ....
>
> Did you file a bug with Canonical against ubuntu or ask for help over
> there yet?
>

No, not yet. My assumption to be honest is that this card support has
maybe been removed in the driver or the kernel?

> > that I have configured in LACP bond0:
> >
> > # cat /proc/net/bonding/bond0
> > Ethernet Channel Bonding Driver: v5.15.0-69-generic
> >
> > Bonding Mode: IEEE 802.3ad Dynamic link aggregation
> > Transmit Hash Policy: layer2+3 (2)
> > MII Status: down
> > MII Polling Interval (ms): 100
> > Up Delay (ms): 100
> > Down Delay (ms): 100
> > Peer Notification Delay (ms): 0
> >
> > 802.3ad info
> > LACP active: on
> > LACP rate: fast
> > Min links: 0
> > Aggregator selection policy (ad_select): stable
> > System priority: 65535
> > System MAC address: MAC_BOND0
> > bond bond0 has no active aggregator
>
> Did you try bonding without MII link monitoring? I'm wondering if you're
> getting caught up in the ethtool transition to netlink for some reason.
>

No but that's a good idea.

>
> >
> > Slave Interface: eth1
> > MII Status: down
> > Speed: 1000 Mbps
> > Duplex: full
> > Link Failure Count: 0
> > Permanent HW addr: MAC_ETH1
> > Slave queue ID: 0
> > Aggregator ID: 1
> > Actor Churn State: churned
> > Partner Churn State: churned
> > Actor Churned Count: 1
> > Partner Churned Count: 1
> > details actor lacp pdu:
> >     system priority: 65535
> >     system mac address: MAC_BOND0
> >     port key: 0
> >     port priority: 255
> >     port number: 1
> >     port state: 71
> > details partner lacp pdu:
> >     system priority: 65535
> >     system mac address: 00:00:00:00:00:00
> >     oper key: 1
> >     port priority: 255
> >     port number: 1
> >     port state: 1
> >
> > Slave Interface: eth2
> > MII Status: down
> > Speed: 1000 Mbps
> > Duplex: full
> > Link Failure Count: 0
> > Permanent HW addr: MAC_ETH2
> > Slave queue ID: 0
> > Aggregator ID: 2
> > Actor Churn State: churned
> > Partner Churn State: churned
> > Actor Churned Count: 1
> > Partner Churned Count: 1
> > details actor lacp pdu:
> >     system priority: 65535
> >     system mac address: MAC_BOND0
> >     port key: 0
> >     port priority: 255
> >     port number: 2
> >     port state: 71
> > details partner lacp pdu:
> >     system priority: 65535
> >     system mac address: 00:00:00:00:00:00
> >     oper key: 1
> >     port priority: 255
> >     port number: 1
> >     port state: 1
> >
> > that is in state down of course since both interfaces have MII Status:
> > down. The dmesg shows:
> >
> > # dmesg | grep -E "bond0|eth[1|2]"
> > [   42.999281] e1000 0000:01:0a.0 eth1: (PCI:33MHz:32-bit) MAC_ETH1
> > [   42.999292] e1000 0000:01:0a.0 eth1: Intel(R) PRO/1000 Network Connection
> > [   43.323358] e1000 0000:01:0a.1 eth2: (PCI:33MHz:32-bit) MAC_ETH2
> > [   43.323366] e1000 0000:01:0a.1 eth2: Intel(R) PRO/1000 Network Connection
> > [   65.617020] bonding: bond0 is being created...
> > [   65.787883] 8021q: adding VLAN 0 to HW filter on device eth1
> > [   67.790638] 8021q: adding VLAN 0 to HW filter on device eth2
> > [   70.094511] 8021q: adding VLAN 0 to HW filter on device bond0
> > [   70.558364] 8021q: adding VLAN 0 to HW filter on device eth1
> > [   70.558675] bond0: (slave eth1): Enslaving as a backup interface with a
> > down link
> > [   70.560050] 8021q: adding VLAN 0 to HW filter on device eth2
> > [   70.560354] bond0: (slave eth2): Enslaving as a backup interface with a
> > down link
> >
> > So both eth1 and eth2 are UP and recognised, ethtool says "Link detected:
> > yes" but their links are DOWN. I have a confusing port type of FIBRE
> > reported by ethtool (capabilities reported by lshw are capabilities: pm
> > pcix msi cap_list rom ethernet physical fibre 1000bt-fd autonegotiation).
> > It is weird and I suspect some hardware or firmware issue. Any ideas are
> > welcome.
>
> You didn't post your bonding options enabled or bonding config file:
>

Sure, this is the content of /etc/network/interfaces that I've been
dragging since 12.04 or even maybe earlier, so many times I planned to
migrate to systemd or even better netplan but never got the time:

auto eth1
allow-hotplug eth1
allow-bond0 eth1
iface eth1 inet manual
    bond-master bond0

auto eth2
allow-hotplug eth2
allow-bond0 eth2
iface eth2 inet manual
    bond-master bond0

auto bond0
iface bond0 inet static
    pre-up /sbin/ifconfig eth1 0.0.0.0 up || /bin/true && \
           sleep 2 && /sbin/ifconfig eth2 0.0.0.0 up || /bin/true && sleep 2
    post-up ifenslave bond0 eth1 eth2
    pre-down ifenslave -d bond0 eth1 eth2
    #there are several modes, this is also known as mode 4
    bond-mode 802.3ad
    bond-miimon 100
    bond-lacp_rate fast
    bond-xmit_hash_policy layer2+3
    bond-downdelay 100
    bond-updelay 100
    bond-slaves none
    address 172.128.1.129
    netmask 255.255.255.0

Might be outdated though since there are some pre and post commands I
had to use due to some bug in previous ubuntu versions, not sure if
that hurts or not.

> did you try the use_carrier=1 option, it's the default but you're not
> setting it to zero are you??
>

Nope I haven't. As you say that's the default though and I'm not
setting it to zero.

> >
> > P.S: It is not the switch or the switch ports and it is not the cables
> > already tested that. The same setup, switch+cables+card was working fine up
> > to Ubuntu 18.04
>
> The
>     Supported ports: [ FIBRE ]
>
> thing is strange, but it really shouldn't matter.
>
>
>
Cheers,
Igor


_______________________________________________
E1000-devel mailing list
E1000-devel@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/e1000-devel
To learn more about Intel Ethernet, visit 
https://community.intel.com/t5/Ethernet-Products/bd-p/ethernet-products

Reply via email to