On 3/31/2021 9:12 PM, duluxoz wrote:
Hi Greg,

Nope, not the answer (or not the whole answer) - I changed eno3's IP address to a different subnet and I've still got the exact same issue.

Any other ideas?

It seems like it should work but is not so that's when I drag out
the tcpdump utility and start trying to figure out where the traffic
is getting to (or not).

Run tcpdump sessions on all of the relevant interfaces while attempting
your ping.  Hopefully you'll see some point where an interface is
sending and/or forwarding a packet but the receiving interface isn't
getting it.  That might provide some clue.

You may try upgrading to a more recent version of OVS, just as an
experiment.  Release 2.12 is getting a little long in the tooth.

- Greg



On 01/04/2021 08:56, Gregory Rose wrote:


On 3/30/2021 7:01 PM, duluxoz wrote:
Hi Greg,

Thanks for getting back to me.


Hi,

I think (at least part of) the problem is that you have two IP addresses
on the same subnet assigned to two interfaces.   But in Linux the OS
owns the IP address, not the interface.  Perhaps try a different subnet?

Otherwise I think it should work.

- Greg


ip link show:

~~~
1: lo: <LOOPBACK,UP,LOWER_UP> mtu 65536 qdisc noqueue state UNKNOWN mode DEFAULT group default qlen 1000
     link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00
2: eno1: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc mq master team1 state UP mode DEFAULT group default qlen 1000
     link/ether ac:1f:6b:3c:cc:e8 brd ff:ff:ff:ff:ff:ff
3: eno2: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc mq master team1 state UP mode DEFAULT group default qlen 1000
     link/ether ac:1f:6b:3c:cc:e8 brd ff:ff:ff:ff:ff:ff
4: eno3: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc mq state UP mode DEFAULT group default qlen 1000
     link/ether 00:1b:21:e4:a6:f4 brd ff:ff:ff:ff:ff:ff
5: ovs-system: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN mode DEFAULT group default qlen 1000
     link/ether 6a:e9:90:3b:d0:47 brd ff:ff:ff:ff:ff:ff
6: ovs1: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc noqueue state UNKNOWN mode DEFAULT group default qlen 1000
     link/ether ac:1f:6b:3c:cc:e8 brd ff:ff:ff:ff:ff:ff
7: team1: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc noqueue master ovs-system state UP mode DEFAULT group default qlen 1000
     link/ether ac:1f:6b:3c:cc:e8 brd ff:ff:ff:ff:ff:ff
~~~

ip addr show:

~~~
1: lo: <LOOPBACK,UP,LOWER_UP> mtu 65536 qdisc noqueue state UNKNOWN group default qlen 1000
     link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00
     inet 127.0.0.1/8 scope host lo
        valid_lft forever preferred_lft forever
     inet6 ::1/128 scope host
        valid_lft forever preferred_lft forever
2: eno1: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc mq master team1 state UP group default qlen 1000
     link/ether ac:1f:6b:3c:cc:e8 brd ff:ff:ff:ff:ff:ff
3: eno2: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc mq master team1 state UP group default qlen 1000
     link/ether ac:1f:6b:3c:cc:e8 brd ff:ff:ff:ff:ff:ff
4: eno3: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc mq state UP group default qlen 1000
     link/ether 00:1b:21:e4:a6:f4 brd ff:ff:ff:ff:ff:ff
     inet 192.168.0.16/24 brd 192.168.0.255 scope global noprefixroute eno3
        valid_lft forever preferred_lft forever
     inet6 fe80::21b:21ff:fee4:a6f4/64 scope link
        valid_lft forever preferred_lft forever
5: ovs-system: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN group default qlen 1000
     link/ether 6a:e9:90:3b:d0:47 brd ff:ff:ff:ff:ff:ff
6: ovs1: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc noqueue state UNKNOWN group default qlen 1000
     link/ether ac:1f:6b:3c:cc:e8 brd ff:ff:ff:ff:ff:ff
7: team1: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc noqueue master ovs-system state UP group default qlen 1000
     link/ether ac:1f:6b:3c:cc:e8 brd ff:ff:ff:ff:ff:ff
     inet6 fe80::ae1f:6bff:fe3c:cce8/64 scope link
        valid_lft forever preferred_lft forever
~~~

ovs-vsctl show

~~~
49b8d126-5e44-4ebc-8daa-8a73683d08a5
     Bridge "ovs1"
         Port "ovs1"
             Interface "ovs1"
                 type: internal
         Port "team1"
             Interface "team1"
     ovs_version: "2.12.0"
~~~

Cheers

Dulux-Oz

On 31/03/2021 03:21, Gregory Rose wrote:


On 3/29/2021 12:00 AM, duluxoz wrote:
Hi All,

Sorry if this is the wrong place to post this and/or I missed the answer in the doco.

I'm new to OVS and have been struggling for days with this issue - I hope someone can help.

I have a Centos 8 box loaded with kvm/qemu/libvert, NetworkManager-ovs, and ovs (v12.2). The box has three nics in it (en01, en02, en03). I am using eno3 as a management-nic (with an ip address of 192.168.0.16) for the box itself, and the other two nics set up as a lacp team (team1, no ip address, with team1-slave1 (eno1) and team1-slave2 (eno2)).

Everything is working correctly, and my physical switch has the two team nics both active and load-balanced.

I created an ovs (ovs1) and set it up with a static ipv4 address (192.168.0.17). I also modified the ifcfg-team1 file to include the references to ovs1.

But I can't ping ovs1 and nmcli says that ovs1 is down - and yes, I have tried bringing it up (nmcli c up ovs1) - which gives an error.

What is the output of 'ip link show', 'ip addr show' and 'ovs-vsctl
show'?

- Greg


I have include the relevant ifcfg-* files below.

Can someone please tell/show me what I'm doing wrong - thanks.

ifcfg-ovs1:

~~~
DEVICE=ovs1
ONBOOT=yes
DEVICETYPE=ovs
TYPE=OVSBridge
BOOTPROTO=none
IPADDR=192.168.0.17
PREFIX=24
DEFROUTE=yes
GATEWAY=192.168.0.1
DNS1=192.168.0.2
DNS2=192.168.0.3
IPV6INIT=no
HOTPLUG=no
~~~

ifcfg-team1:

~~~
TEAM_CONFIG=$'{\"runner\": {\"name\": \"lacp\", \"active\": true, \"fast_rate\": false, \"agg_select_policy\": \"lacp_prio\"}, \"link_watch\": {\"name\": \"ethtool\"}}'
PROXY_METHOD=none
BROWSER_ONLY=no
BOOTPROTO=none
DEFROUTE=yes
IPV4_FAILURE_FATAL=no
IPV6INIT=no
NAME=team1
DEVICE=team1
ONBOOT=yes
DEVICETYPE=ovs
TYPE=OVSPort
OVS_BRIDGE=ovs1
~~~

ifcfg-team1-slave1:

~~~
ETHTOOL_OPTS="autoneg off speed 1000 duplex full"
NAME=team1-slave1
DEVICE=eno1
ONBOOT=yes
TEAM_MASTER=team1
DEVICETYPE=TeamPort
~~~

ifcfg-team1-slave2:

~~~
ETHTOOL_OPTS="autoneg off speed 1000 duplex full"
NAME=team1-slave2
DEVICE=eno2
ONBOOT=yes
TEAM_MASTER=team1
DEVICETYPE=TeamPort
~~~

Finally, nmcli d status give:

~~~

DEVICE TYPE           STATE         CONNECTION
eno1         ethernet          connected       team1-slave1
eno2         ethernet          connected       team1-slave2
eno3 ethernet          connected       eno3
team1 team               connected       team1
ovs1         ovs-interface  disconnected  --
lo             loopback         unmanaged    --
ovs1        ovs-bridge       unmanaged    --
team1      ovs-port           unmanaged    --
ovs1        ovs-port           unmanaged    --
~~~

Cheers

Dulux-Oz



_______________________________________________
discuss mailing list
disc...@openvswitch.org
https://mail.openvswitch.org/mailman/listinfo/ovs-discuss

_______________________________________________
discuss mailing list
disc...@openvswitch.org
https://mail.openvswitch.org/mailman/listinfo/ovs-discuss

Reply via email to