On Mon, Nov 19, 2018 at 7:17 PM Gregory Rose <gvrose8...@gmail.com> wrote:

>
> Hi Siva,
>
> One more request  - I need to see the underlying network configuration
> of the hypervisor running the two VMs.
> Are both VMs on the same machine?  If so then just the network
> configuration of the base machine
> running the VMs, otherwise the network configuration of each base
> machine running their perspective
> VM.
>
> This is turning into quite the investigation and I apologize that it is
> taking so long.  Please bear with me
> if you can and we'll see if we can't get this problem solved.  I've seen
> some puzzling bugs before and
> this one is turning out to be one of the best.  Or worst.... depends on
> your outlook.  :)
>
> Thanks for all your help so far!
>
> - Greg
>

Hi Greg,

Both the VMs run on same hypervisor in my setup. Created VMs and virtual
networks using virsh commands. Virsh XMLs for networks look like below

[user@hyp1 ] virsh net-dumpxml route1
<network connections='2'>
  <name>route1</name>
  <uuid>2c935aaf-ebde-5b76-a903-4fccb115ff75</uuid>
  <forward mode='route'/>
  <bridge name='testbr1' stp='on' delay='0'/>
  <mac address='42:54:00:84:4e:04'/>
  <ip address='20.20.0.1' netmask='255.255.255.0'>
    <dhcp>
      <range start='20.20.0.2' end='20.20.0.254'/>
    </dhcp>
  </ip>
</network>

[user@hyp1 ]  network virsh net-dumpxml route2
<network connections='2'>
  <name>route2</name>
  <uuid>2c935baf-ebde-5b76-a903-4fccb115ff75</uuid>
  <forward mode='route'/>
  <bridge name='testbr2' stp='on' delay='0'/>
  <mac address='42:54:10:84:4e:04'/>
  <ip address='30.30.0.1' netmask='255.255.255.0'>
    <dhcp>
      <range start='30.30.0.2' end='30.30.0.254'/>
    </dhcp>
  </ip>
</network>

Each VM is connected to both the networks.

Some network configuration of the hypervisor.

[user@hyp-1] ip a
1: lo: <LOOPBACK,UP,LOWER_UP> mtu 65536 qdisc noqueue state UNKNOWN group
default qlen 1
    link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00
    inet 127.0.0.1/8 scope host lo
       valid_lft forever preferred_lft forever
    inet6 ::1/128 scope host
       valid_lft forever preferred_lft forever
2: enp5s0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1400 qdisc pfifo_fast
state UP group default qlen 1000
    link/ether <mac> brd ff:ff:ff:ff:ff:ff
    inet A.B.C.D/24 brd X.Y.Z.W scope global dynamic enp5s0
       valid_lft 318349sec preferred_lft 318349sec
3: virbr0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc noqueue state
UP group default qlen 1000
    link/ether fe:54:00:0a:d3:70 brd ff:ff:ff:ff:ff:ff
    inet 192.168.122.1/24 brd 192.168.122.255 scope global virbr0
       valid_lft forever preferred_lft forever
4: virbr0-nic: <BROADCAST,MULTICAST> mtu 1500 qdisc pfifo_fast state DOWN
group default qlen 1000
    link/ether 52:54:00:94:4e:04 brd ff:ff:ff:ff:ff:ff
11: docker0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc noqueue state
UP group default
    link/ether 02:42:89:28:db:a5 brd ff:ff:ff:ff:ff:ff
    inet 172.17.0.1/16 scope global docker0
       valid_lft forever preferred_lft forever
    inet6 fe80::42:89ff:fe28:dba5/64 scope link
       valid_lft forever preferred_lft forever
96: vboxnet0: <NO-CARRIER,BROADCAST,MULTICAST,UP> mtu 1500 qdisc pfifo_fast
state DOWN group default qlen 1000
    link/ether 0a:00:27:00:00:00 brd ff:ff:ff:ff:ff:ff
    inet 192.168.99.1/24 brd 192.168.99.255 scope global vboxnet0
       valid_lft forever preferred_lft forever
    inet6 fe80::800:27ff:fe00:0/64 scope link
       valid_lft forever preferred_lft forever
193: testbr0: <NO-CARRIER,BROADCAST,MULTICAST,UP> mtu 1500 qdisc noqueue
state DOWN group default qlen 1000
    link/ether 00:00:00:00:00:00 brd ff:ff:ff:ff:ff:ff
    inet 10.10.0.1/24 brd 10.10.0.255 scope global testbr0
       valid_lft forever preferred_lft forever
194: testbr0-nic: <BROADCAST,MULTICAST> mtu 1500 qdisc pfifo_fast state
DOWN group default qlen 1000
    link/ether 42:54:00:94:4e:04 brd ff:ff:ff:ff:ff:ff
227: testbr1: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc noqueue
state UP group default qlen 1000
    link/ether fe:54:00:05:93:7c brd ff:ff:ff:ff:ff:ff
    inet 20.20.0.1/24 brd 20.20.0.255 scope global testbr1
       valid_lft forever preferred_lft forever
228: testbr1-nic: <BROADCAST,MULTICAST> mtu 1500 qdisc pfifo_fast state
DOWN group default qlen 1000
    link/ether 42:54:00:84:4e:04 brd ff:ff:ff:ff:ff:ff
229: testbr2: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc noqueue
state UP group default qlen 1000
    link/ether fe:54:00:79:ef:92 brd ff:ff:ff:ff:ff:ff
    inet 30.30.0.1/24 brd 30.30.0.255 scope global testbr2
       valid_lft forever preferred_lft forever
230: testbr2-nic: <BROADCAST,MULTICAST> mtu 1500 qdisc pfifo_fast state
DOWN group default qlen 1000
    link/ether 42:54:10:84:4e:04 brd ff:ff:ff:ff:ff:ff
231: vnet0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc pfifo_fast
master virbr0 state UNKNOWN group default qlen 1000
    link/ether fe:54:00:0a:d3:70 brd ff:ff:ff:ff:ff:ff
    inet6 fe80::fc54:ff:fe0a:d370/64 scope link
       valid_lft forever preferred_lft forever
232: vnet1: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc pfifo_fast
master testbr2 state UNKNOWN group default qlen 1000
    link/ether fe:54:00:b8:05:be brd ff:ff:ff:ff:ff:ff
    inet6 fe80::fc54:ff:feb8:5be/64 scope link
       valid_lft forever preferred_lft forever
233: vnet2: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc pfifo_fast
master testbr1 state UNKNOWN group default qlen 1000
    link/ether fe:54:00:f0:64:37 brd ff:ff:ff:ff:ff:ff
    inet6 fe80::fc54:ff:fef0:6437/64 scope link
       valid_lft forever preferred_lft forever
234: vnet3: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc pfifo_fast
master virbr0 state UNKNOWN group default qlen 1000
    link/ether fe:54:00:56:cb:89 brd ff:ff:ff:ff:ff:ff
    inet6 fe80::fc54:ff:fe56:cb89/64 scope link
       valid_lft forever preferred_lft forever
235: vnet4: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc pfifo_fast
master testbr2 state UNKNOWN group default qlen 1000
    link/ether fe:54:00:79:ef:92 brd ff:ff:ff:ff:ff:ff
    inet6 fe80::fc54:ff:fe79:ef92/64 scope link
       valid_lft forever preferred_lft forever
236: vnet5: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc pfifo_fast
master testbr1 state UNKNOWN group default qlen 1000
    link/ether fe:54:00:05:93:7c brd ff:ff:ff:ff:ff:ff
    inet6 fe80::fc54:ff:fe05:937c/64 scope link
       valid_lft forever preferred_lft forever
[user@hyp-1] ip route
default via A.B.C.D dev enp5s0  proto static  metric 100
10.10.0.0/24 dev testbr0  proto kernel  scope link  src 10.10.0.1 linkdown
20.20.0.0/24 dev testbr1  proto kernel  scope link  src 20.20.0.1
30.30.0.0/24 dev testbr2  proto kernel  scope link  src 30.30.0.1
A.B.C.D via P.Q.R.S dev enp5s0  proto dhcp  metric 100
X.Y.Z.W dev enp5s0  proto kernel  scope link  src A.B.C.D  metric 100
172.17.0.0/16 dev docker0  proto kernel  scope link  src 172.17.0.1
192.168.99.0/24 dev vboxnet0  proto kernel  scope link  src 192.168.99.1
linkdown
192.168.122.0/24 dev virbr0  proto kernel  scope link  src 192.168.122.1

I am not completely sure what you meant by network configuration. Kindly
let me know if you are looking for something more specific.

There is one strange behavior that I observed on the VMs. I am able to ping
across networks if I use IP address but it does not work if I use interface
name directly.

[root@vm2 ~]# ip addr  show eth1
3: eth1: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc pfifo_fast state
UP qlen 1000
    link/ether 52:54:00:79:ef:92 brd ff:ff:ff:ff:ff:ff
    inet 30.30.0.193/24 brd 30.30.0.255 scope global dynamic eth1
       valid_lft 2728sec preferred_lft 2728sec
    inet6 fe80::5054:ff:fe79:ef92/64 scope link
       valid_lft forever preferred_lft forever
[root@vm2 ~]# ping 20.20.0.183 -I eth1
PING 20.20.0.183 (20.20.0.183) from 30.30.0.193 eth1: 56(84) bytes of data.
^C
--- 20.20.0.183 ping statistics ---
7 packets transmitted, 0 received, 100% packet loss, time 6000ms

[root@vm2 ~]# ping 20.20.0.183 -I 30.30.0.193
PING 20.20.0.183 (20.20.0.183) from 30.30.0.193 : 56(84) bytes of data.
64 bytes from 20.20.0.183: icmp_seq=1 ttl=64 time=0.766 ms
64 bytes from 20.20.0.183: icmp_seq=2 ttl=64 time=0.561 ms
64 bytes from 20.20.0.183: icmp_seq=3 ttl=64 time=0.605 ms
64 bytes from 20.20.0.183: icmp_seq=4 ttl=64 time=0.537 ms
64 bytes from 20.20.0.183: icmp_seq=5 ttl=64 time=0.607 ms
64 bytes from 20.20.0.183: icmp_seq=6 ttl=64 time=0.618 ms
64 bytes from 20.20.0.183: icmp_seq=7 ttl=64 time=0.624 ms
^C
--- 20.20.0.183 ping statistics ---
7 packets transmitted, 7 received, 0% packet loss, time 6000ms
rtt min/avg/max/mdev = 0.537/0.616/0.766/0.075 ms
[root@vm2 ~]#

I don't the reason behind this and I will need to understand this when I
get some time. Just letting you know if it might make any difference in
your setup.

Siva Teja.
_______________________________________________
discuss mailing list
disc...@openvswitch.org
https://mail.openvswitch.org/mailman/listinfo/ovs-discuss

Reply via email to