Configuration with versions Next email will have log files. 2 sites First site: Bayview 4 nodes BL460 gen9 with 4 x 10G nics Node 1-3 have not been changed since 4.3.2 upgrade. These nodes have the network sync issue and cannot migrate VMs. OS Version: RHEL - 7 - 6.1810.2.el7.centos OS Description: CentOS Linux 7 (Core) Kernel Version: 3.10.0 - 957.10.1.el7.x86_64 KVM Version: 2.12.0 - 18.el7_6.3.1 LIBVIRT Version: libvirt-4.5.0-10.el7_6.6 VDSM Version: vdsm-4.30.11-1.el7 SPICE Version: 0.14.0 - 6.el7_6.1 CEPH Version: librbd1-10.2.5-4.el7 Open vSwitch Version: openvswitch-2.10.1-3.el7 Kernel Features: PTI: 1, IBRS: 0, RETP: 1 VNC Encryption: Disabled
I evacuated node 4 and did update to 4.3.3, still had issues so tried ISO OVNode install. It does not have network sync issue. However I cannot upgrade all the nodes without scheduling cluster downtime. OS Version: RHEL - 7 - 6.1810.2.el7.centos OS Description: oVirt Node 4.3.3.1 Kernel Version: 3.10.0 - 957.10.1.el7.x86_64 KVM Version: 2.12.0 - 18.el7_6.3.1 LIBVIRT Version: libvirt-4.5.0-10.el7_6.6 VDSM Version: vdsm-4.30.13-1.el7 SPICE Version: 0.14.0 - 6.el7_6.1 CEPH Version: librbd1-10.2.5-4.el7 Open vSwitch Version: openvswitch-2.10.1-3.el7 Kernel Features: PTI: 1, IBRS: 0, RETP: 1, SSBD: 3 VNC Encryption: Disabled 1 engine, DL360g10. 4.3.3.5-1.el7 CentOS 7 patched up to today Storage: Dell Unity ISCSI What I see in the engine is that the first 3 nodes all have the public network listed as out of sync with the DC. I can migrate form node 4 to the other 3 nodes but cannot migrate off the other 3 nodes. I also cannot sync the network on the other three nodes. There have been no recent network changes. After the 4.3.3 upgrade I initially found some curious OVN errors in the logfile on nodes 3 and 4. Nodes 1 and 2 do not have these errors. However the engine did have 2 extra OVN ports defined. ovs-vsctl show be10cd3d-85fe-4985-9635-f447bfbc5e25 Bridge br-int fail_mode: secure Port "ovn-877214-0" Interface "ovn-877214-0" type: geneve options: {csum="true", key=flow, remote_ip="137.187.160.14"} error: "could not add network device ovn-877214-0 to ofproto (File exists)" Port "ovn-48e040-0" Interface "ovn-48e040-0" type: geneve options: {csum="true", key=flow, remote_ip="137.187.160.18"} Port "ovn-d6eaa1-0" Interface "ovn-d6eaa1-0" type: geneve options: {csum="true", key=flow, remote_ip="137.187.160.13"} error: "could not add network device ovn-d6eaa1-0 to ofproto (File exists)" Port br-int Interface br-int type: internal Port "ovn-f0f789-0" Interface "ovn-f0f789-0" type: geneve options: {csum="true", key=flow, remote_ip="137.187.160.13"} Second site:Harbor I upgraded to 4.3.2 and stopped any upgrades until I could deal with the migration and restarting VDSMD issues. Nodes are 3 supermicro 1U with 2x10G nics All 3 nodes are the same and from the 4.3.2 update. OS Version: RHEL - 7 - 6.1810.2.el7.centos OS Description: CentOS Linux 7 (Core) Kernel Version: 3.10.0 - 957.10.1.el7.x86_64 KVM Version: 2.12.0 - 18.el7_6.3.1 LIBVIRT Version: libvirt-4.5.0-10.el7_6.6 VDSM Version: vdsm-4.30.11-1.el7 SPICE Version: 0.14.0 - 6.el7_6.1 GlusterFS Version: [N/A] CEPH Version: librbd1-10.2.5-4.el7 Open vSwitch Version: openvswitch-2.10.1-3.el7 Kernel Features: PTI: 1, IBRS: 0, RETP: 1 VNC Encryption: Disabled 1 engine DL360 g10 with 2x10G nics running 4.3.2.1-1.el7 Storage: Dell Unity ISCSI The only issue with Harbor site while running 4.3.2 is that when I migrate a node it never finishes in the engine until I restart VDSMD on the original host. It did not exhibit this issue with 4.2.x
_______________________________________________ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/3XORQANVT7TZCOBXMBO7SH3IZ2XWDHX4/