On Mon, 10 Dec 2018 18:09:40 +0000 Douglas Duckworth <dod2...@med.cornell.edu> wrote:
> Hi Dominik, > > I have added LACP bond network to all hosts and renamed the Hosted Engine > using "/usr/share/ovirt-engine/setup/bin/ovirt-engine-rename." > > However, I am still missing the option to assign Migration and Management > network roles to this new bond. > > Can you advise where I can find this option? > You cannot assign this role to the host interface directly, but to the network, which is assigned to the interface, in "Compute > Clusters > Clustername > Logical Networks > Manage Networks" > Thanks, > > Douglas Duckworth, MSc, LFCS > HPC System Administrator > Scientific Computing Unit<https://scu.med.cornell.edu> > Weill Cornell Medicine > 1300 York Avenue > New York, NY 10065 > E: d...@med.cornell.edu<mailto:d...@med.cornell.edu> > O: 212-746-6305 > F: 212-746-8690 > > > On Fri, Aug 24, 2018 at 11:52 AM Dominik Holler > <dhol...@redhat.com<mailto:dhol...@redhat.com>> wrote: > On Fri, 24 Aug 2018 09:46:25 -0400 > Douglas Duckworth <dod2...@med.cornell.edu<mailto:dod2...@med.cornell.edu>> > wrote: > > > Sorry, I mean "migration network" for moving live migration traffic. > > > > You have to create a new logical network in > "Network > Networks > New" > and assign this to ib0 in > "Compute > Hosts > hostname > Network Interfaces > Setup Host Networks". > After this you can assign a role to this network in > "Compute > Clusters > Clustername > Logical Networks > Manage Networks" > > > > FDR infiniband much faster than 1Gb network which currently acts as > > migration network, vm network, display network, mgmt network, etc. > > > > Thanks, > > > > Douglas Duckworth, MSc, LFCS > > HPC System Administrator > > Scientific Computing Unit > > Weill Cornell Medicine > > 1300 York - LC-502 > > E: d...@med.cornell.edu<mailto:d...@med.cornell.edu> > > O: 212-746-6305 > > F: 212-746-8690 > > > > > > On Fri, Aug 24, 2018 at 9:36 AM, Dominik Holler > > <dhol...@redhat.com<mailto:dhol...@redhat.com>> > > wrote: > > > > > On Thu, 23 Aug 2018 13:51:39 -0400 > > > Douglas Duckworth > > > <dod2...@med.cornell.edu<mailto:dod2...@med.cornell.edu>> wrote: > > > > > > > THANKS! > > > > > > > > ib0 now up with NFS storage back on this hypervisor > > > > > > > > > > Thanks for letting us know. > > > > > > > Though how do I make it a transfer network? I don't see an > > > > option. > > > > > > I do not understand the meaning of "transfer network". > > > The network interface to use for NFS results from the routing > > > tables of the host. > > > In "Compute > Clusters > Clustername > Logical Networks > Manage > > > Networks" network roles for some kind of loads can be assigned, but > > > not for NFS access. > > > > > > > > > > Thanks, > > > > > > > > Douglas Duckworth, MSc, LFCS > > > > HPC System Administrator > > > > Scientific Computing Unit > > > > Weill Cornell Medicine > > > > 1300 York - LC-502 > > > > E: d...@med.cornell.edu<mailto:d...@med.cornell.edu> > > > > O: 212-746-6305 > > > > F: 212-746-8690 > > > > > > > > > > > > On Thu, Aug 23, 2018 at 11:12 AM, Douglas Duckworth > > > > <dod2...@med.cornell.edu<mailto:dod2...@med.cornell.edu> > > > > > wrote: > > > > > > > > > Hi Dominik > > > > > > > > > > Yes, the network-script was created by our Ansible role that > > > > > deploys CentOS hosts. It pulls the IP from DNS then templates > > > > > the script and copies to host. > > > > > > > > > > I will try this oVirt step then see if it works! > > > > > > > > > > Thanks, > > > > > > > > > > Douglas Duckworth, MSc, LFCS > > > > > HPC System Administrator > > > > > Scientific Computing Unit > > > > > Weill Cornell Medicine > > > > > 1300 York - LC-502 > > > > > E: d...@med.cornell.edu<mailto:d...@med.cornell.edu> > > > > > O: 212-746-6305 > > > > > F: 212-746-8690 > > > > > > > > > > > > > > > On Thu, Aug 23, 2018 at 11:09 AM, Dominik Holler > > > > > <dhol...@redhat.com<mailto:dhol...@redhat.com>> wrote: > > > > > > > > > >> Is ifcfg-ib0 created before adding the host? > > > > >> Can ib0 be reconfigured using engine, e.g. by > > > > >> "Compute > Hosts > hostx > Network Interfaces > Setup Host > > > > >> Networks"? If this some kind of self-hosted engine? > > > > >> > > > > >> On Thu, 23 Aug 2018 09:30:59 -0400 > > > > >> Douglas Duckworth > > > > >> <dod2...@med.cornell.edu<mailto:dod2...@med.cornell.edu>> wrote: > > > > >> > > > > >> > Here's a link to the files: > > > > >> > > > > > >> > https://urldefense.proofpoint.com/v2/url?u=https-3A__bit.ly_ > > > > >> 2wjZ6Vo&d=DwICAg&c=lb62iw4YL4RFalcE2hQUQealT9-RXrryqt9KZX2qu > > > > >> 2s&r=2Fzhh_78OGspKQpl_e-CbhH6xUjnRkaqPFUS2wTJ2cw&m=Y25- > > > > >> OOvgu58jlC82-fzBeNIpQ7ZscoHznffUhqE6EBM&s=QQXlC9Tisa60TvimyS > > > > >> 3BnFDCaDF7VPD8eCzT-Fke-p0&e= > > > > >> > > > > > >> > Thank you! > > > > >> > > > > > >> > Thanks, > > > > >> > > > > > >> > Douglas Duckworth, MSc, LFCS > > > > >> > HPC System Administrator > > > > >> > Scientific Computing Unit > > > > >> > Weill Cornell Medicine > > > > >> > 1300 York - LC-502 > > > > >> > E: d...@med.cornell.edu<mailto:d...@med.cornell.edu> > > > > >> > O: 212-746-6305 > > > > >> > F: 212-746-8690 > > > > >> > > > > > >> > > > > > >> > On Thu, Aug 23, 2018 at 6:51 AM, Dominik Holler > > > > >> > <dhol...@redhat.com<mailto:dhol...@redhat.com>> wrote: > > > > >> > > > > > >> > > Would you please share the vdsm.log and the supervdsm.log > > > > >> > > from this host? > > > > >> > > > > > > >> > > On Wed, 22 Aug 2018 11:36:09 -0400 > > > > >> > > Douglas Duckworth > > > > >> > > <dod2...@med.cornell.edu<mailto:dod2...@med.cornell.edu>> wrote: > > > > >> > > > > > > >> > > > Hi > > > > >> > > > > > > > >> > > > I keep losing ib0 connection on hypervisor after adding > > > > >> > > > host to engine. This makes the host not really work > > > > >> > > > since NFS will be mounted over ib0. > > > > >> > > > > > > > >> > > > I don't really understand why this occurs. > > > > >> > > > > > > > >> > > > OS: > > > > >> > > > > > > > >> > > > [root@ovirt-hv2 ~]# cat /etc/redhat-release > > > > >> > > > CentOS Linux release 7.5.1804 (Core) > > > > >> > > > > > > > >> > > > Here's the network script: > > > > >> > > > > > > > >> > > > [root@ovirt-hv2 ~]# > > > > >> > > > cat /etc/sysconfig/network-scripts/ifcfg-ib0 DEVICE=ib0 > > > > >> > > > BOOTPROTO=static > > > > >> > > > IPADDR=172.16.0.207 > > > > >> > > > NETMASK=255.255.255.0 > > > > >> > > > ONBOOT=yes > > > > >> > > > ZONE=public > > > > >> > > > > > > > >> > > > When I try "ifup" > > > > >> > > > > > > > >> > > > [root@ovirt-hv2 ~]# ifup ib0 > > > > >> > > > Error: Connection activation failed: No suitable device > > > > >> > > > found for this connection. > > > > >> > > > > > > > >> > > > The error in syslog: > > > > >> > > > > > > > >> > > > Aug 22 11:31:50 ovirt-hv2 kernel: IPv4: martian source > > > > >> > > > 172.16.0.87 from 172.16.0.49, on dev ib0 > > > > >> > > > Aug 22 11:31:53 ovirt-hv2 NetworkManager[1070]: <info> > > > > >> > > > [1534951913.7486] audit: op="connection-activate" > > > > >> > > > uuid="2ab4abde-b8a5-6cbc-19b1-2bfb193e4e89" name="System > > > > >> > > > ib0" result="fail" reason="No suitable device found for > > > > >> > > > this connection. > > > > >> > > > > > > > >> > > > As you can see media state up: > > > > >> > > > > > > > >> > > > [root@ovirt-hv2 ~]# ip a > > > > >> > > > 1: lo: <LOOPBACK,UP,LOWER_UP> mtu 65536 qdisc noqueue > > > > >> > > > state UNKNOWN group default qlen 1000 > > > > >> > > > link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00 > > > > >> > > > inet 127.0.0.1/8<http://127.0.0.1/8> scope host lo > > > > >> > > > valid_lft forever preferred_lft forever > > > > >> > > > 2: em1: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc > > > > >> > > > mq master ovirtmgmt state UP group default qlen 1000 > > > > >> > > > link/ether 50:9a:4c:89:d3:81 brd ff:ff:ff:ff:ff:ff > > > > >> > > > 3: em2: <NO-CARRIER,BROADCAST,MULTICAST,UP> mtu 1500 > > > > >> > > > qdisc mq state DOWN group default qlen 1000 > > > > >> > > > link/ether 50:9a:4c:89:d3:82 brd ff:ff:ff:ff:ff:ff > > > > >> > > > 4: p1p1: <NO-CARRIER,BROADCAST,MULTICAST,UP> mtu 1500 > > > > >> > > > qdisc mq state DOWN group default qlen 1000 > > > > >> > > > link/ether b4:96:91:13:ea:68 brd ff:ff:ff:ff:ff:ff > > > > >> > > > 5: p1p2: <NO-CARRIER,BROADCAST,MULTICAST,UP> mtu 1500 > > > > >> > > > qdisc mq state DOWN group default qlen 1000 > > > > >> > > > link/ether b4:96:91:13:ea:6a brd ff:ff:ff:ff:ff:ff > > > > >> > > > 6: idrac: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 > > > > >> > > > qdisc pfifo_fast state UNKNOWN group default qlen 1000 > > > > >> > > > link/ether 50:9a:4c:89:d3:84 brd ff:ff:ff:ff:ff:ff > > > > >> > > > inet 169.254.0.2/16<http://169.254.0.2/16> brd > > > > >> > > > 169.254.255.255 scope global > > > > >> > > > idrac valid_lft forever preferred_lft forever > > > > >> > > > 7: ib0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 2044 qdisc > > > > >> > > > mq state UP group default qlen 256 > > > > >> > > > link/infiniband > > > > >> > > > a0:00:02:08:fe:80:00:00:00:00:00:00:ec:0d:9a:03:00:1d:13:41 > > > > >> > > > brd > > > > >> > > > 00:ff:ff:ff:ff:12:40:1b:ff:ff:00:00:00:00:00:00:ff:ff:ff:ff > > > > >> > > > 8: ovs-system: <BROADCAST,MULTICAST> mtu 1500 qdisc noop > > > > >> > > > state DOWN group default qlen 1000 link/ether > > > > >> > > > 12:b4:30:22:39:5b brd ff:ff:ff:ff:ff:ff 9: br-int: > > > > >> > > > <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN > > > > >> > > > group default qlen 1000 link/ether 3e:32:e6:66:98:49 brd > > > > >> > > > ff:ff:ff:ff:ff:ff 25: ovirtmgmt: > > > > >> > > > <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc noqueue > > > > >> > > > state UP group default qlen 1000 link/ether > > > > >> > > > 50:9a:4c:89:d3:81 brd ff:ff:ff:ff:ff:ff inet > > > > >> > > > 10.0.0.183/16<http://10.0.0.183/16> brd 10.0.255.255 scope > > > > >> > > > global ovirtmgmt > > > > >> > > > valid_lft forever preferred_lft forever 26: > > > > >> > > > genev_sys_6081: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu > > > > >> > > > 65000 qdisc noqueue master ovs-system state UNKNOWN > > > > >> > > > group default qlen 1000 link/ether aa:32:82:1b:01:d9 brd > > > > >> > > > ff:ff:ff:ff:ff:ff 27: ;vdsmdummy;: <BROADCAST,MULTICAST> > > > > >> > > > mtu 1500 qdisc noop state DOWN group default qlen 1000 > > > > >> > > > link/ether 32:ff:5d:b8:c2:b4 brd ff:ff:ff:ff:ff:ff > > > > >> > > > > > > > >> > > > The card is FDR: > > > > >> > > > > > > > >> > > > [root@ovirt-hv2 ~]# lspci -v | grep Mellanox > > > > >> > > > 01:00.0 Network controller: Mellanox Technologies MT27500 > > > > >> > > > Family [ConnectX-3] > > > > >> > > > Subsystem: Mellanox Technologies Device 0051 > > > > >> > > > > > > > >> > > > Latest OFED driver: > > > > >> > > > > > > > >> > > > [root@ovirt-hv2 ~]# /etc/init.d/openibd status > > > > >> > > > > > > > >> > > > HCA driver loaded > > > > >> > > > > > > > >> > > > Configured IPoIB devices: > > > > >> > > > ib0 > > > > >> > > > > > > > >> > > > Currently active IPoIB devices: > > > > >> > > > ib0 > > > > >> > > > Configured Mellanox EN devices: > > > > >> > > > > > > > >> > > > Currently active Mellanox devices: > > > > >> > > > ib0 > > > > >> > > > > > > > >> > > > The following OFED modules are loaded: > > > > >> > > > > > > > >> > > > rdma_ucm > > > > >> > > > rdma_cm > > > > >> > > > ib_ipoib > > > > >> > > > mlx4_core > > > > >> > > > mlx4_ib > > > > >> > > > mlx4_en > > > > >> > > > mlx5_core > > > > >> > > > mlx5_ib > > > > >> > > > ib_uverbs > > > > >> > > > ib_umad > > > > >> > > > ib_ucm > > > > >> > > > ib_cm > > > > >> > > > ib_core > > > > >> > > > mlxfw > > > > >> > > > mlx5_fpga_tools > > > > >> > > > > > > > >> > > > I can add an IP to ib0 using "ip addr" though I need > > > > >> > > > Network Manager to work with ib0. > > > > >> > > > > > > > >> > > > > > > > >> > > > Thanks, > > > > >> > > > > > > > >> > > > Douglas Duckworth, MSc, LFCS > > > > >> > > > HPC System Administrator > > > > >> > > > Scientific Computing Unit > > > > >> > > > Weill Cornell Medicine > > > > >> > > > 1300 York - LC-502 > > > > >> > > > E: d...@med.cornell.edu<mailto:d...@med.cornell.edu> > > > > >> > > > O: 212-746-6305 > > > > >> > > > F: 212-746-8690 > > > > >> > > > > > > >> > > > > > > >> > > > > >> > > > > > > > > > > > > _______________________________________________ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/HJWPDXXSO2FXL633LVG4ZXG2NFLOSMK7/