On Thu, 23 Aug 2018 13:51:39 -0400
Douglas Duckworth <dod2...@med.cornell.edu> wrote:

> THANKS!
> 
> ib0 now up with NFS storage back on this hypervisor
> 

Thanks for letting us know.

> Though how do I make it a transfer network?  I don't see an option.
> 

I do not understand the meaning of "transfer network".
The network interface to use for NFS results from the routing tables of
the host.
In "Compute > Clusters > Clustername > Logical Networks > Manage
Networks" network roles for some kind of loads can be assigned, but not
for NFS access.


> Thanks,
> 
> Douglas Duckworth, MSc, LFCS
> HPC System Administrator
> Scientific Computing Unit
> Weill Cornell Medicine
> 1300 York - LC-502
> E: d...@med.cornell.edu
> O: 212-746-6305
> F: 212-746-8690
> 
> 
> On Thu, Aug 23, 2018 at 11:12 AM, Douglas Duckworth
> <dod2...@med.cornell.edu
> > wrote:  
> 
> > Hi Dominik
> >
> > Yes, the network-script was created by our Ansible role that deploys
> > CentOS hosts.  It pulls the IP from DNS then templates the script
> > and copies to host.
> >
> > I will try this oVirt step then see if it works!
> >
> > Thanks,
> >
> > Douglas Duckworth, MSc, LFCS
> > HPC System Administrator
> > Scientific Computing Unit
> > Weill Cornell Medicine
> > 1300 York - LC-502
> > E: d...@med.cornell.edu
> > O: 212-746-6305
> > F: 212-746-8690
> >
> >
> > On Thu, Aug 23, 2018 at 11:09 AM, Dominik Holler
> > <dhol...@redhat.com> wrote:
> >  
> >> Is ifcfg-ib0 created before adding the host?
> >> Can ib0 be reconfigured using engine, e.g. by
> >> "Compute > Hosts > hostx > Network Interfaces > Setup Host
> >> Networks"? If this some kind of self-hosted engine?
> >>
> >> On Thu, 23 Aug 2018 09:30:59 -0400
> >> Douglas Duckworth <dod2...@med.cornell.edu> wrote:
> >>  
> >> > Here's a link to the files:
> >> >
> >> > https://urldefense.proofpoint.com/v2/url?u=https-3A__bit.ly_  
> >> 2wjZ6Vo&d=DwICAg&c=lb62iw4YL4RFalcE2hQUQealT9-RXrryqt9KZX2qu
> >> 2s&r=2Fzhh_78OGspKQpl_e-CbhH6xUjnRkaqPFUS2wTJ2cw&m=Y25-
> >> OOvgu58jlC82-fzBeNIpQ7ZscoHznffUhqE6EBM&s=QQXlC9Tisa60TvimyS
> >> 3BnFDCaDF7VPD8eCzT-Fke-p0&e=  
> >> >
> >> > Thank you!
> >> >
> >> > Thanks,
> >> >
> >> > Douglas Duckworth, MSc, LFCS
> >> > HPC System Administrator
> >> > Scientific Computing Unit
> >> > Weill Cornell Medicine
> >> > 1300 York - LC-502
> >> > E: d...@med.cornell.edu
> >> > O: 212-746-6305
> >> > F: 212-746-8690
> >> >
> >> >
> >> > On Thu, Aug 23, 2018 at 6:51 AM, Dominik Holler
> >> > <dhol...@redhat.com> wrote:
> >> >  
> >> > > Would you please share the vdsm.log and the supervdsm.log from
> >> > > this host?
> >> > >
> >> > > On Wed, 22 Aug 2018 11:36:09 -0400
> >> > > Douglas Duckworth <dod2...@med.cornell.edu> wrote:
> >> > >  
> >> > > > Hi
> >> > > >
> >> > > > I keep losing ib0 connection on hypervisor after adding host
> >> > > > to engine. This makes the host not really work since NFS
> >> > > > will be mounted over ib0.
> >> > > >
> >> > > > I don't really understand why this occurs.
> >> > > >
> >> > > > OS:
> >> > > >
> >> > > > [root@ovirt-hv2 ~]# cat /etc/redhat-release
> >> > > > CentOS Linux release 7.5.1804 (Core)
> >> > > >
> >> > > > Here's the network script:
> >> > > >
> >> > > > [root@ovirt-hv2 ~]#
> >> > > > cat /etc/sysconfig/network-scripts/ifcfg-ib0 DEVICE=ib0
> >> > > > BOOTPROTO=static
> >> > > > IPADDR=172.16.0.207
> >> > > > NETMASK=255.255.255.0
> >> > > > ONBOOT=yes
> >> > > > ZONE=public
> >> > > >
> >> > > > When I try "ifup"
> >> > > >
> >> > > > [root@ovirt-hv2 ~]# ifup ib0
> >> > > > Error: Connection activation failed: No suitable device
> >> > > > found for this connection.
> >> > > >
> >> > > > The error in syslog:
> >> > > >
> >> > > > Aug 22 11:31:50 ovirt-hv2 kernel: IPv4: martian source
> >> > > > 172.16.0.87 from 172.16.0.49, on dev ib0
> >> > > > Aug 22 11:31:53 ovirt-hv2 NetworkManager[1070]: <info>
> >> > > > [1534951913.7486] audit: op="connection-activate"
> >> > > > uuid="2ab4abde-b8a5-6cbc-19b1-2bfb193e4e89" name="System ib0"
> >> > > > result="fail" reason="No suitable device found for this
> >> > > > connection.
> >> > > >
> >> > > > As you can see media state up:
> >> > > >
> >> > > > [root@ovirt-hv2 ~]# ip a
> >> > > > 1: lo: <LOOPBACK,UP,LOWER_UP> mtu 65536 qdisc noqueue state
> >> > > > UNKNOWN group default qlen 1000
> >> > > >     link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00
> >> > > >     inet 127.0.0.1/8 scope host lo
> >> > > >        valid_lft forever preferred_lft forever
> >> > > > 2: em1: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc mq
> >> > > > master ovirtmgmt state UP group default qlen 1000
> >> > > >     link/ether 50:9a:4c:89:d3:81 brd ff:ff:ff:ff:ff:ff
> >> > > > 3: em2: <NO-CARRIER,BROADCAST,MULTICAST,UP> mtu 1500 qdisc mq
> >> > > > state DOWN group default qlen 1000
> >> > > >     link/ether 50:9a:4c:89:d3:82 brd ff:ff:ff:ff:ff:ff
> >> > > > 4: p1p1: <NO-CARRIER,BROADCAST,MULTICAST,UP> mtu 1500 qdisc
> >> > > > mq state DOWN group default qlen 1000
> >> > > >     link/ether b4:96:91:13:ea:68 brd ff:ff:ff:ff:ff:ff
> >> > > > 5: p1p2: <NO-CARRIER,BROADCAST,MULTICAST,UP> mtu 1500 qdisc
> >> > > > mq state DOWN group default qlen 1000
> >> > > >     link/ether b4:96:91:13:ea:6a brd ff:ff:ff:ff:ff:ff
> >> > > > 6: idrac: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc
> >> > > > pfifo_fast state UNKNOWN group default qlen 1000
> >> > > >     link/ether 50:9a:4c:89:d3:84 brd ff:ff:ff:ff:ff:ff
> >> > > >     inet 169.254.0.2/16 brd 169.254.255.255 scope global
> >> > > > idrac valid_lft forever preferred_lft forever
> >> > > > 7: ib0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 2044 qdisc mq
> >> > > > state UP group default qlen 256
> >> > > >     link/infiniband
> >> > > > a0:00:02:08:fe:80:00:00:00:00:00:00:ec:0d:9a:03:00:1d:13:41
> >> > > > brd
> >> > > > 00:ff:ff:ff:ff:12:40:1b:ff:ff:00:00:00:00:00:00:ff:ff:ff:ff
> >> > > > 8: ovs-system: <BROADCAST,MULTICAST> mtu 1500 qdisc noop
> >> > > > state DOWN group default qlen 1000 link/ether
> >> > > > 12:b4:30:22:39:5b brd ff:ff:ff:ff:ff:ff 9: br-int:
> >> > > > <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN group
> >> > > > default qlen 1000 link/ether 3e:32:e6:66:98:49 brd
> >> > > > ff:ff:ff:ff:ff:ff 25: ovirtmgmt:
> >> > > > <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc noqueue
> >> > > > state UP group default qlen 1000 link/ether
> >> > > > 50:9a:4c:89:d3:81 brd ff:ff:ff:ff:ff:ff inet 10.0.0.183/16
> >> > > > brd 10.0.255.255 scope global ovirtmgmt valid_lft forever
> >> > > > preferred_lft forever 26: genev_sys_6081:
> >> > > > <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 65000 qdisc noqueue
> >> > > > master ovs-system state UNKNOWN group default qlen 1000
> >> > > > link/ether aa:32:82:1b:01:d9 brd ff:ff:ff:ff:ff:ff
> >> > > > 27: ;vdsmdummy;: <BROADCAST,MULTICAST> mtu 1500 qdisc noop
> >> > > > state DOWN group default qlen 1000 link/ether
> >> > > > 32:ff:5d:b8:c2:b4 brd ff:ff:ff:ff:ff:ff
> >> > > >
> >> > > > The card is FDR:
> >> > > >
> >> > > > [root@ovirt-hv2 ~]# lspci -v | grep Mellanox
> >> > > > 01:00.0 Network controller: Mellanox Technologies MT27500
> >> > > > Family [ConnectX-3]
> >> > > > Subsystem: Mellanox Technologies Device 0051
> >> > > >
> >> > > > Latest OFED driver:
> >> > > >
> >> > > > [root@ovirt-hv2 ~]# /etc/init.d/openibd status
> >> > > >
> >> > > >   HCA driver loaded
> >> > > >
> >> > > > Configured IPoIB devices:
> >> > > > ib0
> >> > > >
> >> > > > Currently active IPoIB devices:
> >> > > > ib0
> >> > > > Configured Mellanox EN devices:
> >> > > >
> >> > > > Currently active Mellanox devices:
> >> > > > ib0
> >> > > >
> >> > > > The following OFED modules are loaded:
> >> > > >
> >> > > >   rdma_ucm
> >> > > >   rdma_cm
> >> > > >   ib_ipoib
> >> > > >   mlx4_core
> >> > > >   mlx4_ib
> >> > > >   mlx4_en
> >> > > >   mlx5_core
> >> > > >   mlx5_ib
> >> > > >   ib_uverbs
> >> > > >   ib_umad
> >> > > >   ib_ucm
> >> > > >   ib_cm
> >> > > >   ib_core
> >> > > >   mlxfw
> >> > > >   mlx5_fpga_tools
> >> > > >
> >> > > > I can add an IP to ib0 using "ip addr" though I need Network
> >> > > > Manager to work with ib0.
> >> > > >
> >> > > >
> >> > > > Thanks,
> >> > > >
> >> > > > Douglas Duckworth, MSc, LFCS
> >> > > > HPC System Administrator
> >> > > > Scientific Computing Unit
> >> > > > Weill Cornell Medicine
> >> > > > 1300 York - LC-502
> >> > > > E: d...@med.cornell.edu
> >> > > > O: 212-746-6305
> >> > > > F: 212-746-8690  
> >> > >
> >> > >  
> >>
> >>  
> >  
_______________________________________________
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/BMOIFY6IVK4LFBD2MOXP3ZEBY37TCIWD/

Reply via email to