Hello all!

I've decided to try IP over IB. There are two servers (blade08 and blade09) which have eth0 and ib0 NICs. The problem is
that two CTs deployed on different servers can't see each other (check by ping 
command).
CT 152 deployed on host blade08 with ib0 IP address 10.1.36.18 can ping that 
remote host but can't
ping any CT deployed on that remote host.
Whereas CT 120 deployed on host blade09 with ib0 IP address 10.1.36.19 can't 
ping neither remote
host blade08 nor CT deployed on that remote host (CT152). Two servers can ping 
each other via ib0
interface (checked by 'ping -I ib0 <remote server IP>' command). Iptables is 
turned off on both
servers and all CTs.

The configuration is the following.
[root@blade09 ~]# ifconfig
eth0      Link encap:Ethernet  HWaddr **************************
          inet addr:192.93.36.239  Bcast:192.93.36.255  Mask:255.255.255.0
          inet6 addr: **********************/64 Scope:Link
          UP BROADCAST RUNNING MULTICAST  MTU:1500  Metric:1
          RX packets:364749 errors:0 dropped:0 overruns:0 frame:0
          TX packets:146504 errors:0 dropped:0 overruns:0 carrier:0
          collisions:0 txqueuelen:1000
          RX bytes:204560009 (195.0 MiB)  TX bytes:148608765 (141.7 MiB)
          Memory:c7420000-c743ffff

ib0       Link encap:InfiniBand  HWaddr 
*****************************************
          inet addr:10.1.36.19  Bcast:10.1.36.255  Mask:255.255.255.0
          inet6 addr: **********************/64 Scope:Link
          UP BROADCAST RUNNING MULTICAST  MTU:65520  Metric:1
          RX packets:15567403 errors:0 dropped:0 overruns:0 frame:0
          TX packets:34644634 errors:0 dropped:2098 overruns:0 carrier:0
          collisions:0 txqueuelen:256
          RX bytes:315873847681 (294.1 GiB)  TX bytes:360607196041 (335.8 GiB)

lo        Link encap:Local Loopback
          inet addr:127.0.0.1  Mask:255.0.0.0
          inet6 addr: ::1/128 Scope:Host
          UP LOOPBACK RUNNING  MTU:65536  Metric:1
          RX packets:12 errors:0 dropped:0 overruns:0 frame:0
          TX packets:12 errors:0 dropped:0 overruns:0 carrier:0
          collisions:0 txqueuelen:0
          RX bytes:1004 (1004.0 b)  TX bytes:1004 (1004.0 b)

venet0    Link encap:UNSPEC  HWaddr 
00-00-00-00-00-00-00-00-00-00-00-00-00-00-00-00
          inet6 addr: fe80::1/128 Scope:Link
          UP BROADCAST POINTOPOINT RUNNING NOARP  MTU:1500  Metric:1
          RX packets:46172203 errors:0 dropped:0 overruns:0 frame:0
          TX packets:46172173 errors:0 dropped:3 overruns:0 carrier:0
          collisions:0 txqueuelen:0
          RX bytes:49753326008 (46.3 GiB)  TX bytes:49753323488 (46.3 GiB)

The same configuration on the second server (blade08) but it has different IPs: 
eth0 -
192.93.36.238, ib0 - 10.1.36.18.

On each server I deployed CTs:
[root@blade08 ~]# vzlist -a
      CTID      NPROC STATUS    IP_ADDR         HOSTNAME
       152         11 running   10.1.36.52      ct152

[root@blade09 ~]# vzlist -a
      CTID      NPROC STATUS    IP_ADDR         HOSTNAME
       120         11 running   10.1.36.50      ct150
       121         11 running   10.1.36.51      ct151

[root@blade09 ~]# route -n
Kernel IP routing table
Destination     Gateway         Genmask         Flags Metric Ref    Use Iface
10.1.36.51      0.0.0.0         255.255.255.255 UH    0      0        0 venet0
10.1.36.50      0.0.0.0         255.255.255.255 UH    0      0        0 venet0
10.1.36.0       0.0.0.0         255.255.255.0   U     0      0        0 ib0
192.93.36.0     0.0.0.0         255.255.255.0   U     0      0        0 eth0
169.254.0.0     0.0.0.0         255.255.0.0     U     1002   0        0 eth0
0.0.0.0         192.93.36.1     0.0.0.0         UG    0      0        0 eth0

[root@blade08 ~]# route -n
Kernel IP routing table
Destination     Gateway         Genmask         Flags Metric Ref    Use Iface
10.1.36.52      0.0.0.0         255.255.255.255 UH    0      0        0 venet0
10.1.36.0       0.0.0.0         255.255.255.0   U     0      0        0 ib0
192.93.36.0     0.0.0.0         255.255.255.0   U     0      0        0 eth0
169.254.0.0     0.0.0.0         255.255.0.0     U     1002   0        0 eth0
0.0.0.0         192.93.36.1     0.0.0.0         UG    0      0        0 eth0


[root@blade08 ~]# grep -i forward /etc/sysctl.conf
# Controls IP packet forwarding
net.ipv4.ip_forward = 1

[root@blade09 ~]# grep -i forward /etc/sysctl.conf
# Controls IP packet forwarding
net.ipv4.ip_forward = 1

[root@blade08 ~]# grep ^VE_ROUTE_SRC_DEV /etc/vz/vz.conf
VE_ROUTE_SRC_DEV="ib0"

[root@blade09 ~]# grep ^VE_ROUTE_SRC_DEV /etc/vz/vz.conf
VE_ROUTE_SRC_DEV="ib0"


$ rpm -qa|egrep "vzctl|vzkernel"
vzkernel-2.6.32-042stab116.2.x86_64
vzctl-core-4.9.4-1.x86_64
vzctl-4.9.4-1.x86_64

All servers (blade08-09) and all CTs have CentOS 6.8 x86_64 deployed.

I would appreciate any hint on that issue.

Best regards,
Nikolay.


kna...@gmail.com wrote on 01.07.2016 10:36:
Hello all!

I am trying to evaluate a possibility to dynamically deploy HPC clusters in the 
cloud using
OpenVZ containers and Infiniband (IB). I couldn't found on the web a proper way 
to do that. What
I would like to achieve is the same functionality as one can get with ethernet 
NIC but with IB
device, i.e. share the same IB device for venet CTs' network interfaces. So I 
wonder if it's
possible at all to use IB for such kind of purposes. And if it is possible then 
what  steps needs
to be performed.

Best regards, Nikolay.
_______________________________________________
Users mailing list
Users@openvz.org
https://lists.openvz.org/mailman/listinfo/users

Reply via email to