Re: does i need compat-rdma package

2014-12-24 Thread Hiroyuki Sato
Hello Vasilly Tolstov

This URL might be help.
http://www.rdmamojo.com

There are RDMA configuration for many OSs. (RHEL, Ubuntu, SLES..)

--
Hiroyuki Sato



2014-12-24 17:19 GMT+09:00 Or Gerlitz :
> On 12/23/2014 4:47 PM, Vasiliy Tolstov wrote:
>>
>> Hello. I want to develop some software under linux 3.14 and using
>> infiniband libraries.
>
>
> You can use the inbox libraries and install them through your distro package
> installer, with RHEL, Fedora and such
> it would be just
>
> $ yum groupinstall "infiniband support"
>
> to get you the set of required RPMS [1] SLES and Ubuntu should have similar
> means, no need to use OFED, compat, etc.
>
> Or.
>
> [1] you mentioned ConnectX-3, so your basic needs are libmlx4, libibverbs
> and librdmacm
>
>> I'm download latest ofed (3.12 ?) and see that
>> some packages i don't need (i'm use only mellanox connect x3)
>> But i can't understand in case of linux 3.14 does i need compat-rdma ?
>>
>>
>
> --
> To unsubscribe from this list: send the line "unsubscribe linux-rdma" in
> the body of a message to majord...@vger.kernel.org
> More majordomo info at  http://vger.kernel.org/majordomo-info.html



-- 
Hiroyuki Sato
--
To unsubscribe from this list: send the line "unsubscribe linux-rdma" in
the body of a message to majord...@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html


Re: [Q] How to tranfer a file which is over 2GB(2^31) size in RDMA network?

2012-07-03 Thread Hiroyuki Sato
Hello Sean

Thank you for your information.

I'll try it.

P.S.
I'm reading rockets(librdmacm) commit log for study Infiniband. :-)


2012/7/4 Hefty, Sean :
>> Hello Parav.Pandit
>>
>> Thank you for your advice.
>>
>> I'll try it.
>
> You can also look at rsockets in the latest librdmacm library.  You'd need to 
> download and build the library yourself, since rsockets is not yet available 
> in any release.  But there's a sample program (rcopy) that will copy a file 
> between two systems over that API.
>
> I haven't tried using it to copy a very large files, but in theory it should 
> work.  (Hey, maybe you'll get lucky!)  If you do try it and hit into any 
> problems, please let me know.
>
> - Sean



-- 
Hiroyuki Sato
--
To unsubscribe from this list: send the line "unsubscribe linux-rdma" in
the body of a message to majord...@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html


Re: [Q] How to tranfer a file which is over 2GB(2^31) size in RDMA network?

2012-07-03 Thread Hiroyuki Sato
Hello Parav.Pandit

Thank you for your advice.

I'll try it.

--
Hiroyuki Sato

2012/7/3  :
>
>
>> -Original Message-
>> From: linux-rdma-ow...@vger.kernel.org [mailto:linux-rdma-
>> ow...@vger.kernel.org] On Behalf Of Hiroyuki Sato
>> Sent: Monday, July 02, 2012 7:20 PM
>> To: linux-rdma
>> Subject: [Q] How to tranfer a file which is over 2GB(2^31) size in RDMA
>> network?
>>
>> Dear developers.
>>
>> I'm writing simple file transfer program.
>>
>> I would like to know about the following.
>> Q: How to tranfer a file which is over 2GB(2^31) size  in RDMA network?
>>
>> Please imagine to transfer whole DVD(4.7GB) file via RDMA networok.
>>
>> The maximum message size of RC is 2^31.
>> so I can't transfer it with one RDMA message.
>>
>> Maybe It must split multiple parts.
>> When I sent first part with the following sequence, how to transfer second
>> part?
>>
>> Do I have to call qp ibv_destroy_qp and recreate new qp?
>> Or can I reuse current qp?
>>
> You can use the same QP to send multiple messages of same or different size.
> Receive side needs to have sufficient memory to receive 2GB of data, which 
> you should have posted using post_recv().
> Since you are going to divide a file/consumer application data to multiple 
> RDMA messages, make sure sender and receiver agrees to a message size of 
> interest.
> This is the simplest way to do via RDMA_SEND and RECV buffers.
>
> You can further do it via sharing the stag before invoking post_send/recv and 
> use RDMA Write or READ operations and don't need to synchronize the message 
> size.
> Sender can send/split data into one or more WRITE messages followed by ending 
> it with write_immidiate or send_immidiate to notify the peer of file transfer 
> done.
>
> In a third way, you can void writing the application and use SDP protocol to 
> send your file via FTP/SFTP, in which FTP server and client application 
> sockets use the SDP sockets instead of TCP sockets.
>
>
>> I'm looking for similar example but I can't find it.
>> Any information is welcome.
>>
>> Thank you for your advice.
>>
>> Sincerely
>>
>> --
>> Hiroyuki Sato.
>>
>>
>> File transfer sequence (1st part)
>>
>>   ibv_open_device
>>   ibv_alloc_pd
>>   ibv_reg_mr
>>   ibv_create_cq
>>   ibv_create_qp
>>   ibv_modify_qp(RESET->INIT)
>>   ibv_post_recv
>>   exchange lid, sid,qpn
>>   ibv_connect
>>   ibv_modify_qp(INIT->RTR)
>>   ibv_modify_qp(RTR->RTS)
>>   ibv_post_send
>>   ...
>>
>>
>>
>> Environment
>>  OS Scientific Linux 6.2
>>  OFED: 1.5.4.1
>>
>> --
>> Hiroyuki Sato
>> --
>> To unsubscribe from this list: send the line "unsubscribe linux-rdma" in the
>> body of a message to majord...@vger.kernel.org More majordomo info at
>> http://vger.kernel.org/majordomo-info.html



-- 
Hiroyuki Sato
--
To unsubscribe from this list: send the line "unsubscribe linux-rdma" in
the body of a message to majord...@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html


[Q] How to tranfer a file which is over 2GB(2^31) size in RDMA network?

2012-07-02 Thread Hiroyuki Sato
Dear developers.

I'm writing simple file transfer program.

I would like to know about the following.
Q: How to tranfer a file which is over 2GB(2^31) size  in RDMA network?

Please imagine to transfer whole DVD(4.7GB) file via RDMA networok.

The maximum message size of RC is 2^31.
so I can't transfer it with one RDMA message.

Maybe It must split multiple parts.
When I sent first part with the following sequence,
how to transfer second part?

Do I have to call qp ibv_destroy_qp and recreate new qp?
Or can I reuse current qp?

I'm looking for similar example but I can't find it.
Any information is welcome.

Thank you for your advice.

Sincerely

--
Hiroyuki Sato.


File transfer sequence (1st part)

  ibv_open_device
  ibv_alloc_pd
  ibv_reg_mr
  ibv_create_cq
  ibv_create_qp
  ibv_modify_qp(RESET->INIT)
  ibv_post_recv
  exchange lid, sid,qpn
  ibv_connect
  ibv_modify_qp(INIT->RTR)
  ibv_modify_qp(RTR->RTS)
  ibv_post_send
  ...



Environment
 OS Scientific Linux 6.2
 OFED: 1.5.4.1

-- 
Hiroyuki Sato
--
To unsubscribe from this list: send the line "unsubscribe linux-rdma" in
the body of a message to majord...@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html


[Q] Mellanox ConnectX QDR on SandyBridge motherboard works only PCIex GEN1 speed.

2011-05-16 Thread Hiroyuki Sato
Dears developers.

We are now investigating Mellanox ConnectX QDR on SandyBridge
motherboard(P67/H67/C206),
and have a serious performance problem on this configuration.

Any information is appreciated.

* Questions

  (1) kerenel parameters.

What parameter should I change to detect ConnectX GEN2 speed??
with vanilla kernel??

It is mean that, enable CONFIG_ parameters, or change module
parameters, apply patch .. and so on

  (2) What kernel function should I check to work ConnectX with PCIex
GEN2 speed?? on C206 chipset.

check kernel output, execute some commands, add printk in functions

* Problem

  (1) ConnectX works only PCI express GEN1 speed (2.5GT/s) on
CentOS5.5 and Ubuntu 10.10 on P67/H67 chipset.

This is lspci output. (complete output attached below)

 LnkSta: Speed 2.5GT/s, Width x8, TrErr- Train- SlotClk-
DLActive- BWMgmt- ABWMgmt-

But, Scientific Linux6, and Ubuntu 10.10 works GEN2(5.0GT/s) speed
on same motherboard.

And I tried to build vanilla kernel on Ubuntu 10.04 with default
configuration.
The kerenel newer than Ubuntu 10.10, but It still works only GEN1 speeed.


  (2) ConnectX works only PCI express GEN1 speed on Scientific Linux6,
CentOS 5.5 Ubuntu 10.10 and 10.04


* Environment (one of example)

  - OFED 1.5.3.1
  - Ubuntu 10.4 x64 server, and CentOS 5.5
  - ASUS P8P67-WS-Revolution
  - 1-port ConnectX QDR : Mellanox Technologies MT26428 [ConnectX VPI
PCIe 2.0 5GT/s - IB QDR / 10GigE] (rev b0)


* lspci output

  04:00.0 InfiniBand: Mellanox Technologies MT26428 [ConnectX VPI PCIe
2.0 5GT/s - IB QDR / 10GigE] (rev b0)
  Subsystem: Mellanox Technologies Device 0022
  Control: I/O- Mem+ BusMaster+ SpecCycle- MemWINV- VGASnoop-
ParErr- Stepping- SERR- FastB2B- DisINTx+
  Status: Cap+ 66MHz- UDF- FastB2B- ParErr- DEVSEL=fast >TAbort-
SERR- 
  Capabilities: [9c] MSI-X: Enable+ Mask- TabSize=256
   Vector table: BAR=0 offset=0007c000
   PBA: BAR=0 offset=0007d000
  Capabilities: [60] Express (v2) Endpoint, MSI 00
   DevCap: MaxPayload 256 bytes, PhantFunc 0, Latency L0s
<64ns, L1 unlimited
ExtTag- AttnBtn- AttnInd- PwrInd- RBE+ FLReset-
   DevCtl: Report errors: Correctable- Non-Fatal- Fatal-
Unsupported-
RlxdOrd- ExtTag- PhantFunc- AuxPwr- NoSnoop-
MaxPayload 128 bytes, MaxReadReq 512 bytes
   DevSta: CorrErr- UncorrErr- FatalErr- UnsuppReq-
AuxPwr- TransPend-
   LnkCap: Port #8, Speed 5GT/s, Width x8, ASPM L0s,
Latency L0 unlimited, L1 unlimited
ClockPM- Suprise- LLActRep- BwNot-
   LnkCtl: ASPM Disabled; RCB 64 bytes Disabled- Retrain- CommClk-
ExtSynch- ClockPM- AutWidDis- BWInt- AutBWInt-
   LnkSta: Speed 2.5GT/s, Width x8, TrErr- Train- SlotClk-
DLActive- BWMgmt- ABWMgmt-
  Capabilities: [100] #0e
  Kernel driver in use: mlx4_core
  Kernel modules: mlx4_en, mlx4_core
  00: b3 15 3c 67 06 04 10 00 b0 00 06 0c 10 00 00 00
  10: 04 00 10 fe 00 00 00 00 0c 00 00 d0 00 00 00 00
  20: 00 00 00 00 00 00 00 00 00 00 00 00 b3 15 22 00
  30: 00 00 00 00 40 00 00 00 00 00 00 00 05 01 00 00
  40: 01 48 03 00 00 00 00 00 03 9c ff 7f 11 11 00 00
  50: 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00
  60: 10 00 02 00 01 8e 00 00 00 20 00 00 82 f4 03 08
  70: 00 00 81 00 00 00 00 00 00 00 00 00 00 00 00 00
  80: 00 00 00 00 1f 00 00 00 00 00 00 00 00 00 00 00
  90: 02 00 00 00 00 00 00 00 00 00 00 00 11 60 ff 80
  a0: 00 c0 07 00 00 d0 07 00 05 00 8a 00 00 00 00 00
  b0: 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00
  c0: 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00
  d0: 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00
  e0: 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00
  f0: 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00

Sincerely

--
Hiroyuki Sato.
--
To unsubscribe from this list: send the line "unsubscribe linux-rdma" in
the body of a message to majord...@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html


Re: IPoIB to Ethernet routing performance

2010-12-07 Thread Hiroyuki Sato
Basic Question.
  MTU size matched?? on 10GbE and Infiniband

BTW, could you tell me more detail about 10GbE configuration.
what kind of teaming are you using?? (ex IEEE802.3ad)
(It is my interest.)

--
Hiroyuki Sato.


2010/12/7 sebastien dugue :
> On Tue, 7 Dec 2010 12:33:32 +0200
> Or Gerlitz  wrote:
>
>> On 12/7/2010 12:27 PM, sebastien dugue wrote:
>> >
>> >    Huh? How do you do that?
>> see Documentation/infiniband/ipoib.txt on your clone of Linus tree or if
>> you're not a developer, see
>> http://lxr.linux.no/#linux+v2.6.36/Documentation/infiniband/ipoib.txt
>> the HCA you're using does support offloads, so you're probably running
>> connected mode, use datagram
>
>  Right, I'm running in connected mode. Last time I checked, datagram mode
> was far from connected mode performance wise.
>
>  Sébastien.
> --
> To unsubscribe from this list: send the line "unsubscribe linux-rdma" in
> the body of a message to majord...@vger.kernel.org
> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>
--
To unsubscribe from this list: send the line "unsubscribe linux-rdma" in
the body of a message to majord...@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html


Re: [Q] I tried to mount nfs_server:/mnt volume, but It mounted another volume.

2010-11-21 Thread Hiroyuki Sato
Thank you useful information. I'll study about NFSv4

--
Hiroyuki Sato

2010/11/22 J. Bruce Fields :
> On Sun, Nov 21, 2010 at 03:41:18PM +0200, Boaz Harrosh wrote:
>> On 11/21/2010 02:23 PM, Hiroyuki Sato wrote:
>> > Hello lists
>> >
>> > OS: CentOS 5.5
>> > kernel: 2.6.36 rebuild myself.
>> >
>> >
>> >   I'm trying to test NFS/RDMA.
>> >   I tried to mount nfs_server:/mnt volume on on nfs_client
>> >   but It mounted nfs_server:/nfstest volume
>> >
>> >   note: /nfstest is tmpfs
>> >
>> >  this is mount output
>> >     tmpfs on /nfstest type tmpfs (rw,size=4g)
>> >     /dev/sdb1 on /mnt type ext3 (rw)
>> >
>> >   Is this bug??
>> >
>> > NFS server config
>> >
>> >   # ls -1 /mnt
>> >   This_is_mnt_volume
>> >
>> >   # ls -1 /nfstest
>> >   This_is_nfstest_volume
>> >
>> >   # cat /etc/exports
>> >   /nfstest   
>> > 192.168.100.0/255.255.255.0(fsid=0,rw,async,insecure,no_root_squash)
>> >   /mnt       
>> > 192.168.100.0/255.255.255.0(fsid=0,rw,async,insecure,no_root_squash)
>> >
>>
>> You must not have two exports with fsid=0. First one is picked.
>>
>> nfs4 will only export a single name space point, other exports are
>> subdirs of that root export. (use bind mounts to present a single
>> directory tree)
>>
>> <http://www.citi.umich.edu/projects/nfsv4/linux/using-nfsv4.html>
>
> We should update and/or replace that documentation; with new nfs-utils
> and kernel, if you omit the "fsid=0", you'll end up with an nfsv4
> namespace that's the same as v2/v3's always was.
>
> --b.
>
>>
>> NFSv4 exports on linux
>> ~~
>>
>> NFSv4 no longer has a separate "mount" protocol. Instead of exporting a 
>> number of distinct exports, an NFSv4 client sees the NFSv4 server's exports 
>> as existing inside a single filesystem, called the nfsv4 "pseudofilesystem".
>>
>> On the current linux implementation, the pseudofilesystem is a single real 
>> filesystem, identified at export with the fsid=0 option.
>>
>> In the example above, we exported only a single filesystem, which the client 
>> mounted as "/". You can provide clients with multiple filesystems to mount, 
>> producing NFSv3-like-behavior, by creative use of mount --bind. For example, 
>> you could export /usr/local/bin to clients as /bin and /usr/local/etc as 
>> /etc as follows:
>>
>> mkdir /export
>> mkdir /export/bin
>> mkdir /export/etc
>> mount --bind /usr/local/bin /export/bin
>> mount --bind /usr/local/etc /export/etc
>> exportfs -ofsid=0,insecure,no_subtree_check *:/export
>> exportfs -orw,nohide,insecure,no_subtree_check *:/export/bin
>> exportfs -orw,nohide,insecure,no_subtree_check *:/export/etc
>>
>> Note that the paths returned by the "showmount" program are meaningful only 
>> to clients using nfs versions 2 and 3; in the above example, "showmount" 
>> will list the paths /export, /export/bin/, and /export/etc, but nfsv4 
>> clients should mount yourserver:/, yourserver:/bin, or yourserver:/etc.
>>
>> http://www.citi.umich.edu/projects/nfsv4/linux/using-nfsv4.html>
>>
>> Boaz
>> >
>> >   # modprobe svcrdma
>> >
>> >   # /sbin/service nfs start
>> >
>> >   # echo rdma 20049 > /proc/fs/nfsd/portlist
>> >
>> >
>> > Client Setting
>> >
>> >   /sbin/modprobe xprtrdma
>> >   /sbin/mount.rnfs 192.168.100.231:/mnt /mnt -i -o rdma,port=20049
>> >
>> >   # ls -1 /mnt
>> >   This_is_nfstest_volume
>> >
>> > NFS Server log
>> >   sysctl -w sunrpc.nfsd_debug=1023
>> >
>> >   Nov 21 20:47:37 dell1435 mountd[3575]: authenticated mount request
>> > from 192.168.100.232:766 for /mnt (/mnt)
>> >   Nov 21 20:47:37 dell1435 mountd[3575]: /nfstest and /mnt have same
>> > filehandle for 192.168.100.0/255.255.255.0, using first
>> >   Nov 21 20:48:55 dell1435 mountd[3575]: authenticated unmount request
>> > from 192.168.100.232:912 for /mnt (/mnt)
>> >   Nov 21 20:48:55 dell1435 mountd[3575]: authenticated unmount request
>> > from 192.168.100.232:913 for /mnt (/mnt)
>> >   Nov 21 20:49:00 dell1435 mountd[3575]: authenticated unmount request
>> > from 192.168.100.232:917 for /mnt (/mnt)
>> >   Nov 21 20:49:16 dell1435 mountd[3

[Q] I tried to mount nfs_server:/mnt volume, but It mounted another volume.

2010-11-21 Thread Hiroyuki Sato
Hello lists

OS: CentOS 5.5
kernel: 2.6.36 rebuild myself.


  I'm trying to test NFS/RDMA.
  I tried to mount nfs_server:/mnt volume on on nfs_client
  but It mounted nfs_server:/nfstest volume

  note: /nfstest is tmpfs

 this is mount output
tmpfs on /nfstest type tmpfs (rw,size=4g)
/dev/sdb1 on /mnt type ext3 (rw)

  Is this bug??

NFS server config

  # ls -1 /mnt
  This_is_mnt_volume

  # ls -1 /nfstest
  This_is_nfstest_volume

  # cat /etc/exports
  /nfstest   
192.168.100.0/255.255.255.0(fsid=0,rw,async,insecure,no_root_squash)
  /mnt   
192.168.100.0/255.255.255.0(fsid=0,rw,async,insecure,no_root_squash)


  # modprobe svcrdma

  # /sbin/service nfs start

  # echo rdma 20049 > /proc/fs/nfsd/portlist


Client Setting

  /sbin/modprobe xprtrdma
  /sbin/mount.rnfs 192.168.100.231:/mnt /mnt -i -o rdma,port=20049

  # ls -1 /mnt
  This_is_nfstest_volume

NFS Server log
  sysctl -w sunrpc.nfsd_debug=1023

  Nov 21 20:47:37 dell1435 mountd[3575]: authenticated mount request
from 192.168.100.232:766 for /mnt (/mnt)
  Nov 21 20:47:37 dell1435 mountd[3575]: /nfstest and /mnt have same
filehandle for 192.168.100.0/255.255.255.0, using first
  Nov 21 20:48:55 dell1435 mountd[3575]: authenticated unmount request
from 192.168.100.232:912 for /mnt (/mnt)
  Nov 21 20:48:55 dell1435 mountd[3575]: authenticated unmount request
from 192.168.100.232:913 for /mnt (/mnt)
  Nov 21 20:49:00 dell1435 mountd[3575]: authenticated unmount request
from 192.168.100.232:917 for /mnt (/mnt)
  Nov 21 20:49:16 dell1435 mountd[3575]: authenticated mount request
from 192.168.100.232:865 for /mnt (/mnt)
  Nov 21 21:02:22 dell1435 mountd[3575]: authenticated unmount request
from 192.168.100.232:955 for /mnt (/mnt)
  Nov 21 21:02:26 dell1435 mountd[3575]: authenticated mount request
from 192.168.100.232:884 for /mnt (/mnt)
  Nov 21 21:02:26 dell1435 kernel: nfsd: exp_rootfh(/mnt
[88011f586740] 192.168.100.0/255.255.255.0:sdb1/2)


--
Hiroyuki Sato
--
To unsubscribe from this list: send the line "unsubscribe linux-rdma" in
the body of a message to majord...@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html


[Q] I would like to test NFS/RDMA with multiple ports on same fabric

2010-11-21 Thread Hiroyuki Sato
Dear members.

I'm newbie about Infiniband.

I would like to test NFS/RDMA on the following environment.
I succeed to setup NFS/RDMA between NFS Server and NFS Client1

Probelm:

  I want to add second NFS client (NFS client2)
  but I can't setup IP over Infiniband between NFS server and NFS client2
  (I don't have Infiniband switch so I want to use multiple ports on
same fabric)

  NFS Serer% ping nfs_client2
  no response

Question:

  (1) Is this environment work fine if I setup correctly??
ex) I have to change subnet address
I have to use Infiniband Switch

  (2) What is the problem??

Could you tell me how to check what is the problem.
or Please tell me good example if avairable .

Thank you for your information.

Sincrely

--
Hiroyuki Sato.



opensmd
NFS Server
192.168.100.231(port1)
192.168.200.231(port2)NFS client1(192.168.100.232)
+--+  +--+
|  | 192.168.100.0/24 |  |
|  port1port1|
|  port2--+ port2|
|  |  |   |  |
+--+|   +--+
  | 
  |
  |   NFS client2(192.168.100.233)
  |   +--+
  |   |  |
  +--port1   |
192.168.200.0/24 port2   |
  |  |
  +--+

Environment
  OS: CentOS 5.5
  Infiniband Card: MHEA28-XTC
  Kernel: 2.6.36


This is ibstat output on NFS Server
# ibstat
CA 'mthca0'
CA type: MT25208
Number of ports: 2
Firmware version: 5.3.0
Hardware version: a0
Node GUID: 0x0002c90200220eb8
System image GUID: 0x0005ad000100d050
Port 1:
State: Active
Physical state: LinkUp
Rate: 10
Base lid: 1
LMC: 0
SM lid: 1
Capability mask: 0x02590a6a
Port GUID: 0x0002c90200220eb9
Link layer: IB
Port 2:
State: Active
Physical state: LinkUp
Rate: 10
Base lid: 1
LMC: 0
SM lid: 3
Capability mask: 0x02590a68
Port GUID: 0x0002c90200220eba
Link layer: IB
--
To unsubscribe from this list: send the line "unsubscribe linux-rdma" in
the body of a message to majord...@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html