[OMPI users] mpirun hang up randomly

2010-07-09 Thread Harichand M V
Hi, I am getting hang ups in mpi job randomly. .. ... IT:20760 CF: 0.7743 Time: 1540.0 MaxMin:20.69/5 :20.12/12 IT:20770 CF: 0.7734 Time: 1560.2 MaxMin:20.50/1 :19.31/5 -- mpirun noticed that

[OMPI users] Low Open MPI performance on InfiniBand and shared memory?

2010-07-09 Thread Andreas Schäfer
Hi, I'm evaluating Open MPI 1.4.2 on one of our BladeCenters and I'm getting via InfiniBand about 1550 MB/s and via shared memory about 1770 for the PingPong benchmark in Intel's MPI benchmark. (That benchmark is just an example, I'm seeing similar numbers for my own codes.) Each node has two AMD

Re: [OMPI users] Low Open MPI performance on InfiniBand and shared memory?

2010-07-09 Thread Andreas Schäfer
Maybe I should add that for tests I ran the benchmarks with two MPI processes: for InfiniBand one process per node and for shared memory both processes were located on one node. -- == Andreas Schäfer HPC and Grid Computing Chair of Computer

Re: [OMPI users] mpirun hang up randomly

2010-07-09 Thread Ralph Castain
While your method starts mpirun itself nohup, the mpi processes themselves are not launched that way and therefore run in the foreground. This message indicates that at least one of those mpi processes received a hangup signal and aborted. Even though mpirun won't get the signal itself, it does

Re: [OMPI users] Low Open MPI performance on InfiniBand and shared memory?

2010-07-09 Thread Ralph Castain
Did you remember to set --bind-to-core or --bind-to-socket on the cmd line? Otherwise, the processes are running unbound, which makes a significant difference to performance. On Jul 9, 2010, at 3:15 AM, Andreas Schäfer wrote: > Maybe I should add that for tests I ran the benchmarks with two MP

Re: [OMPI users] Low Open MPI performance on InfiniBand and shared memory?

2010-07-09 Thread Andreas Schäfer
Just tried these options and was able to increased shared memory throughput to approx. 1950 MB/s. Still I wonder what the bottleneck is. After all, the memory bandwidth of the systems is much higher. Thanks -Andreas On 03:22 Fri 09 Jul , Ralph Castain wrote: > Did you remember to set --bind-

Re: [OMPI users] Low Open MPI performance on InfiniBand and shared memory?

2010-07-09 Thread Peter Kjellstrom
On Friday 09 July 2010, Andreas Schäfer wrote: > Hi, > > I'm evaluating Open MPI 1.4.2 on one of our BladeCenters and I'm > getting via InfiniBand about 1550 MB/s and via shared memory about > 1770 for the PingPong benchmark in Intel's MPI benchmark. (That > benchmark is just an example, I'm seeing

Re: [OMPI users] ipath_userinit: userinit command failed: Cannot allocate memory

2010-07-09 Thread Anton Shterenlikht
On Thu, Jul 08, 2010 at 11:04:09AM -0700, Avneesh Pant wrote: > Anton, > On the node that you saw the failure (u02n065) > can you verify what the max locked memory limit > is set to? In a bash shell you can do this with > ulimit -l. It should be set to at least 128K. > Also please verify that th

Re: [OMPI users] Low Open MPI performance on InfiniBand and shared memory?

2010-07-09 Thread Andreas Schäfer
Thanks, those were good suggestions. On 11:53 Fri 09 Jul , Peter Kjellstrom wrote: > On an E5520 (nehalem) node I get ~5 GB/s ping-pong for >64K sizes. I just tried a Core i7 system which maxes at 6550 MB/s for the ping-pong test. > On QDR IB on similar nodes I get ~3 GB/s ping-pong for >

Re: [OMPI users] Low Open MPI performance on InfiniBand and shared memory?

2010-07-09 Thread Peter Kjellstrom
On Friday 09 July 2010, Andreas Schäfer wrote: > Thanks, those were good suggestions. > > On 11:53 Fri 09 Jul , Peter Kjellstrom wrote: > > On an E5520 (nehalem) node I get ~5 GB/s ping-pong for >64K sizes. > > I just tried a Core i7 system which maxes at 6550 MB/s for the > ping-pong test. It

Re: [OMPI users] Low Open MPI performance on InfiniBand and shared memory?

2010-07-09 Thread Andreas Schäfer
On 14:39 Fri 09 Jul , Peter Kjellstrom wrote: > 8x pci-express gen2 5GT/s should show figures like mine. If it's pci-express > gen1 or gen2 2.5GT/s or 4x or if the IB only came up with two lanes then 1500 > is expected. lspci and ibv_devinfo tell me it's PCIe 2.0 x8 and InfiniBand 4x QDR (ac

[OMPI users] Error while compiling openMPI 1.4.2 in Cygwin 1.7.5-1. Library missing?

2010-07-09 Thread Miguel Rubio-Roy
Hi all,   I'm trying to compile openMPI 1.4.2 in Cygwin 1.7.5-1. After ./configure I do make and after some time I always get this error. I've tried "make clean" and "make" again, but that doesn't help. It looks to me like I have all the requirements of the README.Windows file (Cygwin and libtool

[OMPI users] Dynamic process tutorials?

2010-07-09 Thread Brian Budge
Hi all - I've been looking at the dynamic process features of mpi-2. I have managed to actually launch processes using spawn, but haven't seen examples for actually communicating once these processes are launched. I am additionally interested in how processes created through multiple spawn calls

[OMPI users] first cluster [was trouble using openmpi under slurm]

2010-07-09 Thread Douglas Guptill
On Thu, Jul 08, 2010 at 09:43:48AM -0400, Gus Correa wrote: > Douglas Guptill wrote: >> On Wed, Jul 07, 2010 at 12:37:54PM -0600, Ralph Castain wrote: >> >>> Noafraid not. Things work pretty well, but there are places >>> where things just don't mesh. Sub-node allocation in particular is >>> an

[OMPI users] Install OpenMPI on Win 7 machine

2010-07-09 Thread Alexandru Blidaru
Hey guys, This is my first time posting on an email list, so if I don't follow certain protocols just tell me. Anyway, I am working at a research lab as a High School Summer Intern, and part of my project involves installing OpenMPI on a Windows machine to run some applications in virtual clusters