Hi,
I am getting hang ups in mpi job randomly.
..
...
IT:20760 CF: 0.7743 Time: 1540.0 MaxMin:20.69/5 :20.12/12
IT:20770 CF: 0.7734 Time: 1560.2 MaxMin:20.50/1 :19.31/5
--
mpirun noticed that
Hi,
I'm evaluating Open MPI 1.4.2 on one of our BladeCenters and I'm
getting via InfiniBand about 1550 MB/s and via shared memory about
1770 for the PingPong benchmark in Intel's MPI benchmark. (That
benchmark is just an example, I'm seeing similar numbers for my own
codes.)
Each node has two AMD
Maybe I should add that for tests I ran the benchmarks with two MPI
processes: for InfiniBand one process per node and for shared memory
both processes were located on one node.
--
==
Andreas Schäfer
HPC and Grid Computing
Chair of Computer
While your method starts mpirun itself nohup, the mpi processes themselves are
not launched that way and therefore run in the foreground. This message
indicates that at least one of those mpi processes received a hangup signal and
aborted. Even though mpirun won't get the signal itself, it does
Did you remember to set --bind-to-core or --bind-to-socket on the cmd line?
Otherwise, the processes are running unbound, which makes a significant
difference to performance.
On Jul 9, 2010, at 3:15 AM, Andreas Schäfer wrote:
> Maybe I should add that for tests I ran the benchmarks with two MP
Just tried these options and was able to increased shared memory
throughput to approx. 1950 MB/s. Still I wonder what the bottleneck
is. After all, the memory bandwidth of the systems is much higher.
Thanks
-Andreas
On 03:22 Fri 09 Jul , Ralph Castain wrote:
> Did you remember to set --bind-
On Friday 09 July 2010, Andreas Schäfer wrote:
> Hi,
>
> I'm evaluating Open MPI 1.4.2 on one of our BladeCenters and I'm
> getting via InfiniBand about 1550 MB/s and via shared memory about
> 1770 for the PingPong benchmark in Intel's MPI benchmark. (That
> benchmark is just an example, I'm seeing
On Thu, Jul 08, 2010 at 11:04:09AM -0700, Avneesh Pant wrote:
> Anton,
> On the node that you saw the failure (u02n065)
> can you verify what the max locked memory limit
> is set to? In a bash shell you can do this with
> ulimit -l. It should be set to at least 128K.
> Also please verify that th
Thanks, those were good suggestions.
On 11:53 Fri 09 Jul , Peter Kjellstrom wrote:
> On an E5520 (nehalem) node I get ~5 GB/s ping-pong for >64K sizes.
I just tried a Core i7 system which maxes at 6550 MB/s for the
ping-pong test.
> On QDR IB on similar nodes I get ~3 GB/s ping-pong for >
On Friday 09 July 2010, Andreas Schäfer wrote:
> Thanks, those were good suggestions.
>
> On 11:53 Fri 09 Jul , Peter Kjellstrom wrote:
> > On an E5520 (nehalem) node I get ~5 GB/s ping-pong for >64K sizes.
>
> I just tried a Core i7 system which maxes at 6550 MB/s for the
> ping-pong test.
It
On 14:39 Fri 09 Jul , Peter Kjellstrom wrote:
> 8x pci-express gen2 5GT/s should show figures like mine. If it's pci-express
> gen1 or gen2 2.5GT/s or 4x or if the IB only came up with two lanes then 1500
> is expected.
lspci and ibv_devinfo tell me it's PCIe 2.0 x8 and InfiniBand 4x QDR
(ac
Hi all,
I'm trying to compile openMPI 1.4.2 in Cygwin 1.7.5-1.
After ./configure I do make and after some time I always get this
error. I've tried "make clean" and "make" again, but that doesn't
help. It looks to me like I have all the requirements of the
README.Windows file (Cygwin and libtool
Hi all -
I've been looking at the dynamic process features of mpi-2. I have managed
to actually launch processes using spawn, but haven't seen examples for
actually communicating once these processes are launched. I am additionally
interested in how processes created through multiple spawn calls
On Thu, Jul 08, 2010 at 09:43:48AM -0400, Gus Correa wrote:
> Douglas Guptill wrote:
>> On Wed, Jul 07, 2010 at 12:37:54PM -0600, Ralph Castain wrote:
>>
>>> Noafraid not. Things work pretty well, but there are places
>>> where things just don't mesh. Sub-node allocation in particular is
>>> an
Hey guys,
This is my first time posting on an email list, so if I don't follow certain
protocols just tell me. Anyway, I am working at a research lab as a High
School Summer Intern, and part of my project involves installing OpenMPI on
a Windows machine to run some applications in virtual clusters
15 matches
Mail list logo