Hi,
I have stumbled upon a similar issue, so I wonder those might be
related. On one of our systems I get the following error message, both
when using openmpi 1.8.8 and 1.10.4
$ mpirun -debug-daemons --mca btl tcp,self --mca mca_base_verbose 100
--mca btl_base_verbose 100 ls
[...]
[compute
something else did as
I seem to recall we handled this okay before (but I could be wrong).
Fixing that will take some time that I honestly won’t have for awhile.
On Oct 9, 2015, at 6:14 AM, Marcin Krotkiewski
wrote:
Ralph,
Here is the result running
mpirun --map-by slot:pe=4 -display
Ralph,
Here is the result running
mpirun --map-by slot:pe=4 -display-allocation ./affinity
== ALLOCATED NODES ==
c12-29: slots=4 max_slots=0 slots_inuse=0 state=UP
=
rank 0 @ compute-
Hi, Gilles,
I have briefly tested your patch with master. So far everything works. I
must say what I really like about this version is that it with
--report-bindings it actually shows how the heterogeneous architectures
looks like, i.e., varying number of cores/sockets per compute node. This
Hi,
I fail to make OpenMPI bind to cores correctly when running from within
SLURM-allocated CPU resources spread over a range of compute nodes in an
otherwise homogeneous cluster. I have found this thread
http://www.open-mpi.org/community/lists/users/2014/06/24682.php
and did try to use what
Hi,
I am trying to compile the 2.x branch with libfabric support, but get
this error during configure:
configure:100708: checking rdma/fi_ext_usnic.h presence
configure:100708: gcc -E
-I/cluster/software/VERSIONS/openmpi.gnu.2.x/include
-I/usit/abel/u1/marcink/software/ompi-release-2.x/opal/
Thanks, Dave.
I have verified the memory locality and IB card locality, all's fine.
Quite accidentally I have found that there is a huge penalty if I mmap
the shm with PROT_READ only. Using PROT_READ | PROT_WRITE yields good
results, although I must look at this further. I'll report when I am
But where would I put it? If I put it in the while(1), then
MPI_Comm_Accept cannot be called for the second time. If I put it
outside of the loop it will never be called.
On 09/16/2015 04:18 PM, Jalel Chergui wrote:
Can you check with an MPI_Finalize in the receiver ?
Jalel
Le 16/09/2015 16:
Brilliant! Thank you, Rolf. This works: all ranks have reported using
the expected port number, and performance is twice of what I was
observing before :)
I can certainly live with this workaround, but I will be happy to do
some debugging to find the problem. If you tell me what is needed /