Jeff thanks for the clarification,
What I am trying to do is run N concurrent copies of a 1 to N data movement
program to affect an N to N solution. The actual mechanism I am using is to
spawn N copies of mpirun from PVM across the cluster. So each 1 to N MPI
application starts at the same
Hi Jeff
Thank you for opening a ticket and taking care of this.
Jeff Squyres wrote:
On Jul 28, 2010, at 5:07 PM, Gus Correa wrote:
Still, the alignment under Intel may or may not be right.
And this may or may not explain the errors that Hugo has got.
FYI, the ompi_info from my OpenMPI 1.3.2
Thank you, Cristobal.
That is good news.
Gus Correa
Cristobal Navarro wrote:
i have good news.
after updating to a newer kernel on ubuntu server nodes, sm is not a
problem anymore for the nehalem CPUs!!!
my older kernel, was
Linux 2.6.32-22-server #36-Ubuntu SMP Thu Jun 3 20:38:33 UTC 2010
Solved: The process to core locking was due to affinity being set at
the PSM layer. So I added -x IPATH_NO_CPUAFFINITY=1 to the mpirun
command.
Dave
On Wed, Aug 4, 2010 at 12:13 PM, Eugene Loh wrote:
>
> David Akin wrote:
>
>> All,
>> I'm trying to get the OpenMP portion
+1 on Eugene's comment that I don't fully understand what you are trying to do.
Can you send a short example code?
Some random points:
- Edgar already chimed in about how MPI-2 allows the use of intercommunicators
with bcast. Open MPI is MPI-2.1 complaint, so you can use intercommunicators
On Tue, Aug 10, 2010 at 9:49 PM, Alexandru Blidaru wrote:
> Are the Boost.MPI send and recv functions as fast as the standard ones when
> using Open-MPI?
Boost.MPI is layered on top of plain MPI; it basically provides a
mapping from complex and user-defined C++ data types to
Hi Riccardo,
Are the Boost.MPI send and recv functions as fast as the standard ones when
using Open-MPI?
Best regards,
Alexandru Blidaru
University of Waterloo - Electrical Engineering '15
University email: asbli...@uwaterloo.ca
Twitter handle: @G_raph
Blog: http://alexblidaru.wordpress.com/
Hi,
sorry, i just forgot to add the values of the function parameters:
(gdb) print reg->cbdata
$1 = (void *) 0x0
(gdb) print openib_btl->super
$2 = {btl_component = 0x2b341edd7380, btl_eager_limit = 12288,
btl_rndv_eager_limit = 12288, btl_max_send_size = 65536,
btl_rdma_pipeline_send_length =
Hi,
Here is the output of a core file generated during a segmentation fault
observed during a collective call (using openib):
#0 0x in ?? ()
(gdb) where
#0 0x in ?? ()
#1 0x2aedbc4e05f4 in btl_openib_handle_incoming (openib_btl=0x1902f9b0,
ep=0x1908a1c0,
Hi Mike,
The HCA card is a Mellanox Technologies MT25418 (ConnectX IB DDR, PCIe 2.0
2.5GT/s, rev a0).
I cannot post code/instructions how to reproduce these errors as they randomly
appeared during some tests I've performed to locate the origin of a
segmentation fault during an MPI collective
Hey Eloi,
What HCA card do you have ? Can you post code/instructions howto reproduce
it?
10x
Mike
On Mon, Aug 9, 2010 at 5:22 PM, Eloi Gaudry wrote:
> Hi,
>
> Could someone have a look on these two different error messages ? I'd like
> to know the reason(s) why they were displayed
Hi,
I just stumped into the following behavior of Open MPI 1.4.2. Used jobscript:
***
#!/bin/sh
export PATH=~/local/openmpi-1.4.2/bin:$PATH
cat $PE_HOSTFILE
mpiexec ./dummy.sh
***
with dummy.sh:
***
#!/bin/sh
env | grep TMPDIR
sleep 30
***
===
Situation 1: getting 4 slots in total from 2
12 matches
Mail list logo