[OMPI users] bug of openmpi-1.3b3r20000

2008-12-12 Thread Catrina Nguyen
Hi all, I use openmpi-1.3b3r2 and blcr-0.7.3 to run my application on 2 nodes. I configure openmpi to run on 2 nodes for default.  I want to use checkpoint/restart functionalities, so I use this command to configure openmpi: # .configure --with-devel-headers --with-ft=cr --with-blcr= First:

Re: [OMPI users] Onesided + derived datatypes

2008-12-12 Thread Dorian Krause
Thanks George (and Brian :)). The MPI_Put error is gone. Did you take a look at the problem that with the block_indexed type the PUT doesn't work? I'm still getting the following output (V1 corresponds to the datatype created with MPI_Type_create_indexed_block while the V2 type is created with MPI

Re: [OMPI users] Axon BTL for OpenMPI?

2008-12-12 Thread Jeff Squyres
On Dec 11, 2008, at 7:01 PM, Marcus G. Daniels wrote: I've heard some experimental work has been done to run OpenMPI over the Axon driver as found in IBM triblades. Seems like that should work fine, as it's just another RDMA interface, no? I personally don't know anything about Axon; perh

Re: [OMPI users] Onesided + derived datatypes

2008-12-12 Thread Dorian Krause
Hi again. I adapted my testing program by overwriting the window buffer complete with 1. This allows me to see at which places OpenMPI writes. The result is: *** -DO1=1 -DV1=1 *** (displ 3,2,1,0 , MPI_Type_create_indexed_block) mem[0] = { 0., 0., 0.} mem[1] = { 0., 0., 0

Re: [OMPI users] trouble using --mca mpi_yield_when_idle 1

2008-12-12 Thread Jeff Squyres
FWIW, Open MPI does have on its long-term roadmap to have "blocking" progress -- meaning that it'll (probably) spin aggressively for a while and if nothing "interesting" is happening, it'll go into a blocking mode and let the process block in some kind of OS call. Although we have some inte

Re: [OMPI users] Axon BTL for OpenMPI?

2008-12-12 Thread Ralph Castain
Just to clarify: are you talking about using this on RoadRunner to move data from the Opteron to the Cell blades? Or are you talking about using OpenMPI on some other machine? If you are interested in RR, we should probably continue the discussion off-list. The short answer is that we have

Re: [OMPI users] Axon BTL for OpenMPI?

2008-12-12 Thread Marcus G. Daniels
Hi Jeff, I'd like to use MPI for features like derived types, and moving around non-flat data. You should be able to do that today. The Axon PCIe-to-PCIe RDMA interface has a raw device and an ethernet device in the Linux kernel, and I can indeed using the latter as a workaround. It's mor

[OMPI users] Interactive OMPI job, stdin problem

2008-12-12 Thread Harald Anlauf
Hello, I am having problems with OMPI-1.3beta with an interactive job where rank 0 reads stdin from a terminal. The problem does not show up when stdin is redirected from a file. The problem also does not exist with OMPI 1.2.[5-9]. Has there been any change in OMPI between 1.2 and 1.3 that I sho

Re: [OMPI users] trouble using --mca mpi_yield_when_idle 1

2008-12-12 Thread Eugene Loh
Jeff Squyres wrote: FWIW, Open MPI does have on its long-term roadmap to have "blocking" progress -- meaning that it'll (probably) spin aggressively for a while and if nothing "interesting" is happening, it'll go into a blocking mode and let the process block in some kind of OS call. Alth

Re: [OMPI users] Interactive OMPI job, stdin problem

2008-12-12 Thread Ralph Castain
Hi Harald There is a patch for the IOF in 1.3 "soaking" in the trunk right now. I'll check to ensure it fixes this issue too. Hopefully, it will come over to the 1.3 branch early next week. Ralph On Dec 12, 2008, at 8:21 AM, Harald Anlauf wrote: Hello, I am having problems with OMPI-1.

Re: [OMPI users] trouble using --mca mpi_yield_when_idle 1

2008-12-12 Thread douglas . guptill
Hello Eugene: On Mon, Dec 08, 2008 at 11:14:10AM -0800, Eugene Loh wrote: .. >> Proceeding from that, it seems that "mpi_recv" is implemented as >> "poll forever until the message comes" >> and NOT as >>"sleep until the message comes" >> >> I had assumed, until now, that mpi_recv w

Re: [OMPI users] trouble using --mca mpi_yield_when_idle 1

2008-12-12 Thread douglas . guptill
Hello Jeff: On Fri, Dec 12, 2008 at 08:37:14AM -0500, Jeff Squyres wrote: > FWIW, Open MPI does have on its long-term roadmap to have "blocking" > progress -- meaning that it'll (probably) spin aggressively for a > while and if nothing "interesting" is happening, it'll go into a > blocking

Re: [OMPI users] trouble using --mca mpi_yield_when_idle 1

2008-12-12 Thread Jeff Squyres
On Dec 12, 2008, at 11:46 AM, Eugene Loh wrote: FWIW, I've run into the need for this a few times due to HPCC tests on large (>100 MPI procs) nodes or multicore systems. HPCC (among other things) looks at the performance of a single process while all other np-1 processes spinwait -- or of

Re: [OMPI users] trouble using --mca mpi_yield_when_idle 1

2008-12-12 Thread Jeff Squyres
On Dec 12, 2008, at 3:22 PM, douglas.gupt...@dal.ca wrote: I could imagine another alternative.  Construct a wrapper function that intercepts MPI_Recv and turn it into something like PMPI_Irecv while ( ! done ) {     nanosleep(short);     PMPI_Test(&done); } I don't know how useful this woul

Re: [OMPI users] trouble using --mca mpi_yield_when_idle 1

2008-12-12 Thread Jeff Squyres
On Dec 12, 2008, at 3:46 PM, douglas.gupt...@dal.ca wrote: This is probably because most users running MPI jobs tend to devote the entire core/CPU/server to the MPI job and don't try to run other jobs concurrently on the same resources. Our situation is different. While our number-cruncher ap

Re: [OMPI users] trouble using --mca mpi_yield_when_idle 1

2008-12-12 Thread Eugene Loh
Jeff Squyres wrote: On Dec 12, 2008, at 11:46 AM, Eugene Loh wrote: FWIW, I've run into the need for this a few times due to HPCC tests on large (>100 MPI procs) nodes or multicore systems. HPCC (among other things) looks at the performance of a single process while all other np-1 proces

Re: [OMPI users] Onesided + derived datatypes

2008-12-12 Thread George Bosilca
Dorian, I looked into this again. So far I can confirm that the datatype is correctly created, and always contain the correct values (internally). If instead of one sided you use send/recv then the output is exactly what you expect. With the one sided there are several strange things. Wha