[OMPI devel] mpirun -host does not work from r27879 and forward on trunk

2013-01-31 Thread Rolf vandeVaart
I have stumbled into a problem with the -host argument. This problem appears to be introduced with changeset r27879 on 1/19/2013 by rhc. With r27877, things work: [rolf@node]$ which mpirun /home/rolf/ompi-trunk-r27877/64/bin/mpirun [rolf@node]$ mpirun -np 2 -host c0-0,c0-3 hostname c0-3 c0-0

Re: [OMPI devel] mpirun -host does not work from r27879 and forward on trunk

2013-01-31 Thread Ralph Castain
Ummm...that was fixed a long time ago. You might try a later version. Or are you saying the head of the trunk doesn't work too? On Jan 31, 2013, at 7:31 AM, Rolf vandeVaart wrote: > I have stumbled into a problem with the -host argument. This problem appears > to be introduced with changeset

Re: [OMPI devel] mpirun -host does not work from r27879 and forward on trunk

2013-01-31 Thread Ralph Castain
FWIW: I just tried it on the trunk head and it worked fine On Jan 31, 2013, at 8:20 AM, Ralph Castain wrote: > Ummm...that was fixed a long time ago. You might try a later version. > > Or are you saying the head of the trunk doesn't work too? > > On Jan 31, 2013, at 7:31 AM, Rolf vandeVaart w

Re: [OMPI devel] mpirun -host does not work from r27879 and forward on trunk

2013-01-31 Thread Rolf vandeVaart
Interesting. Yes, I was saying that the latest trunk does not work for me. I just retested the trunk also, and no luck. Are you launching the MPI processes on remote nodes from the HNP? >-Original Message- >From: devel-boun...@open-mpi.org [mailto:devel-boun...@open-mpi.org] >On Behalf

Re: [OMPI devel] mpirun -host does not work from r27879 and forward on trunk

2013-01-31 Thread Ralph Castain
Yes - no hostfile and no RM allocation, just -host. What is your setup? On Jan 31, 2013, at 8:44 AM, Rolf vandeVaart wrote: > Interesting. Yes, I was saying that the latest trunk does not work for me. > I just retested the trunk also, and no luck. > Are you launching the MPI processes on re

Re: [OMPI devel] mpirun -host does not work from r27879 and forward on trunk

2013-01-31 Thread Rolf vandeVaart
Ralph and I talked off-list about the issue. He figured it and fixed it with changelist 27955. See that changelist for the details. >-Original Message- >From: devel-boun...@open-mpi.org [mailto:devel-boun...@open-mpi.org] >On Behalf Of Ralph Castain >Sent: Thursday, January 31, 2013 11:5

[OMPI devel] "pml_ob1_sendreq.c:188 FATAL" errors

2013-01-31 Thread Jeff Squyres (jsquyres)
I'm seeing a LOT of these on errors on the trunk: pml_ob1_sendreq.c:188 FATAL The job then hangs. I see this starting at np=6 across 2 nodes, using only the TCP and SM BTLs. This is not happening on v1.6 or v1.7. Line 188 in pml_ob1_sendreq.c is when someone calls mca_pml_ob1_match_compl

Re: [OMPI devel] "pml_ob1_sendreq.c:188 FATAL" errors

2013-01-31 Thread Ralph Castain
I fixed it so that "abort" really aborts the job - see r28004 On Jan 31, 2013, at 2:02 PM, Jeff Squyres (jsquyres) wrote: > I'm seeing a LOT of these on errors on the trunk: > >pml_ob1_sendreq.c:188 FATAL > > The job then hangs. I see this starting at np=6 across 2 nodes, using only > th

Re: [OMPI devel] "pml_ob1_sendreq.c:188 FATAL" errors

2013-01-31 Thread Jeff Squyres (jsquyres)
The show help bit doesn't look right -- opal_output on stream 0 will put the hostname and PID as the prefix. On Jan 31, 2013, at 6:13 PM, Ralph Castain wrote: > I fixed it so that "abort" really aborts the job - see r28004 > > On Jan 31, 2013, at 2:02 PM, Jeff Squyres (jsquyres) > wrote: >

[OMPI devel] RFC: shiny new variable subsystem

2013-01-31 Thread Nathan Hjelm
What: Introduce the MCA variable system. This system is meant as a replacement for the MCA parameter system. It offers a number of improvements over the old system including: - Cleaner, expandable API. Instead of providing multiple variable registration functions (reg_int, reg_int_name, reg_s