I tried this on a humble PC and it works there.
I see in the --mca mpi_show_mca_params 1 print out that there is a
[bb17:06646] paffinity=
entry, so I expect that sets the value entry back to 0?

I'll get to the SLES10 cluster when I can (other people doing
benchmarks) and see what I can. I see there's no stdbool.h there,
so maybe this is an artifact of defining the bool type on an
operton. I'll get back to you when I can.

The test of boundness was a perl program invoked via system() in a
C MPI program. The /proc/<pid>/stat result shows the CPU you are
bound to (3rd number from the end) and a taskset call gets back the
mask to show if you are bound or not.

Regards,
Mostyn


On Mon, 23 Apr 2007, Jeff Squyres wrote:

On Apr 22, 2007, at 8:46 PM, Mostyn Lewis wrote:

More information.

--mca mpi_paffinity_alone 0

Can you describe how you're verifying that the process is actually
bound?

May I ask I question about the code, which I scanned.

I see in ompi/runtime/ompi_mpi_params.c you do a:

     mca_base_param_reg_int_name("mpi", "paffinity_alone",
                                 "If nonzero, assume that this job
is the only (set of) process(es) r
unning on each node and bind processes to processors, starting with
processor ID 0",
                                 false, false,
                                 (int) ompi_mpi_paffinity_alone,
&value);
     ompi_mpi_paffinity_alone = OPAL_INT_TO_BOOL(value);

You're right that we're not checking the return value here, which is
bad.  However, this is a pretty stable portion of the code -- I would
be surprised if it is failing.

Have you stepped through this with a debugger to verify if it is
actually failing or not?

--
Jeff Squyres
Cisco Systems

_______________________________________________
users mailing list
us...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/users

Reply via email to