Setting these environment variables did indeed change the way mpirun maps
things, and I didn't have to specify a hostfile.  However, setting these
for my MPI_Comm_spawn code still resulted in the same segmentation fault.

Evan

On Tue, Feb 3, 2015 at 10:09 AM, Ralph Castain <r...@open-mpi.org> wrote:

> If you add the following to your environment, you should run on multiple
> nodes:
>
> OMPI_MCA_rmaps_base_mapping_policy=node
> OMPI_MCA_orte_default_hostfile=<your hostfile>
>
> The first tells OMPI to map-by node. The second passes in your default
> hostfile so you don't need to specify it as an Info key.
>
> HTH
> Ralph
>
>
> On Tue, Feb 3, 2015 at 9:23 AM, Evan Samanas <evan.sama...@gmail.com>
> wrote:
>
>> Hi Ralph,
>>
>> Good to know you've reproduced it.  I was experiencing this using both
>> the hostfile and host key.  A simple comm_spawn was working for me as well,
>> but it was only launching locally, and I'm pretty sure each node only has 4
>> slots given past behavior (the mpirun -np 8 example I gave in my first
>> email launches on both hosts).  Is there a way to specify the hosts I want
>> to launch on without the hostfile or host key so I can test remote launch?
>>
>> And to the "hostname" response...no wonder it was hanging!  I just
>> constructed that as a basic example.  In my real use I'm launching
>> something that calls MPI_Init.
>>
>> Evan
>>
>>
>>
>> _______________________________________________
>> users mailing list
>> us...@open-mpi.org
>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
>> Link to this post:
>> http://www.open-mpi.org/community/lists/users/2015/02/26271.php
>>
>
>
> _______________________________________________
> users mailing list
> us...@open-mpi.org
> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
> Link to this post:
> http://www.open-mpi.org/community/lists/users/2015/02/26272.php
>

Reply via email to