Setting these environment variables did indeed change the way mpirun maps things, and I didn't have to specify a hostfile. However, setting these for my MPI_Comm_spawn code still resulted in the same segmentation fault.
Evan On Tue, Feb 3, 2015 at 10:09 AM, Ralph Castain <r...@open-mpi.org> wrote: > If you add the following to your environment, you should run on multiple > nodes: > > OMPI_MCA_rmaps_base_mapping_policy=node > OMPI_MCA_orte_default_hostfile=<your hostfile> > > The first tells OMPI to map-by node. The second passes in your default > hostfile so you don't need to specify it as an Info key. > > HTH > Ralph > > > On Tue, Feb 3, 2015 at 9:23 AM, Evan Samanas <evan.sama...@gmail.com> > wrote: > >> Hi Ralph, >> >> Good to know you've reproduced it. I was experiencing this using both >> the hostfile and host key. A simple comm_spawn was working for me as well, >> but it was only launching locally, and I'm pretty sure each node only has 4 >> slots given past behavior (the mpirun -np 8 example I gave in my first >> email launches on both hosts). Is there a way to specify the hosts I want >> to launch on without the hostfile or host key so I can test remote launch? >> >> And to the "hostname" response...no wonder it was hanging! I just >> constructed that as a basic example. In my real use I'm launching >> something that calls MPI_Init. >> >> Evan >> >> >> >> _______________________________________________ >> users mailing list >> us...@open-mpi.org >> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users >> Link to this post: >> http://www.open-mpi.org/community/lists/users/2015/02/26271.php >> > > > _______________________________________________ > users mailing list > us...@open-mpi.org > Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users > Link to this post: > http://www.open-mpi.org/community/lists/users/2015/02/26272.php >