Try using slots in hostfile ? -- Sent from my iPhone
On Apr 24, 2012, at 2:52 PM, Kyle Boe <boex0...@umn.edu> wrote: > I'm having a problem trying to use OpenMPI on some multicore machines I have. > The code I am running was giving me errors which suggested that MPI was > assigning multiple processes to the same core (which I do not want). So, I > tried launching my job using the -nooversubscribe option, and I get this > error: > > bash-3.2$ mpirun -np 2 -nooversubscribe <executable and options here> > -------------------------------------------------------------------------- > There are not enough slots available in the system to satisfy the 2 slots > that were requested by the application: > <executable name> > > Either request fewer slots for your application, or make more slots available > for use. > -------------------------------------------------------------------------- > -------------------------------------------------------------------------- > A daemon (pid unknown) died unexpectedly on signal 1 while attempting to > launch so we are aborting. > > There may be more information reported by the environment (see above). > > This may be because the daemon was unable to find all the needed shared > libraries on the remote node. You may set your LD_LIBRARY_PATH to have the > location of the shared libraries on the remote nodes and this will > automatically be forwarded to the remote nodes. > -------------------------------------------------------------------------- > -------------------------------------------------------------------------- > mpirun noticed that the job aborted, but has no info as to the process > that caused that situation. > -------------------------------------------------------------------------- > mpirun: clean termination accomplished > > I am just trying to run on the localhost, not on any remote machines. This > happens on both my 8 (2*4) core and 24 (4*6) core machines. Relevant info: I > am not using any type of scheduler here, although from the searching I've > done that doesn't seem like a requirement. The only thing I can think is > there must be some type of configuration or option I'm not setting for using > on shared memory machines (either at compile or run time), but I can't find > anyone else who has come across this error. Any thoughts? > > Thanks, > > Kyle > _______________________________________________ > users mailing list > us...@open-mpi.org > http://www.open-mpi.org/mailman/listinfo.cgi/users