Yeah, you're using too much memory for the shared memory system. Run with -mca 
btl ^sm on your cmd line - it'll run slower, but you probably don't have a 
choice.


On Oct 30, 2012, at 11:38 AM, Sandra Guija <sgu...@hotmail.com> wrote:

> yes I think is related with my program too, when I run 1000x1000 matrix 
> multiplication, the program works.
> when I run the 10,000 matrix only on one machine  I got this:
> mca_common_sm_mmap_init: mmap failed with errno=12
> mca_mpool_sm_init: unable to shared memory mapping ( 
> /tmp/openmpi-sessions-mpiu@tango_0/default-universe-1529/1/shared_mem_pool 
> .tango)
> mca_common_sm_mmap_init: 
> /tmp/openmpi-sessions-mpiu@tango_0/default-universe-1529/1/shared_mem_pool 
> .tango failed with errno=2
> mca_mpool_sm_init: unable to shared memory mapping ( 
> /tmp/openmpi-sessions-mpiu@tango_0/default-universe-1529/1/shared_mem_pool 
> .tango)
> PML add procs failed
> -->Returned "0ut of resource" (-2) instead of " Success" (0)
> 
> this is the result when I run free -m
>                   total   used   free   shared  buffers   cached
> Mem:          2026    54    1972      0         6           25
> -/+ buffer cache:    22      511      
>  Swap:         511     0       511
> 
> Sandra Guija
> 
> From: r...@open-mpi.org
> Date: Tue, 30 Oct 2012 10:33:02 -0700
> To: de...@open-mpi.org
> Subject: Re: [OMPI devel] process kill signal 59
> 
> Ummm...not sure what I can say about that with so little info. It looks like 
> your process died for some reason that has nothing to do with us - a bug in 
> your "magic10000" program?
> 
> 
> On Oct 30, 2012, at 10:24 AM, Sandra Guija <sgu...@hotmail.com> wrote:
> 
> Hello, 
> I am running a 10,000x10,000 matrix multiplication in 4 processors/1 core and 
> I get the following error:
> mpirun -np 4 --hostfile nodes --bynode magic10000
> 
> mpirun noticed that job rank1 with PID 635 on node slave1 exited on signal 
> 509(Real-time signal 25).
> 2 additional process aborted (not shown)
> 1 process killed (possibly by open MPI)
> 
> node file contains:
> master
> slave1
> slave2
> slave3
> 
> 
> _______________________________________________
> devel mailing list
> de...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/devel
> 
> 
> _______________________________________________ devel mailing list 
> de...@open-mpi.org http://www.open-mpi.org/mailman/listinfo.cgi/devel
> _______________________________________________
> devel mailing list
> de...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/devel

Reply via email to