On Tue, 2006-11-28 at 10:00 -0700, Li-Ta Lo wrote: > On Mon, 2006-11-27 at 17:21 -0800, Matt Leininger wrote: > > On Mon, 2006-11-27 at 16:45 -0800, Matt Leininger wrote: > > > Has anyone testing OMPI's alltoall at > 2000 MPI tasks? I'm seeing each > > > MPI task eat up > 1GB of memory (just for OMPI - not the app). > > > > I gathered some more data using the alltoall benchmark in mpiBench. > > mpiBench is pretty smart about how large its buffers are. I set it to > > use <= 100MB. > > > > num nodes num MPI tasks system mem mpibench buffer mem > > 128 1024 1 GB 65 MB > > 160 1280 1.2 GB 82 MB > > 192 1536 1.4 GB 98 MB > > 224 1792 1.6 GB 57 MB > > 256 2048 1.6-1.8 GB < 100 MB > > > > The 256 node run was killed by the OOM for using too much memory. For > > all these tests the OMPI alltoall is using 1 GB or more of system > > memory. I know LANL is looking into optimized alltoall, but is anyone > > looking into the scalability of the memory footprint? > > > > I am the one who is looking into those collective communications. Which > mca/coll are you using for alltoall?
The ompi_info output had some mca/coll information in it. I'm not sure which mca/coll parameter you are interested in. > Does the OOM killer kick in when > calling other collective routines? I've tested Bcast, Barrier, Allreduce, Gather, Scatter, Reduce, Allgather, and Alltoall. So far only the Alltoall has this problem. - Matt > If it is a problem caused by SM > files, all collectives should be affected. > > Ollie > > > _______________________________________________ > devel mailing list > de...@open-mpi.org > http://www.open-mpi.org/mailman/listinfo.cgi/devel >