Sorry to continue off-topic but going to System V shm would be for me like going back in the past.

System V shared memory used to be the main way to do shared memory on MPICH and from my (little) experience, this was truly painful : - Cleanup issues : does shmctl(IPC_RMID) solve _all_ cases ? (even kill -9 ?) - Naming issues : shm segments identified as 32 bits key potentially causing conflicts between applications or layers of the same application on one node - Space issues : the total shm size on a system is bound to /proc/sys/kernel/shmmax, needing admin configuration and causing conflicts between MPI applications running on the same node

Mmap'ed files can have a comprehensive name like <component : MPI>-<Layer : Opal>-<Jobib>-<Rank>, preventing naming issues. If we are on linux, they can be allocated in /dev/shm to prevent filesystem trafic, and space is not limited.

Sylvain

On Mon, 30 Mar 2009, Tim Mattox wrote:

I've been lurking on this conversation, and I am again left with the impression
that the underlying shared memory configuration based on sharing a file
is flawed.  Why not use a System V shared memory segment without a
backing file as I described in ticket #1320?

On Mon, Mar 30, 2009 at 1:34 PM, George Bosilca <bosi...@eecs.utk.edu> wrote:
Then it looks like the safest solution is the use either ftruncate or the
lseek method and then touch the first byte of all memory pages.
Unfortunately, I see two problems with this. First, there is a clear
performance hit on the startup time. And second, we will have to find a
pretty smart way to do this or we will completely break the memory affinity
stuff.

 george.

On Mar 30, 2009, at 13:24 , Iain Bason wrote:


On Mar 30, 2009, at 12:05 PM, Jeff Squyres wrote:

But don't we need the whole area to be zero filled?

It will be zero-filled on demand using the lseek/touch method.  However,
the OS may not reserve space for the skipped pages or disk blocks.  Thus one
could still get out of memory or file system full errors at arbitrary
points.  Presumably one could also get segfaults from an mmap'ed segment
whose pages couldn't be allocated when the demand came.

Iain

_______________________________________________
devel mailing list
de...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/devel

_______________________________________________
devel mailing list
de...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/devel




--
Tim Mattox, Ph.D. - http://homepage.mac.com/tmattox/
tmat...@gmail.com || timat...@open-mpi.org
   I'm a bright... http://www.the-brights.net/

_______________________________________________
devel mailing list
de...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/devel

Reply via email to