On 08/08/2014 03:27 AM, houssen wrote:
In short : are there things to know / make sure of / be aware of to get
good performance with P-HDF5 ?
- turn on collective I/O. it's not enabled by default
- HDF5 metadata might be a factor if you have very many small datasets,
but for most applications it's not important
- consult your MPI library for any file-system specific tuning you might
be able to do. For example, Intel-MPI needs you to set an environment
variable before it will use any of the GPFS or Panasas optimizations it
has written.
- be mindful of type conversions: if your data in memory is a 4-byte
float, but they are 8-byte doubles on disk, HDF5 will "break collective"
and do that I/O independently.
To test this I wrote a MPI code. ... I expected to get better
performance with MPI-IO and P-HDF5 than with the sequential approach.
The spirit of this test code is very simple / basic (each MPI process
writes his own block of data in the same file, or, in separate files in
the sequential approach).
Note : in each case (sequential, MPI-IO, P-HDF5), when I say "write data
in file", I mean writing big blocks / bunch of data at once (I do not
write data one by one - I write the biggest block of data, but smaller
than 2Gb, that is possible to write).
Note : I tried with N = 1, 2, 4, 8, 16.
in 2014, 16 is not very parallel. serial I/O has many benefits at
modest levels of parallelism: caching, mostly.
Note : I generated files (MPI-IO, P-HDF5) whose size scaled from 1Gb to
16 Gb (which looks like a "very big" file to me).
that's adequate, yes
Note : I followed the P-HDF5 documentation (use H5P_FILE_ACCESS and
H5P_DATASET_XFER property list + use hyperslab "by chunks")
Note : the file system is "GPFS" (it has been installed by the cluster
vendor : this is supposed to be ready to get performance out of P-HDF5 -
I am an "application" guy that try to use HDF5, I am not an "admin sys"
that would be familiar with complex related stuffs related to the file
system)
Now we are getting somewhere.
Note : I compiled the HDF5 package like this "./configure
--enable-parallel".
Note : I use CentOS + GNU compilers (for both HDF5 package and my test
code) + hdf5-1.8.13
Note : I use mpic++ (not h5pxx compilers - actually I didn't get why
HDF5 provides compilers) to compile my test code, is this a problem ?
just makes it easier to pick up any libraries needed. I don't use the
wrappers, either, which means sometimes I need to figure out what new
library (like -ldl) HDF5 needs.
Any relevant clue / information would be appreciated. If what I observe
is logical I would just understand why, and, how / when it is possible
to get performance out of P-HDF5. I just would like to get some logic
out of this.
If you are using GPFS, there is one optimization that goes a long way
towards improving performance: aligning writes to file system block
boundaries. See this email from a few weeks ago:
http://mail.lists.hdfgroup.org/pipermail/hdf-forum_lists.hdfgroup.org/2014-July/007963.html
==rob
Thanks for help,
FH
PS : I can give more information and the code, if needed (?)
_______________________________________________
Hdf-forum is for HDF software users discussion.
[email protected]
http://mail.lists.hdfgroup.org/mailman/listinfo/hdf-forum_lists.hdfgroup.org
Twitter: https://twitter.com/hdf5
--
Rob Latham
Mathematics and Computer Science Division
Argonne National Lab, IL USA
_______________________________________________
Hdf-forum is for HDF software users discussion.
[email protected]
http://mail.lists.hdfgroup.org/mailman/listinfo/hdf-forum_lists.hdfgroup.org
Twitter: https://twitter.com/hdf5