Hi, I need to implement a storage for data with the following properties:
1) multi-dimensional unlimited size data set of variable-length records 2) may be highly sparsed 3) usually randomly accessed one record at a time 4) each record may vary in size from tens of kilobytes to tens of megabytes I am thinking of unlimited chunked data space. However to make it efficient in terms of disk space and access time I need to have my chunks as small as one element. Could you please save me performance test and tell if such configuration is practical with HDF5? Thanks, Efim ------------------- This e-mail, including any attached files, may contain confidential and privileged information for the sole use of the intended recipient. Any review, use, distribution, or disclosure by others is strictly prohibited. If you are not the intended recipient (or authorized to receive information for the intended recipient), please contact the sender by reply e-mail and delete all copies of this message.
_______________________________________________ Hdf-forum is for HDF software users discussion. [email protected] http://lists.hdfgroup.org/mailman/listinfo/hdf-forum_lists.hdfgroup.org Twitter: https://twitter.com/hdf5
