Hello,

When testing different chunk size for dataset compression, I obtain the
following results for the writing of a dataset of size 384 x 256 x 1024:

Chunk size        Global max memory      File size (MB)      Write time
(seconds)
2 ^ 15                22%                            15                      
24
2 ^ 12                35%                            16                      
22
2 ^ 10                58%                            26                      
22

Tests realized on Intel Xeon CPU E5530 @ 2.40GHz 4 cores, 4G RAM, Linux SLED
11, HDF5 1.8.3. I'm using the Nexus library that is a thin layer on top of
HDF.

The memory footprint is increasing when the chunk size is decreasing. That
is not an intuitive behaviour. Has somebody an explanation?

Yannick





Same post



--
View this message in context: 
http://hdf-forum.184993.n3.nabble.com/HDF5-chunk-profiling-tp4025718.html
Sent from the hdf-forum mailing list archive at Nabble.com.

_______________________________________________
Hdf-forum is for HDF software users discussion.
[email protected]
http://mail.hdfgroup.org/mailman/listinfo/hdf-forum_hdfgroup.org

Reply via email to