Thanks.. I've read the pertinent sections and what I'm coming away with is that the chunk *sizes* should be designed around the I/O bandwidth of your disk subsystem, and the *shapes* should be designed around the access patterns for the data and around the data set itself (avoid mostly empty chunks and so on, as per 5.1.2 guidelines)...

What this doesn't really get into, it seems to me, is the role of the raw data chunk cache in all of this.

I don't think contiguous data is even an option for us, as we would have several multi-terabyte data sets which take quite some time just to initialize on disk.

Ruth Aydt wrote:
You may find some of the chunking discussions in this paper of interest:

http://www.hdfgroup.org/pubs/papers/2008-06_netcdf4_perf_report.pdf
in particular, section 3.2 and port6ions of sections 4 & 5.



_______________________________________________
Hdf-forum is for HDF software users discussion.
[email protected]
http://mail.hdfgroup.org/mailman/listinfo/hdf-forum_hdfgroup.org

Reply via email to