Another piece of information that would be useful for the h5perf benchmark/test would be the resulting file sizes for the write operation. From this, one could get a clear idea of just how much overhead a particular dataset layout was costing, both in the relative sense (by comparing one HDF5 layout to another) and in an "absolute" sense (by comparing an HDF5 layout to the raw posix file size).

Quincey Koziol wrote:
Hi John,

On May 9, 2011, at 3:17 PM, John Knutson wrote:

I've been using h5perf_serial in the process of testing the iRODS VFD that I've 
been working on, and have ultimately found myself thinking it might be time to 
revisit my choices for chunking and transfer sizes for our data.  Eventually I 
remembered that throughput wasn't the only or even the primary measure of 
performance in our situation.

My suggestion is that there also be a performance benchmark that measures 
latency, i.e., the amount of time it takes to do a single write or read.  When 
you're streaming data like we are, the ideal chunk size for latency may yield 
very low throughput.

Additionally, while h5perf_serial supports the option of only doing write 
tests, it would be really handy to have a similar option for only doing read 
tests - that is, decouple reading and writing to have a better chance of 
getting read throughput measurements that are really only showing the 
throughput of the cache (disk, system, whatever it may be).

        We're planning to update the parallel I/O performance benchmark and 
will keep this in mind, trying to migrate the changes back to the serial 
version also.

        Thanks,
                Quincey


_______________________________________________
Hdf-forum is for HDF software users discussion.
Hdf-forum@hdfgroup.org
http://mail.hdfgroup.org/mailman/listinfo/hdf-forum_hdfgroup.org

_______________________________________________
Hdf-forum is for HDF software users discussion.
Hdf-forum@hdfgroup.org
http://mail.hdfgroup.org/mailman/listinfo/hdf-forum_hdfgroup.org

Reply via email to