Hi Izaak,

On Aug 12, 2011, at 11:32 AM, Izaak Beekman wrote:

> One more question about hdf5 parallel data set creation: here two 
> alternatives are presenting for writing data from a subset of MPI ranks. I 
> will be writting to a lustre filesystem, and and was wondering about the 
> performance differences between the two techniques. I have ordered data in a 
> 3D array, and the domain decomposition is done over the 2nd and 3rd indices 
> (Fortran code). So each MPI rank has a 3d array. The domain is decomposed, 
> and the processors layed out, as a 2D grid. I want to write 2D planar slices 
> of data out with high frequency. If the slice is taken as j=const or k=const 
> (where j and k are the 2nd and 3rd indices respectively, the indeces which 
> the separate subdomains) then only a relatively small (compared to the total 
> #) but still O(10-100) MPI ranks need to write data. I would say in general 
> less than 50% but more than 10% of the MPI ranks will be involved in this 
> write. Should I do it as a collective write where I guess there is some 
> over-head as the MPI ranks are cycled through in a predetermined order, or 
> should I hammer the lustre fs with uncoordinated individual writes, 
> concurrently? Any insight here is appreciated.

        I'm guessing that collective I/O will be a win still, but you may have 
to experiment a little...

                Quincey

_______________________________________________
Hdf-forum is for HDF software users discussion.
[email protected]
http://mail.hdfgroup.org/mailman/listinfo/hdf-forum_hdfgroup.org

Reply via email to