Marty Barnaby wrote:
I'm attempting to establish an absolute maximum, byte-rate performance
value, running a bare bones MPI_File_write_all_at benchmark program,
for our Cray XT3 installation, RedStorm, here at Sandia National
Laboratories. Processor time is at a premium, and I only run in the
standard queue, so I'm not able to do everything I would imagine,
though maybe what I can run is adequate.
I have a directory under our Lustre, redstorm:/scratch_grande which I
have defined with:
lfs setstripe -1 0 -1
Though there are 320 OST's comprising the FS, these defaults give me a
stripe_count of 160 (I'm sure someone could explain that), and I don't
know the stripe_size. With a job of 160 processors, each of which has
a contiguous chuck of 20 MB of memory, respectively, to append to an
open file in an iterative series of singular, atomic, write_all
operations, I can normally average 25 GB/s. To curb any confusion
here, that represents only an experimental maximum to me; none of our
many, complex, science and engineering simulation applications perform
their output dumping with per-processor blocks as large as a single MB.
I would like any succinct suggestions on explicitly setting my lfs
stripe_size, given the configuration and parameters I've mentioned
here, to optimize it and, perhaps see a decrease in the time spent
storing my data on the FS.
Try setting your stripesize to 20MB. As Kalpak mentioned, we currently
have a limit of 160 OSTs for any 1 file (although, of course, there are
plans to remove this limitation soon).
Would you mind posting your test prog? I can imagine others (besides
me) might be interested in such experimental maximums.
_______________________________________________
Lustre-discuss mailing list
[email protected]
https://mail.clusterfs.com/mailman/listinfo/lustre-discuss