Barry Smith <bsm...@petsc.dev> writes:

>> On Jan 22, 2022, at 12:15 PM, Jed Brown <j...@jedbrown.org> wrote:
>> Barry, when you did the tech reports, did you make an example to reproduce 
>> on other architectures? Like, run this one example (it'll run all the 
>> benchmarks across different sizes) and then run this script on the output to 
>> make all the figures?
>
>    It is documented in 
> https://www.overleaf.com/project/5ff8f7aca589b2f7eb81c579    You may need to 
> dig through the submit scripts etc to find out exactly.

This runs a ton of small jobs and each job doesn't really preload, but instead 
of loops in job submission scripts, the loops could be inside the C code and it 
could directly output tabular data. This would run faster and be easier to 
submit and analyze.

https://gitlab.com/hannah_mairs/summit-performance/-/blob/master/summit-submissions/submit_gpu1.lsf

It would hopefully also avoid writing the size range manually over here in the 
analysis script where it has to match exactly the job submission.

https://gitlab.com/hannah_mairs/summit-performance/-/blob/master/python/graphs.py#L8-9


We'd make our lives a lot easier understanding new machines if we put into the 
design of performance studies just a fraction of the kind of thought we put 
into public library interfaces.

Reply via email to