Barry Smith <bsm...@petsc.dev> writes: >> On Jan 22, 2022, at 12:15 PM, Jed Brown <j...@jedbrown.org> wrote: >> Barry, when you did the tech reports, did you make an example to reproduce >> on other architectures? Like, run this one example (it'll run all the >> benchmarks across different sizes) and then run this script on the output to >> make all the figures? > > It is documented in > https://www.overleaf.com/project/5ff8f7aca589b2f7eb81c579 You may need to > dig through the submit scripts etc to find out exactly.
This runs a ton of small jobs and each job doesn't really preload, but instead of loops in job submission scripts, the loops could be inside the C code and it could directly output tabular data. This would run faster and be easier to submit and analyze. https://gitlab.com/hannah_mairs/summit-performance/-/blob/master/summit-submissions/submit_gpu1.lsf It would hopefully also avoid writing the size range manually over here in the analysis script where it has to match exactly the job submission. https://gitlab.com/hannah_mairs/summit-performance/-/blob/master/python/graphs.py#L8-9 We'd make our lives a lot easier understanding new machines if we put into the design of performance studies just a fraction of the kind of thought we put into public library interfaces.