2017-03-16 2:04 GMT+01:00 Wang, Peter Xihong <peter.xihong.w...@intel.com>:
> Understood on the obsolete benchmark part.  This was the work done before the 
> new benchmark was created on github.

I strongly advice you to move to performance. It also has a nice a
API. It now produces a JSON file with *all* data, instead of just
writing into summaries into stdout.

> I thought this is related, and thus didn't open a new thread.

The other thread was a discussion about statistics, how to summarize
all timing into two numbers :-)

> Maybe you could point me to one single micro-benchmark for the time being, 
> and then we could compare result across?

The "new" performance project is a fork of the old "benchmark"
project. Benchmark names are very close or even the same for many
benchmarks.

If you would like to validate that your benchmark runner is stable:
run call_method and call_simple microbenchmarks on different revisions
of CPython, reboot sometimes the computer used to run benchmarks, and
make sure that results are stable.

Compare them with results of speed.python.org.

call_method:
https://speed.python.org/timeline/#/?exe=5&ben=call_method&env=1&revs=50&equid=off&quarts=on&extr=on

call_simple:
https://speed.python.org/timeline/#/?exe=5&ben=call_simple&env=1&revs=50&equid=off&quarts=on&extr=on

Around november and december 2016, you should notice a significant
speedup on call_method.

The best is to be able to avoid "temporary spikes" like this one:
https://haypo.github.io/analysis-python-performance-issue.html

The API of the perf project, PGO and LTO compilation, new performance
using perf, "perf system tune" for system tuning, etc. helped to get
more stable results.

Victor
_______________________________________________
Speed mailing list
Speed@python.org
https://mail.python.org/mailman/listinfo/speed

Reply via email to