On Sun, Jul 18, 2010, Elazar Leibovich wrote about "Re: New Freecell Solver 
gcc-4.5.0 vs. LLVM+clang Benchmark":
> The standard deviation can give you an estimation of the minimal running
> time. (99.xxxx of the samples are within X standard deviations below the
> average. Pick a high enough xxxx relative to the number of times you'll run
> the software, and you'll get an estimation of the minimum running time
> you'll get).
> 
> The scalar representing the minimum running time have the obnoxious
> mis-feature that it doesn't tend to a certain value. So with a small
>...

Usually, the run time of deterministic programs does not behave the way
you describe (i.e., with a gaussian distribution, having an average and
distributed around it). A deterministic program *DOES* have a clearly
defined minimum run time, which it will theoretically achieve on every
run, unless delayed by something else - other processes, daemons in the
background, or whatever.

Imagine, for example, that you run a certain program 5 times and get the
times: 20.0, 18.0, 18.1, 27.0, 18.1
Evidently, the first run was slower because things were not in the cache,
and the run that took 27.0 was delayed by some other process in the background
taking up the CPU or disk. The minimum run time, 18.0, is the most interesting
one - it is the time a run would take every time, if things were perfect.
If you average the above numbers, or find the standard deviation, etc.,
the numbers would not be very interesting...

-- 
Nadav Har'El                        |           Monday, Jul 19 2010, 8 Av 5770
n...@math.technion.ac.il             |-----------------------------------------
Phone +972-523-790466, ICQ 13349191 |How long a minute depends on what side of
http://nadav.harel.org.il           |the bathroom door you're on.

_______________________________________________
Linux-il mailing list
Linux-il@cs.huji.ac.il
http://mailman.cs.huji.ac.il/mailman/listinfo/linux-il

Reply via email to