[
https://issues.apache.org/jira/browse/FLINK-1195?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Alexander Alexandrov reassigned FLINK-1195:
-------------------------------------------
Assignee: Alexander Alexandrov
> Improvement of benchmarking infrastructure
> ------------------------------------------
>
> Key: FLINK-1195
> URL: https://issues.apache.org/jira/browse/FLINK-1195
> Project: Flink
> Issue Type: Wish
> Reporter: Till Rohrmann
> Assignee: Alexander Alexandrov
>
> I noticed while running my ALS benchmarks that we still have some potential
> to improve our benchmarking infrastructure. The current state is that we
> execute the benchmark jobs by writing a script with a single set of
> parameters. The runtime is then manually retrieved from the web interface of
> Flink and Spark, respectively.
> I think we need the following extensions:
> * Automatic runtime retrieval and storage in a file
> * Repeated execution of jobs to gather some "advanced" statistics such as
> mean and standard deviation of the runtimes
> * Support for value sets for the individual parameters
> The automatic runtime retrieval would allow us to execute several benchmarks
> consecutively without having to lookup the runtimes in the logs or in the web
> interface, which btw only stores the runtimes of the last 5 jobs.
> What I mean with value sets is that would be nice to specify a set of
> parameter values for which the benchmark is run without having to write for
> every single parameter combination a benchmark script. I believe that this
> feature would become very handy when we want to look at the runtime behaviour
> of Flink for different input sizes or degrees of parallelism, for example. To
> illustrate what I mean:
> {code}
> INPUTSIZE = 1000, 2000, 4000, 8000
> DOP = 1, 2, 4, 8
> OUTPUT=benchmarkResults
> repetitions=10
> command=benchmark.jar -p $DOP $INPUTSIZE
> {code}
> Something like that would execute the benchmark job with (DOP=1,
> INPUTSIZE=1000), (DOP=2, INPUTSIZE=2000),.... 10 times each, calculate for
> each parameter combination runtime statistics and store the results in the
> file benchmarkResults.
> I believe that spending some effort now will pay off in the long run because
> we will benchmark Flink continuously. What do you guys think?
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)