Hi Areg,
That sounds like a good idea to me. Note our benchmarks are currently scattered accross the various implementations. The two that I know of: - the C++ benchmarks are standalone executables created using the Google Benchmark library, aptly named "*-benchmark" (or "*-benchmark.exe" on Windows) - the Python benchmarks use the ASV utility: https://github.com/apache/arrow/blob/master/docs/source/python/benchmarks.rst There may be more in the other implementations. Regards Antoine. Le 18/01/2019 à 07:13, Melik-Adamyan, Areg a écrit : > Hello, > > I want to restart/attach to the discussions for creating Arrow benchmarking > dashboard. I want to propose performance benchmark run per commit to track > the changes. > The proposal includes building infrastructure for per-commit tracking > comprising of the following parts: > - Hosted JetBrains for OSS https://teamcity.jetbrains.com/ as a build system > - Agents running in cloud both VM/container (DigitalOcean, or others) and > bare-metal (Packet.net/AWS) and on-premise(Nvidia boxes?) > - JFrog artifactory storage and management for OSS projects > https://jfrog.com/open-source/#artifactory2 > - Codespeed as a frontend https://github.com/tobami/codespeed > > I am volunteering to build such system (if needed more Intel folks will be > involved) so we can start tracking performance on various platforms and > understand how changes affect it. > > Please, let me know your thoughts! > > Thanks, > -Areg. > > >