Lukasz Gajowy created BEAM-7772:
-----------------------------------

             Summary: Stop using Perfkit Benchmarker tool in all tests
                 Key: BEAM-7772
                 URL: https://issues.apache.org/jira/browse/BEAM-7772
             Project: Beam
          Issue Type: Bug
          Components: testing
            Reporter: Lukasz Gajowy
            Assignee: Lukasz Gajowy


[Devlist thread 
link|https://lists.apache.org/thread.html/dab1c093799248787e8b75e63b66d7389b594b649a4d9a4a5db1cfbb@%3Cdev.beam.apache.org%3E]

 

Currently Python, IOIT and some Dataflow and Spark performance tests are 
relying on Perfkit Benchmarker tool. Due to the reasons discussed on the 
devlist it was decided to remove it from Beam's tests. 

Problems that we face currently:
 # Changes to Gradle tasks/build configuration in the Beam codebase have to be 
reflected in Perfkit code. This required PRs to Perfkit which can last and the 
tests break due to this sometimes (no change in Perfkit + change already there 
in beam = incompatibility). This is what happened in PR 8919 (above),
 # Can't run in Python3 (depends on python 2 only library like functools32),
 # Black box testing which hard to collect pipeline related metrics,
 # Measurement of run time is inaccurate,
 # It offers relatively small elasticity in comparison with eg. Jenkins tasks 
in terms of setting up the testing infrastructure (runners, databases). For 
example, if we'd like to setup Flink runner, and reuse it in consequent tests 
in one go, that would be impossible. We can easily do this in Jenkins.

Tests that use Perfkit:
 # IO integration tests,
 # Python performance tests,
 # beam_PerformanceTests_Dataflow (disabled),
 # beam_PerformanceTests_Spark (failing constantly - looks not maintained).


 



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)

Reply via email to