> > Are there features in Perfkit that we would like to be using that we > aren't? >
Besides the Kubernetes related code I mentioned above (that, I believe, can be easily replaced) I don't see any added value in having Perfkit. The Kubernetes parts could be replaced with a set of fine-grained Gradle tasks invoked by other high-level tasks and Jenkins job's steps. There also seem to be some Gradle + Kubernetes plugins out there that might prove useful here (no solid research in that area). > Can we make the integration with Perfkit less brittle? > There was an idea to move all beam benchmark's code from Perfkit ( beam_benchmark_helper.py <https://github.com/GoogleCloudPlatform/PerfKitBenchmarker/blob/5680e174ad1799056b4b6d4a6600ef9f93fe39ad/perfkitbenchmarker/beam_benchmark_helper.py> , beam_integration_benchmark.py <https://github.com/GoogleCloudPlatform/PerfKitBenchmarker/blob/7cdcea2561c66baa838e3ce4d776236a248e6700/perfkitbenchmarker/linux_benchmarks/beam_integration_benchmark.py>) to beam repository and inject it to Perfkit every time we use it. However, that would require investing time and effort in doing that and it will still not solve the problems I listed above. It will also still require knowledge of how Perfkit works from Beam developers while we can avoid that and use the existing tools (gradle, jenkins). Thanks! pt., 28 cze 2019 o 17:31 Lukasz Cwik <lc...@google.com> napisał(a): > +1 for removing tests that are not maintained. > > Are there features in Perfkit that we would like to be using that we > aren't? > Can we make the integration with Perfkit less brittle? > > If we aren't getting much and don't plan to get much value in the short > term, removal makes sense to me. > > On Thu, Jun 27, 2019 at 3:16 AM Łukasz Gajowy <lgaj...@apache.org> wrote: > >> Hi all, >> >> moving the discussion to the dev list: >> https://github.com/apache/beam/pull/8919. I think that Perfkit >> Benchmarker should be removed from all our tests. >> >> Problems that we face currently: >> >> 1. Changes to Gradle tasks/build configuration in the Beam codebase >> have to be reflected in Perfkit code. This required PRs to Perfkit which >> can last and the tests break due to this sometimes (no change in Perfkit + >> change already there in beam = incompatibility). This is what happened in >> PR 8919 (above), >> 2. Can't run in Python3 (depends on python 2 only library like >> functools32), >> 3. Black box testing which hard to collect pipeline related metrics, >> 4. Measurement of run time is inaccurate, >> 5. It offers relatively small elasticity in comparison with eg. >> Jenkins tasks in terms of setting up the testing infrastructure (runners, >> databases). For example, if we'd like to setup Flink runner, and reuse it >> in consequent tests in one go, that would be impossible. We can easily do >> this in Jenkins. >> >> Tests that use Perfkit: >> >> 1. IO integration tests, >> 2. Python performance tests, >> 3. beam_PerformanceTests_Dataflow (disabled), >> 4. beam_PerformanceTests_Spark (failing constantly - looks not >> maintained). >> >> From the IOIT perspective (1), only the code that setups/tears down >> Kubernetes resources is useful right now but these parts can be easily >> implemented in Jenkins/Gradle code. That would make Perfkit obsolete in >> IOIT because we already collect metrics using Metrics API and store them in >> BigQuery directly. >> >> As for point 2: I have no knowledge of how complex the task would be >> (help needed). >> >> Regarding 3, 4: Those tests seem to be not maintained - should we remove >> them? >> >> Opinions? >> >> Thank you, >> Łukasz >> >> >> >> >>