Thanks! I'm following up with this PR to display the Flink Pardo streaming data: https://github.com/apache/beam/pull/12408

Streaming data appears to be missing for Dataflow. We can revise the Jenkins jobs to add those.

-Max

On 29.07.20 17:01, Tyson Hamilton wrote:
Max,

The runner dimension are present when hovering over a particular graph. For some more info, the load test configurations can be found here [1]. I didn't get a chance to look into them but there are tests for all the runners there, possibly not for every loadtest.

[1]: https://github.com/apache/beam/tree/master/.test-infra/jenkins

-Tyson

On Wed, Jul 29, 2020 at 3:46 AM Maximilian Michels <[email protected] <mailto:[email protected]>> wrote:

    Looks like the permissions won't be necessary because backup data gets
    loaded into the local InfluxDb instance which makes writing queries
    locally possible.

    On 29.07.20 12:21, Maximilian Michels wrote:
     > Thanks Michał!
     >
     > It is a bit tricky to verify the exported query works if I don't
    have
     > access to the data stored in InfluxDb.
     >
     > ==> Could somebody give me permissions to [email protected]
    <mailto:[email protected]> for
     > apache-beam-testing such that I can setup a ssh port-forwarding
    from the
     > InfluxDb pod to my machine? I do have access to see the pods but
    that is
     > not enough.
     >
     >> I think that the only test data is from Python streaming tests,
    which
     >> are not implemented right now (check out
     >>
    
http://metrics.beam.apache.org/d/MOi-kf3Zk/pardo-load-tests?orgId=1&var-processingType=batch&var-sdk=python

     >> )
     >
     > Additionally, there is an entire dimension missing: Runners. I'm
     > assuming this data is for Dataflow?
     >
     > -Max
     >
     > On 29.07.20 11:55, Michał Walenia wrote:
     >> Hi there,
     >>
     >>  > Indeed the Python load test data appears to be missing:
     >>  >
     >>
    
http://metrics.beam.apache.org/d/MOi-kf3Zk/pardo-load-tests?orgId=1&var-processingType=streaming&var-sdk=python

     >>
     >>
     >> I think that the only test data is from Python streaming tests,
    which
     >> are not implemented right now (check out
     >>
    
http://metrics.beam.apache.org/d/MOi-kf3Zk/pardo-load-tests?orgId=1&var-processingType=batch&var-sdk=python

     >> )
     >>
     >> As for updating the dashboards, the manual for doing this is here:
     >>
    
https://cwiki.apache.org/confluence/display/BEAM/Community+Metrics#CommunityMetrics-UpdatingDashboards

     >>
     >>
     >> I hope this helps,
     >>
     >> Michal
     >>
     >> On Mon, Jul 27, 2020 at 4:31 PM Maximilian Michels
    <[email protected] <mailto:[email protected]>
     >> <mailto:[email protected] <mailto:[email protected]>>> wrote:
     >>
     >>     Indeed the Python load test data appears to be missing:
     >>
     >>
    
http://metrics.beam.apache.org/d/MOi-kf3Zk/pardo-load-tests?orgId=1&var-processingType=streaming&var-sdk=python

     >>
     >>
     >>     How do we typically modify the dashboards?
     >>
     >>     It looks like we need to edit this json file:
     >>
     >>
    
https://github.com/apache/beam/blob/8d460db620d2ff1257b0e092218294df15b409a1/.test-infra/metrics/grafana/dashboards/perftests_metrics/ParDo_Load_Tests.json#L81

     >>
     >>
     >>     I found some documentation on the deployment:
     >>
     >>
    https://cwiki.apache.org/confluence/display/BEAM/Test+Results+Monitoring
     >>
     >>
     >>     +1 for alerting or weekly emails including performance
    numbers for
     >>     fixed
     >>     intervals (1d, 1w, 1m, previous release).
     >>
     >>     +1 for linking the dashboards in the release guide to allow
    for a
     >>     comparison as part of the release process.
     >>
     >>     As a first step, consolidating all the data seems like the most
     >>     pressing
     >>     problem to solve.
     >>
     >>     @Kamil I could need some advice regarding how to proceed
    updating the
     >>     dashboards.
     >>
     >>     -Max
     >>
     >>     On 22.07.20 20:20, Robert Bradshaw wrote:
     >>      > On Tue, Jul 21, 2020 at 9:58 AM Thomas Weise
    <[email protected] <mailto:[email protected]>
     >>     <mailto:[email protected] <mailto:[email protected]>>
     >>      > <mailto:[email protected] <mailto:[email protected]>
    <mailto:[email protected] <mailto:[email protected]>>>> wrote:
     >>      >
     >>      >     It appears that there is coverage missing in the Grafana
     >>     dashboards
     >>      >     (it could also be that I just don't find it).
     >>      >
     >>      >     For example:
     >>      >
     >>
     >>
    https://apache-beam-testing.appspot.com/explore?dashboard=5751884853805056

     >>
     >>      >
     >>      >     The GBK and ParDo tests have a selection for {batch,
     >>     streaming} and
     >>      >     SDK. No coverage for streaming and python? There is
    also no
     >>     runner
     >>      >     option currently.
     >>      >
     >>      >     We have seen repeated regressions with streaming, Python,
     >>     Flink. The
     >>      >     test has been contributed. It would be great if the
    results
     >>     can be
     >>      >     covered as part of release verification.
     >>      >
     >>      >
     >>      > Even better would be if we can use these dashboards (plus
     >>     alerting or
     >>      > similar?) to find issues before release verification.
    It's much
     >>     easier
     >>      > to fix things earlier.
     >>      >
     >>      >
     >>      >     Thomas
     >>      >
     >>      >
     >>      >
     >>      >     On Tue, Jul 21, 2020 at 7:55 AM Kamil Wasilewski
     >>      >     <[email protected]
    <mailto:[email protected]>
     >>     <mailto:[email protected]
    <mailto:[email protected]>>
     >>     <mailto:[email protected]
    <mailto:[email protected]>
     >>     <mailto:[email protected]
    <mailto:[email protected]>>>>
     >>      >     wrote:
     >>      >
     >>      >             The prerequisite is that we have all the
    stats in one
     >>     place.
     >>      >             They seem
     >>      >             to be scattered across
     >> http://metrics.beam.apache.org and
     >>      > https://apache-beam-testing.appspot.com.
     >>      >
     >>      >             Would it be possible to consolidate the two,
    i.e.
     >> use the
     >>      >             Grafana-based
     >>      >             dashboard to load the legacy stats?
     >>      >
     >>      >
     >>      >         I'm pretty sure that all dashboards have been
    moved to
     >>      > http://metrics.beam.apache.org. Let me know if I missed
     >>      >         something during the migration.
     >>      >
     >>      >         I think we should turn off
     >>      > https://apache-beam-testing.appspot.com in the near
    future. New
     >>      >         Grafana-based dashboards have been working
    seamlessly for
     >>     some
     >>      >         time now and there's no point in maintaining the
    older
     >>     solution.
     >>      >         We'd also avoid ambiguity in where the stats
    should be
     >>     looked for.
     >>      >
     >>      >         Kamil
     >>      >
     >>      >         On Tue, Jul 21, 2020 at 4:17 PM Maximilian Michels
     >>      >         <[email protected] <mailto:[email protected]>
    <mailto:[email protected] <mailto:[email protected]>>
     >>     <mailto:[email protected] <mailto:[email protected]>
    <mailto:[email protected] <mailto:[email protected]>>>> wrote:
     >>      >
     >>      >              > It doesn't support https. I had to add an
     >> exception to
     >>      >             the HTTPS Everywhere extension for
     >>     "metrics.beam.apache.org <http://metrics.beam.apache.org>
    <http://metrics.beam.apache.org>
     >>      >             <http://metrics.beam.apache.org>".
     >>      >
     >>      >             *facepalm* Thanks Udi! It would always hang on me
     >>     because I
     >>      >             use HTTPS
     >>      >             Everywhere.
     >>      >
     >>      >              > To be explicit, I am supporting the idea of
     >>     reviewing the
     >>      >             release guide but not changing the release
    process
     >>     for the
     >>      >             already in-progress release.
     >>      >
     >>      >             I consider the release guide immutable for the
     >>     process of a
     >>      >             release.
     >>      >             Thus, a change to the release guide can only
    affect
     >> new
     >>      >             upcoming
     >>      >             releases, not an in-process release.
     >>      >
     >>      >              > +1 and I think we can also evaluate
    whether flaky
     >>     tests
     >>      >             should be reviewed as release blockers or
    not. Some
     >> flaky
     >>      >             tests would be hiding real issues our users
    could
     >> face.
     >>      >
     >>      >             Flaky tests are also worth to take into
    account when
     >>      >             releasing, but a
     >>      >             little harder to find because may just happen
    to pass
     >>     during
     >>      >             building
     >>      >             the release. It is possible though if we
    strictly
     >> capture
     >>      >             flaky tests
     >>      >             via JIRA and mark them with the Fix Version
    for the
     >>     release.
     >>      >
     >>      >              > We keep accumulating dashboards and
     >>      >              > tests that few people care about, so it is
     >>     probably worth
     >>      >             that we use
     >>      >              > them or get a way to alert us of regressions
     >>     during the
     >>      >             release cycle
     >>      >              > to catch this even before the RCs.
     >>      >
     >>      >             +1 The release guide should be explicit about
    which
     >>      >             performance test
     >>      >             results to evaluate.
     >>      >
     >>      >             The prerequisite is that we have all the
    stats in one
     >>     place.
     >>      >             They seem
     >>      >             to be scattered across
     >> http://metrics.beam.apache.org and
     >>      > https://apache-beam-testing.appspot.com.
     >>      >
     >>      >             Would it be possible to consolidate the two,
    i.e.
     >> use the
     >>      >             Grafana-based
     >>      >             dashboard to load the legacy stats?
     >>      >
     >>      >             For the evaluation during the release process, I
     >>     suggest to
     >>      >             use a
     >>      >             standardized set of performance tests for all
     >>     runners, e.g.:
     >>      >
     >>      >             - Nexmark
     >>      >             - ParDo (Classic/Portable)
     >>      >             - GroupByKey
     >>      >             - IO
     >>      >
     >>      >
     >>      >             -Max
     >>      >
     >>      >             On 21.07.20 01:23, Ahmet Altay wrote:
     >>      >              >
     >>      >              > On Mon, Jul 20, 2020 at 3:07 PM Ismaël Mejía
     >>      >             <[email protected] <mailto:[email protected]>
    <mailto:[email protected] <mailto:[email protected]>>
     >>     <mailto:[email protected] <mailto:[email protected]>
    <mailto:[email protected] <mailto:[email protected]>>>
     >>      >              > <mailto:[email protected]
    <mailto:[email protected]>
     >>     <mailto:[email protected] <mailto:[email protected]>>
    <mailto:[email protected] <mailto:[email protected]>
     >>     <mailto:[email protected] <mailto:[email protected]>>>>> wrote:
     >>      >              >
     >>      >              >     +1
     >>      >              >
     >>      >              >     This is not in the release guide and
    we should
     >>      >             probably re evaluate if
     >>      >              >     this should be a release blocking reason.
     >>      >              >     Of course exceptionally a performance
     >> regression
     >>      >             could be motivated by
     >>      >              >     a correctness fix or a worth refactor,
    so we
     >>     should
     >>      >             consider this.
     >>      >              >
     >>      >              >
     >>      >              > +1 and I think we can also evaluate whether
     >>     flaky tests
     >>      >             should be
     >>      >              > reviewed as release blockers or not. Some
    flaky
     >> tests
     >>      >             would be hiding
     >>      >              > real issues our users could face.
     >>      >              >
     >>      >              > To be explicit, I am supporting the idea of
     >>     reviewing the
     >>      >             release guide
     >>      >              > but not changing the release process for the
     >> already
     >>      >             in-progress release.
     >>      >              >
     >>      >              >
     >>      >              >     We have been tracking and fixing
    performance
     >>      >             regressions multiple
     >>      >              >     times found simply by checking the
    nexmark
     >> tests
     >>      >             including on the
     >>      >              >     ongoing 2.23.0 release so value is there.
     >> Nexmark
     >>      >             does not cover yet
     >>      >              >     python and portable runners so we are
    probably
     >>     still
     >>      >             missing many
     >>      >              >     issues and it is worth to work on
    this. In any
     >>     case
     >>      >             we should probably
     >>      >              >     decide what validations matter. We keep
     >>     accumulating
     >>      >             dashboards and
     >>      >              >     tests that few people care about, so it is
     >>     probably
     >>      >             worth that we use
     >>      >              >     them or get a way to alert us of
    regressions
     >>     during
     >>      >             the release cycle
     >>      >              >     to catch this even before the RCs.
     >>      >              >
     >>      >              >
     >>      >              > I agree. And if we cannot use
    dashboards/tests in a
     >>      >             meaningful way, IMO
     >>      >              > we can remove them. There is not much value to
     >>     maintain
     >>      >             them if they do
     >>      >              > not provide important signals.
     >>      >              >
     >>      >              >
     >>      >              >     On Fri, Jul 10, 2020 at 9:30 PM Udi Meiri
     >>      >             <[email protected] <mailto:[email protected]>
    <mailto:[email protected] <mailto:[email protected]>>
     >>     <mailto:[email protected] <mailto:[email protected]>
    <mailto:[email protected] <mailto:[email protected]>>>
     >>      >              >     <mailto:[email protected]
    <mailto:[email protected]>
     >>     <mailto:[email protected] <mailto:[email protected]>>
    <mailto:[email protected] <mailto:[email protected]>
     >>     <mailto:[email protected] <mailto:[email protected]>>>>>
     >>      >             wrote:
     >>      >              >      >
     >>      >              >      > On Thu, Jul 9, 2020 at 12:48 PM
    Maximilian
     >>     Michels
     >>      >              >     <[email protected]
    <mailto:[email protected]> <mailto:[email protected] <mailto:[email protected]>>
     >>     <mailto:[email protected] <mailto:[email protected]>
    <mailto:[email protected] <mailto:[email protected]>>>
     >>      >             <mailto:[email protected]
    <mailto:[email protected]> <mailto:[email protected] <mailto:[email protected]>>
     >>     <mailto:[email protected] <mailto:[email protected]>
    <mailto:[email protected] <mailto:[email protected]>>>>> wrote:
     >>      >              >      >>
     >>      >              >      >> Not yet, I just learned about the
     >>     migration to a
     >>      >             new frontend,
     >>      >              >     including
     >>      >              >      >> a new backend (InfluxDB instead of
     >> BigQuery).
     >>      >              >      >>
     >>      >              >      >> >  - Are the metrics available on
     >>      > metrics.beam.apache.org <http://metrics.beam.apache.org>
    <http://metrics.beam.apache.org>
     >>     <http://metrics.beam.apache.org>
     >>      >              >     <http://metrics.beam.apache.org>?
     >>      >              >      >>
     >>      >              >      >> Is http://metrics.beam.apache.org
    online?
     >>     I was
     >>      >             never able to
     >>      >              >     access it.
     >>      >              >      >
     >>      >              >      >
     >>      >              >      > It doesn't support https. I had to
    add an
     >>      >             exception to the HTTPS
     >>      >              >     Everywhere extension for
     >>     "metrics.beam.apache.org <http://metrics.beam.apache.org>
    <http://metrics.beam.apache.org>
     >>      >             <http://metrics.beam.apache.org>
     >>      >              >     <http://metrics.beam.apache.org>".
     >>      >              >      >
     >>      >              >      >>
     >>      >              >      >>
     >>      >              >      >> >  - What is the feature delta
    between
     >> usinig
     >>      >              > metrics.beam.apache.org
    <http://metrics.beam.apache.org>
     >>     <http://metrics.beam.apache.org>
    <http://metrics.beam.apache.org>
     >>      >             <http://metrics.beam.apache.org> (much
     >>      >              >     better UI) and using
     >> apache-beam-testing.appspot.com
    <http://apache-beam-testing.appspot.com>
     >> <http://apache-beam-testing.appspot.com>
     >>      >             <http://apache-beam-testing.appspot.com>
     >>      >              >     <http://apache-beam-testing.appspot.com>?
     >>      >              >      >>
     >>      >              >      >> AFAIK it is an ongoing migration
    and the
     >> delta
     >>      >             appears to be high.
     >>      >              >      >>
     >>      >              >      >> >  - Can we notice regressions
    faster than
     >>      >             release cadence?
     >>      >              >      >>
     >>      >              >      >> Absolutely! A report with the latest
     >> numbers
     >>      >             including
     >>      >              >     statistics about
     >>      >              >      >> the growth of metrics would be useful.
     >>      >              >      >>
     >>      >              >      >> >  - Can we get automated alerts?
     >>      >              >      >>
     >>      >              >      >> I think we could setup a Jenkins
    job to do
     >>     this.
     >>      >              >      >>
     >>      >              >      >> -Max
     >>      >              >      >>
     >>      >              >      >> On 09.07.20 20:26, Kenneth Knowles
    wrote:
     >>      >              >      >> > Questions:
     >>      >              >      >> >
     >>      >              >      >> >   - Are the metrics available on
     >>      > metrics.beam.apache.org <http://metrics.beam.apache.org>
    <http://metrics.beam.apache.org>
     >>     <http://metrics.beam.apache.org>
     >>      >              >     <http://metrics.beam.apache.org>
     >>      >              >      >> > <http://metrics.beam.apache.org>?
     >>      >              >      >> >   - What is the feature delta
    between
     >> usinig
     >>      >              > metrics.beam.apache.org
    <http://metrics.beam.apache.org>
     >>     <http://metrics.beam.apache.org>
    <http://metrics.beam.apache.org>
     >>      >             <http://metrics.beam.apache.org>
     >>      >              >      >> > <http://metrics.beam.apache.org>
    (much
     >>     better
     >>      >             UI) and using
     >>      >              >      >> > apache-beam-testing.appspot.com
    <http://apache-beam-testing.appspot.com>
     >>     <http://apache-beam-testing.appspot.com>
     >>      >             <http://apache-beam-testing.appspot.com>
     >>      >              >     <http://apache-beam-testing.appspot.com>
     >>      >              >     <http://apache-beam-testing.appspot.com>?
     >>      >              >      >> >   - Can we notice regressions
    faster than
     >>      >             release cadence?
     >>      >              >      >> >   - Can we get automated alerts?
     >>      >              >      >> >
     >>      >              >      >> > Kenn
     >>      >              >      >> >
     >>      >              >      >> > On Thu, Jul 9, 2020 at 10:21 AM
     >>     Maximilian Michels
     >>      >              >     <[email protected]
    <mailto:[email protected]> <mailto:[email protected] <mailto:[email protected]>>
     >>     <mailto:[email protected] <mailto:[email protected]>
    <mailto:[email protected] <mailto:[email protected]>>>
     >>      >             <mailto:[email protected]
    <mailto:[email protected]> <mailto:[email protected] <mailto:[email protected]>>
     >>     <mailto:[email protected] <mailto:[email protected]>
    <mailto:[email protected] <mailto:[email protected]>>>>
     >>      >              >      >> > <mailto:[email protected]
    <mailto:[email protected]>
     >>     <mailto:[email protected] <mailto:[email protected]>>
    <mailto:[email protected] <mailto:[email protected]>
     >> <mailto:[email protected] <mailto:[email protected]>>>
     >>      >             <mailto:[email protected]
    <mailto:[email protected]> <mailto:[email protected] <mailto:[email protected]>>
     >>     <mailto:[email protected] <mailto:[email protected]>
    <mailto:[email protected] <mailto:[email protected]>>>>>> wrote:
     >>      >              >      >> >
     >>      >              >      >> >     Hi,
     >>      >              >      >> >
     >>      >              >      >> >     We recently saw an increase in
     >> latency
     >>      >             migrating from Beam
     >>      >              >     2.18.0 to
     >>      >              >      >> >     2.21.0 (Python SDK with Flink
     >>     Runner). This
     >>      >             proofed very
     >>      >              >     hard to debug
     >>      >              >      >> >     and it looks like each
    version in
     >>     between
     >>      >             the two versions
     >>      >              >     let to
     >>      >              >      >> >     increased latency.
     >>      >              >      >> >
     >>      >              >      >> >     This is not the first time
    we saw
     >> issues
     >>      >             when migrating,
     >>      >              >     another
     >>      >              >      >> >     time we
     >>      >              >      >> >     had a decline in checkpointing
     >>     performance
     >>      >             and thus added a
     >>      >              >      >> >     checkpointing test [1] and
    dashboard
     >>     [2] (see
     >>      >              >     checkpointing widget).
     >>      >              >      >> >
     >>      >              >      >> >     That makes me wonder if we
    should
     >>     monitor
     >>      >             performance
     >>      >              >     (throughput /
     >>      >              >      >> >     latency) for basic use cases
    as part
     >>     of the
     >>      >             release
     >>      >              >     testing. Currently,
     >>      >              >      >> >     our release guide [3] mentions
     >> running
     >>      >             examples but not
     >>      >              >     evaluating the
     >>      >              >      >> >     performance. I think it
    would be good
     >>      >             practice to check
     >>      >              >     relevant charts
     >>      >              >      >> >     with performance measurements as
     >>     part of of
     >>      >             the release
     >>      >              >     process. The
     >>      >              >      >> >     release guide should reflect
    that.
     >>      >              >      >> >
     >>      >              >      >> >     WDYT?
     >>      >              >      >> >
     >>      >              >      >> >     -Max
     >>      >              >      >> >
     >>      >              >      >> >     PS: Of course, this requires
    tests
     >> and
     >>      >             metrics to be
     >>      >              >     available. This PR
     >>      >              >      >> >     adds latency measurements to the
     >>     load tests
     >>      >             [4].
     >>      >              >      >> >
     >>      >              >      >> >
     >>      >              >      >> >     [1]
     >> https://github.com/apache/beam/pull/11558
     >>      >              >      >> >     [2]
     >>      >              >      >> >
     >>      >              >
     >>      >
     >>
     >>
    https://apache-beam-testing.appspot.com/explore?dashboard=5751884853805056

     >>
     >>      >              >      >> >     [3]
     >>      > https://beam.apache.org/contribute/release-guide/
     >>      >              >      >> >     [4]
     >> https://github.com/apache/beam/pull/12065
     >>      >              >      >> >
     >>      >              >
     >>      >
     >>
     >>
     >>
     >> --
     >>
     >> Michał Walenia
     >> Polidea <https://www.polidea.com/> | Software Engineer
     >>
     >> M: +48 791 432 002 <tel:+48%20791%20432%20002> <tel:+48791432002
    <tel:+48%20791%20432%20002>>
     >> E: [email protected]
    <mailto:[email protected]>
    <mailto:[email protected] <mailto:[email protected]>>
     >>
     >> Unique Tech
     >> Check out our projects! <https://www.polidea.com/our-work>
     >>

Reply via email to