Chromium's buildbot now provides *automated monitoring of performance test regressions*. This new monitoring system alerts committers and sheriffs on the buildbot waterfall of regressions and speedups in select performance tests.
*What this means if you're a committer:* When you land a CL that affects Chromium's performance today, tests that measure performance could turn red or orange because of your CL's performance impact. If you or anyone else notices a perf regression from your change (eg. the sheriff), please revert your change asap. Performance is a key feature in Chromium, and regressing performance robs Chromium -- and your fellow developers -- of the work that's gone into making that feature a reality. In some cases reverting might not be straightforward, but reverting early on ensures Chromium's performance doesn't regress while you get the space and time necessary to address the unexpected performance impact of your CL before landing again. *What this means if you update reference builds:* Changing a reference build for a platform now also requires reconfiguring all of the expected performance values for that platform. See the Performance Test Plots<http://dev.chromium.org/developers/testing/chromium-build-infrastructure/performance-test-plots>page for more info. In the best case this will simply mean setting each affected value to 0 with the appropriate variances. *Not all perf test failures are perf regressions:* Performance tests can and will still fail if the test harness or slave itself fails running a test. When diagnosing why a test is red, regressions or speedups are denoted with the text "PERF_REGRESS" or "PERF_IMPROVE" in the test's status detail. (An example of a XP Perf morejs regression.<http://build.chromium.org/buildbot/waterfall/builders/XP%20Perf/builds/9740>) If you don't see the regress/improve line, look at the test's output for whether the failure was due to a crash in the harness or some other test component. *False positives can happen:* Most performance test results vary between test runs. This new perf monitoring system comes with independent variance settings to keep tests from triggering a failure/warning unless the results surpass those variance limits. Should a test run produce abnormal results, though, a false positive will occur. We will do our best to keep the noise down. The good news is that the handful of times noise appeared it was later found to be caused by a subtle-yet-very-real underlying performance change. *Where to learn more:* Read Chromium's Performance Test Plots<http://dev.chromium.org/developers/testing/chromium-build-infrastructure/performance-test-plots> docs for info along with a description for how we configure the expectations the system uses. The number of performance expectations will increase over time -- we're currently watching 6 traces with plans to add another 10 soon -- so feel free to familiarize yourself with the new system. And feel free to mail me (ch...@chromium.org) if you have any questions. Thanks to Nicolas, Marc-Antoine, Pam, Darin, and Steven for the help and input they provided to me while adding this feature. Chase --~--~---------~--~----~------------~-------~--~----~ Chromium Developers mailing list: chromium-dev@googlegroups.com View archives, change email options, or unsubscribe: http://groups.google.com/group/chromium-dev -~----------~----~----~----~------~----~------~--~---