Wow, this is really cool, and something I've wanted for a long time. Nice work! Patrick
On Fri, Oct 16, 2009 at 12:35 AM, Chase Phillips <ch...@chromium.org> wrote: > Chromium's buildbot now provides *automated monitoring of performance test > regressions*. This new monitoring system alerts committers and sheriffs > on the buildbot waterfall of regressions and speedups in select performance > tests. > > *What this means if you're a committer:* > When you land a CL that affects Chromium's performance today, tests that > measure performance could turn red or orange because of your CL's > performance impact. If you or anyone else notices a perf regression from > your change (eg. the sheriff), please revert your change asap. Performance > is a key feature in Chromium, and regressing performance robs Chromium -- > and your fellow developers -- of the work that's gone into making that > feature a reality. > > In some cases reverting might not be straightforward, but reverting early > on ensures Chromium's performance doesn't regress while you get the space > and time necessary to address the unexpected performance impact of your CL > before landing again. > > *What this means if you update reference builds:* > Changing a reference build for a platform now also requires reconfiguring > all of the expected performance values for that platform. See the Performance > Test > Plots<http://dev.chromium.org/developers/testing/chromium-build-infrastructure/performance-test-plots>page > for more info. In the best case this will simply mean setting each > affected value to 0 with the appropriate variances. > > *Not all perf test failures are perf regressions:* > Performance tests can and will still fail if the test harness or slave > itself fails running a test. When diagnosing why a test is red, regressions > or speedups are denoted with the text "PERF_REGRESS" or "PERF_IMPROVE" in > the test's status detail. (An example of a XP Perf morejs > regression.<http://build.chromium.org/buildbot/waterfall/builders/XP%20Perf/builds/9740>) > If you don't see the regress/improve line, look at the test's output for > whether the failure was due to a crash in the harness or some other test > component. > > *False positives can happen:* > Most performance test results vary between test runs. This new perf > monitoring system comes with independent variance settings to keep tests > from triggering a failure/warning unless the results surpass those variance > limits. Should a test run produce abnormal results, though, a false positive > will occur. We will do our best to keep the noise down. The good news is > that the handful of times noise appeared it was later found to be caused by > a subtle-yet-very-real underlying performance change. > > *Where to learn more:* > Read Chromium's Performance Test > Plots<http://dev.chromium.org/developers/testing/chromium-build-infrastructure/performance-test-plots> > docs > for info along with a description for how we configure the expectations the > system uses. The number of performance expectations will increase over time > -- we're currently watching 6 traces with plans to add another 10 soon -- so > feel free to familiarize yourself with the new system. And feel free to > mail me (ch...@chromium.org) if you have any questions. > > Thanks to Nicolas, Marc-Antoine, Pam, Darin, and Steven for the help and > input they provided to me while adding this feature. > > Chase > > > > --~--~---------~--~----~------------~-------~--~----~ Chromium Developers mailing list: chromium-dev@googlegroups.com View archives, change email options, or unsubscribe: http://groups.google.com/group/chromium-dev -~----------~----~----~----~------~----~------~--~---