I've been doing stable-branch release of mesa for close to a year now.

In all of that time, there's one thing that I've never been very
comfortable with. Namely, release candidates are not tested very
thoroughly prior to release.[*] I'm glad that we haven't had any major
problems yet with broken releases. But I'd like to improve our release
testing rather than just trust to future luck.

Here are the changes that I'm proposing now for comments:

  1. The release schedule needs to be predictable.

        The current ideal is "every 2 weeks" for stable releases. I
        believe that is healthy and feasible.

        I propose that releases occur consistently every 2 weeks on
        Friday, (in the release manager's timezone---currently
        America/Los_Angeles).

  2. The release candidate needs to be made available in advance to
     allow for testing.

        I propose that the release manager push a candidate branch to
        the upstream repository by the end of the Monday prior to each
        scheduled Friday release.

  3. I'd like to receive a testing report from each driver team

        This is the meat of my proposal. I'm requesting that each driver
        team designate one (or more) people that will be responsible for
        running (or collecting) tests for each release-candidate and
        reporting the results to me.

        With a new release-candidate pushed by the end of the day on
        Monday, and me starting the actual release work on Friday, that
        gives 72 hours for each team to perform testing.

        I'm happy to let each driver team decide its own testing
        strategy. I would hope it would be based on running piglit
        across a set of "interesting" hardware configurations, (and
        augmenting piglit as necessary as bug fixes are performed). But
        I do not plan to be involved in the specification of what those
        configurations are. All I need is something along the lines of:

                "The radeon team is OK with releasing commit <foo>"

        sent to me before the scheduled day of the release.

        Obviously, any negative report from any team can trigger changes
        to the patches to be included in the release.

        And in order to put some teeth into this scheme:

        I propose that on the day of the release, the release manager
        drop all driver-specific patches for any driver for which the
        driver team has not provided an affirmative testing report.

Does the above sound like a reasonable scheme? If so,who from each driver
team is willing to volunteer to send me testing reports? I'll be happy
to send personal reminders to such volunteers as releases are getting
close and testing reports are missing, (before yanking out any
driver-specific patches).

Thanks in advance for any thoughts or comments,

-Carl

PS. This same testing scheme can obviously be used for major releases as
well, presumably with the same volunteers, (but different details on
release timing).

[*] For background, here's the testing scheme I've been using so far:

  1. Running each release candidate through piglit on my laptop

  2. Trying to leave the stable branch pushed out for at least a day or
     so to allow testing prior to release.

The testing I do on my laptop at least gives some sanity checking for
core, but very little driver-specific testing. Meanwhile, the majority
of stable-branch fixes seem to be in driver-specific code.

And I don't know if there is any serious testing of the stable branch by
anyone else between when I push it and when I release, (I haven't heard
much either way). I know it doesn't help that the timing of my releases
has not often been predictable.
_______________________________________________
mesa-dev mailing list
mesa-dev@lists.freedesktop.org
http://lists.freedesktop.org/mailman/listinfo/mesa-dev

Reply via email to