On Tue, Oct 6, 2015 at 7:21 PM, Eli Perelman <[email protected]> wrote:

> On Tue, Oct 6, 2015 at 11:58 AM, Etienne Segonzac <[email protected]>
> wrote:
>
>>
>> So we're fine with the system that didn't work for 2.5 and we're making
>> no promise for the future.
>> Nice commitment to performance.
>>
>
> I'd hardly say that what we've built doesn't have a positive impact
> towards performance. The fact that this conversation can even exist with
> real data is a testament to how far we've come. The intent of my email was
> not to back people into corners and play blame games, but just to shine a
> light as to what things look like right now so owners and peers have
> ammunition to make decisions. Let me repeat what I just said, because it is
> the crux of the problem:
>
>
>
> *The current scope of performance automation is not in the state that it
> is something automatically sheriffable. We've built up the tools and
> infrastructure from almost the ground-up and it has accomplished exactly
> what it was set out to do: to put the knowledge of performance and problems
> into the hands of owners and peers to make their own decisions.*
> Automating performance is usually not a binary decision like a unit test.
> It takes analysis and guesswork, and even then it still needs human eyes.
> Rob and I are working towards making this better, to automate as much as
> possible, but right now the burden of making the tough calls still lies
> with those landing patches. We equip you to make those determinations until
> we have more tooling and automation in place for the sheriffing to actually
> be an option, because right now *it is not*.
>


We're back to my first message on the thread.
We don't have the adequate tooling to achieve our performance goal.

Every release we talk about performance like if the issue was a "developer
awareness" issue, and we take strong stance on how "we should never
regress".
But if we meant it we'd have more that 2 people working on the very
challenging tooling work required. And believe me I'm fully aware of how
challenging it is.

We can't hand every gaia and gecko developer a link to moztrap (manual test
case tracker), remove all automated tests, and then be all high-minded
about how we should never regress a feature. But it's exactly what we're
doing with launch time performance.
_______________________________________________
dev-fxos mailing list
[email protected]
https://lists.mozilla.org/listinfo/dev-fxos

Reply via email to