Hi Alberto --

Can you file an issue at https://github.com/stasm/raptor-compare/issues and
let me know which version of node you're using?  Perhaps you could also
attach the metrics.ldjson file you tried to run raptor-compare on?

Also, thanks, Zibi, for reminding the group about raptor-compare and happy
World Statistics Day! (Yes, it's a thing and it's today:
https://worldstatisticsday.org/ & http://www.un.org/en/events/statisticsday/
).

-stas

On Tue, Oct 20, 2015 at 11:04 AM, Alberto Pastor <[email protected]>
wrote:

> That sounds great!
>
> I’ve tried to follow the steps, but when running raptor-compare I got
>
> $ raptor-compare ./metrics.ldjson
> [TypeError: undefined is not a function]
>
> Is there anything I can attach to make more clear what’s going wrong?
>
> Thanks!
>
> > On 20 Oct 2015, at 09:04, [email protected] wrote:
> >
> > Hi all!
> >
> > I've been seeing a lot of people starting using raptor for testing
> performance of their patches/code, especially in the context of 2.2 -> 2.5
> regressions.
> >
> > That's awesome!
> >
> > Now, on top of that, :stas has developed a neat app that helps you get
> *more* out of those tests. In particular, it helps you learn if the
> difference you see is statistically significant[0].
> >
> > That's important. Not perfect yet, but super important. What it means is
> that it answers a question of wherever the change you see can be explained
> by fluctuations in results within your test.
> >
> > So instead of trying to guess, if the 100ms visuallyLoaded you see
> between two test results is real, install raptor-compare and follow the
> steps below:
> >
> > 1) Remove "metrics.ldjson" from the directory you are in
> > 2) Run your raptor test with as many runs as you can
> > 3) Apply your change
> > 4) Run your raptor test with the same amount of runes
> > 5) raptor-compare ./metrics.ldjson
> >
> > zbraniecki@rivia:~$ raptor-compare ./metrics.ldjson
> > fm.gaiamobile.org      base: mean  1: mean  1: delta  1: p-value
> > ---------------------  ----------  -------  --------  ----------
> > navigationLoaded              528      524        -4        0.72
> > navigationInteractive         738      721       -17        0.77
> > visuallyLoaded                738      721       -17        0.77
> > contentInteractive            738      722       -17        0.76
> > fullyLoaded                   923      903       -19        0.59
> > rss                        29.595   29.412    -0.183      * 0.02
> > uss                        11.098   11.001    -0.098      * 0.04
> > pss                        15.050   14.970    -0.080      * 0.03
> >
> > Reading the results - the most important thing is the little asterisk
> next to p-value[1]. If p-value is below 5% it suggests that the data
> observed is not consistent with the assumption that there are no difference
> between those two groups.
> >
> > In this example, it states, that there's less than 4% chance, the USS
> difference of almost 100kb is random.
> > At the same time the 20ms difference in fullyLoaded can be totally
> random.
> >
> > If you are getting p-value above 5%, you should reduce your trust in
> your results and consider rerunning your tests with more runs.
> >
> > Hope that helps!
> > zb.
> >
> >
> >
> > [0] https://en.wikipedia.org/wiki/Statistical_significance
> > [1] https://en.wikipedia.org/wiki/P-value
> > _______________________________________________
> > dev-fxos mailing list
> > [email protected]
> > https://lists.mozilla.org/listinfo/dev-fxos
>
> _______________________________________________
> dev-fxos mailing list
> [email protected]
> https://lists.mozilla.org/listinfo/dev-fxos
>
_______________________________________________
dev-fxos mailing list
[email protected]
https://lists.mozilla.org/listinfo/dev-fxos

Reply via email to