At 09:56 AM 6/10/2005, Fred Baker wrote:
you saw marshall's comment. If you're interested in a moving average, he's
pretty close.
If I understood your question, though, you simply wanted to quantify the
jitter in a set of samples. I should think there are two obvious
definitions there.
A statistician would look, I should think, at the variance of the set.
Reaching for my CRC book of standard math formulae and tables, it defines
the variance as the square of the standard deviation of the set, which is
to say
That is one thing I have never understood, if you can pretty much just look
at a standard dev and see it is high, and yeah that means your numbers are
flopping all over the place, then what good is the square of it? Does it
just make graphing better in some way?
Thanks,
Eric