On Mon, Jun 08, 2009 at 06:28:06AM -0700, Keith Goodman wrote:
> On Mon, Jun 8, 2009 at 6:17 AM, Gael Varoquaux
> <gael.varoqu...@normalesup.org> wrote:
> > On Mon, Jun 08, 2009 at 09:02:12AM -0400, josef.p...@gmail.com wrote:
> >> whats the actual shape of the array/data you run your PCA on.

> > 50 000 dimensions, 820 datapoints.

> Have you tried shuffling each time series, performing PCA, looking at
> the magnitude of the largest eigenvalue, then repeating many times?
> That will give you an idea of how large the noise can be. Then you can
> see how many eigenvectors of the unshuffled data have eigenvalues
> greater than the noise. It would be kind of the empirical approach to
> random matrix theory.

Yes, that's the kind of things that is done in the paper I pointed out
and I use to infer the number of PCAs I retain.

Gaël
_______________________________________________
Numpy-discussion mailing list
Numpy-discussion@scipy.org
http://mail.scipy.org/mailman/listinfo/numpy-discussion

Reply via email to