And this you have likely seen already in
Wikipedia:https://en.wikipedia.org/wiki/Principal_component_analysis"...PCA is
mostly used as a tool in exploratory data analysis and for making predictive
models. It's often used to visualize genetic distance and relatedness between
populations. PCA can be done by eigenvalue decomposition of a data covariance
(or correlation) matrix or singular value decomposition of a data matrix,
usually after mean centering[clarification needed] (and normalizing or using
Z-scores) the data matrix for each attribute.[4] The results of a PCA are
usually discussed in terms of component scores, sometimes called factor scores
(the transformed variable values corresponding to a particular data point), and
loadings (the weight by which each standardized original variable should be
multiplied to get the component score)..."
On Saturday, May 26, 2018, 10:10:32 PM PDT, Shiheng Duan
<[email protected]> wrote:
Thanks.
Do you mean that if feature one has a larger derivation than feature two, after
zscore they will have the same weight? In that case, it is a bias, right? The
feature one should be more important than feature two in the PCA.
On Thu, May 24, 2018 at 5:09 PM, Michael Eickenberg
<[email protected]> wrote:
Hi,
that totally depends on the nature of your data and whether the standard
deviation of individual feature axes/columns of your data carry some form of
importance measure. Note that PCA will bias its loadings towards columns with
large standard deviations all else being held equal (meaning that if you have
zscored columns, and then you choose one column and multiply it by, say 1000,
then that component will likely show up as your first component [if 1000 is
comparable or large wrt the number of features you are using])
Does this help?Michael
On Thu, May 24, 2018 at 4:39 PM, Shiheng Duan <[email protected]> wrote:
Hello all,
I wonder is it necessary or correct to do z score transformation before PCA? I
didn't see any preprocessing for face image in the example of Faces recognition
example using eigenfaces and SVMs, link:http://scikit-learn.org/s
table/auto_examples/applicatio ns/plot_face_recognition.html#
sphx-glr-auto-examples- applications-plot-face- recognition-py
I am doing on a similar dataset and got a weird result if I standardized data
before PCA. The components figure will have a strong gradient and it doesn't
make any sense. Any ideas about the reason?
Thanks.
______________________________ _________________
scikit-learn mailing list
[email protected]
https://mail.python.org/mailma n/listinfo/scikit-learn
______________________________ _________________
scikit-learn mailing list
[email protected]
https://mail.python.org/ mailman/listinfo/scikit-learn
_______________________________________________
scikit-learn mailing list
[email protected]
https://mail.python.org/mailman/listinfo/scikit-learn
_______________________________________________
scikit-learn mailing list
[email protected]
https://mail.python.org/mailman/listinfo/scikit-learn