[ 
https://issues.apache.org/jira/browse/SPARK-7856?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tarek Elgamal updated SPARK-7856:
---------------------------------
    Description: 
Currently the PCA implementation has a limitation of fitting d^2 
covariance/grammian matrix entries in memory (d is the number of 
columns/dimensions of the matrix). We often need only the largest k principal 
components. To make pca really scalable, I suggest an implementation where the 
memory usage is proportional to the principal components k rather than the full 
dimensionality d. 

I suggest adopting the solution described in this paper that is published in 
SIGMOD 2015 (http://ds.qcri.org/images/profile/tarek_elgamal/sigmod2015.pdf). 
The paper offers an implementation for Probabilistic PCA (PPCA) which has less 
memory and time complexity and could potentially scale to tall and fat matrices 
rather than tall and skinny matrices that is supported by the current PCA 
impelmentation. 
Probablistic PCA could be potentially added to the set of algorithms supported 
by MLlib and it does not necessarily replace the old PCA implementation.

PPCA implementation is adopted in Matlab's Statistics and Machine Learning 
Toolbox (http://www.mathworks.com/help/stats/ppca.html)

  was:
Currently the PCA implementation has a limitation of fitting d^2 
covariance/grammian matrix entries in memory (d is the number of 
columns/dimensions of the matrix). We often need only the largest k principal 
components. To make pca really scalable, I suggest an implementation where the 
memory usage is proportional to the principal components k rather than the full 
dimensionality d. 

I suggest adopting the solution described in this paper that is published in 
SIGMOD 2015 (http://ds.qcri.org/images/profile/tarek_elgamal/sigmod2015.pdf). 
The paper offers an implementation for Probabilistic PCA (PPCA) which has less 
memory and time complexity and could potentially scale to tall and fat matrices 
rather than tall and skinny matrices that is supported by the current PCA 
impelmentation. 
Probablistic PCA could be potentially added to the set of algorithms supported 
by MLlib and it does not necessarily replace the old PCA implementation.


> Scalable PCA implementation for tall and fat matrices
> -----------------------------------------------------
>
>                 Key: SPARK-7856
>                 URL: https://issues.apache.org/jira/browse/SPARK-7856
>             Project: Spark
>          Issue Type: Bug
>          Components: MLlib
>            Reporter: Tarek Elgamal
>
> Currently the PCA implementation has a limitation of fitting d^2 
> covariance/grammian matrix entries in memory (d is the number of 
> columns/dimensions of the matrix). We often need only the largest k principal 
> components. To make pca really scalable, I suggest an implementation where 
> the memory usage is proportional to the principal components k rather than 
> the full dimensionality d. 
> I suggest adopting the solution described in this paper that is published in 
> SIGMOD 2015 (http://ds.qcri.org/images/profile/tarek_elgamal/sigmod2015.pdf). 
> The paper offers an implementation for Probabilistic PCA (PPCA) which has 
> less memory and time complexity and could potentially scale to tall and fat 
> matrices rather than tall and skinny matrices that is supported by the 
> current PCA impelmentation. 
> Probablistic PCA could be potentially added to the set of algorithms 
> supported by MLlib and it does not necessarily replace the old PCA 
> implementation.
> PPCA implementation is adopted in Matlab's Statistics and Machine Learning 
> Toolbox (http://www.mathworks.com/help/stats/ppca.html)



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org

Reply via email to