http://ocw.mit.edu/courses/mathematics/18-06-linear-algebra-spring-2010/video-lectures/
On 29.08.2011 09:15, myn wrote:
i want to study Singular Value Decomposition algorithms;
I also have a book called mahout in action,but i can`t found sth about this
algorithm;
is there someplace introduce
Command line arguments are found here:
https://cwiki.apache.org/MAHOUT/dimensional-reduction.html
I wrote a quick tutorial on how to prepare sparse matrices as input to
Mahout SVD here:
http://bickson.blogspot.com/2011/02/mahout-svd-matrix-factorization.html
Let me know if you have further
thanks
But could you send the content
ofhttp://bickson.blogspot.com/2011/02/mahout-svd-matrix-factorization.html to
me ?
I can`t open it in china .
在 2011-08-29 15:29:46,Sebastian Schelter s...@apache.org 写道:
thanks
But could you send the content
ofhttp://bickson.blogspot.com/2011/02/mahout-svd-matrix-factorization.html to
me ?
I can`t open it in china .
At 2011-08-29 15:29:40,Danny Bickson danny.bick...@gmail.com wrote:
Command line arguments are found here:
Mahout - SVD matrix factorization - formatting input matrix
Converting Input Format into Mahout's SVD Distributed Matrix Factorization
Solver
Purpose
The code below, converts a matrix from csv format:
from row,to col,value\n
Into Mahout's SVD solver format.
For example,
The 3x3 matrix:
0
There are a few problems with integrating with Lucene and Solr:
1) There are different Lucene index file formats. The Mahout tools have a
version of Lucene bound to them, and the workflows assume that generated
index files will be thrown away. Different Solr versions use different
Lucene index
'R' also has an svd implementation, directly in the base package.
There are a few answers to your question:
1) What is SVD? The video lecture above will help. Also, searching for
'singular value decomposition' on Baidu finds a lot of basic explanations.
2) Why do you want it? It creates in on
2011/8/29 myn m...@163.com:
thanks
But could you send the content
ofhttp://bickson.blogspot.com/2011/02/mahout-svd-matrix-factorization.html to
me ?
(You asked the same thing twice with only 6 minutes between)
Try this:
why not change the clusterID from int to long
I have a data about 30 billion rows,when i used createCanopyFromVectors in
meanshift.
the clusterid,is not big enough.
second ,in MeanShiftCanopyCreatorMapper class,
nextCanopyId = ((1 31) / 5) * (Integer.parseInt(parts[4])%5); in
thanks a lot ,that is a quit good example for my study.
At 2011-08-29 15:50:36,Danny Bickson danny.bick...@gmail.com wrote:
Mahout - SVD matrix factorization - formatting input matrix
Converting Input Format into Mahout's SVD Distributed Matrix Factorization
Solver
Purpose
The code below,
the best way is to read the sorce code ;
@_@
At 2011-08-29 16:02:57,Lance Norskog goks...@gmail.com wrote:
'R' also has an svd implementation, directly in the base package.
There are a few answers to your question:
1) What is SVD? The video lecture above will help. Also, searching for
2011/8/29 myn m...@163.com:
the best way is to read the sorce code ;
Talking of view source, has anyone taken a look at these (public
domain) Javascript demos:
http://users.telenet.be/paul.larmuseau/SVD.htm /
http://www.stasegem.be/shop2/SVD.htm
or http://metamerist.com/excanvas/example23a.htm
Funny somebody mentioned the MIT Strang lectures from 99. I just spent the
weekend watching the first 13. In case anybody doesn't like being
restricted to watching them in their browser and doesn't have easy access to
download them with iTunesU, you can download them directly from blip (where
Out of Solr, that is.
On Mon, Aug 29, 2011 at 12:55 AM, Lance Norskog goks...@gmail.com wrote:
You are best off pulling what you want to analyze out of Mahout into
sequence files of some sort. This would require your own program. Failing
that, recent Solr has a CSV output that you could
Hello,
I was playing with Mahout and found that the FileDataModel accepts data in
the format
userId,itemId,pref(long,long,Double).
The data that i want to experiment with is of the format
String,long,double
What is the best/easiest method to work with this dataset on Mahout,
Really, the best thing is to use numeric IDs. Hash the string or otherwise
turn them into numbers first.
if you really need to work with Strings, see the IDMigrator class which
provides a little automatic help in doing so.
On Mon, Aug 29, 2011 at 3:04 PM, Amit Mahale mahale.a...@gmail.com wrote:
My sample code that I wrote for a magazin article that will shortly be
published might help you with that issue.
The essence is that you need to preprocess your data into two files. One
holds all preferences using longs only, the other one has the original
strings. Be aware that you need to
Friday I finally got around to reading Ted's paper accurate methods for
statistics of surprise and coincidence for a better understanding of how to
apply log likelihood.
Can somebody validate if I'm understanding/applying the idea correctly in
this case?
If we have a item/feature matrix
Jeff,
I think that this is a much simpler exposition:
http://tdunning.blogspot.com/2008/03/surprise-and-coincidence.html
It makes the connection with entropy clear and allows a very simple
implementation for more than 2x2 situations.
More comments in-line:
On Mon, Aug 29, 2011 at 1:34 PM, Jeff
I've done some work (i.e. prototyping) on this at:
https://github.com/gsingers/ApacheCon2010 and discussed some at
http://www.lucidimagination.com/blog/2010/03/16/integrating-apache-mahout-with-apache-lucene-and-solr-part-i-of-3/
(yes, I know, I need to finish part II and III)
The gist of it
20 matches
Mail list logo