Re: spark-rowsimilarity java.lang.OutOfMemoryError: Java heap space

2015-05-18 Thread Xavier Rampino
space but try the above first. > > BTW setting the executor memory twice, is not necessary. > > > On May 13, 2015, at 2:21 AM, Xavier Rampino > wrote: > > Hello, > > I've tried spark-rowsimilarity with out-of-the-box setup (downloaded mahout > distribution and

spark-rowsimilarity java.lang.OutOfMemoryError: Java heap space

2015-05-13 Thread Xavier Rampino
Hello, I've tried spark-rowsimilarity with out-of-the-box setup (downloaded mahout distribution and spark, and set up the PATH), and I stumble upon a Java Heap space error. My input file is ~100MB. It seems the various parameters I tried to give won't change this. I do : ~/mahout-distribution-0.1

Re: java.lang.UnsatisfiedLinkError: no snappyjava in java.library.path

2015-05-07 Thread Xavier Rampino
Late to the party but I had the same problem and solved like that : 1. Download libsnappyjava.jnilib 2. Copy to a directory on the java.library.path, e.g. /usr/lib/java/ 3. Rename to libsnappyjava.dylib On Wed, Jan 28, 2015 at 1:07 AM, Dmitriy Lyubimov wrote: > This looks

Re: Indicator Matrix and Mahout + Solr recommender

2014-06-06 Thread Xavier Rampino
I have a related question about the Indicator Matrix. Is it possible to compute it using either quantitative ratings; or maybe just good ratings taken a single action (user 1 "liked" product 1). I am referring to "Practical Machine Learning Innovations in Recommendation" where you say that "The bes

Re: Question - Mahout Taste - User-Based Recommendations...

2013-01-29 Thread Xavier Rampino
Just to add that you can also use UserNeighborhood.getUserNeighborhood(userId) to find the most similar users to a given one, should you want to. On Tue, Jan 22, 2013 at 9:02 PM, Henning Kuich wrote: > ok, thanks! > > > On Tue, Jan 22, 2013 at 8:59 PM, Sean Owen wrote: > > > That's a question o

Re: Extracting data from websites

2012-07-30 Thread Xavier Rampino
If you want to develop scrapers, I suggest you take a look at jsoup ( http://jsoup.org/), which allows you to parse HTML easily. If you need subsequent classification of the websites, then maybe you'll need Mahout On Mon, Jul 30, 2012 at 2:26 PM, Sean Owen wrote: > Extract as in web crawl? No it