Hmm I just ran into this, thanks for the research.
This may cause problems on cluster machines unless it is Mac specific so
putting into /usr/lib/java may need to be on all nodes. Not sure that is the
best solution. Let me know if you run into this on a ’nix type cluster.
On Mar 19, 2015, at
OK, fixed the snappy issue (that happens on Mac/jre 1.7) by downloading
https://wso2.org/jira/secure/attachment/32013/libsnappyjava.jnilib and placing
the file in /usr/lib/java/
Now, when I run ./bin/mahout spark-itemsimilarity -i demoItems.csv -o output4
-fc 1 -ic 2 --filter1 purchase
Looks like you don’t have the native snappy code installed correctly. That’s a
Hadoop thing I think, for fast compressed serialization methinks.
On Mar 18, 2015, at 4:08 PM, Jeff Isenhart jeffi...@yahoo.com.INVALID wrote:
Thanks for the input Pat. I ran the following command
./bin/mahout
Thanks for the input Pat. I ran the following command
./bin/mahout spark-itemsimilarity -i demoItems.csv -o output4 -fc 1 -ic 2
--filter1 purchase --filter2 view
on data
u1,purchase,iphoneu1,purchase,ipadu2,purchase,nexus
and now seeing this error
java.lang.reflect.InvocationTargetException at
spark-itemsimilarity takes tuples
user-id,item-id
You are looking at the collected input as a matrix. it would be collected from
something of the form:
u1,item1
u1,item10
u1,item500
u2,item2
u2,item500
...
On Mar 11, 2015, at 8:24 PM, Jeff Isenhart jeffi...@yahoo.com.INVALID wrote:
I am
There are many ways to structure the input. The spark-itemsimilarity driver can
take only two actions, though the internal code, if you want to use it as a
library, will take any number. The CLI driver can optionally take input of the
for you mention but will extract a primary and single
preHmmm, then what about the How to Use Multiple Actions section that states
For a mixed action log of the form:u1,purchase,iphone
u1,purchase,ipad
u2,purchase,nexus/pre
On Thursday, March 12, 2015 9:39 AM, Pat Ferrel p...@occamsmachete.com
wrote:
spark-itemsimilarity takes tuples
I am trying to run the example found here:
http://mahout.apache.org/users/recommender/intro-cooccurrence-spark.html
The data (demoItems.csv added to hdfs) is just copied from the example:
u1,purchase,iphoneu1,purchase,ipadu2,purchase,nexus..
But when I run
mahout spark-itemsimilarity -i