Error Running mahout-core-0.5-job.jar

2012-03-21 Thread jeanbabyxu
I tried to run mahout in Hadoop using the following command, [jxu13@lppma692 hadoop-0.20.2]$ bin/hadoop jar /opt/mapr/mahout/mahout-0.5/core/target/mahout-core-0.5-job.jar org.apache.mahout.cf.taste.hadoop.ite m.RecommenderJob -Dmapred.input.dir=input/input.txt --Dmapred.output.dir=output --users

Re: Error Running mahout-core-0.5-job.jar

2012-03-21 Thread Sean Owen
It's -Dmapred.output.dir=output not --Dmapred.output.dir=output (one dash), but, that's not even the problem. I don't think you can specify -D options this way, as they are JVM arguments. You need to configure these in Hadoop's config files. This is not specific to Mahout. On Wed, Mar 21, 2012 at

Re: Error Running mahout-core-0.5-job.jar

2012-03-22 Thread tianwild
the correct Dmapred is -Dmapred.dir=output, not --Dmapred.dir=output -- View this message in context: http://lucene.472066.n3.nabble.com/Error-Running-mahout-core-0-5-job-jar-tp3846385p3847789.html Sent from the Mahout User List mailing list archive at Nabble.com.

Re: Error Running mahout-core-0.5-job.jar

2012-03-22 Thread jeanbabyxu
Thanks so much tianwild for pointing out the typo. Now it's running but I got a different error msg: Exception in thread "main" org.apache.hadoop.mapred.FileAlreadyExistsException: Output directory temp/itemIDIndex already exists Any idea how to resolve this issue? Many thanks. -- View this mes

Re: Error Running mahout-core-0.5-job.jar

2012-03-22 Thread Sean Owen
That pretty much means what it says = delete temp. On Thu, Mar 22, 2012 at 6:06 PM, jeanbabyxu wrote: > Thanks so much tianwild for pointing out the typo. Now it's running but I got > a different error msg: > > Exception in thread "main" > org.apache.hadoop.mapred.FileAlreadyExistsException: Outp

Re: Error Running mahout-core-0.5-job.jar

2012-03-22 Thread jeanbabyxu
I was able to manually clear out the output directory by using bin/hadoop dfs -rmr output. But do we have to remove all content in the output directory manually every time we run mahout? -- View this message in context: http://lucene.472066.n3.nabble.com/Error-Running-mahout-core-0-5-job-jar-

Re: Error Running mahout-core-0.5-job.jar

2012-03-22 Thread Sean Owen
Yes. This prevents accidental overwrite, and mimics how Hadoop/HDFS generally act. On Thu, Mar 22, 2012 at 6:58 PM, jeanbabyxu wrote: > I was able to manually clear out the output directory by using > > bin/hadoop dfs -rmr output. > > But do we have to remove all content in the output directory m

Re: Error Running mahout-core-0.5-job.jar

2012-03-22 Thread Paritosh Ranjan
You can also use HadoopUtil.delete(conf, paths) api or use the -ow (override) flag ( if available for that job). On 23-03-2012 00:28, jeanbabyxu wrote: I was able to manually clear out the output directory by using bin/hadoop dfs -rmr output. But do we have to remove all content in the output

Re: Error Running mahout-core-0.5-job.jar

2012-03-27 Thread tianwild
yes, I did this every time, and clear the tmp/ folder also hadoop fs -rmr /user/**/output hadoop fs -rmr /user/**/temp/* -- View this message in context: http://lucene.472066.n3.nabble.com/Error-Running-mahout-core-0-5-job-jar-tp3846385p3860869.html Sent from the Mahout User List mailing list ar

Re: Error Running mahout-core-0.5-job.jar

2012-03-27 Thread Isabel Drost
On 22.03.2012 Paritosh Ranjan wrote: > You can also use HadoopUtil.delete(conf, paths) api or use the -ow > (override) flag ( if available for that job). If that flag isn't available for the job you are looking at, that might be a good chance to submit a bug report and mark it as "suitable for be