Hi, I am new to the world of Mahout and Hadoop though I have worked with Lucene.
I am trying to run the clustering example as specified here : http://cwiki.apache.org/MAHOUT/syntheticcontroldata.html I got the job file for examples from http://mirrors.ibiblio.org/pub/mirrors/maven2/org/apache/mahout/mahout-examples/0.1/ I started Hadoop (in a single node configuration) and tried to run the example with following command. $HADOOP_HOME/bin/hadoop jar $MAHOUT_HOME/examples/target/mahout-examples-0.1.job org.apache.mahout.clustering.syntheticcontrol.kmeans.Job It starts and displays following message. INFO mapred.FileInputFormat: Total input paths to process : 1 INFO mapred.FileInputFormat: Total input paths to process : 1 mapred.JobClient: Running job: job_200904281825_0005 INFO mapred.JobClient: map 0% reduce 0% Then immediately, it throws following exception multiple times and dies. INFO mapred.JobClient: Task Id : attempt_200904281825_0004_m_000001_2, Status : FAILED java.lang.UnsupportedClassVersionError: Bad version number in .class file Initially, I got the version number error at the beginning. I found that JDK version was 1.5. It has been upgraded it to 1.6. Now JAVA_HOME points to /usr/java/jdk1.6.0_13/ and I am using Hadoop 0.18.3. 1. What could possibly be wrong? I checked the Hadoop script. Value of JAVA_HOME is correct (ie 1.6). Is it possible that somehow it is still using 1.5? 2. The last step the clustering tutorial says "Get the data out of HDFS and have a look." Can you please point me to the documentation of Hadoop about how to read this data? Thanks, --shashi
