I peeked in the examples job jar and it definitely does have this class,
along with the other dependencies (after my patch). Double-check that you've
done the clean build an "install" again? and maybe even print out MAHOUT_JOB
in the script to double-check what it is using?

On Tue, May 10, 2011 at 12:40 AM, Jake Mannix <[email protected]> wrote:

> wah.  Even trying to do seq2sparse doesn't work for me:
>
> [jake@smf1-ady-15-sr1 mahout-distribution-0.5-SNAPSHOT]$ ./bin/mahout
> seq2sparse -i hdfs://<namenode>/user/jake/text_temp -o
> hdfs://<namenode>/user/jake/text_vectors_temp
> Running on hadoop, using HADOOP_HOME=/usr/lib/hadoop-0.20
> No HADOOP_CONF_DIR set, using /usr/lib/hadoop-0.20/src/conf
> 11/05/09 23:36:01 WARN driver.MahoutDriver: No seq2sparse.props found on
> classpath, will use command-line arguments only
> 11/05/09 23:36:01 INFO vectorizer.SparseVectorsFromSequenceFiles: Maximum
> n-gram size is: 1
> 11/05/09 23:36:01 INFO vectorizer.SparseVectorsFromSequenceFiles: Minimum
> LLR value: 1.0
> 11/05/09 23:36:01 INFO vectorizer.SparseVectorsFromSequenceFiles: Number of
> reduce tasks: 1
> 11/05/09 23:36:04 INFO input.FileInputFormat: Total input paths to process
> :
> 1
> 11/05/09 23:36:10 INFO mapred.JobClient: Running job:
> job_201104300433_126621
> 11/05/09 23:36:12 INFO mapred.JobClient:  map 0% reduce 0%
> 11/05/09 23:36:47 INFO mapred.JobClient: Task Id :
> attempt_201104300433_126621_m_000000_0, Status : FAILED
> 11/05/09 23:37:07 INFO mapred.JobClient: Task Id :
> attempt_201104300433_126621_m_000000_1, Status : FAILED
> Error: java.lang.ClassNotFoundException:
> org.apache.lucene.analysis.Analyzer
>
> ----
>
> Note I'm not specifying any fancy analyzer.  Just trying to run with the
> defaults. :\
>
>  -jake

Reply via email to