Yes. As far as I can tell, Spark seems to be including Hadoop 1 via its transitive dependency: http://mvnrepository.com/artifact/org.apache.spark/spark-core_2.10/1.0.0 - shows a dependency on Hadoop 1.0.4, which I'm perplexed by.
On 6/26/14, Sean Owen <so...@cloudera.com> wrote: > You seem to have the binary for Hadoop 2, since it was compiled > expecting that TaskAttemptContext is an interface. So the error > indicates that Spark is also seeing Hadoop 1 classes somewhere. > > On Wed, Jun 25, 2014 at 4:41 PM, Robert James <srobertja...@gmail.com> > wrote: >> After upgrading to Spark 1.0.0, I get this error: >> >> ERROR org.apache.spark.executor.ExecutorUncaughtExceptionHandler - >> Uncaught exception in thread Thread[Executor task launch >> worker-2,5,main] >> java.lang.IncompatibleClassChangeError: Found interface >> org.apache.hadoop.mapreduce.TaskAttemptContext, but class was expected >> >> I thought this was caused by a dependency on Hadoop 1.0.4 (even though >> I downloaded the Spark 1.0.0 for Hadoop 2), but I can't seem to fix >> it. Any advice? >