Yes.  As far as I can tell, Spark seems to be including Hadoop 1 via
its transitive dependency:
http://mvnrepository.com/artifact/org.apache.spark/spark-core_2.10/1.0.0
- shows a dependency on Hadoop 1.0.4, which I'm perplexed by.

On 6/26/14, Sean Owen <so...@cloudera.com> wrote:
> You seem to have the binary for Hadoop 2, since it was compiled
> expecting that TaskAttemptContext is an interface. So the error
> indicates that Spark is also seeing Hadoop 1 classes somewhere.
>
> On Wed, Jun 25, 2014 at 4:41 PM, Robert James <srobertja...@gmail.com>
> wrote:
>> After upgrading to Spark 1.0.0, I get this error:
>>
>>  ERROR org.apache.spark.executor.ExecutorUncaughtExceptionHandler -
>> Uncaught exception in thread Thread[Executor task launch
>> worker-2,5,main]
>> java.lang.IncompatibleClassChangeError: Found interface
>> org.apache.hadoop.mapreduce.TaskAttemptContext, but class was expected
>>
>> I thought this was caused by a dependency on Hadoop 1.0.4 (even though
>> I downloaded the Spark 1.0.0 for Hadoop 2), but I can't seem to fix
>> it.  Any advice?
>

Reply via email to