Delete the checkpoint directory, you might have modified your driver
program.

Thanks
Best Regards

On Wed, Jun 10, 2015 at 9:44 PM, Ashish Nigam <ashnigamt...@gmail.com>
wrote:

> Hi,
> If checkpoint data is already present in HDFS, driver fails to load as it
> is performing lookup on previous application directory. As that folder
> already exists, it fails to start context.
> Failed job's application id was application_1432284018452_0635 and job
> was performing lookup on application_1432284018452_0633 folder.
>
> Here's snippet of exception stack trace-
>
> 15/06/10 05:28:36 INFO yarn.ApplicationMaster: Final app status: FAILED,
> exitCode: 15, (reason: User class threw exception: Log directory
> hdfs://x.x.x.x:8020/user/myuser/spark/applicationHistory/application_1432284018452_0633
> already exists!)
>
> Exception in thread "Driver" java.io.IOException: Log directory hdfs://
> 172.16.201.171:8020/user/shn/spark/applicationHistory/application_1432284018452_0633
> already exists!
>
>         at
> org.apache.spark.util.FileLogger.createLogDir(FileLogger.scala:129)
>
>         at org.apache.spark.util.FileLogger.start(FileLogger.scala:115)
>
>         at
> org.apache.spark.scheduler.EventLoggingListener.start(EventLoggingListener.scala:74)
>
>         at org.apache.spark.SparkContext.<init>(SparkContext.scala:368)
>
>         at
> org.apache.spark.streaming.StreamingContext.<init>(StreamingContext.scala:118)
>
>         at
> org.apache.spark.streaming.StreamingContext$$anonfun$getOrCreate$1.apply(StreamingContext.scala:561)
>
>         at
> org.apache.spark.streaming.StreamingContext$$anonfun$getOrCreate$1.apply(StreamingContext.scala:561)
>
>         at scala.Option.map(Option.scala:145)
>
>         at
> org.apache.spark.streaming.StreamingContext$.getOrCreate(StreamingContext.scala:561)
>
>         at
>
>
>
> Any idea on how to fix this issue?
>
>
> Thanks
>
> Ashish
>

Reply via email to