it seems to be an old problem : http://mail-archives.apache.org/mod_mbox/spark-user/201311.mbox/%3c7f6aa9e820f55d4a96946a87e086ef4a4bcdf...@eagh-erfpmbx41.erf.thomson.com%3E
https://groups.google.com/forum/#!topic/spark-users/Q66UOeA2u-I Does anyone got the solution ? On Wed, Mar 26, 2014 at 5:50 PM, Yana Kadiyska <yana.kadiy...@gmail.com>wrote: > I might be way off here but are you looking at the logs on the worker > machines? I am running an older version (0.8) and when I look at the > error log for the executor process I see the exact location where the > executor process tries to load the jar from...with a line like this: > > 14/03/26 13:57:11 INFO executor.Executor: Adding > file:/dirs/dirs/spark/work/app-20140326135710-0029/0/./spark-test.jar > to class loader > > You said "The jar file is present in each node", do you see any > information on the executor indicating that it's trying to load the > jar or where it's loading it from? I can't tell for sure by looking at > your logs but they seem to be logs from the master and driver, not > from the executor itself? > > On Wed, Mar 26, 2014 at 11:46 AM, Ognen Duzlevski > <og...@plainvanillagames.com> wrote: > > Have you looked through the logs fully? I have seen this (in my limited > > experience) pop up as a result of previous exceptions/errors, also as a > > result of being unable to serialize objects etc. > > Ognen > > > > > > On 3/26/14, 10:39 AM, Jaonary Rabarisoa wrote: > > > > I notice that I get this error when I'm trying to load an objectFile with > > val viperReloaded = context.objectFile[ReIdDataSetEntry]("data") > > > > > > On Wed, Mar 26, 2014 at 3:58 PM, Jaonary Rabarisoa <jaon...@gmail.com> > > wrote: > >> > >> Here the output that I get : > >> > >> [error] (run-main-0) org.apache.spark.SparkException: Job aborted: Task > >> 1.0:1 failed 4 times (most recent failure: Exception failure in TID 6 on > >> host 172.166.86.36: java.lang.ClassNotFoundException: > >> value.models.ReIdDataSetEntry) > >> org.apache.spark.SparkException: Job aborted: Task 1.0:1 failed 4 times > >> (most recent failure: Exception failure in TID 6 on host 172.166.86.36: > >> java.lang.ClassNotFoundException: value.models.ReIdDataSetEntry) > >> at > >> > org.apache.spark.scheduler.DAGScheduler$$anonfun$org$apache$spark$scheduler$DAGScheduler$$abortStage$1.apply(DAGScheduler.scala:1011) > >> at > >> > org.apache.spark.scheduler.DAGScheduler$$anonfun$org$apache$spark$scheduler$DAGScheduler$$abortStage$1.apply(DAGScheduler.scala:1009) > >> at > >> > scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59) > >> at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:47) > >> at > >> org.apache.spark.scheduler.DAGScheduler.org > $apache$spark$scheduler$DAGScheduler$$abortStage(DAGScheduler.scala:1009) > >> at > >> > org.apache.spark.scheduler.DAGScheduler$$anonfun$processEvent$10.apply(DAGScheduler.scala:596) > >> at > >> > org.apache.spark.scheduler.DAGScheduler$$anonfun$processEvent$10.apply(DAGScheduler.scala:596) > >> at scala.Option.foreach(Option.scala:236) > >> at > >> > org.apache.spark.scheduler.DAGScheduler.processEvent(DAGScheduler.scala:596) > >> at > >> > org.apache.spark.scheduler.DAGScheduler$$anonfun$start$1$$anon$2$$anonfun$receive$1.applyOrElse(DAGScheduler.scala:146) > >> at akka.actor.ActorCell.receiveMessage(ActorCell.scala:498) > >> at akka.actor.ActorCell.invoke(ActorCell.scala:456) > >> at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:237) > >> at akka.dispatch.Mailbox.run(Mailbox.scala:219) > >> at > >> > akka.dispatch.ForkJoinExecutorConfigurator$AkkaForkJoinTask.exec(AbstractDispatcher.scala:386) > >> at scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260) > >> at > >> > scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339) > >> at > >> scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979) > >> at > >> > scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107) > >> > >> Spark says that the jar is added : > >> > >> 14/03/26 15:49:18 INFO SparkContext: Added JAR > >> target/scala-2.10/value-spark_2.10-1.0.jar > >> > >> > >> > >> > >> > >> On Wed, Mar 26, 2014 at 3:34 PM, Ognen Duzlevski > >> <og...@plainvanillagames.com> wrote: > >>> > >>> Have you looked at the individual nodes logs? Can you post a bit more > of > >>> the exception's output? > >>> > >>> > >>> On 3/26/14, 8:42 AM, Jaonary Rabarisoa wrote: > >>>> > >>>> Hi all, > >>>> > >>>> I got java.lang.ClassNotFoundException even with "addJar" called. The > >>>> jar file is present in each node. > >>>> > >>>> I use the version of spark from github master. > >>>> > >>>> Any ideas ? > >>>> > >>>> > >>>> Jaonary > > > > >