My guess, and it's just a guess, is that there is some change between versions which you got bit by as it chsnged the class path.
El lunes, 2 de noviembre de 2015, Babar Tareen <babartar...@gmail.com> escribió: > I am using *'sbt run'* to execute the code. Detailed sbt output is here ( > https://drive.google.com/open?id=0B2dlA_DzEohVakpValRjRS1zVG8). > > I had scala 2.11.7 installed on my machine. But even after uninstalling > it, I am still getting the exception with 2.11.6. > > Changing the scala version to 2.11.7 in build.sbt fixes the exception as > you suggested. I am unclear as to why it works with 2.11.7 and not 2.11.6. > > Thanks, > Babar > > On Mon, Nov 2, 2015 at 2:10 PM Jonathan Coveney <jcove...@gmail.com > <javascript:_e(%7B%7D,'cvml','jcove...@gmail.com');>> wrote: > >> Caused by: java.lang.ClassNotFoundException: scala.Some >> >> indicates that you don't have the scala libs present. How are you >> executing this? My guess is the issue is a conflict between scala 2.11.6 in >> your build and 2.11.7? Not sure...try setting your scala to 2.11.7? >> >> But really, first it'd be good to see what command you're using to invoke >> this. >> >> 2015-11-02 14:48 GMT-05:00 Babar Tareen <babartar...@gmail.com >> <javascript:_e(%7B%7D,'cvml','babartar...@gmail.com');>>: >> >>> Resending, haven't found a workaround. Any help is highly appreciated. >>> >>> ---------- Forwarded message ---------- >>> From: Babar Tareen <babartar...@gmail.com >>> <javascript:_e(%7B%7D,'cvml','babartar...@gmail.com');>> >>> Date: Thu, Oct 22, 2015 at 2:47 PM >>> Subject: Getting ClassNotFoundException: scala.Some on Spark 1.5.x >>> To: user@spark.apache.org >>> <javascript:_e(%7B%7D,'cvml','user@spark.apache.org');> >>> >>> >>> Hi, >>> >>> I am getting following exception when submitting a job to Spark 1.5.x >>> from Scala. The same code works with Spark 1.4.1. Any clues as to what >>> might causing the exception. >>> >>> >>> >>> *Code:App.scala*import org.apache.spark.SparkContext >>> >>> object App { >>> def main(args: Array[String]) = { >>> val l = List(1,2,3,4,5,6,7,8,9,0) >>> val sc = new SparkContext("local[4]", "soark-test") >>> val rdd = sc.parallelize(l) >>> rdd.foreach(println) >>> println(rdd.collect()) >>> } >>> } >>> >>> *build.sbt* >>> lazy val sparkjob = (project in file(".")) >>> .settings( >>> name := "SparkJob", >>> version := "1.0", >>> scalaVersion := "2.11.6", >>> libraryDependencies := libs >>> ) >>> >>> lazy val libs = Seq( >>> "org.apache.spark" %% "spark-core" % "1.5.1" >>> ) >>> >>> >>> *Exception:*15/10/22 14:32:42 INFO DAGScheduler: Job 0 failed: foreach >>> at app.scala:9, took 0.689832 s >>> [error] (run-main-0) org.apache.spark.SparkException: Job aborted due to >>> stage failure: Task 2 in stage 0.0 failed 1 times, most recent failure: >>> Lost task 2.0 in stage 0.0 (TID 2, localhost): java.io.IOException: >>> java.lang.ClassNotFoundException: scala.Some >>> [error] at >>> org.apache.spark.util.Utils$.tryOrIOException(Utils.scala:1163) >>> [error] at >>> org.apache.spark.Accumulable.readObject(Accumulators.scala:151) >>> [error] at sun.reflect.NativeMethodAccessorImpl.invoke0(Native >>> Method) >>> [error] at >>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) >>> [error] at >>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) >>> [error] at java.lang.reflect.Method.invoke(Method.java:497) >>> [error] at >>> java.io.ObjectStreamClass.invokeReadObject(ObjectStreamClass.java:1017) >>> [error] at >>> java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1896) >>> [error] at >>> java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1801) >>> [error] at >>> java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1351) >>> [error] at >>> java.io.ObjectInputStream.skipCustomData(ObjectInputStream.java:1959) >>> [error] at >>> java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1921) >>> [error] at >>> java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1801) >>> [error] at >>> java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1351) >>> [error] at >>> java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1993) >>> [error] at >>> java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1918) >>> [error] at >>> java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1801) >>> [error] at >>> java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1351) >>> [error] at >>> java.io.ObjectInputStream.readObject(ObjectInputStream.java:371) >>> [error] at >>> org.apache.spark.serializer.JavaDeserializationStream.readObject(JavaSerializer.scala:72) >>> [error] at >>> org.apache.spark.serializer.JavaSerializerInstance.deserialize(JavaSerializer.scala:98) >>> [error] at >>> org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:194) >>> [error] at >>> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) >>> [error] at >>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) >>> [error] at java.lang.Thread.run(Thread.java:745) >>> [error] Caused by: java.lang.ClassNotFoundException: scala.Some >>> [error] at java.net.URLClassLoader.findClass(URLClassLoader.java:381) >>> [error] at java.lang.ClassLoader.loadClass(ClassLoader.java:424) >>> [error] at java.lang.ClassLoader.loadClass(ClassLoader.java:357) >>> [error] at java.lang.Class.forName0(Native Method) >>> [error] at java.lang.Class.forName(Class.java:348) >>> [error] at >>> org.apache.spark.serializer.JavaDeserializationStream$$anon$1.resolveClass(JavaSerializer.scala:67) >>> [error] at >>> java.io.ObjectInputStream.readNonProxyDesc(ObjectInputStream.java:1613) >>> [error] at >>> java.io.ObjectInputStream.readClassDesc(ObjectInputStream.java:1518) >>> [error] at >>> java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1774) >>> [error] at >>> java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1351) >>> [error] at >>> java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1993) >>> [error] at >>> java.io.ObjectInputStream.defaultReadObject(ObjectInputStream.java:501) >>> [error] at >>> org.apache.spark.Accumulable$$anonfun$readObject$1.apply$mcV$sp(Accumulators.scala:152) >>> [error] at >>> org.apache.spark.util.Utils$.tryOrIOException(Utils.scala:1160) >>> [error] ... 24 more >>> [error] >>> [error] Driver stacktrace: >>> org.apache.spark.SparkException: Job aborted due to stage failure: Task >>> 2 in stage 0.0 failed 1 times, most recent failure: Lost task 2.0 in stage >>> 0.0 (TID 2, localhost): java.io.IOException: >>> java.lang.ClassNotFoundException: scala.Some >>> at org.apache.spark.util.Utils$.tryOrIOException(Utils.scala:1163) >>> at org.apache.spark.Accumulable.readObject(Accumulators.scala:151) >>> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) >>> at >>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) >>> at >>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) >>> at java.lang.reflect.Method.invoke(Method.java:497) >>> at >>> java.io.ObjectStreamClass.invokeReadObject(ObjectStreamClass.java:1017) >>> at >>> java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1896) >>> at >>> java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1801) >>> at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1351) >>> at >>> java.io.ObjectInputStream.skipCustomData(ObjectInputStream.java:1959) >>> at >>> java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1921) >>> at >>> java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1801) >>> at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1351) >>> at >>> java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1993) >>> at >>> java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1918) >>> at >>> java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1801) >>> at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1351) >>> at java.io.ObjectInputStream.readObject(ObjectInputStream.java:371) >>> at >>> org.apache.spark.serializer.JavaDeserializationStream.readObject(JavaSerializer.scala:72) >>> at >>> org.apache.spark.serializer.JavaSerializerInstance.deserialize(JavaSerializer.scala:98) >>> at >>> org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:194) >>> at >>> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) >>> at >>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) >>> at java.lang.Thread.run(Thread.java:745) >>> >>> Thanks, >>> Babar Tareen >>> >> >>