That was actually an issue with our Mesos. On Wed, Nov 18, 2015 at 5:29 PM, Tathagata Das <t...@databricks.com> wrote:
> If possible, could you give us the root cause and solution for future > readers of this thread. > > On Wed, Nov 18, 2015 at 6:37 AM, swetha kasireddy < > swethakasire...@gmail.com> wrote: > >> It works fine after some changes. >> >> -Thanks, >> Swetha >> >> On Tue, Nov 17, 2015 at 10:22 PM, Tathagata Das <t...@databricks.com> >> wrote: >> >>> Can you verify that the cluster is running the correct version of Spark. >>> 1.5.2. >>> >>> On Tue, Nov 17, 2015 at 7:23 PM, swetha kasireddy < >>> swethakasire...@gmail.com> wrote: >>> >>>> Sorry <scope>compile</scope> makes it work locally. But, the cluster >>>> still seems to have issues with <scope>provided</scope>. Basically it >>>> does not seem to process any records, no data is shown in any of the tabs >>>> of the Streaming UI except the Streaming tab. Executors, Storage, Stages >>>> etc show empty RDDs. >>>> >>>> On Tue, Nov 17, 2015 at 7:19 PM, swetha kasireddy < >>>> swethakasire...@gmail.com> wrote: >>>> >>>>> Hi TD, >>>>> >>>>> Basically, I see two issues. With <scope>provided</scope> the job >>>>> does not start localy. It does start in Cluster but seems no data is >>>>> getting processed. >>>>> >>>>> Thanks, >>>>> Swetha >>>>> >>>>> On Tue, Nov 17, 2015 at 7:04 PM, Tim Barthram <tim.barth...@iag.com.au >>>>> > wrote: >>>>> >>>>>> If you are running a local context, could it be that you should use: >>>>>> >>>>>> >>>>>> >>>>>> <scope>provided</scope> >>>>>> >>>>>> >>>>>> >>>>>> ? >>>>>> >>>>>> >>>>>> >>>>>> Thanks, >>>>>> >>>>>> Tim >>>>>> >>>>>> >>>>>> >>>>>> *From:* swetha kasireddy [mailto:swethakasire...@gmail.com] >>>>>> *Sent:* Wednesday, 18 November 2015 2:01 PM >>>>>> *To:* Tathagata Das >>>>>> *Cc:* user >>>>>> *Subject:* Re: Streaming Job gives error after changing to version >>>>>> 1.5.2 >>>>>> >>>>>> >>>>>> >>>>>> This error I see locally. >>>>>> >>>>>> >>>>>> >>>>>> On Tue, Nov 17, 2015 at 5:44 PM, Tathagata Das <t...@databricks.com> >>>>>> wrote: >>>>>> >>>>>> Are you running 1.5.2-compiled jar on a Spark 1.5.2 cluster? >>>>>> >>>>>> >>>>>> >>>>>> On Tue, Nov 17, 2015 at 5:34 PM, swetha <swethakasire...@gmail.com> >>>>>> wrote: >>>>>> >>>>>> >>>>>> >>>>>> Hi, >>>>>> >>>>>> I see java.lang.NoClassDefFoundError after changing the Streaming job >>>>>> version to 1.5.2. Any idea as to why this is happening? Following are >>>>>> my >>>>>> dependencies and the error that I get. >>>>>> >>>>>> <dependency> >>>>>> <groupId>org.apache.spark</groupId> >>>>>> <artifactId>spark-core_2.10</artifactId> >>>>>> <version>${sparkVersion}</version> >>>>>> <scope>provided</scope> >>>>>> </dependency> >>>>>> >>>>>> >>>>>> <dependency> >>>>>> <groupId>org.apache.spark</groupId> >>>>>> <artifactId>spark-streaming_2.10</artifactId> >>>>>> <version>${sparkVersion}</version> >>>>>> <scope>provided</scope> >>>>>> </dependency> >>>>>> >>>>>> >>>>>> <dependency> >>>>>> <groupId>org.apache.spark</groupId> >>>>>> <artifactId>spark-sql_2.10</artifactId> >>>>>> <version>${sparkVersion}</version> >>>>>> <scope>provided</scope> >>>>>> </dependency> >>>>>> >>>>>> >>>>>> <dependency> >>>>>> <groupId>org.apache.spark</groupId> >>>>>> <artifactId>spark-hive_2.10</artifactId> >>>>>> <version>${sparkVersion}</version> >>>>>> <scope>provided</scope> >>>>>> </dependency> >>>>>> >>>>>> >>>>>> >>>>>> <dependency> >>>>>> <groupId>org.apache.spark</groupId> >>>>>> <artifactId>spark-streaming-kafka_2.10</artifactId> >>>>>> <version>${sparkVersion}</version> >>>>>> </dependency> >>>>>> >>>>>> >>>>>> Exception in thread "main" java.lang.NoClassDefFoundError: >>>>>> org/apache/spark/streaming/StreamingContext >>>>>> at java.lang.Class.getDeclaredMethods0(Native Method) >>>>>> at java.lang.Class.privateGetDeclaredMethods(Class.java:2693) >>>>>> at java.lang.Class.privateGetMethodRecursive(Class.java:3040) >>>>>> at java.lang.Class.getMethod0(Class.java:3010) >>>>>> at java.lang.Class.getMethod(Class.java:1776) >>>>>> at >>>>>> com.intellij.rt.execution.application.AppMain.main(AppMain.java:125) >>>>>> Caused by: java.lang.ClassNotFoundException: >>>>>> org.apache.spark.streaming.StreamingContext >>>>>> at java.net.URLClassLoader$1.run(URLClassLoader.java:372) >>>>>> at java.net.URLClassLoader$1.run(URLClassLoader.java:361) >>>>>> at java.security.AccessController.doPrivileged(Native Method) >>>>>> at java.net.URLClassLoader.findClass(URLClassLoader.java:360) >>>>>> at java.lang.ClassLoader.loadClass(ClassLoader.java:424) >>>>>> at >>>>>> sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:308) >>>>>> at java.lang.ClassLoader.loadClass(ClassLoader.java:357) >>>>>> >>>>>> >>>>>> >>>>>> -- >>>>>> View this message in context: >>>>>> http://apache-spark-user-list.1001560.n3.nabble.com/Streaming-Job-gives-error-after-changing-to-version-1-5-2-tp25406.html >>>>>> Sent from the Apache Spark User List mailing list archive at >>>>>> Nabble.com. >>>>>> >>>>>> --------------------------------------------------------------------- >>>>>> To unsubscribe, e-mail: user-unsubscr...@spark.apache.org >>>>>> For additional commands, e-mail: user-h...@spark.apache.org >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> _____________________________________________________________________ >>>>>> >>>>>> The information transmitted in this message and its attachments (if >>>>>> any) is intended >>>>>> only for the person or entity to which it is addressed. >>>>>> The message may contain confidential and/or privileged material. Any >>>>>> review, >>>>>> retransmission, dissemination or other use of, or taking of any >>>>>> action in reliance >>>>>> upon this information, by persons or entities other than the intended >>>>>> recipient is >>>>>> prohibited. >>>>>> >>>>>> If you have received this in error, please contact the sender and >>>>>> delete this e-mail >>>>>> and associated material from any computer. >>>>>> >>>>>> The intended recipient of this e-mail may only use, reproduce, >>>>>> disclose or distribute >>>>>> the information contained in this e-mail and any attached files, with >>>>>> the permission >>>>>> of the sender. >>>>>> >>>>>> This message has been scanned for viruses. >>>>>> _____________________________________________________________________ >>>>>> >>>>> >>>>> >>>> >>> >> >