Hi, Thanks for quick input, now I am getting class not found error *Command:*
dse spark-submit --master spark://10.246.43.15:7077 --class HelloWorld --jars ///home/missingmerch/postgresql-9.4-1201.jdbc41.jar ///home/missingmerch/dse.jar ///home/missingmerch/spark-cassandra-connector-java_2.10-1.1.1.jar ///home/missingmerch/etl-0.0.1-SNAPSHOT.jar *Error:* java.lang.ClassNotFoundException: HelloWorld at java.net.URLClassLoader$1.run(URLClassLoader.java:366) at java.net.URLClassLoader$1.run(URLClassLoader.java:355) at java.security.AccessController.doPrivileged(Native Method) at java.net.URLClassLoader.findClass(URLClassLoader.java:354) at java.lang.ClassLoader.loadClass(ClassLoader.java:425) at java.lang.ClassLoader.loadClass(ClassLoader.java:358) at java.lang.Class.forName0(Native Method) at java.lang.Class.forName(Class.java:270) at org.apache.spark.deploy.SparkSubmit$.launch(SparkSubmit.scala:342) at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:75) at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala) Previously I could fix the issue by changing the order of arguments passing in DSE command line interface but now I am not sure why the issue again Please let me know if still I am missing anything in my Command as mentioned above(as insisted I have added dse.jar and spark-cassandra-connector-java_2.10.1.1.1.jar) Thanks for support Satish Chandra On Mon, Aug 10, 2015 at 6:19 PM, Dean Wampler <deanwamp...@gmail.com> wrote: > Add the other Cassandra dependencies (dse.jar, > spark-cassandra-connect-java_2.10) to your --jars argument on the command > line. > > Dean Wampler, Ph.D. > Author: Programming Scala, 2nd Edition > <http://shop.oreilly.com/product/0636920033073.do> (O'Reilly) > Typesafe <http://typesafe.com> > @deanwampler <http://twitter.com/deanwampler> > http://polyglotprogramming.com > > On Mon, Aug 10, 2015 at 7:44 AM, satish chandra j < > jsatishchan...@gmail.com> wrote: > >> HI All, >> Please help me to fix Spark Cassandra Connector issue, find the details >> below >> >> *Command:* >> >> dse spark-submit --master spark://10.246.43.15:7077 --class HelloWorld >> --jars ///home/missingmerch/postgresql-9.4-1201.jdbc41.jar >> ///home/missingmerch/etl-0.0.1-SNAPSHOT.jar >> >> >> *Error:* >> >> >> WARN 2015-08-10 06:33:35 org.apache.spark.util.Utils: Service 'SparkUI' >> could not bind on port 4040. Attempting port 4041. >> >> Exception in thread "main" java.lang.NoSuchMethodError: >> com.datastax.spark.connector.package$.toRDDFunctions(Lorg/apache/spark/rdd/RDD;Lscala/reflect/ClassTag;)Lcom/datastax/spark/connector/RDDFunctions; >> >> at HelloWorld$.main(HelloWorld.scala:29) >> >> at HelloWorld.main(HelloWorld.scala) >> >> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) >> >> at >> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) >> >> at >> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) >> >> at java.lang.reflect.Method.invoke(Method.java:606) >> >> at >> org.apache.spark.deploy.SparkSubmit$.launch(SparkSubmit.scala:358) >> >> at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:75) >> >> at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala) >> >> >> *Code:* >> >> *import* *org.apache*.spark.SparkContext >> >> *import* *org.apache*.spark.SparkContext._ >> >> *import* *org.apache*.spark.SparkConf >> >> *import* *org.apache*.spark.rdd.JdbcRDD >> >> *import* *com.datastax*.spark.connector._ >> >> *import* com.datastax.spark.connector.cql.CassandraConnector >> >> *import* com.datastax.bdp.spark.DseSparkConfHelper._ >> >> *import* java.sql.{Connection, DriverManager, ResultSet, >> PreparedStatement, SQLException, Statement} >> >> *object* HelloWorld { >> >> *def* main(args: Array[String]) { >> >> *def* createSparkContext() = { >> >> *val** myJar = >> *getClass.getProtectionDomain.getCodeSource.getLocation.getPath >> >> >> *val* conf = *new* SparkConf().set( >> "spark.cassandra.connection.host", "10.246.43.15") >> >> .setAppName("First Spark App") >> >> .setMaster("local") >> >> * .s*etJars(Array(myJar)) >> >> .set("cassandra.username", "username") >> >> .set("cassandra.password", "password") >> >> .forDse >> >> *new* SparkContext(conf) >> >> } >> >> >> >> *val* sc = createSparkContext() >> >> *val* user="hkonak0" >> >> *val** pass=*"Winter18" >> >> Class.forName("org.postgresql.Driver").newInstance >> >> *val* url = "jdbc:postgresql://gptester:5432/db_test" >> >> *val* myRDD27 = *new* JdbcRDD( sc, ()=> >> DriverManager.getConnection(url,user,pass),"select * from >> wmax_vmax.arm_typ_txt LIMIT ? OFFSET ?",5,0,1,(r: ResultSet) => >> {(r.getInt("alarm_type_code"),r.getString("language_code"),r.getString( >> "alrm_type_cd_desc"))}) >> >> myRDD27.saveToCassandra("keyspace","arm_typ_txt",SomeColumns( >> "alarm_type_code","language_code","alrm_type_cd_desc")) >> >> println(myRDD27.count()) >> >> println(myRDD27.first) >> >> sc.stop() >> >> sys.exit() >> >> >> >> } >> >> } >> >> >> >> *POM XML:* >> >> >> <dependencies> >> >> <dependency> >> >> <groupId>org.apache.spark</groupId> >> >> <artifactId>spark-core_2.10</artifactId> >> >> <version>1.2.2</version> >> >> </dependency> >> >> <dependency> >> >> <groupId>org.apache.hadoop</groupId> >> >> <artifactId>*hadoop*-client</artifactId> >> >> <version>1.2.1</version> >> >> </dependency> >> >> <dependency> >> >> <groupId>org.scala-*lang*</groupId> >> >> <artifactId>*scala*-library</artifactId> >> >> <version>2.10.5</version> >> >> </dependency> >> >> <dependency> >> >> <groupId>*junit*</groupId> >> >> <artifactId>*junit*</artifactId> >> >> <version>3.8.1</version> >> >> <scope>test</scope> >> >> </dependency> >> >> <dependency> >> >> <groupId>com.datastax.dse</groupId> >> >> <artifactId>*dse*</artifactId> >> >> <version>4.7.2</version> >> >> <scope>system</scope> >> >> <systemPath>C:\workspace\*etl*\*lib*\dse.jar</ >> systemPath> >> >> </dependency> >> >> <dependency> >> >> <groupId>com.datastax.spark</groupId> >> >> <artifactId>spark-*cassandra*-connector-java_2.10</ >> artifactId> >> >> <version>1.1.1</version> >> >> </dependency> >> >> </dependencies> >> >> >> Please let me know if any further details required to analyze the issue >> >> >> Regards, >> >> Satish Chandra >> > >