HI, Please let me know if I am missing anything in the below mail, to get the issue fixed
Regards, Satish Chandra On Wed, Aug 12, 2015 at 6:59 PM, satish chandra j <jsatishchan...@gmail.com> wrote: > HI, > > The below mentioned code is working very well fine in Spark Shell but when > the same is placed in Spark Application it is errors as mentioned below: > > *Exception in thread "main" java.lang.NoSuchMethodError: > com.datastax.spark.connector.package$.toRDDFunctions(Lorg/apache/spark/rdd/RDD;Lscala/reflect/ClassTag;)Lcom/datastax/spark/connector/RDDFunctions* > > > *Code:* > > *import* *org.apache*.spark.SparkContext > > *import* *org.apache*.spark.SparkContext._ > > *import* *org.apache*.spark.SparkConf > > *import* *org.apache*.spark.rdd.JdbcRDD > > *import* *com.datastax*.spark.connector._ > > *import* com.datastax.spark.connector.cql.CassandraConnector > > *import* com.datastax.bdp.spark.DseSparkConfHelper._ > > *import* java.sql.{Connection, DriverManager, ResultSet, > PreparedStatement, SQLException, Statement} > > *object* HelloWorld { > > *def* main(args: Array[String]) { > > *def* createSparkContext() = { > > *val* conf = *new* SparkConf().set( > "spark.cassandra.connection.host", "10.246.43.15") > > .setAppName("First Spark App") > > .setMaster("local") > > .set("cassandra.username", "username") > > .set("cassandra.password", "password") > > .forDse > > *new* SparkContext(conf) > > } > > > > *val* sc = createSparkContext() > > *val* user="user" > > *val** pass=*"password" > > Class.forName("org.postgresql.Driver").newInstance > > *val* url = "jdbc:postgresql://gptester:5432/db_test" > > *val* myRDD27 = *new* JdbcRDD( sc, ()=> > DriverManager.getConnection(url,user,pass),"select * from > wmax_vmax.arm_typ_txt LIMIT ? OFFSET ?",5,0,1,(r: ResultSet) => (r.getInt( > "alarm_type_code"),r.getString("language_code"),r.getString( > "alrm_type_cd_desc"))) > > * myRDD27.saveToCassandra(* > *"keyspace","arm_typ_txt",SomeColumns("alarm_type_code","language_code","alrm_type_cd_desc"))* > > println(myRDD27.count()) > > println(myRDD27.first) > > sc.stop() > > sys.exit() > > > > } > > } > > > > *Command: * > dse spark-submit --master spark://10.246.43.15:7077 --class HelloWorld > --jars /home/missingmerch/postgresql-9.4-1201.jdbc41.jar etl-0.0. > 1-SNAPSHOT.jar > > Please let me know if any solutions for this issue > > Regards, > Satish Chandra >