Hi Satish, The problem is that `--jars` accepts a comma-delimited list of jars! E.g.
spark-submit ... --jars lib1.jar,lib2.jar,lib3.jar main.jar where main.jar is your main application jar (the one that starts a SparkContext), and lib*.jar refer to additional libraries that your main application jar uses. -Andrew 2015-08-13 3:22 GMT-07:00 Javier Domingo Cansino <javier.domi...@fon.com>: > Please notice that 'jars: null' > > I don't know why you put ///..... but I would propose you just put normal > absolute paths. > > dse spark-submit --master spark://10.246.43.15:7077 --class HelloWorld > --jars /home/missingmerch/postgresql-9.4-1201.jdbc41.jar > /home/missingmerch/dse.jar > /home/missingmerch/spark-cassandra-connector-java_2.10-1.1.1.jar > /home/missingmerch/etl-0.0.1-SNAPSHOT.jar > > Hope this is helpful! > > [image: Fon] <http://www.fon.com/>Javier Domingo CansinoResearch & > Development Engineer+34 946545847Skype: javier.domingo.fonAll information > in this email is confidential <http://corp.fon.com/legal/email-disclaimer> > > On Tue, Aug 11, 2015 at 3:42 PM, satish chandra j < > jsatishchan...@gmail.com> wrote: > >> HI, >> >> Please find the log details below: >> >> >> dse spark-submit --verbose --master local --class HelloWorld >> etl-0.0.1-SNAPSHOT.jar --jars >> file:/home/missingmerch/postgresql-9.4-1201.jdbc41.jar >> file:/home/missingmerch/dse.jar >> file:/home/missingmerch/postgresql-9.4-1201.jdbc41.jar >> >> Using properties file: /etc/dse/spark/spark-defaults.conf >> >> Adding default property: >> spark.cassandra.connection.factory=com.datastax.bdp.spark.DseCassandraConnectionFactory >> >> Adding default property: spark.ssl.keyStore=.keystore >> >> Adding default property: spark.ssl.enabled=false >> >> Adding default property: spark.ssl.trustStore=.truststore >> >> Adding default property: >> spark.cassandra.auth.conf.factory=com.datastax.bdp.spark.DseAuthConfFactory >> >> Adding default property: spark.ssl.keyPassword=cassandra >> >> Adding default property: spark.ssl.keyStorePassword=cassandra >> >> Adding default property: spark.ssl.protocol=TLS >> >> Adding default property: spark.ssl.useNodeLocalConf=true >> >> Adding default property: spark.ssl.trustStorePassword=cassandra >> >> Adding default property: >> spark.ssl.enabledAlgorithms=TLS_RSA_WITH_AES_128_CBC_SHA,TLS_RSA_WITH_AES_256_CBC_SHA >> >> Parsed arguments: >> >> master local >> >> deployMode null >> >> executorMemory null >> >> executorCores null >> >> totalExecutorCores null >> >> propertiesFile /etc/dse/spark/spark-defaults.conf >> >> driverMemory 512M >> >> driverCores null >> >> driverExtraClassPath null >> >> driverExtraLibraryPath null >> >> driverExtraJavaOptions -Dcassandra.username=missingmerch >> -Dcassandra.password=STMbrjrlb -XX:MaxPermSize=256M >> >> supervise false >> >> queue null >> >> numExecutors null >> >> files null >> >> pyFiles null >> >> archives null >> >> mainClass HelloWorld >> >> primaryResource file:/home/missingmerch/etl-0.0.1-SNAPSHOT.jar >> >> name HelloWorld >> >> childArgs [--jars >> file:/home/missingmerch/postgresql-9.4-1201.jdbc41.jar >> file:/home/missingmerch/dse.jar >> file:/home/missingmerch/postgresql-9.4-1201.jdbc41.jar] >> >> jars null >> >> verbose true >> >> >> >> Spark properties used, including those specified through >> >> --conf and those from the properties file >> /etc/dse/spark/spark-defaults.conf: >> >> spark.cassandra.connection.factory -> >> com.datastax.bdp.spark.DseCassandraConnectionFactory >> >> spark.ssl.useNodeLocalConf -> true >> >> spark.ssl.enabled -> false >> >> spark.executor.extraJavaOptions -> -XX:MaxPermSize=256M >> >> spark.ssl.keyStore -> .keystore >> >> spark.ssl.trustStore -> .truststore >> >> spark.ssl.trustStorePassword -> cassandra >> >> spark.ssl.enabledAlgorithms -> >> TLS_RSA_WITH_AES_128_CBC_SHA,TLS_RSA_WITH_AES_256_CBC_SHA >> >> spark.cassandra.auth.conf.factory -> >> com.datastax.bdp.spark.DseAuthConfFactory >> >> spark.ssl.protocol -> TLS >> >> spark.ssl.keyPassword -> cassandra >> >> spark.ssl.keyStorePassword -> cassandra >> >> >> >> >> >> Main class: >> >> HelloWorld >> >> Arguments: >> >> --jars >> >> file:/home/missingmerch/postgresql-9.4-1201.jdbc41.jar >> >> file:/home/missingmerch/dse.jar >> >> file:/home/missingmerch/postgresql-9.4-1201.jdbc41.jar >> >> System properties: >> >> spark.cassandra.connection.factory -> >> com.datastax.bdp.spark.DseCassandraConnectionFactory >> >> spark.driver.memory -> 512M >> >> spark.ssl.useNodeLocalConf -> true >> >> spark.ssl.enabled -> false >> >> SPARK_SUBMIT -> true >> >> spark.executor.extraJavaOptions -> -XX:MaxPermSize=256M >> >> spark.app.name -> HelloWorld >> >> spark.ssl.enabledAlgorithms -> >> TLS_RSA_WITH_AES_128_CBC_SHA,TLS_RSA_WITH_AES_256_CBC_SHA >> >> spark.ssl.trustStorePassword -> cassandra >> >> spark.driver.extraJavaOptions -> -Dcassandra.username=missingmerch >> -Dcassandra.password=STMbrjrlb -XX:MaxPermSize=256M >> >> spark.ssl.keyStore -> .keystore >> >> spark.ssl.trustStore -> .truststore >> >> spark.jars -> file:/home/missingmerch/etl-0.0.1-SNAPSHOT.jar >> >> spark.cassandra.auth.conf.factory -> >> com.datastax.bdp.spark.DseAuthConfFactory >> >> spark.master -> local >> >> spark.ssl.protocol -> TLS >> >> spark.ssl.keyPassword -> cassandra >> >> spark.ssl.keyStorePassword -> cassandra >> >> Classpath elements: >> >> file:/home/missingmerch/etl-0.0.1-SNAPSHOT.jar >> >> >> >> >> >> WARN 2015-08-11 08:23:25 org.apache.spark.util.Utils: Service 'SparkUI' >> could not bind on port 4040. Attempting port 4041. >> >> Exception in thread "main" java.lang.ClassNotFoundException: >> org.postgresql.Driver >> >> at java.net.URLClassLoader$1.run(URLClassLoader.java:366) >> >> at java.net.URLClassLoader$1.run(URLClassLoader.java:355) >> >> at java.security.AccessController.doPrivileged(Native Method) >> >> at java.net.URLClassLoader.findClass(URLClassLoader.java:354) >> >> at java.lang.ClassLoader.loadClass(ClassLoader.java:425) >> >> at java.lang.ClassLoader.loadClass(ClassLoader.java:358) >> >> at java.lang.Class.forName0(Native Method) >> >> at java.lang.Class.forName(Class.java:190) >> >> at HelloWorld$.main(HelloWorld.scala:26) >> >> at HelloWorld.main(HelloWorld.scala) >> >> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) >> >> at >> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) >> >> at >> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) >> >> at java.lang.reflect.Method.invoke(Method.java:606) >> >> at >> org.apache.spark.deploy.SparkSubmit$.launch(SparkSubmit.scala:358) >> >> at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:75) >> >> at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala) >> >> >> Regards, >> >> Satish Chandra J >> >> On Tue, Aug 11, 2015 at 6:15 PM, Javier Domingo Cansino < >> javier.domi...@fon.com> wrote: >> >>> use --verbose, it might give you some insights on what0s happening,.... >>> >>> [image: Fon] <http://www.fon.com/>Javier Domingo CansinoResearch & >>> Development Engineer+34 946545847Skype: javier.domingo.fonAll >>> information in this email is confidential >>> <http://corp.fon.com/legal/email-disclaimer> >>> >>> On Tue, Aug 11, 2015 at 2:44 PM, satish chandra j < >>> jsatishchan...@gmail.com> wrote: >>> >>>> HI , >>>> I have used --jars option as well, please find the command below >>>> >>>> *Command:* >>>> >>>> dse spark-submit --master spark://10.246.43.15:7077 --class HelloWorld >>>> *--jars* ///home/missingmerch/postgresql-9.4-1201.jdbc41.jar >>>> ///home/missingmerch/dse.jar >>>> ///home/missingmerch/spark-cassandra-connector-java_2.10-1.1.1.jar >>>> ///home/missingmerch/etl-0.0.1-SNAPSHOT.jar >>>> >>>> Regards, >>>> Satish >>>> >>>> On Tue, Aug 11, 2015 at 4:08 PM, Javier Domingo Cansino < >>>> javier.domi...@fon.com> wrote: >>>> >>>>> I have no real idea (not java user), but have you tried with the >>>>> --jars option? >>>>> >>>>> >>>>> http://spark.apache.org/docs/latest/submitting-applications.html#advanced-dependency-management >>>>> >>>>> AFAIK, you are currently submitting the jar names as arguments to the >>>>> called Class instead of the jars themselves >>>>> >>>>> [image: Fon] <http://www.fon.com/>Javier Domingo CansinoResearch & >>>>> Development Engineer+34 946545847Skype: javier.domingo.fonAll >>>>> information in this email is confidential >>>>> <http://corp.fon.com/legal/email-disclaimer> >>>>> >>>>> On Tue, Aug 11, 2015 at 12:29 PM, satish chandra j < >>>>> jsatishchan...@gmail.com> wrote: >>>>> >>>>>> >>>>>> *HI,* >>>>>> >>>>>> Please let me know if i am missing anything in the command below >>>>>> >>>>>> *Command:* >>>>>> >>>>>> dse spark-submit --master spark://10.246.43.15:7077 --class >>>>>> HelloWorld --jars ///home/missingmerch/postgresql-9.4-1201.jdbc41.jar >>>>>> ///home/missingmerch/dse.jar >>>>>> ///home/missingmerch/spark-cassandra-connector-java_2.10-1.1.1.jar >>>>>> ///home/missingmerch/etl-0.0.1-SNAPSHOT.jar >>>>>> >>>>>> >>>>>> *Error:* >>>>>> >>>>>> *java.lang.ClassNotFoundException: HelloWorld* >>>>>> >>>>>> at java.net.URLClassLoader$1.run(URLClassLoader.java:366) >>>>>> >>>>>> at java.net.URLClassLoader$1.run(URLClassLoader.java:355) >>>>>> >>>>>> at java.security.AccessController.doPrivileged(Native Method) >>>>>> >>>>>> at java.net.URLClassLoader.findClass(URLClassLoader.java:354) >>>>>> >>>>>> at java.lang.ClassLoader.loadClass(ClassLoader.java:425) >>>>>> >>>>>> at java.lang.ClassLoader.loadClass(ClassLoader.java:358) >>>>>> >>>>>> at java.lang.Class.forName0(Native Method) >>>>>> >>>>>> at java.lang.Class.forName(Class.java:270) >>>>>> >>>>>> at >>>>>> org.apache.spark.deploy.SparkSubmit$.launch(SparkSubmit.scala:342) >>>>>> >>>>>> at >>>>>> org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:75) >>>>>> >>>>>> at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala) >>>>>> >>>>>> >>>>>> I understand the way I am giving multiple jar file paths in the >>>>>> command is an issue, please provide an appropriate format for providing >>>>>> multiple jars in the command >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> Thanks for support >>>>>> >>>>>> >>>>>> Satish Chandra >>>>>> >>>>> >>>>> >>>> >>> >> >