`--jars` accepts a comma-separated list of jars. See the usage about `--jars`
--jars JARS Comma-separated list of local jars to include on the driver and executor classpaths. Best Regards, Shixiong Zhu 2015-01-08 19:23 GMT+08:00 Guillermo Ortiz <konstt2...@gmail.com>: > I'm trying to execute Spark from a Hadoop Cluster, I have created this > script to try it: > > #!/bin/bash > > export HADOOP_CONF_DIR=/etc/hadoop/conf > SPARK_CLASSPATH="" > for lib in `ls /user/local/etc/lib/*.jar` > do > SPARK_CLASSPATH=$SPARK_CLASSPATH:$lib > done > /home/spark-1.1.1-bin-hadoop2.4/bin/spark-submit --name "Streaming" > --master yarn-cluster --class com.sparkstreaming.Executor --jars > $SPARK_CLASSPATH --executor-memory 10g > /user/local/etc/lib/my-spark-streaming-scala.jar > > When I execute the script I get this error: > > Spark assembly has been built with Hive, including Datanucleus jars on > classpath > Exception in thread "main" java.net.URISyntaxException: Expected > scheme name at index 0: > > :/user/local/etc/lib/akka-actor_2.10-2.2.3-shaded-protobuf.jar:/user/local/etc/lib/akka-remote_2.10-.. > .... > .... > > -maths-1.2.2a.jar:/user/local/etc/lib/xmlenc-0.52.jar:/user/local/etc/lib/zkclient-0.3.jar:/user/local/etc/lib/zookeeper-3.4.5.jar > at java.net.URI$Parser.fail(URI.java:2829) > at java.net.URI$Parser.failExpecting(URI.java:2835) > at java.net.URI$Parser.parse(URI.java:3027) > at java.net.URI.<init>(URI.java:595) > at org.apache.spark.util.Utils$.resolveURI(Utils.scala:1396) > at > org.apache.spark.util.Utils$$anonfun$resolveURIs$1.apply(Utils.scala:1419) > at > org.apache.spark.util.Utils$$anonfun$resolveURIs$1.apply(Utils.scala:1419) > at > scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:244) > at > scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:244) > at > scala.collection.IndexedSeqOptimized$class.foreach(IndexedSeqOptimized.scala:33) > at > scala.collection.mutable.ArrayOps$ofRef.foreach(ArrayOps.scala:108) > at > scala.collection.TraversableLike$class.map(TraversableLike.scala:244) > at scala.collection.mutable.ArrayOps$ofRef.map(ArrayOps.scala:108) > at org.apache.spark.util.Utils$.resolveURIs(Utils.scala:1419) > at > org.apache.spark.deploy.SparkSubmitArguments.parse$1(SparkSubmitArguments.scala:308) > at > org.apache.spark.deploy.SparkSubmitArguments.parseOpts(SparkSubmitArguments.scala:221) > at > org.apache.spark.deploy.SparkSubmitArguments.<init>(SparkSubmitArguments.scala:65) > at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:70) > at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala) > > > > Why do I get this error? I have no idea. Any clue? > > --------------------------------------------------------------------- > To unsubscribe, e-mail: user-unsubscr...@spark.apache.org > For additional commands, e-mail: user-h...@spark.apache.org > >