thanks!

2015-01-08 12:59 GMT+01:00 Shixiong Zhu <zsxw...@gmail.com>:
> `--jars` accepts a comma-separated list of jars. See the usage about
> `--jars`
>
> --jars JARS Comma-separated list of local jars to include on the driver and
> executor classpaths.
>
>
>
> Best Regards,
>
> Shixiong Zhu
>
> 2015-01-08 19:23 GMT+08:00 Guillermo Ortiz <konstt2...@gmail.com>:
>>
>> I'm trying to execute Spark from a Hadoop Cluster, I have created this
>> script to try it:
>>
>> #!/bin/bash
>>
>> export HADOOP_CONF_DIR=/etc/hadoop/conf
>> SPARK_CLASSPATH=""
>> for lib in `ls /user/local/etc/lib/*.jar`
>> do
>>         SPARK_CLASSPATH=$SPARK_CLASSPATH:$lib
>> done
>> /home/spark-1.1.1-bin-hadoop2.4/bin/spark-submit --name "Streaming"
>> --master yarn-cluster --class com.sparkstreaming.Executor --jars
>> $SPARK_CLASSPATH --executor-memory 10g
>> /user/local/etc/lib/my-spark-streaming-scala.jar
>>
>> When I execute the script I get this error:
>>
>> Spark assembly has been built with Hive, including Datanucleus jars on
>> classpath
>> Exception in thread "main" java.net.URISyntaxException: Expected
>> scheme name at index 0:
>>
>> :/user/local/etc/lib/akka-actor_2.10-2.2.3-shaded-protobuf.jar:/user/local/etc/lib/akka-remote_2.10-..
>> ....
>> ....
>>
>> -maths-1.2.2a.jar:/user/local/etc/lib/xmlenc-0.52.jar:/user/local/etc/lib/zkclient-0.3.jar:/user/local/etc/lib/zookeeper-3.4.5.jar
>>         at java.net.URI$Parser.fail(URI.java:2829)
>>         at java.net.URI$Parser.failExpecting(URI.java:2835)
>>         at java.net.URI$Parser.parse(URI.java:3027)
>>         at java.net.URI.<init>(URI.java:595)
>>         at org.apache.spark.util.Utils$.resolveURI(Utils.scala:1396)
>>         at
>> org.apache.spark.util.Utils$$anonfun$resolveURIs$1.apply(Utils.scala:1419)
>>         at
>> org.apache.spark.util.Utils$$anonfun$resolveURIs$1.apply(Utils.scala:1419)
>>         at
>> scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:244)
>>         at
>> scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:244)
>>         at
>> scala.collection.IndexedSeqOptimized$class.foreach(IndexedSeqOptimized.scala:33)
>>         at
>> scala.collection.mutable.ArrayOps$ofRef.foreach(ArrayOps.scala:108)
>>         at
>> scala.collection.TraversableLike$class.map(TraversableLike.scala:244)
>>         at scala.collection.mutable.ArrayOps$ofRef.map(ArrayOps.scala:108)
>>         at org.apache.spark.util.Utils$.resolveURIs(Utils.scala:1419)
>>         at
>> org.apache.spark.deploy.SparkSubmitArguments.parse$1(SparkSubmitArguments.scala:308)
>>         at
>> org.apache.spark.deploy.SparkSubmitArguments.parseOpts(SparkSubmitArguments.scala:221)
>>         at
>> org.apache.spark.deploy.SparkSubmitArguments.<init>(SparkSubmitArguments.scala:65)
>>         at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:70)
>>         at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
>>
>>
>>
>> Why do I get this error? I have no idea. Any clue?
>>
>> ---------------------------------------------------------------------
>> To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
>> For additional commands, e-mail: user-h...@spark.apache.org
>>
>

---------------------------------------------------------------------
To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
For additional commands, e-mail: user-h...@spark.apache.org

Reply via email to