Hi, No,you don't need to. However,when submitting jobs certain resources will be uploaded to hdfs,which could be a performance issue read the log and you will understand:
15/12/29 11:10:06 INFO Client: Uploading resource file:/data/spark/spark152/lib/spark-assembly-1.5.2-hadoop2.6.0.jar -> hdfs 15/12/29 11:10:08 INFO Client: Uploading resource file:/data/spark/spark152/python/lib/pyspark.zip -> hdfs 15/12/29 11:10:08 INFO Client: Uploading resource file:/data/spark/spark152/python/lib/py4j-0.8.2.1-src.zip -> hdfs 15/12/29 11:10:08 INFO Client: Uploading resource file:/data/tmp/spark-86791975-2cef-4663-aacd-5da95e58cd91/__spark_conf__6261788210225867171.zip -> hdfs 2016-01-19 19:43 GMT+08:00 Sanjeev Verma <sanjeev.verm...@gmail.com>: > Hi > > Do I need to install spark on all the yarn cluster node if I want to > submit the job to yarn client? > is there any way exists in which I can spawn a spark job executors on the > cluster nodes where I have not installed spark. > > Thanks > Sanjeev >