Re: Requirements for Spark cluster

2014-07-09 Thread Krishna Sankar
I rsync the spark-1.0.1 directory to all the nodes. Yep, one needs Spark in all the nodes irrespective of Hadoop/YARN. Cheers On Tue, Jul 8, 2014 at 6:24 PM, Robert James wrote: > I have a Spark app which runs well on local master. I'm now ready to > put it on a cluster. What needs to be ins

Re: Requirements for Spark cluster

2014-07-09 Thread Sandy Ryza
Hi Robert, If you're running Spark against YARN, you don't need to install anything Spark-specific on all the nodes. For each application, the client will copy the Spark jar to HDFS where the Spark processes can fetch it. For faster app startup, you can copy the Spark jar to a public location on

Re: Requirements for Spark cluster

2014-07-08 Thread Akhil Das
You can use the spark-ec2/bdutil scripts to set it up on the AWS/GCE cloud quickly. If you want to set it up on your own then these are the things that you will need to do: 1. Make sure you have java (7) installed on all machines. 2. Install and configure spark (add all slave nodes in conf/slaves