Sorry, I can't help with this issue, but if you are interested in a simple way to launch a Spark cluster on Amazon, Spark is now offered as an application in Amazon EMR. With this you can have a full cluster with a few clicks:
https://aws.amazon.com/blogs/aws/new-apache-spark-on-amazon-emr/ - Arun On Tue, Jul 7, 2015 at 4:34 PM, Pagliari, Roberto <rpagli...@appcomsci.com> wrote: > > > > > I'm following the tutorial about Apache Spark on EC2. The output is the > following: > > > > > > $ ./spark-ec2 -i ../spark.pem -k spark --copy launch spark-training > > Setting up security groups... > > Searching for existing cluster spark-training... > > Latest Spark AMI: ami-19474270 > > Launching instances... > > Launched 5 slaves in us-east-1d, regid = r-59a0d4b6 > > Launched master in us-east-1d, regid = r-9ba2d674 > > Waiting for instances to start up... > > Waiting 120 more seconds... > > Copying SSH key ../spark.pem to master... > > ssh: connect to host ec2-54-152-15-165.compute-1.amazonaws.com port > 22: Connection refused > > Error connecting to host Command 'ssh -t -o StrictHostKeyChecking=no > -i ../spark.pem r...@ec2-54-152-15-165.compute-1.amazonaws.com 'mkdir -p > ~/.ssh'' returned non-zero exit status 255, sleeping 30 > > ssh: connect to host ec2-54-152-15-165.compute-1.amazonaws.com port > 22: Connection refused > > Error connecting to host Command 'ssh -t -o StrictHostKeyChecking=no > -i ../spark.pem r...@ec2-54-152-15-165.compute-1.amazonaws.com 'mkdir -p > ~/.ssh'' returned non-zero exit status 255, sleeping 30 > > ssh: Could not resolve hostname > ec2-54-152-15-165.compute-1.amazonaws.com: Name or service not known > > Error connecting to host Command 'ssh -t -o StrictHostKeyChecking=no > -i ../spark.pem r...@ec2-54-152-15-165.compute-1.amazonaws.com 'mkdir -p > ~/.ssh'' returned non-zero exit status 255, sleeping 30 > > ssh: connect to host ec2-54-152-15-165.compute-1.amazonaws.com port > 22: Connection refused > > Traceback (most recent call last): > > File "./spark_ec2.py", line 925, in <module> > > main() > > File "./spark_ec2.py", line 766, in main > > setup_cluster(conn, master_nodes, slave_nodes, zoo_nodes, opts, > True) > > File "./spark_ec2.py", line 406, in setup_cluster > > ssh(master, opts, 'mkdir -p ~/.ssh') > > File "./spark_ec2.py", line 712, in ssh > > raise e > > subprocess.CalledProcessError: Command 'ssh -t -o > StrictHostKeyChecking=no -i ../spark.pem > r...@ec2-54-152-15-165.compute-1.amazonaws.com 'mkdir -p ~/.ssh'' > returned non-zero exit status 255 > > > > > > However, I can see the six instances created on my EC2 console, and I > could even get the name of the master. I'm not sure how to fix the ssh > issue (my region is US EST). > > >