Sorry, I can't help with this issue, but if you are interested in a simple
way to launch a Spark cluster on Amazon, Spark is now offered as an
application in Amazon EMR.  With this you can have a full cluster with a
few clicks:

https://aws.amazon.com/blogs/aws/new-apache-spark-on-amazon-emr/

- Arun

On Tue, Jul 7, 2015 at 4:34 PM, Pagliari, Roberto <rpagli...@appcomsci.com>
wrote:

>
>
>
>
> I'm following the tutorial about Apache Spark on EC2. The output is the
> following:
>
>
>
>
>
>     $ ./spark-ec2 -i ../spark.pem -k spark --copy launch spark-training
>
>     Setting up security groups...
>
>     Searching for existing cluster spark-training...
>
>     Latest Spark AMI: ami-19474270
>
>     Launching instances...
>
>     Launched 5 slaves in us-east-1d, regid = r-59a0d4b6
>
>     Launched master in us-east-1d, regid = r-9ba2d674
>
>     Waiting for instances to start up...
>
>     Waiting 120 more seconds...
>
>     Copying SSH key ../spark.pem to master...
>
>     ssh: connect to host ec2-54-152-15-165.compute-1.amazonaws.com port
> 22: Connection refused
>
>     Error connecting to host Command 'ssh -t -o StrictHostKeyChecking=no
> -i ../spark.pem r...@ec2-54-152-15-165.compute-1.amazonaws.com 'mkdir -p
> ~/.ssh'' returned non-zero exit status 255, sleeping 30
>
>     ssh: connect to host ec2-54-152-15-165.compute-1.amazonaws.com port
> 22: Connection refused
>
>     Error connecting to host Command 'ssh -t -o StrictHostKeyChecking=no
> -i ../spark.pem r...@ec2-54-152-15-165.compute-1.amazonaws.com 'mkdir -p
> ~/.ssh'' returned non-zero exit status 255, sleeping 30
>
>     ssh: Could not resolve hostname
> ec2-54-152-15-165.compute-1.amazonaws.com: Name or service not known
>
>     Error connecting to host Command 'ssh -t -o StrictHostKeyChecking=no
> -i ../spark.pem r...@ec2-54-152-15-165.compute-1.amazonaws.com 'mkdir -p
> ~/.ssh'' returned non-zero exit status 255, sleeping 30
>
>     ssh: connect to host ec2-54-152-15-165.compute-1.amazonaws.com port
> 22: Connection refused
>
>    Traceback (most recent call last):
>
>       File "./spark_ec2.py", line 925, in <module>
>
>         main()
>
>       File "./spark_ec2.py", line 766, in main
>
>         setup_cluster(conn, master_nodes, slave_nodes, zoo_nodes, opts,
> True)
>
>       File "./spark_ec2.py", line 406, in setup_cluster
>
>         ssh(master, opts, 'mkdir -p ~/.ssh')
>
>       File "./spark_ec2.py", line 712, in ssh
>
>         raise e
>
>     subprocess.CalledProcessError: Command 'ssh -t -o
> StrictHostKeyChecking=no -i ../spark.pem
> r...@ec2-54-152-15-165.compute-1.amazonaws.com 'mkdir -p ~/.ssh''
> returned non-zero exit status 255
>
>
>
>
>
> However, I can see the six instances created on my EC2 console, and I
> could even get the name of the master. I'm not sure how to fix the ssh
> issue (my region is US EST).
>
>
>

Reply via email to