Re: Launching Spark cluster on EC2 with Ubuntu AMI

2015-02-23 Thread Nicholas Chammas
I know that Spark EC2 scripts are not guaranteed to work with custom AMIs
but still, it should work…

Nope, it shouldn’t, unfortunately. The Spark base AMIs are custom-built for
spark-ec2. No other AMI will work unless it was built with that goal in
mind. Using a random AMI from the Amazon marketplace is unlikely to work
because there are several tools and packages (e.g. like git) that need to
be on the AMI.

Furthermore, the spark-ec2 scripts all assume a yum-based Linux
distribution, so you won’t be able to use Ubuntu (and apt-get-based distro)
without some significant changes to the shell scripts used to build the AMI.

There is some work ongoing as part of SPARK-3821
<https://issues.apache.org/jira/browse/SPARK-3821> to make it easier to
generate AMIs that work with spark-ec2.

Nick
​

On Sun Feb 22 2015 at 7:42:52 PM Ted Yu  wrote:

> bq. bash: git: command not found
>
> Looks like the AMI doesn't have git pre-installed.
>
> Cheers
>
> On Sun, Feb 22, 2015 at 4:29 PM, olegshirokikh  wrote:
>
>> I'm trying to launch Spark cluster on AWS EC2 with custom AMI (Ubuntu)
>> using
>> the following:
>>
>> ./ec2/spark-ec2 --key-pair=*** --identity-file='/home/***.pem'
>> --region=us-west-2 --zone=us-west-2b --spark-version=1.2.1 --slaves=2
>> --instance-type=t2.micro --ami=ami-29ebb519 --user=ubuntu launch
>> spark-ubuntu-cluster
>>
>> Everything starts OK and instances are launched:
>>
>> Found 1 master(s), 2 slaves
>> Waiting for all instances in cluster to enter 'ssh-ready' state.
>> Generating cluster's SSH key on master.
>>
>> But then I'm getting the following SSH errors until it stops trying and
>> quits:
>>
>> bash: git: command not found
>> Connection to ***.us-west-2.compute.amazonaws.com closed.
>> Error executing remote command, retrying after 30 seconds: Command
>> '['ssh',
>> '-o', 'StrictHostKeyChecking=no', '-i', '/home/***t.pem', '-o',
>> 'UserKnownHostsFile=/dev/null', '-t', '-t',
>> u'ubuntu@***.us-west-2.compute.amazonaws.com', 'rm -rf spark-ec2 && git
>> clone https://github.com/mesos/spark-ec2.git -b v4']' returned non-zero
>> exit
>> status 127
>>
>> I know that Spark EC2 scripts are not guaranteed to work with custom AMIs
>> but still, it should work... Any advice would be greatly appreciated!
>>
>>
>>
>>
>> --
>> View this message in context:
>> http://apache-spark-user-list.1001560.n3.nabble.com/Launching-Spark-cluster-on-EC2-with-Ubuntu-AMI-tp21757.html
>> Sent from the Apache Spark User List mailing list archive at Nabble.com.
>>
>> -
>> To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
>> For additional commands, e-mail: user-h...@spark.apache.org
>>
>>
>


Re: Launching Spark cluster on EC2 with Ubuntu AMI

2015-02-22 Thread Ted Yu
bq. bash: git: command not found

Looks like the AMI doesn't have git pre-installed.

Cheers

On Sun, Feb 22, 2015 at 4:29 PM, olegshirokikh  wrote:

> I'm trying to launch Spark cluster on AWS EC2 with custom AMI (Ubuntu)
> using
> the following:
>
> ./ec2/spark-ec2 --key-pair=*** --identity-file='/home/***.pem'
> --region=us-west-2 --zone=us-west-2b --spark-version=1.2.1 --slaves=2
> --instance-type=t2.micro --ami=ami-29ebb519 --user=ubuntu launch
> spark-ubuntu-cluster
>
> Everything starts OK and instances are launched:
>
> Found 1 master(s), 2 slaves
> Waiting for all instances in cluster to enter 'ssh-ready' state.
> Generating cluster's SSH key on master.
>
> But then I'm getting the following SSH errors until it stops trying and
> quits:
>
> bash: git: command not found
> Connection to ***.us-west-2.compute.amazonaws.com closed.
> Error executing remote command, retrying after 30 seconds: Command '['ssh',
> '-o', 'StrictHostKeyChecking=no', '-i', '/home/***t.pem', '-o',
> 'UserKnownHostsFile=/dev/null', '-t', '-t',
> u'ubuntu@***.us-west-2.compute.amazonaws.com', 'rm -rf spark-ec2 && git
> clone https://github.com/mesos/spark-ec2.git -b v4']' returned non-zero
> exit
> status 127
>
> I know that Spark EC2 scripts are not guaranteed to work with custom AMIs
> but still, it should work... Any advice would be greatly appreciated!
>
>
>
>
> --
> View this message in context:
> http://apache-spark-user-list.1001560.n3.nabble.com/Launching-Spark-cluster-on-EC2-with-Ubuntu-AMI-tp21757.html
> Sent from the Apache Spark User List mailing list archive at Nabble.com.
>
> -
> To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
> For additional commands, e-mail: user-h...@spark.apache.org
>
>


Launching Spark cluster on EC2 with Ubuntu AMI

2015-02-22 Thread olegshirokikh
I'm trying to launch Spark cluster on AWS EC2 with custom AMI (Ubuntu) using
the following:

./ec2/spark-ec2 --key-pair=*** --identity-file='/home/***.pem'
--region=us-west-2 --zone=us-west-2b --spark-version=1.2.1 --slaves=2
--instance-type=t2.micro --ami=ami-29ebb519 --user=ubuntu launch
spark-ubuntu-cluster

Everything starts OK and instances are launched:

Found 1 master(s), 2 slaves
Waiting for all instances in cluster to enter 'ssh-ready' state.
Generating cluster's SSH key on master.

But then I'm getting the following SSH errors until it stops trying and
quits:

bash: git: command not found
Connection to ***.us-west-2.compute.amazonaws.com closed.
Error executing remote command, retrying after 30 seconds: Command '['ssh',
'-o', 'StrictHostKeyChecking=no', '-i', '/home/***t.pem', '-o',
'UserKnownHostsFile=/dev/null', '-t', '-t',
u'ubuntu@***.us-west-2.compute.amazonaws.com', 'rm -rf spark-ec2 && git
clone https://github.com/mesos/spark-ec2.git -b v4']' returned non-zero exit
status 127

I know that Spark EC2 scripts are not guaranteed to work with custom AMIs
but still, it should work... Any advice would be greatly appreciated!




--
View this message in context: 
http://apache-spark-user-list.1001560.n3.nabble.com/Launching-Spark-cluster-on-EC2-with-Ubuntu-AMI-tp21757.html
Sent from the Apache Spark User List mailing list archive at Nabble.com.

-
To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
For additional commands, e-mail: user-h...@spark.apache.org