Re: Error while launching ec2 spark cluster with HVM (r3.large)

2014-05-23 Thread Mayur Rustagi
I am not sure if EC2 script was updated for R3, R3 doesnt provide formatted
instance store also requires newer version of AMI for the same.

Regards
Mayur

Mayur Rustagi
Ph: +1 (760) 203 3257
http://www.sigmoidanalytics.com
@mayur_rustagi <https://twitter.com/mayur_rustagi>



On Fri, May 23, 2014 at 4:32 AM, Xiangrui Meng  wrote:

> Was the error message the same as you posted when you used `root` as
> the user id? Could you try this:
>
> 1) Do not specify user id. (Default would be `root`.)
> 2) If it fails in the middle, try `spark-ec2  --resume launch
> ` to continue launching the cluster.
>
> Best,
> Xiangrui
>
> On Thu, May 22, 2014 at 12:44 PM, adparker  wrote:
> > I had this problem too and fixed it by setting the wait time-out to a
> larger
> > value: --wait
> >
> > For example, in "stand alone" mode with default values, a time out of 480
> > seconds worked for me:
> >
> > $ cd spark-0.9.1/ec2
> > $ ./spark-ec2 --key-pair= --identity-file=
> --instance-type=r3.large
> > --wait=480  launch 
> >
> >
> >
> >
> > --
> > View this message in context:
> http://apache-spark-user-list.1001560.n3.nabble.com/Error-while-launching-ec2-spark-cluster-with-HVM-r3-large-tp5862p6276.html
> > Sent from the Apache Spark User List mailing list archive at Nabble.com.
>


Re: Error while launching ec2 spark cluster with HVM (r3.large)

2014-05-22 Thread Xiangrui Meng
Was the error message the same as you posted when you used `root` as
the user id? Could you try this:

1) Do not specify user id. (Default would be `root`.)
2) If it fails in the middle, try `spark-ec2  --resume launch
` to continue launching the cluster.

Best,
Xiangrui

On Thu, May 22, 2014 at 12:44 PM, adparker  wrote:
> I had this problem too and fixed it by setting the wait time-out to a larger
> value: --wait
>
> For example, in "stand alone" mode with default values, a time out of 480
> seconds worked for me:
>
> $ cd spark-0.9.1/ec2
> $ ./spark-ec2 --key-pair= --identity-file= --instance-type=r3.large
> --wait=480  launch 
>
>
>
>
> --
> View this message in context: 
> http://apache-spark-user-list.1001560.n3.nabble.com/Error-while-launching-ec2-spark-cluster-with-HVM-r3-large-tp5862p6276.html
> Sent from the Apache Spark User List mailing list archive at Nabble.com.


Re: Error while launching ec2 spark cluster with HVM (r3.large)

2014-05-22 Thread adparker
I had this problem too and fixed it by setting the wait time-out to a larger
value: --wait

For example, in "stand alone" mode with default values, a time out of 480
seconds worked for me:

$ cd spark-0.9.1/ec2
$ ./spark-ec2 --key-pair= --identity-file= --instance-type=r3.large
--wait=480  launch 




--
View this message in context: 
http://apache-spark-user-list.1001560.n3.nabble.com/Error-while-launching-ec2-spark-cluster-with-HVM-r3-large-tp5862p6276.html
Sent from the Apache Spark User List mailing list archive at Nabble.com.


Error while launching ec2 spark cluster with HVM (r3.large)

2014-05-16 Thread Usman Ghani
Hi All,

I am trying to use amazon memory optimized (R3) instances in the Oregon
region. I am getting 'Connection refused' during the SSH setup phase. I
tried using both root and ec2-user as user ids.


sh: connect to host ec2-54-185-57-74.us-west-2.compute.amazonaws.com port
22: Connection refused
Error executing remote command, retrying after 30 seconds: Command '['ssh',
'-o', 'StrictHostKeyChecking=no', '-i',
'/Users/usman/.ssh/ughani_imp_us_west_2.pem', '-t', '-t',
u'r...@ec2-54-185-57-74.us-west-2.compute.amazonaws.com', "\n  [ -f
~/.ssh/id_rsa ] ||\n(ssh-keygen -q -t rsa -N '' -f ~/.ssh/id_rsa
&&\n cat ~/.ssh/id_rsa.pub >> ~/.ssh/authorized_keys)\n"]'
returned non-zero exit status 255
ssh: connect to host ec2-54-185-57-74.us-west-2.compute.amazonaws.com port
22: Connection refused


Generating cluster's SSH key on master...
ssh: connect to host ec2-54-185-58-135.us-west-2.compute.amazonaws.com port
22: Connection refused
Error executing remote command, retrying after 30 seconds: Command '['ssh',
'-o', 'StrictHostKeyChecking=no', '-i',
'/Users/usman/.ssh/ughani_imp_us_west_2.pem', '-t', '-t',
u'ec2-u...@ec2-54-185-58-135.us-west-2.compute.amazonaws.com', "\n  [
-f ~/.ssh/id_rsa ] ||\n(ssh-keygen -q -t rsa -N '' -f ~/.ssh/id_rsa
&&\n cat ~/.ssh/id_rsa.pub >> ~/.ssh/authorized_keys)\n"]'
returned non-zero exit status 255
ssh: connect to host ec2-54-185-58-135.us-west-2.compute.amazonaws.com port
22: Connection refused
Error executing remote command, retrying after 30 seconds: Command '['ssh',
'-o', 'StrictHostKeyChecking=no', '-i',
'/Users/usman/.ssh/ughani_imp_us_west_2.pem', '-t', '-t',
u'ec2-u...@ec2-54-185-58-135.us-west-2.compute.amazonaws.com', "\n  [
-f ~/.ssh/id_rsa ] ||\n(ssh-keygen -q -t rsa -N '' -f ~/.ssh/id_rsa
&&\n cat ~/.ssh/id_rsa.pub >> ~/.ssh/authorized_keys)\n"]'
returned non-zero exit status 255
ssh: connect to host ec2-54-185-58-135.us-west-2.compute.amazonaws.com port
22: Connection refused
Error executing remote command, retrying after 30 seconds: Command '['ssh',
'-o', 'StrictHostKeyChecking=no', '-i',
'/Users/usman/.ssh/ughani_imp_us_west_2.pem', '-t', '-t',
u'ec2-u...@ec2-54-185-58-135.us-west-2.compute.amazonaws.com', "\n  [
-f ~/.ssh/id_rsa ] ||\n(ssh-keygen -q -t rsa -N '' -f ~/.ssh/id_rsa
&&\n cat ~/.ssh/id_rsa.pub >> ~/.ssh/authorized_keys)\n"]'
returned non-zero exit status 255
Warning: Permanently added '
ec2-54-185-58-135.us-west-2.compute.amazonaws.com,54.185.58.135' (RSA) to
the list of known hosts.
Connection to ec2-54-185-58-135.us-west-2.compute.amazonaws.com closed.
Transferring cluster's SSH key to slaves...
ec2-54-185-57-126.us-west-2.compute.amazonaws.com
ssh: connect to host ec2-54-185-57-126.us-west-2.compute.amazonaws.com port
22: Connection refused
Error 255 while executing remote command, retrying after 30 seconds
ssh: connect to host ec2-54-185-57-126.us-west-2.compute.amazonaws.com port
22: Connection refused
Error 255 while executing remote command, retrying after 30 seconds
ssh: connect to host ec2-54-185-57-126.us-west-2.compute.amazonaws.com port
22: Connection refused
Error 255 while executing remote command, retrying after 30 seconds
Warning: Permanently added '
ec2-54-185-57-126.us-west-2.compute.amazonaws.com,54.185.57.126' (RSA) to
the list of known hosts.
ec2-54-188-56-91.us-west-2.compute.amazonaws.com
Warning: Permanently added
'ec2-54-188-56-91.us-west-2.compute.amazonaws.com,54.188.56.91'
(RSA) to the list of known hosts.
ec2-54-188-61-59.us-west-2.compute.amazonaws.com
Warning: Permanently added
'ec2-54-188-61-59.us-west-2.compute.amazonaws.com,54.188.61.59'
(RSA) to the list of known hosts.
ec2-54-188-21-245.us-west-2.compute.amazonaws.com
Warning: Permanently added '
ec2-54-188-21-245.us-west-2.compute.amazonaws.com,54.188.21.245' (RSA) to
the list of known hosts.
Cloning into 'spark-ec2'...
remote: Counting objects: 1407, done.
remote: Compressing objects: 100% (673/673), done.
remote: Total 1407 (delta 457), reused 1398 (delta 454)
Receiving objects: 100% (1407/1407), 219.00 KiB | 340 KiB/s, done.
Resolving deltas: 100% (457/457), done.
Connection to ec2-54-185-58-135.us-west-2.compute.amazonaws.com closed.
Deploying files to master...
WARNING: Don't know number of disks on instance type r3.large; assuming 1
building file list ... done
root/spark-ec2/
rsync: recv_generator: mkdir "/root/spark-ec2" failed: Permission denied
(13)
*** Skipping any contents from this failed directory ***

sent 101 bytes  received 26 bytes  254.00 bytes/sec
total size is 1617  speedup is 12.73
rsync error: some files could not be transferred (code 23) at
/SourceCache/rsync/rsync-42/rsync/main.c(992) [sender=2.6.9]
Traceback (most recent call last):
  File "./spark_ec2.py", line 822, in 
main()
  File "./spark_ec2.py", line 814, in main
real_main()
  File "./spark_ec2.py", line 699, in real_main
setup_cluster(conn, master_nodes, slave_nodes, opts, True