RE: spark 1.2 ec2 launch script hang

2015-01-28 Thread ey-chih chow
We found the problem and already fixed it.  Basically, spark-ec2 requires ec2 
instances to have external ip addresses. You need to specify this in the ASW 
console.  

From: nicholas.cham...@gmail.com
Date: Tue, 27 Jan 2015 17:19:21 +
Subject: Re: spark 1.2 ec2 launch script hang
To: charles.fed...@gmail.com; pzybr...@gmail.com; eyc...@hotmail.com
CC: user@spark.apache.org

For those who found that absolute vs. relative path for the pem file mattered, 
what OS and shell are you using? What version of Spark are you using?
~/ vs. absolute path shouldn’t matter. Your shell will expand the ~/ to the 
absolute path before sending it to spark-ec2. (i.e. tilde expansion.)
Absolute vs. relative path (e.g. ../../path/to/pem) also shouldn’t matter, 
since we fixed that for Spark 1.2.0. Maybe there’s some case that we missed?
Nick
On Tue Jan 27 2015 at 10:10:29 AM Charles Feduke charles.fed...@gmail.com 
wrote:

Absolute path means no ~ and also verify that you have the path to the file 
correct. For some reason the Python code does not validate that the file exists 
and will hang (this is the same reason why ~ hangs).
On Mon, Jan 26, 2015 at 10:08 PM Pete Zybrick pzybr...@gmail.com wrote:
Try using an absolute path to the pem file







 On Jan 26, 2015, at 8:57 PM, ey-chih chow eyc...@hotmail.com wrote:



 Hi,



 I used the spark-ec2 script of spark 1.2 to launch a cluster.  I have

 modified the script according to



 https://github.com/grzegorz-dubicki/spark/commit/5dd8458d2ab9753aae939b3bb33be953e2c13a70



 But the script was still hung at the following message:



 Waiting for cluster to enter 'ssh-ready'

 state.



 Any additional thing I should do to make it succeed?  Thanks.





 Ey-Chih Chow







 --

 View this message in context: 
 http://apache-spark-user-list.1001560.n3.nabble.com/spark-1-2-ec2-launch-script-hang-tp21381.html

 Sent from the Apache Spark User List mailing list archive at Nabble.com.



 -

 To unsubscribe, e-mail: user-unsubscr...@spark.apache.org

 For additional commands, e-mail: user-h...@spark.apache.org





-

To unsubscribe, e-mail: user-unsubscr...@spark.apache.org

For additional commands, e-mail: user-h...@spark.apache.org






​ 

Re: spark 1.2 ec2 launch script hang

2015-01-28 Thread Nicholas Chammas
Hmm, I can’t see why using ~ would be problematic, especially if you
confirm that echo ~/path/to/pem expands to the correct path to your
identity file.

If you have a simple reproduction of the problem, please send it over. I’d
love to look into this. When I pass paths with ~ to spark-ec2 on my system,
it works fine. I’m using bash, but zsh handles tilde expansion the same as
bash.

Nick
​

On Wed Jan 28 2015 at 3:30:08 PM Charles Feduke charles.fed...@gmail.com
wrote:

 It was only hanging when I specified the path with ~ I never tried
 relative.

 Hanging on the waiting for ssh to be ready on all hosts. I let it sit for
 about 10 minutes then I found the StackOverflow answer that suggested
 specifying an absolute path, cancelled, and re-run with --resume and the
 absolute path and all slaves were up in a couple minutes.

 (I've stood up 4 integration clusters and 2 production clusters on EC2
 since with no problems.)

 On Wed Jan 28 2015 at 12:05:43 PM Nicholas Chammas 
 nicholas.cham...@gmail.com wrote:

 Ey-chih,

 That makes more sense. This is a known issue that will be fixed as part
 of SPARK-5242 https://issues.apache.org/jira/browse/SPARK-5242.

 Charles,

 Thanks for the info. In your case, when does spark-ec2 hang? Only when
 the specified path to the identity file doesn't exist? Or also when you
 specify the path as a relative path or with ~?

 Nick


 On Wed Jan 28 2015 at 9:29:34 AM ey-chih chow eyc...@hotmail.com wrote:

 We found the problem and already fixed it.  Basically, spark-ec2
 requires ec2 instances to have external ip addresses. You need to specify
 this in the ASW console.
 --
 From: nicholas.cham...@gmail.com
 Date: Tue, 27 Jan 2015 17:19:21 +
 Subject: Re: spark 1.2 ec2 launch script hang
 To: charles.fed...@gmail.com; pzybr...@gmail.com; eyc...@hotmail.com
 CC: user@spark.apache.org


 For those who found that absolute vs. relative path for the pem file
 mattered, what OS and shell are you using? What version of Spark are you
 using?

 ~/ vs. absolute path shouldn’t matter. Your shell will expand the ~/ to
 the absolute path before sending it to spark-ec2. (i.e. tilde
 expansion.)

 Absolute vs. relative path (e.g. ../../path/to/pem) also shouldn’t
 matter, since we fixed that for Spark 1.2.0
 https://issues.apache.org/jira/browse/SPARK-4137. Maybe there’s some
 case that we missed?

 Nick

 On Tue Jan 27 2015 at 10:10:29 AM Charles Feduke 
 charles.fed...@gmail.com wrote:


 Absolute path means no ~ and also verify that you have the path to the
 file correct. For some reason the Python code does not validate that the
 file exists and will hang (this is the same reason why ~ hangs).
 On Mon, Jan 26, 2015 at 10:08 PM Pete Zybrick pzybr...@gmail.com
 wrote:

 Try using an absolute path to the pem file



  On Jan 26, 2015, at 8:57 PM, ey-chih chow eyc...@hotmail.com wrote:
 
  Hi,
 
  I used the spark-ec2 script of spark 1.2 to launch a cluster.  I have
  modified the script according to
 
  https://github.com/grzegorz-dubicki/spark/commit/5dd8458d2ab
 9753aae939b3bb33be953e2c13a70
 
  But the script was still hung at the following message:
 
  Waiting for cluster to enter 'ssh-ready'
  state.
 
  Any additional thing I should do to make it succeed?  Thanks.
 
 
  Ey-Chih Chow
 
 
 
  --
  View this message in context: http://apache-spark-user-list.
 1001560.n3.nabble.com/spark-1-2-ec2-launch-script-hang-tp21381.html
  Sent from the Apache Spark User List mailing list archive at
 Nabble.com.
 
  -
  To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
  For additional commands, e-mail: user-h...@spark.apache.org
 

 -
 To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
 For additional commands, e-mail: user-h...@spark.apache.org


 ​




Re: spark 1.2 ec2 launch script hang

2015-01-28 Thread Charles Feduke
Yeah, I agree ~ should work. And it could have been [read: probably was]
the fact that one of the EC2 hosts was in my known_hosts (don't know, never
saw an error message, but the behavior is no error message for that state),
which I had fixed later with Pete's patch. But the second execution when
things worked with an absolute path could have worked because the random
hosts that came up on EC2 were never in my known_hosts.

On Wed Jan 28 2015 at 3:45:36 PM Nicholas Chammas 
nicholas.cham...@gmail.com wrote:

 Hmm, I can’t see why using ~ would be problematic, especially if you
 confirm that echo ~/path/to/pem expands to the correct path to your
 identity file.

 If you have a simple reproduction of the problem, please send it over. I’d
 love to look into this. When I pass paths with ~ to spark-ec2 on my system,
 it works fine. I’m using bash, but zsh handles tilde expansion the same as
 bash.

 Nick
 ​

 On Wed Jan 28 2015 at 3:30:08 PM Charles Feduke charles.fed...@gmail.com
 wrote:

 It was only hanging when I specified the path with ~ I never tried
 relative.

 Hanging on the waiting for ssh to be ready on all hosts. I let it sit for
 about 10 minutes then I found the StackOverflow answer that suggested
 specifying an absolute path, cancelled, and re-run with --resume and the
 absolute path and all slaves were up in a couple minutes.

 (I've stood up 4 integration clusters and 2 production clusters on EC2
 since with no problems.)

 On Wed Jan 28 2015 at 12:05:43 PM Nicholas Chammas 
 nicholas.cham...@gmail.com wrote:

 Ey-chih,

 That makes more sense. This is a known issue that will be fixed as part
 of SPARK-5242 https://issues.apache.org/jira/browse/SPARK-5242.

 Charles,

 Thanks for the info. In your case, when does spark-ec2 hang? Only when
 the specified path to the identity file doesn't exist? Or also when you
 specify the path as a relative path or with ~?

 Nick


 On Wed Jan 28 2015 at 9:29:34 AM ey-chih chow eyc...@hotmail.com
 wrote:

 We found the problem and already fixed it.  Basically, spark-ec2
 requires ec2 instances to have external ip addresses. You need to specify
 this in the ASW console.
 --
 From: nicholas.cham...@gmail.com
 Date: Tue, 27 Jan 2015 17:19:21 +
 Subject: Re: spark 1.2 ec2 launch script hang
 To: charles.fed...@gmail.com; pzybr...@gmail.com; eyc...@hotmail.com
 CC: user@spark.apache.org


 For those who found that absolute vs. relative path for the pem file
 mattered, what OS and shell are you using? What version of Spark are you
 using?

 ~/ vs. absolute path shouldn’t matter. Your shell will expand the ~/
 to the absolute path before sending it to spark-ec2. (i.e. tilde
 expansion.)

 Absolute vs. relative path (e.g. ../../path/to/pem) also shouldn’t
 matter, since we fixed that for Spark 1.2.0
 https://issues.apache.org/jira/browse/SPARK-4137. Maybe there’s some
 case that we missed?

 Nick

 On Tue Jan 27 2015 at 10:10:29 AM Charles Feduke 
 charles.fed...@gmail.com wrote:


 Absolute path means no ~ and also verify that you have the path to the
 file correct. For some reason the Python code does not validate that the
 file exists and will hang (this is the same reason why ~ hangs).
 On Mon, Jan 26, 2015 at 10:08 PM Pete Zybrick pzybr...@gmail.com
 wrote:

 Try using an absolute path to the pem file



  On Jan 26, 2015, at 8:57 PM, ey-chih chow eyc...@hotmail.com wrote:
 
  Hi,
 
  I used the spark-ec2 script of spark 1.2 to launch a cluster.  I have
  modified the script according to
 
  https://github.com/grzegorz-dubicki/spark/commit/5dd8458d2ab
 9753aae939b3bb33be953e2c13a70
 
  But the script was still hung at the following message:
 
  Waiting for cluster to enter 'ssh-ready'
  state.
 
  Any additional thing I should do to make it succeed?  Thanks.
 
 
  Ey-Chih Chow
 
 
 
  --
  View this message in context: http://apache-spark-user-list.
 1001560.n3.nabble.com/spark-1-2-ec2-launch-script-hang-tp21381.html
  Sent from the Apache Spark User List mailing list archive at
 Nabble.com.
 
  -
  To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
  For additional commands, e-mail: user-h...@spark.apache.org
 

 -
 To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
 For additional commands, e-mail: user-h...@spark.apache.org


 ​




Re: spark 1.2 ec2 launch script hang

2015-01-28 Thread Charles Feduke
It was only hanging when I specified the path with ~ I never tried relative.

Hanging on the waiting for ssh to be ready on all hosts. I let it sit for
about 10 minutes then I found the StackOverflow answer that suggested
specifying an absolute path, cancelled, and re-run with --resume and the
absolute path and all slaves were up in a couple minutes.

(I've stood up 4 integration clusters and 2 production clusters on EC2
since with no problems.)

On Wed Jan 28 2015 at 12:05:43 PM Nicholas Chammas 
nicholas.cham...@gmail.com wrote:

 Ey-chih,

 That makes more sense. This is a known issue that will be fixed as part of
 SPARK-5242 https://issues.apache.org/jira/browse/SPARK-5242.

 Charles,

 Thanks for the info. In your case, when does spark-ec2 hang? Only when the
 specified path to the identity file doesn't exist? Or also when you specify
 the path as a relative path or with ~?

 Nick


 On Wed Jan 28 2015 at 9:29:34 AM ey-chih chow eyc...@hotmail.com wrote:

 We found the problem and already fixed it.  Basically, spark-ec2 requires
 ec2 instances to have external ip addresses. You need to specify this in
 the ASW console.
 --
 From: nicholas.cham...@gmail.com
 Date: Tue, 27 Jan 2015 17:19:21 +
 Subject: Re: spark 1.2 ec2 launch script hang
 To: charles.fed...@gmail.com; pzybr...@gmail.com; eyc...@hotmail.com
 CC: user@spark.apache.org


 For those who found that absolute vs. relative path for the pem file
 mattered, what OS and shell are you using? What version of Spark are you
 using?

 ~/ vs. absolute path shouldn’t matter. Your shell will expand the ~/ to
 the absolute path before sending it to spark-ec2. (i.e. tilde expansion.)

 Absolute vs. relative path (e.g. ../../path/to/pem) also shouldn’t
 matter, since we fixed that for Spark 1.2.0
 https://issues.apache.org/jira/browse/SPARK-4137. Maybe there’s some
 case that we missed?

 Nick

 On Tue Jan 27 2015 at 10:10:29 AM Charles Feduke 
 charles.fed...@gmail.com wrote:


 Absolute path means no ~ and also verify that you have the path to the
 file correct. For some reason the Python code does not validate that the
 file exists and will hang (this is the same reason why ~ hangs).
 On Mon, Jan 26, 2015 at 10:08 PM Pete Zybrick pzybr...@gmail.com wrote:

 Try using an absolute path to the pem file



  On Jan 26, 2015, at 8:57 PM, ey-chih chow eyc...@hotmail.com wrote:
 
  Hi,
 
  I used the spark-ec2 script of spark 1.2 to launch a cluster.  I have
  modified the script according to
 
  https://github.com/grzegorz-dubicki/spark/commit/5dd8458d2ab
 9753aae939b3bb33be953e2c13a70
 
  But the script was still hung at the following message:
 
  Waiting for cluster to enter 'ssh-ready'
  state.
 
  Any additional thing I should do to make it succeed?  Thanks.
 
 
  Ey-Chih Chow
 
 
 
  --
  View this message in context: http://apache-spark-user-list.
 1001560.n3.nabble.com/spark-1-2-ec2-launch-script-hang-tp21381.html
  Sent from the Apache Spark User List mailing list archive at Nabble.com.
 
  -
  To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
  For additional commands, e-mail: user-h...@spark.apache.org
 

 -
 To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
 For additional commands, e-mail: user-h...@spark.apache.org


 ​




Re: spark 1.2 ec2 launch script hang

2015-01-28 Thread Nicholas Chammas
Thanks for sending this over, Peter.

What if you try this? (i.e. Remove the = after --identity-file.)

ec2/spark-ec2 --key-pair=spark-streaming-kp --identity-file
~/.pzkeys/spark-streaming-kp.pem  --region=us-east-1 login
pz-spark-cluster

If that works, then I think the problem in this case is simply that Bash
cannot expand the tilde because it’s stuck to the --identity-file=. This
isn’t a problem with spark-ec2.

Bash sees the --identity-file=~/.pzkeys/spark-streaming-kp.pem as one big
argument, so it can’t do tilde expansion.

Nick
​

On Wed Jan 28 2015 at 9:17:06 PM Peter Zybrick pzybr...@gmail.com wrote:

 Below is trace from trying to access with ~/path.  I also did the echo as
 per Nick (see the last line), looks ok to me.  This is my development box
 with Spark 1.2.0 running CentOS 6.5, Python 2.6.6

 [pete.zybrick@pz-lt2-ipc spark-1.2.0]$ ec2/spark-ec2
 --key-pair=spark-streaming-kp
 --identity-file=~/.pzkeys/spark-streaming-kp.pem  --region=us-east-1 login
 pz-spark-cluster
 Searching for existing cluster pz-spark-cluster...
 Found 1 master(s), 3 slaves
 Logging into master ec2-54-152-95-129.compute-1.amazonaws.com...
 Warning: Identity file ~/.pzkeys/spark-streaming-kp.pem not accessible: No
 such file or directory.
 Permission denied (publickey).
 Traceback (most recent call last):
   File ec2/spark_ec2.py, line 1082, in module
 main()
   File ec2/spark_ec2.py, line 1074, in main
 real_main()
   File ec2/spark_ec2.py, line 1007, in real_main
 ssh_command(opts) + proxy_opt + ['-t', '-t', %s@%s % (opts.user,
 master)])
   File /usr/lib64/python2.6/subprocess.py, line 505, in check_call
 raise CalledProcessError(retcode, cmd)
 subprocess.CalledProcessError: Command '['ssh', '-o',
 'StrictHostKeyChecking=no', '-i', '~/.pzkeys/spark-streaming-kp.pem', '-t',
 '-t', u'r...@ec2-54-152-95-129.compute-1.amazonaws.com']' returned
 non-zero exit status 255
 [pete.zybrick@pz-lt2-ipc spark-1.2.0]$ echo
 ~/.pzkeys/spark-streaming-kp.pem
 /home/pete.zybrick/.pzkeys/spark-streaming-kp.pem


 On Wed, Jan 28, 2015 at 3:49 PM, Charles Feduke charles.fed...@gmail.com
 wrote:

 Yeah, I agree ~ should work. And it could have been [read: probably was]
 the fact that one of the EC2 hosts was in my known_hosts (don't know, never
 saw an error message, but the behavior is no error message for that state),
 which I had fixed later with Pete's patch. But the second execution when
 things worked with an absolute path could have worked because the random
 hosts that came up on EC2 were never in my known_hosts.


 On Wed Jan 28 2015 at 3:45:36 PM Nicholas Chammas 
 nicholas.cham...@gmail.com wrote:

 Hmm, I can’t see why using ~ would be problematic, especially if you
 confirm that echo ~/path/to/pem expands to the correct path to your
 identity file.

 If you have a simple reproduction of the problem, please send it over.
 I’d love to look into this. When I pass paths with ~ to spark-ec2 on my
 system, it works fine. I’m using bash, but zsh handles tilde expansion the
 same as bash.

 Nick
 ​

 On Wed Jan 28 2015 at 3:30:08 PM Charles Feduke 
 charles.fed...@gmail.com wrote:

 It was only hanging when I specified the path with ~ I never tried
 relative.

 Hanging on the waiting for ssh to be ready on all hosts. I let it sit
 for about 10 minutes then I found the StackOverflow answer that suggested
 specifying an absolute path, cancelled, and re-run with --resume and the
 absolute path and all slaves were up in a couple minutes.

 (I've stood up 4 integration clusters and 2 production clusters on EC2
 since with no problems.)

 On Wed Jan 28 2015 at 12:05:43 PM Nicholas Chammas 
 nicholas.cham...@gmail.com wrote:

 Ey-chih,

 That makes more sense. This is a known issue that will be fixed as
 part of SPARK-5242 https://issues.apache.org/jira/browse/SPARK-5242.

 Charles,

 Thanks for the info. In your case, when does spark-ec2 hang? Only when
 the specified path to the identity file doesn't exist? Or also when you
 specify the path as a relative path or with ~?

 Nick


 On Wed Jan 28 2015 at 9:29:34 AM ey-chih chow eyc...@hotmail.com
 wrote:

 We found the problem and already fixed it.  Basically, spark-ec2
 requires ec2 instances to have external ip addresses. You need to specify
 this in the ASW console.
 --
 From: nicholas.cham...@gmail.com
 Date: Tue, 27 Jan 2015 17:19:21 +
 Subject: Re: spark 1.2 ec2 launch script hang
 To: charles.fed...@gmail.com; pzybr...@gmail.com; eyc...@hotmail.com
 CC: user@spark.apache.org


 For those who found that absolute vs. relative path for the pem file
 mattered, what OS and shell are you using? What version of Spark are you
 using?

 ~/ vs. absolute path shouldn’t matter. Your shell will expand the ~/
 to the absolute path before sending it to spark-ec2. (i.e. tilde
 expansion.)

 Absolute vs. relative path (e.g. ../../path/to/pem) also shouldn’t
 matter, since we fixed that for Spark 1.2.0
 https://issues.apache.org/jira/browse/SPARK

Re: spark 1.2 ec2 launch script hang

2015-01-28 Thread Nicholas Chammas
If that was indeed the problem, I suggest updating your answer on SO
http://stackoverflow.com/a/28005151/877069 to help others who may run
into this same problem.
​

On Wed Jan 28 2015 at 9:40:39 PM Nicholas Chammas 
nicholas.cham...@gmail.com wrote:

 Thanks for sending this over, Peter.

 What if you try this? (i.e. Remove the = after --identity-file.)

 ec2/spark-ec2 --key-pair=spark-streaming-kp --identity-file 
 ~/.pzkeys/spark-streaming-kp.pem  --region=us-east-1 login pz-spark-cluster

 If that works, then I think the problem in this case is simply that Bash
 cannot expand the tilde because it’s stuck to the --identity-file=. This
 isn’t a problem with spark-ec2.

 Bash sees the --identity-file=~/.pzkeys/spark-streaming-kp.pem as one big
 argument, so it can’t do tilde expansion.

 Nick
 ​

 On Wed Jan 28 2015 at 9:17:06 PM Peter Zybrick pzybr...@gmail.com wrote:

 Below is trace from trying to access with ~/path.  I also did the echo as
 per Nick (see the last line), looks ok to me.  This is my development box
 with Spark 1.2.0 running CentOS 6.5, Python 2.6.6

 [pete.zybrick@pz-lt2-ipc spark-1.2.0]$ ec2/spark-ec2
 --key-pair=spark-streaming-kp 
 --identity-file=~/.pzkeys/spark-streaming-kp.pem
 --region=us-east-1 login pz-spark-cluster
 Searching for existing cluster pz-spark-cluster...
 Found 1 master(s), 3 slaves
 Logging into master ec2-54-152-95-129.compute-1.amazonaws.com...
 Warning: Identity file ~/.pzkeys/spark-streaming-kp.pem not accessible:
 No such file or directory.
 Permission denied (publickey).
 Traceback (most recent call last):
   File ec2/spark_ec2.py, line 1082, in module
 main()
   File ec2/spark_ec2.py, line 1074, in main
 real_main()
   File ec2/spark_ec2.py, line 1007, in real_main
 ssh_command(opts) + proxy_opt + ['-t', '-t', %s@%s % (opts.user,
 master)])
   File /usr/lib64/python2.6/subprocess.py, line 505, in check_call
 raise CalledProcessError(retcode, cmd)
 subprocess.CalledProcessError: Command '['ssh', '-o',
 'StrictHostKeyChecking=no', '-i', '~/.pzkeys/spark-streaming-kp.pem',
 '-t', '-t', u'r...@ec2-54-152-95-129.compute-1.amazonaws.com']' returned
 non-zero exit status 255
 [pete.zybrick@pz-lt2-ipc spark-1.2.0]$ echo ~/.pzkeys/spark-streaming-kp.
 pem
 /home/pete.zybrick/.pzkeys/spark-streaming-kp.pem


 On Wed, Jan 28, 2015 at 3:49 PM, Charles Feduke charles.fed...@gmail.com
  wrote:

 Yeah, I agree ~ should work. And it could have been [read: probably was]
 the fact that one of the EC2 hosts was in my known_hosts (don't know, never
 saw an error message, but the behavior is no error message for that state),
 which I had fixed later with Pete's patch. But the second execution when
 things worked with an absolute path could have worked because the random
 hosts that came up on EC2 were never in my known_hosts.


 On Wed Jan 28 2015 at 3:45:36 PM Nicholas Chammas 
 nicholas.cham...@gmail.com wrote:

 Hmm, I can’t see why using ~ would be problematic, especially if you
 confirm that echo ~/path/to/pem expands to the correct path to your
 identity file.

 If you have a simple reproduction of the problem, please send it over.
 I’d love to look into this. When I pass paths with ~ to spark-ec2 on my
 system, it works fine. I’m using bash, but zsh handles tilde expansion the
 same as bash.

 Nick
 ​

 On Wed Jan 28 2015 at 3:30:08 PM Charles Feduke 
 charles.fed...@gmail.com wrote:

 It was only hanging when I specified the path with ~ I never tried
 relative.

 Hanging on the waiting for ssh to be ready on all hosts. I let it sit
 for about 10 minutes then I found the StackOverflow answer that suggested
 specifying an absolute path, cancelled, and re-run with --resume and the
 absolute path and all slaves were up in a couple minutes.

 (I've stood up 4 integration clusters and 2 production clusters on EC2
 since with no problems.)

 On Wed Jan 28 2015 at 12:05:43 PM Nicholas Chammas 
 nicholas.cham...@gmail.com wrote:

 Ey-chih,

 That makes more sense. This is a known issue that will be fixed as
 part of SPARK-5242 https://issues.apache.org/jira/browse/SPARK-5242
 .

 Charles,

 Thanks for the info. In your case, when does spark-ec2 hang? Only
 when the specified path to the identity file doesn't exist? Or also when
 you specify the path as a relative path or with ~?

 Nick


 On Wed Jan 28 2015 at 9:29:34 AM ey-chih chow eyc...@hotmail.com
 wrote:

 We found the problem and already fixed it.  Basically, spark-ec2
 requires ec2 instances to have external ip addresses. You need to 
 specify
 this in the ASW console.
 --
 From: nicholas.cham...@gmail.com
 Date: Tue, 27 Jan 2015 17:19:21 +
 Subject: Re: spark 1.2 ec2 launch script hang
 To: charles.fed...@gmail.com; pzybr...@gmail.com; eyc...@hotmail.com
 CC: user@spark.apache.org


 For those who found that absolute vs. relative path for the pem file
 mattered, what OS and shell are you using? What version of Spark are you
 using?

 ~/ vs. absolute path shouldn’t matter

Re: spark 1.2 ec2 launch script hang

2015-01-28 Thread Peter Zybrick
Below is trace from trying to access with ~/path.  I also did the echo as
per Nick (see the last line), looks ok to me.  This is my development box
with Spark 1.2.0 running CentOS 6.5, Python 2.6.6

[pete.zybrick@pz-lt2-ipc spark-1.2.0]$ ec2/spark-ec2
--key-pair=spark-streaming-kp
--identity-file=~/.pzkeys/spark-streaming-kp.pem  --region=us-east-1 login
pz-spark-cluster
Searching for existing cluster pz-spark-cluster...
Found 1 master(s), 3 slaves
Logging into master ec2-54-152-95-129.compute-1.amazonaws.com...
Warning: Identity file ~/.pzkeys/spark-streaming-kp.pem not accessible: No
such file or directory.
Permission denied (publickey).
Traceback (most recent call last):
  File ec2/spark_ec2.py, line 1082, in module
main()
  File ec2/spark_ec2.py, line 1074, in main
real_main()
  File ec2/spark_ec2.py, line 1007, in real_main
ssh_command(opts) + proxy_opt + ['-t', '-t', %s@%s % (opts.user,
master)])
  File /usr/lib64/python2.6/subprocess.py, line 505, in check_call
raise CalledProcessError(retcode, cmd)
subprocess.CalledProcessError: Command '['ssh', '-o',
'StrictHostKeyChecking=no', '-i', '~/.pzkeys/spark-streaming-kp.pem', '-t',
'-t', u'r...@ec2-54-152-95-129.compute-1.amazonaws.com']' returned non-zero
exit status 255
[pete.zybrick@pz-lt2-ipc spark-1.2.0]$ echo ~/.pzkeys/spark-streaming-kp.pem
/home/pete.zybrick/.pzkeys/spark-streaming-kp.pem


On Wed, Jan 28, 2015 at 3:49 PM, Charles Feduke charles.fed...@gmail.com
wrote:

 Yeah, I agree ~ should work. And it could have been [read: probably was]
 the fact that one of the EC2 hosts was in my known_hosts (don't know, never
 saw an error message, but the behavior is no error message for that state),
 which I had fixed later with Pete's patch. But the second execution when
 things worked with an absolute path could have worked because the random
 hosts that came up on EC2 were never in my known_hosts.


 On Wed Jan 28 2015 at 3:45:36 PM Nicholas Chammas 
 nicholas.cham...@gmail.com wrote:

 Hmm, I can’t see why using ~ would be problematic, especially if you
 confirm that echo ~/path/to/pem expands to the correct path to your
 identity file.

 If you have a simple reproduction of the problem, please send it over.
 I’d love to look into this. When I pass paths with ~ to spark-ec2 on my
 system, it works fine. I’m using bash, but zsh handles tilde expansion the
 same as bash.

 Nick
 ​

 On Wed Jan 28 2015 at 3:30:08 PM Charles Feduke charles.fed...@gmail.com
 wrote:

 It was only hanging when I specified the path with ~ I never tried
 relative.

 Hanging on the waiting for ssh to be ready on all hosts. I let it sit
 for about 10 minutes then I found the StackOverflow answer that suggested
 specifying an absolute path, cancelled, and re-run with --resume and the
 absolute path and all slaves were up in a couple minutes.

 (I've stood up 4 integration clusters and 2 production clusters on EC2
 since with no problems.)

 On Wed Jan 28 2015 at 12:05:43 PM Nicholas Chammas 
 nicholas.cham...@gmail.com wrote:

 Ey-chih,

 That makes more sense. This is a known issue that will be fixed as part
 of SPARK-5242 https://issues.apache.org/jira/browse/SPARK-5242.

 Charles,

 Thanks for the info. In your case, when does spark-ec2 hang? Only when
 the specified path to the identity file doesn't exist? Or also when you
 specify the path as a relative path or with ~?

 Nick


 On Wed Jan 28 2015 at 9:29:34 AM ey-chih chow eyc...@hotmail.com
 wrote:

 We found the problem and already fixed it.  Basically, spark-ec2
 requires ec2 instances to have external ip addresses. You need to specify
 this in the ASW console.
 --
 From: nicholas.cham...@gmail.com
 Date: Tue, 27 Jan 2015 17:19:21 +
 Subject: Re: spark 1.2 ec2 launch script hang
 To: charles.fed...@gmail.com; pzybr...@gmail.com; eyc...@hotmail.com
 CC: user@spark.apache.org


 For those who found that absolute vs. relative path for the pem file
 mattered, what OS and shell are you using? What version of Spark are you
 using?

 ~/ vs. absolute path shouldn’t matter. Your shell will expand the ~/
 to the absolute path before sending it to spark-ec2. (i.e. tilde
 expansion.)

 Absolute vs. relative path (e.g. ../../path/to/pem) also shouldn’t
 matter, since we fixed that for Spark 1.2.0
 https://issues.apache.org/jira/browse/SPARK-4137. Maybe there’s
 some case that we missed?

 Nick

 On Tue Jan 27 2015 at 10:10:29 AM Charles Feduke 
 charles.fed...@gmail.com wrote:


 Absolute path means no ~ and also verify that you have the path to the
 file correct. For some reason the Python code does not validate that the
 file exists and will hang (this is the same reason why ~ hangs).
 On Mon, Jan 26, 2015 at 10:08 PM Pete Zybrick pzybr...@gmail.com
 wrote:

 Try using an absolute path to the pem file



  On Jan 26, 2015, at 8:57 PM, ey-chih chow eyc...@hotmail.com
 wrote:
 
  Hi,
 
  I used the spark-ec2 script of spark 1.2 to launch a cluster.  I have
  modified the script

Re: spark 1.2 ec2 launch script hang

2015-01-27 Thread Charles Feduke
Absolute path means no ~ and also verify that you have the path to the file
correct. For some reason the Python code does not validate that the file
exists and will hang (this is the same reason why ~ hangs).
On Mon, Jan 26, 2015 at 10:08 PM Pete Zybrick pzybr...@gmail.com wrote:

 Try using an absolute path to the pem file



  On Jan 26, 2015, at 8:57 PM, ey-chih chow eyc...@hotmail.com wrote:
 
  Hi,
 
  I used the spark-ec2 script of spark 1.2 to launch a cluster.  I have
  modified the script according to
 
  https://github.com/grzegorz-dubicki/spark/commit/
 5dd8458d2ab9753aae939b3bb33be953e2c13a70
 
  But the script was still hung at the following message:
 
  Waiting for cluster to enter 'ssh-ready'
  state.
 
  Any additional thing I should do to make it succeed?  Thanks.
 
 
  Ey-Chih Chow
 
 
 
  --
  View this message in context: http://apache-spark-user-list.
 1001560.n3.nabble.com/spark-1-2-ec2-launch-script-hang-tp21381.html
  Sent from the Apache Spark User List mailing list archive at Nabble.com.
 
  -
  To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
  For additional commands, e-mail: user-h...@spark.apache.org
 

 -
 To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
 For additional commands, e-mail: user-h...@spark.apache.org




Re: spark 1.2 ec2 launch script hang

2015-01-27 Thread Nicholas Chammas
For those who found that absolute vs. relative path for the pem file
mattered, what OS and shell are you using? What version of Spark are you
using?

~/ vs. absolute path shouldn’t matter. Your shell will expand the ~/ to the
absolute path before sending it to spark-ec2. (i.e. tilde expansion.)

Absolute vs. relative path (e.g. ../../path/to/pem) also shouldn’t matter,
since we fixed that for Spark 1.2.0
https://issues.apache.org/jira/browse/SPARK-4137. Maybe there’s some case
that we missed?

Nick

On Tue Jan 27 2015 at 10:10:29 AM Charles Feduke charles.fed...@gmail.com
wrote:

Absolute path means no ~ and also verify that you have the path to the file
 correct. For some reason the Python code does not validate that the file
 exists and will hang (this is the same reason why ~ hangs).
 On Mon, Jan 26, 2015 at 10:08 PM Pete Zybrick pzybr...@gmail.com wrote:

 Try using an absolute path to the pem file



  On Jan 26, 2015, at 8:57 PM, ey-chih chow eyc...@hotmail.com wrote:
 
  Hi,
 
  I used the spark-ec2 script of spark 1.2 to launch a cluster.  I have
  modified the script according to
 
  https://github.com/grzegorz-dubicki/spark/commit/5dd8458d2ab
 9753aae939b3bb33be953e2c13a70
 
  But the script was still hung at the following message:
 
  Waiting for cluster to enter 'ssh-ready'
  state.
 
  Any additional thing I should do to make it succeed?  Thanks.
 
 
  Ey-Chih Chow
 
 
 
  --
  View this message in context: http://apache-spark-user-list.
 1001560.n3.nabble.com/spark-1-2-ec2-launch-script-hang-tp21381.html
  Sent from the Apache Spark User List mailing list archive at Nabble.com.
 
  -
  To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
  For additional commands, e-mail: user-h...@spark.apache.org
 

 -
 To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
 For additional commands, e-mail: user-h...@spark.apache.org

  ​


Re: spark 1.2 ec2 launch script hang

2015-01-26 Thread Pete Zybrick
Try using an absolute path to the pem file



 On Jan 26, 2015, at 8:57 PM, ey-chih chow eyc...@hotmail.com wrote:
 
 Hi,
 
 I used the spark-ec2 script of spark 1.2 to launch a cluster.  I have
 modified the script according to 
 
 https://github.com/grzegorz-dubicki/spark/commit/5dd8458d2ab9753aae939b3bb33be953e2c13a70
 
 But the script was still hung at the following message:
 
 Waiting for cluster to enter 'ssh-ready'
 state.
 
 Any additional thing I should do to make it succeed?  Thanks.
 
 
 Ey-Chih Chow
 
 
 
 --
 View this message in context: 
 http://apache-spark-user-list.1001560.n3.nabble.com/spark-1-2-ec2-launch-script-hang-tp21381.html
 Sent from the Apache Spark User List mailing list archive at Nabble.com.
 
 -
 To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
 For additional commands, e-mail: user-h...@spark.apache.org
 

-
To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
For additional commands, e-mail: user-h...@spark.apache.org



spark 1.2 ec2 launch script hang

2015-01-26 Thread ey-chih chow
Hi,

I used the spark-ec2 script of spark 1.2 to launch a cluster.  I have
modified the script according to 

https://github.com/grzegorz-dubicki/spark/commit/5dd8458d2ab9753aae939b3bb33be953e2c13a70

But the script was still hung at the following message:

Waiting for cluster to enter 'ssh-ready'
state.

Any additional thing I should do to make it succeed?  Thanks.


Ey-Chih Chow



--
View this message in context: 
http://apache-spark-user-list.1001560.n3.nabble.com/spark-1-2-ec2-launch-script-hang-tp21381.html
Sent from the Apache Spark User List mailing list archive at Nabble.com.

-
To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
For additional commands, e-mail: user-h...@spark.apache.org