I was able to keep the "workaround" ...around... by overwriting the generated '/root/.ssh/authorized_keys' file with a known good one, in the '/etc/rc.local' file
On Tue, Apr 8, 2014 at 10:12 AM, Marco Costantini < silvio.costant...@granatads.com> wrote: > Another thing I didn't mention. The AMI and user used: naturally I've > created several of my own AMIs with the following characteristics. None of > which worked. > > 1) Enabling ssh as root as per this guide ( > http://blog.tiger-workshop.com/enable-root-access-on-amazon-ec2-instance/). > When doing this, I do not specify a user for the spark-ec2 script. What > happens is that, it works! But only while it's alive. If I stop the > instance, create an AMI, and launch a new instance based from the new AMI, > the change I made in the '/root/.ssh/authorized_keys' file is overwritten > > 2) adding the 'ec2-user' to the 'root' group. This means that the ec2-user > does not have to use sudo to perform any operations needing root > privilidges. When doing this, I specify the user 'ec2-user' for the > spark-ec2 script. An error occurs: rsync fails with exit code 23. > > I believe HVMs still work. But it would be valuable to the community to > know that the root user work-around does/doesn't work any more for > paravirtual instances. > > Thanks, > Marco. > > > On Tue, Apr 8, 2014 at 9:51 AM, Marco Costantini < > silvio.costant...@granatads.com> wrote: > >> As requested, here is the script I am running. It is a simple shell >> script which calls spark-ec2 wrapper script. I execute it from the 'ec2' >> directory of spark, as usual. The AMI used is the raw one from the AWS >> Quick Start section. It is the first option (an Amazon Linux paravirtual >> image). Any ideas or confirmation would be GREATLY appreciated. Please and >> thank you. >> >> >> #!/bin/sh >> >> export AWS_ACCESS_KEY_ID=MyCensoredKey >> export AWS_SECRET_ACCESS_KEY=MyCensoredKey >> >> AMI_ID=ami-2f726546 >> >> ./spark-ec2 -k gds-generic -i ~/.ssh/gds-generic.pem -u ec2-user -s 10 -v >> 0.9.0 -w 300 --no-ganglia -a ${AMI_ID} -m m3.2xlarge -t m3.2xlarge launch >> marcotest >> >> >> >> On Mon, Apr 7, 2014 at 6:21 PM, Shivaram Venkataraman < >> shivaram.venkatara...@gmail.com> wrote: >> >>> Hmm -- That is strange. Can you paste the command you are using to >>> launch the instances ? The typical workflow is to use the spark-ec2 wrapper >>> script using the guidelines at >>> http://spark.apache.org/docs/latest/ec2-scripts.html >>> >>> Shivaram >>> >>> >>> On Mon, Apr 7, 2014 at 1:53 PM, Marco Costantini < >>> silvio.costant...@granatads.com> wrote: >>> >>>> Hi Shivaram, >>>> >>>> OK so let's assume the script CANNOT take a different user and that it >>>> must be 'root'. The typical workaround is as you said, allow the ssh with >>>> the root user. Now, don't laugh, but, this worked last Friday, but today >>>> (Monday) it no longer works. :D Why? ... >>>> >>>> ...It seems that NOW, when you launch a 'paravirtual' ami, the root >>>> user's 'authorized_keys' file is always overwritten. This means the >>>> workaround doesn't work anymore! I would LOVE for someone to verify this. >>>> >>>> Just to point out, I am trying to make this work with a paravirtual >>>> instance and not an HVM instance. >>>> >>>> Please and thanks, >>>> Marco. >>>> >>>> >>>> On Mon, Apr 7, 2014 at 4:40 PM, Shivaram Venkataraman < >>>> shivaram.venkatara...@gmail.com> wrote: >>>> >>>>> Right now the spark-ec2 scripts assume that you have root access and a >>>>> lot of internal scripts assume have the user's home directory hard coded >>>>> as >>>>> /root. However all the Spark AMIs we build should have root ssh access >>>>> -- >>>>> Do you find this not to be the case ? >>>>> >>>>> You can also enable root ssh access in a vanilla AMI by editing >>>>> /etc/ssh/sshd_config and setting "PermitRootLogin" to yes >>>>> >>>>> Thanks >>>>> Shivaram >>>>> >>>>> >>>>> >>>>> On Mon, Apr 7, 2014 at 11:14 AM, Marco Costantini < >>>>> silvio.costant...@granatads.com> wrote: >>>>> >>>>>> Hi all, >>>>>> On the old Amazon Linux EC2 images, the user 'root' was enabled for >>>>>> ssh. Also, it is the default user for the Spark-EC2 script. >>>>>> >>>>>> Currently, the Amazon Linux images have an 'ec2-user' set up for ssh >>>>>> instead of 'root'. >>>>>> >>>>>> I can see that the Spark-EC2 script allows you to specify which user >>>>>> to log in with, but even when I change this, the script fails for various >>>>>> reasons. And the output SEEMS that the script is still based on the >>>>>> specified user's home directory being '/root'. >>>>>> >>>>>> Am I using this script wrong? >>>>>> Has anyone had success with this 'ec2-user' user? >>>>>> Any ideas? >>>>>> >>>>>> Please and thank you, >>>>>> Marco. >>>>>> >>>>> >>>>> >>>> >>> >> >