Is there any reason why you want to start with a vanilla amazon AMI rather
than the ones we build and provide as a part of Spark EC2 scripts ? The
AMIs we provide are close to the vanilla AMI but have the root account
setup properly and install packages like java that are used by Spark.

If you wish to customize the AMI, you could always start with our AMI and
add more packages you like -- I have definitely done this recently and it
works with HVM and PVM as far as I can tell.

Shivaram


On Tue, Apr 8, 2014 at 8:50 AM, Marco Costantini <
silvio.costant...@granatads.com> wrote:

> I was able to keep the "workaround" ...around... by overwriting the
> generated '/root/.ssh/authorized_keys' file with a known good one, in the
> '/etc/rc.local' file
>
>
> On Tue, Apr 8, 2014 at 10:12 AM, Marco Costantini <
> silvio.costant...@granatads.com> wrote:
>
>> Another thing I didn't mention. The AMI and user used: naturally I've
>> created several of my own AMIs with the following characteristics. None of
>> which worked.
>>
>> 1) Enabling ssh as root as per this guide (
>> http://blog.tiger-workshop.com/enable-root-access-on-amazon-ec2-instance/).
>> When doing this, I do not specify a user for the spark-ec2 script. What
>> happens is that, it works! But only while it's alive. If I stop the
>> instance, create an AMI, and launch a new instance based from the new AMI,
>> the change I made in the '/root/.ssh/authorized_keys' file is overwritten
>>
>> 2) adding the 'ec2-user' to the 'root' group. This means that the
>> ec2-user does not have to use sudo to perform any operations needing root
>> privilidges. When doing this, I specify the user 'ec2-user' for the
>> spark-ec2 script. An error occurs: rsync fails with exit code 23.
>>
>> I believe HVMs still work. But it would be valuable to the community to
>> know that the root user work-around does/doesn't work any more for
>> paravirtual instances.
>>
>> Thanks,
>> Marco.
>>
>>
>> On Tue, Apr 8, 2014 at 9:51 AM, Marco Costantini <
>> silvio.costant...@granatads.com> wrote:
>>
>>> As requested, here is the script I am running. It is a simple shell
>>> script which calls spark-ec2 wrapper script. I execute it from the 'ec2'
>>> directory of spark, as usual. The AMI used is the raw one from the AWS
>>> Quick Start section. It is the first option (an Amazon Linux paravirtual
>>> image). Any ideas or confirmation would be GREATLY appreciated. Please and
>>> thank you.
>>>
>>>
>>> #!/bin/sh
>>>
>>> export AWS_ACCESS_KEY_ID=MyCensoredKey
>>> export AWS_SECRET_ACCESS_KEY=MyCensoredKey
>>>
>>> AMI_ID=ami-2f726546
>>>
>>> ./spark-ec2 -k gds-generic -i ~/.ssh/gds-generic.pem -u ec2-user -s 10
>>> -v 0.9.0 -w 300 --no-ganglia -a ${AMI_ID} -m m3.2xlarge -t m3.2xlarge
>>> launch marcotest
>>>
>>>
>>>
>>> On Mon, Apr 7, 2014 at 6:21 PM, Shivaram Venkataraman <
>>> shivaram.venkatara...@gmail.com> wrote:
>>>
>>>> Hmm -- That is strange. Can you paste the command you are using to
>>>> launch the instances ? The typical workflow is to use the spark-ec2 wrapper
>>>> script using the guidelines at
>>>> http://spark.apache.org/docs/latest/ec2-scripts.html
>>>>
>>>> Shivaram
>>>>
>>>>
>>>> On Mon, Apr 7, 2014 at 1:53 PM, Marco Costantini <
>>>> silvio.costant...@granatads.com> wrote:
>>>>
>>>>> Hi Shivaram,
>>>>>
>>>>> OK so let's assume the script CANNOT take a different user and that it
>>>>> must be 'root'. The typical workaround is as you said, allow the ssh with
>>>>> the root user. Now, don't laugh, but, this worked last Friday, but today
>>>>> (Monday) it no longer works. :D Why? ...
>>>>>
>>>>> ...It seems that NOW, when you launch a 'paravirtual' ami, the root
>>>>> user's 'authorized_keys' file is always overwritten. This means the
>>>>> workaround doesn't work anymore! I would LOVE for someone to verify this.
>>>>>
>>>>> Just to point out, I am trying to make this work with a paravirtual
>>>>> instance and not an HVM instance.
>>>>>
>>>>> Please and thanks,
>>>>> Marco.
>>>>>
>>>>>
>>>>> On Mon, Apr 7, 2014 at 4:40 PM, Shivaram Venkataraman <
>>>>> shivaram.venkatara...@gmail.com> wrote:
>>>>>
>>>>>> Right now the spark-ec2 scripts assume that you have root access and
>>>>>> a lot of internal scripts assume have the user's home directory hard 
>>>>>> coded
>>>>>> as /root.   However all the Spark AMIs we build should have root ssh 
>>>>>> access
>>>>>> -- Do you find this not to be the case ?
>>>>>>
>>>>>> You can also enable root ssh access in a vanilla AMI by editing
>>>>>> /etc/ssh/sshd_config and setting "PermitRootLogin" to yes
>>>>>>
>>>>>> Thanks
>>>>>> Shivaram
>>>>>>
>>>>>>
>>>>>>
>>>>>> On Mon, Apr 7, 2014 at 11:14 AM, Marco Costantini <
>>>>>> silvio.costant...@granatads.com> wrote:
>>>>>>
>>>>>>> Hi all,
>>>>>>> On the old Amazon Linux EC2 images, the user 'root' was enabled for
>>>>>>> ssh. Also, it is the default user for the Spark-EC2 script.
>>>>>>>
>>>>>>> Currently, the Amazon Linux images have an 'ec2-user' set up for ssh
>>>>>>> instead of 'root'.
>>>>>>>
>>>>>>> I can see that the Spark-EC2 script allows you to specify which user
>>>>>>> to log in with, but even when I change this, the script fails for 
>>>>>>> various
>>>>>>> reasons. And the output SEEMS that the script is still based on the
>>>>>>> specified user's home directory being '/root'.
>>>>>>>
>>>>>>> Am I using this script wrong?
>>>>>>> Has anyone had success with this 'ec2-user' user?
>>>>>>> Any ideas?
>>>>>>>
>>>>>>> Please and thank you,
>>>>>>> Marco.
>>>>>>>
>>>>>>
>>>>>>
>>>>>
>>>>
>>>
>>
>

Reply via email to