Ah, tried that. I believe this is an HVM AMI? We are exploring paravirtual
AMIs.


On Wed, Apr 9, 2014 at 11:17 AM, Nicholas Chammas <
nicholas.cham...@gmail.com> wrote:

> And for the record, that AMI is ami-35b1885c. Again, you don't need to
> specify it explicitly; spark-ec2 will default to it.
>
>
> On Wed, Apr 9, 2014 at 11:08 AM, Nicholas Chammas <
> nicholas.cham...@gmail.com> wrote:
>
>> Marco,
>>
>> If you call spark-ec2 launch without specifying an AMI, it will default
>> to the Spark-provided AMI.
>>
>> Nick
>>
>>
>> On Wed, Apr 9, 2014 at 9:43 AM, Marco Costantini <
>> silvio.costant...@granatads.com> wrote:
>>
>>> Hi there,
>>> To answer your question; no there is no reason NOT to use an AMI that
>>> Spark has prepared. The reason we haven't is that we were not aware such
>>> AMIs existed. Would you kindly point us to the documentation where we can
>>> read about this further?
>>>
>>> Many many thanks, Shivaram.
>>> Marco.
>>>
>>>
>>> On Tue, Apr 8, 2014 at 4:42 PM, Shivaram Venkataraman <
>>> shiva...@eecs.berkeley.edu> wrote:
>>>
>>>> Is there any reason why you want to start with a vanilla amazon AMI
>>>> rather than the ones we build and provide as a part of Spark EC2 scripts ?
>>>> The AMIs we provide are close to the vanilla AMI but have the root account
>>>> setup properly and install packages like java that are used by Spark.
>>>>
>>>> If you wish to customize the AMI, you could always start with our AMI
>>>> and add more packages you like -- I have definitely done this recently and
>>>> it works with HVM and PVM as far as I can tell.
>>>>
>>>> Shivaram
>>>>
>>>>
>>>> On Tue, Apr 8, 2014 at 8:50 AM, Marco Costantini <
>>>> silvio.costant...@granatads.com> wrote:
>>>>
>>>>> I was able to keep the "workaround" ...around... by overwriting the
>>>>> generated '/root/.ssh/authorized_keys' file with a known good one, in the
>>>>> '/etc/rc.local' file
>>>>>
>>>>>
>>>>> On Tue, Apr 8, 2014 at 10:12 AM, Marco Costantini <
>>>>> silvio.costant...@granatads.com> wrote:
>>>>>
>>>>>> Another thing I didn't mention. The AMI and user used: naturally I've
>>>>>> created several of my own AMIs with the following characteristics. None 
>>>>>> of
>>>>>> which worked.
>>>>>>
>>>>>> 1) Enabling ssh as root as per this guide (
>>>>>> http://blog.tiger-workshop.com/enable-root-access-on-amazon-ec2-instance/).
>>>>>> When doing this, I do not specify a user for the spark-ec2 script. What
>>>>>> happens is that, it works! But only while it's alive. If I stop the
>>>>>> instance, create an AMI, and launch a new instance based from the new 
>>>>>> AMI,
>>>>>> the change I made in the '/root/.ssh/authorized_keys' file is overwritten
>>>>>>
>>>>>> 2) adding the 'ec2-user' to the 'root' group. This means that the
>>>>>> ec2-user does not have to use sudo to perform any operations needing root
>>>>>> privilidges. When doing this, I specify the user 'ec2-user' for the
>>>>>> spark-ec2 script. An error occurs: rsync fails with exit code 23.
>>>>>>
>>>>>> I believe HVMs still work. But it would be valuable to the community
>>>>>> to know that the root user work-around does/doesn't work any more for
>>>>>> paravirtual instances.
>>>>>>
>>>>>> Thanks,
>>>>>> Marco.
>>>>>>
>>>>>>
>>>>>> On Tue, Apr 8, 2014 at 9:51 AM, Marco Costantini <
>>>>>> silvio.costant...@granatads.com> wrote:
>>>>>>
>>>>>>> As requested, here is the script I am running. It is a simple shell
>>>>>>> script which calls spark-ec2 wrapper script. I execute it from the 'ec2'
>>>>>>> directory of spark, as usual. The AMI used is the raw one from the AWS
>>>>>>> Quick Start section. It is the first option (an Amazon Linux paravirtual
>>>>>>> image). Any ideas or confirmation would be GREATLY appreciated. Please 
>>>>>>> and
>>>>>>> thank you.
>>>>>>>
>>>>>>>
>>>>>>> #!/bin/sh
>>>>>>>
>>>>>>> export AWS_ACCESS_KEY_ID=MyCensoredKey
>>>>>>> export AWS_SECRET_ACCESS_KEY=MyCensoredKey
>>>>>>>
>>>>>>> AMI_ID=ami-2f726546
>>>>>>>
>>>>>>> ./spark-ec2 -k gds-generic -i ~/.ssh/gds-generic.pem -u ec2-user -s
>>>>>>> 10 -v 0.9.0 -w 300 --no-ganglia -a ${AMI_ID} -m m3.2xlarge -t m3.2xlarge
>>>>>>> launch marcotest
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> On Mon, Apr 7, 2014 at 6:21 PM, Shivaram Venkataraman <
>>>>>>> shivaram.venkatara...@gmail.com> wrote:
>>>>>>>
>>>>>>>> Hmm -- That is strange. Can you paste the command you are using to
>>>>>>>> launch the instances ? The typical workflow is to use the spark-ec2 
>>>>>>>> wrapper
>>>>>>>> script using the guidelines at
>>>>>>>> http://spark.apache.org/docs/latest/ec2-scripts.html
>>>>>>>>
>>>>>>>> Shivaram
>>>>>>>>
>>>>>>>>
>>>>>>>> On Mon, Apr 7, 2014 at 1:53 PM, Marco Costantini <
>>>>>>>> silvio.costant...@granatads.com> wrote:
>>>>>>>>
>>>>>>>>> Hi Shivaram,
>>>>>>>>>
>>>>>>>>> OK so let's assume the script CANNOT take a different user and
>>>>>>>>> that it must be 'root'. The typical workaround is as you said, allow 
>>>>>>>>> the
>>>>>>>>> ssh with the root user. Now, don't laugh, but, this worked last 
>>>>>>>>> Friday, but
>>>>>>>>> today (Monday) it no longer works. :D Why? ...
>>>>>>>>>
>>>>>>>>> ...It seems that NOW, when you launch a 'paravirtual' ami, the
>>>>>>>>> root user's 'authorized_keys' file is always overwritten. This means 
>>>>>>>>> the
>>>>>>>>> workaround doesn't work anymore! I would LOVE for someone to verify 
>>>>>>>>> this.
>>>>>>>>>
>>>>>>>>> Just to point out, I am trying to make this work with a
>>>>>>>>> paravirtual instance and not an HVM instance.
>>>>>>>>>
>>>>>>>>> Please and thanks,
>>>>>>>>> Marco.
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> On Mon, Apr 7, 2014 at 4:40 PM, Shivaram Venkataraman <
>>>>>>>>> shivaram.venkatara...@gmail.com> wrote:
>>>>>>>>>
>>>>>>>>>> Right now the spark-ec2 scripts assume that you have root access
>>>>>>>>>> and a lot of internal scripts assume have the user's home directory 
>>>>>>>>>> hard
>>>>>>>>>> coded as /root.   However all the Spark AMIs we build should have 
>>>>>>>>>> root ssh
>>>>>>>>>> access -- Do you find this not to be the case ?
>>>>>>>>>>
>>>>>>>>>> You can also enable root ssh access in a vanilla AMI by editing
>>>>>>>>>> /etc/ssh/sshd_config and setting "PermitRootLogin" to yes
>>>>>>>>>>
>>>>>>>>>> Thanks
>>>>>>>>>> Shivaram
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> On Mon, Apr 7, 2014 at 11:14 AM, Marco Costantini <
>>>>>>>>>> silvio.costant...@granatads.com> wrote:
>>>>>>>>>>
>>>>>>>>>>> Hi all,
>>>>>>>>>>> On the old Amazon Linux EC2 images, the user 'root' was enabled
>>>>>>>>>>> for ssh. Also, it is the default user for the Spark-EC2 script.
>>>>>>>>>>>
>>>>>>>>>>> Currently, the Amazon Linux images have an 'ec2-user' set up for
>>>>>>>>>>> ssh instead of 'root'.
>>>>>>>>>>>
>>>>>>>>>>> I can see that the Spark-EC2 script allows you to specify which
>>>>>>>>>>> user to log in with, but even when I change this, the script fails 
>>>>>>>>>>> for
>>>>>>>>>>> various reasons. And the output SEEMS that the script is still 
>>>>>>>>>>> based on the
>>>>>>>>>>> specified user's home directory being '/root'.
>>>>>>>>>>>
>>>>>>>>>>> Am I using this script wrong?
>>>>>>>>>>> Has anyone had success with this 'ec2-user' user?
>>>>>>>>>>> Any ideas?
>>>>>>>>>>>
>>>>>>>>>>> Please and thank you,
>>>>>>>>>>> Marco.
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>
>>>>>>>>
>>>>>>>
>>>>>>
>>>>>
>>>>
>>>
>>
>

Reply via email to