And for the record, that AMI is ami-35b1885c. Again, you don't need to specify it explicitly; spark-ec2 will default to it.
On Wed, Apr 9, 2014 at 11:08 AM, Nicholas Chammas < nicholas.cham...@gmail.com> wrote: > Marco, > > If you call spark-ec2 launch without specifying an AMI, it will default to > the Spark-provided AMI. > > Nick > > > On Wed, Apr 9, 2014 at 9:43 AM, Marco Costantini < > silvio.costant...@granatads.com> wrote: > >> Hi there, >> To answer your question; no there is no reason NOT to use an AMI that >> Spark has prepared. The reason we haven't is that we were not aware such >> AMIs existed. Would you kindly point us to the documentation where we can >> read about this further? >> >> Many many thanks, Shivaram. >> Marco. >> >> >> On Tue, Apr 8, 2014 at 4:42 PM, Shivaram Venkataraman < >> shiva...@eecs.berkeley.edu> wrote: >> >>> Is there any reason why you want to start with a vanilla amazon AMI >>> rather than the ones we build and provide as a part of Spark EC2 scripts ? >>> The AMIs we provide are close to the vanilla AMI but have the root account >>> setup properly and install packages like java that are used by Spark. >>> >>> If you wish to customize the AMI, you could always start with our AMI >>> and add more packages you like -- I have definitely done this recently and >>> it works with HVM and PVM as far as I can tell. >>> >>> Shivaram >>> >>> >>> On Tue, Apr 8, 2014 at 8:50 AM, Marco Costantini < >>> silvio.costant...@granatads.com> wrote: >>> >>>> I was able to keep the "workaround" ...around... by overwriting the >>>> generated '/root/.ssh/authorized_keys' file with a known good one, in the >>>> '/etc/rc.local' file >>>> >>>> >>>> On Tue, Apr 8, 2014 at 10:12 AM, Marco Costantini < >>>> silvio.costant...@granatads.com> wrote: >>>> >>>>> Another thing I didn't mention. The AMI and user used: naturally I've >>>>> created several of my own AMIs with the following characteristics. None of >>>>> which worked. >>>>> >>>>> 1) Enabling ssh as root as per this guide ( >>>>> http://blog.tiger-workshop.com/enable-root-access-on-amazon-ec2-instance/). >>>>> When doing this, I do not specify a user for the spark-ec2 script. What >>>>> happens is that, it works! But only while it's alive. If I stop the >>>>> instance, create an AMI, and launch a new instance based from the new AMI, >>>>> the change I made in the '/root/.ssh/authorized_keys' file is overwritten >>>>> >>>>> 2) adding the 'ec2-user' to the 'root' group. This means that the >>>>> ec2-user does not have to use sudo to perform any operations needing root >>>>> privilidges. When doing this, I specify the user 'ec2-user' for the >>>>> spark-ec2 script. An error occurs: rsync fails with exit code 23. >>>>> >>>>> I believe HVMs still work. But it would be valuable to the community >>>>> to know that the root user work-around does/doesn't work any more for >>>>> paravirtual instances. >>>>> >>>>> Thanks, >>>>> Marco. >>>>> >>>>> >>>>> On Tue, Apr 8, 2014 at 9:51 AM, Marco Costantini < >>>>> silvio.costant...@granatads.com> wrote: >>>>> >>>>>> As requested, here is the script I am running. It is a simple shell >>>>>> script which calls spark-ec2 wrapper script. I execute it from the 'ec2' >>>>>> directory of spark, as usual. The AMI used is the raw one from the AWS >>>>>> Quick Start section. It is the first option (an Amazon Linux paravirtual >>>>>> image). Any ideas or confirmation would be GREATLY appreciated. Please >>>>>> and >>>>>> thank you. >>>>>> >>>>>> >>>>>> #!/bin/sh >>>>>> >>>>>> export AWS_ACCESS_KEY_ID=MyCensoredKey >>>>>> export AWS_SECRET_ACCESS_KEY=MyCensoredKey >>>>>> >>>>>> AMI_ID=ami-2f726546 >>>>>> >>>>>> ./spark-ec2 -k gds-generic -i ~/.ssh/gds-generic.pem -u ec2-user -s >>>>>> 10 -v 0.9.0 -w 300 --no-ganglia -a ${AMI_ID} -m m3.2xlarge -t m3.2xlarge >>>>>> launch marcotest >>>>>> >>>>>> >>>>>> >>>>>> On Mon, Apr 7, 2014 at 6:21 PM, Shivaram Venkataraman < >>>>>> shivaram.venkatara...@gmail.com> wrote: >>>>>> >>>>>>> Hmm -- That is strange. Can you paste the command you are using to >>>>>>> launch the instances ? The typical workflow is to use the spark-ec2 >>>>>>> wrapper >>>>>>> script using the guidelines at >>>>>>> http://spark.apache.org/docs/latest/ec2-scripts.html >>>>>>> >>>>>>> Shivaram >>>>>>> >>>>>>> >>>>>>> On Mon, Apr 7, 2014 at 1:53 PM, Marco Costantini < >>>>>>> silvio.costant...@granatads.com> wrote: >>>>>>> >>>>>>>> Hi Shivaram, >>>>>>>> >>>>>>>> OK so let's assume the script CANNOT take a different user and that >>>>>>>> it must be 'root'. The typical workaround is as you said, allow the ssh >>>>>>>> with the root user. Now, don't laugh, but, this worked last Friday, but >>>>>>>> today (Monday) it no longer works. :D Why? ... >>>>>>>> >>>>>>>> ...It seems that NOW, when you launch a 'paravirtual' ami, the root >>>>>>>> user's 'authorized_keys' file is always overwritten. This means the >>>>>>>> workaround doesn't work anymore! I would LOVE for someone to verify >>>>>>>> this. >>>>>>>> >>>>>>>> Just to point out, I am trying to make this work with a paravirtual >>>>>>>> instance and not an HVM instance. >>>>>>>> >>>>>>>> Please and thanks, >>>>>>>> Marco. >>>>>>>> >>>>>>>> >>>>>>>> On Mon, Apr 7, 2014 at 4:40 PM, Shivaram Venkataraman < >>>>>>>> shivaram.venkatara...@gmail.com> wrote: >>>>>>>> >>>>>>>>> Right now the spark-ec2 scripts assume that you have root access >>>>>>>>> and a lot of internal scripts assume have the user's home directory >>>>>>>>> hard >>>>>>>>> coded as /root. However all the Spark AMIs we build should have >>>>>>>>> root ssh >>>>>>>>> access -- Do you find this not to be the case ? >>>>>>>>> >>>>>>>>> You can also enable root ssh access in a vanilla AMI by editing >>>>>>>>> /etc/ssh/sshd_config and setting "PermitRootLogin" to yes >>>>>>>>> >>>>>>>>> Thanks >>>>>>>>> Shivaram >>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> On Mon, Apr 7, 2014 at 11:14 AM, Marco Costantini < >>>>>>>>> silvio.costant...@granatads.com> wrote: >>>>>>>>> >>>>>>>>>> Hi all, >>>>>>>>>> On the old Amazon Linux EC2 images, the user 'root' was enabled >>>>>>>>>> for ssh. Also, it is the default user for the Spark-EC2 script. >>>>>>>>>> >>>>>>>>>> Currently, the Amazon Linux images have an 'ec2-user' set up for >>>>>>>>>> ssh instead of 'root'. >>>>>>>>>> >>>>>>>>>> I can see that the Spark-EC2 script allows you to specify which >>>>>>>>>> user to log in with, but even when I change this, the script fails >>>>>>>>>> for >>>>>>>>>> various reasons. And the output SEEMS that the script is still based >>>>>>>>>> on the >>>>>>>>>> specified user's home directory being '/root'. >>>>>>>>>> >>>>>>>>>> Am I using this script wrong? >>>>>>>>>> Has anyone had success with this 'ec2-user' user? >>>>>>>>>> Any ideas? >>>>>>>>>> >>>>>>>>>> Please and thank you, >>>>>>>>>> Marco. >>>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>> >>>>>>> >>>>>> >>>>> >>>> >>> >> >