Perfect. Now I know what to do. Thanks to your help! Many thanks, Marco.
On Wed, Apr 9, 2014 at 12:27 PM, Shivaram Venkataraman < shiva...@eecs.berkeley.edu> wrote: > The AMI should automatically switch between PVM and HVM based on the > instance type you specify on the command line. For reference (note you > don't need to specify this on the command line), the PVM ami id > is ami-5bb18832 in us-east-1. > > FWIW we maintain the list of AMI Ids (across regions and pvm, hvm) at > https://github.com/mesos/spark-ec2/tree/v2/ami-list > > Thanks > Shivaram > > > On Wed, Apr 9, 2014 at 9:12 AM, Marco Costantini < > silvio.costant...@granatads.com> wrote: > >> Ah, tried that. I believe this is an HVM AMI? We are exploring >> paravirtual AMIs. >> >> >> On Wed, Apr 9, 2014 at 11:17 AM, Nicholas Chammas < >> nicholas.cham...@gmail.com> wrote: >> >>> And for the record, that AMI is ami-35b1885c. Again, you don't need to >>> specify it explicitly; spark-ec2 will default to it. >>> >>> >>> On Wed, Apr 9, 2014 at 11:08 AM, Nicholas Chammas < >>> nicholas.cham...@gmail.com> wrote: >>> >>>> Marco, >>>> >>>> If you call spark-ec2 launch without specifying an AMI, it will default >>>> to the Spark-provided AMI. >>>> >>>> Nick >>>> >>>> >>>> On Wed, Apr 9, 2014 at 9:43 AM, Marco Costantini < >>>> silvio.costant...@granatads.com> wrote: >>>> >>>>> Hi there, >>>>> To answer your question; no there is no reason NOT to use an AMI that >>>>> Spark has prepared. The reason we haven't is that we were not aware such >>>>> AMIs existed. Would you kindly point us to the documentation where we can >>>>> read about this further? >>>>> >>>>> Many many thanks, Shivaram. >>>>> Marco. >>>>> >>>>> >>>>> On Tue, Apr 8, 2014 at 4:42 PM, Shivaram Venkataraman < >>>>> shiva...@eecs.berkeley.edu> wrote: >>>>> >>>>>> Is there any reason why you want to start with a vanilla amazon AMI >>>>>> rather than the ones we build and provide as a part of Spark EC2 scripts >>>>>> ? >>>>>> The AMIs we provide are close to the vanilla AMI but have the root >>>>>> account >>>>>> setup properly and install packages like java that are used by Spark. >>>>>> >>>>>> If you wish to customize the AMI, you could always start with our AMI >>>>>> and add more packages you like -- I have definitely done this recently >>>>>> and >>>>>> it works with HVM and PVM as far as I can tell. >>>>>> >>>>>> Shivaram >>>>>> >>>>>> >>>>>> On Tue, Apr 8, 2014 at 8:50 AM, Marco Costantini < >>>>>> silvio.costant...@granatads.com> wrote: >>>>>> >>>>>>> I was able to keep the "workaround" ...around... by overwriting the >>>>>>> generated '/root/.ssh/authorized_keys' file with a known good one, in >>>>>>> the >>>>>>> '/etc/rc.local' file >>>>>>> >>>>>>> >>>>>>> On Tue, Apr 8, 2014 at 10:12 AM, Marco Costantini < >>>>>>> silvio.costant...@granatads.com> wrote: >>>>>>> >>>>>>>> Another thing I didn't mention. The AMI and user used: naturally >>>>>>>> I've created several of my own AMIs with the following characteristics. >>>>>>>> None of which worked. >>>>>>>> >>>>>>>> 1) Enabling ssh as root as per this guide ( >>>>>>>> http://blog.tiger-workshop.com/enable-root-access-on-amazon-ec2-instance/). >>>>>>>> When doing this, I do not specify a user for the spark-ec2 script. What >>>>>>>> happens is that, it works! But only while it's alive. If I stop the >>>>>>>> instance, create an AMI, and launch a new instance based from the new >>>>>>>> AMI, >>>>>>>> the change I made in the '/root/.ssh/authorized_keys' file is >>>>>>>> overwritten >>>>>>>> >>>>>>>> 2) adding the 'ec2-user' to the 'root' group. This means that the >>>>>>>> ec2-user does not have to use sudo to perform any operations needing >>>>>>>> root >>>>>>>> privilidges. When doing this, I specify the user 'ec2-user' for the >>>>>>>> spark-ec2 script. An error occurs: rsync fails with exit code 23. >>>>>>>> >>>>>>>> I believe HVMs still work. But it would be valuable to the >>>>>>>> community to know that the root user work-around does/doesn't work any >>>>>>>> more >>>>>>>> for paravirtual instances. >>>>>>>> >>>>>>>> Thanks, >>>>>>>> Marco. >>>>>>>> >>>>>>>> >>>>>>>> On Tue, Apr 8, 2014 at 9:51 AM, Marco Costantini < >>>>>>>> silvio.costant...@granatads.com> wrote: >>>>>>>> >>>>>>>>> As requested, here is the script I am running. It is a simple >>>>>>>>> shell script which calls spark-ec2 wrapper script. I execute it from >>>>>>>>> the >>>>>>>>> 'ec2' directory of spark, as usual. The AMI used is the raw one from >>>>>>>>> the >>>>>>>>> AWS Quick Start section. It is the first option (an Amazon Linux >>>>>>>>> paravirtual image). Any ideas or confirmation would be GREATLY >>>>>>>>> appreciated. >>>>>>>>> Please and thank you. >>>>>>>>> >>>>>>>>> >>>>>>>>> #!/bin/sh >>>>>>>>> >>>>>>>>> export AWS_ACCESS_KEY_ID=MyCensoredKey >>>>>>>>> export AWS_SECRET_ACCESS_KEY=MyCensoredKey >>>>>>>>> >>>>>>>>> AMI_ID=ami-2f726546 >>>>>>>>> >>>>>>>>> ./spark-ec2 -k gds-generic -i ~/.ssh/gds-generic.pem -u ec2-user >>>>>>>>> -s 10 -v 0.9.0 -w 300 --no-ganglia -a ${AMI_ID} -m m3.2xlarge -t >>>>>>>>> m3.2xlarge >>>>>>>>> launch marcotest >>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> On Mon, Apr 7, 2014 at 6:21 PM, Shivaram Venkataraman < >>>>>>>>> shivaram.venkatara...@gmail.com> wrote: >>>>>>>>> >>>>>>>>>> Hmm -- That is strange. Can you paste the command you are using >>>>>>>>>> to launch the instances ? The typical workflow is to use the >>>>>>>>>> spark-ec2 >>>>>>>>>> wrapper script using the guidelines at >>>>>>>>>> http://spark.apache.org/docs/latest/ec2-scripts.html >>>>>>>>>> >>>>>>>>>> Shivaram >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> On Mon, Apr 7, 2014 at 1:53 PM, Marco Costantini < >>>>>>>>>> silvio.costant...@granatads.com> wrote: >>>>>>>>>> >>>>>>>>>>> Hi Shivaram, >>>>>>>>>>> >>>>>>>>>>> OK so let's assume the script CANNOT take a different user and >>>>>>>>>>> that it must be 'root'. The typical workaround is as you said, >>>>>>>>>>> allow the >>>>>>>>>>> ssh with the root user. Now, don't laugh, but, this worked last >>>>>>>>>>> Friday, but >>>>>>>>>>> today (Monday) it no longer works. :D Why? ... >>>>>>>>>>> >>>>>>>>>>> ...It seems that NOW, when you launch a 'paravirtual' ami, the >>>>>>>>>>> root user's 'authorized_keys' file is always overwritten. This >>>>>>>>>>> means the >>>>>>>>>>> workaround doesn't work anymore! I would LOVE for someone to verify >>>>>>>>>>> this. >>>>>>>>>>> >>>>>>>>>>> Just to point out, I am trying to make this work with a >>>>>>>>>>> paravirtual instance and not an HVM instance. >>>>>>>>>>> >>>>>>>>>>> Please and thanks, >>>>>>>>>>> Marco. >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> On Mon, Apr 7, 2014 at 4:40 PM, Shivaram Venkataraman < >>>>>>>>>>> shivaram.venkatara...@gmail.com> wrote: >>>>>>>>>>> >>>>>>>>>>>> Right now the spark-ec2 scripts assume that you have root >>>>>>>>>>>> access and a lot of internal scripts assume have the user's home >>>>>>>>>>>> directory >>>>>>>>>>>> hard coded as /root. However all the Spark AMIs we build should >>>>>>>>>>>> have root >>>>>>>>>>>> ssh access -- Do you find this not to be the case ? >>>>>>>>>>>> >>>>>>>>>>>> You can also enable root ssh access in a vanilla AMI by editing >>>>>>>>>>>> /etc/ssh/sshd_config and setting "PermitRootLogin" to yes >>>>>>>>>>>> >>>>>>>>>>>> Thanks >>>>>>>>>>>> Shivaram >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> On Mon, Apr 7, 2014 at 11:14 AM, Marco Costantini < >>>>>>>>>>>> silvio.costant...@granatads.com> wrote: >>>>>>>>>>>> >>>>>>>>>>>>> Hi all, >>>>>>>>>>>>> On the old Amazon Linux EC2 images, the user 'root' was >>>>>>>>>>>>> enabled for ssh. Also, it is the default user for the Spark-EC2 >>>>>>>>>>>>> script. >>>>>>>>>>>>> >>>>>>>>>>>>> Currently, the Amazon Linux images have an 'ec2-user' set up >>>>>>>>>>>>> for ssh instead of 'root'. >>>>>>>>>>>>> >>>>>>>>>>>>> I can see that the Spark-EC2 script allows you to specify >>>>>>>>>>>>> which user to log in with, but even when I change this, the >>>>>>>>>>>>> script fails >>>>>>>>>>>>> for various reasons. And the output SEEMS that the script is >>>>>>>>>>>>> still based on >>>>>>>>>>>>> the specified user's home directory being '/root'. >>>>>>>>>>>>> >>>>>>>>>>>>> Am I using this script wrong? >>>>>>>>>>>>> Has anyone had success with this 'ec2-user' user? >>>>>>>>>>>>> Any ideas? >>>>>>>>>>>>> >>>>>>>>>>>>> Please and thank you, >>>>>>>>>>>>> Marco. >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>> >>>>>>>>> >>>>>>>> >>>>>>> >>>>>> >>>>> >>>> >>> >> >