And for the record, that AMI is ami-35b1885c. Again, you don't need to
specify it explicitly; spark-ec2 will default to it.


On Wed, Apr 9, 2014 at 11:08 AM, Nicholas Chammas <
nicholas.cham...@gmail.com> wrote:

> Marco,
>
> If you call spark-ec2 launch without specifying an AMI, it will default to
> the Spark-provided AMI.
>
> Nick
>
>
> On Wed, Apr 9, 2014 at 9:43 AM, Marco Costantini <
> silvio.costant...@granatads.com> wrote:
>
>> Hi there,
>> To answer your question; no there is no reason NOT to use an AMI that
>> Spark has prepared. The reason we haven't is that we were not aware such
>> AMIs existed. Would you kindly point us to the documentation where we can
>> read about this further?
>>
>> Many many thanks, Shivaram.
>> Marco.
>>
>>
>> On Tue, Apr 8, 2014 at 4:42 PM, Shivaram Venkataraman <
>> shiva...@eecs.berkeley.edu> wrote:
>>
>>> Is there any reason why you want to start with a vanilla amazon AMI
>>> rather than the ones we build and provide as a part of Spark EC2 scripts ?
>>> The AMIs we provide are close to the vanilla AMI but have the root account
>>> setup properly and install packages like java that are used by Spark.
>>>
>>> If you wish to customize the AMI, you could always start with our AMI
>>> and add more packages you like -- I have definitely done this recently and
>>> it works with HVM and PVM as far as I can tell.
>>>
>>> Shivaram
>>>
>>>
>>> On Tue, Apr 8, 2014 at 8:50 AM, Marco Costantini <
>>> silvio.costant...@granatads.com> wrote:
>>>
>>>> I was able to keep the "workaround" ...around... by overwriting the
>>>> generated '/root/.ssh/authorized_keys' file with a known good one, in the
>>>> '/etc/rc.local' file
>>>>
>>>>
>>>> On Tue, Apr 8, 2014 at 10:12 AM, Marco Costantini <
>>>> silvio.costant...@granatads.com> wrote:
>>>>
>>>>> Another thing I didn't mention. The AMI and user used: naturally I've
>>>>> created several of my own AMIs with the following characteristics. None of
>>>>> which worked.
>>>>>
>>>>> 1) Enabling ssh as root as per this guide (
>>>>> http://blog.tiger-workshop.com/enable-root-access-on-amazon-ec2-instance/).
>>>>> When doing this, I do not specify a user for the spark-ec2 script. What
>>>>> happens is that, it works! But only while it's alive. If I stop the
>>>>> instance, create an AMI, and launch a new instance based from the new AMI,
>>>>> the change I made in the '/root/.ssh/authorized_keys' file is overwritten
>>>>>
>>>>> 2) adding the 'ec2-user' to the 'root' group. This means that the
>>>>> ec2-user does not have to use sudo to perform any operations needing root
>>>>> privilidges. When doing this, I specify the user 'ec2-user' for the
>>>>> spark-ec2 script. An error occurs: rsync fails with exit code 23.
>>>>>
>>>>> I believe HVMs still work. But it would be valuable to the community
>>>>> to know that the root user work-around does/doesn't work any more for
>>>>> paravirtual instances.
>>>>>
>>>>> Thanks,
>>>>> Marco.
>>>>>
>>>>>
>>>>> On Tue, Apr 8, 2014 at 9:51 AM, Marco Costantini <
>>>>> silvio.costant...@granatads.com> wrote:
>>>>>
>>>>>> As requested, here is the script I am running. It is a simple shell
>>>>>> script which calls spark-ec2 wrapper script. I execute it from the 'ec2'
>>>>>> directory of spark, as usual. The AMI used is the raw one from the AWS
>>>>>> Quick Start section. It is the first option (an Amazon Linux paravirtual
>>>>>> image). Any ideas or confirmation would be GREATLY appreciated. Please 
>>>>>> and
>>>>>> thank you.
>>>>>>
>>>>>>
>>>>>> #!/bin/sh
>>>>>>
>>>>>> export AWS_ACCESS_KEY_ID=MyCensoredKey
>>>>>> export AWS_SECRET_ACCESS_KEY=MyCensoredKey
>>>>>>
>>>>>> AMI_ID=ami-2f726546
>>>>>>
>>>>>> ./spark-ec2 -k gds-generic -i ~/.ssh/gds-generic.pem -u ec2-user -s
>>>>>> 10 -v 0.9.0 -w 300 --no-ganglia -a ${AMI_ID} -m m3.2xlarge -t m3.2xlarge
>>>>>> launch marcotest
>>>>>>
>>>>>>
>>>>>>
>>>>>> On Mon, Apr 7, 2014 at 6:21 PM, Shivaram Venkataraman <
>>>>>> shivaram.venkatara...@gmail.com> wrote:
>>>>>>
>>>>>>> Hmm -- That is strange. Can you paste the command you are using to
>>>>>>> launch the instances ? The typical workflow is to use the spark-ec2 
>>>>>>> wrapper
>>>>>>> script using the guidelines at
>>>>>>> http://spark.apache.org/docs/latest/ec2-scripts.html
>>>>>>>
>>>>>>> Shivaram
>>>>>>>
>>>>>>>
>>>>>>> On Mon, Apr 7, 2014 at 1:53 PM, Marco Costantini <
>>>>>>> silvio.costant...@granatads.com> wrote:
>>>>>>>
>>>>>>>> Hi Shivaram,
>>>>>>>>
>>>>>>>> OK so let's assume the script CANNOT take a different user and that
>>>>>>>> it must be 'root'. The typical workaround is as you said, allow the ssh
>>>>>>>> with the root user. Now, don't laugh, but, this worked last Friday, but
>>>>>>>> today (Monday) it no longer works. :D Why? ...
>>>>>>>>
>>>>>>>> ...It seems that NOW, when you launch a 'paravirtual' ami, the root
>>>>>>>> user's 'authorized_keys' file is always overwritten. This means the
>>>>>>>> workaround doesn't work anymore! I would LOVE for someone to verify 
>>>>>>>> this.
>>>>>>>>
>>>>>>>> Just to point out, I am trying to make this work with a paravirtual
>>>>>>>> instance and not an HVM instance.
>>>>>>>>
>>>>>>>> Please and thanks,
>>>>>>>> Marco.
>>>>>>>>
>>>>>>>>
>>>>>>>> On Mon, Apr 7, 2014 at 4:40 PM, Shivaram Venkataraman <
>>>>>>>> shivaram.venkatara...@gmail.com> wrote:
>>>>>>>>
>>>>>>>>> Right now the spark-ec2 scripts assume that you have root access
>>>>>>>>> and a lot of internal scripts assume have the user's home directory 
>>>>>>>>> hard
>>>>>>>>> coded as /root.   However all the Spark AMIs we build should have 
>>>>>>>>> root ssh
>>>>>>>>> access -- Do you find this not to be the case ?
>>>>>>>>>
>>>>>>>>> You can also enable root ssh access in a vanilla AMI by editing
>>>>>>>>> /etc/ssh/sshd_config and setting "PermitRootLogin" to yes
>>>>>>>>>
>>>>>>>>> Thanks
>>>>>>>>> Shivaram
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> On Mon, Apr 7, 2014 at 11:14 AM, Marco Costantini <
>>>>>>>>> silvio.costant...@granatads.com> wrote:
>>>>>>>>>
>>>>>>>>>> Hi all,
>>>>>>>>>> On the old Amazon Linux EC2 images, the user 'root' was enabled
>>>>>>>>>> for ssh. Also, it is the default user for the Spark-EC2 script.
>>>>>>>>>>
>>>>>>>>>> Currently, the Amazon Linux images have an 'ec2-user' set up for
>>>>>>>>>> ssh instead of 'root'.
>>>>>>>>>>
>>>>>>>>>> I can see that the Spark-EC2 script allows you to specify which
>>>>>>>>>> user to log in with, but even when I change this, the script fails 
>>>>>>>>>> for
>>>>>>>>>> various reasons. And the output SEEMS that the script is still based 
>>>>>>>>>> on the
>>>>>>>>>> specified user's home directory being '/root'.
>>>>>>>>>>
>>>>>>>>>> Am I using this script wrong?
>>>>>>>>>> Has anyone had success with this 'ec2-user' user?
>>>>>>>>>> Any ideas?
>>>>>>>>>>
>>>>>>>>>> Please and thank you,
>>>>>>>>>> Marco.
>>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>
>>>>>>>
>>>>>>
>>>>>
>>>>
>>>
>>
>

Reply via email to