Inline.

On Wed, Jan 11, 2012 at 3:52 AM, Madhu Ramanna <ma...@buysight.com> wrote:

> So, launching the cluster with non-root user was successful. Namenode UI
> url was printed to the logs.
>
> However, namenode was shutdown
>
> No directory, logging in with HOME=/
> No directory, logging in with HOME=/
> No directory, logging in with HOME=/
> No directory, logging in with HOME=/
> , error=************************************************************/
> 12/01/11 01:20:52 INFO namenode.FSNamesystem: fsOwner=hadoop,hadoop
> 12/01/11 01:20:52 INFO namenode.FSNamesystem: supergroup=supergroup
> 12/01/11 01:20:52 INFO namenode.FSNamesystem: isPermissionEnabled=true
> 12/01/11 01:20:52 INFO common.Storage: Image file of size 96 saved in 0
> seconds.
> 12/01/11 01:20:52 INFO common.Storage: Storage directory
> /data/hadoop/hdfs/name has been successfully formatted.
> 12/01/11 01:20:52 INFO namenode.NameNode: SHUTDOWN_MSG:
> /************************************************************
> SHUTDOWN_MSG: Shutting down NameNode at ip-10-8-39-133.ec2.internal/
> 10.8.39.133
> ************************************************************/
> , exitCode=0]
>

This is the expected message you get when formatting HDFS (first time
only).


>
>
> Not able to access the UI from browser.
>
> Any ideas ?
>

Have you tried to login on the remote machines? Are the daemons running as
expected? (check with jps)


>
> I want to setup a long running hadoop cluster that is never down. What is
> the recommended whirr setup for this configuration ?
>

Right now Whirr is best for short-lived clusters for periodical data
processing. Are you building something like a warehouse in Amazon with
Hadoop? Why not push data in S3 and only start Hadoop for processing?


>
> Thanks,
> Madhu
>
>
>
> From: Madhu Ramanna <ma...@buysight.com>
> Date: Tue, 10 Jan 2012 15:19:27 -0800
> To: Alex Heneveld <alex.henev...@cloudsoftcorp.com>, Madhu Ramanna <
> ma...@buysight.com>
>
> Cc: "user@whirr.apache.org" <user@whirr.apache.org>
> Subject: Re: [newbie] Unable to setup hadoop cluster on ec2
>
> Will give it a shot then
>
> Thanks,
> Madhu
>
> From: Alex Heneveld <alex.henev...@cloudsoftcorp.com>
> Date: Tue, 10 Jan 2012 15:15:37 -0800
> To: Madhu Ramanna <ma...@buysight.com>
> Cc: "user@whirr.apache.org" <user@whirr.apache.org>
> Subject: Re: [newbie] Unable to setup hadoop cluster on ec2
>
>
> Hi Madhu,
>
> Looks like you might be running as root -- there are known issues with
> this.  Does it work if you are a different user?
>
> Best,
> Alex
>
>
>
> On 10/01/2012 23:10, Madhu Ramanna wrote:
>
> Here you go:
>
>  whirr.cluster-name=test_bs_hadoop_cluster
> whirr.instance-templates=1 hadoop-jobtracker+hadoop-namenode,1
> hadoop-datanode+hadoop-tasktracker
> whirr.provider=aws-ec2
> whirr.identity=${env:AWS_ACCESS_KEY_ID}
> whirr.credential=${env:AWS_SECRET_ACCESS_KEY}
> whirr.private-key-file=/root/.ssh/id_rsa_whirr
> whirr.public-key-file=/root/.ssh/id_rsa_whirr.pub
> whirr.cluster-user=tuser
>
>  Thanks,
> Madhu
>
>   From: Andrei Savu <savu.and...@gmail.com>
> Reply-To: "user@whirr.apache.org" <user@whirr.apache.org>
> Date: Tue, 10 Jan 2012 14:59:17 -0800
> To: "user@whirr.apache.org" <user@whirr.apache.org>
> Subject: Re: [newbie] Unable to setup hadoop cluster on ec2
>
>  Can you share the .properties file you are using? I want to give it a
> try on my machine.
>
>  Thanks,
>
> -- Andrei Savu / andreisavu.ro
>
> On Wed, Jan 11, 2012 at 12:55 AM, Madhu Ramanna <ma...@buysight.com>wrote:
>
>>  Hello everyone,
>>
>>  I wanted to setup a hadoop cluster on ec2 and downloaded 0.7.0 version.
>>
>>  I setup the relevant files as per the quick start guide. However the
>> cluster setup failed with the following exceptions.  What could be going
>> wrong ? I'm sure (almost) my aws keys are correct
>>
>>  Appreciate your response !
>>
>>  2012-01-06 10:38:06,690 DEBUG [jclouds.compute] (user thread 2) <<
>> images(3758)
>> 2012-01-06 10:38:07,210 DEBUG [jclouds.compute] (pool-4-thread-2) <<
>> images(3758)
>> 2012-01-06 10:38:07,212 DEBUG [jclouds.compute] (pool-4-thread-1) >>
>> blocking on socket [address=107.21.80.86, port=22] for 600000 MILLISECONDS
>> 2012-01-06 10:38:07,212 DEBUG [jclouds.compute] (pool-4-thread-2) >>
>> blocking on socket [address=107.22.64.12, port=22] for 600000 MILLISECONDS
>> 2012-01-06 10:38:07,292 DEBUG [jclouds.compute] (pool-4-thread-2) <<
>> socket [address=107.22.64.12, port=22] opened
>> 2012-01-06 10:38:10,293 DEBUG [jclouds.compute] (pool-4-thread-1) <<
>> socket [address=107.21.80.86, port=22] opened
>> 2012-01-06 10:38:13,012 INFO
>>  [org.apache.whirr.actions.ScriptBasedClusterAction] (pool-4-thread-2)
>> configure phase script run completed on: us-east-1/i-94bccef6
>> 2012-01-06 10:38:16,003 INFO
>>  [org.apache.whirr.actions.ScriptBasedClusterAction] (pool-4-thread-1)
>> configure phase script run completed on: us-east-1/i-9abccef8
>> 2012-01-06 10:38:16,004 ERROR [org.apache.whirr.ClusterController] (main)
>> Unable to start the cluster. Terminating all nodes.
>> java.io.IOException: org.jclouds.rest.AuthorizationException: (root:
>> <snip> ) error acquiring SSHClient(timeout=60000): Exhausted available
>> authentication methods
>>         at
>> org.apache.whirr.actions.ScriptBasedClusterAction.runScripts(ScriptBasedClusterAction.java:215)
>>         at
>> org.apache.whirr.actions.ScriptBasedClusterAction.doAction(ScriptBasedClusterAction.java:128)
>>         at
>> org.apache.whirr.actions.ScriptBasedClusterAction.execute(ScriptBasedClusterAction.java:107)
>>         at
>> org.apache.whirr.ClusterController.launchCluster(ClusterController.java:109)
>>         at
>> org.apache.whirr.cli.command.LaunchClusterCommand.run(LaunchClusterCommand.java:63)
>>         at org.apache.whirr.cli.Main.run(Main.java:64)
>>         at org.apache.whirr.cli.Main.main(Main.java:97)
>> Caused by: org.jclouds.rest.AuthorizationException: (root:rsa[<snip>]@
>> 107.21.80.86:22) error acquiring SSHClient(timeout=60000): Exhausted
>> available authentication methods
>>         at
>> org.jclouds.sshj.SshjSshClient.propagate(SshjSshClient.java:413)
>>         at org.jclouds.sshj.SshjSshClient.acquire(SshjSshClient.java:244)
>>         at org.jclouds.sshj.SshjSshClient.connect(SshjSshClient.java:255)
>>         at
>> org.jclouds.compute.callables.RunScriptOnNodeAsInitScriptUsingSsh.call(RunScriptOnNodeAsInitScriptUsingSsh.java:89)
>>         at
>> org.jclouds.compute.internal.BaseComputeService.runScriptOnNode(BaseComputeService.java:612)
>>         at
>> org.apache.whirr.actions.ScriptBasedClusterAction$2.call(ScriptBasedClusterAction.java:190)
>>         at
>> org.apache.whirr.actions.ScriptBasedClusterAction$2.call(ScriptBasedClusterAction.java:178)
>>         at
>> java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303)
>>         at java.util.concurrent.FutureTask.run(FutureTask.java:138)
>>         at
>> java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886)
>>         at
>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908)
>>         at java.lang.Thread.run(Thread.java:619)
>> Caused by: net.schmizz.sshj.userauth.UserAuthException: Exhausted
>> available authentication methods
>>         at
>> net.schmizz.sshj.userauth.UserAuthImpl.authenticate(UserAuthImpl.java:114)
>>         at net.schmizz.sshj.SSHClient.auth(SSHClient.java:204)
>>         at net.schmizz.sshj.SSHClient.authPublickey(SSHClient.java:304)
>>         at net.schmizz.sshj.SSHClient.authPublickey(SSHClient.java:323)
>>         at org.jclouds.sshj.SshjSshClient$1.create(SshjSshClient.java:199)
>>         at org.jclouds.sshj.SshjSshClient$1.create(SshjSshClient.java:171)
>>         at org.jclouds.sshj.SshjSshClient.acquire(SshjSshClient.java:220)
>>         ... 10 more
>> Caused by: net.schmizz.sshj.userauth.UserAuthException: publickey auth
>> failed
>>         at
>> net.schmizz.sshj.userauth.UserAuthImpl.handle(UserAuthImpl.java:157)
>>         at
>> net.schmizz.sshj.transport.TransportImpl.handle(TransportImpl.java:474)
>>         at net.schmizz.sshj.transport.Decoder.decode(Decoder.java:127)
>>         at net.schmizz.sshj.transport.Decoder.received(Decoder.java:195)
>>         at net.schmizz.sshj.transport.Reader.run(Reader.java:72)
>> 2012-01-06 10:38:16,007 INFO  [org.apache.whirr.state.ClusterStateStore]
>> (main) Unable to load cluster state, assuming it has no running nodes.
>> java.io.FileNotFoundException:
>> /root/.whirr/test_bs_hadoop_cluster/instances (No such file or directory)
>>
>>  Thanks,
>> Madhu
>>
>>
>
>

Reply via email to