Argh -- I just started having a similar problem with whirr-0.7.0 pulling
from the cdh3u2 repo and installing the basic hadoop stack:


Successfully executed configure script: [output=Reading package lists...
Building dependency tree...
Reading state information...
Reading package lists...
Building dependency tree...
Reading state information...
,
error=/tmp/configure-hadoop-datanode_hadoop-tasktracker/configure-hadoop-datanode_hadoop-tasktracker.sh:
line 73: /etc/hadoop-0.20/conf.dist/hadoop-metrics.properties: No such file
or directory
chgrp: invalid group: `hadoop'
chgrp: invalid group: `hadoop'
chgrp: invalid group: `hadoop'
chmod: missing operand after `/var/log/hadoop/logs'
Try `chmod --help' for more information.
E: Couldn't find package hadoop-0.20-datanode
hadoop-0.20-datanode: unrecognized service
E: Couldn't find package hadoop-0.20-tasktracker
hadoop-0.20-tasktracker: unrecognized service
, exitCode=0]


Same whirr config I've been using for a while -- this just started
happening to me today.  Three clusters in a row failed in this way.



On Fri, Feb 17, 2012 at 10:49 AM, Andrei Savu <[email protected]> wrote:

> The trunk should work just fine. I think in your case the download is
> failing for Hadoop or for Mahout.
>
>
> On Fri, Feb 17, 2012 at 6:33 PM, Frank Scholten <[email protected]>wrote:
>
>> Hi all,
>>
>> I am having trouble starting a Hadoop / Mahout cluster with Whirr
>> trunk, commit 44fb39fc8.
>>
>> Several errors are reported. The first one is:
>>
>> Bootstrapping cluster
>> Configuring template
>> Starting 1 node(s) with roles [hadoop-jobtracker, hadoop-namenode,
>> mahout-client]
>> Configuring template
>> Starting 4 node(s) with roles [hadoop-datanode, hadoop-tasktracker]
>> Dying because - net.schmizz.sshj.transport.TransportException: Broken
>> transport; encountered EOF
>> Dying because - net.schmizz.sshj.transport.TransportException: Broken
>> transport; encountered EOF
>> <<
>> (ubuntu:rsa[fingerprint(af:e3:53:27:e0:12:18:54:1c:fc:3b:24:b9:18:39:10),sha1(83:6a:70:2f:c2:d5:3d:e0:05:7a:4a:e5:1a:51:67:dc:2b:56:62:18)]@
>> 50.17.130.132:22)
>> error acquiring SSHClient(timeout=60000) (attempt 1 of 7): Socket
>> closed
>>
>> This repeats several times until I get a stacktrace
>>
>>  call get() on this exception to get access to the task in progress
>>        at
>> org.jclouds.compute.callables.BlockUntilInitScriptStatusIsZeroThenReturnOutput.get(BlockUntilInitScriptStatusIsZeroThenReturnOutput.java:195)
>>        at
>> org.jclouds.compute.callables.RunScriptOnNodeAsInitScriptUsingSshAndBlockUntilComplete.doCall(RunScriptOnNodeAsInitScriptUsingSshAndBlockUntilComplete.java:60)
>>        ... 8 more
>>
>> which is also repeated for several roles
>>
>> and at the end I get
>>
>> Successfully executed configure script: [output=, error=chown: invalid
>> user: `hadoop:hadoop'
>> cp: target `/usr/local/hadoop/conf' is not a directory
>> cp: cannot create regular file `/usr/local/hadoop/conf': No such file
>> or directory
>> chown: invalid user: `hadoop:hadoop'
>> chown: invalid user: `hadoop:hadoop'
>> chown: invalid user: `hadoop:hadoop'
>> Unknown id: hadoop
>> Unknown id: hadoop
>> , exitCode=0]
>>
>> for several roles.
>>
>> Has something changed recently that caused this problem?
>>
>> Cheers,
>>
>> Frank
>>
>
>

Reply via email to