Tom any ideas? I will debug this tomorrow morning.
On Feb 18, 2012 11:46 PM, "Evan Pollan" <[email protected]> wrote:

> I have that in my properties file (recall that I ran into a critical
> mapreduce bug on Ubuntu that was introduced in CDH3U3).  So, it is trying
> to pull down the right bits -- i.e., I've been using cdh3u2 rather than
> cdh3u3 since u3 was released a couple of weeks ago.
>
> However, it doesn't even look like the install automation is being run at
> all.  The first thing that the register_cloudera_repo function does in
> install_cdh_hadoop is cat the CDH repo information into
> /etc/apt/sources.list.d/cloudera.list.  Well, that file doesn't exist on my
> systems.
>
> What is going on!?
>
>
>
> On Sat, Feb 18, 2012 at 3:37 PM, Andrei Savu <[email protected]>wrote:
>
>> Cloudera just released CDH4 - this failure may be related to that. Can
>> you try to specify whirr.env.repo=cdh3u2 ?
>>  On Feb 18, 2012 11:16 PM, "Evan Pollan" <[email protected]> wrote:
>>
>>> Argh -- I just started having a similar problem with whirr-0.7.0 pulling
>>> from the cdh3u2 repo and installing the basic hadoop stack:
>>>
>>>
>>> Successfully executed configure script: [output=Reading package lists...
>>> Building dependency tree...
>>> Reading state information...
>>> Reading package lists...
>>> Building dependency tree...
>>> Reading state information...
>>> ,
>>> error=/tmp/configure-hadoop-datanode_hadoop-tasktracker/configure-hadoop-datanode_hadoop-tasktracker.sh:
>>> line 73: /etc/hadoop-0.20/conf.dist/hadoop-metrics.properties: No such file
>>> or directory
>>> chgrp: invalid group: `hadoop'
>>> chgrp: invalid group: `hadoop'
>>> chgrp: invalid group: `hadoop'
>>> chmod: missing operand after `/var/log/hadoop/logs'
>>> Try `chmod --help' for more information.
>>> E: Couldn't find package hadoop-0.20-datanode
>>> hadoop-0.20-datanode: unrecognized service
>>> E: Couldn't find package hadoop-0.20-tasktracker
>>> hadoop-0.20-tasktracker: unrecognized service
>>> , exitCode=0]
>>>
>>>
>>> Same whirr config I've been using for a while -- this just started
>>> happening to me today.  Three clusters in a row failed in this way.
>>>
>>>
>>>
>>> On Fri, Feb 17, 2012 at 10:49 AM, Andrei Savu <[email protected]>wrote:
>>>
>>>> The trunk should work just fine. I think in your case the download is
>>>> failing for Hadoop or for Mahout.
>>>>
>>>>
>>>> On Fri, Feb 17, 2012 at 6:33 PM, Frank Scholten <[email protected]
>>>> > wrote:
>>>>
>>>>> Hi all,
>>>>>
>>>>> I am having trouble starting a Hadoop / Mahout cluster with Whirr
>>>>> trunk, commit 44fb39fc8.
>>>>>
>>>>> Several errors are reported. The first one is:
>>>>>
>>>>> Bootstrapping cluster
>>>>> Configuring template
>>>>> Starting 1 node(s) with roles [hadoop-jobtracker, hadoop-namenode,
>>>>> mahout-client]
>>>>> Configuring template
>>>>> Starting 4 node(s) with roles [hadoop-datanode, hadoop-tasktracker]
>>>>> Dying because - net.schmizz.sshj.transport.TransportException: Broken
>>>>> transport; encountered EOF
>>>>> Dying because - net.schmizz.sshj.transport.TransportException: Broken
>>>>> transport; encountered EOF
>>>>> <<
>>>>> (ubuntu:rsa[fingerprint(af:e3:53:27:e0:12:18:54:1c:fc:3b:24:b9:18:39:10),sha1(83:6a:70:2f:c2:d5:3d:e0:05:7a:4a:e5:1a:51:67:dc:2b:56:62:18)]@
>>>>> 50.17.130.132:22)
>>>>> error acquiring SSHClient(timeout=60000) (attempt 1 of 7): Socket
>>>>> closed
>>>>>
>>>>> This repeats several times until I get a stacktrace
>>>>>
>>>>>  call get() on this exception to get access to the task in progress
>>>>>        at
>>>>> org.jclouds.compute.callables.BlockUntilInitScriptStatusIsZeroThenReturnOutput.get(BlockUntilInitScriptStatusIsZeroThenReturnOutput.java:195)
>>>>>        at
>>>>> org.jclouds.compute.callables.RunScriptOnNodeAsInitScriptUsingSshAndBlockUntilComplete.doCall(RunScriptOnNodeAsInitScriptUsingSshAndBlockUntilComplete.java:60)
>>>>>        ... 8 more
>>>>>
>>>>> which is also repeated for several roles
>>>>>
>>>>> and at the end I get
>>>>>
>>>>> Successfully executed configure script: [output=, error=chown: invalid
>>>>> user: `hadoop:hadoop'
>>>>> cp: target `/usr/local/hadoop/conf' is not a directory
>>>>> cp: cannot create regular file `/usr/local/hadoop/conf': No such file
>>>>> or directory
>>>>> chown: invalid user: `hadoop:hadoop'
>>>>> chown: invalid user: `hadoop:hadoop'
>>>>> chown: invalid user: `hadoop:hadoop'
>>>>> Unknown id: hadoop
>>>>> Unknown id: hadoop
>>>>> , exitCode=0]
>>>>>
>>>>> for several roles.
>>>>>
>>>>> Has something changed recently that caused this problem?
>>>>>
>>>>> Cheers,
>>>>>
>>>>> Frank
>>>>>
>>>>
>>>>
>>>
>

Reply via email to