It's strange this is happening because the integration tests work as
expected (we actually running MR jobs).

Are you adding any other options?

On Thu, Feb 23, 2012 at 8:50 PM, Andrei Savu <[email protected]> wrote:

> That looks like a change we've made in
> https://issues.apache.org/jira/browse/WHIRR-490
>
> It seems like "unlimited" is not a valid value for mapred.child.ulimit.
> Let me investigate a bit more.
>
> In  the meantime you can add to your .properties file something like:
>
> hadoop-mapreduce.mapred.child.ulimit=<very-large-number>
>
>
> On Thu, Feb 23, 2012 at 8:36 PM, Edmar Ferreira <
> [email protected]> wrote:
>
>> changed it and the cluster is running and I can access the fs and submit
>> jobs, but all jobs aways fail with this strange error:
>>
>> java.lang.NumberFormatException: For input string: "unlimited"
>>      at 
>> java.lang.NumberFormatException.forInputString(NumberFormatException.java:65)
>>      at java.lang.Integer.parseInt(Integer.java:481)
>>      at java.lang.Integer.valueOf(Integer.java:570)
>>      at org.apache.hadoop.util.Shell.getUlimitMemoryCommand(Shell.java:86)
>>      at org.apache.hadoop.mapred.TaskRunner.run(TaskRunner.java:379)
>>
>>
>>
>> Also when I try to access the full error log I see this in the browser:
>>
>> HTTP ERROR: 410
>>
>> Failed to retrieve stdout log for task: attempt_201202232026_0001_m_000005_0
>>
>> RequestURI=/tasklog
>>
>>
>> My proxy is running and I'm using the socks proxy in localhost 6666
>>
>>
>>
>>
>>
>> On Thu, Feb 23, 2012 at 5:25 PM, Andrei Savu <[email protected]>wrote:
>>
>>> That should work but I recommend you to try:
>>>
>>> http://apache.osuosl.org/hadoop/common/hadoop-0.20.2/hadoop-0.20.2.tar.gz
>>>
>>> archive.apache.org  is extremely unreliable.
>>>
>>>
>>> On Thu, Feb 23, 2012 at 7:18 PM, Edmar Ferreira <
>>> [email protected]> wrote:
>>>
>>>> I will destroy this cluster and launch again with these lines in the
>>>> properties:
>>>>
>>>>
>>>> whirr.hadoop.version=0.20.2
>>>> whirr.hadoop.tarball.url=
>>>> http://archive.apache.org/dist/hadoop/core/hadoop-${whirr.hadoop.version}/hadoop-${whirr.hadoop.version}.tar.gz
>>>>
>>>>
>>>> Any other ideas ?
>>>>
>>>>
>>>>
>>>>
>>>> On Thu, Feb 23, 2012 at 5:16 PM, Andrei Savu <[email protected]>wrote:
>>>>
>>>>> Yep, so I think this is the root cause. I'm pretty sure that you need
>>>>> to make sure you are running the same version.
>>>>>
>>>>> On Thu, Feb 23, 2012 at 7:14 PM, Edmar Ferreira <
>>>>> [email protected]> wrote:
>>>>>
>>>>>> When I run :
>>>>>>
>>>>>> hadoop version in one cluster machine I get
>>>>>>
>>>>>> Warning: $HADOOP_HOME is deprecated.
>>>>>>
>>>>>> Hadoop 0.20.205.0
>>>>>> Subversion
>>>>>> https://svn.apache.org/repos/asf/hadoop/common/branches/branch-0.20-security-205-r
>>>>>>  1179940
>>>>>> Compiled by hortonfo on Fri Oct  7 06:20:32 UTC 2011
>>>>>>
>>>>>>
>>>>>> When I run hadoop version in my local machine I get
>>>>>>
>>>>>> Hadoop 0.20.2
>>>>>> Subversion
>>>>>> https://svn.apache.org/repos/asf/hadoop/common/branches/branch-0.20-r 
>>>>>> 911707
>>>>>> Compiled by chrisdo on Fri Feb 19 08:07:34 UTC 2010
>>>>>>
>>>>>>
>>>>>> On Thu, Feb 23, 2012 at 5:05 PM, Andrei Savu 
>>>>>> <[email protected]>wrote:
>>>>>>
>>>>>>> Do the local Hadoop version match the remote one?
>>>>>>>
>>>>>>>
>>>>>>> On Thu, Feb 23, 2012 at 7:00 PM, Edmar Ferreira <
>>>>>>> [email protected]> wrote:
>>>>>>>
>>>>>>>> Yes, I did a
>>>>>>>>
>>>>>>>> export HADOOP_CONF_DIR=~/.whirr/hadoop/
>>>>>>>>
>>>>>>>> before running hadoop fs -ls
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>> On Thu, Feb 23, 2012 at 4:56 PM, Ashish <[email protected]>wrote:
>>>>>>>>
>>>>>>>>> Did you set the HADOOP_CONF_DIR=~/.whirr/<you cluster name> from
>>>>>>>>> the
>>>>>>>>> shell where you are running the hadoop command?
>>>>>>>>>
>>>>>>>>> On Fri, Feb 24, 2012 at 12:23 AM, Andrei Savu <
>>>>>>>>> [email protected]> wrote:
>>>>>>>>> > That looks fine.
>>>>>>>>> >
>>>>>>>>> > Anything interesting in the Hadoop logs on the remote machines?
>>>>>>>>> Are all the
>>>>>>>>> > daemons running as expected?
>>>>>>>>> >
>>>>>>>>> > On Thu, Feb 23, 2012 at 6:48 PM, Edmar Ferreira
>>>>>>>>> > <[email protected]> wrote:
>>>>>>>>> >>
>>>>>>>>> >> last lines
>>>>>>>>> >>
>>>>>>>>> >>
>>>>>>>>> >> 2012-02-23 16:04:30,241 INFO
>>>>>>>>> >>  [org.apache.whirr.actions.ScriptBasedClusterAction] (main)
>>>>>>>>> Finished running
>>>>>>>>> >> configure phase scripts on all cluster instances
>>>>>>>>> >> 2012-02-23 16:04:30,241 INFO
>>>>>>>>> >>
>>>>>>>>>  [org.apache.whirr.service.hadoop.HadoopNameNodeClusterActionHandler] 
>>>>>>>>> (main)
>>>>>>>>> >> Completed configuration of hadoop role hadoop-namenode
>>>>>>>>> >> 2012-02-23 16:04:30,241 INFO
>>>>>>>>> >>
>>>>>>>>>  [org.apache.whirr.service.hadoop.HadoopNameNodeClusterActionHandler] 
>>>>>>>>> (main)
>>>>>>>>> >> Namenode web UI available at
>>>>>>>>> >> http://ec2-23-20-110-12.compute-1.amazonaws.com:50070
>>>>>>>>> >> 2012-02-23 16:04:30,242 INFO
>>>>>>>>> >>
>>>>>>>>>  [org.apache.whirr.service.hadoop.HadoopNameNodeClusterActionHandler] 
>>>>>>>>> (main)
>>>>>>>>> >> Wrote Hadoop site file
>>>>>>>>> >> /Users/edmaroliveiraferreira/.whirr/hadoop/hadoop-site.xml
>>>>>>>>> >> 2012-02-23 16:04:30,246 INFO
>>>>>>>>> >>
>>>>>>>>>  [org.apache.whirr.service.hadoop.HadoopNameNodeClusterActionHandler] 
>>>>>>>>> (main)
>>>>>>>>> >> Wrote Hadoop proxy script
>>>>>>>>> >> /Users/edmaroliveiraferreira/.whirr/hadoop/hadoop-proxy.sh
>>>>>>>>> >> 2012-02-23 16:04:30,246 INFO
>>>>>>>>> >>
>>>>>>>>>  
>>>>>>>>> [org.apache.whirr.service.hadoop.HadoopJobTrackerClusterActionHandler]
>>>>>>>>> >> (main) Completed configuration of hadoop role hadoop-jobtracker
>>>>>>>>> >> 2012-02-23 16:04:30,246 INFO
>>>>>>>>> >>
>>>>>>>>>  
>>>>>>>>> [org.apache.whirr.service.hadoop.HadoopJobTrackerClusterActionHandler]
>>>>>>>>> >> (main) Jobtracker web UI available at
>>>>>>>>> >> http://ec2-23-20-110-12.compute-1.amazonaws.com:50030
>>>>>>>>> >> 2012-02-23 16:04:30,246 INFO
>>>>>>>>> >>
>>>>>>>>>  [org.apache.whirr.service.hadoop.HadoopDataNodeClusterActionHandler] 
>>>>>>>>> (main)
>>>>>>>>> >> Completed configuration of hadoop role hadoop-datanode
>>>>>>>>> >> 2012-02-23 16:04:30,246 INFO
>>>>>>>>> >>
>>>>>>>>>  
>>>>>>>>> [org.apache.whirr.service.hadoop.HadoopTaskTrackerClusterActionHandler]
>>>>>>>>> >> (main) Completed configuration of hadoop role hadoop-tasktracker
>>>>>>>>> >> 2012-02-23 16:04:30,253 INFO
>>>>>>>>> >>  [org.apache.whirr.actions.ScriptBasedClusterAction] (main)
>>>>>>>>> Finished running
>>>>>>>>> >> start phase scripts on all cluster instances
>>>>>>>>> >> 2012-02-23 16:04:30,257 DEBUG
>>>>>>>>> [org.apache.whirr.service.ComputeCache]
>>>>>>>>> >> (Thread-3) closing ComputeServiceContext {provider=aws-ec2,
>>>>>>>>> >> endpoint=https://ec2.us-east-1.amazonaws.com,
>>>>>>>>> apiVersion=2010-06-15,
>>>>>>>>> >> buildVersion=, identity=08WMRG9HQYYGVQDT57R2,
>>>>>>>>> iso3166Codes=[US-VA, US-CA,
>>>>>>>>> >> US-OR, BR-SP, IE, SG, JP-13]}
>>>>>>>>> >>
>>>>>>>>> >>
>>>>>>>>> >>
>>>>>>>>> >>
>>>>>>>>> >> On Thu, Feb 23, 2012 at 4:31 PM, Andrei Savu <
>>>>>>>>> [email protected]>
>>>>>>>>> >> wrote:
>>>>>>>>> >>>
>>>>>>>>> >>> I think it's the first time I see this. Anything interesting
>>>>>>>>> in the
>>>>>>>>> >>> logs?
>>>>>>>>> >>>
>>>>>>>>> >>>
>>>>>>>>> >>> On Thu, Feb 23, 2012 at 6:27 PM, Edmar Ferreira
>>>>>>>>> >>> <[email protected]> wrote:
>>>>>>>>> >>>>
>>>>>>>>> >>>> Hi guys,
>>>>>>>>> >>>>
>>>>>>>>> >>>> When I launch a cluster and run the proxy everything seems to
>>>>>>>>> be right,
>>>>>>>>> >>>> but when I try to use any command in hadoop I get this error:
>>>>>>>>> >>>>
>>>>>>>>> >>>> Bad connection to FS. command aborted.
>>>>>>>>> >>>>
>>>>>>>>> >>>> Any suggestions ?
>>>>>>>>> >>>>
>>>>>>>>> >>>> Thanks
>>>>>>>>> >>>>
>>>>>>>>> >>>> --
>>>>>>>>> >>>> Edmar Ferreira
>>>>>>>>> >>>> Co-Founder at Everwrite
>>>>>>>>> >>>>
>>>>>>>>> >>>
>>>>>>>>> >>
>>>>>>>>> >>
>>>>>>>>> >>
>>>>>>>>> >> --
>>>>>>>>> >> Edmar Ferreira
>>>>>>>>> >> Co-Founder at Everwrite
>>>>>>>>> >>
>>>>>>>>> >
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> --
>>>>>>>>> thanks
>>>>>>>>> ashish
>>>>>>>>>
>>>>>>>>> Blog: http://www.ashishpaliwal.com/blog
>>>>>>>>> My Photo Galleries: http://www.pbase.com/ashishpaliwal
>>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>> --
>>>>>>>> Edmar Ferreira
>>>>>>>> Co-Founder at Everwrite
>>>>>>>>
>>>>>>>>
>>>>>>>
>>>>>>
>>>>>>
>>>>>> --
>>>>>> Edmar Ferreira
>>>>>> Co-Founder at Everwrite
>>>>>>
>>>>>>
>>>>>
>>>>
>>>>
>>>> --
>>>> Edmar Ferreira
>>>> Co-Founder at Everwrite
>>>>
>>>>
>>>
>>
>>
>> --
>> Edmar Ferreira
>> Co-Founder at Everwrite
>>
>>
>

Reply via email to