Hi,

What is your # of slots per TaskTracker? Your ulimit seems pretty
high. I'd set it to 1.5x times heap initially, i.e., 6291456 (6 GB)
and try.

On Sun, Oct 7, 2012 at 3:50 AM, Attila Csordas <attilacsor...@gmail.com> wrote:
> some details to this problem:
>
> 12/10/05 12:13:27 INFO mapred.JobClient:  map 0% reduce 0%
> 12/10/05 12:13:40 INFO mapred.JobClient: Task Id :
> attempt_201210051158_0001_m_000002_0, Status : FAILED
> java.lang.Throwable: Child Error
>         at org.apache.hadoop.mapred.TaskRunner.run(TaskRunner.java:271)
> Caused by: java.io.IOException: Task process exit with nonzero status of 134.
>         at org.apache.hadoop.mapred.TaskRunner.run(TaskRunner.java:258)
>
> attempt_201210051158_0001_m_000002_0: #
> attempt_201210051158_0001_m_000002_0: # There is insufficient memory
> for the Java Runtime Environment to continue.
> attempt_201210051158_0001_m_000002_0: # pthread_getattr_np
>
> in mapred-site.xml the following memory settings were set after a
> couple trials to get rid of the problem this way:
>
> <property>
> <name>mapred.child.java.opts</name>
> <value>-server -Xmx4096M -Djava.net.preferIPv4Stack=true</value>
> </property>
>
> <property>
> <name>mapred.child.ulimit</name>
> <value>16777216</value>
> </property>
>
> Cheers,
> Attila
>
>
>
> On Fri, Oct 5, 2012 at 10:50 AM, Steve Lewis <lordjoe2...@gmail.com> wrote:
>> [We get 'There is insufficient memory for the Java Runtime Environment to
>> continue.'
>> any time we run any job including the most trivial word count process. It is
>> true I am generating a jar for a larger job but only running a version of
>> wordcount that worked well under 0.2
>> Any bright ideas???
>> This is a new 1.03 installation and nothing is known to work
>>
>> Steven M. Lewis PhD
>> 4221 105th Ave NE
>> Kirkland, WA 98033
>> cell 206-384-1340
>> skype lordjoe_com



-- 
Harsh J

Reply via email to