[ 
https://issues.apache.org/jira/browse/HADOOP-4976?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Thibaut updated HADOOP-4976:
----------------------------


Try increasing the child heap size in the hadoop-site.xml configuration file.

> Mapper runs out of memory
> -------------------------
>
>                 Key: HADOOP-4976
>                 URL: https://issues.apache.org/jira/browse/HADOOP-4976
>             Project: Hadoop Core
>          Issue Type: Bug
>          Components: mapred
>    Affects Versions: 0.19.0
>         Environment: Amazon EC2 Extra Large instance (4 cores, 15 GB RAM), 
> Sun Java 6 (1.6.0_10); 1 Master, 4 Slaves (all the same); each Java process 
> takes the argument "-Xmx700m" (2 Java processes per Instance)
>            Reporter: Richard J. Zak
>             Fix For: 0.19.1
>
>
> The hadoop job has the task of processing 4 directories in HDFS, each with 15 
> files.  This is sample data, a test run, before I go to the needed 5 
> directories of about 800 documents each.  The mapper takes in nearly 200 
> pages (not files) and throws an OutOfMemory exception.  The largest file is 
> 17 MB.
> If this problem is something on my end and not truly a bug, I apologize.  
> However, after Googling a bit, I did see many threads of people running out 
> of memory with small data sets.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.

Reply via email to