Error occurred in  "crawldb TestDB/crawldb" reduce phase

i get error msg --- java.lang.OutOfMemoryError: Java heap space

my command
 bin/nutch crawl url -dir TestDB -depth 4 -threads 3

 single fetchlist around in 200000

my settings on the memory

hadoop-env.sh
export HADOOP_HEAPSIZE=800

hadoop-site.xml
<property>
  <name>mapred.tasktracker.map.tasks.maximum</name>
  <value>4</value>
</property>
<property>
  <name>mapred.tasktracker.reduce.tasks.maximum</name>
  <value>4</value>
</property>
<property>
  <name>mapred.map.tasks</name>
  <value>2</value>
</property>
<property>
  <name>mapred.reduce.tasks</name>
  <value>2</value>
</property>
<property>
  <name>mapred.map.max.attempts</name>
  <value>4</value>
</property>
<property>
  <name>mapred.reduce.max.attempts</name>
  <value>4</value>
</property>
<property>
  <name>mapred.child.java.opts</name>
  <value>-Xmx250m</value>
</property>

Reply via email to