hi,
I am encountering a problem when running a hadoop job with a relative large
dataset(about 400M) in single-node hadoop environment.
The error said that dfs fail to create new block. However, the size of physical
disk is large enough. So is there any reason for this failure? any limitation
for the size of disk space a job can occupy? The following is the snippet of
exception stack. Thanks for your attention.
Regards,
Jianmin
2009-07-28 18:00:31,757 INFO org.apache.hadoop.mapred.Merger: Merging 1 sorted
segments
2009-07-28 18:00:31,792 INFO org.apache.hadoop.mapred.Merger: Down to the last
merge-pass, with 1 segments left of total size: 1272809137 bytes
2009-07-28 18:01:06,521 INFO org.apache.hadoop.hdfs.DFSClient: Exception in
createBlockOutputStream java.io.EOFException
2009-07-28 18:01:06,521 INFO org.apache.hadoop.hdfs.DFSClient: Abandoning block
blk_2149418359249628613_12378
2009-07-28 18:01:12,578 INFO org.apache.hadoop.hdfs.DFSClient: Exception in
createBlockOutputStream java.io.EOFException
2009-07-28 18:01:12,578 INFO org.apache.hadoop.hdfs.DFSClient: Abandoning block
blk_-4276450909968435375_12378
2009-07-28 18:01:18,581 INFO org.apache.hadoop.hdfs.DFSClient: Exception in
createBlockOutputStream java.io.EOFException
2009-07-28 18:01:18,581 INFO org.apache.hadoop.hdfs.DFSClient: Abandoning block
blk_1370666846409896923_12378
2009-07-28 18:01:24,584 INFO org.apache.hadoop.hdfs.DFSClient: Exception in
createBlockOutputStream java.io.EOFException
2009-07-28 18:01:24,584 INFO org.apache.hadoop.hdfs.DFSClient: Abandoning block
blk_-8117322104093252360_12378
2009-07-28 18:01:30,621 WARN org.apache.hadoop.hdfs.DFSClient: DataStreamer
Exception: java.io.IOException: Unable to create new block.
at
org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.nextBlockOutputStream(DFSClient.java:2781)
at
org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.access$2000(DFSClient.java:2046)
at
org.apache.hadoop.hdfs.DFSClient$DFSOutputStream$DataStreamer.run(DFSClient.java:2232)
2009-07-28 18:01:30,622 WARN org.apache.hadoop.hdfs.DFSClient: Error Recovery
for block blk_-8117322104093252360_12378 bad datanode[0] nodes == null
2009-07-28 18:01:30,622 WARN org.apache.hadoop.hdfs.DFSClient: Could not get
block locations. Source file "/data/segment/dat_4_8" - Aborting...
2009-07-28 18:01:30,635 WARN org.apache.hadoop.mapred.TaskTracker: Error
running child
java.io.EOFException
at java.io.DataInputStream.readByte(DataInputStream.java:250)
at org.apache.hadoop.io.WritableUtils.readVLong(WritableUtils.java:298)
at org.apache.hadoop.io.WritableUtils.readVInt(WritableUtils.java:319)
at org.apache.hadoop.io.Text.readString(Text.java:400)
at
org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.createBlockOutputStream(DFSClient.java:2837)
at
org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.nextBlockOutputStream(DFSClient.java:2762)
at
org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.access$2000(DFSClient.java:2046)
at
org.apache.hadoop.hdfs.DFSClient$DFSOutputStream$DataStreamer.run(DFSClient.java:2232)
2009-07-28 18:01:30,645 INFO org.apache.hadoop.mapred.TaskRunner: Runnning
cleanup for the task