Gordon Wang created HDFS-6505:
---------------------------------
Summary: Can not close file due to last block is marked as corrupt
Key: HDFS-6505
URL: https://issues.apache.org/jira/browse/HDFS-6505
Project: Hadoop HDFS
Issue Type: Bug
Affects Versions: 2.2.0
Reporter: Gordon Wang
After appending a file, client could not close it. Because namenode could not
complete the last block in file. The UC status of last block remained as COMMIT
and never change.
The namenode log was like this.
{code}
INFO org.apache.hadoop.hdfs.server.namenode.FSNamesystem: BLOCK*
checkFileProgress: blk_1073741920_13948{blockUCState=COMMITTED,
primaryNodeIndex=-1,
replicas=[ReplicaUnderConstruction[172.28.1.2:50010|RBW]]} has not reached
minimal replication 1
{code}
After going through the log of namenode, I found a log like this
{code}
INFO BlockStateChange: BLOCK NameSystem.addToCorruptReplicasMap: blk_1073741920
added as corrupt on 172.28.1.2:50010 by sdw3/172.28.1.3 because client machine
reported it
{code}
But actually, the last block was finished successfully in the data node.
Because I could find the log in datanode
{code}
INFO org.apache.hadoop.hdfs.server.datanode.DataNode: DataTransfer: Transmitted
BP-649434182-172.28.1.251-1401432753616:blk_1073741920_13808
(numBytes=50120352) to /172.28.1.3:50010
INFO org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src:
/172.28.1.2:36860, dest: /172.28.1.2:50010, bytes: 51686616, op: HDFS_WRITE,
cliID: libhdfs3_client_random_741511239_count_1_pid_215802_tid_140085714196576,
offset: 0, srvID: DS-2074102060-172.28.1.2-50010-1401432768690, blockid:
BP-649434182-172.28.1.251-1401432753616:blk_1073741920_13948, duration:
189226453336
INFO org.apache.hadoop.hdfs.server.datanode.DataNode: PacketResponder:
BP-649434182-172.28.1.251-1401432753616:blk_1073741920_13948,
type=LAST_IN_PIPELINE, downstreams=0:[] terminating
{code}
--
This message was sent by Atlassian JIRA
(v6.2#6252)