Brahma Reddy Battula created HDFS-10226:
-------------------------------------------
Summary: Track and use BlockScheduled size for DatanodeDescriptor
instead of count.
Key: HDFS-10226
URL: https://issues.apache.org/jira/browse/HDFS-10226
Project: Hadoop HDFS
Issue Type: Bug
Reporter: Brahma Reddy Battula
Assignee: Brahma Reddy Battula
Tracking count will result in inaccurate estimation of remaining space in case
of different block sized files being written.
*For Example:*
1. Datanode Capacity is 10GB, available is 2GB.
2. For NNBench testing, Low blocksize might be used ( such as 1MB), and
currently 20 blocks are being written to DN. Scheduled counter will be 20.
3. This counter will not give any issue for blocks of NNBench with block size
as 1MB.
but for normal files with 128MB block size, remaining space will be seen as 0.
(because it will calculate based on current file's block size. not the original
scheduled size.
20*128MB = 2.5GB, which is greater than available. So remaining will be 0 for
normal block.
here we'll get,
"Could only be replicated to 0 nodes instead of minReplication (=1 ). There are
2 datanode(s) running and no node(s) are excluded in this operation" exception
will come.
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)