Repository: hadoop Updated Branches: refs/heads/branch-2.6 2586b7500 -> 3f87efc0c
HDFS-9530. ReservedSpace is not cleared for abandoned Blocks (Brahma Reddy Battula) Project: http://git-wip-us.apache.org/repos/asf/hadoop/repo Commit: http://git-wip-us.apache.org/repos/asf/hadoop/commit/3f87efc0 Tree: http://git-wip-us.apache.org/repos/asf/hadoop/tree/3f87efc0 Diff: http://git-wip-us.apache.org/repos/asf/hadoop/diff/3f87efc0 Branch: refs/heads/branch-2.6 Commit: 3f87efc0c8f189253493c927b3c39995c61d6849 Parents: 2586b75 Author: Arpit Agarwal <[email protected]> Authored: Wed Aug 17 14:59:04 2016 -0700 Committer: Arpit Agarwal <[email protected]> Committed: Wed Aug 17 14:59:04 2016 -0700 ---------------------------------------------------------------------- hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt | 3 ++ .../server/datanode/DataNodeFaultInjector.java | 2 + .../hdfs/server/datanode/DataXceiver.java | 1 + .../datanode/fsdataset/impl/FsDatasetImpl.java | 6 ++- .../fsdataset/impl/TestRbwSpaceReservation.java | 48 +++++++++++++++++++- 5 files changed, 58 insertions(+), 2 deletions(-) ---------------------------------------------------------------------- http://git-wip-us.apache.org/repos/asf/hadoop/blob/3f87efc0/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt ---------------------------------------------------------------------- diff --git a/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt b/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt index cc27d77..da6abd9 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt +++ b/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt @@ -30,6 +30,9 @@ Release 2.6.5 - UNRELEASED HDFS-8581. ContentSummary on / skips further counts on yielding lock (J.Andreina via vinayakumarb) + HDFS-9530. ReservedSpace is not cleared for abandoned Blocks. + (Brahma Reddy Battula) + Release 2.6.4 - 2016-02-11 INCOMPATIBLE CHANGES http://git-wip-us.apache.org/repos/asf/hadoop/blob/3f87efc0/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/DataNodeFaultInjector.java ---------------------------------------------------------------------- diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/DataNodeFaultInjector.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/DataNodeFaultInjector.java index e66d66b..2627788 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/DataNodeFaultInjector.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/DataNodeFaultInjector.java @@ -49,4 +49,6 @@ public class DataNodeFaultInjector { public void sendShortCircuitShmResponse() throws IOException {} public void noRegistration() throws IOException { } + + public void failMirrorConnection() throws IOException { } } http://git-wip-us.apache.org/repos/asf/hadoop/blob/3f87efc0/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/DataXceiver.java ---------------------------------------------------------------------- diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/DataXceiver.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/DataXceiver.java index d7551cc..8e9628a 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/DataXceiver.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/DataXceiver.java @@ -665,6 +665,7 @@ class DataXceiver extends Receiver implements Runnable { mirrorTarget = NetUtils.createSocketAddr(mirrorNode); mirrorSock = datanode.newSocket(); try { + DataNodeFaultInjector.get().failMirrorConnection(); int timeoutValue = dnConf.socketTimeout + (HdfsServerConstants.READ_TIMEOUT_EXTENSION * targets.length); int writeTimeout = dnConf.socketWriteTimeout + http://git-wip-us.apache.org/repos/asf/hadoop/blob/3f87efc0/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/impl/FsDatasetImpl.java ---------------------------------------------------------------------- diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/impl/FsDatasetImpl.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/impl/FsDatasetImpl.java index f00416b..2008b3a 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/impl/FsDatasetImpl.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/impl/FsDatasetImpl.java @@ -86,6 +86,7 @@ import org.apache.hadoop.hdfs.server.datanode.UnexpectedReplicaStateException; import org.apache.hadoop.hdfs.server.datanode.fsdataset.FsDatasetSpi; import org.apache.hadoop.hdfs.server.datanode.fsdataset.FsVolumeSpi; import org.apache.hadoop.hdfs.server.datanode.fsdataset.LengthInputStream; +import org.apache.hadoop.hdfs.server.datanode.ReplicaInPipelineInterface; import org.apache.hadoop.hdfs.server.datanode.fsdataset.ReplicaInputStreams; import org.apache.hadoop.hdfs.server.datanode.fsdataset.ReplicaOutputStreams; import org.apache.hadoop.hdfs.server.datanode.fsdataset.RollingLogs; @@ -1606,7 +1607,10 @@ class FsDatasetImpl implements FsDatasetSpi<FsVolumeImpl> { + ". Parent not found for file " + f); continue; } - volumeMap.remove(bpid, invalidBlks[i]); + ReplicaInfo removing = volumeMap.remove(bpid, invalidBlks[i]); + if (removing instanceof ReplicaInPipelineInterface) { + ((ReplicaInPipelineInterface) removing).releaseAllBytesReserved(); + } } if (v.isTransientStorage()) { http://git-wip-us.apache.org/repos/asf/hadoop/blob/3f87efc0/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/impl/TestRbwSpaceReservation.java ---------------------------------------------------------------------- diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/impl/TestRbwSpaceReservation.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/impl/TestRbwSpaceReservation.java index efe03eb..61a849c 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/impl/TestRbwSpaceReservation.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/impl/TestRbwSpaceReservation.java @@ -32,6 +32,8 @@ import org.apache.hadoop.fs.FSDataOutputStream; import org.apache.hadoop.fs.Path; import org.apache.hadoop.hdfs.*; import org.apache.hadoop.hdfs.server.datanode.DataNode; +import org.apache.hadoop.hdfs.server.datanode.DataNodeFaultInjector; +import org.apache.hadoop.hdfs.server.datanode.fsdataset.FsDatasetSpi; import org.apache.hadoop.hdfs.server.datanode.fsdataset.FsVolumeSpi; import org.apache.hadoop.ipc.RemoteException; import org.apache.hadoop.test.GenericTestUtils; @@ -65,7 +67,7 @@ public class TestRbwSpaceReservation { private DistributedFileSystem fs = null; private DFSClient client = null; FsVolumeImpl singletonVolume = null; - + private DataNodeFaultInjector old = null; private static Random rand = new Random(); private void initConfig(int blockSize) { @@ -414,6 +416,50 @@ public class TestRbwSpaceReservation { checkReservedSpace(expectedFile2Reserved); } + @Test(timeout = 30000) + public void testReservedSpaceForPipelineRecovery() + throws Exception { + final short replication = 3; + startCluster(BLOCK_SIZE, replication, -1); + + final String methodName = GenericTestUtils.getMethodName(); + final Path file = new Path("/" + methodName + ".01.dat"); + + old = DataNodeFaultInjector.get(); + // Fault injector to fail connection to mirror first time. + DataNodeFaultInjector.set(new DataNodeFaultInjector() { + private int tries = 0; + + @Override + public void failMirrorConnection() throws IOException { + if (tries++ == 0) { + throw new IOException("Failing Mirror for space reservation"); + } + } + }); + // Write 1 byte to the file and kill the writer. + FSDataOutputStream os = fs.create(file, replication); + os.write(new byte[1]); + os.close(); + // Ensure all space reserved for the replica was released on each + // DataNode. + cluster.triggerBlockReports(); + for (final DataNode dn : cluster.getDataNodes()) { + for (FsVolumeSpi fsVolume : dn.getFSDataset().getVolumes()) { + { + final FsVolumeImpl volume = (FsVolumeImpl) fsVolume; + GenericTestUtils.waitFor(new Supplier<Boolean>() { + @Override public Boolean get() { + LOG.info("dn " + dn.getDisplayName() + " space : " + volume + .getReservedForRbw()); + return (volume.getReservedForRbw() == 0); + } + }, 100, Integer.MAX_VALUE); // Wait until the test times out. + } + } + } + } + private void checkReservedSpace(final long expectedReserved) throws TimeoutException, InterruptedException, IOException { for (final DataNode dn : cluster.getDataNodes()) { --------------------------------------------------------------------- To unsubscribe, e-mail: [email protected] For additional commands, e-mail: [email protected]
