Repository: hadoop Updated Branches: refs/heads/trunk d907fdc3c -> 49f9acae4
YARN-8162. Remove Method DirectoryCollection#verifyDirUsingMkdir. (Yufei Gu via Haibo Chen) Project: http://git-wip-us.apache.org/repos/asf/hadoop/repo Commit: http://git-wip-us.apache.org/repos/asf/hadoop/commit/49f9acae Tree: http://git-wip-us.apache.org/repos/asf/hadoop/tree/49f9acae Diff: http://git-wip-us.apache.org/repos/asf/hadoop/diff/49f9acae Branch: refs/heads/trunk Commit: 49f9acae40a6e99cadee9854824dfdc40be67695 Parents: d907fdc Author: Haibo Chen <haiboc...@apache.org> Authored: Mon Apr 16 15:04:12 2018 -0700 Committer: Haibo Chen <haiboc...@apache.org> Committed: Mon Apr 16 15:05:04 2018 -0700 ---------------------------------------------------------------------- .../server/nodemanager/DirectoryCollection.java | 30 -------------------- 1 file changed, 30 deletions(-) ---------------------------------------------------------------------- http://git-wip-us.apache.org/repos/asf/hadoop/blob/49f9acae/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/java/org/apache/hadoop/yarn/server/nodemanager/DirectoryCollection.java ---------------------------------------------------------------------- diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/java/org/apache/hadoop/yarn/server/nodemanager/DirectoryCollection.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/java/org/apache/hadoop/yarn/server/nodemanager/DirectoryCollection.java index 7a840b1..7debbdc 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/java/org/apache/hadoop/yarn/server/nodemanager/DirectoryCollection.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/java/org/apache/hadoop/yarn/server/nodemanager/DirectoryCollection.java @@ -36,8 +36,6 @@ import java.util.concurrent.locks.ReentrantReadWriteLock.WriteLock; import org.slf4j.Logger; import org.slf4j.LoggerFactory; -import org.apache.commons.io.FileUtils; -import org.apache.commons.lang.RandomStringUtils; import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.fs.FileAlreadyExistsException; import org.apache.hadoop.fs.FileContext; @@ -490,9 +488,6 @@ public class DirectoryCollection { new DiskErrorInformation(DiskErrorCause.DISK_FULL, msg)); continue; } - - // create a random dir to make sure fs isn't in read-only mode - verifyDirUsingMkdir(testDir); } catch (IOException ie) { ret.put(dir, new DiskErrorInformation(DiskErrorCause.OTHER, ie.getMessage())); @@ -501,31 +496,6 @@ public class DirectoryCollection { return ret; } - /** - * Function to test whether a dir is working correctly by actually creating a - * random directory. - * - * @param dir - * the dir to test - */ - private void verifyDirUsingMkdir(File dir) throws IOException { - - String randomDirName = RandomStringUtils.randomAlphanumeric(5); - File target = new File(dir, randomDirName); - int i = 0; - while (target.exists()) { - - randomDirName = RandomStringUtils.randomAlphanumeric(5) + i; - target = new File(dir, randomDirName); - i++; - } - try { - diskValidator.checkStatus(target); - } finally { - FileUtils.deleteQuietly(target); - } - } - private boolean isDiskUsageOverPercentageLimit(File dir, float diskUtilizationPercentageCutoff) { float freePercentage = --------------------------------------------------------------------- To unsubscribe, e-mail: common-commits-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-commits-h...@hadoop.apache.org