hadoop-yetus commented on pull request #2655: URL: https://github.com/apache/hadoop/pull/2655#issuecomment-768192307
:broken_heart: **-1 overall** | Vote | Subsystem | Runtime | Logfile | Comment | |:----:|----------:|--------:|:--------:|:-------:| | +0 :ok: | reexec | 1m 27s | | Docker mode activated. | |||| _ Prechecks _ | | +1 :green_heart: | dupname | 0m 4s | | No case conflicting files found. | | +0 :ok: | buf | 0m 1s | | buf was not available. | | +1 :green_heart: | @author | 0m 0s | | The patch does not contain any @author tags. | | +1 :green_heart: | | 0m 0s | [test4tests](test4tests) | The patch appears to include 79 new or modified test files. | |||| _ HDFS-15714 Compile Tests _ | | +0 :ok: | mvndep | 13m 53s | | Maven dependency ordering for branch | | +1 :green_heart: | mvninstall | 23m 50s | | HDFS-15714 passed | | +1 :green_heart: | compile | 21m 54s | | HDFS-15714 passed with JDK Ubuntu-11.0.9.1+1-Ubuntu-0ubuntu1.18.04 | | +1 :green_heart: | compile | 18m 22s | | HDFS-15714 passed with JDK Private Build-1.8.0_275-8u275-b01-0ubuntu1~18.04-b01 | | +1 :green_heart: | checkstyle | 4m 9s | | HDFS-15714 passed | | +1 :green_heart: | mvnsite | 6m 3s | | HDFS-15714 passed | | +1 :green_heart: | shadedclient | 27m 51s | | branch has no errors when building and testing our client artifacts. | | +1 :green_heart: | javadoc | 4m 30s | | HDFS-15714 passed with JDK Ubuntu-11.0.9.1+1-Ubuntu-0ubuntu1.18.04 | | +1 :green_heart: | javadoc | 5m 53s | | HDFS-15714 passed with JDK Private Build-1.8.0_275-8u275-b01-0ubuntu1~18.04-b01 | | +0 :ok: | spotbugs | 0m 46s | | Used deprecated FindBugs config; considering switching to SpotBugs. | | +1 :green_heart: | findbugs | 11m 26s | | HDFS-15714 passed | |||| _ Patch Compile Tests _ | | +0 :ok: | mvndep | 0m 26s | | Maven dependency ordering for patch | | +1 :green_heart: | mvninstall | 4m 31s | | the patch passed | | +1 :green_heart: | compile | 21m 7s | | the patch passed with JDK Ubuntu-11.0.9.1+1-Ubuntu-0ubuntu1.18.04 | | -1 :x: | cc | 21m 7s | [/diff-compile-cc-root-jdkUbuntu-11.0.9.1+1-Ubuntu-0ubuntu1.18.04.txt](https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-2655/1/artifact/out/diff-compile-cc-root-jdkUbuntu-11.0.9.1+1-Ubuntu-0ubuntu1.18.04.txt) | root-jdkUbuntu-11.0.9.1+1-Ubuntu-0ubuntu1.18.04 with JDK Ubuntu-11.0.9.1+1-Ubuntu-0ubuntu1.18.04 generated 30 new + 142 unchanged - 30 fixed = 172 total (was 172) | | -1 :x: | javac | 21m 7s | [/diff-compile-javac-root-jdkUbuntu-11.0.9.1+1-Ubuntu-0ubuntu1.18.04.txt](https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-2655/1/artifact/out/diff-compile-javac-root-jdkUbuntu-11.0.9.1+1-Ubuntu-0ubuntu1.18.04.txt) | root-jdkUbuntu-11.0.9.1+1-Ubuntu-0ubuntu1.18.04 with JDK Ubuntu-11.0.9.1+1-Ubuntu-0ubuntu1.18.04 generated 67 new + 2006 unchanged - 27 fixed = 2073 total (was 2033) | | +1 :green_heart: | compile | 22m 21s | | the patch passed with JDK Private Build-1.8.0_275-8u275-b01-0ubuntu1~18.04-b01 | | -1 :x: | cc | 22m 21s | [/diff-compile-cc-root-jdkPrivateBuild-1.8.0_275-8u275-b01-0ubuntu1~18.04-b01.txt](https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-2655/1/artifact/out/diff-compile-cc-root-jdkPrivateBuild-1.8.0_275-8u275-b01-0ubuntu1~18.04-b01.txt) | root-jdkPrivateBuild-1.8.0_275-8u275-b01-0ubuntu1~18.04-b01 with JDK Private Build-1.8.0_275-8u275-b01-0ubuntu1~18.04-b01 generated 33 new + 139 unchanged - 33 fixed = 172 total (was 172) | | -1 :x: | javac | 22m 21s | [/diff-compile-javac-root-jdkPrivateBuild-1.8.0_275-8u275-b01-0ubuntu1~18.04-b01.txt](https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-2655/1/artifact/out/diff-compile-javac-root-jdkPrivateBuild-1.8.0_275-8u275-b01-0ubuntu1~18.04-b01.txt) | root-jdkPrivateBuild-1.8.0_275-8u275-b01-0ubuntu1~18.04-b01 with JDK Private Build-1.8.0_275-8u275-b01-0ubuntu1~18.04-b01 generated 67 new + 1901 unchanged - 27 fixed = 1968 total (was 1928) | | -0 :warning: | checkstyle | 4m 43s | [/diff-checkstyle-root.txt](https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-2655/1/artifact/out/diff-checkstyle-root.txt) | root: The patch generated 144 new + 4280 unchanged - 35 fixed = 4424 total (was 4315) | | +1 :green_heart: | mvnsite | 9m 18s | | the patch passed | | -1 :x: | whitespace | 0m 0s | [/whitespace-eol.txt](https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-2655/1/artifact/out/whitespace-eol.txt) | The patch has 6 line(s) that end in whitespace. Use git apply --whitespace=fix <<patch_file>>. Refer https://git-scm.com/docs/git-apply | | +1 :green_heart: | xml | 0m 7s | | The patch has no ill-formed XML file. | | +1 :green_heart: | shadedclient | 19m 31s | | patch has no errors when building and testing our client artifacts. | | -1 :x: | javadoc | 1m 9s | [/patch-javadoc-hadoop-common-project_hadoop-common-jdkUbuntu-11.0.9.1+1-Ubuntu-0ubuntu1.18.04.txt](https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-2655/1/artifact/out/patch-javadoc-hadoop-common-project_hadoop-common-jdkUbuntu-11.0.9.1+1-Ubuntu-0ubuntu1.18.04.txt) | hadoop-common in the patch failed with JDK Ubuntu-11.0.9.1+1-Ubuntu-0ubuntu1.18.04. | | -1 :x: | javadoc | 0m 55s | [/patch-javadoc-hadoop-hdfs-project_hadoop-hdfs-client-jdkUbuntu-11.0.9.1+1-Ubuntu-0ubuntu1.18.04.txt](https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-2655/1/artifact/out/patch-javadoc-hadoop-hdfs-project_hadoop-hdfs-client-jdkUbuntu-11.0.9.1+1-Ubuntu-0ubuntu1.18.04.txt) | hadoop-hdfs-client in the patch failed with JDK Ubuntu-11.0.9.1+1-Ubuntu-0ubuntu1.18.04. | | -1 :x: | javadoc | 1m 19s | [/patch-javadoc-hadoop-hdfs-project_hadoop-hdfs-jdkUbuntu-11.0.9.1+1-Ubuntu-0ubuntu1.18.04.txt](https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-2655/1/artifact/out/patch-javadoc-hadoop-hdfs-project_hadoop-hdfs-jdkUbuntu-11.0.9.1+1-Ubuntu-0ubuntu1.18.04.txt) | hadoop-hdfs in the patch failed with JDK Ubuntu-11.0.9.1+1-Ubuntu-0ubuntu1.18.04. | | -1 :x: | javadoc | 1m 34s | [/diff-javadoc-javadoc-hadoop-hdfs-project_hadoop-hdfs-jdkPrivateBuild-1.8.0_275-8u275-b01-0ubuntu1~18.04-b01.txt](https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-2655/1/artifact/out/diff-javadoc-javadoc-hadoop-hdfs-project_hadoop-hdfs-jdkPrivateBuild-1.8.0_275-8u275-b01-0ubuntu1~18.04-b01.txt) | hadoop-hdfs-project_hadoop-hdfs-jdkPrivateBuild-1.8.0_275-8u275-b01-0ubuntu1~18.04-b01 with JDK Private Build-1.8.0_275-8u275-b01-0ubuntu1~18.04-b01 generated 2 new + 1 unchanged - 0 fixed = 3 total (was 1) | | -1 :x: | findbugs | 2m 31s | [/new-findbugs-hadoop-common-project_hadoop-common.html](https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-2655/1/artifact/out/new-findbugs-hadoop-common-project_hadoop-common.html) | hadoop-common-project/hadoop-common generated 1 new + 0 unchanged - 0 fixed = 1 total (was 0) | | -1 :x: | findbugs | 3m 50s | [/new-findbugs-hadoop-hdfs-project_hadoop-hdfs.html](https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-2655/1/artifact/out/new-findbugs-hadoop-hdfs-project_hadoop-hdfs.html) | hadoop-hdfs-project/hadoop-hdfs generated 25 new + 0 unchanged - 0 fixed = 25 total (was 0) | |||| _ Other Tests _ | | -1 :x: | unit | 17m 9s | [/patch-unit-hadoop-common-project_hadoop-common.txt](https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-2655/1/artifact/out/patch-unit-hadoop-common-project_hadoop-common.txt) | hadoop-common in the patch passed. | | +1 :green_heart: | unit | 2m 30s | | hadoop-hdfs-client in the patch passed. | | -1 :x: | unit | 127m 42s | [/patch-unit-hadoop-hdfs-project_hadoop-hdfs.txt](https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-2655/1/artifact/out/patch-unit-hadoop-hdfs-project_hadoop-hdfs.txt) | hadoop-hdfs in the patch passed. | | -1 :x: | unit | 0m 44s | [/patch-unit-hadoop-hdfs-project_hadoop-hdfs-rbf.txt](https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-2655/1/artifact/out/patch-unit-hadoop-hdfs-project_hadoop-hdfs-rbf.txt) | hadoop-hdfs-rbf in the patch failed. | | -1 :x: | unit | 0m 46s | [/patch-unit-hadoop-tools_hadoop-aws.txt](https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-2655/1/artifact/out/patch-unit-hadoop-tools_hadoop-aws.txt) | hadoop-aws in the patch failed. | | -1 :x: | unit | 16m 27s | [/patch-unit-hadoop-tools_hadoop-fs2img.txt](https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-2655/1/artifact/out/patch-unit-hadoop-tools_hadoop-fs2img.txt) | hadoop-fs2img in the patch passed. | | -1 :x: | asflicense | 1m 52s | [/patch-asflicense-problems.txt](https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-2655/1/artifact/out/patch-asflicense-problems.txt) | The patch generated 1 ASF License warnings. | | | | 409m 22s | | | | Reason | Tests | |-------:|:------| | FindBugs | module:hadoop-common-project/hadoop-common | | | Found reliance on default encoding in org.apache.hadoop.fs.impl.FileSystemMultipartUploader.lambda$innerComplete$3(Map$Entry):in org.apache.hadoop.fs.impl.FileSystemMultipartUploader.lambda$innerComplete$3(Map$Entry): String.getBytes() At FileSystemMultipartUploader.java:[line 217] | | FindBugs | module:hadoop-hdfs-project/hadoop-hdfs | | | Switch statement found in org.apache.hadoop.hdfs.protocolPB.PBHelper.convert(DatanodeProtocolProtos$ProvidedVolCommandProto) where default case is missing At PBHelper.java:where default case is missing At PBHelper.java:[lines 833-838] | | | Switch statement found in org.apache.hadoop.hdfs.protocolPB.PBHelper.convert(ProvidedVolumeCommand) where default case is missing At PBHelper.java:where default case is missing At PBHelper.java:[lines 640-645] | | | org.apache.hadoop.hdfs.server.blockmanagement.ProvidedStorageMap.addBlocksToAliasMap(Map) makes inefficient use of keySet iterator instead of entrySet iterator At ProvidedStorageMap.java:keySet iterator instead of entrySet iterator At ProvidedStorageMap.java:[line 274] | | | Redundant nullcheck of nnProxy, which is known to be non-null in org.apache.hadoop.hdfs.server.common.blockaliasmap.impl.NamenodeInMemoryAliasMapClient.setConf(Configuration) Redundant null check at NamenodeInMemoryAliasMapClient.java:is known to be non-null in org.apache.hadoop.hdfs.server.common.blockaliasmap.impl.NamenodeInMemoryAliasMapClient.setConf(Configuration) Redundant null check at NamenodeInMemoryAliasMapClient.java:[line 51] | | | Redundant nullcheck of org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.ReplicaMap.replicas(String), which is known to be non-null in org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl.deepCopyReplica(String) Redundant null check at FsDatasetImpl.java:is known to be non-null in org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl.deepCopyReplica(String) Redundant null check at FsDatasetImpl.java:[line 229] | | | org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsVolumeImpl.getVolumeMap(RamDiskReplicaTracker) makes inefficient use of keySet iterator instead of entrySet iterator At FsVolumeImpl.java:keySet iterator instead of entrySet iterator At FsVolumeImpl.java:[line 1074] | | | Redundant nullcheck of org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.SynchronousReadThroughInputStream.localReplicaInfo, which is known to be non-null in org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.SynchronousReadThroughInputStream.createLocalReplica(FsVolumeImpl) Redundant null check at SynchronousReadThroughInputStream.java:is known to be non-null in org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.SynchronousReadThroughInputStream.createLocalReplica(FsVolumeImpl) Redundant null check at SynchronousReadThroughInputStream.java:[line 175] | | | instanceof will always return true for all non-null values in org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.VolumeReplicaMap.addAll(VolumeReplicaMap), since all org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.VolumeReplicaMap are instances of org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.VolumeReplicaMap At VolumeReplicaMap.java:for all non-null values in org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.VolumeReplicaMap.addAll(VolumeReplicaMap), since all org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.VolumeReplicaMap are instances of org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.VolumeReplicaMap At VolumeReplicaMap.java:[line 170] | | | Redundant nullcheck of xConfig, which is known to be non-null in org.apache.hadoop.hdfs.server.namenode.FSEditLogOp$AddMountOp.writeFields(DataOutputStream) Redundant null check at FSEditLogOp.java:is known to be non-null in org.apache.hadoop.hdfs.server.namenode.FSEditLogOp$AddMountOp.writeFields(DataOutputStream) Redundant null check at FSEditLogOp.java:[line 4466] | | | Possible null pointer dereference of r in org.apache.hadoop.hdfs.server.namenode.FSNamesystem.truncate(String, long, String, String, long) Dereferenced at FSNamesystem.java:r in org.apache.hadoop.hdfs.server.namenode.FSNamesystem.truncate(String, long, String, String, long) Dereferenced at FSNamesystem.java:[line 2440] | | | Found reliance on default encoding in org.apache.hadoop.hdfs.server.namenode.MountManager.prepareMount(String, String, MountMode, Map, Configuration):in org.apache.hadoop.hdfs.server.namenode.MountManager.prepareMount(String, String, MountMode, Map, Configuration): String.getBytes() At MountManager.java:[line 177] | | | Found reliance on default encoding in org.apache.hadoop.hdfs.server.namenode.SyncMountManager.lambda$getXattrValueByName$3(XAttr):in org.apache.hadoop.hdfs.server.namenode.SyncMountManager.lambda$getXattrValueByName$3(XAttr): new String(byte[]) At SyncMountManager.java:[line 324] | | | Found reliance on default encoding in org.apache.hadoop.hdfs.server.namenode.SyncMountManager.performInitialDiff(String, String):in org.apache.hadoop.hdfs.server.namenode.SyncMountManager.performInitialDiff(String, String): String.getBytes() At SyncMountManager.java:[line 227] | | | Found reliance on default encoding in org.apache.hadoop.hdfs.server.namenode.SyncMountManager.storeSnapshotNameAsXAttr(String, String, String, XAttrSetFlag):in org.apache.hadoop.hdfs.server.namenode.SyncMountManager.storeSnapshotNameAsXAttr(String, String, String, XAttrSetFlag): String.getBytes() At SyncMountManager.java:[line 282] | | | Incorrect lazy initialization of static field org.apache.hadoop.hdfs.server.namenode.SyncMountManager.manager in org.apache.hadoop.hdfs.server.namenode.SyncMountManager.getInstance(Configuration, FSNamesystem) At SyncMountManager.java:field org.apache.hadoop.hdfs.server.namenode.SyncMountManager.manager in org.apache.hadoop.hdfs.server.namenode.SyncMountManager.getInstance(Configuration, FSNamesystem) At SyncMountManager.java:[lines 104-105] | | | org.apache.hadoop.hdfs.server.namenode.mountmanager.SimpleReadCacheManager.findBlocksToEvict(long) makes inefficient use of keySet iterator instead of entrySet iterator At SimpleReadCacheManager.java:keySet iterator instead of entrySet iterator At SimpleReadCacheManager.java:[line 320] | | | Found reliance on default encoding in org.apache.hadoop.hdfs.server.namenode.syncservice.SyncMonitor.getSourceSnapshotId(SnapshotDiffReport):in org.apache.hadoop.hdfs.server.namenode.syncservice.SyncMonitor.getSourceSnapshotId(SnapshotDiffReport): String.getBytes() At SyncMonitor.java:[line 303] | | | Found reliance on default encoding in org.apache.hadoop.hdfs.server.namenode.syncservice.SyncMonitor.getTargetSnapshotId(SnapshotDiffReport):in org.apache.hadoop.hdfs.server.namenode.syncservice.SyncMonitor.getTargetSnapshotId(SnapshotDiffReport): String.getBytes() At SyncMonitor.java:[line 315] | | | Inconsistent synchronization of org.apache.hadoop.hdfs.server.namenode.syncservice.SyncServiceSatisfier.syncServiceSatisfierThread; locked 70% of time Unsynchronized access at SyncServiceSatisfier.java:70% of time Unsynchronized access at SyncServiceSatisfier.java:[line 182] | | | Incorrect lazy initialization of static field org.apache.hadoop.hdfs.server.namenode.syncservice.WriteCacheEvictor.writeCacheEvictor in org.apache.hadoop.hdfs.server.namenode.syncservice.WriteCacheEvictor.getInstance(Configuration, FSNamesystem) At WriteCacheEvictor.java:field org.apache.hadoop.hdfs.server.namenode.syncservice.WriteCacheEvictor.writeCacheEvictor in org.apache.hadoop.hdfs.server.namenode.syncservice.WriteCacheEvictor.getInstance(Configuration, FSNamesystem) At WriteCacheEvictor.java:[lines 84-92] | | | Found reliance on default encoding in org.apache.hadoop.hdfs.server.namenode.syncservice.planner.DirectoryPlanner.convertPathToAbsoluteFile(byte[], Path):in org.apache.hadoop.hdfs.server.namenode.syncservice.planner.DirectoryPlanner.convertPathToAbsoluteFile(byte[], Path): new String(byte[]) At DirectoryPlanner.java:[line 63] | | | Found reliance on default encoding in org.apache.hadoop.hdfs.server.namenode.syncservice.planner.DirectoryPlanner.convertPathToAbsoluteFile(byte[], Path, String):in org.apache.hadoop.hdfs.server.namenode.syncservice.planner.DirectoryPlanner.convertPathToAbsoluteFile(byte[], Path, String): new String(byte[]) At DirectoryPlanner.java:[line 73] | | | Found reliance on default encoding in org.apache.hadoop.hdfs.server.namenode.syncservice.planner.DirectoryPlanner.createPlanForDirectory(SnapshotDiffReport$DiffReportEntry, String, ProvidedVolumeInfo, int):in org.apache.hadoop.hdfs.server.namenode.syncservice.planner.DirectoryPlanner.createPlanForDirectory(SnapshotDiffReport$DiffReportEntry, String, ProvidedVolumeInfo, int): String.getBytes() At DirectoryPlanner.java:[line 103] | | | Format string should use %n rather than n in org.apache.hadoop.hdfs.tools.DFSAdmin.addMount(String[]) At DFSAdmin.java:rather than n in org.apache.hadoop.hdfs.tools.DFSAdmin.addMount(String[]) At DFSAdmin.java:[line 2711] | | | Format string should use %n rather than n in org.apache.hadoop.hdfs.tools.DFSAdmin.listMounts(String[]) At DFSAdmin.java:rather than n in org.apache.hadoop.hdfs.tools.DFSAdmin.listMounts(String[]) At DFSAdmin.java:[line 2746] | | Failed junit tests | hadoop.ha.TestZKFailoverController | | | hadoop.hdfs.server.blockmanagement.TestRBWBlockInvalidation | | | hadoop.hdfs.tools.TestDFSAdminWithHA | | | hadoop.hdfs.TestDatanodeRegistration | | | hadoop.hdfs.TestDFSUpgradeFromImage | | | hadoop.hdfs.server.blockmanagement.TestUnderReplicatedBlocks | | | hadoop.hdfs.TestAppendSnapshotTruncate | | | hadoop.hdfs.server.balancer.TestBalancerWithEncryptedTransfer | | | hadoop.hdfs.TestReplaceDatanodeFailureReplication | | | hadoop.hdfs.TestGetFileChecksum | | | hadoop.hdfs.TestRollingUpgrade | | | hadoop.hdfs.server.blockmanagement.TestHeartbeatHandling | | | hadoop.hdfs.web.TestWebHdfsWithMultipleNameNodes | | | hadoop.hdfs.server.balancer.TestBalancerWithNodeGroup | | | hadoop.tools.TestJMXGet | | | hadoop.hdfs.server.blockmanagement.TestBlockReportRateLimiting | | | hadoop.hdfs.server.balancer.TestBalancerWithSaslDataTransfer | | | hadoop.hdfs.TestBlocksScheduledCounter | | | hadoop.hdfs.TestDistributedFileSystem | | | hadoop.hdfs.TestReadStripedFileWithDecodingCorruptData | | | hadoop.hdfs.security.TestDelegationToken | | | hadoop.hdfs.server.blockmanagement.TestDatanodeManager | | | hadoop.hdfs.TestFileLengthOnClusterRestart | | | hadoop.hdfs.TestFileAppend3 | | | hadoop.hdfs.TestSafeMode | | | hadoop.hdfs.TestAppendDifferentChecksum | | | hadoop.hdfs.TestFileAppend2 | | | hadoop.hdfs.server.balancer.TestBalancerRPCDelay | | | hadoop.hdfs.web.TestWebHDFSAcl | | | hadoop.hdfs.tools.TestDFSAdmin | | | hadoop.tools.TestHdfsConfigFields | | | hadoop.hdfs.server.blockmanagement.TestCorruptionWithFailover | | | hadoop.metrics2.sink.TestRollingFileSystemSinkWithHdfs | | | hadoop.hdfs.server.blockmanagement.TestNodeCount | | | hadoop.hdfs.server.blockmanagement.TestComputeInvalidateWork | | | hadoop.hdfs.TestUnsetAndChangeDirectoryEcPolicy | | | hadoop.hdfs.server.balancer.TestBalancer | | | hadoop.hdfs.server.blockmanagement.TestNameNodePrunesMissingStorages | | | hadoop.hdfs.TestErasureCodingPoliciesWithRandomECPolicy | | | hadoop.hdfs.TestAclsEndToEnd | | | hadoop.hdfs.TestErasureCodingPolicyWithSnapshot | | | hadoop.hdfs.server.blockmanagement.TestBlockReportLease | | | hadoop.fs.TestFcHdfsPermission | | | hadoop.hdfs.server.balancer.TestBalancerWithMultipleNameNodes | | | hadoop.hdfs.TestLeaseRecoveryStriped | | | hadoop.hdfs.security.TestDelegationTokenForProxyUser | | | hadoop.hdfs.TestDecommissionWithStripedBackoffMonitor | | | hadoop.hdfs.server.balancer.TestBalancerService | | | hadoop.hdfs.server.blockmanagement.TestPendingReconstruction | | | hadoop.hdfs.server.blockmanagement.TestReconstructStripedBlocksWithRackAwareness | | | hadoop.hdfs.TestMissingBlocksAlert | | | hadoop.hdfs.TestReconstructStripedFileWithRandomECPolicy | | | hadoop.hdfs.server.blockmanagement.TestBlockTokenWithDFS | | | hadoop.hdfs.server.blockmanagement.TestPendingInvalidateBlock | | | hadoop.hdfs.server.balancer.TestBalancerWithHANameNodes | | | hadoop.hdfs.web.TestWebHdfsFileSystemContract | | | hadoop.hdfs.server.namenode.TestMultiRootProvidedCluster | | | hadoop.hdfs.server.namenode.TestSingleUGIResolver | | | hadoop.hdfs.server.namenode.TestFailuresDuringMount | | Subsystem | Report/Notes | |----------:|:-------------| | Docker | ClientAPI=1.41 ServerAPI=1.41 base: https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-2655/1/artifact/out/Dockerfile | | GITHUB PR | https://github.com/apache/hadoop/pull/2655 | | JIRA Issue | HDFS-15714 | | Optional Tests | dupname asflicense compile javac javadoc mvninstall mvnsite unit shadedclient findbugs checkstyle xml cc buflint bufcompat | | uname | Linux 9a595f5ac4c5 4.15.0-101-generic #102-Ubuntu SMP Mon May 11 10:07:26 UTC 2020 x86_64 x86_64 x86_64 GNU/Linux | | Build tool | maven | | Personality | dev-support/bin/hadoop.sh | | git revision | HDFS-15714 / d82009599a2 | | Default Java | Private Build-1.8.0_275-8u275-b01-0ubuntu1~18.04-b01 | | Multi-JDK versions | /usr/lib/jvm/java-11-openjdk-amd64:Ubuntu-11.0.9.1+1-Ubuntu-0ubuntu1.18.04 /usr/lib/jvm/java-8-openjdk-amd64:Private Build-1.8.0_275-8u275-b01-0ubuntu1~18.04-b01 | | Test Results | https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-2655/1/testReport/ | | Max. process+thread count | 2226 (vs. ulimit of 5500) | | modules | C: hadoop-common-project/hadoop-common hadoop-hdfs-project/hadoop-hdfs-client hadoop-hdfs-project/hadoop-hdfs hadoop-hdfs-project/hadoop-hdfs-rbf hadoop-tools/hadoop-aws hadoop-tools/hadoop-fs2img U: . | | Console output | https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-2655/1/console | | versions | git=2.17.1 maven=3.6.0 findbugs=4.0.6 | | Powered by | Apache Yetus 0.13.0-SNAPSHOT https://yetus.apache.org | This message was automatically generated. ---------------------------------------------------------------- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: [email protected] --------------------------------------------------------------------- To unsubscribe, e-mail: [email protected] For additional commands, e-mail: [email protected]
