[jira] [Commented] (HDFS-17352) Add configuration to control whether DN delete this replica from disk when client requests a missing block
[ https://issues.apache.org/jira/browse/HDFS-17352?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17820288#comment-17820288 ] ASF GitHub Bot commented on HDFS-17352: --- hadoop-yetus commented on PR #6559: URL: https://github.com/apache/hadoop/pull/6559#issuecomment-1962286525 :broken_heart: **-1 overall** | Vote | Subsystem | Runtime | Logfile | Comment | |::|--:|:|::|:---:| | +0 :ok: | reexec | 0m 22s | | Docker mode activated. | _ Prechecks _ | | +1 :green_heart: | dupname | 0m 0s | | No case conflicting files found. | | +0 :ok: | codespell | 0m 1s | | codespell was not available. | | +0 :ok: | detsecrets | 0m 1s | | detect-secrets was not available. | | +0 :ok: | xmllint | 0m 1s | | xmllint was not available. | | +1 :green_heart: | @author | 0m 0s | | The patch does not contain any @author tags. | | +1 :green_heart: | test4tests | 0m 0s | | The patch appears to include 5 new or modified test files. | _ trunk Compile Tests _ | | +1 :green_heart: | mvninstall | 50m 31s | | trunk passed | | +1 :green_heart: | compile | 0m 41s | | trunk passed with JDK Ubuntu-11.0.21+9-post-Ubuntu-0ubuntu120.04 | | +1 :green_heart: | compile | 0m 40s | | trunk passed with JDK Private Build-1.8.0_392-8u392-ga-1~20.04-b08 | | +1 :green_heart: | checkstyle | 0m 40s | | trunk passed | | +1 :green_heart: | mvnsite | 0m 47s | | trunk passed | | +1 :green_heart: | javadoc | 0m 45s | | trunk passed with JDK Ubuntu-11.0.21+9-post-Ubuntu-0ubuntu120.04 | | +1 :green_heart: | javadoc | 1m 2s | | trunk passed with JDK Private Build-1.8.0_392-8u392-ga-1~20.04-b08 | | +1 :green_heart: | spotbugs | 1m 42s | | trunk passed | | +1 :green_heart: | shadedclient | 20m 41s | | branch has no errors when building and testing our client artifacts. | _ Patch Compile Tests _ | | +1 :green_heart: | mvninstall | 0m 37s | | the patch passed | | +1 :green_heart: | compile | 0m 36s | | the patch passed with JDK Ubuntu-11.0.21+9-post-Ubuntu-0ubuntu120.04 | | +1 :green_heart: | javac | 0m 36s | | the patch passed | | +1 :green_heart: | compile | 0m 34s | | the patch passed with JDK Private Build-1.8.0_392-8u392-ga-1~20.04-b08 | | +1 :green_heart: | javac | 0m 34s | | the patch passed | | +1 :green_heart: | blanks | 0m 0s | | The patch has no blanks issues. | | +1 :green_heart: | checkstyle | 0m 31s | | the patch passed | | +1 :green_heart: | mvnsite | 0m 37s | | the patch passed | | +1 :green_heart: | javadoc | 0m 32s | | the patch passed with JDK Ubuntu-11.0.21+9-post-Ubuntu-0ubuntu120.04 | | +1 :green_heart: | javadoc | 0m 59s | | the patch passed with JDK Private Build-1.8.0_392-8u392-ga-1~20.04-b08 | | +1 :green_heart: | spotbugs | 1m 43s | | the patch passed | | +1 :green_heart: | shadedclient | 20m 30s | | patch has no errors when building and testing our client artifacts. | _ Other Tests _ | | -1 :x: | unit | 204m 3s | [/patch-unit-hadoop-hdfs-project_hadoop-hdfs.txt](https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-6559/5/artifact/out/patch-unit-hadoop-hdfs-project_hadoop-hdfs.txt) | hadoop-hdfs in the patch passed. | | +1 :green_heart: | asflicense | 0m 30s | | The patch does not generate ASF License warnings. | | | | 310m 8s | | | | Reason | Tests | |---:|:--| | Failed junit tests | hadoop.hdfs.server.datanode.TestLargeBlockReport | | | hadoop.hdfs.server.datanode.TestDirectoryScanner | | | hadoop.hdfs.protocol.TestBlockListAsLongs | | | hadoop.hdfs.tools.TestDFSAdmin | | Subsystem | Report/Notes | |--:|:-| | Docker | ClientAPI=1.44 ServerAPI=1.44 base: https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-6559/5/artifact/out/Dockerfile | | GITHUB PR | https://github.com/apache/hadoop/pull/6559 | | Optional Tests | dupname asflicense compile javac javadoc mvninstall mvnsite unit shadedclient spotbugs checkstyle codespell detsecrets xmllint | | uname | Linux e1573661312f 5.15.0-94-generic #104-Ubuntu SMP Tue Jan 9 15:25:40 UTC 2024 x86_64 x86_64 x86_64 GNU/Linux | | Build tool | maven | | Personality | dev-support/bin/hadoop.sh | | git revision | trunk / cada170bb5603c7921ae6eae957dff26deb76c5d | | Default Java | Private Build-1.8.0_392-8u392-ga-1~20.04-b08 | | Multi-JDK versions | /usr/lib/jvm/java-11-openjdk-amd64:Ubuntu-11.0.21+9-post-Ubuntu-0ubuntu120.04 /usr/lib/jvm/java-8-openjdk-amd64:Private Build-1.8.0_392-8u392-ga-1~20.04-b08 | | Test Results | https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-6559
[jira] [Commented] (HDFS-17396) BootstrapStandby should download rollback image during RollingUpgrade
[ https://issues.apache.org/jira/browse/HDFS-17396?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17820266#comment-17820266 ] ASF GitHub Bot commented on HDFS-17396: --- hadoop-yetus commented on PR #6583: URL: https://github.com/apache/hadoop/pull/6583#issuecomment-1962238321 :broken_heart: **-1 overall** | Vote | Subsystem | Runtime | Logfile | Comment | |::|--:|:|::|:---:| | +0 :ok: | reexec | 0m 34s | | Docker mode activated. | _ Prechecks _ | | +1 :green_heart: | dupname | 0m 0s | | No case conflicting files found. | | +0 :ok: | codespell | 0m 1s | | codespell was not available. | | +0 :ok: | detsecrets | 0m 1s | | detect-secrets was not available. | | +0 :ok: | buf | 0m 1s | | buf was not available. | | +0 :ok: | buf | 0m 1s | | buf was not available. | | +1 :green_heart: | @author | 0m 0s | | The patch does not contain any @author tags. | | +1 :green_heart: | test4tests | 0m 0s | | The patch appears to include 2 new or modified test files. | _ trunk Compile Tests _ | | +0 :ok: | mvndep | 34m 5s | | Maven dependency ordering for branch | | +1 :green_heart: | mvninstall | 36m 33s | | trunk passed | | +1 :green_heart: | compile | 5m 46s | | trunk passed with JDK Ubuntu-11.0.21+9-post-Ubuntu-0ubuntu120.04 | | +1 :green_heart: | compile | 5m 43s | | trunk passed with JDK Private Build-1.8.0_392-8u392-ga-1~20.04-b08 | | +1 :green_heart: | checkstyle | 1m 30s | | trunk passed | | +1 :green_heart: | mvnsite | 2m 11s | | trunk passed | | +1 :green_heart: | javadoc | 1m 48s | | trunk passed with JDK Ubuntu-11.0.21+9-post-Ubuntu-0ubuntu120.04 | | +1 :green_heart: | javadoc | 2m 10s | | trunk passed with JDK Private Build-1.8.0_392-8u392-ga-1~20.04-b08 | | -1 :x: | spotbugs | 1m 22s | [/branch-spotbugs-hadoop-hdfs-project_hadoop-hdfs-rbf-warnings.html](https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-6583/1/artifact/out/branch-spotbugs-hadoop-hdfs-project_hadoop-hdfs-rbf-warnings.html) | hadoop-hdfs-project/hadoop-hdfs-rbf in trunk has 1 extant spotbugs warnings. | | +1 :green_heart: | shadedclient | 36m 41s | | branch has no errors when building and testing our client artifacts. | _ Patch Compile Tests _ | | +0 :ok: | mvndep | 2m 39s | | Maven dependency ordering for patch | | -1 :x: | mvninstall | 0m 50s | [/patch-mvninstall-hadoop-hdfs-project_hadoop-hdfs.txt](https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-6583/1/artifact/out/patch-mvninstall-hadoop-hdfs-project_hadoop-hdfs.txt) | hadoop-hdfs in the patch failed. | | -1 :x: | mvninstall | 0m 24s | [/patch-mvninstall-hadoop-hdfs-project_hadoop-hdfs-rbf.txt](https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-6583/1/artifact/out/patch-mvninstall-hadoop-hdfs-project_hadoop-hdfs-rbf.txt) | hadoop-hdfs-rbf in the patch failed. | | -1 :x: | compile | 1m 25s | [/patch-compile-hadoop-hdfs-project-jdkUbuntu-11.0.21+9-post-Ubuntu-0ubuntu120.04.txt](https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-6583/1/artifact/out/patch-compile-hadoop-hdfs-project-jdkUbuntu-11.0.21+9-post-Ubuntu-0ubuntu120.04.txt) | hadoop-hdfs-project in the patch failed with JDK Ubuntu-11.0.21+9-post-Ubuntu-0ubuntu120.04. | | -1 :x: | cc | 1m 25s | [/patch-compile-hadoop-hdfs-project-jdkUbuntu-11.0.21+9-post-Ubuntu-0ubuntu120.04.txt](https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-6583/1/artifact/out/patch-compile-hadoop-hdfs-project-jdkUbuntu-11.0.21+9-post-Ubuntu-0ubuntu120.04.txt) | hadoop-hdfs-project in the patch failed with JDK Ubuntu-11.0.21+9-post-Ubuntu-0ubuntu120.04. | | -1 :x: | javac | 1m 25s | [/patch-compile-hadoop-hdfs-project-jdkUbuntu-11.0.21+9-post-Ubuntu-0ubuntu120.04.txt](https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-6583/1/artifact/out/patch-compile-hadoop-hdfs-project-jdkUbuntu-11.0.21+9-post-Ubuntu-0ubuntu120.04.txt) | hadoop-hdfs-project in the patch failed with JDK Ubuntu-11.0.21+9-post-Ubuntu-0ubuntu120.04. | | -1 :x: | compile | 1m 15s | [/patch-compile-hadoop-hdfs-project-jdkPrivateBuild-1.8.0_392-8u392-ga-1~20.04-b08.txt](https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-6583/1/artifact/out/patch-compile-hadoop-hdfs-project-jdkPrivateBuild-1.8.0_392-8u392-ga-1~20.04-b08.txt) | hadoop-hdfs-project in the patch failed with JDK Private Build-1.8.0_392-8u392-ga-1~20.04-b08. | | -1 :x: | cc | 1m 15s | [/patch-compile-hadoop-hdfs-project-jdkPrivateBuild-1.8.0_392-8u392-ga-1~20.04-b08.txt](https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-6583/1/artifact/out/patch-compile-hadoop-hdfs-project-jdkPrivateBuild-1.8.0_392-8u392-ga-1~20.04-b08.txt)
[jira] [Commented] (HDFS-17396) BootstrapStandby should download rollback image during RollingUpgrade
[ https://issues.apache.org/jira/browse/HDFS-17396?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17820250#comment-17820250 ] ASF GitHub Bot commented on HDFS-17396: --- goiri commented on code in PR #6583: URL: https://github.com/apache/hadoop/pull/6583#discussion_r1501301861 ## hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/FSImageTestUtil.java: ## @@ -519,6 +519,23 @@ public static void assertNNHasCheckpoints(MiniDFSCluster cluster, } } + public static void assertNNHasRollbackCheckpoints(MiniDFSCluster cluster, + int nnIdx, List txids) { + +for (File nameDir : getNameNodeCurrentDirs(cluster, nnIdx)) { + LOG.info("examining name dir with files: " + + Joiner.on(",").join(nameDir.listFiles())); + // Should have fsimage_N for the three checkpoints + LOG.info("Examining storage dir " + nameDir + " with contents: " + + StringUtils.join(nameDir.listFiles(), ", ")); + for (long checkpointTxId : txids) { +File image = new File(nameDir, +NNStorage.getRollbackImageFileName(checkpointTxId)); +assertTrue("Expected non-empty " + image, image.length() > 0); Review Comment: OK, I see this pattern already exists. ## hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/ha/TestBootstrapStandby.java: ## @@ -208,12 +210,21 @@ public void testRollingUpgradeBootstrapStandby() throws Exception { // BootstrapStandby should fail if the node has a future version // and the cluster isn't in rolling upgrade -bs.setConf(cluster.getConfiguration(1)); +bs.setConf(cluster.getConfiguration(2)); assertEquals("BootstrapStandby should return ERR_CODE_INVALID_VERSION", ERR_CODE_INVALID_VERSION, bs.run(new String[]{"-force"})); // Start rolling upgrade fs.rollingUpgrade(RollingUpgradeAction.PREPARE); +RollingUpgradeInfo info = fs.rollingUpgrade(RollingUpgradeAction.QUERY); +while (!info.createdRollbackImages()) { Review Comment: This is unbounded. Can you use LambdaTestUtils#wait or similar? ## hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/ha/TestBootstrapStandby.java: ## @@ -253,6 +269,13 @@ public void testRollingUpgradeBootstrapStandby() throws Exception { // We should now be able to start the standby successfully restartNameNodesFromIndex(1, "-rollingUpgrade", "started"); +for (int i = 1; i < maxNNCount; i++) { + assertTrue("NameNodes should all have the rollback FSImage", + cluster.getNameNode(i).getFSImage().hasRollbackFSImage()); Review Comment: Easier to read if you extract the `nn = cluster.getNameNode(i);` ## hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/FSImageTestUtil.java: ## @@ -519,6 +519,23 @@ public static void assertNNHasCheckpoints(MiniDFSCluster cluster, } } + public static void assertNNHasRollbackCheckpoints(MiniDFSCluster cluster, + int nnIdx, List txids) { + +for (File nameDir : getNameNodeCurrentDirs(cluster, nnIdx)) { + LOG.info("examining name dir with files: " + Review Comment: Can you use the logger format? ## hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/FSImageTestUtil.java: ## @@ -519,6 +519,23 @@ public static void assertNNHasCheckpoints(MiniDFSCluster cluster, } } + public static void assertNNHasRollbackCheckpoints(MiniDFSCluster cluster, + int nnIdx, List txids) { + +for (File nameDir : getNameNodeCurrentDirs(cluster, nnIdx)) { + LOG.info("examining name dir with files: " + + Joiner.on(",").join(nameDir.listFiles())); + // Should have fsimage_N for the three checkpoints + LOG.info("Examining storage dir " + nameDir + " with contents: " + + StringUtils.join(nameDir.listFiles(), ", ")); + for (long checkpointTxId : txids) { +File image = new File(nameDir, +NNStorage.getRollbackImageFileName(checkpointTxId)); +assertTrue("Expected non-empty " + image, image.length() > 0); Review Comment: An assert in the regular code? Should this be an exception? > BootstrapStandby should download rollback image during RollingUpgrade > -- > > Key: HDFS-17396 > URL: https://issues.apache.org/jira/browse/HDFS-17396 > Project: Hadoop HDFS > Issue Type: Bug > Components: namenode >Affects Versions: 3.4.0 >Reporter: Danny Becker >Assignee: Danny Becker >Priority: Minor > Labels: pull-requests-available > Fix For: 3.4.0 > > > Due to [HDFS-1
[jira] [Commented] (HDFS-17396) BootstrapStandby should download rollback image during RollingUpgrade
[ https://issues.apache.org/jira/browse/HDFS-17396?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17820249#comment-17820249 ] ASF GitHub Bot commented on HDFS-17396: --- dannytbecker opened a new pull request, #6583: URL: https://github.com/apache/hadoop/pull/6583 ### Description of PR Due to [HDFS-17178](https://issues.apache.org/jira/browse/HDFS-17178), BootstrapStandby can now succeed during a RollingUpgrade. This can lead to an edge case where the bootstrapped NameNode does not have the fsimage_rollback_N file locally so it does not recognize that it is in a RollingUpgrade. This means the bootstrapped NameNode will overwrite the VERSION file during a checkpoint because it does not know that the NameService is in a RollingUpgrade. BootstrapStandby needs to download the fsimage_rollback_N file in addition to the normal fsimage_N file so that the bootstrapped node correctly recognizes the RollingUpgrade state of the NameService. ### How was this patch tested? Added unit tests to cover the scenario and verify the rollback image is downloaded. ### For code changes: - [x] Does the title or this PR starts with the corresponding JIRA issue id (e.g. 'HADOOP-17799. Your PR title ...')? - [ ] Object storage: have the integration tests been executed and the endpoint declared according to the connector-specific documentation? - [ ] If adding new dependencies to the code, are these dependencies licensed in a way that is compatible for inclusion under [ASF 2.0](http://www.apache.org/legal/resolved.html#category-a)? - [ ] If applicable, have you updated the `LICENSE`, `LICENSE-binary`, `NOTICE-binary` files? > BootstrapStandby should download rollback image during RollingUpgrade > -- > > Key: HDFS-17396 > URL: https://issues.apache.org/jira/browse/HDFS-17396 > Project: Hadoop HDFS > Issue Type: Bug > Components: namenode >Affects Versions: 3.4.0 >Reporter: Danny Becker >Assignee: Danny Becker >Priority: Minor > Fix For: 3.4.0 > > > Due to [HDFS-17178|https://issues.apache.org/jira/browse/HDFS-17178], > BootstrapStandby can now succeed during a RollingUpgrade. This can lead to an > edge case where the bootstrapped NameNode does not have the > fsimage_rollback_N file locally so it does not recognize that it is in a > RollingUpgrade. This means the bootstrapped NameNode will overwrite the > VERSION file during a checkpoint because it does not know that the > NameService is in a RollingUpgrade. > > BootstrapStandby needs to download the fsimage_rollback_N file in addition to > the normal fsimage_N file so that the bootstrapped node correctly recognizes > the RollingUpgrade state of the NameService. -- This message was sent by Atlassian Jira (v8.20.10#820010) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-17381) Distcp of EC files should not be limited to DFS.
[ https://issues.apache.org/jira/browse/HDFS-17381?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17820234#comment-17820234 ] ASF GitHub Bot commented on HDFS-17381: --- hadoop-yetus commented on PR #6551: URL: https://github.com/apache/hadoop/pull/6551#issuecomment-1962171606 :broken_heart: **-1 overall** | Vote | Subsystem | Runtime | Logfile | Comment | |::|--:|:|::|:---:| | +0 :ok: | reexec | 19m 9s | | Docker mode activated. | _ Prechecks _ | | +1 :green_heart: | dupname | 0m 0s | | No case conflicting files found. | | +0 :ok: | codespell | 0m 1s | | codespell was not available. | | +0 :ok: | detsecrets | 0m 1s | | detect-secrets was not available. | | +1 :green_heart: | @author | 0m 0s | | The patch does not contain any @author tags. | | -1 :x: | test4tests | 0m 0s | | The patch doesn't appear to include any new or modified tests. Please justify why no new tests are needed for this patch. Also please list what manual steps were performed to verify this patch. | _ trunk Compile Tests _ | | +0 :ok: | mvndep | 14m 18s | | Maven dependency ordering for branch | | +1 :green_heart: | mvninstall | 36m 48s | | trunk passed | | +1 :green_heart: | compile | 19m 48s | | trunk passed with JDK Ubuntu-11.0.21+9-post-Ubuntu-0ubuntu120.04 | | +1 :green_heart: | compile | 18m 51s | | trunk passed with JDK Private Build-1.8.0_392-8u392-ga-1~20.04-b08 | | +1 :green_heart: | checkstyle | 4m 46s | | trunk passed | | +1 :green_heart: | mvnsite | 3m 40s | | trunk passed | | +1 :green_heart: | javadoc | 2m 57s | | trunk passed with JDK Ubuntu-11.0.21+9-post-Ubuntu-0ubuntu120.04 | | +1 :green_heart: | javadoc | 2m 26s | | trunk passed with JDK Private Build-1.8.0_392-8u392-ga-1~20.04-b08 | | -1 :x: | spotbugs | 2m 36s | [/branch-spotbugs-hadoop-common-project_hadoop-common-warnings.html](https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-6551/2/artifact/out/branch-spotbugs-hadoop-common-project_hadoop-common-warnings.html) | hadoop-common-project/hadoop-common in trunk has 1 extant spotbugs warnings. | | -1 :x: | spotbugs | 2m 49s | [/branch-spotbugs-hadoop-hdfs-project_hadoop-hdfs-client-warnings.html](https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-6551/2/artifact/out/branch-spotbugs-hadoop-hdfs-project_hadoop-hdfs-client-warnings.html) | hadoop-hdfs-project/hadoop-hdfs-client in trunk has 1 extant spotbugs warnings. | | +1 :green_heart: | shadedclient | 38m 56s | | branch has no errors when building and testing our client artifacts. | _ Patch Compile Tests _ | | +0 :ok: | mvndep | 0m 31s | | Maven dependency ordering for patch | | +1 :green_heart: | mvninstall | 2m 8s | | the patch passed | | +1 :green_heart: | compile | 18m 55s | | the patch passed with JDK Ubuntu-11.0.21+9-post-Ubuntu-0ubuntu120.04 | | +1 :green_heart: | javac | 18m 55s | | the patch passed | | +1 :green_heart: | compile | 17m 52s | | the patch passed with JDK Private Build-1.8.0_392-8u392-ga-1~20.04-b08 | | +1 :green_heart: | javac | 17m 52s | | the patch passed | | +1 :green_heart: | blanks | 0m 0s | | The patch has no blanks issues. | | -0 :warning: | checkstyle | 4m 42s | [/results-checkstyle-root.txt](https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-6551/2/artifact/out/results-checkstyle-root.txt) | root: The patch generated 7 new + 29 unchanged - 0 fixed = 36 total (was 29) | | +1 :green_heart: | mvnsite | 3m 35s | | the patch passed | | -1 :x: | javadoc | 1m 11s | [/results-javadoc-javadoc-hadoop-common-project_hadoop-common-jdkUbuntu-11.0.21+9-post-Ubuntu-0ubuntu120.04.txt](https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-6551/2/artifact/out/results-javadoc-javadoc-hadoop-common-project_hadoop-common-jdkUbuntu-11.0.21+9-post-Ubuntu-0ubuntu120.04.txt) | hadoop-common-project_hadoop-common-jdkUbuntu-11.0.21+9-post-Ubuntu-0ubuntu120.04 with JDK Ubuntu-11.0.21+9-post-Ubuntu-0ubuntu120.04 generated 11 new + 0 unchanged - 0 fixed = 11 total (was 0) | | +1 :green_heart: | javadoc | 2m 29s | | the patch passed with JDK Private Build-1.8.0_392-8u392-ga-1~20.04-b08 | | +1 :green_heart: | spotbugs | 6m 55s | | the patch passed | | +1 :green_heart: | shadedclient | 38m 48s | | patch has no errors when building and testing our client artifacts. | _ Other Tests _ | | +1 :green_heart: | unit | 19m 23s | | hadoop-common in the patch passed. | | +1 :green_heart: | unit | 2m 42s | | hadoop-hdfs-client in the patch passed. | | +1 :green_heart: | unit | 16m 18s | | hadoop-distcp in the patch passed
[jira] [Commented] (HDFS-17381) Distcp of EC files should not be limited to DFS.
[ https://issues.apache.org/jira/browse/HDFS-17381?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17820228#comment-17820228 ] ASF GitHub Bot commented on HDFS-17381: --- hadoop-yetus commented on PR #6551: URL: https://github.com/apache/hadoop/pull/6551#issuecomment-1962143806 :broken_heart: **-1 overall** | Vote | Subsystem | Runtime | Logfile | Comment | |::|--:|:|::|:---:| | +0 :ok: | reexec | 0m 36s | | Docker mode activated. | _ Prechecks _ | | +1 :green_heart: | dupname | 0m 0s | | No case conflicting files found. | | +0 :ok: | codespell | 0m 0s | | codespell was not available. | | +0 :ok: | detsecrets | 0m 0s | | detect-secrets was not available. | | +1 :green_heart: | @author | 0m 0s | | The patch does not contain any @author tags. | | -1 :x: | test4tests | 0m 0s | | The patch doesn't appear to include any new or modified tests. Please justify why no new tests are needed for this patch. Also please list what manual steps were performed to verify this patch. | _ trunk Compile Tests _ | | +0 :ok: | mvndep | 14m 41s | | Maven dependency ordering for branch | | +1 :green_heart: | mvninstall | 35m 47s | | trunk passed | | +1 :green_heart: | compile | 17m 44s | | trunk passed with JDK Ubuntu-11.0.21+9-post-Ubuntu-0ubuntu120.04 | | +1 :green_heart: | compile | 16m 26s | | trunk passed with JDK Private Build-1.8.0_392-8u392-ga-1~20.04-b08 | | +1 :green_heart: | checkstyle | 4m 24s | | trunk passed | | +1 :green_heart: | mvnsite | 3m 47s | | trunk passed | | +1 :green_heart: | javadoc | 3m 2s | | trunk passed with JDK Ubuntu-11.0.21+9-post-Ubuntu-0ubuntu120.04 | | +1 :green_heart: | javadoc | 2m 37s | | trunk passed with JDK Private Build-1.8.0_392-8u392-ga-1~20.04-b08 | | -1 :x: | spotbugs | 2m 38s | [/branch-spotbugs-hadoop-common-project_hadoop-common-warnings.html](https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-6551/3/artifact/out/branch-spotbugs-hadoop-common-project_hadoop-common-warnings.html) | hadoop-common-project/hadoop-common in trunk has 1 extant spotbugs warnings. | | -1 :x: | spotbugs | 2m 48s | [/branch-spotbugs-hadoop-hdfs-project_hadoop-hdfs-client-warnings.html](https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-6551/3/artifact/out/branch-spotbugs-hadoop-hdfs-project_hadoop-hdfs-client-warnings.html) | hadoop-hdfs-project/hadoop-hdfs-client in trunk has 1 extant spotbugs warnings. | | +1 :green_heart: | shadedclient | 33m 24s | | branch has no errors when building and testing our client artifacts. | _ Patch Compile Tests _ | | +0 :ok: | mvndep | 0m 32s | | Maven dependency ordering for patch | | +1 :green_heart: | mvninstall | 2m 7s | | the patch passed | | +1 :green_heart: | compile | 17m 27s | | the patch passed with JDK Ubuntu-11.0.21+9-post-Ubuntu-0ubuntu120.04 | | +1 :green_heart: | javac | 17m 27s | | the patch passed | | +1 :green_heart: | compile | 16m 35s | | the patch passed with JDK Private Build-1.8.0_392-8u392-ga-1~20.04-b08 | | +1 :green_heart: | javac | 16m 35s | | the patch passed | | +1 :green_heart: | blanks | 0m 0s | | The patch has no blanks issues. | | -0 :warning: | checkstyle | 4m 17s | [/results-checkstyle-root.txt](https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-6551/3/artifact/out/results-checkstyle-root.txt) | root: The patch generated 7 new + 28 unchanged - 0 fixed = 35 total (was 28) | | +1 :green_heart: | mvnsite | 3m 37s | | the patch passed | | -1 :x: | javadoc | 1m 11s | [/results-javadoc-javadoc-hadoop-common-project_hadoop-common-jdkUbuntu-11.0.21+9-post-Ubuntu-0ubuntu120.04.txt](https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-6551/3/artifact/out/results-javadoc-javadoc-hadoop-common-project_hadoop-common-jdkUbuntu-11.0.21+9-post-Ubuntu-0ubuntu120.04.txt) | hadoop-common-project_hadoop-common-jdkUbuntu-11.0.21+9-post-Ubuntu-0ubuntu120.04 with JDK Ubuntu-11.0.21+9-post-Ubuntu-0ubuntu120.04 generated 11 new + 0 unchanged - 0 fixed = 11 total (was 0) | | +1 :green_heart: | javadoc | 2m 35s | | the patch passed with JDK Private Build-1.8.0_392-8u392-ga-1~20.04-b08 | | +1 :green_heart: | spotbugs | 6m 56s | | the patch passed | | +1 :green_heart: | shadedclient | 33m 31s | | patch has no errors when building and testing our client artifacts. | _ Other Tests _ | | +1 :green_heart: | unit | 19m 28s | | hadoop-common in the patch passed. | | +1 :green_heart: | unit | 2m 44s | | hadoop-hdfs-client in the patch passed. | | +1 :green_heart: | unit | 15m 22s | | hadoop-distcp in the patch passed
[jira] [Commented] (HDFS-17382) Add Apache Log4j Extras Library to Hadoop 3.3 for Enhanced Log Rolling Capabilities
[ https://issues.apache.org/jira/browse/HDFS-17382?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17820154#comment-17820154 ] Steve Loughran commented on HDFS-17382: --- we are still shipping reload4j with 3.4.x, so perhaps you should target branch-3.4 first > Add Apache Log4j Extras Library to Hadoop 3.3 for Enhanced Log Rolling > Capabilities > --- > > Key: HDFS-17382 > URL: https://issues.apache.org/jira/browse/HDFS-17382 > Project: Hadoop HDFS > Issue Type: Improvement >Affects Versions: 3.3.6 >Reporter: woosuk.ro >Priority: Minor > Labels: pull-request-available > > In the current Hadoop 3.3 version, the system relies on Log4j 1.x for logging > purposes. This dependency limits the logging functionality, especially when > it comes to rolling log files such as 'hdfs-audit.log'. Rolling log files is > crucial for managing log size and ensuring logs are rotated out over time to > prevent excessive disk space usage. However, the Log4j 1.x version integrated > within Hadoop lacks the necessary capabilities to efficiently handle log > rolling. > This library extends logging capabilities, including more flexible and > configurable log rolling features. By deploying this library, we can enable > advanced rolling strategies such as time-based rolling, size-based rolling, > and compression of rolled logs, which are not supported by the default Log4j > 1.x setup in Hadoop. > The integration of Apache Log4j Extras into Hadoop will significantly improve > log management by allowing for more sophisticated and configurable log > rotation policies. This enhancement is particularly important for maintaining > system performance and reliability, as well as for compliance with log > retention policies. > Although there are plans to upgrade to Log4j 2 in the forthcoming Hadoop 3.5 > version, which will inherently solve these issues by providing enhanced > logging features, there is an immediate need to enable advanced log rolling > capabilities in the current and previous versions of Hadoop. -- This message was sent by Atlassian Jira (v8.20.10#820010) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-17382) Add Apache Log4j Extras Library to Hadoop 3.3 for Enhanced Log Rolling Capabilities
[ https://issues.apache.org/jira/browse/HDFS-17382?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Steve Loughran updated HDFS-17382: -- Affects Version/s: 3.3.6 (was: 3.3.4) > Add Apache Log4j Extras Library to Hadoop 3.3 for Enhanced Log Rolling > Capabilities > --- > > Key: HDFS-17382 > URL: https://issues.apache.org/jira/browse/HDFS-17382 > Project: Hadoop HDFS > Issue Type: Improvement >Affects Versions: 3.3.6 >Reporter: woosuk.ro >Priority: Minor > Labels: pull-request-available > > In the current Hadoop 3.3 version, the system relies on Log4j 1.x for logging > purposes. This dependency limits the logging functionality, especially when > it comes to rolling log files such as 'hdfs-audit.log'. Rolling log files is > crucial for managing log size and ensuring logs are rotated out over time to > prevent excessive disk space usage. However, the Log4j 1.x version integrated > within Hadoop lacks the necessary capabilities to efficiently handle log > rolling. > This library extends logging capabilities, including more flexible and > configurable log rolling features. By deploying this library, we can enable > advanced rolling strategies such as time-based rolling, size-based rolling, > and compression of rolled logs, which are not supported by the default Log4j > 1.x setup in Hadoop. > The integration of Apache Log4j Extras into Hadoop will significantly improve > log management by allowing for more sophisticated and configurable log > rotation policies. This enhancement is particularly important for maintaining > system performance and reliability, as well as for compliance with log > retention policies. > Although there are plans to upgrade to Log4j 2 in the forthcoming Hadoop 3.5 > version, which will inherently solve these issues by providing enhanced > logging features, there is an immediate need to enable advanced log rolling > capabilities in the current and previous versions of Hadoop. -- This message was sent by Atlassian Jira (v8.20.10#820010) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Updated] (HDFS-17382) Add Apache Log4j Extras Library to Hadoop 3.3 for Enhanced Log Rolling Capabilities
[ https://issues.apache.org/jira/browse/HDFS-17382?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Steve Loughran updated HDFS-17382: -- Target Version/s: (was: 3.3.4) > Add Apache Log4j Extras Library to Hadoop 3.3 for Enhanced Log Rolling > Capabilities > --- > > Key: HDFS-17382 > URL: https://issues.apache.org/jira/browse/HDFS-17382 > Project: Hadoop HDFS > Issue Type: Improvement >Affects Versions: 3.3.4 >Reporter: woosuk.ro >Priority: Minor > Labels: pull-request-available > > In the current Hadoop 3.3 version, the system relies on Log4j 1.x for logging > purposes. This dependency limits the logging functionality, especially when > it comes to rolling log files such as 'hdfs-audit.log'. Rolling log files is > crucial for managing log size and ensuring logs are rotated out over time to > prevent excessive disk space usage. However, the Log4j 1.x version integrated > within Hadoop lacks the necessary capabilities to efficiently handle log > rolling. > This library extends logging capabilities, including more flexible and > configurable log rolling features. By deploying this library, we can enable > advanced rolling strategies such as time-based rolling, size-based rolling, > and compression of rolled logs, which are not supported by the default Log4j > 1.x setup in Hadoop. > The integration of Apache Log4j Extras into Hadoop will significantly improve > log management by allowing for more sophisticated and configurable log > rotation policies. This enhancement is particularly important for maintaining > system performance and reliability, as well as for compliance with log > retention policies. > Although there are plans to upgrade to Log4j 2 in the forthcoming Hadoop 3.5 > version, which will inherently solve these issues by providing enhanced > logging features, there is an immediate need to enable advanced log rolling > capabilities in the current and previous versions of Hadoop. -- This message was sent by Atlassian Jira (v8.20.10#820010) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-17394) [FGL] Remove unused WriteHoldCount of FSNamesystemLock
[ https://issues.apache.org/jira/browse/HDFS-17394?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17820076#comment-17820076 ] ASF GitHub Bot commented on HDFS-17394: --- xinglin commented on PR #6571: URL: https://github.com/apache/hadoop/pull/6571#issuecomment-1961467118 can we trigger a new build? > [FGL] Remove unused WriteHoldCount of FSNamesystemLock > -- > > Key: HDFS-17394 > URL: https://issues.apache.org/jira/browse/HDFS-17394 > Project: Hadoop HDFS > Issue Type: Sub-task >Reporter: ZanderXu >Assignee: ZanderXu >Priority: Major > Labels: pull-request-available > > {code:java} > public int getWriteHoldCount() { > return this.fsLock.getWriteHoldCount(FSNamesystemLockMode.GLOBAL); > } > @Deprecated // dirLock is obsolete, use namesystem.fsLock instead > public int getWriteHoldCount() { > return namesystem.getWriteHoldCount(); > } > // sanity check. > if (!hadDirReadLock || !hadFsnReadLock || hadDirWriteLock || > hadFsnWriteLock || dir.getReadHoldCount() != 1 || > fsn.getReadHoldCount() != 1) { > // cannot relinquish > return false; > } {code} > getWriteHoldCount in FSNamesystem.java and FSDirectory.java is unused. > dir.getReadHoldCount() is useless as it's same as fsn.getReadHoldCount(). -- This message was sent by Atlassian Jira (v8.20.10#820010) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-17365) EC: Add extra redunency configuration in checkStreamerFailures to prevent data loss.
[ https://issues.apache.org/jira/browse/HDFS-17365?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17820051#comment-17820051 ] ASF GitHub Bot commented on HDFS-17365: --- hadoop-yetus commented on PR #6517: URL: https://github.com/apache/hadoop/pull/6517#issuecomment-1961295707 :broken_heart: **-1 overall** | Vote | Subsystem | Runtime | Logfile | Comment | |::|--:|:|::|:---:| | +0 :ok: | reexec | 6m 47s | | Docker mode activated. | _ Prechecks _ | | +1 :green_heart: | dupname | 0m 0s | | No case conflicting files found. | | +0 :ok: | codespell | 0m 1s | | codespell was not available. | | +0 :ok: | detsecrets | 0m 1s | | detect-secrets was not available. | | +0 :ok: | xmllint | 0m 1s | | xmllint was not available. | | +1 :green_heart: | @author | 0m 0s | | The patch does not contain any @author tags. | | -1 :x: | test4tests | 0m 0s | | The patch doesn't appear to include any new or modified tests. Please justify why no new tests are needed for this patch. Also please list what manual steps were performed to verify this patch. | _ trunk Compile Tests _ | | +0 :ok: | mvndep | 13m 58s | | Maven dependency ordering for branch | | +1 :green_heart: | mvninstall | 19m 45s | | trunk passed | | +1 :green_heart: | compile | 2m 57s | | trunk passed with JDK Ubuntu-11.0.21+9-post-Ubuntu-0ubuntu120.04 | | +1 :green_heart: | compile | 2m 50s | | trunk passed with JDK Private Build-1.8.0_392-8u392-ga-1~20.04-b08 | | +1 :green_heart: | checkstyle | 0m 46s | | trunk passed | | +1 :green_heart: | mvnsite | 1m 18s | | trunk passed | | +1 :green_heart: | javadoc | 1m 8s | | trunk passed with JDK Ubuntu-11.0.21+9-post-Ubuntu-0ubuntu120.04 | | +1 :green_heart: | javadoc | 1m 35s | | trunk passed with JDK Private Build-1.8.0_392-8u392-ga-1~20.04-b08 | | -1 :x: | spotbugs | 1m 28s | [/branch-spotbugs-hadoop-hdfs-project_hadoop-hdfs-client-warnings.html](https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-6517/4/artifact/out/branch-spotbugs-hadoop-hdfs-project_hadoop-hdfs-client-warnings.html) | hadoop-hdfs-project/hadoop-hdfs-client in trunk has 1 extant spotbugs warnings. | | +1 :green_heart: | shadedclient | 20m 29s | | branch has no errors when building and testing our client artifacts. | _ Patch Compile Tests _ | | +0 :ok: | mvndep | 0m 21s | | Maven dependency ordering for patch | | +1 :green_heart: | mvninstall | 1m 2s | | the patch passed | | +1 :green_heart: | compile | 2m 51s | | the patch passed with JDK Ubuntu-11.0.21+9-post-Ubuntu-0ubuntu120.04 | | +1 :green_heart: | javac | 2m 51s | | the patch passed | | +1 :green_heart: | compile | 2m 43s | | the patch passed with JDK Private Build-1.8.0_392-8u392-ga-1~20.04-b08 | | +1 :green_heart: | javac | 2m 43s | | the patch passed | | -1 :x: | blanks | 0m 0s | [/blanks-eol.txt](https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-6517/4/artifact/out/blanks-eol.txt) | The patch has 2 line(s) that end in blanks. Use git apply --whitespace=fix <>. Refer https://git-scm.com/docs/git-apply | | -0 :warning: | checkstyle | 0m 39s | [/results-checkstyle-hadoop-hdfs-project.txt](https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-6517/4/artifact/out/results-checkstyle-hadoop-hdfs-project.txt) | hadoop-hdfs-project: The patch generated 28 new + 33 unchanged - 0 fixed = 61 total (was 33) | | +1 :green_heart: | mvnsite | 1m 7s | | the patch passed | | +1 :green_heart: | javadoc | 0m 55s | | the patch passed with JDK Ubuntu-11.0.21+9-post-Ubuntu-0ubuntu120.04 | | +1 :green_heart: | javadoc | 1m 25s | | the patch passed with JDK Private Build-1.8.0_392-8u392-ga-1~20.04-b08 | | +1 :green_heart: | spotbugs | 3m 18s | | the patch passed | | +1 :green_heart: | shadedclient | 23m 32s | | patch has no errors when building and testing our client artifacts. | _ Other Tests _ | | +1 :green_heart: | unit | 1m 49s | | hadoop-hdfs-client in the patch passed. | | -1 :x: | unit | 84m 49s | [/patch-unit-hadoop-hdfs-project_hadoop-hdfs.txt](https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-6517/4/artifact/out/patch-unit-hadoop-hdfs-project_hadoop-hdfs.txt) | hadoop-hdfs in the patch passed. | | +0 :ok: | asflicense | 0m 24s | | ASF License check generated no output? | | | | 201m 0s | | | | Reason | Tests | |---:|:--| | Failed junit tests | hadoop.hdfs.shortcircuit.TestShortCircuitCache | | | hadoop.hdfs.server.blockmanagement.TestReplicationPolicyExcludeSlowNodes | | | hadoop.hdfs.server.mover.Test
[jira] [Commented] (HDFS-17299) HDFS is not rack failure tolerant while creating a new file.
[ https://issues.apache.org/jira/browse/HDFS-17299?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17819990#comment-17819990 ] ASF GitHub Bot commented on HDFS-17299: --- hadoop-yetus commented on PR #6566: URL: https://github.com/apache/hadoop/pull/6566#issuecomment-1960990799 :broken_heart: **-1 overall** | Vote | Subsystem | Runtime | Logfile | Comment | |::|--:|:|::|:---:| | +0 :ok: | reexec | 20m 49s | | Docker mode activated. | _ Prechecks _ | | +1 :green_heart: | dupname | 0m 0s | | No case conflicting files found. | | +0 :ok: | codespell | 0m 0s | | codespell was not available. | | +0 :ok: | detsecrets | 0m 0s | | detect-secrets was not available. | | +0 :ok: | shelldocs | 0m 0s | | Shelldocs was not available. | | +0 :ok: | xmllint | 0m 0s | | xmllint was not available. | | +1 :green_heart: | @author | 0m 0s | | The patch does not contain any @author tags. | | +1 :green_heart: | test4tests | 0m 0s | | The patch appears to include 4 new or modified test files. | _ trunk Compile Tests _ | | +0 :ok: | mvndep | 14m 45s | | Maven dependency ordering for branch | | +1 :green_heart: | mvninstall | 37m 15s | | trunk passed | | +1 :green_heart: | compile | 21m 0s | | trunk passed with JDK Ubuntu-11.0.21+9-post-Ubuntu-0ubuntu120.04 | | +1 :green_heart: | compile | 18m 51s | | trunk passed with JDK Private Build-1.8.0_392-8u392-ga-1~20.04-b08 | | +1 :green_heart: | checkstyle | 5m 14s | | trunk passed | | +1 :green_heart: | mvnsite | 3m 26s | | trunk passed | | +1 :green_heart: | javadoc | 2m 56s | | trunk passed with JDK Ubuntu-11.0.21+9-post-Ubuntu-0ubuntu120.04 | | +1 :green_heart: | javadoc | 3m 56s | | trunk passed with JDK Private Build-1.8.0_392-8u392-ga-1~20.04-b08 | | +0 :ok: | spotbugs | 0m 41s | | branch/hadoop-project no spotbugs output file (spotbugsXml.xml) | | -1 :x: | spotbugs | 2m 52s | [/branch-spotbugs-hadoop-hdfs-project_hadoop-hdfs-client-warnings.html](https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-6566/7/artifact/out/branch-spotbugs-hadoop-hdfs-project_hadoop-hdfs-client-warnings.html) | hadoop-hdfs-project/hadoop-hdfs-client in trunk has 1 extant spotbugs warnings. | | -1 :x: | shadedclient | 5m 15s | | branch has errors when building and testing our client artifacts. | _ Patch Compile Tests _ | | +0 :ok: | mvndep | 0m 33s | | Maven dependency ordering for patch | | +1 :green_heart: | mvninstall | 2m 24s | | the patch passed | | +1 :green_heart: | compile | 18m 31s | | the patch passed with JDK Ubuntu-11.0.21+9-post-Ubuntu-0ubuntu120.04 | | +1 :green_heart: | javac | 18m 31s | | the patch passed | | +1 :green_heart: | compile | 17m 22s | | the patch passed with JDK Private Build-1.8.0_392-8u392-ga-1~20.04-b08 | | +1 :green_heart: | javac | 17m 22s | | the patch passed | | +1 :green_heart: | blanks | 0m 0s | | The patch has no blanks issues. | | -0 :warning: | checkstyle | 4m 41s | [/results-checkstyle-root.txt](https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-6566/7/artifact/out/results-checkstyle-root.txt) | root: The patch generated 17 new + 244 unchanged - 2 fixed = 261 total (was 246) | | +1 :green_heart: | mvnsite | 3m 31s | | the patch passed | | +1 :green_heart: | shellcheck | 0m 1s | | No new issues. | | +1 :green_heart: | javadoc | 2m 53s | | the patch passed with JDK Ubuntu-11.0.21+9-post-Ubuntu-0ubuntu120.04 | | +1 :green_heart: | javadoc | 3m 27s | | the patch passed with JDK Private Build-1.8.0_392-8u392-ga-1~20.04-b08 | | +0 :ok: | spotbugs | 0m 37s | | hadoop-project has no data from spotbugs | | +1 :green_heart: | shadedclient | 39m 1s | | patch has no errors when building and testing our client artifacts. | _ Other Tests _ | | +1 :green_heart: | unit | 0m 37s | | hadoop-project in the patch passed. | | +1 :green_heart: | unit | 2m 47s | | hadoop-hdfs-client in the patch passed. | | -1 :x: | unit | 275m 42s | [/patch-unit-hadoop-hdfs-project_hadoop-hdfs.txt](https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-6566/7/artifact/out/patch-unit-hadoop-hdfs-project_hadoop-hdfs.txt) | hadoop-hdfs in the patch passed. | | +1 :green_heart: | asflicense | 1m 14s | | The patch does not generate ASF License warnings. | | | | 525m 10s | | | | Reason | Tests | |---:|:--| | Failed junit tests | hadoop.hdfs.protocol.TestBlockListAsLongs | | | hadoop.hdfs.server.datanode.TestLargeBlockReport | | | hadoop.hdfs.tools.TestDFSAdmin | | | hadoop.h