Apache Hadoop qbt Report: trunk+JDK8 on Linux/x86
For more details, see https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/1477/ [Apr 21, 2020 6:31:33 AM] (pjoseph) YARN-10240. Prevent Fatal CancelledException in TimelineV2Client when [Apr 21, 2020 4:27:29 PM] (github) HADOOP-16965. Refactor abfs stream configuration. (#1956) [Apr 21, 2020 4:37:40 PM] (github) HADOOP-16922. ABFS: Change User-Agent header (#1938) -1 overall The following subsystems voted -1: asflicense compile findbugs mvnsite pathlen unit xml The following subsystems voted -1 but were configured to be filtered/ignored: cc checkstyle javac javadoc pylint shellcheck shelldocs whitespace The following subsystems are considered long running: (runtime bigger than 1h 0m 0s) unit Specific tests: XML : Parsing Error(s): hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/test/resources/nvidia-smi-output-excerpt.xml hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/test/resources/nvidia-smi-output-missing-tags.xml hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/test/resources/nvidia-smi-output-missing-tags2.xml hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/test/resources/nvidia-smi-sample-output.xml hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/resources/fair-scheduler-invalid.xml hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/resources/yarn-site-with-invalid-allocation-file-ref.xml FindBugs : module:hadoop-hdfs-project/hadoop-hdfs Possible null pointer dereference of effectiveDirective in org.apache.hadoop.hdfs.server.namenode.FSNamesystem.addCacheDirective(CacheDirectiveInfo, EnumSet, boolean) Dereferenced at FSNamesystem.java:effectiveDirective in org.apache.hadoop.hdfs.server.namenode.FSNamesystem.addCacheDirective(CacheDirectiveInfo, EnumSet, boolean) Dereferenced at FSNamesystem.java:[line 7444] Possible null pointer dereference of ret in org.apache.hadoop.hdfs.server.namenode.FSNamesystem.renameTo(String, String, boolean) Dereferenced at FSNamesystem.java:ret in org.apache.hadoop.hdfs.server.namenode.FSNamesystem.renameTo(String, String, boolean) Dereferenced at FSNamesystem.java:[line 3213] Possible null pointer dereference of res in org.apache.hadoop.hdfs.server.namenode.FSNamesystem.renameTo(String, String, boolean, Options$Rename[]) Dereferenced at FSNamesystem.java:res in org.apache.hadoop.hdfs.server.namenode.FSNamesystem.renameTo(String, String, boolean, Options$Rename[]) Dereferenced at FSNamesystem.java:[line 3248] FindBugs : module:hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-common org.apache.hadoop.yarn.server.webapp.WebServiceClient.sslFactory should be package protected At WebServiceClient.java: At WebServiceClient.java:[line 42] FindBugs : module:hadoop-cloud-storage-project/hadoop-cos org.apache.hadoop.fs.cosn.CosNInputStream$ReadBuffer.getBuffer() may expose internal representation by returning CosNInputStream$ReadBuffer.buffer At CosNInputStream.java:by returning CosNInputStream$ReadBuffer.buffer At CosNInputStream.java:[line 87] Failed junit tests : hadoop.hdfs.server.namenode.ha.TestConfiguredFailoverProxyProvider hadoop.hdfs.server.balancer.TestBalancer hadoop.yarn.applications.distributedshell.TestDistributedShell compile: https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/1477/artifact/out/patch-compile-root.txt [724K] cc: https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/1477/artifact/out/patch-compile-root.txt [724K] javac: https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/1477/artifact/out/patch-compile-root.txt [724K] checkstyle: https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/1477/artifact/out/diff-checkstyle-root.txt [16M] mvnsite: https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/1477/artifact/out/patch-mvnsite-root.txt [284K] pathlen: https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/1477/artifact/out/pathlen.txt [12K] pylint: https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/1477/artifact/out/diff-patch-pylint.txt [24K] shellcheck: https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/1477/artifact/out/diff-patch-shellcheck.txt [16K] shelldocs: https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/1477/artifact/out/diff-patch-shelldocs.txt [96K] whitespace: https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/1477/artifact/out/whitespace-eol.txt [9.9M]
Apache Hadoop qbt Report: branch2.10+JDK7 on Linux/x86
For more details, see https://builds.apache.org/job/hadoop-qbt-branch-2.10-java7-linux-x86/663/ [Apr 21, 2020 12:13:27 PM] (aajisaka) HADOOP-16944. Use Yetus 0.12.0 in GitHub PR (#1917) -1 overall The following subsystems voted -1: asflicense findbugs hadolint pathlen unit xml The following subsystems voted -1 but were configured to be filtered/ignored: cc checkstyle javac javadoc pylint shellcheck shelldocs whitespace The following subsystems are considered long running: (runtime bigger than 1h 0m 0s) unit Specific tests: XML : Parsing Error(s): hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/conf/empty-configuration.xml hadoop-tools/hadoop-azure/src/config/checkstyle-suppressions.xml hadoop-yarn-project/hadoop-yarn/hadoop-yarn-ui/public/crossdomain.xml hadoop-yarn-project/hadoop-yarn/hadoop-yarn-ui/src/main/webapp/public/crossdomain.xml FindBugs : module:hadoop-common-project/hadoop-minikdc Possible null pointer dereference in org.apache.hadoop.minikdc.MiniKdc.delete(File) due to return value of called method Dereferenced at MiniKdc.java:org.apache.hadoop.minikdc.MiniKdc.delete(File) due to return value of called method Dereferenced at MiniKdc.java:[line 515] FindBugs : module:hadoop-common-project/hadoop-auth org.apache.hadoop.security.authentication.server.MultiSchemeAuthenticationHandler.authenticate(HttpServletRequest, HttpServletResponse) makes inefficient use of keySet iterator instead of entrySet iterator At MultiSchemeAuthenticationHandler.java:of keySet iterator instead of entrySet iterator At MultiSchemeAuthenticationHandler.java:[line 192] FindBugs : module:hadoop-common-project/hadoop-common org.apache.hadoop.crypto.CipherSuite.setUnknownValue(int) unconditionally sets the field unknownValue At CipherSuite.java:unknownValue At CipherSuite.java:[line 44] org.apache.hadoop.crypto.CryptoProtocolVersion.setUnknownValue(int) unconditionally sets the field unknownValue At CryptoProtocolVersion.java:unknownValue At CryptoProtocolVersion.java:[line 67] Possible null pointer dereference in org.apache.hadoop.fs.FileUtil.fullyDeleteOnExit(File) due to return value of called method Dereferenced at FileUtil.java:org.apache.hadoop.fs.FileUtil.fullyDeleteOnExit(File) due to return value of called method Dereferenced at FileUtil.java:[line 118] Possible null pointer dereference in org.apache.hadoop.fs.RawLocalFileSystem.handleEmptyDstDirectoryOnWindows(Path, File, Path, File) due to return value of called method Dereferenced at RawLocalFileSystem.java:org.apache.hadoop.fs.RawLocalFileSystem.handleEmptyDstDirectoryOnWindows(Path, File, Path, File) due to return value of called method Dereferenced at RawLocalFileSystem.java:[line 383] Useless condition:lazyPersist == true at this point At CommandWithDestination.java:[line 502] org.apache.hadoop.io.DoubleWritable.compareTo(DoubleWritable) incorrectly handles double value At DoubleWritable.java: At DoubleWritable.java:[line 78] org.apache.hadoop.io.DoubleWritable$Comparator.compare(byte[], int, int, byte[], int, int) incorrectly handles double value At DoubleWritable.java:int) incorrectly handles double value At DoubleWritable.java:[line 97] org.apache.hadoop.io.FloatWritable.compareTo(FloatWritable) incorrectly handles float value At FloatWritable.java: At FloatWritable.java:[line 71] org.apache.hadoop.io.FloatWritable$Comparator.compare(byte[], int, int, byte[], int, int) incorrectly handles float value At FloatWritable.java:int) incorrectly handles float value At FloatWritable.java:[line 89] Possible null pointer dereference in org.apache.hadoop.io.IOUtils.listDirectory(File, FilenameFilter) due to return value of called method Dereferenced at IOUtils.java:org.apache.hadoop.io.IOUtils.listDirectory(File, FilenameFilter) due to return value of called method Dereferenced at IOUtils.java:[line 389] Possible bad parsing of shift operation in org.apache.hadoop.io.file.tfile.Utils$Version.hashCode() At Utils.java:operation in org.apache.hadoop.io.file.tfile.Utils$Version.hashCode() At Utils.java:[line 398] org.apache.hadoop.metrics2.lib.DefaultMetricsFactory.setInstance(MutableMetricsFactory) unconditionally sets the field mmfImpl At DefaultMetricsFactory.java:mmfImpl At DefaultMetricsFactory.java:[line 49] org.apache.hadoop.metrics2.lib.DefaultMetricsSystem.setMiniClusterMode(boolean) unconditionally sets the field miniClusterMode At DefaultMetricsSystem.java:miniClusterMode At DefaultMetricsSystem.java:[line 92] Useless object stored in variable seqOs of method org.apache.hadoop.security.token.delegation.ZKDelegationTokenSecretManager.addOrUpdateToken(AbstractDelegationTokenIdentifier, AbstractDelegationTokenSecretManager$DelegationTokenInformation,
Hadoop 3.1.4 Release Code Freeze
Hi All, The code freeze for Hadoop 3.1.4 release is planned for today. There some issues still open for that version. These are the open issues with 3.1.4 target/fix version: * https://issues.apache.org/jira/browse/YARN-8234 * https://issues.apache.org/jira/browse/HDFS-15205 * https://issues.apache.org/jira/browse/HADOOP-16341 * https://issues.apache.org/jira/browse/HDFS-15272 * https://issues.apache.org/jira/browse/HADOOP-16193 * https://issues.apache.org/jira/browse/YARN-8257 * https://issues.apache.org/jira/browse/YARN-8417 * https://issues.apache.org/jira/browse/HADOOP-16917 Please check if your issue is a blocker. I will move these issues to 3.1.5 target/fix version after 04.23. 10AM CET. I also commented on each issue separately. Release info: * https://cwiki.apache.org/confluence/display/HADOOP/Hadoop+3.1+Release * https://cwiki.apache.org/confluence/display/HADOOP/Roadmap Regards, Gabor - To unsubscribe, e-mail: mapreduce-dev-unsubscr...@hadoop.apache.org For additional commands, e-mail: mapreduce-dev-h...@hadoop.apache.org