[
https://issues.apache.org/jira/browse/HADOOP-10480?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14242167#comment-14242167
]
Hadoop QA commented on HADOOP-10480:
------------------------------------
{color:red}-1 overall{color}. Here are the results of testing the latest
attachment
http://issues.apache.org/jira/secure/attachment/12686383/HADOOP-10480.003.patch
against trunk revision 92916ae.
{color:green}+1 @author{color}. The patch does not contain any @author
tags.
{color:green}+1 tests included{color}. The patch appears to include 3 new
or modified test files.
{color:green}+1 javac{color}. The applied patch does not increase the
total number of javac compiler warnings.
{color:green}+1 javadoc{color}. There were no new javadoc warning messages.
{color:green}+1 eclipse:eclipse{color}. The patch built with
eclipse:eclipse.
{color:green}+1 findbugs{color}. The patch does not introduce any new
Findbugs (version 2.0.3) warnings.
{color:green}+1 release audit{color}. The applied patch does not increase
the total number of release audit warnings.
{color:red}-1 core tests{color}. The patch failed these unit tests in
hadoop-hdfs-project/hadoop-hdfs:
org.apache.hadoop.hdfs.server.balancer.TestBalancer
org.apache.hadoop.hdfs.web.TestWebHDFS
Test results:
https://builds.apache.org/job/PreCommit-HADOOP-Build/5239//testReport/
Console output:
https://builds.apache.org/job/PreCommit-HADOOP-Build/5239//console
This message is automatically generated.
> Fix new findbugs warnings in hadoop-hdfs
> ----------------------------------------
>
> Key: HADOOP-10480
> URL: https://issues.apache.org/jira/browse/HADOOP-10480
> Project: Hadoop Common
> Issue Type: Sub-task
> Reporter: Haohui Mai
> Assignee: Haohui Mai
> Labels: newbie
> Attachments: HADOOP-10480.000.patch, HADOOP-10480.001.patch,
> HADOOP-10480.002.patch, HADOOP-10480.003.patch, HADOOP-10480.2.patch,
> HADOOP-10480.patch
>
>
> The following findbugs warnings need to be fixed:
> {noformat}
> [INFO] --- findbugs-maven-plugin:2.5.3:check (default-cli) @ hadoop-hdfs ---
> [INFO] BugInstance size is 14
> [INFO] Error size is 0
> [INFO] Total bugs: 14
> [INFO] Redundant nullcheck of curPeer, which is known to be non-null in
> org.apache.hadoop.hdfs.BlockReaderFactory.getRemoteBlockReaderFromTcp()
> ["org.apache.hadoop.hdfs.BlockReaderFactory"] At
> BlockReaderFactory.java:[lines 68-808]
> [INFO] Increment of volatile field
> org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.restartingNodeIndex in
> org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.setupPipelineForAppendOrRecovery()
> ["org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer"] At
> DFSOutputStream.java:[lines 308-1492]
> [INFO] Found reliance on default encoding in
> org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(DataOutputStream,
> DataInputStream, DataOutputStream, String, DataTransferThrottler,
> DatanodeInfo[]): new java.io.FileWriter(File)
> ["org.apache.hadoop.hdfs.server.datanode.BlockReceiver"] At
> BlockReceiver.java:[lines 66-905]
> [INFO] b must be nonnull but is marked as nullable
> ["org.apache.hadoop.hdfs.server.datanode.DatanodeJspHelper$2"] At
> DatanodeJspHelper.java:[lines 546-549]
> [INFO] Found reliance on default encoding in
> org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.BlockPoolSlice.addToReplicasMap(ReplicaMap,
> File, boolean): new java.util.Scanner(File)
> ["org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.BlockPoolSlice"] At
> BlockPoolSlice.java:[lines 58-427]
> [INFO] Found reliance on default encoding in
> org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.BlockPoolSlice.loadDfsUsed():
> new java.util.Scanner(File)
> ["org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.BlockPoolSlice"] At
> BlockPoolSlice.java:[lines 58-427]
> [INFO] Found reliance on default encoding in
> org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.BlockPoolSlice.saveDfsUsed():
> new java.io.FileWriter(File)
> ["org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.BlockPoolSlice"] At
> BlockPoolSlice.java:[lines 58-427]
> [INFO] Redundant nullcheck of f, which is known to be non-null in
> org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl.invalidate(String,
> Block[])
> ["org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl"] At
> FsDatasetImpl.java:[lines 60-1910]
> [INFO] Found reliance on default encoding in
> org.apache.hadoop.hdfs.server.namenode.FSImageUtil.<static initializer for
> FSImageUtil>(): String.getBytes()
> ["org.apache.hadoop.hdfs.server.namenode.FSImageUtil"] At
> FSImageUtil.java:[lines 34-89]
> [INFO] Found reliance on default encoding in
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getListingInt(String,
> byte[], boolean): new String(byte[])
> ["org.apache.hadoop.hdfs.server.namenode.FSNamesystem"] At
> FSNamesystem.java:[lines 301-7701]
> [INFO] Found reliance on default encoding in
> org.apache.hadoop.hdfs.server.namenode.INode.dumpTreeRecursively(PrintStream):
> new java.io.PrintWriter(OutputStream, boolean)
> ["org.apache.hadoop.hdfs.server.namenode.INode"] At INode.java:[lines 51-744]
> [INFO] Redundant nullcheck of fos, which is known to be non-null in
> org.apache.hadoop.hdfs.server.namenode.NamenodeFsck.copyBlocksToLostFound(String,
> HdfsFileStatus, LocatedBlocks)
> ["org.apache.hadoop.hdfs.server.namenode.NamenodeFsck"] At
> NamenodeFsck.java:[lines 94-710]
> [INFO] Found reliance on default encoding in
> org.apache.hadoop.hdfs.tools.offlineImageViewer.OfflineImageViewerPB.run(String[]):
> new java.io.PrintWriter(File)
> ["org.apache.hadoop.hdfs.tools.offlineImageViewer.OfflineImageViewerPB"] At
> OfflineImageViewerPB.java:[lines 45-181]
> [INFO] Found reliance on default encoding in
> org.apache.hadoop.hdfs.tools.offlineImageViewer.OfflineImageViewerPB.run(String[]):
> new java.io.PrintWriter(OutputStream)
> ["org.apache.hadoop.hdfs.tools.offlineImageViewer.OfflineImageViewerPB"] At
> OfflineImageViewerPB.java:[lines 45-181]
> {noformat}
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)