[ https://issues.apache.org/jira/browse/HBASE-5878?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14660755#comment-14660755 ]
Hadoop QA commented on HBASE-5878: ---------------------------------- {color:red}-1 overall{color}. Here are the results of testing the latest attachment http://issues.apache.org/jira/secure/attachment/12749093/HBASE-5878-v5-0.98.patch against 0.98 branch at commit 7a9e10dc11877420c53245c403897d746bebc077. ATTACHMENT ID: 12749093 {color:green}+1 @author{color}. The patch does not contain any @author tags. {color:red}-1 tests included{color}. The patch doesn't appear to include any new or modified tests. Please justify why no new tests are needed for this patch. Also please list what manual steps were performed to verify this patch. {color:green}+1 hadoop versions{color}. The patch compiles with all supported hadoop versions (2.4.0 2.4.1 2.5.0 2.5.1 2.5.2 2.6.0 2.7.0) {color:green}+1 javac{color}. The applied patch does not increase the total number of javac compiler warnings. {color:green}+1 protoc{color}. The applied patch does not increase the total number of protoc compiler warnings. {color:red}-1 javadoc{color}. The javadoc tool appears to have generated 22 warning messages. {color:red}-1 checkstyle{color}. The applied patch generated 3879 checkstyle errors (more than the master's current 3877 errors). {color:green}+1 findbugs{color}. The patch does not introduce any new Findbugs (version 2.0.3) warnings. {color:green}+1 release audit{color}. The applied patch does not increase the total number of release audit warnings. {color:green}+1 lineLengths{color}. The patch does not introduce lines longer than 100 {color:red}-1 site{color}. The patch appears to cause mvn post-site goal to fail. {color:red}-1 core tests{color}. The patch failed these unit tests: {color:red}-1 core zombie tests{color}. There are 1 zombie test(s): at org.apache.hadoop.hbase.rest.TestTableResource.testTableListXML(TestTableResource.java:207) Test results: https://builds.apache.org/job/PreCommit-HBASE-Build/14992//testReport/ Release Findbugs (version 2.0.3) warnings: https://builds.apache.org/job/PreCommit-HBASE-Build/14992//artifact/patchprocess/newFindbugsWarnings.html Checkstyle Errors: https://builds.apache.org/job/PreCommit-HBASE-Build/14992//artifact/patchprocess/checkstyle-aggregate.html Javadoc warnings: https://builds.apache.org/job/PreCommit-HBASE-Build/14992//artifact/patchprocess/patchJavadocWarnings.txt Console output: https://builds.apache.org/job/PreCommit-HBASE-Build/14992//console This message is automatically generated. > Use getVisibleLength public api from HdfsDataInputStream from Hadoop-2. > ----------------------------------------------------------------------- > > Key: HBASE-5878 > URL: https://issues.apache.org/jira/browse/HBASE-5878 > Project: HBase > Issue Type: Bug > Components: wal > Reporter: Uma Maheswara Rao G > Assignee: Ashish Singhi > Fix For: 2.0.0, 1.3.0 > > Attachments: HBASE-5878-v2.patch, HBASE-5878-v3.patch, > HBASE-5878-v4.patch, HBASE-5878-v5-0.98.patch, HBASE-5878-v5.patch, > HBASE-5878-v5.patch, HBASE-5878-v6-0.98.patch, HBASE-5878-v6.patch, > HBASE-5878.patch > > > SequencFileLogReader: > Currently Hbase using getFileLength api from DFSInputStream class by > reflection. DFSInputStream is not exposed as public. So, this may change in > future. Now HDFS exposed HdfsDataInputStream as public API. > We can make use of it, when we are not able to find the getFileLength api > from DFSInputStream as a else condition. So, that we will not have any sudden > surprise like we are facing today. > Also, it is just logging one warn message and proceeding if it throws any > exception while getting the length. I think we can re-throw the exception > because there is no point in continuing with dataloss. > {code} > long adjust = 0; > try { > Field fIn = FilterInputStream.class.getDeclaredField("in"); > fIn.setAccessible(true); > Object realIn = fIn.get(this.in); > // In hadoop 0.22, DFSInputStream is a standalone class. Before > this, > // it was an inner class of DFSClient. > if (realIn.getClass().getName().endsWith("DFSInputStream")) { > Method getFileLength = realIn.getClass(). > getDeclaredMethod("getFileLength", new Class<?> []{}); > getFileLength.setAccessible(true); > long realLength = ((Long)getFileLength. > invoke(realIn, new Object []{})).longValue(); > assert(realLength >= this.length); > adjust = realLength - this.length; > } else { > LOG.info("Input stream class: " + realIn.getClass().getName() + > ", not adjusting length"); > } > } catch(Exception e) { > SequenceFileLogReader.LOG.warn( > "Error while trying to get accurate file length. " + > "Truncation / data loss may occur if RegionServers die.", e); > } > return adjust + super.getPos(); > {code} -- This message was sent by Atlassian JIRA (v6.3.4#6332)