[ https://issues.apache.org/jira/browse/PHOENIX-5842?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17100531#comment-17100531 ]
Hadoop QA commented on PHOENIX-5842: ------------------------------------ {color:red}-1 overall{color}. Here are the results of testing the latest attachment http://issues.apache.org/jira/secure/attachment/13002126/PHOENIX-5842.4.x.v2.patch against 4.x branch at commit 364b62c34547a3a4e4b5496e1e50eee1d6daa514. ATTACHMENT ID: 13002126 {color:green}+1 @author{color}. The patch does not contain any @author tags. {color:green}+1 tests included{color}. The patch appears to include 0 new or modified tests. {color:green}+1 javac{color}. The applied patch does not increase the total number of javac compiler warnings. {color:green}+1 release audit{color}. The applied patch does not increase the total number of release audit warnings. {color:red}-1 lineLengths{color}. The patch introduces the following lines longer than 100: +JIRA_COMMENT_FOOTER="Code Coverage results: $BUILD_URL/artifact/phoenix-core/target/site/jacoco/index.html + <argLine>@{jacocoArgLine} -Xmx2000m -XX:MaxPermSize=256m -Djava.security.egd=file:/dev/./urandom "-Djava.library.path=${hadoop.library.path}${path.separator}${java.library.path}" -XX:NewRatio=4 -XX:SurvivorRatio=8 -XX:+UseCompressedOops -XX:+UseConcMarkSweepGC -XX:+UseParNewGC -XX:+DisableExplicitGC -XX:+UseCMSInitiatingOccupancyOnly -XX:+CMSClassUnloadingEnabled -XX:+CMSScavengeBeforeRemark -XX:CMSInitiatingOccupancyFraction=68 -XX:+HeapDumpOnOutOfMemoryError -XX:HeapDumpPath=./target/</argLine> + <argLine>@{jacocoArgLine} -Xmx3000m -XX:MaxPermSize=256m -Djava.security.egd=file:/dev/./urandom "-Djava.library.path=${hadoop.library.path}${path.separator}${java.library.path}" -XX:NewRatio=4 -XX:SurvivorRatio=8 -XX:+UseCompressedOops -XX:+UseConcMarkSweepGC -XX:+UseParNewGC -XX:+DisableExplicitGC -XX:+UseCMSInitiatingOccupancyOnly -XX:+CMSClassUnloadingEnabled -XX:+CMSScavengeBeforeRemark -XX:CMSInitiatingOccupancyFraction=68 -XX:+HeapDumpOnOutOfMemoryError -XX:HeapDumpPath=./target/</argLine> + <argLine>@{jacocoArgLine} -enableassertions -Xmx2000m -XX:MaxPermSize=128m -Djava.security.egd=file:/dev/./urandom "-Djava.library.path=${hadoop.library.path}${path.separator}${java.library.path}" -XX:+HeapDumpOnOutOfMemoryError -XX:HeapDumpPath=./target/</argLine> + <argLine>@{jacocoArgLine} -enableassertions -Xmx2000m -XX:MaxPermSize=256m -Djava.security.egd=file:/dev/./urandom "-Djava.library.path=${hadoop.library.path}${path.separator}${java.library.path}" -XX:+HeapDumpOnOutOfMemoryError -XX:HeapDumpPath=./target/</argLine> + <argLine>@{jacocoArgLine} -enableassertions -Xmx2000m -XX:MaxPermSize=256m -Djava.security.egd=file:/dev/./urandom "-Djava.library.path=${hadoop.library.path}${path.separator}${java.library.path}" -XX:+HeapDumpOnOutOfMemoryError -XX:HeapDumpPath=./target/</argLine> {color:red}-1 core tests{color}. The patch failed these unit tests: ./phoenix-core/target/failsafe-reports/TEST-org.apache.phoenix.end2end.join.HashJoinLocalIndexIT ./phoenix-core/target/failsafe-reports/TEST-org.apache.phoenix.end2end.DropTableWithViewsIT ./phoenix-core/target/failsafe-reports/TEST-org.apache.phoenix.end2end.salted.SaltedTableUpsertSelectIT ./phoenix-core/target/failsafe-reports/TEST-org.apache.phoenix.trace.PhoenixTableMetricsWriterIT ./phoenix-core/target/failsafe-reports/TEST-org.apache.phoenix.iterate.ChunkedResultIteratorIT ./phoenix-core/target/failsafe-reports/TEST-org.apache.phoenix.iterate.PhoenixQueryTimeoutIT ./phoenix-core/target/failsafe-reports/TEST-org.apache.phoenix.end2end.join.SubqueryIT ./phoenix-core/target/failsafe-reports/TEST-org.apache.phoenix.end2end.index.LocalImmutableNonTxIndexIT ./phoenix-core/target/failsafe-reports/TEST-org.apache.phoenix.end2end.index.LocalMutableTxIndexIT ./phoenix-core/target/failsafe-reports/TEST-org.apache.phoenix.end2end.index.txn.MutableRollbackIT ./phoenix-core/target/failsafe-reports/TEST-org.apache.phoenix.rpc.UpdateCacheConnectionLevelPropIT ./phoenix-core/target/failsafe-reports/TEST-org.apache.phoenix.end2end.salted.SaltedTableVarLengthRowKeyIT ./phoenix-core/target/failsafe-reports/TEST-org.apache.phoenix.util.IndexScrutinyIT ./phoenix-core/target/failsafe-reports/TEST-org.apache.phoenix.end2end.join.HashJoinPersistentCacheIT {color:red}-1 core zombie tests{color}. There are 1 zombie test(s): at org.apache.phoenix.end2end.DropTableWithViewsIT.testDropTableWithChildViews(DropTableWithViewsIT.java:140) Test results: https://builds.apache.org/job/PreCommit-PHOENIX-Build/3815//testReport/ Console output: https://builds.apache.org/job/PreCommit-PHOENIX-Build/3815//console This message is automatically generated. > Code Coverage tool for Phoenix > ------------------------------ > > Key: PHOENIX-5842 > URL: https://issues.apache.org/jira/browse/PHOENIX-5842 > Project: Phoenix > Issue Type: Improvement > Reporter: Sandeep Guggilam > Assignee: Sandeep Guggilam > Priority: Major > Labels: quality-improvement > Attachments: PHOENIX-5842.4.x.v1.patch, PHOENIX-5842.4.x.v2.patch > > > Currently we don't have any code coverage tool for Phoenix. This is required > for us to measure the test coverage and helps us in improving the test suite > further till we reach may be 80% coverage > The test coverage results can also be integrated into the Hadoop QA run of > the precommit build such that the reviewers could take a look at the report > and see if the added code doesn't have enough coverage -- This message was sent by Atlassian Jira (v8.3.4#803005)