This is an automated email from the ASF dual-hosted git repository. stoty pushed a commit to branch 4.x in repository https://gitbox.apache.org/repos/asf/phoenix.git
The following commit(s) were added to refs/heads/4.x by this push: new 4b4a366 PHOENIX-6327 Consolidate Junit test categories 4b4a366 is described below commit 4b4a366709a3ffda337c08d88ca51f9c3a36ea25 Author: Istvan Toth <st...@apache.org> AuthorDate: Tue Jan 19 08:28:47 2021 +0100 PHOENIX-6327 Consolidate Junit test categories --- .../phoenix/end2end/BaseHBaseManagedTimeIT.java | 69 ---------------------- .../phoenix/end2end/HBaseManagedTimeTest.java | 43 -------------- .../phoenix/end2end/NeedsOwnMiniClusterTest.java | 8 +-- .../phoenix/end2end/ParallelStatsDisabledTest.java | 4 +- .../phoenix/end2end/ParallelStatsEnabledTest.java | 4 +- .../phoenix/end2end/SplitSystemCatalogIT.java | 2 +- .../phoenix/end2end/SplitSystemCatalogTests.java | 28 --------- .../java/org/apache/phoenix/query/BaseTest.java | 17 ++++-- .../org/apache/phoenix/pherf/ResultBaseTestIT.java | 26 +++++++- .../org/apache/phoenix/pherf/SchemaReaderIT.java | 32 +++++++++- pom.xml | 40 +------------ 11 files changed, 74 insertions(+), 199 deletions(-) diff --git a/phoenix-core/src/it/java/org/apache/phoenix/end2end/BaseHBaseManagedTimeIT.java b/phoenix-core/src/it/java/org/apache/phoenix/end2end/BaseHBaseManagedTimeIT.java deleted file mode 100644 index 0f455fd..0000000 --- a/phoenix-core/src/it/java/org/apache/phoenix/end2end/BaseHBaseManagedTimeIT.java +++ /dev/null @@ -1,69 +0,0 @@ -/* - * Licensed to the Apache Software Foundation (ASF) under one - * or more contributor license agreements. See the NOTICE file - * distributed with this work for additional information - * regarding copyright ownership. The ASF licenses this file - * to you under the Apache License, Version 2.0 (the - * "License"); you may not use this file except in compliance - * with the License. You may obtain a copy of the License at - * - * http://www.apache.org/licenses/LICENSE-2.0 - * - * Unless required by applicable law or agreed to in writing, software - * distributed under the License is distributed on an "AS IS" BASIS, - * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. - * See the License for the specific language governing permissions and - * limitations under the License. - */ -package org.apache.phoenix.end2end; - -import javax.annotation.concurrent.NotThreadSafe; - -import org.apache.hadoop.conf.Configuration; -import org.apache.hadoop.hbase.HConstants; -import org.apache.phoenix.query.BaseTest; -import org.apache.phoenix.util.ReadOnlyProps; -import org.junit.After; -import org.junit.AfterClass; -import org.junit.BeforeClass; -import org.junit.experimental.categories.Category; - -/** - * Base class for tests that let HBase set timestamps. - * We need to separate these from tests that rely on clients - * to set timestamps, because we create/destroy the Phoenix tables - * between tests and only allow a table time stamp to increase. - * Without this separation table deletion/creation would fail. - * - * Try and use (@link BaseHBaseManagedTimeTableReuseIT) and only - * extend this class if really necessary. It is very slow as - * we have to delete tables which takes up to 10 seconds per test case. - * All tests extending this class use the mini cluster that is - * different from the mini cluster used by test classes extending - * {@link BaseClientManagedTimeIT}. - * - * @since 0.1 - */ -@NotThreadSafe -@Category(HBaseManagedTimeTest.class) -public abstract class BaseHBaseManagedTimeIT extends BaseTest { - protected static Configuration getTestClusterConfig() { - // don't want callers to modify config. - return new Configuration(config); - } - - @BeforeClass - public static synchronized void doSetup() throws Exception { - setUpTestDriver(ReadOnlyProps.EMPTY_PROPS); - } - - @AfterClass - public static synchronized void doTeardown() throws Exception { - dropNonSystemTables(); - } - - @After - public void cleanUpAfterTest() throws Exception { - deletePriorMetaData(HConstants.LATEST_TIMESTAMP, getUrl()); - } -} diff --git a/phoenix-core/src/it/java/org/apache/phoenix/end2end/HBaseManagedTimeTest.java b/phoenix-core/src/it/java/org/apache/phoenix/end2end/HBaseManagedTimeTest.java deleted file mode 100644 index c9eee86..0000000 --- a/phoenix-core/src/it/java/org/apache/phoenix/end2end/HBaseManagedTimeTest.java +++ /dev/null @@ -1,43 +0,0 @@ -/* - * Licensed to the Apache Software Foundation (ASF) under one - * or more contributor license agreements. See the NOTICE file - * distributed with this work for additional information - * regarding copyright ownership. The ASF licenses this file - * to you under the Apache License, Version 2.0 (the - * "License"); you maynot use this file except in compliance - * with the License. You may obtain a copy of the License at - * - * http://www.apache.org/licenses/LICENSE-2.0 - * - * Unless required by applicablelaw or agreed to in writing, software - * distributed under the License is distributed on an "AS IS" BASIS, - * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. - * See the License for the specific language governing permissions and - * limitations under the License. - */ -package org.apache.phoenix.end2end; - -import java.lang.annotation.ElementType; -import java.lang.annotation.Inherited; -import java.lang.annotation.Retention; -import java.lang.annotation.RetentionPolicy; -import java.lang.annotation.Target; - -/** - * - * Annotation to denote that the test lets HBase manage timestamp - * itself. - * - * Tests using a mini cluster need to be classified either - * as {@link HBaseManagedTimeTest} or {@link ClientManagedTimeTest} - * or {@link NeedsOwnMiniClusterTest} or {@link ParallelStatsDisabledTest} - * otherwise they won't be run when one runs mvn verify or mvn install. - * - * @since 4.1 - */ -@Retention(RetentionPolicy.RUNTIME) -@Target(ElementType.TYPE) -@Inherited -public @interface HBaseManagedTimeTest { - -} diff --git a/phoenix-core/src/it/java/org/apache/phoenix/end2end/NeedsOwnMiniClusterTest.java b/phoenix-core/src/it/java/org/apache/phoenix/end2end/NeedsOwnMiniClusterTest.java index 6ea4c4e..aa37c88 100644 --- a/phoenix-core/src/it/java/org/apache/phoenix/end2end/NeedsOwnMiniClusterTest.java +++ b/phoenix-core/src/it/java/org/apache/phoenix/end2end/NeedsOwnMiniClusterTest.java @@ -26,12 +26,12 @@ import java.lang.annotation.Target; * * Annotation to denote that the test needs to run its own * mini cluster that is separate from the clusters used by - * tests annotated as {@link HBaseManagedTimeTest} or - * {@link ClientManagedTimeTest}. + * tests annotated as {@link ParallelStatsDisabledTest} or + * {@link ParallelStatsEnabledTest}. * * As much as possible, tests should be able to run in one of the - * mini clusters used by {@link HBaseManagedTimeTest} or - * {@link ClientManagedTimeTest}. In the *rare* case it can't + * mini clusters used by {@link ParallelStatsDisabledTest} or + * {@link ParallelStatsEnabledTest}. In the *rare* case it can't * you would need to annotate the test as {@link NeedsOwnMiniClusterTest} * otherwise the test won't be executed when you run mvn verify or mvn install. * diff --git a/phoenix-core/src/it/java/org/apache/phoenix/end2end/ParallelStatsDisabledTest.java b/phoenix-core/src/it/java/org/apache/phoenix/end2end/ParallelStatsDisabledTest.java index bb8c19f..d807b8d 100644 --- a/phoenix-core/src/it/java/org/apache/phoenix/end2end/ParallelStatsDisabledTest.java +++ b/phoenix-core/src/it/java/org/apache/phoenix/end2end/ParallelStatsDisabledTest.java @@ -32,8 +32,8 @@ import org.apache.phoenix.query.BaseTest; * these tests use random table names generated by {@link BaseTest#generateRandomString}. * * Tests using a mini cluster need to be classified either - * as {@link ParallelStatsDisabledTest} or {@link ClientManagedTimeTest} - * or {@link NeedsOwnMiniClusterTest} or {@link HBaseManagedTimeTest} otherwise they won't be run + * as {@link ParallelStatsDisabledTest} or {@link ParallelStatsEnabledTest} + * or {@link NeedsOwnMiniClusterTest} otherwise they won't be run * when one runs mvn verify or mvn install. * * @since 4.1 diff --git a/phoenix-core/src/it/java/org/apache/phoenix/end2end/ParallelStatsEnabledTest.java b/phoenix-core/src/it/java/org/apache/phoenix/end2end/ParallelStatsEnabledTest.java index 6607b05..9fcffbc 100644 --- a/phoenix-core/src/it/java/org/apache/phoenix/end2end/ParallelStatsEnabledTest.java +++ b/phoenix-core/src/it/java/org/apache/phoenix/end2end/ParallelStatsEnabledTest.java @@ -32,8 +32,8 @@ import org.apache.phoenix.query.BaseTest; * these tests use random table names generated by {@link BaseTest#generateRandomString}. * * Tests using a mini cluster need to be classified either - * as {@link ParallelStatsEnabledTest} or {@link ClientManagedTimeTest} - * or {@link NeedsOwnMiniClusterTest} or {@link HBaseManagedTimeTest} otherwise they won't be run + * as {@link ParallelStatsDisabledTest} or {@link ParallelStatsEnabledTest} + * or {@link NeedsOwnMiniClusterTest} otherwise they won't be run * when one runs mvn verify or mvn install. * * @since 4.1 diff --git a/phoenix-core/src/it/java/org/apache/phoenix/end2end/SplitSystemCatalogIT.java b/phoenix-core/src/it/java/org/apache/phoenix/end2end/SplitSystemCatalogIT.java index b2075a7..ccf2eb2 100644 --- a/phoenix-core/src/it/java/org/apache/phoenix/end2end/SplitSystemCatalogIT.java +++ b/phoenix-core/src/it/java/org/apache/phoenix/end2end/SplitSystemCatalogIT.java @@ -37,7 +37,7 @@ import com.google.common.collect.Maps; * Base class for tests that run with split SYSTEM.CATALOG. * */ -@Category(SplitSystemCatalogTests.class) +@Category(NeedsOwnMiniClusterTest.class) public class SplitSystemCatalogIT extends BaseTest { protected static String SCHEMA1 = "SCHEMA1"; diff --git a/phoenix-core/src/it/java/org/apache/phoenix/end2end/SplitSystemCatalogTests.java b/phoenix-core/src/it/java/org/apache/phoenix/end2end/SplitSystemCatalogTests.java deleted file mode 100644 index 67d3fd3..0000000 --- a/phoenix-core/src/it/java/org/apache/phoenix/end2end/SplitSystemCatalogTests.java +++ /dev/null @@ -1,28 +0,0 @@ -/* - * Licensed to the Apache Software Foundation (ASF) under one - * or more contributor license agreements. See the NOTICE file - * distributed with this work for additional information - * regarding copyright ownership. The ASF licenses this file - * to you under the Apache License, Version 2.0 (the - * "License"); you may not use this file except in compliance - * with the License. You may obtain a copy of the License at - * - * http://www.apache.org/licenses/LICENSE-2.0 - * - * Unless required by applicable law or agreed to in writing, software - * distributed under the License is distributed on an "AS IS" BASIS, - * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. - * See the License for the specific language governing permissions and - * limitations under the License. - */ -package org.apache.phoenix.end2end; - -import java.lang.annotation.ElementType; -import java.lang.annotation.Retention; -import java.lang.annotation.RetentionPolicy; -import java.lang.annotation.Target; - -@Retention(RetentionPolicy.RUNTIME) -@Target(ElementType.TYPE) -public @interface SplitSystemCatalogTests { -} diff --git a/phoenix-core/src/test/java/org/apache/phoenix/query/BaseTest.java b/phoenix-core/src/test/java/org/apache/phoenix/query/BaseTest.java index f326f9b..1537ce6 100644 --- a/phoenix-core/src/test/java/org/apache/phoenix/query/BaseTest.java +++ b/phoenix-core/src/test/java/org/apache/phoenix/query/BaseTest.java @@ -134,7 +134,6 @@ import org.apache.hadoop.hbase.regionserver.RSRpcServices; import org.apache.hadoop.hbase.util.Bytes; import org.apache.hadoop.hbase.util.JVMClusterUtil.RegionServerThread; import org.apache.phoenix.SystemExitRule; -import org.apache.phoenix.end2end.BaseHBaseManagedTimeIT; import org.apache.phoenix.exception.SQLExceptionCode; import org.apache.phoenix.exception.SQLExceptionInfo; import org.apache.phoenix.jdbc.PhoenixConnection; @@ -170,13 +169,19 @@ import com.google.common.util.concurrent.ThreadFactoryBuilder; * Base class that contains all the methods needed by * client-time and hbase-time managed tests. * + * * Tests using a mini cluster need to be classified either + * as {@link ParallelStatsDisabledTest} or {@link ParallelStatsEnabledTest} + * or {@link NeedsOwnMiniClusterTest} otherwise they won't be run + * when one runs mvn verify or mvn install. + * * For tests needing connectivity to a cluster, please use - * {@link BaseHBaseManagedTimeIT}. + * {@link ParallelStatsDisabledIt} or {@link ParallelStatsEnabledIt}. * - * In the rare case when a test can't share the same mini cluster as the - * ones used by {@link BaseHBaseManagedTimeIT}, - * one could extend this class and spin up your own mini cluster. Please - * make sure to shutdown the mini cluster in a method annotated by @AfterClass. + * In the case when a test can't share the same mini cluster as the + * ones used by {@link ParallelStatsDisabledIt} or {@link ParallelStatsEnabledIt}, + * one could extend this class and spin up your own mini cluster. Please + * make sure to annotate such clesses with {@link NeedsOwnMiniClusterTest} and + * shutdown the mini cluster in a method annotated by @AfterClass. * */ diff --git a/phoenix-pherf/src/it/java/org/apache/phoenix/pherf/ResultBaseTestIT.java b/phoenix-pherf/src/it/java/org/apache/phoenix/pherf/ResultBaseTestIT.java index a1ca6fa..fe1f2ea 100644 --- a/phoenix-pherf/src/it/java/org/apache/phoenix/pherf/ResultBaseTestIT.java +++ b/phoenix-pherf/src/it/java/org/apache/phoenix/pherf/ResultBaseTestIT.java @@ -22,15 +22,22 @@ import java.nio.file.Path; import java.util.List; import java.util.Properties; -import org.apache.phoenix.end2end.BaseHBaseManagedTimeIT; +import org.apache.hadoop.conf.Configuration; +import org.apache.hadoop.hbase.HConstants; +import org.apache.phoenix.end2end.NeedsOwnMiniClusterTest; import org.apache.phoenix.pherf.configuration.XMLConfigParser; import org.apache.phoenix.pherf.result.ResultUtil; import org.apache.phoenix.pherf.schema.SchemaReader; import org.apache.phoenix.pherf.util.PhoenixUtil; +import org.apache.phoenix.query.BaseTest; +import org.apache.phoenix.util.ReadOnlyProps; +import org.junit.After; import org.junit.AfterClass; import org.junit.BeforeClass; +import org.junit.experimental.categories.Category; -public class ResultBaseTestIT extends BaseHBaseManagedTimeIT { +@Category(NeedsOwnMiniClusterTest.class) +public class ResultBaseTestIT extends BaseTest { protected static final String matcherScenario = ".*scenario/.*test.*xml"; protected static final String matcherSchema = ".*datamodel/.*test.*sql"; @@ -41,6 +48,11 @@ public class ResultBaseTestIT extends BaseHBaseManagedTimeIT { protected static List<Path> resources; protected static ResultUtil resultUtil = new ResultUtil(); + protected static Configuration getTestClusterConfig() { + // don't want callers to modify config. + return new Configuration(config); + } + @BeforeClass public static synchronized void setUp() throws Exception { PherfConstants constants = PherfConstants.create(); @@ -51,9 +63,17 @@ public class ResultBaseTestIT extends BaseHBaseManagedTimeIT { PhoenixUtil.setZookeeper("localhost"); reader = new SchemaReader(util, matcherSchema); parser = new XMLConfigParser(matcherScenario); + + setUpTestDriver(ReadOnlyProps.EMPTY_PROPS); } @AfterClass public static synchronized void tearDown() throws Exception { - resultUtil.deleteDir(properties.getProperty("pherf.default.results.dir")); + dropNonSystemTables(); + resultUtil.deleteDir(properties.getProperty("pherf.default.results.dir")); + } + + @After + public void cleanUpAfterTest() throws Exception { + deletePriorMetaData(HConstants.LATEST_TIMESTAMP, getUrl()); } } diff --git a/phoenix-pherf/src/it/java/org/apache/phoenix/pherf/SchemaReaderIT.java b/phoenix-pherf/src/it/java/org/apache/phoenix/pherf/SchemaReaderIT.java index a897a46..901c92f 100644 --- a/phoenix-pherf/src/it/java/org/apache/phoenix/pherf/SchemaReaderIT.java +++ b/phoenix-pherf/src/it/java/org/apache/phoenix/pherf/SchemaReaderIT.java @@ -30,19 +30,45 @@ import java.sql.SQLException; import java.util.ArrayList; import java.util.List; -import org.apache.phoenix.end2end.BaseHBaseManagedTimeIT; +import org.apache.hadoop.conf.Configuration; +import org.apache.hadoop.hbase.HConstants; +import org.apache.phoenix.end2end.NeedsOwnMiniClusterTest; import org.apache.phoenix.pherf.configuration.Column; import org.apache.phoenix.pherf.configuration.DataModel; import org.apache.phoenix.pherf.configuration.Scenario; import org.apache.phoenix.pherf.configuration.XMLConfigParser; import org.apache.phoenix.pherf.schema.SchemaReader; import org.apache.phoenix.pherf.util.PhoenixUtil; +import org.apache.phoenix.query.BaseTest; +import org.apache.phoenix.util.ReadOnlyProps; +import org.junit.After; +import org.junit.AfterClass; +import org.junit.BeforeClass; import org.junit.Test; +import org.junit.experimental.categories.Category; - -public class SchemaReaderIT extends BaseHBaseManagedTimeIT { +@Category(NeedsOwnMiniClusterTest.class) +public class SchemaReaderIT extends BaseTest { protected static PhoenixUtil util = PhoenixUtil.create(true); + protected static Configuration getTestClusterConfig() { + // don't want callers to modify config. + return new Configuration(config); + } + + @BeforeClass public static synchronized void setUp() throws Exception { + setUpTestDriver(ReadOnlyProps.EMPTY_PROPS); + } + + @AfterClass public static synchronized void tearDown() throws Exception { + dropNonSystemTables(); + } + + @After + public void cleanUpAfterTest() throws Exception { + deletePriorMetaData(HConstants.LATEST_TIMESTAMP, getUrl()); + } + @Test public void testSchemaReader() { // Test for the unit test version of the schema files. diff --git a/pom.xml b/pom.xml index 202445f..8916df1 100644 --- a/pom.xml +++ b/pom.xml @@ -315,9 +315,7 @@ <id>ParallelStatsEnabledTest</id> <configuration> <reuseForks>true</reuseForks> - <!--parallel>methods</parallel> - <threadCount>20</threadCount--> - <argLine>@{jacocoArgLine} -Xmx2000m -XX:MaxPermSize=256m -Djava.security.egd=file:/dev/./urandom "-Djava.library.path=${hadoop.library.path}${path.separator}${java.library.path}" -XX:NewRatio=4 -XX:SurvivorRatio=8 -XX:+UseCompressedOops -XX:+UseConcMarkSweepGC -XX:+UseParNewGC -XX:+DisableExplicitGC -XX:+UseCMSInitiatingOccupancyOnly -XX:+CMSClassUnloadingEnabled -XX:+CMSScavengeBeforeRemark -XX:CMSInitiatingOccupancyFraction=68 -XX:+HeapDumpOnOutOfMemoryError -XX:HeapDum [...] + <argLine>@{jacocoArgLine} -Xmx2000m -Djava.security.egd=file:/dev/./urandom "-Djava.library.path=${hadoop.library.path}${path.separator}${java.library.path}" -XX:+HeapDumpOnOutOfMemoryError -XX:HeapDumpPath=./target/</argLine> <groups>org.apache.phoenix.end2end.ParallelStatsEnabledTest</groups> </configuration> <goals> @@ -329,17 +327,7 @@ <id>ParallelStatsDisabledTest</id> <configuration> <reuseForks>true</reuseForks> - <!--parallel>methods</parallel> - <threadCount>20</threadCount--> - <!-- We're intermittantly hitting this assertion when running in parallel: - Caused by: java.lang.AssertionError: we should never remove a different context - at org.apache.hadoop.hbase.regionserver.HRegion$RowLockContext.cleanUp(HRegion.java:5206) - at org.apache.hadoop.hbase.regionserver.HRegion$RowLockImpl.release(HRegion.java:5246) - at org.apache.phoenix.coprocessor.MetaDataEndpointImpl.doGetTable(MetaDataEndpointImpl.java:2898) - at org.apache.phoenix.coprocessor.MetaDataEndpointImpl.doGetTable(MetaDataEndpointImpl.java:2835) - at org.apache.phoenix.coprocessor.MetaDataEndpointImpl.getTable(MetaDataEndpointImpl.java:490) --> - <!--enableAssertions>false</enableAssertions--> - <argLine>@{jacocoArgLine} -Xmx3000m -XX:MaxPermSize=256m -Djava.security.egd=file:/dev/./urandom "-Djava.library.path=${hadoop.library.path}${path.separator}${java.library.path}" -XX:NewRatio=4 -XX:SurvivorRatio=8 -XX:+UseCompressedOops -XX:+UseConcMarkSweepGC -XX:+UseParNewGC -XX:+DisableExplicitGC -XX:+UseCMSInitiatingOccupancyOnly -XX:+CMSClassUnloadingEnabled -XX:+CMSScavengeBeforeRemark -XX:CMSInitiatingOccupancyFraction=68 -XX:+HeapDumpOnOutOfMemoryError -XX:HeapDum [...] + <argLine>@{jacocoArgLine} -Xmx2000m -XX:MaxPermSize=256m -Djava.security.egd=file:/dev/./urandom "-Djava.library.path=${hadoop.library.path}${path.separator}${java.library.path}" -XX:+HeapDumpOnOutOfMemoryError -XX:HeapDumpPath=./target/</argLine> <groups>org.apache.phoenix.end2end.ParallelStatsDisabledTest</groups> </configuration> <goals> @@ -348,18 +336,6 @@ </goals> </execution> <execution> - <id>HBaseManagedTimeTests</id> - <configuration> - <reuseForks>true</reuseForks> - <argLine>@{jacocoArgLine} -enableassertions -Xmx2000m -XX:MaxPermSize=128m -Djava.security.egd=file:/dev/./urandom "-Djava.library.path=${hadoop.library.path}${path.separator}${java.library.path}" -XX:+HeapDumpOnOutOfMemoryError -XX:HeapDumpPath=./target/</argLine> - <groups>org.apache.phoenix.end2end.HBaseManagedTimeTest</groups> - </configuration> - <goals> - <goal>integration-test</goal> - <goal>verify</goal> - </goals> - </execution> - <execution> <id>NeedTheirOwnClusterTests</id> <configuration> <reuseForks>false</reuseForks> @@ -371,18 +347,6 @@ <goal>verify</goal> </goals> </execution> -<execution> - <id>SplitSystemCatalogTests</id> - <configuration> - <reuseForks>false</reuseForks> - <argLine>@{jacocoArgLine} -enableassertions -Xmx2000m -XX:MaxPermSize=256m -Djava.security.egd=file:/dev/./urandom "-Djava.library.path=${hadoop.library.path}${path.separator}${java.library.path}" -XX:+HeapDumpOnOutOfMemoryError -XX:HeapDumpPath=./target/</argLine> - <groups>org.apache.phoenix.end2end.SplitSystemCatalogTests</groups> - </configuration> - <goals> - <goal>integration-test</goal> - <goal>verify</goal> - </goals> - </execution> </executions> </plugin> <plugin>