http://git-wip-us.apache.org/repos/asf/hbase/blob/9246af8d/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestOpenTableInCoprocessor.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestOpenTableInCoprocessor.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestOpenTableInCoprocessor.java index c5af6d5..4c3594b 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestOpenTableInCoprocessor.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestOpenTableInCoprocessor.java @@ -18,23 +18,12 @@ */ package org.apache.hadoop.hbase.coprocessor; -import static org.junit.Assert.assertEquals; -import static org.junit.Assert.assertTrue; - -import java.io.IOException; -import java.util.Collections; -import java.util.concurrent.ExecutorService; -import java.util.concurrent.SynchronousQueue; -import java.util.concurrent.ThreadPoolExecutor; -import java.util.concurrent.TimeUnit; - import org.apache.hadoop.hbase.HBaseTestingUtility; import org.apache.hadoop.hbase.HColumnDescriptor; import org.apache.hadoop.hbase.HTableDescriptor; import org.apache.hadoop.hbase.TableName; import org.apache.hadoop.hbase.client.Admin; import org.apache.hadoop.hbase.client.Durability; -import org.apache.hadoop.hbase.client.HTable; import org.apache.hadoop.hbase.client.Put; import org.apache.hadoop.hbase.client.Result; import org.apache.hadoop.hbase.client.ResultScanner; @@ -50,6 +39,16 @@ import org.junit.BeforeClass; import org.junit.Test; import org.junit.experimental.categories.Category; +import java.io.IOException; +import java.util.Collections; +import java.util.concurrent.ExecutorService; +import java.util.concurrent.SynchronousQueue; +import java.util.concurrent.ThreadPoolExecutor; +import java.util.concurrent.TimeUnit; + +import static org.junit.Assert.assertEquals; +import static org.junit.Assert.assertTrue; + /** * Test that a coprocessor can open a connection and write to another table, inside a hook. */ @@ -161,13 +160,13 @@ public class TestOpenTableInCoprocessor { admin.createTable(primary); admin.createTable(other); - Table table = new HTable(UTIL.getConfiguration(), TableName.valueOf("primary")); + Table table = UTIL.getConnection().getTable(TableName.valueOf("primary")); Put p = new Put(new byte[] { 'a' }); p.add(family, null, new byte[] { 'a' }); table.put(p); table.close(); - Table target = new HTable(UTIL.getConfiguration(), otherTable); + Table target = UTIL.getConnection().getTable(otherTable); assertTrue("Didn't complete update to target table!", completeCheck[0]); assertEquals("Didn't find inserted row", 1, getKeyValueCount(target)); target.close();
http://git-wip-us.apache.org/repos/asf/hbase/blob/9246af8d/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRegionObserverBypass.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRegionObserverBypass.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRegionObserverBypass.java index 3e41859..a02758dac 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRegionObserverBypass.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRegionObserverBypass.java @@ -93,7 +93,7 @@ public class TestRegionObserverBypass { */ @Test public void testSimple() throws Exception { - Table t = new HTable(util.getConfiguration(), tableName); + Table t = util.getConnection().getTable(tableName); Put p = new Put(row1); p.add(test,dummy,dummy); // before HBASE-4331, this would throw an exception @@ -112,7 +112,7 @@ public class TestRegionObserverBypass { //previous deletes will eclipse successive puts having the same timestamp EnvironmentEdgeManagerTestHelper.injectEdge(new IncrementingEnvironmentEdge()); - Table t = new HTable(util.getConfiguration(), tableName); + Table t = util.getConnection().getTable(tableName); List<Put> puts = new ArrayList<Put>(); Put p = new Put(row1); p.add(dummy,dummy,dummy); http://git-wip-us.apache.org/repos/asf/hbase/blob/9246af8d/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRegionObserverInterface.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRegionObserverInterface.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRegionObserverInterface.java index 0f161d9..6895fbe 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRegionObserverInterface.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRegionObserverInterface.java @@ -340,7 +340,7 @@ public class TestRegionObserverInterface { new Boolean[] {false, false, false, false} ); - Table table = new HTable(util.getConfiguration(), tableName); + Table table = util.getConnection().getTable(tableName); Put put = new Put(ROW); put.add(A, A, A); table.put(put); @@ -390,7 +390,7 @@ public class TestRegionObserverInterface { new Boolean[] {false, false} ); - Table table = new HTable(util.getConfiguration(), tableName); + Table table = util.getConnection().getTable(tableName); Put put = new Put(ROW); put.add(A, A, A); table.put(put); @@ -497,7 +497,7 @@ public class TestRegionObserverInterface { htd.addCoprocessor(EvenOnlyCompactor.class.getName()); admin.createTable(htd); - Table table = new HTable(util.getConfiguration(), compactTable); + Table table = util.getConnection().getTable(compactTable); for (long i=1; i<=10; i++) { byte[] iBytes = Bytes.toBytes(i); Put put = new Put(iBytes); http://git-wip-us.apache.org/repos/asf/hbase/blob/9246af8d/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRegionObserverScannerOpenHook.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRegionObserverScannerOpenHook.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRegionObserverScannerOpenHook.java index 8742066..88b001d 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRegionObserverScannerOpenHook.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRegionObserverScannerOpenHook.java @@ -259,7 +259,7 @@ public class TestRegionObserverScannerOpenHook { Admin admin = UTIL.getHBaseAdmin(); admin.createTable(desc); - Table table = new HTable(conf, desc.getTableName()); + Table table = UTIL.getConnection().getTable(desc.getTableName()); // put a row and flush it to disk Put put = new Put(ROW); http://git-wip-us.apache.org/repos/asf/hbase/blob/9246af8d/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRegionServerCoprocessorEndpoint.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRegionServerCoprocessorEndpoint.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRegionServerCoprocessorEndpoint.java index df85004..7ae6787 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRegionServerCoprocessorEndpoint.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRegionServerCoprocessorEndpoint.java @@ -73,7 +73,7 @@ public class TestRegionServerCoprocessorEndpoint { new BlockingRpcCallback<DummyRegionServerEndpointProtos.DummyResponse>(); DummyRegionServerEndpointProtos.DummyService service = ProtobufUtil.newServiceStub(DummyRegionServerEndpointProtos.DummyService.class, - new HBaseAdmin(CONF).coprocessorService(serverName)); + TEST_UTIL.getHBaseAdmin().coprocessorService(serverName)); service.dummyCall(controller, DummyRegionServerEndpointProtos.DummyRequest.getDefaultInstance(), rpcCallback); assertEquals(DUMMY_VALUE, rpcCallback.get().getValue()); http://git-wip-us.apache.org/repos/asf/hbase/blob/9246af8d/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRegionServerObserver.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRegionServerObserver.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRegionServerObserver.java index 0c30bb2..2e6eabc 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRegionServerObserver.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRegionServerObserver.java @@ -76,7 +76,7 @@ public class TestRegionServerObserver { // Start the cluster HBaseTestingUtility TEST_UTIL = new HBaseTestingUtility(conf); TEST_UTIL.startMiniCluster(NUM_MASTERS, NUM_RS); - Admin admin = new HBaseAdmin(conf); + Admin admin = TEST_UTIL.getHBaseAdmin(); try { MiniHBaseCluster cluster = TEST_UTIL.getHBaseCluster(); HRegionServer regionServer = cluster.getRegionServer(0); http://git-wip-us.apache.org/repos/asf/hbase/blob/9246af8d/hbase-server/src/test/java/org/apache/hadoop/hbase/filter/FilterTestingCluster.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/filter/FilterTestingCluster.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/filter/FilterTestingCluster.java index 76290fb..982b977 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/filter/FilterTestingCluster.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/filter/FilterTestingCluster.java @@ -37,6 +37,9 @@ import org.apache.hadoop.hbase.HTableDescriptor; import org.apache.hadoop.hbase.MasterNotRunningException; import org.apache.hadoop.hbase.TableName; import org.apache.hadoop.hbase.ZooKeeperConnectionException; +import org.apache.hadoop.hbase.client.Admin; +import org.apache.hadoop.hbase.client.Connection; +import org.apache.hadoop.hbase.client.ConnectionFactory; import org.apache.hadoop.hbase.client.HBaseAdmin; import org.apache.hadoop.hbase.client.HTable; import org.apache.hadoop.hbase.client.ScannerCallable; @@ -58,13 +61,13 @@ import org.junit.experimental.categories.Category; @Category({FilterTests.class, MediumTests.class}) public class FilterTestingCluster { private static final HBaseTestingUtility TEST_UTIL = new HBaseTestingUtility(); - private static Configuration conf = null; - private static HBaseAdmin admin = null; - private static List<String> createdTables = new ArrayList<>(); + private static Connection connection; + private static Admin admin = null; + private static List<TableName> createdTables = new ArrayList<>(); - protected static void createTable(String tableName, String columnFamilyName) { + protected static void createTable(TableName tableName, String columnFamilyName) { assertNotNull("HBaseAdmin is not initialized successfully.", admin); - HTableDescriptor desc = new HTableDescriptor(TableName.valueOf(tableName)); + HTableDescriptor desc = new HTableDescriptor(tableName); HColumnDescriptor colDef = new HColumnDescriptor(Bytes.toBytes(columnFamilyName)); desc.addFamily(colDef); @@ -77,15 +80,15 @@ public class FilterTestingCluster { } } - protected static Table openTable(String tableName) throws IOException { - Table table = new HTable(conf, tableName); + protected static Table openTable(TableName tableName) throws IOException { + Table table = connection.getTable(tableName); assertTrue("Fail to create the table", admin.tableExists(tableName)); return table; } private static void deleteTables() { if (admin != null) { - for (String tableName: createdTables){ + for (TableName tableName: createdTables){ try { if (admin.tableExists(tableName)) { admin.disableTable(tableName); @@ -99,10 +102,11 @@ public class FilterTestingCluster { } private static void initialize(Configuration conf) { - FilterTestingCluster.conf = HBaseConfiguration.create(conf); - FilterTestingCluster.conf.setInt(HConstants.HBASE_CLIENT_RETRIES_NUMBER, 1); + conf = HBaseConfiguration.create(conf); + conf.setInt(HConstants.HBASE_CLIENT_RETRIES_NUMBER, 1); try { - admin = new HBaseAdmin(conf); + connection = ConnectionFactory.createConnection(conf); + admin = connection.getAdmin(); } catch (MasterNotRunningException e) { assertNull("Master is not running", e); } catch (ZooKeeperConnectionException e) { @@ -124,7 +128,7 @@ public class FilterTestingCluster { @AfterClass public static void tearDown() throws Exception { deleteTables(); + connection.close(); TEST_UTIL.shutdownMiniCluster(); } - } http://git-wip-us.apache.org/repos/asf/hbase/blob/9246af8d/hbase-server/src/test/java/org/apache/hadoop/hbase/filter/TestFilterWithScanLimits.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/filter/TestFilterWithScanLimits.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/filter/TestFilterWithScanLimits.java index 142b15a..78a4d1f 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/filter/TestFilterWithScanLimits.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/filter/TestFilterWithScanLimits.java @@ -30,6 +30,7 @@ import java.util.List; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; import org.apache.hadoop.hbase.Cell; +import org.apache.hadoop.hbase.TableName; import org.apache.hadoop.hbase.client.Put; import org.apache.hadoop.hbase.client.Result; import org.apache.hadoop.hbase.client.ResultScanner; @@ -50,7 +51,7 @@ public class TestFilterWithScanLimits extends FilterTestingCluster { private static final Log LOG = LogFactory .getLog(TestFilterWithScanLimits.class); - private static final String tableName = "scanWithLimit"; + private static final TableName tableName = TableName.valueOf("scanWithLimit"); private static final String columnFamily = "f1"; @Test http://git-wip-us.apache.org/repos/asf/hbase/blob/9246af8d/hbase-server/src/test/java/org/apache/hadoop/hbase/filter/TestFilterWrapper.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/filter/TestFilterWrapper.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/filter/TestFilterWrapper.java index 1cffe1d..8ce0b76 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/filter/TestFilterWrapper.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/filter/TestFilterWrapper.java @@ -36,6 +36,8 @@ import org.apache.hadoop.hbase.HTableDescriptor; import org.apache.hadoop.hbase.TableName; import org.apache.hadoop.hbase.MasterNotRunningException; import org.apache.hadoop.hbase.ZooKeeperConnectionException; +import org.apache.hadoop.hbase.client.Connection; +import org.apache.hadoop.hbase.client.ConnectionFactory; import org.apache.hadoop.hbase.client.HBaseAdmin; import org.apache.hadoop.hbase.client.HTable; import org.apache.hadoop.hbase.client.Put; @@ -66,6 +68,7 @@ public class TestFilterWrapper { private static Configuration conf = null; private static HBaseAdmin admin = null; private static TableName name = TableName.valueOf("test"); + private static Connection connection; @Test public void testFilterWrapper() { @@ -84,7 +87,7 @@ public class TestFilterWrapper { FilterList filter = new FilterList(fs); scan.setFilter(filter); - Table table = new HTable(conf, name); + Table table = connection.getTable(name); ResultScanner scanner = table.getScanner(scan); // row2 (c1-c4) and row3(c1-c4) are returned @@ -111,7 +114,7 @@ public class TestFilterWrapper { private static void prepareData() { try { - Table table = new HTable(TestFilterWrapper.conf, name); + Table table = connection.getTable(name); assertTrue("Fail to create the table", admin.tableExists(name)); List<Put> puts = new ArrayList<Put>(); @@ -173,7 +176,8 @@ public class TestFilterWrapper { TestFilterWrapper.conf = HBaseConfiguration.create(conf); TestFilterWrapper.conf.setInt(HConstants.HBASE_CLIENT_RETRIES_NUMBER, 1); try { - admin = new HBaseAdmin(conf); + connection = ConnectionFactory.createConnection(TestFilterWrapper.conf); + admin = TEST_UTIL.getHBaseAdmin(); } catch (MasterNotRunningException e) { assertNull("Master is not running", e); } catch (ZooKeeperConnectionException e) { @@ -187,7 +191,6 @@ public class TestFilterWrapper { @BeforeClass public static void setUp() throws Exception { - Configuration config = TEST_UTIL.getConfiguration(); TEST_UTIL.startMiniCluster(1); initialize(TEST_UTIL.getConfiguration()); } @@ -195,6 +198,7 @@ public class TestFilterWrapper { @AfterClass public static void tearDown() throws Exception { deleteTable(); + connection.close(); TEST_UTIL.shutdownMiniCluster(); } http://git-wip-us.apache.org/repos/asf/hbase/blob/9246af8d/hbase-server/src/test/java/org/apache/hadoop/hbase/filter/TestScanRowPrefix.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/filter/TestScanRowPrefix.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/filter/TestScanRowPrefix.java index 100f26d..3be10ec 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/filter/TestScanRowPrefix.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/filter/TestScanRowPrefix.java @@ -22,6 +22,7 @@ package org.apache.hadoop.hbase.filter; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; import org.apache.commons.codec.binary.Hex; +import org.apache.hadoop.hbase.TableName; import org.apache.hadoop.hbase.client.Put; import org.apache.hadoop.hbase.client.Result; import org.apache.hadoop.hbase.client.ResultScanner; @@ -48,7 +49,7 @@ public class TestScanRowPrefix extends FilterTestingCluster { @Test public void testPrefixScanning() throws IOException { - String tableName = "prefixScanning"; + TableName tableName = TableName.valueOf("prefixScanning"); createTable(tableName,"F"); Table table = openTable(tableName); http://git-wip-us.apache.org/repos/asf/hbase/blob/9246af8d/hbase-server/src/test/java/org/apache/hadoop/hbase/io/encoding/TestChangingEncoding.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/io/encoding/TestChangingEncoding.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/io/encoding/TestChangingEncoding.java index fdbef44..918e0f1 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/io/encoding/TestChangingEncoding.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/io/encoding/TestChangingEncoding.java @@ -16,15 +16,6 @@ */ package org.apache.hadoop.hbase.io.encoding; -import static org.junit.Assert.assertTrue; - -import java.io.IOException; -import java.util.ArrayList; -import java.util.Arrays; -import java.util.Collections; -import java.util.List; -import java.util.Random; - import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; import org.apache.hadoop.conf.Configuration; @@ -41,7 +32,6 @@ import org.apache.hadoop.hbase.client.ConnectionFactory; import org.apache.hadoop.hbase.client.Durability; import org.apache.hadoop.hbase.client.Get; import org.apache.hadoop.hbase.client.HBaseAdmin; -import org.apache.hadoop.hbase.client.HTable; import org.apache.hadoop.hbase.client.Put; import org.apache.hadoop.hbase.client.Result; import org.apache.hadoop.hbase.client.Table; @@ -55,6 +45,15 @@ import org.junit.BeforeClass; import org.junit.Test; import org.junit.experimental.categories.Category; +import java.io.IOException; +import java.util.ArrayList; +import java.util.Arrays; +import java.util.Collections; +import java.util.List; +import java.util.Random; + +import static org.junit.Assert.assertTrue; + /** * Tests changing data block encoding settings of a column family. */ @@ -127,7 +126,7 @@ public class TestChangingEncoding { + "_col" + j); } - static void writeTestDataBatch(Configuration conf, TableName tableName, + static void writeTestDataBatch(TableName tableName, int batchId) throws Exception { LOG.debug("Writing test data batch " + batchId); List<Put> puts = new ArrayList<>(); @@ -146,10 +145,10 @@ public class TestChangingEncoding { } } - static void verifyTestDataBatch(Configuration conf, TableName tableName, + static void verifyTestDataBatch(TableName tableName, int batchId) throws Exception { LOG.debug("Verifying test data batch " + batchId); - Table table = new HTable(conf, tableName); + Table table = TEST_UTIL.getConnection().getTable(tableName); for (int i = 0; i < NUM_ROWS_PER_BATCH; ++i) { Get get = new Get(getRowKey(batchId, i)); Result result = table.get(get); @@ -162,13 +161,13 @@ public class TestChangingEncoding { } private void writeSomeNewData() throws Exception { - writeTestDataBatch(conf, tableName, numBatchesWritten); + writeTestDataBatch(tableName, numBatchesWritten); ++numBatchesWritten; } private void verifyAllData() throws Exception { for (int i = 0; i < numBatchesWritten; ++i) { - verifyTestDataBatch(conf, tableName, i); + verifyTestDataBatch(tableName, i); } } http://git-wip-us.apache.org/repos/asf/hbase/blob/9246af8d/hbase-server/src/test/java/org/apache/hadoop/hbase/io/encoding/TestLoadAndSwitchEncodeOnDisk.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/io/encoding/TestLoadAndSwitchEncodeOnDisk.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/io/encoding/TestLoadAndSwitchEncodeOnDisk.java index 26183ac..d6088ec 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/io/encoding/TestLoadAndSwitchEncodeOnDisk.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/io/encoding/TestLoadAndSwitchEncodeOnDisk.java @@ -68,14 +68,14 @@ public class TestLoadAndSwitchEncodeOnDisk extends @Test(timeout=TIMEOUT_MS) public void loadTest() throws Exception { - HBaseAdmin admin = new HBaseAdmin(conf); + HBaseAdmin admin = TEST_UTIL.getHBaseAdmin(); compression = Compression.Algorithm.GZ; // used for table setup super.loadTest(); HColumnDescriptor hcd = getColumnDesc(admin); System.err.println("\nDisabling encode-on-disk. Old column descriptor: " + hcd + "\n"); - HTable t = new HTable(this.conf, TABLE); + HTable t = (HTable) TEST_UTIL.getConnection().getTable(TABLE); assertAllOnLine(t); admin.disableTable(TABLE); http://git-wip-us.apache.org/repos/asf/hbase/blob/9246af8d/hbase-server/src/test/java/org/apache/hadoop/hbase/mapred/TestTableMapReduce.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/mapred/TestTableMapReduce.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/mapred/TestTableMapReduce.java index 107837e..35e6ae5 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/mapred/TestTableMapReduce.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/mapred/TestTableMapReduce.java @@ -26,6 +26,7 @@ import java.io.IOException; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; import org.apache.hadoop.fs.FileUtil; +import org.apache.hadoop.hbase.client.Table; import org.apache.hadoop.hbase.testclassification.LargeTests; import org.apache.hadoop.hbase.testclassification.MapReduceTests; import org.apache.hadoop.hbase.client.HTable; @@ -73,7 +74,7 @@ public class TestTableMapReduce extends TestTableMapReduceBase { } @Override - protected void runTestOnTable(HTable table) throws IOException { + protected void runTestOnTable(Table table) throws IOException { JobConf jobConf = null; try { LOG.info("Before map/reduce startup"); http://git-wip-us.apache.org/repos/asf/hbase/blob/9246af8d/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TableSnapshotInputFormatTestBase.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TableSnapshotInputFormatTestBase.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TableSnapshotInputFormatTestBase.java index 1f8ccae..43e29ad 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TableSnapshotInputFormatTestBase.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TableSnapshotInputFormatTestBase.java @@ -28,6 +28,7 @@ import org.apache.hadoop.hbase.TableName; import org.apache.hadoop.hbase.client.Admin; import org.apache.hadoop.hbase.client.HTable; import org.apache.hadoop.hbase.client.Result; +import org.apache.hadoop.hbase.client.Table; import org.apache.hadoop.hbase.io.ImmutableBytesWritable; import org.apache.hadoop.hbase.master.snapshot.SnapshotManager; import org.apache.hadoop.hbase.snapshot.SnapshotTestingUtils; @@ -152,7 +153,7 @@ public abstract class TableSnapshotInputFormatTestBase { Admin admin = util.getHBaseAdmin(); // put some stuff in the table - HTable table = new HTable(util.getConfiguration(), tableName); + Table table = util.getConnection().getTable(tableName); util.loadTable(table, FAMILIES); Path rootDir = FSUtils.getRootDir(util.getConfiguration()); http://git-wip-us.apache.org/repos/asf/hbase/blob/9246af8d/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestHFileOutputFormat.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestHFileOutputFormat.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestHFileOutputFormat.java index c447895..92a7b42 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestHFileOutputFormat.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestHFileOutputFormat.java @@ -24,6 +24,7 @@ import static org.junit.Assert.assertNotNull; import static org.junit.Assert.assertNotSame; import static org.junit.Assert.assertTrue; import static org.junit.Assert.fail; +import static org.mockito.Mockito.when; import java.io.IOException; import java.util.Arrays; @@ -55,6 +56,7 @@ import org.apache.hadoop.hbase.KeyValue; import org.apache.hadoop.hbase.PerformanceEvaluation; import org.apache.hadoop.hbase.TableDescriptor; import org.apache.hadoop.hbase.TableName; +import org.apache.hadoop.hbase.client.Admin; import org.apache.hadoop.hbase.client.HBaseAdmin; import org.apache.hadoop.hbase.client.HRegionLocator; import org.apache.hadoop.hbase.client.HTable; @@ -378,7 +380,7 @@ public class TestHFileOutputFormat { try { util.startMiniCluster(); Path testDir = util.getDataTestDirOnTestFS("testLocalMRIncrementalLoad"); - admin = new HBaseAdmin(conf); + admin = util.getHBaseAdmin(); HTable table = util.createTable(TABLE_NAME, FAMILIES); assertEquals("Should start with empty table", 0, util.countRows(table)); @@ -408,7 +410,7 @@ public class TestHFileOutputFormat { // handle the split case if (shouldChangeRegions) { LOG.info("Changing regions in table"); - admin.disableTable(table.getTableName()); + admin.disableTable(table.getName()); while(util.getMiniHBaseCluster().getMaster().getAssignmentManager(). getRegionStates().isRegionsInTransition()) { Threads.sleep(200); @@ -417,9 +419,9 @@ public class TestHFileOutputFormat { byte[][] newStartKeys = generateRandomStartKeys(15); util.createMultiRegions( util.getConfiguration(), table, FAMILIES[0], newStartKeys); - admin.enableTable(table.getTableName()); + admin.enableTable(table.getName()); while (table.getRegionLocations().size() != 15 || - !admin.isTableAvailable(table.getTableName())) { + !admin.isTableAvailable(table.getName())) { Thread.sleep(200); LOG.info("Waiting for new region assignment to happen"); } @@ -967,7 +969,7 @@ public class TestHFileOutputFormat { util.startMiniCluster(); Path testDir = util.getDataTestDirOnTestFS("testExcludeMinorCompaction"); final FileSystem fs = util.getDFSCluster().getFileSystem(); - HBaseAdmin admin = new HBaseAdmin(conf); + Admin admin = util.getHBaseAdmin(); HTable table = util.createTable(TABLE_NAME, FAMILIES); assertEquals("Should start with empty table", 0, util.countRows(table)); @@ -982,7 +984,7 @@ public class TestHFileOutputFormat { Put p = new Put(Bytes.toBytes("test")); p.add(FAMILIES[0], Bytes.toBytes("1"), Bytes.toBytes("1")); table.put(p); - admin.flush(TABLE_NAME.getName()); + admin.flush(TABLE_NAME); assertEquals(1, util.countRows(table)); quickPoll(new Callable<Boolean>() { public Boolean call() throws Exception { @@ -1008,7 +1010,7 @@ public class TestHFileOutputFormat { assertEquals(2, fs.listStatus(storePath).length); // minor compactions shouldn't get rid of the file - admin.compact(TABLE_NAME.getName()); + admin.compact(TABLE_NAME); try { quickPoll(new Callable<Boolean>() { public Boolean call() throws Exception { @@ -1021,7 +1023,7 @@ public class TestHFileOutputFormat { } // a major compaction should work though - admin.majorCompact(TABLE_NAME.getName()); + admin.majorCompact(TABLE_NAME); quickPoll(new Callable<Boolean>() { public Boolean call() throws Exception { return fs.listStatus(storePath).length == 1; @@ -1063,7 +1065,7 @@ public class TestHFileOutputFormat { admin.enableTable(tname); } else if ("incremental".equals(args[0])) { TableName tname = TableName.valueOf(args[1]); - HTable table = new HTable(conf, tname); + HTable table = (HTable) util.getConnection().getTable(tname); Path outDir = new Path("incremental-out"); runIncrementalPELoad(conf, table, outDir); } else { http://git-wip-us.apache.org/repos/asf/hbase/blob/9246af8d/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestImportExport.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestImportExport.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestImportExport.java index e828fa6..935d462 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestImportExport.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestImportExport.java @@ -255,7 +255,7 @@ public class TestImportExport { .setMaxVersions(1) ); UTIL.getHBaseAdmin().createTable(desc); - Table t = new HTable(UTIL.getConfiguration(), desc.getTableName()); + Table t = UTIL.getConnection().getTable(desc.getTableName()); Put p = new Put(ROW1); p.add(FAMILYA, QUAL, now, QUAL); @@ -286,7 +286,7 @@ public class TestImportExport { .setKeepDeletedCells(true) ); UTIL.getHBaseAdmin().createTable(desc); - Table t = new HTable(UTIL.getConfiguration(), desc.getTableName()); + Table t = UTIL.getConnection().getTable(desc.getTableName()); Put p = new Put(ROW1); p.add(FAMILYA, QUAL, now, QUAL); @@ -318,7 +318,7 @@ public class TestImportExport { ); UTIL.getHBaseAdmin().createTable(desc); t.close(); - t = new HTable(UTIL.getConfiguration(), desc.getTableName()); + t = UTIL.getConnection().getTable(desc.getTableName()); args = new String[] { IMPORT_TABLE, FQ_OUTPUT_DIR @@ -344,14 +344,16 @@ public class TestImportExport { @Test public void testWithMultipleDeleteFamilyMarkersOfSameRowSameFamily() throws Exception { - String EXPORT_TABLE = "exportWithMultipleDeleteFamilyMarkersOfSameRowSameFamily"; - HTableDescriptor desc = new HTableDescriptor(TableName.valueOf(EXPORT_TABLE)); + TableName EXPORT_TABLE = + TableName.valueOf("exportWithMultipleDeleteFamilyMarkersOfSameRowSameFamily"); + HTableDescriptor desc = new HTableDescriptor(EXPORT_TABLE); desc.addFamily(new HColumnDescriptor(FAMILYA) .setMaxVersions(5) .setKeepDeletedCells(true) ); UTIL.getHBaseAdmin().createTable(desc); - HTable exportT = new HTable(UTIL.getConfiguration(), EXPORT_TABLE); + + Table exportT = UTIL.getConnection().getTable(EXPORT_TABLE); //Add first version of QUAL Put p = new Put(ROW1); @@ -373,8 +375,7 @@ public class TestImportExport { String[] args = new String[] { - "-D" + Export.RAW_SCAN + "=true", - EXPORT_TABLE, + "-D" + Export.RAW_SCAN + "=true", EXPORT_TABLE.getNameAsString(), FQ_OUTPUT_DIR, "1000", // max number of key versions per key to export }; @@ -387,8 +388,8 @@ public class TestImportExport { .setKeepDeletedCells(true) ); UTIL.getHBaseAdmin().createTable(desc); - - HTable importT = new HTable(UTIL.getConfiguration(), IMPORT_TABLE); + + Table importT = UTIL.getConnection().getTable(TableName.valueOf(IMPORT_TABLE)); args = new String[] { IMPORT_TABLE, FQ_OUTPUT_DIR @@ -429,7 +430,7 @@ public class TestImportExport { HTableDescriptor desc = new HTableDescriptor(TableName.valueOf(EXPORT_TABLE)); desc.addFamily(new HColumnDescriptor(FAMILYA).setMaxVersions(5)); UTIL.getHBaseAdmin().createTable(desc); - Table exportTable = new HTable(UTIL.getConfiguration(), desc.getTableName()); + Table exportTable = UTIL.getConnection().getTable(desc.getTableName()); Put p1 = new Put(ROW1); p1.add(FAMILYA, QUAL, now, QUAL); @@ -454,7 +455,7 @@ public class TestImportExport { desc.addFamily(new HColumnDescriptor(FAMILYA).setMaxVersions(5)); UTIL.getHBaseAdmin().createTable(desc); - Table importTable = new HTable(UTIL.getConfiguration(), desc.getTableName()); + Table importTable = UTIL.getConnection().getTable(desc.getTableName()); args = new String[] { "-D" + Import.FILTER_CLASS_CONF_KEY + "=" + PrefixFilter.class.getName(), "-D" + Import.FILTER_ARGS_CONF_KEY + "=" + Bytes.toString(ROW1), IMPORT_TABLE, FQ_OUTPUT_DIR, "1000" }; http://git-wip-us.apache.org/repos/asf/hbase/blob/9246af8d/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestImportTSVWithOperationAttributes.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestImportTSVWithOperationAttributes.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestImportTSVWithOperationAttributes.java index eddee5a..8bd6771 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestImportTSVWithOperationAttributes.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestImportTSVWithOperationAttributes.java @@ -39,6 +39,8 @@ import org.apache.hadoop.hbase.HBaseTestingUtility; import org.apache.hadoop.hbase.HConstants; import org.apache.hadoop.hbase.TableName; import org.apache.hadoop.hbase.client.Admin; +import org.apache.hadoop.hbase.client.Connection; +import org.apache.hadoop.hbase.client.ConnectionFactory; import org.apache.hadoop.hbase.testclassification.LargeTests; import org.apache.hadoop.hbase.testclassification.MapReduceTests; import org.apache.hadoop.hbase.client.Durability; @@ -100,7 +102,6 @@ public class TestImportTSVWithOperationAttributes implements Configurable { conf.set("hbase.coprocessor.master.classes", OperationAttributesTestController.class.getName()); conf.set("hbase.coprocessor.region.classes", OperationAttributesTestController.class.getName()); util.startMiniCluster(); - Admin admin = new HBaseAdmin(util.getConfiguration()); util.startMiniMapReduceCluster(); } @@ -196,7 +197,8 @@ public class TestImportTSVWithOperationAttributes implements Configurable { int valueMultiplier, boolean dataAvailable) throws IOException { LOG.debug("Validating table."); - Table table = new HTable(conf, tableName); + Connection connection = ConnectionFactory.createConnection(conf); + Table table = connection.getTable(tableName); boolean verified = false; long pause = conf.getLong("hbase.client.pause", 5 * 1000); int numRetries = conf.getInt(HConstants.HBASE_CLIENT_RETRIES_NUMBER, 5); @@ -238,6 +240,7 @@ public class TestImportTSVWithOperationAttributes implements Configurable { } } table.close(); + connection.close(); assertTrue(verified); } http://git-wip-us.apache.org/repos/asf/hbase/blob/9246af8d/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestImportTSVWithVisibilityLabels.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestImportTSVWithVisibilityLabels.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestImportTSVWithVisibilityLabels.java index 0ca0f8f..6754ce9 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestImportTSVWithVisibilityLabels.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestImportTSVWithVisibilityLabels.java @@ -122,7 +122,6 @@ public class TestImportTSVWithVisibilityLabels implements Configurable { // Wait for the labels table to become available util.waitTableEnabled(VisibilityConstants.LABELS_TABLE_NAME.getName(), 50000); createLabels(); - Admin admin = new HBaseAdmin(util.getConfiguration()); util.startMiniMapReduceCluster(); } @@ -185,7 +184,7 @@ public class TestImportTSVWithVisibilityLabels implements Configurable { private void issueDeleteAndVerifyData(TableName tableName) throws IOException { LOG.debug("Validating table after delete."); - Table table = new HTable(conf, tableName); + Table table = util.getConnection().getTable(tableName); boolean verified = false; long pause = conf.getLong("hbase.client.pause", 5 * 1000); int numRetries = conf.getInt(HConstants.HBASE_CLIENT_RETRIES_NUMBER, 5); @@ -370,7 +369,7 @@ public class TestImportTSVWithVisibilityLabels implements Configurable { int valueMultiplier) throws IOException { LOG.debug("Validating table."); - Table table = new HTable(conf, tableName); + Table table = util.getConnection().getTable(tableName); boolean verified = false; long pause = conf.getLong("hbase.client.pause", 5 * 1000); int numRetries = conf.getInt(HConstants.HBASE_CLIENT_RETRIES_NUMBER, 5); http://git-wip-us.apache.org/repos/asf/hbase/blob/9246af8d/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestImportTsv.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestImportTsv.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestImportTsv.java index 3844a64..9e2e4be 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestImportTsv.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestImportTsv.java @@ -43,6 +43,8 @@ import org.apache.hadoop.hbase.HBaseTestingUtility; import org.apache.hadoop.hbase.HConstants; import org.apache.hadoop.hbase.TableName; import org.apache.hadoop.hbase.TableNotFoundException; +import org.apache.hadoop.hbase.client.Connection; +import org.apache.hadoop.hbase.client.ConnectionFactory; import org.apache.hadoop.hbase.client.HTable; import org.apache.hadoop.hbase.client.Result; import org.apache.hadoop.hbase.client.ResultScanner; @@ -346,7 +348,8 @@ public class TestImportTsv implements Configurable { String family, int valueMultiplier) throws IOException { LOG.debug("Validating table."); - Table table = new HTable(conf, tableName); + Connection connection = ConnectionFactory.createConnection(conf); + Table table = connection.getTable(tableName); boolean verified = false; long pause = conf.getLong("hbase.client.pause", 5 * 1000); int numRetries = conf.getInt(HConstants.HBASE_CLIENT_RETRIES_NUMBER, 5); @@ -378,6 +381,7 @@ public class TestImportTsv implements Configurable { } } table.close(); + connection.close(); assertTrue(verified); } http://git-wip-us.apache.org/repos/asf/hbase/blob/9246af8d/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestLoadIncrementalHFiles.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestLoadIncrementalHFiles.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestLoadIncrementalHFiles.java index fff0200..d68d55c 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestLoadIncrementalHFiles.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestLoadIncrementalHFiles.java @@ -255,7 +255,7 @@ public class TestLoadIncrementalHFiles { String [] args= {dir.toString(), tableName.toString()}; loader.run(args); - Table table = new HTable(util.getConfiguration(), tableName); + Table table = util.getConnection().getTable(tableName); try { assertEquals(expectedRows, util.countRows(table)); } finally { http://git-wip-us.apache.org/repos/asf/hbase/blob/9246af8d/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestLoadIncrementalHFilesSplitRecovery.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestLoadIncrementalHFilesSplitRecovery.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestLoadIncrementalHFilesSplitRecovery.java index 4d4043b..e07353f 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestLoadIncrementalHFilesSplitRecovery.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestLoadIncrementalHFilesSplitRecovery.java @@ -242,7 +242,7 @@ public class TestLoadIncrementalHFilesSplitRecovery { assertEquals(htds.length, 1); Table t = null; try { - t = new HTable(util.getConfiguration(), table); + t = util.getConnection().getTable(table); Scan s = new Scan(); ResultScanner sr = t.getScanner(s); int i = 0; http://git-wip-us.apache.org/repos/asf/hbase/blob/9246af8d/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestMultithreadedTableMapper.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestMultithreadedTableMapper.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestMultithreadedTableMapper.java index 20d577d..99684e9 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestMultithreadedTableMapper.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestMultithreadedTableMapper.java @@ -126,11 +126,10 @@ public class TestMultithreadedTableMapper { @Test public void testMultithreadedTableMapper() throws IOException, InterruptedException, ClassNotFoundException { - runTestOnTable(new HTable(new Configuration(UTIL.getConfiguration()), - MULTI_REGION_TABLE_NAME)); + runTestOnTable(UTIL.getConnection().getTable(MULTI_REGION_TABLE_NAME)); } - private void runTestOnTable(HTable table) + private void runTestOnTable(Table table) throws IOException, InterruptedException, ClassNotFoundException { Job job = null; try { @@ -140,16 +139,16 @@ public class TestMultithreadedTableMapper { Scan scan = new Scan(); scan.addFamily(INPUT_FAMILY); TableMapReduceUtil.initTableMapperJob( - table.getTableName(), scan, + table.getName(), scan, MultithreadedTableMapper.class, ImmutableBytesWritable.class, Put.class, job); MultithreadedTableMapper.setMapperClass(job, ProcessContentsMapper.class); MultithreadedTableMapper.setNumberOfThreads(job, NUMBER_OF_THREADS); TableMapReduceUtil.initTableReducerJob( - Bytes.toString(table.getTableName()), + table.getName().getNameAsString(), IdentityTableReducer.class, job); FileOutputFormat.setOutputPath(job, new Path("test")); - LOG.info("Started " + table.getTableName()); + LOG.info("Started " + table.getName()); assertTrue(job.waitForCompletion(true)); LOG.info("After map/reduce completion"); // verify map-reduce results @@ -164,7 +163,7 @@ public class TestMultithreadedTableMapper { } private void verify(TableName tableName) throws IOException { - Table table = new HTable(new Configuration(UTIL.getConfiguration()), tableName); + Table table = UTIL.getConnection().getTable(tableName); boolean verified = false; long pause = UTIL.getConfiguration().getLong("hbase.client.pause", 5 * 1000); int numRetries = UTIL.getConfiguration().getInt(HConstants.HBASE_CLIENT_RETRIES_NUMBER, 5); http://git-wip-us.apache.org/repos/asf/hbase/blob/9246af8d/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestTableMapReduce.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestTableMapReduce.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestTableMapReduce.java index 11a35f0..1dcbd2a 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestTableMapReduce.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestTableMapReduce.java @@ -33,6 +33,7 @@ import org.apache.hadoop.hbase.client.HTable; import org.apache.hadoop.hbase.client.Put; import org.apache.hadoop.hbase.client.Result; import org.apache.hadoop.hbase.client.Scan; +import org.apache.hadoop.hbase.client.Table; import org.apache.hadoop.hbase.io.ImmutableBytesWritable; import org.apache.hadoop.hbase.testclassification.LargeTests; import org.apache.hadoop.hbase.testclassification.VerySlowMapReduceTests; @@ -90,7 +91,7 @@ public class TestTableMapReduce extends TestTableMapReduceBase { } } - protected void runTestOnTable(HTable table) throws IOException { + protected void runTestOnTable(Table table) throws IOException { Job job = null; try { LOG.info("Before map/reduce startup"); @@ -99,14 +100,14 @@ public class TestTableMapReduce extends TestTableMapReduceBase { Scan scan = new Scan(); scan.addFamily(INPUT_FAMILY); TableMapReduceUtil.initTableMapperJob( - Bytes.toString(table.getTableName()), scan, + table.getName().getNameAsString(), scan, ProcessContentsMapper.class, ImmutableBytesWritable.class, Put.class, job); TableMapReduceUtil.initTableReducerJob( - Bytes.toString(table.getTableName()), + table.getName().getNameAsString(), IdentityTableReducer.class, job); FileOutputFormat.setOutputPath(job, new Path("test")); - LOG.info("Started " + Bytes.toString(table.getTableName())); + LOG.info("Started " + table.getName().getNameAsString()); assertTrue(job.waitForCompletion(true)); LOG.info("After map/reduce completion"); http://git-wip-us.apache.org/repos/asf/hbase/blob/9246af8d/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestTableMapReduceBase.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestTableMapReduceBase.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestTableMapReduceBase.java index c82d82f..b42966c 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestTableMapReduceBase.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestTableMapReduceBase.java @@ -71,7 +71,7 @@ public abstract class TestTableMapReduceBase { /** * Handles API-specifics for setting up and executing the job. */ - protected abstract void runTestOnTable(HTable table) throws IOException; + protected abstract void runTestOnTable(Table table) throws IOException; @BeforeClass public static void beforeClass() throws Exception { @@ -95,7 +95,7 @@ public abstract class TestTableMapReduceBase { */ @Test public void testMultiRegionTable() throws IOException { - runTestOnTable(new HTable(UTIL.getConfiguration(), MULTI_REGION_TABLE_NAME)); + runTestOnTable(UTIL.getConnection().getTable(MULTI_REGION_TABLE_NAME)); } @Test @@ -103,7 +103,7 @@ public abstract class TestTableMapReduceBase { Configuration conf = new Configuration(UTIL.getConfiguration()); // force use of combiner for testing purposes conf.setInt("mapreduce.map.combine.minspills", 1); - runTestOnTable(new HTable(conf, MULTI_REGION_TABLE_NAME)); + runTestOnTable(UTIL.getConnection().getTable(MULTI_REGION_TABLE_NAME)); } /** @@ -134,7 +134,7 @@ public abstract class TestTableMapReduceBase { } protected void verify(TableName tableName) throws IOException { - Table table = new HTable(UTIL.getConfiguration(), tableName); + Table table = UTIL.getConnection().getTable(tableName); boolean verified = false; long pause = UTIL.getConfiguration().getLong("hbase.client.pause", 5 * 1000); int numRetries = UTIL.getConfiguration().getInt(HConstants.HBASE_CLIENT_RETRIES_NUMBER, 5); http://git-wip-us.apache.org/repos/asf/hbase/blob/9246af8d/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestTimeRangeMapRed.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestTimeRangeMapRed.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestTimeRangeMapRed.java index f138348..03da1ed 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestTimeRangeMapRed.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestTimeRangeMapRed.java @@ -18,28 +18,25 @@ */ package org.apache.hadoop.hbase.mapreduce; -import java.io.File; -import java.io.IOException; -import java.util.ArrayList; -import java.util.List; -import java.util.Map; -import java.util.NavigableMap; -import java.util.TreeMap; - import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; import org.apache.hadoop.conf.Configurable; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.FileUtil; -import org.apache.hadoop.hbase.*; +import org.apache.hadoop.hbase.Cell; +import org.apache.hadoop.hbase.CellUtil; +import org.apache.hadoop.hbase.HBaseTestingUtility; +import org.apache.hadoop.hbase.HColumnDescriptor; +import org.apache.hadoop.hbase.HTableDescriptor; +import org.apache.hadoop.hbase.TableName; import org.apache.hadoop.hbase.client.Admin; -import org.apache.hadoop.hbase.client.HBaseAdmin; -import org.apache.hadoop.hbase.client.HTable; +import org.apache.hadoop.hbase.client.Connection; +import org.apache.hadoop.hbase.client.ConnectionFactory; +import org.apache.hadoop.hbase.client.Durability; import org.apache.hadoop.hbase.client.Put; import org.apache.hadoop.hbase.client.Result; import org.apache.hadoop.hbase.client.ResultScanner; import org.apache.hadoop.hbase.client.Scan; -import org.apache.hadoop.hbase.client.Durability; import org.apache.hadoop.hbase.client.Table; import org.apache.hadoop.hbase.io.ImmutableBytesWritable; import org.apache.hadoop.hbase.testclassification.LargeTests; @@ -49,13 +46,20 @@ import org.apache.hadoop.io.MapWritable; import org.apache.hadoop.io.Text; import org.apache.hadoop.mapreduce.Job; import org.apache.hadoop.mapreduce.lib.output.NullOutputFormat; -import org.junit.After; import org.junit.AfterClass; import org.junit.Before; import org.junit.BeforeClass; import org.junit.Test; import org.junit.experimental.categories.Category; +import java.io.File; +import java.io.IOException; +import java.util.ArrayList; +import java.util.List; +import java.util.Map; +import java.util.NavigableMap; +import java.util.TreeMap; + @Category({MapReduceTests.class, LargeTests.class}) public class TestTimeRangeMapRed { private final static Log log = LogFactory.getLog(TestTimeRangeMapRed.class); @@ -94,12 +98,7 @@ public class TestTimeRangeMapRed { @Before public void before() throws Exception { - this.admin = new HBaseAdmin(UTIL.getConfiguration()); - } - - @After - public void after() throws IOException { - this.admin.close(); + this.admin = UTIL.getHBaseAdmin(); } private static class ProcessTimeRangeMapper @@ -137,7 +136,8 @@ public class TestTimeRangeMapRed { public void setConf(Configuration configuration) { this.conf = configuration; try { - table = new HTable(HBaseConfiguration.create(conf), TABLE_NAME); + Connection connection = ConnectionFactory.createConnection(conf); + table = connection.getTable(TABLE_NAME); } catch (IOException e) { e.printStackTrace(); } @@ -159,7 +159,7 @@ public class TestTimeRangeMapRed { put.add(FAMILY_NAME, COLUMN_NAME, entry.getKey(), Bytes.toBytes(false)); puts.add(put); } - Table table = new HTable(UTIL.getConfiguration(), desc.getTableName()); + Table table = UTIL.getConnection().getTable(desc.getTableName()); table.put(puts); runTestOnTable(); verify(table); @@ -204,7 +204,7 @@ public class TestTimeRangeMapRed { + "\t" + Bytes.toString(CellUtil.cloneQualifier(kv)) + "\t" + kv.getTimestamp() + "\t" + Bytes.toBoolean(CellUtil.cloneValue(kv))); org.junit.Assert.assertEquals(TIMESTAMP.get(kv.getTimestamp()), - (Boolean)Bytes.toBoolean(CellUtil.cloneValue(kv))); + Bytes.toBoolean(CellUtil.cloneValue(kv))); } } scanner.close(); http://git-wip-us.apache.org/repos/asf/hbase/blob/9246af8d/hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestAssignmentListener.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestAssignmentListener.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestAssignmentListener.java index fccff59..f482dfa 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestAssignmentListener.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestAssignmentListener.java @@ -205,7 +205,7 @@ public class TestAssignmentListener { assertEquals(0, listener.getCloseCount()); // Add some data - Table table = new HTable(TEST_UTIL.getConfiguration(), TABLE_NAME); + Table table = TEST_UTIL.getConnection().getTable(TABLE_NAME); try { for (int i = 0; i < 10; ++i) { byte[] key = Bytes.toBytes("row-" + i); http://git-wip-us.apache.org/repos/asf/hbase/blob/9246af8d/hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestAssignmentManagerOnCluster.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestAssignmentManagerOnCluster.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestAssignmentManagerOnCluster.java index e6d08b9..d1769e8 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestAssignmentManagerOnCluster.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestAssignmentManagerOnCluster.java @@ -181,7 +181,7 @@ public class TestAssignmentManagerOnCluster { desc.addFamily(new HColumnDescriptor(FAMILY)); admin.createTable(desc); - Table meta = new HTable(conf, TableName.META_TABLE_NAME); + Table meta = TEST_UTIL.getConnection().getTable(TableName.META_TABLE_NAME); HRegionInfo hri = new HRegionInfo( desc.getTableName(), Bytes.toBytes("A"), Bytes.toBytes("Z")); MetaTableAccessor.addRegionToMeta(meta, hri); @@ -224,7 +224,7 @@ public class TestAssignmentManagerOnCluster { desc.addFamily(new HColumnDescriptor(FAMILY)); admin.createTable(desc); - Table meta = new HTable(conf, TableName.META_TABLE_NAME); + Table meta = TEST_UTIL.getConnection().getTable(TableName.META_TABLE_NAME); final HRegionInfo hri = new HRegionInfo( desc.getTableName(), Bytes.toBytes("A"), Bytes.toBytes("Z")); MetaTableAccessor.addRegionToMeta(meta, hri); @@ -436,7 +436,7 @@ public class TestAssignmentManagerOnCluster { desc.addFamily(new HColumnDescriptor(FAMILY)); admin.createTable(desc); - Table meta = new HTable(conf, TableName.META_TABLE_NAME); + Table meta = TEST_UTIL.getConnection().getTable(TableName.META_TABLE_NAME); HRegionInfo hri = new HRegionInfo( desc.getTableName(), Bytes.toBytes("A"), Bytes.toBytes("Z")); MetaTableAccessor.addRegionToMeta(meta, hri); @@ -485,7 +485,7 @@ public class TestAssignmentManagerOnCluster { desc.addFamily(new HColumnDescriptor(FAMILY)); admin.createTable(desc); - Table meta = new HTable(conf, TableName.META_TABLE_NAME); + Table meta = TEST_UTIL.getConnection().getTable(TableName.META_TABLE_NAME); HRegionInfo hri = new HRegionInfo( desc.getTableName(), Bytes.toBytes("A"), Bytes.toBytes("Z")); MetaTableAccessor.addRegionToMeta(meta, hri); @@ -531,7 +531,7 @@ public class TestAssignmentManagerOnCluster { desc.addFamily(new HColumnDescriptor(FAMILY)); admin.createTable(desc); - Table meta = new HTable(conf, TableName.META_TABLE_NAME); + Table meta = TEST_UTIL.getConnection().getTable(TableName.META_TABLE_NAME); HRegionInfo hri = new HRegionInfo( desc.getTableName(), Bytes.toBytes("A"), Bytes.toBytes("Z")); MetaTableAccessor.addRegionToMeta(meta, hri); @@ -573,7 +573,7 @@ public class TestAssignmentManagerOnCluster { desc.addFamily(new HColumnDescriptor(FAMILY)); admin.createTable(desc); - Table meta = new HTable(conf, TableName.META_TABLE_NAME); + Table meta = TEST_UTIL.getConnection().getTable(TableName.META_TABLE_NAME); HRegionInfo hri = new HRegionInfo( desc.getTableName(), Bytes.toBytes("A"), Bytes.toBytes("Z")); MetaTableAccessor.addRegionToMeta(meta, hri); @@ -658,7 +658,7 @@ public class TestAssignmentManagerOnCluster { desc.addFamily(new HColumnDescriptor(FAMILY)); admin.createTable(desc); - Table meta = new HTable(conf, TableName.META_TABLE_NAME); + Table meta = TEST_UTIL.getConnection().getTable(TableName.META_TABLE_NAME); HRegionInfo hri = new HRegionInfo( desc.getTableName(), Bytes.toBytes("A"), Bytes.toBytes("Z")); MetaTableAccessor.addRegionToMeta(meta, hri); @@ -703,7 +703,7 @@ public class TestAssignmentManagerOnCluster { desc.addFamily(new HColumnDescriptor(FAMILY)); admin.createTable(desc); - Table meta = new HTable(conf, TableName.META_TABLE_NAME); + Table meta = TEST_UTIL.getConnection().getTable(TableName.META_TABLE_NAME); HRegionInfo hri = new HRegionInfo( desc.getTableName(), Bytes.toBytes("A"), Bytes.toBytes("Z")); MetaTableAccessor.addRegionToMeta(meta, hri); @@ -776,7 +776,7 @@ public class TestAssignmentManagerOnCluster { desc.addFamily(new HColumnDescriptor(FAMILY)); admin.createTable(desc); - Table meta = new HTable(conf, TableName.META_TABLE_NAME); + Table meta = TEST_UTIL.getConnection().getTable(TableName.META_TABLE_NAME); HRegionInfo hri = new HRegionInfo( desc.getTableName(), Bytes.toBytes("A"), Bytes.toBytes("Z")); MetaTableAccessor.addRegionToMeta(meta, hri); @@ -904,7 +904,7 @@ public class TestAssignmentManagerOnCluster { desc.addFamily(new HColumnDescriptor(FAMILY)); admin.createTable(desc); - Table meta = new HTable(conf, TableName.META_TABLE_NAME); + Table meta = TEST_UTIL.getConnection().getTable(TableName.META_TABLE_NAME); HRegionInfo hri = new HRegionInfo( desc.getTableName(), Bytes.toBytes("A"), Bytes.toBytes("Z")); MetaTableAccessor.addRegionToMeta(meta, hri); @@ -945,7 +945,7 @@ public class TestAssignmentManagerOnCluster { desc.addFamily(new HColumnDescriptor(FAMILY)); admin.createTable(desc); - Table meta = new HTable(conf, TableName.META_TABLE_NAME); + Table meta = TEST_UTIL.getConnection().getTable(TableName.META_TABLE_NAME); HRegionInfo hri = new HRegionInfo( desc.getTableName(), Bytes.toBytes("A"), Bytes.toBytes("Z")); MetaTableAccessor.addRegionToMeta(meta, hri); @@ -1024,7 +1024,7 @@ public class TestAssignmentManagerOnCluster { desc.addFamily(new HColumnDescriptor(FAMILY)); admin.createTable(desc); - Table meta = new HTable(conf, TableName.META_TABLE_NAME); + Table meta = TEST_UTIL.getConnection().getTable(TableName.META_TABLE_NAME); HRegionInfo hri = new HRegionInfo( desc.getTableName(), Bytes.toBytes("A"), Bytes.toBytes("Z")); MetaTableAccessor.addRegionToMeta(meta, hri); @@ -1099,7 +1099,7 @@ public class TestAssignmentManagerOnCluster { HTableDescriptor desc = new HTableDescriptor(TableName.valueOf(table)); desc.addFamily(new HColumnDescriptor(FAMILY)); admin.createTable(desc); - Table meta = new HTable(conf, TableName.META_TABLE_NAME); + Table meta = TEST_UTIL.getConnection().getTable(TableName.META_TABLE_NAME); HRegionInfo hri = new HRegionInfo(desc.getTableName(), Bytes.toBytes("A"), Bytes.toBytes("Z")); MetaTableAccessor.addRegionToMeta(meta, hri); http://git-wip-us.apache.org/repos/asf/hbase/blob/9246af8d/hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestMasterOperationsForRegionReplicas.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestMasterOperationsForRegionReplicas.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestMasterOperationsForRegionReplicas.java index 846f8e6..6307c4c 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestMasterOperationsForRegionReplicas.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestMasterOperationsForRegionReplicas.java @@ -253,7 +253,7 @@ public class TestMasterOperationsForRegionReplicas { ADMIN.disableTable(table); // now delete one replica info from all the rows // this is to make the meta appear to be only partially updated - Table metaTable = new HTable(TableName.META_TABLE_NAME, ADMIN.getConnection()); + Table metaTable = ADMIN.getConnection().getTable(TableName.META_TABLE_NAME); for (byte[] row : tableRows) { Delete deleteOneReplicaLocation = new Delete(row); deleteOneReplicaLocation.deleteColumns(HConstants.CATALOG_FAMILY, http://git-wip-us.apache.org/repos/asf/hbase/blob/9246af8d/hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestMasterRestartAfterDisablingTable.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestMasterRestartAfterDisablingTable.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestMasterRestartAfterDisablingTable.java index 56961d5..1594f80 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestMasterRestartAfterDisablingTable.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestMasterRestartAfterDisablingTable.java @@ -99,7 +99,7 @@ public class TestMasterRestartAfterDisablingTable { TableState.State.DISABLING)); log("Enabling table\n"); // Need a new Admin, the previous one is on the old master - Admin admin = new HBaseAdmin(TEST_UTIL.getConfiguration()); + Admin admin = TEST_UTIL.getHBaseAdmin(); admin.enableTable(table); admin.close(); log("Waiting for no more RIT\n"); http://git-wip-us.apache.org/repos/asf/hbase/blob/9246af8d/hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestMasterTransitions.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestMasterTransitions.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestMasterTransitions.java index 374366e..7ee69f9 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestMasterTransitions.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestMasterTransitions.java @@ -65,7 +65,7 @@ public class TestMasterTransitions { TEST_UTIL.startMiniCluster(2); // Create a table of three families. This will assign a region. TEST_UTIL.createTable(TABLENAME, FAMILIES); - HTable t = new HTable(TEST_UTIL.getConfiguration(), TABLENAME); + HTable t = (HTable) TEST_UTIL.getConnection().getTable(TABLENAME); int countOfRegions = TEST_UTIL.createMultiRegions(t, getTestFamily()); TEST_UTIL.waitUntilAllRegionsAssigned(TABLENAME); addToEachStartKey(countOfRegions); @@ -480,9 +480,8 @@ public class TestMasterTransitions { * @throws IOException */ private static int addToEachStartKey(final int expected) throws IOException { - Table t = new HTable(TEST_UTIL.getConfiguration(), TABLENAME); - Table meta = new HTable(TEST_UTIL.getConfiguration(), - TableName.META_TABLE_NAME); + Table t = TEST_UTIL.getConnection().getTable(TABLENAME); + Table meta = TEST_UTIL.getConnection().getTable(TableName.META_TABLE_NAME); int rows = 0; Scan scan = new Scan(); scan.addColumn(HConstants.CATALOG_FAMILY, HConstants.REGIONINFO_QUALIFIER); http://git-wip-us.apache.org/repos/asf/hbase/blob/9246af8d/hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestRestartCluster.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestRestartCluster.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestRestartCluster.java index ad22fe9..b51f7c7 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestRestartCluster.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestRestartCluster.java @@ -35,8 +35,6 @@ import org.apache.hadoop.hbase.MiniHBaseCluster; import org.apache.hadoop.hbase.ServerName; import org.apache.hadoop.hbase.TableExistsException; import org.apache.hadoop.hbase.TableName; -import org.apache.hadoop.hbase.client.Connection; -import org.apache.hadoop.hbase.client.ConnectionFactory; import org.apache.hadoop.hbase.client.MetaScanner; import org.apache.hadoop.hbase.testclassification.LargeTests; import org.apache.hadoop.hbase.testclassification.MasterTests; @@ -66,8 +64,6 @@ public class TestRestartCluster { @Test (timeout=300000) public void testClusterRestart() throws Exception { UTIL.startMiniCluster(3); - Connection connection = UTIL.getConnection(); - while (!UTIL.getMiniHBaseCluster().getMaster().isInitialized()) { Threads.sleep(1); } @@ -80,7 +76,7 @@ public class TestRestartCluster { } List<HRegionInfo> allRegions = - MetaScanner.listAllRegions(UTIL.getConfiguration(), connection, true); + MetaScanner.listAllRegions(UTIL.getConfiguration(), UTIL.getConnection(), true); assertEquals(4, allRegions.size()); LOG.info("\n\nShutting down cluster"); @@ -95,8 +91,8 @@ public class TestRestartCluster { // Need to use a new 'Configuration' so we make a new HConnection. // Otherwise we're reusing an HConnection that has gone stale because // the shutdown of the cluster also called shut of the connection. - allRegions = - MetaScanner.listAllRegions(new Configuration(UTIL.getConfiguration()), connection, true); + allRegions = MetaScanner + .listAllRegions(new Configuration(UTIL.getConfiguration()), UTIL.getConnection(), true); assertEquals(4, allRegions.size()); LOG.info("\n\nWaiting for tables to be available"); for(TableName TABLE: TABLES) { http://git-wip-us.apache.org/repos/asf/hbase/blob/9246af8d/hbase-server/src/test/java/org/apache/hadoop/hbase/master/cleaner/TestSnapshotFromMaster.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/cleaner/TestSnapshotFromMaster.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/cleaner/TestSnapshotFromMaster.java index 9a72e77..a2f08ab 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/cleaner/TestSnapshotFromMaster.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/cleaner/TestSnapshotFromMaster.java @@ -290,10 +290,10 @@ public class TestSnapshotFromMaster { htd.setCompactionEnabled(false); UTIL.createTable(htd, new byte[][] { TEST_FAM }, UTIL.getConfiguration()); // load the table (creates 4 hfiles) - UTIL.loadTable(new HTable(UTIL.getConfiguration(), TABLE_NAME), TEST_FAM); + UTIL.loadTable(UTIL.getConnection().getTable(TABLE_NAME), TEST_FAM); UTIL.flush(TABLE_NAME); // Put some more data into the table so for sure we get more storefiles. - UTIL.loadTable(new HTable(UTIL.getConfiguration(), TABLE_NAME), TEST_FAM); + UTIL.loadTable((HTable) UTIL.getConnection().getTable(TABLE_NAME), TEST_FAM); // disable the table so we can take a snapshot admin.disableTable(TABLE_NAME); http://git-wip-us.apache.org/repos/asf/hbase/blob/9246af8d/hbase-server/src/test/java/org/apache/hadoop/hbase/master/handler/TestTableDeleteFamilyHandler.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/handler/TestTableDeleteFamilyHandler.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/handler/TestTableDeleteFamilyHandler.java index ce6abda..5b2f4f6 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/handler/TestTableDeleteFamilyHandler.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/handler/TestTableDeleteFamilyHandler.java @@ -36,6 +36,7 @@ import org.apache.hadoop.hbase.HColumnDescriptor; import org.apache.hadoop.hbase.HTableDescriptor; import org.apache.hadoop.hbase.client.Admin; import org.apache.hadoop.hbase.client.HTable; +import org.apache.hadoop.hbase.client.Table; import org.apache.hadoop.hbase.testclassification.LargeTests; import org.apache.hadoop.hbase.testclassification.MasterTests; import org.apache.hadoop.hbase.util.Bytes; @@ -69,7 +70,7 @@ public class TestTableDeleteFamilyHandler { // Create a table of three families. This will assign a region. TEST_UTIL.createTable(TABLENAME, FAMILIES); - HTable t = new HTable(TEST_UTIL.getConfiguration(), TABLENAME); + Table t = TEST_UTIL.getConnection().getTable(TABLENAME); while(TEST_UTIL.getMiniHBaseCluster().getMaster().getAssignmentManager() .getRegionStates().getRegionsInTransition().size() > 0) { Thread.sleep(100); http://git-wip-us.apache.org/repos/asf/hbase/blob/9246af8d/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestCompactionState.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestCompactionState.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestCompactionState.java index 3ba01bc..3f5f905 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestCompactionState.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestCompactionState.java @@ -89,7 +89,7 @@ public class TestCompactionState { Table ht = null; try { ht = TEST_UTIL.createTable(table, family); - HBaseAdmin admin = new HBaseAdmin(TEST_UTIL.getConfiguration()); + HBaseAdmin admin = TEST_UTIL.getHBaseAdmin(); try { admin.compact(table, fakecf); } catch (IOException ioe) { @@ -137,7 +137,7 @@ public class TestCompactionState { int countBefore = countStoreFilesInFamilies(regions, families); int countBeforeSingleFamily = countStoreFilesInFamily(regions, family); assertTrue(countBefore > 0); // there should be some data files - HBaseAdmin admin = new HBaseAdmin(TEST_UTIL.getConfiguration()); + HBaseAdmin admin = TEST_UTIL.getHBaseAdmin(); if (expectedState == CompactionState.MINOR) { if (singleFamily) { admin.compact(table.getName(), family); @@ -154,10 +154,10 @@ public class TestCompactionState { long curt = System.currentTimeMillis(); long waitTime = 5000; long endt = curt + waitTime; - CompactionState state = admin.getCompactionState(table.getName()); + CompactionState state = admin.getCompactionState(table); while (state == CompactionState.NONE && curt < endt) { Thread.sleep(10); - state = admin.getCompactionState(table.getName()); + state = admin.getCompactionState(table); curt = System.currentTimeMillis(); } // Now, should have the right compaction state, @@ -169,10 +169,10 @@ public class TestCompactionState { } } else { // Wait until the compaction is done - state = admin.getCompactionState(table.getName()); + state = admin.getCompactionState(table); while (state != CompactionState.NONE && curt < endt) { Thread.sleep(10); - state = admin.getCompactionState(table.getName()); + state = admin.getCompactionState(table); } // Now, compaction should be done. assertEquals(CompactionState.NONE, state); http://git-wip-us.apache.org/repos/asf/hbase/blob/9246af8d/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestEncryptionKeyRotation.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestEncryptionKeyRotation.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestEncryptionKeyRotation.java index 941f6d2..af86b4e 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestEncryptionKeyRotation.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestEncryptionKeyRotation.java @@ -208,7 +208,7 @@ public class TestEncryptionKeyRotation { TEST_UTIL.getHBaseAdmin().createTable(htd); TEST_UTIL.waitTableAvailable(htd.getName(), 5000); // Create a store file - Table table = new HTable(conf, htd.getTableName()); + Table table = TEST_UTIL.getConnection().getTable(htd.getTableName()); try { table.put(new Put(Bytes.toBytes("testrow")) .add(hcd.getName(), Bytes.toBytes("q"), Bytes.toBytes("value"))); http://git-wip-us.apache.org/repos/asf/hbase/blob/9246af8d/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestEncryptionRandomKeying.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestEncryptionRandomKeying.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestEncryptionRandomKeying.java index efae472..29a58a6 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestEncryptionRandomKeying.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestEncryptionRandomKeying.java @@ -103,7 +103,7 @@ public class TestEncryptionRandomKeying { TEST_UTIL.waitTableAvailable(htd.getName(), 5000); // Create a store file - Table table = new HTable(conf, htd.getTableName()); + Table table = TEST_UTIL.getConnection().getTable(htd.getTableName()); try { table.put(new Put(Bytes.toBytes("testrow")) .add(hcd.getName(), Bytes.toBytes("q"), Bytes.toBytes("value"))); http://git-wip-us.apache.org/repos/asf/hbase/blob/9246af8d/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestEndToEndSplitTransaction.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestEndToEndSplitTransaction.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestEndToEndSplitTransaction.java index 8b7edeb..4a6ba41 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestEndToEndSplitTransaction.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestEndToEndSplitTransaction.java @@ -331,7 +331,7 @@ public class TestEndToEndSplitTransaction { HTable table = null; try { //HTable.getStartEndKeys() - table = new HTable(conf, tableName); + table = (HTable) connection.getTable(tableName); Pair<byte[][], byte[][]> keys = table.getStartEndKeys(); verifyStartEndKeys(keys); @@ -433,7 +433,8 @@ public class TestEndToEndSplitTransaction { long start = System.currentTimeMillis(); log("blocking until region is split:" + Bytes.toStringBinary(regionName)); HRegionInfo daughterA = null, daughterB = null; - Table metaTable = new HTable(conf, TableName.META_TABLE_NAME); + Connection connection = ConnectionFactory.createConnection(conf); + Table metaTable = connection.getTable(TableName.META_TABLE_NAME); try { Result result = null; @@ -476,6 +477,7 @@ public class TestEndToEndSplitTransaction { } } finally { IOUtils.closeQuietly(metaTable); + IOUtils.closeQuietly(connection); } } @@ -505,7 +507,8 @@ public class TestEndToEndSplitTransaction { throws IOException, InterruptedException { log("blocking until region is opened for reading:" + hri.getRegionNameAsString()); long start = System.currentTimeMillis(); - Table table = new HTable(conf, hri.getTable()); + Connection connection = ConnectionFactory.createConnection(conf); + Table table = connection.getTable(hri.getTable()); try { byte [] row = hri.getStartKey(); @@ -523,7 +526,7 @@ public class TestEndToEndSplitTransaction { } } finally { IOUtils.closeQuietly(table); + IOUtils.closeQuietly(connection); } } } - http://git-wip-us.apache.org/repos/asf/hbase/blob/9246af8d/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestFSErrorsExposed.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestFSErrorsExposed.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestFSErrorsExposed.java index 9a2cc82..c9608c9 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestFSErrorsExposed.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestFSErrorsExposed.java @@ -42,6 +42,7 @@ import org.apache.hadoop.hbase.HColumnDescriptor; import org.apache.hadoop.hbase.HConstants; import org.apache.hadoop.hbase.HTableDescriptor; import org.apache.hadoop.hbase.KeyValue; +import org.apache.hadoop.hbase.client.Table; import org.apache.hadoop.hbase.testclassification.MediumTests; import org.apache.hadoop.hbase.testclassification.RegionServerTests; import org.apache.hadoop.hbase.TableName; @@ -190,7 +191,7 @@ public class TestFSErrorsExposed { TableName tableName = TableName.valueOf("table"); byte[] fam = Bytes.toBytes("fam"); - Admin admin = new HBaseAdmin(util.getConfiguration()); + Admin admin = util.getHBaseAdmin(); HTableDescriptor desc = new HTableDescriptor(tableName); desc.addFamily(new HColumnDescriptor(fam) .setMaxVersions(1) @@ -201,7 +202,7 @@ public class TestFSErrorsExposed { util.getConfiguration().setInt(HConstants.HBASE_CLIENT_RETRIES_NUMBER, 1); // Make a new Configuration so it makes a new connection that has the // above configuration on it; else we use the old one w/ 10 as default. - HTable table = new HTable(new Configuration(util.getConfiguration()), tableName); + Table table = util.getConnection().getTable(tableName); // Load some data util.loadTable(table, fam, false); http://git-wip-us.apache.org/repos/asf/hbase/blob/9246af8d/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestHRegionOnCluster.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestHRegionOnCluster.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestHRegionOnCluster.java index ce2869b..9b8dabf 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestHRegionOnCluster.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestHRegionOnCluster.java @@ -76,7 +76,7 @@ public class TestHRegionOnCluster { // Put data: r1->v1 Log.info("Loading r1 to v1 into " + TABLENAME); - HTable table = new HTable(TEST_UTIL.getConfiguration(), TABLENAME); + HTable table = (HTable) TEST_UTIL.getConnection().getTable(TABLENAME); putDataAndVerify(table, "r1", FAMILY, "v1", 1); TEST_UTIL.waitUntilAllRegionsAssigned(table.getName()); http://git-wip-us.apache.org/repos/asf/hbase/blob/9246af8d/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestHRegionServerBulkLoad.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestHRegionServerBulkLoad.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestHRegionServerBulkLoad.java index 0e94e68..d2bfb52 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestHRegionServerBulkLoad.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestHRegionServerBulkLoad.java @@ -44,6 +44,7 @@ import org.apache.hadoop.hbase.client.ResultScanner; import org.apache.hadoop.hbase.client.RpcRetryingCaller; import org.apache.hadoop.hbase.client.RpcRetryingCallerFactory; import org.apache.hadoop.hbase.client.Scan; +import org.apache.hadoop.hbase.client.Table; import org.apache.hadoop.hbase.io.compress.Compression; import org.apache.hadoop.hbase.io.compress.Compression.Algorithm; import org.apache.hadoop.hbase.io.hfile.CacheConfig; @@ -206,7 +207,7 @@ public class TestHRegionServerBulkLoad { */ public static class AtomicScanReader extends RepeatingTestThread { byte targetFamilies[][]; - HTable table; + Table table; AtomicLong numScans = new AtomicLong(); AtomicLong numRowsScanned = new AtomicLong(); TableName TABLE_NAME; @@ -216,7 +217,7 @@ public class TestHRegionServerBulkLoad { super(ctx); this.TABLE_NAME = TABLE_NAME; this.targetFamilies = targetFamilies; - table = new HTable(conf, TABLE_NAME); + table = UTIL.getConnection().getTable(TABLE_NAME); } public void doAnAction() throws Exception { http://git-wip-us.apache.org/repos/asf/hbase/blob/9246af8d/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestJoinedScanners.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestJoinedScanners.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestJoinedScanners.java index b8e6382..d30b5f8 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestJoinedScanners.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestJoinedScanners.java @@ -101,7 +101,7 @@ public class TestJoinedScanners { desc.addFamily(hcd); } htu.getHBaseAdmin().createTable(desc); - Table ht = new HTable(htu.getConfiguration(), tableName); + Table ht = htu.getConnection().getTable(tableName); long rows_to_insert = 1000; int insert_batch = 20; http://git-wip-us.apache.org/repos/asf/hbase/blob/9246af8d/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestRegionReplicas.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestRegionReplicas.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestRegionReplicas.java index cbe79fe..538e47e 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestRegionReplicas.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestRegionReplicas.java @@ -133,7 +133,7 @@ public class TestRegionReplicas { openRegion(HTU, getRS(), hriSecondary); Table meta = null; try { - meta = new HTable(HTU.getConfiguration(), TableName.META_TABLE_NAME); + meta = HTU.getConnection().getTable(TableName.META_TABLE_NAME); TestMetaTableAccessor.assertMetaLocation(meta, hriPrimary.getRegionName() , getRS().getServerName(), -1, 1, false); } finally {
