http://git-wip-us.apache.org/repos/asf/hive/blob/f7dea106/hcatalog/webhcat/java-client/src/test/java/org/apache/hive/hcatalog/api/repl/commands/TestCommands.java ---------------------------------------------------------------------- diff --git a/hcatalog/webhcat/java-client/src/test/java/org/apache/hive/hcatalog/api/repl/commands/TestCommands.java b/hcatalog/webhcat/java-client/src/test/java/org/apache/hive/hcatalog/api/repl/commands/TestCommands.java index 6ae3d92..0417750 100644 --- a/hcatalog/webhcat/java-client/src/test/java/org/apache/hive/hcatalog/api/repl/commands/TestCommands.java +++ b/hcatalog/webhcat/java-client/src/test/java/org/apache/hive/hcatalog/api/repl/commands/TestCommands.java @@ -25,7 +25,6 @@ import org.apache.hadoop.fs.Path; import org.apache.hadoop.hive.cli.CliSessionState; import org.apache.hadoop.hive.conf.HiveConf; import org.apache.hadoop.hive.metastore.api.MetaException; -import org.apache.hadoop.hive.ql.CommandNeedRetryException; import org.apache.hadoop.hive.ql.DriverFactory; import org.apache.hadoop.hive.ql.IDriver; import org.apache.hadoop.hive.ql.processors.CommandProcessorResponse; @@ -98,7 +97,7 @@ public class TestCommands { } @Test - public void testDropDatabaseCommand() throws HCatException, CommandNeedRetryException { + public void testDropDatabaseCommand() throws HCatException { String dbName = "cmd_testdb"; int evid = 999; Command testCmd = new DropDatabaseCommand(dbName, evid); @@ -130,7 +129,7 @@ public class TestCommands { } @Test - public void testDropTableCommand() throws HCatException, CommandNeedRetryException { + public void testDropTableCommand() throws HCatException { String dbName = "cmd_testdb"; String tableName = "cmd_testtable"; int evid = 789; @@ -210,7 +209,7 @@ public class TestCommands { } @Test - public void testDropPartitionCommand() throws HCatException, CommandNeedRetryException, MetaException { + public void testDropPartitionCommand() throws HCatException, MetaException { String dbName = "cmd_testdb"; String tableName = "cmd_testtable"; int evid = 789; @@ -302,7 +301,7 @@ public class TestCommands { } @Test - public void testDropTableCommand2() throws HCatException, CommandNeedRetryException, MetaException { + public void testDropTableCommand2() throws HCatException, MetaException { // Secondary DropTableCommand test for testing repl-drop-tables' effect on partitions inside a partitioned table // when there exist partitions inside the table which are older than the drop event. // Our goal is this : Create a table t, with repl.last.id=157, say. @@ -373,7 +372,7 @@ public class TestCommands { @Test - public void testBasicReplEximCommands() throws IOException, CommandNeedRetryException { + public void testBasicReplEximCommands() throws IOException { // repl export, has repl.last.id and repl.scope=all in it // import repl dump, table has repl.last.id on it (will likely be 0) int evid = 111; @@ -454,7 +453,7 @@ public class TestCommands { } @Test - public void testMetadataReplEximCommands() throws IOException, CommandNeedRetryException { + public void testMetadataReplEximCommands() throws IOException { // repl metadata export, has repl.last.id and repl.scope=metadata // import repl metadata dump, table metadata changed, allows override, has repl.last.id int evid = 222; @@ -534,7 +533,7 @@ public class TestCommands { @Test - public void testNoopReplEximCommands() throws CommandNeedRetryException, IOException { + public void testNoopReplEximCommands() throws Exception { // repl noop export on non-existant table, has repl.noop, does not error // import repl noop dump, no error
http://git-wip-us.apache.org/repos/asf/hive/blob/f7dea106/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/TestDDLWithRemoteMetastoreSecondNamenode.java ---------------------------------------------------------------------- diff --git a/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/TestDDLWithRemoteMetastoreSecondNamenode.java b/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/TestDDLWithRemoteMetastoreSecondNamenode.java index cde728e..63a7313 100644 --- a/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/TestDDLWithRemoteMetastoreSecondNamenode.java +++ b/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/TestDDLWithRemoteMetastoreSecondNamenode.java @@ -18,7 +18,6 @@ package org.apache.hadoop.hive.ql; import java.io.IOException; import java.net.URI; -import java.net.URISyntaxException; import java.util.HashMap; import junit.framework.JUnit4TestAdapter; import junit.framework.TestCase; @@ -31,7 +30,6 @@ import org.apache.hadoop.hive.conf.HiveConf; import org.apache.hadoop.hive.metastore.MetaStoreTestUtils; import org.apache.hadoop.hive.metastore.api.Database; import org.apache.hadoop.hive.metastore.api.Index; -import org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge; import org.apache.hadoop.hive.ql.exec.mr.ExecDriver; import org.apache.hadoop.hive.ql.metadata.*; import org.apache.hadoop.hive.ql.processors.CommandProcessorResponse; @@ -143,7 +141,7 @@ public class TestDDLWithRemoteMetastoreSecondNamenode extends TestCase { } } - private void cleanup() throws CommandNeedRetryException { + private void cleanup() throws Exception { String[] srcidx = {Index1Name, Index2Name}; for (String src : srcidx) { driver.run("DROP INDEX IF EXISTS " + src + " ON " + Table1Name); @@ -159,7 +157,7 @@ public class TestDDLWithRemoteMetastoreSecondNamenode extends TestCase { } } - private void executeQuery(String query) throws CommandNeedRetryException { + private void executeQuery(String query) throws Exception { CommandProcessorResponse result = driver.run(query); assertNotNull("driver.run() was expected to return result for query: " + query, result); assertEquals("Execution of (" + query + ") failed with exit status: " @@ -173,7 +171,7 @@ public class TestDDLWithRemoteMetastoreSecondNamenode extends TestCase { } private void addPartitionAndCheck(Table table, String column, - String value, String location) throws CommandNeedRetryException, HiveException { + String value, String location) throws Exception { executeQuery("ALTER TABLE " + table.getTableName() + " ADD PARTITION (" + column + "='" + value + "')" + buildLocationClause(location)); @@ -197,8 +195,7 @@ public class TestDDLWithRemoteMetastoreSecondNamenode extends TestCase { } } - private void alterPartitionAndCheck(Table table, String column, - String value, String location) throws CommandNeedRetryException, HiveException { + private void alterPartitionAndCheck(Table table, String column, String value, String location) throws Exception { assertNotNull(location); executeQuery("ALTER TABLE " + table.getTableName() + " PARTITION (" + column + "='" + value + "')" + @@ -219,12 +216,11 @@ public class TestDDLWithRemoteMetastoreSecondNamenode extends TestCase { } private Table createTableAndCheck(String tableName, String tableLocation) - throws CommandNeedRetryException, HiveException, URISyntaxException { + throws Exception { return createTableAndCheck(null, tableName, tableLocation); } - private Table createTableAndCheck(Table baseTable, String tableName, String tableLocation) - throws CommandNeedRetryException, HiveException, URISyntaxException { + private Table createTableAndCheck(Table baseTable, String tableName, String tableLocation) throws Exception { executeQuery("CREATE TABLE " + tableName + (baseTable == null ? " (col1 string, col2 string) PARTITIONED BY (p string) " : " LIKE " + baseTable.getTableName()) @@ -244,8 +240,7 @@ public class TestDDLWithRemoteMetastoreSecondNamenode extends TestCase { return table; } - private void createIndexAndCheck(Table table, String indexName, String indexLocation) - throws CommandNeedRetryException, HiveException, URISyntaxException { + private void createIndexAndCheck(Table table, String indexName, String indexLocation) throws Exception { executeQuery("CREATE INDEX " + indexName + " ON TABLE " + table.getTableName() + " (col1) AS 'COMPACT' WITH DEFERRED REBUILD " + buildLocationClause(indexLocation)); @@ -263,8 +258,7 @@ public class TestDDLWithRemoteMetastoreSecondNamenode extends TestCase { } } - private void createDatabaseAndCheck(String databaseName, String databaseLocation) - throws CommandNeedRetryException, HiveException, URISyntaxException { + private void createDatabaseAndCheck(String databaseName, String databaseLocation) throws Exception { executeQuery("CREATE DATABASE " + databaseName + buildLocationClause(databaseLocation)); Database database = db.getDatabase(databaseName); assertNotNull("Database object is expected for " + databaseName , database); http://git-wip-us.apache.org/repos/asf/hive/blob/f7dea106/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/parse/TestReplicationScenarios.java ---------------------------------------------------------------------- diff --git a/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/parse/TestReplicationScenarios.java b/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/parse/TestReplicationScenarios.java index d1d2d1f..d763666 100644 --- a/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/parse/TestReplicationScenarios.java +++ b/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/parse/TestReplicationScenarios.java @@ -50,8 +50,6 @@ import org.apache.hadoop.hive.metastore.messaging.event.filters.AndFilter; import org.apache.hadoop.hive.metastore.messaging.event.filters.DatabaseAndTableFilter; import org.apache.hadoop.hive.metastore.messaging.event.filters.EventBoundaryFilter; import org.apache.hadoop.hive.metastore.messaging.event.filters.MessageFormatFilter; -import org.apache.hadoop.hive.metastore.utils.MetaStoreUtils; -import org.apache.hadoop.hive.ql.CommandNeedRetryException; import org.apache.hadoop.hive.ql.DriverFactory; import org.apache.hadoop.hive.ql.IDriver; import org.apache.hadoop.hive.ql.exec.repl.ReplDumpWork; @@ -80,10 +78,8 @@ import java.io.IOException; import java.util.ArrayList; import java.util.Arrays; import java.util.List; -import java.util.Map; - -import static junit.framework.Assert.assertFalse; -import static junit.framework.Assert.assertTrue; +import static org.junit.Assert.assertFalse; +import static org.junit.Assert.assertTrue; import static org.junit.Assert.assertEquals; import static org.junit.Assert.assertNotNull; import static org.junit.Assert.assertNull; @@ -587,22 +583,19 @@ public class TestReplicationScenarios { // getTable is invoked after fetching the table names injectionPathCalled = true; Thread t = new Thread(new Runnable() { + @Override public void run() { - try { - LOG.info("Entered new thread"); - IDriver driver2 = DriverFactory.newDriver(hconf); - SessionState.start(new CliSessionState(hconf)); - CommandProcessorResponse ret = driver2.run("ALTER TABLE " + dbName + ".ptned PARTITION (b=1) RENAME TO PARTITION (b=10)"); - success = (ret.getException() == null); - assertFalse(success); - ret = driver2.run("ALTER TABLE " + dbName + ".ptned RENAME TO " + dbName + ".ptned_renamed"); - success = (ret.getException() == null); - assertFalse(success); - LOG.info("Exit new thread success - {}", success); - } catch (CommandNeedRetryException e) { - LOG.info("Hit Exception {} from new thread", e.getMessage()); - throw new RuntimeException(e); - } + LOG.info("Entered new thread"); + IDriver driver2 = DriverFactory.newDriver(hconf); + SessionState.start(new CliSessionState(hconf)); + CommandProcessorResponse ret = + driver2.run("ALTER TABLE " + dbName + ".ptned PARTITION (b=1) RENAME TO PARTITION (b=10)"); + success = (ret.getException() == null); + assertFalse(success); + ret = driver2.run("ALTER TABLE " + dbName + ".ptned RENAME TO " + dbName + ".ptned_renamed"); + success = (ret.getException() == null); + assertFalse(success); + LOG.info("Exit new thread success - {}", success); } }); t.start(); @@ -662,19 +655,15 @@ public class TestReplicationScenarios { // getTable is invoked after fetching the table names injectionPathCalled = true; Thread t = new Thread(new Runnable() { + @Override public void run() { - try { - LOG.info("Entered new thread"); - IDriver driver2 = DriverFactory.newDriver(hconf); - SessionState.start(new CliSessionState(hconf)); - CommandProcessorResponse ret = driver2.run("DROP TABLE " + dbName + ".ptned"); - success = (ret.getException() == null); - assertTrue(success); - LOG.info("Exit new thread success - {}", success); - } catch (CommandNeedRetryException e) { - LOG.info("Hit Exception {} from new thread", e.getMessage()); - throw new RuntimeException(e); - } + LOG.info("Entered new thread"); + IDriver driver2 = DriverFactory.newDriver(hconf); + SessionState.start(new CliSessionState(hconf)); + CommandProcessorResponse ret = driver2.run("DROP TABLE " + dbName + ".ptned"); + success = (ret.getException() == null); + assertTrue(success); + LOG.info("Exit new thread success - {}", success); } }); t.start(); @@ -3124,7 +3113,7 @@ public class TestReplicationScenarios { List<SQLNotNullConstraint> nns = metaStoreClientMirror.getNotNullConstraints(new NotNullConstraintsRequest(dbName+ "_dupe" , "tbl6")); assertEquals(nns.size(), 1); nnName = nns.get(0).getNn_name(); - + } catch (TException te) { assertNull(te); } @@ -3616,12 +3605,7 @@ public class TestReplicationScenarios { private String getResult(int rowNum, int colNum, boolean reuse, IDriver myDriver) throws IOException { if (!reuse) { lastResults = new ArrayList<String>(); - try { - myDriver.getResults(lastResults); - } catch (CommandNeedRetryException e) { - e.printStackTrace(); - throw new RuntimeException(e); - } + myDriver.getResults(lastResults); } // Split around the 'tab' character return (lastResults.get(rowNum).split("\\t"))[colNum]; @@ -3646,12 +3630,7 @@ public class TestReplicationScenarios { private List<String> getOutput(IDriver myDriver) throws IOException { List<String> results = new ArrayList<>(); - try { - myDriver.getResults(results); - } catch (CommandNeedRetryException e) { - LOG.warn(e.getMessage(),e); - throw new RuntimeException(e); - } + myDriver.getResults(results); return results; } @@ -3772,19 +3751,10 @@ public class TestReplicationScenarios { private static boolean run(String cmd, boolean errorOnFail, IDriver myDriver) throws RuntimeException { boolean success = false; - try { - CommandProcessorResponse ret = myDriver.run(cmd); - success = ((ret.getException() == null) && (ret.getErrorMessage() == null)); - if (!success){ - LOG.warn("Error {} : {} running [{}].", ret.getErrorCode(), ret.getErrorMessage(), cmd); - } - } catch (CommandNeedRetryException e) { - if (errorOnFail){ - throw new RuntimeException(e); - } else { - LOG.warn(e.getMessage(),e); - // do nothing else - } + CommandProcessorResponse ret = myDriver.run(cmd); + success = ((ret.getException() == null) && (ret.getErrorMessage() == null)); + if (!success) { + LOG.warn("Error {} : {} running [{}].", ret.getErrorCode(), ret.getErrorMessage(), cmd); } return success; } http://git-wip-us.apache.org/repos/asf/hive/blob/f7dea106/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/parse/WarehouseInstance.java ---------------------------------------------------------------------- diff --git a/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/parse/WarehouseInstance.java b/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/parse/WarehouseInstance.java index dd6fa42..33e5157 100644 --- a/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/parse/WarehouseInstance.java +++ b/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/parse/WarehouseInstance.java @@ -30,7 +30,6 @@ import org.apache.hadoop.hive.common.FileUtils; import org.apache.hadoop.hive.conf.HiveConf; import org.apache.hadoop.hive.metastore.HiveMetaStoreClient; import org.apache.hadoop.hive.metastore.MetaStoreTestUtils; -import org.apache.hadoop.hive.ql.CommandNeedRetryException; import org.apache.hadoop.hive.ql.DriverFactory; import org.apache.hadoop.hive.ql.IDriver; import org.apache.hadoop.hive.ql.exec.repl.ReplDumpWork; @@ -160,12 +159,7 @@ class WarehouseInstance implements Closeable { private String row0Result(int colNum, boolean reuse) throws IOException { if (!reuse) { lastResults = new ArrayList<>(); - try { - driver.getResults(lastResults); - } catch (CommandNeedRetryException e) { - e.printStackTrace(); - throw new RuntimeException(e); - } + driver.getResults(lastResults); } // Split around the 'tab' character return (lastResults.get(0).split("\\t"))[colNum]; @@ -265,12 +259,7 @@ class WarehouseInstance implements Closeable { List<String> getOutput() throws IOException { List<String> results = new ArrayList<>(); - try { - driver.getResults(results); - } catch (CommandNeedRetryException e) { - logger.warn(e.getMessage(), e); - throw new RuntimeException(e); - } + driver.getResults(results); return results; } http://git-wip-us.apache.org/repos/asf/hive/blob/f7dea106/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/security/authorization/plugin/TestHiveAuthorizerCheckInvocation.java ---------------------------------------------------------------------- diff --git a/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/security/authorization/plugin/TestHiveAuthorizerCheckInvocation.java b/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/security/authorization/plugin/TestHiveAuthorizerCheckInvocation.java index 6f2405c..bc2a34a 100644 --- a/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/security/authorization/plugin/TestHiveAuthorizerCheckInvocation.java +++ b/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/security/authorization/plugin/TestHiveAuthorizerCheckInvocation.java @@ -37,7 +37,6 @@ import org.slf4j.Logger; import org.slf4j.LoggerFactory; import org.apache.hadoop.hive.conf.HiveConf; import org.apache.hadoop.hive.conf.HiveConf.ConfVars; -import org.apache.hadoop.hive.ql.CommandNeedRetryException; import org.apache.hadoop.hive.ql.DriverFactory; import org.apache.hadoop.hive.ql.IDriver; import org.apache.hadoop.hive.ql.lockmgr.DbTxnManager; @@ -114,7 +113,7 @@ public class TestHiveAuthorizerCheckInvocation { "stored as orc TBLPROPERTIES ('transactional'='true')"); } - private static void runCmd(String cmd) throws CommandNeedRetryException { + private static void runCmd(String cmd) throws Exception { CommandProcessorResponse resp = driver.run(cmd); assertEquals(0, resp.getResponseCode()); } @@ -131,8 +130,7 @@ public class TestHiveAuthorizerCheckInvocation { } @Test - public void testInputSomeColumnsUsed() throws HiveAuthzPluginException, HiveAccessControlException, - CommandNeedRetryException { + public void testInputSomeColumnsUsed() throws Exception { reset(mockedAuthorizer); int status = driver.compile("select i from " + tableName @@ -148,8 +146,7 @@ public class TestHiveAuthorizerCheckInvocation { } @Test - public void testInputSomeColumnsUsedView() throws HiveAuthzPluginException, HiveAccessControlException, - CommandNeedRetryException { + public void testInputSomeColumnsUsedView() throws Exception { reset(mockedAuthorizer); int status = driver.compile("select i from " + viewName @@ -165,15 +162,14 @@ public class TestHiveAuthorizerCheckInvocation { } @Test - public void testInputSomeColumnsUsedJoin() throws HiveAuthzPluginException, HiveAccessControlException, - CommandNeedRetryException { - + public void testInputSomeColumnsUsedJoin() throws Exception { + reset(mockedAuthorizer); int status = driver.compile("select " + viewName + ".i, " + tableName + ".city from " + viewName + " join " + tableName + " on " + viewName + ".city = " + tableName + ".city where " + tableName + ".k = 'X'"); assertEquals(0, status); - + List<HivePrivilegeObject> inputs = getHivePrivilegeObjectInputs().getLeft(); Collections.sort(inputs); assertEquals(inputs.size(), 2); @@ -194,8 +190,7 @@ public class TestHiveAuthorizerCheckInvocation { } @Test - public void testInputAllColumnsUsed() throws HiveAuthzPluginException, HiveAccessControlException, - CommandNeedRetryException { + public void testInputAllColumnsUsed() throws Exception { reset(mockedAuthorizer); int status = driver.compile("select * from " + tableName + " order by i"); @@ -210,15 +205,13 @@ public class TestHiveAuthorizerCheckInvocation { } @Test - public void testCreateTableWithDb() throws HiveAuthzPluginException, HiveAccessControlException, - CommandNeedRetryException { + public void testCreateTableWithDb() throws Exception { final String newTable = "ctTableWithDb"; checkCreateViewOrTableWithDb(newTable, "create table " + dbName + "." + newTable + "(i int)"); } @Test - public void testCreateViewWithDb() throws HiveAuthzPluginException, HiveAccessControlException, - CommandNeedRetryException { + public void testCreateViewWithDb() throws Exception { final String newTable = "ctViewWithDb"; checkCreateViewOrTableWithDb(newTable, "create table " + dbName + "." + newTable + "(i int)"); } @@ -251,8 +244,7 @@ public class TestHiveAuthorizerCheckInvocation { } @Test - public void testInputNoColumnsUsed() throws HiveAuthzPluginException, HiveAccessControlException, - CommandNeedRetryException { + public void testInputNoColumnsUsed() throws Exception { reset(mockedAuthorizer); int status = driver.compile("describe " + tableName); @@ -265,8 +257,7 @@ public class TestHiveAuthorizerCheckInvocation { } @Test - public void testPermFunction() throws HiveAuthzPluginException, HiveAccessControlException, - CommandNeedRetryException { + public void testPermFunction() throws Exception { reset(mockedAuthorizer); final String funcName = "testauthfunc1"; @@ -296,8 +287,7 @@ public class TestHiveAuthorizerCheckInvocation { } @Test - public void testTempFunction() throws HiveAuthzPluginException, HiveAccessControlException, - CommandNeedRetryException { + public void testTempFunction() throws Exception { reset(mockedAuthorizer); final String funcName = "testAuthFunc2"; @@ -313,8 +303,7 @@ public class TestHiveAuthorizerCheckInvocation { } @Test - public void testUpdateSomeColumnsUsed() throws HiveAuthzPluginException, - HiveAccessControlException, CommandNeedRetryException { + public void testUpdateSomeColumnsUsed() throws Exception { reset(mockedAuthorizer); int status = driver.compile("update " + acidTableName + " set i = 5 where j = 3"); assertEquals(0, status); @@ -333,8 +322,7 @@ public class TestHiveAuthorizerCheckInvocation { } @Test - public void testUpdateSomeColumnsUsedExprInSet() throws HiveAuthzPluginException, - HiveAccessControlException, CommandNeedRetryException { + public void testUpdateSomeColumnsUsedExprInSet() throws Exception { reset(mockedAuthorizer); int status = driver.compile("update " + acidTableName + " set i = 5, j = k where j = 3"); assertEquals(0, status); @@ -355,8 +343,7 @@ public class TestHiveAuthorizerCheckInvocation { } @Test - public void testDelete() throws HiveAuthzPluginException, - HiveAccessControlException, CommandNeedRetryException { + public void testDelete() throws Exception { reset(mockedAuthorizer); int status = driver.compile("delete from " + acidTableName + " where j = 3"); assertEquals(0, status); @@ -370,8 +357,7 @@ public class TestHiveAuthorizerCheckInvocation { } @Test - public void testShowTables() throws HiveAuthzPluginException, - HiveAccessControlException, CommandNeedRetryException { + public void testShowTables() throws Exception { reset(mockedAuthorizer); int status = driver.compile("show tables"); assertEquals(0, status); @@ -384,8 +370,7 @@ public class TestHiveAuthorizerCheckInvocation { } @Test - public void testDescDatabase() throws HiveAuthzPluginException, - HiveAccessControlException, CommandNeedRetryException { + public void testDescDatabase() throws Exception { reset(mockedAuthorizer); int status = driver.compile("describe database " + dbName); assertEquals(0, status); @@ -411,8 +396,7 @@ public class TestHiveAuthorizerCheckInvocation { } @Test - public void testReplDump() throws HiveAuthzPluginException, HiveAccessControlException, - CommandNeedRetryException { + public void testReplDump() throws Exception { resetAuthorizer(); int status = driver.compile("repl dump " + dbName); http://git-wip-us.apache.org/repos/asf/hive/blob/f7dea106/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/security/authorization/plugin/TestHiveAuthorizerShowFilters.java ---------------------------------------------------------------------- diff --git a/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/security/authorization/plugin/TestHiveAuthorizerShowFilters.java b/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/security/authorization/plugin/TestHiveAuthorizerShowFilters.java index 2be86f8..8981223 100644 --- a/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/security/authorization/plugin/TestHiveAuthorizerShowFilters.java +++ b/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/security/authorization/plugin/TestHiveAuthorizerShowFilters.java @@ -21,7 +21,6 @@ package org.apache.hadoop.hive.ql.security.authorization.plugin; import static org.junit.Assert.assertEquals; import static org.mockito.Matchers.any; -import java.io.IOException; import java.util.ArrayList; import java.util.Arrays; import java.util.Collections; @@ -30,7 +29,6 @@ import java.util.List; import org.apache.hadoop.hive.UtilsForTest; import org.apache.hadoop.hive.conf.HiveConf; import org.apache.hadoop.hive.conf.HiveConf.ConfVars; -import org.apache.hadoop.hive.ql.CommandNeedRetryException; import org.apache.hadoop.hive.ql.DriverFactory; import org.apache.hadoop.hive.ql.IDriver; import org.apache.hadoop.hive.ql.processors.CommandProcessorResponse; @@ -90,7 +88,7 @@ public class TestHiveAuthorizerShowFilters { return filteredResults; } } - + @Override public HiveAuthorizer createHiveAuthorizer(HiveMetastoreClientFactory metastoreClientFactory, HiveConf conf, HiveAuthenticationProvider authenticator, HiveAuthzSessionContext ctx) { @@ -153,46 +151,40 @@ public class TestHiveAuthorizerShowFilters { } @Test - public void testShowDatabasesAll() throws HiveAuthzPluginException, HiveAccessControlException, - CommandNeedRetryException, IOException { + public void testShowDatabasesAll() throws Exception { runShowDbTest(AllDbs); } @Test - public void testShowDatabasesSelected() throws HiveAuthzPluginException, - HiveAccessControlException, CommandNeedRetryException, IOException { + public void testShowDatabasesSelected() throws Exception { setFilteredResults(HivePrivilegeObjectType.DATABASE, dbName2); runShowDbTest(Arrays.asList(dbName2)); } - private void runShowDbTest(List<String> expectedDbList) throws HiveAuthzPluginException, - HiveAccessControlException, CommandNeedRetryException, IOException { + private void runShowDbTest(List<String> expectedDbList) throws Exception { runCmd("show databases"); verifyAllDb(); assertEquals("filtered result check ", expectedDbList, getSortedResults()); } @Test - public void testShowTablesAll() throws HiveAuthzPluginException, HiveAccessControlException, - CommandNeedRetryException, IOException { + public void testShowTablesAll() throws Exception { runShowTablesTest(AllTables); } @Test - public void testShowTablesSelected() throws HiveAuthzPluginException, HiveAccessControlException, - CommandNeedRetryException, IOException { + public void testShowTablesSelected() throws Exception { setFilteredResults(HivePrivilegeObjectType.TABLE_OR_VIEW, tableName2); runShowTablesTest(Arrays.asList(tableName2)); } - private void runShowTablesTest(List<String> expectedTabs) throws IOException, - CommandNeedRetryException, HiveAuthzPluginException, HiveAccessControlException { + private void runShowTablesTest(List<String> expectedTabs) throws Exception { runCmd("show tables"); verifyAllTables(); assertEquals("filtered result check ", expectedTabs, getSortedResults()); } - private List<String> getSortedResults() throws IOException, CommandNeedRetryException { + private List<String> getSortedResults() throws Exception { List<String> res = new ArrayList<String>(); // set results to be returned driver.getResults(res); @@ -262,7 +254,7 @@ public class TestHiveAuthorizerShowFilters { } } - private static void runCmd(String cmd) throws CommandNeedRetryException { + private static void runCmd(String cmd) throws Exception { CommandProcessorResponse resp = driver.run(cmd); assertEquals(0, resp.getResponseCode()); } http://git-wip-us.apache.org/repos/asf/hive/blob/f7dea106/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/txn/compactor/TestCompactor.java ---------------------------------------------------------------------- diff --git a/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/txn/compactor/TestCompactor.java b/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/txn/compactor/TestCompactor.java index 1305902..a5e6293 100644 --- a/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/txn/compactor/TestCompactor.java +++ b/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/txn/compactor/TestCompactor.java @@ -55,7 +55,6 @@ import org.apache.hadoop.hive.metastore.txn.CompactionInfo; import org.apache.hadoop.hive.metastore.txn.TxnDbUtil; import org.apache.hadoop.hive.metastore.txn.TxnStore; import org.apache.hadoop.hive.metastore.txn.TxnUtils; -import org.apache.hadoop.hive.ql.CommandNeedRetryException; import org.apache.hadoop.hive.ql.DriverFactory; import org.apache.hadoop.hive.ql.IDriver; import org.apache.hadoop.hive.ql.io.AcidInputFormat; @@ -1425,7 +1424,7 @@ public class TestCompactor { * Execute Hive CLI statement * @param cmd arbitrary statement to execute */ - static void executeStatementOnDriver(String cmd, IDriver driver) throws IOException, CommandNeedRetryException { + static void executeStatementOnDriver(String cmd, IDriver driver) throws Exception { LOG.debug("Executing: " + cmd); CommandProcessorResponse cpr = driver.run(cmd); if(cpr.getResponseCode() != 0) { http://git-wip-us.apache.org/repos/asf/hive/blob/f7dea106/itests/util/src/main/java/org/apache/hadoop/hive/ql/QTestUtil.java ---------------------------------------------------------------------- diff --git a/itests/util/src/main/java/org/apache/hadoop/hive/ql/QTestUtil.java b/itests/util/src/main/java/org/apache/hadoop/hive/ql/QTestUtil.java index 2bc33bd..2d0aca0 100644 --- a/itests/util/src/main/java/org/apache/hadoop/hive/ql/QTestUtil.java +++ b/itests/util/src/main/java/org/apache/hadoop/hive/ql/QTestUtil.java @@ -1313,13 +1313,7 @@ public class QTestUtil { } public int execute(String tname) { - try { - return drv.run(qMap.get(tname)).getResponseCode(); - } catch (CommandNeedRetryException e) { - LOG.error("driver failed to run the command: " + tname + " due to the exception: ", e); - e.printStackTrace(); - return -1; - } + return drv.run(qMap.get(tname)).getResponseCode(); } public int executeClient(String tname1, String tname2) { http://git-wip-us.apache.org/repos/asf/hive/blob/f7dea106/ql/src/java/org/apache/hadoop/hive/ql/CommandNeedRetryException.java ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/CommandNeedRetryException.java b/ql/src/java/org/apache/hadoop/hive/ql/CommandNeedRetryException.java deleted file mode 100644 index 1b60005..0000000 --- a/ql/src/java/org/apache/hadoop/hive/ql/CommandNeedRetryException.java +++ /dev/null @@ -1,40 +0,0 @@ -/* - * Licensed to the Apache Software Foundation (ASF) under one - * or more contributor license agreements. See the NOTICE file - * distributed with this work for additional information - * regarding copyright ownership. The ASF licenses this file - * to you under the Apache License, Version 2.0 (the - * "License"); you may not use this file except in compliance - * with the License. You may obtain a copy of the License at - * - * http://www.apache.org/licenses/LICENSE-2.0 - * - * Unless required by applicable law or agreed to in writing, software - * distributed under the License is distributed on an "AS IS" BASIS, - * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. - * See the License for the specific language governing permissions and - * limitations under the License. - */ - -package org.apache.hadoop.hive.ql; - -public class CommandNeedRetryException extends Exception { - - private static final long serialVersionUID = 1L; - - public CommandNeedRetryException() { - super(); - } - - public CommandNeedRetryException(String message) { - super(message); - } - - public CommandNeedRetryException(Throwable cause) { - super(cause); - } - - public CommandNeedRetryException(String message, Throwable cause) { - super(message, cause); - } -} http://git-wip-us.apache.org/repos/asf/hive/blob/f7dea106/ql/src/java/org/apache/hadoop/hive/ql/Context.java ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/Context.java b/ql/src/java/org/apache/hadoop/hive/ql/Context.java index 820fbf0..dba2dbb 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/Context.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/Context.java @@ -100,8 +100,6 @@ public class Context { protected String cboInfo; protected boolean cboSucceeded; protected String cmd = ""; - // number of previous attempts - protected int tryCount = 0; private TokenRewriteStream tokenRewriteStream; // Holds the qualified name to tokenRewriteStream for the views // referenced by the query. This is used to rewrite the view AST @@ -178,6 +176,7 @@ public class Context { DestClausePrefix(String prefix) { this.prefix = prefix; } + @Override public String toString() { return prefix; } @@ -941,14 +940,6 @@ public class Context { this.needLockMgr = needLockMgr; } - public int getTryCount() { - return tryCount; - } - - public void setTryCount(int tryCount) { - this.tryCount = tryCount; - } - public String getCboInfo() { return cboInfo; } http://git-wip-us.apache.org/repos/asf/hive/blob/f7dea106/ql/src/java/org/apache/hadoop/hive/ql/Driver.java ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/Driver.java b/ql/src/java/org/apache/hadoop/hive/ql/Driver.java index 74595b0..2d7e459 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/Driver.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/Driver.java @@ -158,7 +158,6 @@ public class Driver implements IDriver { // A limit on the number of threads that can be launched private int maxthreads; - private int tryCount = Integer.MAX_VALUE; private String userName; @@ -577,7 +576,6 @@ public class Driver implements IDriver { setTriggerContext(queryId); } - ctx.setTryCount(getTryCount()); ctx.setCmd(command); ctx.setHDFSCleanup(true); @@ -1377,19 +1375,16 @@ public class Driver implements IDriver { @Override - public CommandProcessorResponse run(String command) - throws CommandNeedRetryException { + public CommandProcessorResponse run(String command) { return run(command, false); } @Override - public CommandProcessorResponse run() - throws CommandNeedRetryException { + public CommandProcessorResponse run() { return run(null, true); } - public CommandProcessorResponse run(String command, boolean alreadyCompiled) - throws CommandNeedRetryException { + public CommandProcessorResponse run(String command, boolean alreadyCompiled) { try { runInternal(command, alreadyCompiled); @@ -1579,8 +1574,7 @@ public class Driver implements IDriver { return compileLock; } - private void runInternal(String command, boolean alreadyCompiled) - throws CommandNeedRetryException, CommandProcessorResponse { + private void runInternal(String command, boolean alreadyCompiled) throws CommandProcessorResponse { errorMessage = null; SQLState = null; downstreamError = null; @@ -1794,7 +1788,7 @@ public class Driver implements IDriver { return new CommandProcessorResponse(ret, errorMessage, SQLState, downstreamError); } - private void execute() throws CommandNeedRetryException, CommandProcessorResponse { + private void execute() throws CommandProcessorResponse { PerfLogger perfLogger = SessionState.getPerfLogger(); perfLogger.PerfLogBegin(CLASS_NAME, PerfLogger.DRIVER_EXECUTE); @@ -1949,13 +1943,6 @@ public class Driver implements IDriver { checkInterrupted("when checking the execution result.", hookContext, perfLogger); if (exitVal != 0) { - if (tsk.ifRetryCmdWhenFail()) { - driverCxt.shutdown(); - // in case we decided to run everything in local mode, restore the - // the jobtracker setting to its initial value - ctx.restoreOriginalTracker(); - throw new CommandNeedRetryException(); - } Task<? extends Serializable> backupTask = tsk.getAndInitBackupTask(); if (backupTask != null) { setErrorMsgAndDetail(exitVal, result.getTaskError(), tsk); @@ -2057,9 +2044,6 @@ public class Driver implements IDriver { SessionState.get().getHiveHistory().printRowCount(queryId); } releasePlan(plan); - } catch (CommandNeedRetryException e) { - executionError = true; - throw e; } catch (CommandProcessorResponse cpr) { executionError = true; throw cpr; @@ -2276,7 +2260,7 @@ public class Driver implements IDriver { @SuppressWarnings("unchecked") @Override - public boolean getResults(List res) throws IOException, CommandNeedRetryException { + public boolean getResults(List res) throws IOException { if (lDrvState.driverState == DriverState.DESTROYED || lDrvState.driverState == DriverState.CLOSED) { throw new IOException("FAILED: query has been cancelled, closed, or destroyed."); } @@ -2359,15 +2343,6 @@ public class Driver implements IDriver { } } - public int getTryCount() { - return tryCount; - } - - @Override - public void setTryCount(int tryCount) { - this.tryCount = tryCount; - } - // DriverContext could be released in the query and close processes at same // time, which needs to be thread protected. private void releaseDriverContext() { http://git-wip-us.apache.org/repos/asf/hive/blob/f7dea106/ql/src/java/org/apache/hadoop/hive/ql/IDriver.java ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/IDriver.java b/ql/src/java/org/apache/hadoop/hive/ql/IDriver.java index 6280be0..d4494cc 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/IDriver.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/IDriver.java @@ -41,15 +41,13 @@ public interface IDriver extends CommandProcessor { void setOperationId(String guid64); - void setTryCount(int maxValue); - - CommandProcessorResponse run() throws CommandNeedRetryException; + CommandProcessorResponse run(); @Override - CommandProcessorResponse run(String command) throws CommandNeedRetryException; + CommandProcessorResponse run(String command); // create some "cover" to the result? - boolean getResults(List res) throws IOException, CommandNeedRetryException; + boolean getResults(List res) throws IOException; void setMaxRows(int maxRows); http://git-wip-us.apache.org/repos/asf/hive/blob/f7dea106/ql/src/java/org/apache/hadoop/hive/ql/QueryDisplay.java ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/QueryDisplay.java b/ql/src/java/org/apache/hadoop/hive/ql/QueryDisplay.java index cc67b87..9a77c29 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/QueryDisplay.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/QueryDisplay.java @@ -79,7 +79,6 @@ public class QueryDisplay { private StageType taskType; private String name; private boolean requireLock; - private boolean retryIfFail; private String statusMessage; // required for jackson @@ -92,7 +91,6 @@ public class QueryDisplay { taskType = task.getType(); name = task.getName(); requireLock = task.requireLock(); - retryIfFail = task.ifRetryCmdWhenFail(); } @JsonIgnore public synchronized String getStatus() { @@ -147,10 +145,6 @@ public class QueryDisplay { public synchronized boolean isRequireLock() { return requireLock; } - @JsonIgnore - public synchronized boolean isRetryIfFail() { - return retryIfFail; - } public synchronized String getExternalHandle() { return externalHandle; http://git-wip-us.apache.org/repos/asf/hive/blob/f7dea106/ql/src/java/org/apache/hadoop/hive/ql/exec/FetchTask.java ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/exec/FetchTask.java b/ql/src/java/org/apache/hadoop/hive/ql/exec/FetchTask.java index a7dace9..090a188 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/exec/FetchTask.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/exec/FetchTask.java @@ -24,7 +24,6 @@ import java.util.List; import org.slf4j.Logger; import org.slf4j.LoggerFactory; -import org.apache.hadoop.hive.ql.CommandNeedRetryException; import org.apache.hadoop.hive.ql.CompilationOpContext; import org.apache.hadoop.hive.ql.DriverContext; import org.apache.hadoop.hive.ql.QueryPlan; @@ -130,7 +129,7 @@ public class FetchTask extends Task<FetchWork> implements Serializable { this.maxRows = maxRows; } - public boolean fetch(List res) throws IOException, CommandNeedRetryException { + public boolean fetch(List res) throws IOException { sink.reset(res); int rowsRet = work.getLeastNumRows(); if (rowsRet <= 0) { @@ -145,7 +144,7 @@ public class FetchTask extends Task<FetchWork> implements Serializable { while (sink.getNumRows() < rowsRet) { if (!fetch.pushRow()) { if (work.getLeastNumRows() > 0) { - throw new CommandNeedRetryException(); + throw new HiveException("leastNumRows check failed"); } // Closing the operator can sometimes yield more rows (HIVE-11892) @@ -156,8 +155,6 @@ public class FetchTask extends Task<FetchWork> implements Serializable { fetched = true; } return true; - } catch (CommandNeedRetryException e) { - throw e; } catch (IOException e) { throw e; } catch (Exception e) { http://git-wip-us.apache.org/repos/asf/hive/blob/f7dea106/ql/src/java/org/apache/hadoop/hive/ql/exec/Task.java ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/exec/Task.java b/ql/src/java/org/apache/hadoop/hive/ql/exec/Task.java index 7124c89..a8d851f 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/exec/Task.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/exec/Task.java @@ -579,14 +579,6 @@ public abstract class Task<T extends Serializable> implements Serializable, Node return false; } - public boolean ifRetryCmdWhenFail() { - return retryCmdWhenFail; - } - - public void setRetryCmdWhenFail(boolean retryCmdWhenFail) { - this.retryCmdWhenFail = retryCmdWhenFail; - } - public QueryPlan getQueryPlan() { return queryPlan; } http://git-wip-us.apache.org/repos/asf/hive/blob/f7dea106/ql/src/java/org/apache/hadoop/hive/ql/optimizer/GlobalLimitOptimizer.java ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/GlobalLimitOptimizer.java b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/GlobalLimitOptimizer.java index e656eb0..bd0cbab 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/GlobalLimitOptimizer.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/GlobalLimitOptimizer.java @@ -75,7 +75,7 @@ public class GlobalLimitOptimizer extends Transform { // The query only qualifies when there are only one top operator // and there is no transformer or UDTF and no block sampling // is used. - if (ctx.getTryCount() == 0 && topOps.size() == 1 + if (topOps.size() == 1 && !globalLimitCtx.ifHasTransformOrUDTF() && nameToSplitSample.isEmpty()) { http://git-wip-us.apache.org/repos/asf/hive/blob/f7dea106/ql/src/java/org/apache/hadoop/hive/ql/parse/ExplainSemanticAnalyzer.java ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/parse/ExplainSemanticAnalyzer.java b/ql/src/java/org/apache/hadoop/hive/ql/parse/ExplainSemanticAnalyzer.java index f01b497..63b13c8 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/parse/ExplainSemanticAnalyzer.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/parse/ExplainSemanticAnalyzer.java @@ -33,7 +33,6 @@ import org.apache.hadoop.hive.common.FileUtils; import org.apache.hadoop.hive.common.StatsSetupConst; import org.apache.hadoop.hive.conf.HiveConf; import org.apache.hadoop.hive.metastore.api.FieldSchema; -import org.apache.hadoop.hive.ql.CommandNeedRetryException; import org.apache.hadoop.hive.ql.Context; import org.apache.hadoop.hive.ql.Driver; import org.apache.hadoop.hive.ql.QueryState; @@ -147,8 +146,6 @@ public class ExplainSemanticAnalyzer extends BaseSemanticAnalyzer { config.setOpIdToRuntimeNumRows(aggregateStats(config.getExplainRootPath())); } catch (IOException e1) { throw new SemanticException(e1); - } catch (CommandNeedRetryException e) { - throw new SemanticException(e); } ctx.resetOpContext(); ctx.resetStream(); http://git-wip-us.apache.org/repos/asf/hive/blob/f7dea106/ql/src/java/org/apache/hadoop/hive/ql/parse/TaskCompiler.java ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/parse/TaskCompiler.java b/ql/src/java/org/apache/hadoop/hive/ql/parse/TaskCompiler.java index 0c1c4e0..92d29e3 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/parse/TaskCompiler.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/parse/TaskCompiler.java @@ -47,8 +47,6 @@ import org.apache.hadoop.hive.ql.exec.TableScanOperator; import org.apache.hadoop.hive.ql.exec.Task; import org.apache.hadoop.hive.ql.exec.TaskFactory; import org.apache.hadoop.hive.ql.exec.Utilities; -import org.apache.hadoop.hive.ql.exec.mr.ExecDriver; -import org.apache.hadoop.hive.ql.exec.spark.SparkTask; import org.apache.hadoop.hive.ql.hooks.ReadEntity; import org.apache.hadoop.hive.ql.hooks.WriteEntity; import org.apache.hadoop.hive.ql.io.orc.OrcInputFormat; @@ -72,7 +70,6 @@ import org.apache.hadoop.hive.ql.plan.MoveWork; import org.apache.hadoop.hive.ql.plan.PlanUtils; import org.apache.hadoop.hive.ql.plan.StatsWork; import org.apache.hadoop.hive.ql.plan.TableDesc; -import org.apache.hadoop.hive.ql.session.LineageState; import org.apache.hadoop.hive.ql.session.SessionState; import org.apache.hadoop.hive.ql.session.SessionState.LogHelper; import org.apache.hadoop.hive.serde.serdeConstants; @@ -362,14 +359,6 @@ public abstract class TaskCompiler { if (globalLimitCtx.isEnable() && globalLimitCtx.getLastReduceLimitDesc() != null) { LOG.info("set least row check for LimitDesc: " + globalLimitCtx.getGlobalLimit()); globalLimitCtx.getLastReduceLimitDesc().setLeastRows(globalLimitCtx.getGlobalLimit()); - List<ExecDriver> mrTasks = Utilities.getMRTasks(rootTasks); - for (ExecDriver tsk : mrTasks) { - tsk.setRetryCmdWhenFail(true); - } - List<SparkTask> sparkTasks = Utilities.getSparkTasks(rootTasks); - for (SparkTask sparkTask : sparkTasks) { - sparkTask.setRetryCmdWhenFail(true); - } } Interner<TableDesc> interner = Interners.newStrongInterner(); http://git-wip-us.apache.org/repos/asf/hive/blob/f7dea106/ql/src/java/org/apache/hadoop/hive/ql/processors/CommandProcessor.java ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/processors/CommandProcessor.java b/ql/src/java/org/apache/hadoop/hive/ql/processors/CommandProcessor.java index 3624d08..c753264 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/processors/CommandProcessor.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/processors/CommandProcessor.java @@ -18,8 +18,6 @@ package org.apache.hadoop.hive.ql.processors; -import org.apache.hadoop.hive.ql.CommandNeedRetryException; - public interface CommandProcessor { - CommandProcessorResponse run(String command) throws CommandNeedRetryException; + CommandProcessorResponse run(String command); } http://git-wip-us.apache.org/repos/asf/hive/blob/f7dea106/ql/src/java/org/apache/hadoop/hive/ql/processors/CompileProcessor.java ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/processors/CompileProcessor.java b/ql/src/java/org/apache/hadoop/hive/ql/processors/CompileProcessor.java index 07d70ab..fad4f52 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/processors/CompileProcessor.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/processors/CompileProcessor.java @@ -33,7 +33,6 @@ import org.apache.commons.compress.archivers.jar.JarArchiveOutputStream; import org.apache.commons.io.IOUtils; import org.apache.hadoop.hive.conf.HiveVariableSource; import org.apache.hadoop.hive.conf.VariableSubstitution; -import org.apache.hadoop.hive.ql.CommandNeedRetryException; import org.apache.hadoop.hive.ql.metadata.HiveException; import org.apache.hadoop.hive.ql.security.authorization.plugin.HiveOperationType; import org.apache.hadoop.hive.ql.session.SessionState; @@ -104,7 +103,7 @@ public class CompileProcessor implements CommandProcessor { * @return CommandProcessorResponse with 0 for success and 1 for failure */ @Override - public CommandProcessorResponse run(String command) throws CommandNeedRetryException { + public CommandProcessorResponse run(String command) { SessionState ss = SessionState.get(); this.command = command; http://git-wip-us.apache.org/repos/asf/hive/blob/f7dea106/ql/src/java/org/apache/hadoop/hive/ql/processors/CryptoProcessor.java ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/processors/CryptoProcessor.java b/ql/src/java/org/apache/hadoop/hive/ql/processors/CryptoProcessor.java index 6825dd8..d1202f9 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/processors/CryptoProcessor.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/processors/CryptoProcessor.java @@ -28,7 +28,6 @@ import org.slf4j.Logger; import org.slf4j.LoggerFactory; import org.apache.hadoop.fs.Path; import org.apache.hadoop.hive.conf.HiveConf; -import org.apache.hadoop.hive.ql.CommandNeedRetryException; import org.apache.hadoop.hive.ql.session.SessionState; import org.apache.hadoop.hive.shims.HadoopShims; @@ -82,7 +81,7 @@ public class CryptoProcessor implements CommandProcessor { } @Override - public CommandProcessorResponse run(String command) throws CommandNeedRetryException { + public CommandProcessorResponse run(String command) { String[] args = command.split("\\s+"); if (args.length < 1) { http://git-wip-us.apache.org/repos/asf/hive/blob/f7dea106/ql/src/java/org/apache/hadoop/hive/ql/processors/DfsProcessor.java ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/processors/DfsProcessor.java b/ql/src/java/org/apache/hadoop/hive/ql/processors/DfsProcessor.java index 2f288ce..62a1725 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/processors/DfsProcessor.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/processors/DfsProcessor.java @@ -31,7 +31,7 @@ import org.apache.hadoop.hive.conf.HiveVariableSource; import org.apache.hadoop.hive.conf.VariableSubstitution; import org.apache.hadoop.hive.metastore.api.FieldSchema; import org.apache.hadoop.hive.metastore.api.Schema; -import org.apache.hadoop.hive.ql.CommandNeedRetryException; +import org.apache.hadoop.hive.ql.metadata.HiveException; import org.apache.hadoop.hive.ql.security.authorization.plugin.HiveOperationType; import org.apache.hadoop.hive.ql.session.SessionState; import org.apache.hadoop.hive.ql.session.SessionState.LogHelper; @@ -102,7 +102,7 @@ public class DfsProcessor implements CommandProcessor { } } - private String[] splitCmd(String command) throws CommandNeedRetryException { + private String[] splitCmd(String command) throws HiveException { ArrayList<String> paras = new ArrayList<String>(); int cmdLng = command.length(); @@ -151,8 +151,9 @@ public class DfsProcessor implements CommandProcessor { } if ((int) y != 0) { - console.printError("Syntax error on hadoop options: dfs " + command); - throw new CommandNeedRetryException(); + String message = "Syntax error on hadoop options: dfs " + command; + console.printError(message); + throw new HiveException(message); } return paras.toArray(new String[paras.size()]); http://git-wip-us.apache.org/repos/asf/hive/blob/f7dea106/ql/src/java/org/apache/hadoop/hive/ql/processors/ListResourceProcessor.java ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/processors/ListResourceProcessor.java b/ql/src/java/org/apache/hadoop/hive/ql/processors/ListResourceProcessor.java index 7ec36be..91a6aba 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/processors/ListResourceProcessor.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/processors/ListResourceProcessor.java @@ -21,7 +21,6 @@ package org.apache.hadoop.hive.ql.processors; import org.apache.commons.lang.StringUtils; import org.apache.hadoop.hive.metastore.api.FieldSchema; import org.apache.hadoop.hive.metastore.api.Schema; -import org.apache.hadoop.hive.ql.CommandNeedRetryException; import org.apache.hadoop.hive.ql.session.SessionState; import java.util.Arrays; @@ -44,7 +43,7 @@ public class ListResourceProcessor implements CommandProcessor { } @Override - public CommandProcessorResponse run(String command) throws CommandNeedRetryException { + public CommandProcessorResponse run(String command) { SessionState ss = SessionState.get(); String[] tokens = command.split("\\s+"); SessionState.ResourceType t; http://git-wip-us.apache.org/repos/asf/hive/blob/f7dea106/ql/src/java/org/apache/hadoop/hive/ql/processors/ReloadProcessor.java ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/processors/ReloadProcessor.java b/ql/src/java/org/apache/hadoop/hive/ql/processors/ReloadProcessor.java index b82bd5c..4caab91 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/processors/ReloadProcessor.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/processors/ReloadProcessor.java @@ -22,7 +22,6 @@ import java.io.IOException; import org.slf4j.Logger; import org.slf4j.LoggerFactory; -import org.apache.hadoop.hive.ql.CommandNeedRetryException; import org.apache.hadoop.hive.ql.session.SessionState; /** @@ -32,7 +31,7 @@ public class ReloadProcessor implements CommandProcessor{ private static final Logger LOG = LoggerFactory.getLogger(ReloadProcessor.class); @Override - public CommandProcessorResponse run(String command) throws CommandNeedRetryException { + public CommandProcessorResponse run(String command) { SessionState ss = SessionState.get(); try { ss.loadReloadableAuxJars(); http://git-wip-us.apache.org/repos/asf/hive/blob/f7dea106/ql/src/java/org/apache/hadoop/hive/ql/processors/ResetProcessor.java ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/processors/ResetProcessor.java b/ql/src/java/org/apache/hadoop/hive/ql/processors/ResetProcessor.java index 144f522..ca39ff9 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/processors/ResetProcessor.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/processors/ResetProcessor.java @@ -32,7 +32,6 @@ import org.apache.hadoop.hive.conf.HiveConf; import org.apache.hadoop.hive.conf.HiveVariableSource; import org.apache.hadoop.hive.conf.SystemVariables; import org.apache.hadoop.hive.conf.VariableSubstitution; -import org.apache.hadoop.hive.ql.CommandNeedRetryException; import org.apache.hadoop.hive.ql.metadata.Hive; import org.apache.hadoop.hive.ql.security.authorization.plugin.HiveOperationType; import org.apache.hadoop.hive.ql.session.SessionState; @@ -43,12 +42,12 @@ public class ResetProcessor implements CommandProcessor { private final static String DEFAULT_ARG = "-d"; @Override - public CommandProcessorResponse run(String command) throws CommandNeedRetryException { + public CommandProcessorResponse run(String command) { return run(SessionState.get(), command); } @VisibleForTesting - CommandProcessorResponse run(SessionState ss, String command) throws CommandNeedRetryException { + CommandProcessorResponse run(SessionState ss, String command) { CommandProcessorResponse authErrResp = CommandUtil.authorizeCommand(ss, HiveOperationType.RESET, Arrays.asList(command)); if (authErrResp != null) { http://git-wip-us.apache.org/repos/asf/hive/blob/f7dea106/ql/src/java/org/apache/hadoop/hive/ql/txn/compactor/Worker.java ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/txn/compactor/Worker.java b/ql/src/java/org/apache/hadoop/hive/ql/txn/compactor/Worker.java index 4508e59..26c6700 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/txn/compactor/Worker.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/txn/compactor/Worker.java @@ -30,7 +30,6 @@ import org.apache.hadoop.hive.metastore.api.StorageDescriptor; import org.apache.hadoop.hive.metastore.api.Table; import org.apache.hadoop.hive.metastore.txn.CompactionInfo; import org.apache.hadoop.hive.metastore.txn.TxnUtils; -import org.apache.hadoop.hive.ql.CommandNeedRetryException; import org.apache.hadoop.hive.ql.Driver; import org.apache.hadoop.hive.ql.processors.CommandProcessorResponse; import org.apache.hadoop.hive.ql.session.SessionState; @@ -300,10 +299,6 @@ public class Worker extends CompactorThread { (ci.partName == null ? "" : "/" + ci.partName) + " due to: " + cpr); } } - catch(CommandNeedRetryException cnre) { - throw new IOException("Could not update stats for table " + ci.getFullTableName() + - (ci.partName == null ? "" : "/" + ci.partName) + " due to: " + cnre.getMessage()); - } finally { if(localSession != null) { localSession.close(); http://git-wip-us.apache.org/repos/asf/hive/blob/f7dea106/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDTFGetSplits.java ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDTFGetSplits.java b/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDTFGetSplits.java index d56002d..1e83799 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDTFGetSplits.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDTFGetSplits.java @@ -58,7 +58,6 @@ import org.apache.hadoop.hive.llap.security.LlapTokenIdentifier; import org.apache.hadoop.hive.llap.security.LlapTokenLocalClient; import org.apache.hadoop.hive.llap.tez.Converters; import org.apache.hadoop.hive.metastore.api.FieldSchema; -import org.apache.hadoop.hive.ql.CommandNeedRetryException; import org.apache.hadoop.hive.ql.Context; import org.apache.hadoop.hive.ql.Driver; import org.apache.hadoop.hive.ql.QueryPlan; @@ -269,13 +268,9 @@ public class GenericUDTFGetSplits extends GenericUDTF { String ctas = "create temporary table " + tableName + " as " + query; LOG.info("Materializing the query for LLAPIF; CTAS: " + ctas); - try { - driver.resetQueryState(); - HiveConf.setVar(conf, ConfVars.HIVE_EXECUTION_MODE, originalMode); - cpr = driver.run(ctas, false); - } catch (CommandNeedRetryException e) { - throw new HiveException(e); - } + driver.resetQueryState(); + HiveConf.setVar(conf, ConfVars.HIVE_EXECUTION_MODE, originalMode); + cpr = driver.run(ctas, false); if(cpr.getResponseCode() != 0) { throw new HiveException("Failed to create temp table: " + cpr.getException()); http://git-wip-us.apache.org/repos/asf/hive/blob/f7dea106/ql/src/test/org/apache/hadoop/hive/ql/hooks/TestQueryHooks.java ---------------------------------------------------------------------- diff --git a/ql/src/test/org/apache/hadoop/hive/ql/hooks/TestQueryHooks.java b/ql/src/test/org/apache/hadoop/hive/ql/hooks/TestQueryHooks.java index 0662875..06a96d5 100644 --- a/ql/src/test/org/apache/hadoop/hive/ql/hooks/TestQueryHooks.java +++ b/ql/src/test/org/apache/hadoop/hive/ql/hooks/TestQueryHooks.java @@ -21,7 +21,6 @@ package org.apache.hadoop.hive.ql.hooks; import com.google.common.collect.Lists; import org.apache.hadoop.hive.conf.HiveConf; -import org.apache.hadoop.hive.ql.CommandNeedRetryException; import org.apache.hadoop.hive.ql.Driver; import org.apache.hadoop.hive.ql.session.SessionState; @@ -55,7 +54,7 @@ public class TestQueryHooks { } @Test - public void testAllQueryLifeTimeWithParseHooks() throws IllegalAccessException, ClassNotFoundException, InstantiationException, CommandNeedRetryException { + public void testAllQueryLifeTimeWithParseHooks() throws Exception { String query = "select 1"; ArgumentMatcher<QueryLifeTimeHookContext> argMatcher = new QueryLifeTimeHookContextMatcher(query); QueryLifeTimeHookWithParseHooks mockHook = mock(QueryLifeTimeHookWithParseHooks.class); @@ -71,7 +70,7 @@ public class TestQueryHooks { } @Test - public void testQueryLifeTimeWithParseHooksWithParseError() throws IllegalAccessException, ClassNotFoundException, InstantiationException, CommandNeedRetryException { + public void testQueryLifeTimeWithParseHooksWithParseError() throws Exception { String query = "invalidquery"; ArgumentMatcher<QueryLifeTimeHookContext> argMatcher = new QueryLifeTimeHookContextMatcher(query); QueryLifeTimeHookWithParseHooks mockHook = mock(QueryLifeTimeHookWithParseHooks.class); @@ -87,7 +86,7 @@ public class TestQueryHooks { } @Test - public void testQueryLifeTimeWithParseHooksWithCompileError() throws IllegalAccessException, ClassNotFoundException, InstantiationException, CommandNeedRetryException { + public void testQueryLifeTimeWithParseHooksWithCompileError() throws Exception { String query = "select * from foo"; ArgumentMatcher<QueryLifeTimeHookContext> argMatcher = new QueryLifeTimeHookContextMatcher(query); QueryLifeTimeHookWithParseHooks mockHook = mock(QueryLifeTimeHookWithParseHooks.class); @@ -103,7 +102,7 @@ public class TestQueryHooks { } @Test - public void testAllQueryLifeTimeHooks() throws IllegalAccessException, ClassNotFoundException, InstantiationException, CommandNeedRetryException { + public void testAllQueryLifeTimeHooks() throws Exception { String query = "select 1"; ArgumentMatcher<QueryLifeTimeHookContext> argMatcher = new QueryLifeTimeHookContextMatcher(query); QueryLifeTimeHook mockHook = mock(QueryLifeTimeHook.class); @@ -117,7 +116,7 @@ public class TestQueryHooks { } @Test - public void testQueryLifeTimeWithCompileError() throws IllegalAccessException, ClassNotFoundException, InstantiationException, CommandNeedRetryException { + public void testQueryLifeTimeWithCompileError() throws Exception { String query = "select * from foo"; ArgumentMatcher<QueryLifeTimeHookContext> argMatcher = new QueryLifeTimeHookContextMatcher(query); QueryLifeTimeHook mockHook = mock(QueryLifeTimeHook.class); http://git-wip-us.apache.org/repos/asf/hive/blob/f7dea106/ql/src/test/org/apache/hadoop/hive/ql/parse/TestColumnAccess.java ---------------------------------------------------------------------- diff --git a/ql/src/test/org/apache/hadoop/hive/ql/parse/TestColumnAccess.java b/ql/src/test/org/apache/hadoop/hive/ql/parse/TestColumnAccess.java index 4da7d92..d2b9327 100644 --- a/ql/src/test/org/apache/hadoop/hive/ql/parse/TestColumnAccess.java +++ b/ql/src/test/org/apache/hadoop/hive/ql/parse/TestColumnAccess.java @@ -23,10 +23,9 @@ import java.util.HashSet; import java.util.List; import java.util.Map; -import junit.framework.Assert; +import org.junit.Assert; import org.apache.hadoop.hive.conf.HiveConf; -import org.apache.hadoop.hive.ql.CommandNeedRetryException; import org.apache.hadoop.hive.ql.Driver; import org.apache.hadoop.hive.ql.QueryPlan; import org.apache.hadoop.hive.ql.hooks.ReadEntity; @@ -38,7 +37,7 @@ import org.junit.Test; public class TestColumnAccess { @BeforeClass - public static void Setup() throws CommandNeedRetryException { + public static void Setup() throws Exception { Driver driver = createDriver(); int ret = driver.run("create table t1(id1 int, name1 string)").getResponseCode(); Assert.assertEquals("Checking command success", 0, ret); http://git-wip-us.apache.org/repos/asf/hive/blob/f7dea106/ql/src/test/org/apache/hadoop/hive/ql/plan/TestReadEntityDirect.java ---------------------------------------------------------------------- diff --git a/ql/src/test/org/apache/hadoop/hive/ql/plan/TestReadEntityDirect.java b/ql/src/test/org/apache/hadoop/hive/ql/plan/TestReadEntityDirect.java index 4ad821c..88edc12 100644 --- a/ql/src/test/org/apache/hadoop/hive/ql/plan/TestReadEntityDirect.java +++ b/ql/src/test/org/apache/hadoop/hive/ql/plan/TestReadEntityDirect.java @@ -27,7 +27,6 @@ import java.util.List; import java.util.Set; import org.apache.hadoop.hive.conf.HiveConf; -import org.apache.hadoop.hive.ql.CommandNeedRetryException; import org.apache.hadoop.hive.ql.Driver; import org.apache.hadoop.hive.ql.exec.Task; import org.apache.hadoop.hive.ql.hooks.ReadEntity; @@ -49,7 +48,7 @@ import org.junit.Test; public class TestReadEntityDirect { @BeforeClass - public static void onetimeSetup() throws CommandNeedRetryException { + public static void onetimeSetup() throws Exception { Driver driver = createDriver(); int ret = driver.run("create table t1(i int)").getResponseCode(); assertEquals("Checking command success", 0, ret); http://git-wip-us.apache.org/repos/asf/hive/blob/f7dea106/ql/src/test/org/apache/hadoop/hive/ql/processors/TestResetProcessor.java ---------------------------------------------------------------------- diff --git a/ql/src/test/org/apache/hadoop/hive/ql/processors/TestResetProcessor.java b/ql/src/test/org/apache/hadoop/hive/ql/processors/TestResetProcessor.java index 99f3f6d..26e9083 100644 --- a/ql/src/test/org/apache/hadoop/hive/ql/processors/TestResetProcessor.java +++ b/ql/src/test/org/apache/hadoop/hive/ql/processors/TestResetProcessor.java @@ -22,7 +22,6 @@ import java.util.HashMap; import java.util.Map; import org.apache.hadoop.hive.conf.HiveConf; -import org.apache.hadoop.hive.ql.CommandNeedRetryException; import org.apache.hadoop.hive.ql.session.SessionState; import org.junit.Test; @@ -35,14 +34,14 @@ import static org.mockito.Mockito.when; public class TestResetProcessor { @Test - public void testResetClosesSparkSession() throws CommandNeedRetryException { + public void testResetClosesSparkSession() throws Exception { SessionState mockSessionState = createMockSparkSessionState(); new ResetProcessor().run(mockSessionState, ""); verify(mockSessionState).closeSparkSession(); } @Test - public void testResetExecutionEngineClosesSparkSession() throws CommandNeedRetryException { + public void testResetExecutionEngineClosesSparkSession() throws Exception { SessionState mockSessionState = createMockSparkSessionState(); new ResetProcessor().run(mockSessionState, HiveConf.ConfVars.HIVE_EXECUTION_ENGINE.varname); verify(mockSessionState).closeSparkSession(); http://git-wip-us.apache.org/repos/asf/hive/blob/f7dea106/service/src/jamon/org/apache/hive/tmpl/QueryProfileTmpl.jamon ---------------------------------------------------------------------- diff --git a/service/src/jamon/org/apache/hive/tmpl/QueryProfileTmpl.jamon b/service/src/jamon/org/apache/hive/tmpl/QueryProfileTmpl.jamon index ff7476e..366198e 100644 --- a/service/src/jamon/org/apache/hive/tmpl/QueryProfileTmpl.jamon +++ b/service/src/jamon/org/apache/hive/tmpl/QueryProfileTmpl.jamon @@ -179,7 +179,6 @@ org.apache.hadoop.hive.conf.HiveConf; <th>End Time</th> <th>Elapsed Time (s)</th> <th>Requires Lock</th> - <th>Retry If Fail</th> </tr> <%if queryInfo.getQueryDisplay() != null && queryInfo.getQueryDisplay().getTaskDisplays() != null %> @@ -191,7 +190,6 @@ org.apache.hadoop.hive.conf.HiveConf; <td><% taskDisplay.getEndTime() == null ? "" : new Date(taskDisplay.getEndTime()) %></td> <td><% taskDisplay.getElapsedTime() == null ? "" : taskDisplay.getElapsedTime()/1000 %> (s) </td> <td><% taskDisplay.isRequireLock() %></td> - <td><% taskDisplay.isRetryIfFail() %></td> </tr> </%for> </%if> http://git-wip-us.apache.org/repos/asf/hive/blob/f7dea106/service/src/java/org/apache/hive/service/cli/operation/SQLOperation.java ---------------------------------------------------------------------- diff --git a/service/src/java/org/apache/hive/service/cli/operation/SQLOperation.java b/service/src/java/org/apache/hive/service/cli/operation/SQLOperation.java index 9af0e35..5865abe 100644 --- a/service/src/java/org/apache/hive/service/cli/operation/SQLOperation.java +++ b/service/src/java/org/apache/hive/service/cli/operation/SQLOperation.java @@ -48,7 +48,6 @@ import org.apache.hadoop.hive.common.metrics.common.MetricsScope; import org.apache.hadoop.hive.conf.HiveConf; import org.apache.hadoop.hive.metastore.api.FieldSchema; import org.apache.hadoop.hive.metastore.api.Schema; -import org.apache.hadoop.hive.ql.CommandNeedRetryException; import org.apache.hadoop.hive.ql.DriverFactory; import org.apache.hadoop.hive.ql.IDriver; import org.apache.hadoop.hive.ql.QueryDisplay; @@ -88,7 +87,6 @@ import org.codehaus.jackson.map.ObjectMapper; * SQLOperation. * */ -@SuppressWarnings("deprecation") public class SQLOperation extends ExecuteStatementOperation { private IDriver driver = null; private CommandProcessorResponse response; @@ -197,8 +195,6 @@ public class SQLOperation extends ExecuteStatementOperation { // In Hive server mode, we are not able to retry in the FetchTask // case, when calling fetch queries since execute() has returned. // For now, we disable the test attempts. - driver.setTryCount(Integer.MAX_VALUE); - response = driver.compileAndRespond(statement); if (0 != response.getResponseCode()) { throw toSQLException("Error while compiling statement", response); @@ -249,7 +245,6 @@ public class SQLOperation extends ExecuteStatementOperation { // In Hive server mode, we are not able to retry in the FetchTask // case, when calling fetch queries since execute() has returned. // For now, we disable the test attempts. - driver.setTryCount(Integer.MAX_VALUE); response = driver.run(); if (0 != response.getResponseCode()) { throw toSQLException("Error while processing statement", response); @@ -494,8 +489,6 @@ public class SQLOperation extends ExecuteStatementOperation { return rowSet; } catch (IOException e) { throw new HiveSQLException(e); - } catch (CommandNeedRetryException e) { - throw new HiveSQLException(e); } catch (Exception e) { throw new HiveSQLException(e); } finally {