hive git commit: HIVE-19127: Concurrency fixes in QueryResultsCache (Jason Dere, reviewed by Deepak Jaiswal)
Repository: hive Updated Branches: refs/heads/master 76b696c26 -> a1034102d HIVE-19127: Concurrency fixes in QueryResultsCache (Jason Dere, reviewed by Deepak Jaiswal) Project: http://git-wip-us.apache.org/repos/asf/hive/repo Commit: http://git-wip-us.apache.org/repos/asf/hive/commit/a1034102 Tree: http://git-wip-us.apache.org/repos/asf/hive/tree/a1034102 Diff: http://git-wip-us.apache.org/repos/asf/hive/diff/a1034102 Branch: refs/heads/master Commit: a1034102d3580922f6c8f9d186272280d6917802 Parents: 76b696c Author: Jason Dere Authored: Mon Apr 9 16:48:23 2018 -0700 Committer: Jason Dere Committed: Mon Apr 9 16:48:23 2018 -0700 -- .../ql/cache/results/QueryResultsCache.java | 112 +++ 1 file changed, 68 insertions(+), 44 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hive/blob/a1034102/ql/src/java/org/apache/hadoop/hive/ql/cache/results/QueryResultsCache.java -- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/cache/results/QueryResultsCache.java b/ql/src/java/org/apache/hadoop/hive/ql/cache/results/QueryResultsCache.java index ac5ae57..b1a3646 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/cache/results/QueryResultsCache.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/cache/results/QueryResultsCache.java @@ -502,32 +502,39 @@ public final class QueryResultsCache { return false; } - if (requiresMove) { -// Move the query results to the query cache directory. -cachedResultsPath = moveResultsToCacheDirectory(queryResultsPath); -dataDirMoved = true; - } - LOG.info("Moved query results from {} to {} (size {}) for query '{}'", - queryResultsPath, cachedResultsPath, resultSize, queryText); - - // Create a new FetchWork to reference the new cache location. - FetchWork fetchWorkForCache = - new FetchWork(cachedResultsPath, fetchWork.getTblDesc(), fetchWork.getLimit()); - fetchWorkForCache.setCachedResult(true); - cacheEntry.fetchWork = fetchWorkForCache; - cacheEntry.cachedResultsPath = cachedResultsPath; - cacheEntry.size = resultSize; - this.cacheSize += resultSize; - cacheEntry.createTime = System.currentTimeMillis(); - - cacheEntry.setStatus(CacheEntryStatus.VALID); - // Mark this entry as being in use. Caller will need to release later. - cacheEntry.addReader(); - - scheduleEntryInvalidation(cacheEntry); - - // Notify any queries waiting on this cacheEntry to become valid. + // Synchronize on the cache entry so that no one else can invalidate this entry + // while we are in the process of setting it to valid. synchronized (cacheEntry) { +if (cacheEntry.getStatus() == CacheEntryStatus.INVALID) { + // Entry either expired, or was invalidated due to table updates + return false; +} + +if (requiresMove) { + // Move the query results to the query cache directory. + cachedResultsPath = moveResultsToCacheDirectory(queryResultsPath); + dataDirMoved = true; +} +LOG.info("Moved query results from {} to {} (size {}) for query '{}'", +queryResultsPath, cachedResultsPath, resultSize, queryText); + +// Create a new FetchWork to reference the new cache location. +FetchWork fetchWorkForCache = +new FetchWork(cachedResultsPath, fetchWork.getTblDesc(), fetchWork.getLimit()); +fetchWorkForCache.setCachedResult(true); +cacheEntry.fetchWork = fetchWorkForCache; +cacheEntry.cachedResultsPath = cachedResultsPath; +cacheEntry.size = resultSize; +this.cacheSize += resultSize; +cacheEntry.createTime = System.currentTimeMillis(); + +cacheEntry.setStatus(CacheEntryStatus.VALID); +// Mark this entry as being in use. Caller will need to release later. +cacheEntry.addReader(); + +scheduleEntryInvalidation(cacheEntry); + +// Notify any queries waiting on this cacheEntry to become valid. cacheEntry.notifyAll(); } @@ -564,7 +571,11 @@ public final class QueryResultsCache { try { writeLock.lock(); LOG.info("Clearing the results cache"); - for (CacheEntry entry : lru.keySet().toArray(EMPTY_CACHEENTRY_ARRAY)) { + CacheEntry[] allEntries = null; + synchronized (lru) { +allEntries = lru.keySet().toArray(EMPTY_CACHEENTRY_ARRAY); + } + for (CacheEntry entry : allEntries) { try { removeEntry(entry); } catch (Exception err) { @@ -611,10 +622,15 @@ public final class QueryResultsCache { public void removeEntry(CacheEntry entry) { entry.invalidate(); -removeFromLookup(entry); -lru.remove(entry); -// Should the cache si
hive git commit: HIVE-19127: Concurrency fixes in QueryResultsCache (Jason Dere, reviewed by Deepak Jaiswal)
Repository: hive Updated Branches: refs/heads/branch-3 1cd74b451 -> 3f56b44fb HIVE-19127: Concurrency fixes in QueryResultsCache (Jason Dere, reviewed by Deepak Jaiswal) Project: http://git-wip-us.apache.org/repos/asf/hive/repo Commit: http://git-wip-us.apache.org/repos/asf/hive/commit/3f56b44f Tree: http://git-wip-us.apache.org/repos/asf/hive/tree/3f56b44f Diff: http://git-wip-us.apache.org/repos/asf/hive/diff/3f56b44f Branch: refs/heads/branch-3 Commit: 3f56b44fb9dcd427ae1a93e9739da84a33eb6aed Parents: 1cd74b4 Author: Jason Dere Authored: Mon Apr 9 16:48:23 2018 -0700 Committer: Jason Dere Committed: Mon Apr 9 16:49:17 2018 -0700 -- .../ql/cache/results/QueryResultsCache.java | 112 +++ 1 file changed, 68 insertions(+), 44 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hive/blob/3f56b44f/ql/src/java/org/apache/hadoop/hive/ql/cache/results/QueryResultsCache.java -- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/cache/results/QueryResultsCache.java b/ql/src/java/org/apache/hadoop/hive/ql/cache/results/QueryResultsCache.java index ac5ae57..b1a3646 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/cache/results/QueryResultsCache.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/cache/results/QueryResultsCache.java @@ -502,32 +502,39 @@ public final class QueryResultsCache { return false; } - if (requiresMove) { -// Move the query results to the query cache directory. -cachedResultsPath = moveResultsToCacheDirectory(queryResultsPath); -dataDirMoved = true; - } - LOG.info("Moved query results from {} to {} (size {}) for query '{}'", - queryResultsPath, cachedResultsPath, resultSize, queryText); - - // Create a new FetchWork to reference the new cache location. - FetchWork fetchWorkForCache = - new FetchWork(cachedResultsPath, fetchWork.getTblDesc(), fetchWork.getLimit()); - fetchWorkForCache.setCachedResult(true); - cacheEntry.fetchWork = fetchWorkForCache; - cacheEntry.cachedResultsPath = cachedResultsPath; - cacheEntry.size = resultSize; - this.cacheSize += resultSize; - cacheEntry.createTime = System.currentTimeMillis(); - - cacheEntry.setStatus(CacheEntryStatus.VALID); - // Mark this entry as being in use. Caller will need to release later. - cacheEntry.addReader(); - - scheduleEntryInvalidation(cacheEntry); - - // Notify any queries waiting on this cacheEntry to become valid. + // Synchronize on the cache entry so that no one else can invalidate this entry + // while we are in the process of setting it to valid. synchronized (cacheEntry) { +if (cacheEntry.getStatus() == CacheEntryStatus.INVALID) { + // Entry either expired, or was invalidated due to table updates + return false; +} + +if (requiresMove) { + // Move the query results to the query cache directory. + cachedResultsPath = moveResultsToCacheDirectory(queryResultsPath); + dataDirMoved = true; +} +LOG.info("Moved query results from {} to {} (size {}) for query '{}'", +queryResultsPath, cachedResultsPath, resultSize, queryText); + +// Create a new FetchWork to reference the new cache location. +FetchWork fetchWorkForCache = +new FetchWork(cachedResultsPath, fetchWork.getTblDesc(), fetchWork.getLimit()); +fetchWorkForCache.setCachedResult(true); +cacheEntry.fetchWork = fetchWorkForCache; +cacheEntry.cachedResultsPath = cachedResultsPath; +cacheEntry.size = resultSize; +this.cacheSize += resultSize; +cacheEntry.createTime = System.currentTimeMillis(); + +cacheEntry.setStatus(CacheEntryStatus.VALID); +// Mark this entry as being in use. Caller will need to release later. +cacheEntry.addReader(); + +scheduleEntryInvalidation(cacheEntry); + +// Notify any queries waiting on this cacheEntry to become valid. cacheEntry.notifyAll(); } @@ -564,7 +571,11 @@ public final class QueryResultsCache { try { writeLock.lock(); LOG.info("Clearing the results cache"); - for (CacheEntry entry : lru.keySet().toArray(EMPTY_CACHEENTRY_ARRAY)) { + CacheEntry[] allEntries = null; + synchronized (lru) { +allEntries = lru.keySet().toArray(EMPTY_CACHEENTRY_ARRAY); + } + for (CacheEntry entry : allEntries) { try { removeEntry(entry); } catch (Exception err) { @@ -611,10 +622,15 @@ public final class QueryResultsCache { public void removeEntry(CacheEntry entry) { entry.invalidate(); -removeFromLookup(entry); -lru.remove(entry); -// Should the cach
hive git commit: HIVE-19143 : Update golden files for negative tests
Repository: hive Updated Branches: refs/heads/master a1034102d -> 55fb0a196 HIVE-19143 : Update golden files for negative tests Project: http://git-wip-us.apache.org/repos/asf/hive/repo Commit: http://git-wip-us.apache.org/repos/asf/hive/commit/55fb0a19 Tree: http://git-wip-us.apache.org/repos/asf/hive/tree/55fb0a19 Diff: http://git-wip-us.apache.org/repos/asf/hive/diff/55fb0a19 Branch: refs/heads/master Commit: 55fb0a196d870ff4452137bf2d34c7e415b77b76 Parents: a103410 Author: Ashutosh Chauhan Authored: Mon Apr 9 17:50:32 2018 -0700 Committer: Ashutosh Chauhan Committed: Mon Apr 9 17:52:22 2018 -0700 -- .../results/clientnegative/authorization_caseinsensitivity.q.out | 2 +- ql/src/test/results/clientnegative/authorization_fail_1.q.out | 2 +- .../results/clientnegative/authorization_grant_table_dup.q.out | 2 +- ql/src/test/results/clientnegative/authorization_role_case.q.out | 2 +- .../clientnegative/authorization_role_grant_nosuchrole.q.out | 2 +- .../clientnegative/authorization_table_grant_nosuchrole.q.out | 2 +- ql/src/test/results/clientnegative/subquery_subquery_chain.q.out | 2 +- 7 files changed, 7 insertions(+), 7 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hive/blob/55fb0a19/ql/src/test/results/clientnegative/authorization_caseinsensitivity.q.out -- diff --git a/ql/src/test/results/clientnegative/authorization_caseinsensitivity.q.out b/ql/src/test/results/clientnegative/authorization_caseinsensitivity.q.out index 1dbc3e2..8bc747e 100644 --- a/ql/src/test/results/clientnegative/authorization_caseinsensitivity.q.out +++ b/ql/src/test/results/clientnegative/authorization_caseinsensitivity.q.out @@ -55,4 +55,4 @@ public testrole PREHOOK: query: create role TESTRoLE PREHOOK: type: CREATEROLE -FAILED: Execution Error, return code 1 from org.apache.hadoop.hive.ql.exec.DDLTask. Role testrole already exists. +FAILED: Execution Error, return code 1 from org.apache.hadoop.hive.ql.exec.DDLTask. Error create role: Got exception: org.apache.hadoop.hive.metastore.api.InvalidObjectException Role testrole already exists. http://git-wip-us.apache.org/repos/asf/hive/blob/55fb0a19/ql/src/test/results/clientnegative/authorization_fail_1.q.out -- diff --git a/ql/src/test/results/clientnegative/authorization_fail_1.q.out b/ql/src/test/results/clientnegative/authorization_fail_1.q.out index 5c78f2a..fc52cb3 100644 --- a/ql/src/test/results/clientnegative/authorization_fail_1.q.out +++ b/ql/src/test/results/clientnegative/authorization_fail_1.q.out @@ -15,4 +15,4 @@ POSTHOOK: Output: default@authorization_fail_1 PREHOOK: query: grant Create on table authorization_fail_1 to user hive_test_user PREHOOK: type: GRANT_PRIVILEGE PREHOOK: Output: default@authorization_fail_1 -FAILED: Execution Error, return code 1 from org.apache.hadoop.hive.ql.exec.DDLTask. org.apache.hadoop.hive.ql.metadata.HiveException: java.lang.RuntimeException: InvalidObjectException(message:CREATE is already granted on table [default,authorization_fail_1] by hive_test_user) +FAILED: Execution Error, return code 1 from org.apache.hadoop.hive.ql.exec.DDLTask. org.apache.hadoop.hive.ql.metadata.HiveException: MetaException(message:Got exception: org.apache.hadoop.hive.metastore.api.InvalidObjectException CREATE is already granted on table [default,authorization_fail_1] by hive_test_user) http://git-wip-us.apache.org/repos/asf/hive/blob/55fb0a19/ql/src/test/results/clientnegative/authorization_grant_table_dup.q.out -- diff --git a/ql/src/test/results/clientnegative/authorization_grant_table_dup.q.out b/ql/src/test/results/clientnegative/authorization_grant_table_dup.q.out index 9449474..795dc83 100644 --- a/ql/src/test/results/clientnegative/authorization_grant_table_dup.q.out +++ b/ql/src/test/results/clientnegative/authorization_grant_table_dup.q.out @@ -22,4 +22,4 @@ default tauth_gdup user1 USER UPDATE true-1 user1 PREHOOK: query: GRANT INSERT ON tauth_gdup TO USER user1 PREHOOK: type: GRANT_PRIVILEGE PREHOOK: Output: default@tauth_gdup -FAILED: Execution Error, return code 1 from org.apache.hadoop.hive.ql.exec.DDLTask. Error granting privileges: InvalidObjectException(message:INSERT is already granted on table [default,tauth_gdup] by user1) +FAILED: Execution Error, return code 1 from org.apache.hadoop.hive.ql.exec.DDLTask. Error granting privileges: Got exception: org.apache.hadoop.hive.metastore.api.InvalidObjectException INSERT is already granted on table [default,tauth_gdup] by user1 http://git-wip-us.apache.org/repos/asf/hive/blob/55fb0a19/ql/src/test/results/clientnegative/authoriz
hive git commit: HIVE-19143 : Update golden files for negative tests
Repository: hive Updated Branches: refs/heads/branch-3 3f56b44fb -> 946f619e7 HIVE-19143 : Update golden files for negative tests Project: http://git-wip-us.apache.org/repos/asf/hive/repo Commit: http://git-wip-us.apache.org/repos/asf/hive/commit/946f619e Tree: http://git-wip-us.apache.org/repos/asf/hive/tree/946f619e Diff: http://git-wip-us.apache.org/repos/asf/hive/diff/946f619e Branch: refs/heads/branch-3 Commit: 946f619e78fbe9e4ab32d77eacc672135157253c Parents: 3f56b44 Author: Ashutosh Chauhan Authored: Mon Apr 9 17:50:32 2018 -0700 Committer: Ashutosh Chauhan Committed: Mon Apr 9 17:50:32 2018 -0700 -- .../results/clientnegative/authorization_caseinsensitivity.q.out | 2 +- ql/src/test/results/clientnegative/authorization_fail_1.q.out | 2 +- .../results/clientnegative/authorization_grant_table_dup.q.out | 2 +- ql/src/test/results/clientnegative/authorization_role_case.q.out | 2 +- .../clientnegative/authorization_role_grant_nosuchrole.q.out | 2 +- .../clientnegative/authorization_table_grant_nosuchrole.q.out | 2 +- ql/src/test/results/clientnegative/subquery_subquery_chain.q.out | 2 +- 7 files changed, 7 insertions(+), 7 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hive/blob/946f619e/ql/src/test/results/clientnegative/authorization_caseinsensitivity.q.out -- diff --git a/ql/src/test/results/clientnegative/authorization_caseinsensitivity.q.out b/ql/src/test/results/clientnegative/authorization_caseinsensitivity.q.out index 1dbc3e2..8bc747e 100644 --- a/ql/src/test/results/clientnegative/authorization_caseinsensitivity.q.out +++ b/ql/src/test/results/clientnegative/authorization_caseinsensitivity.q.out @@ -55,4 +55,4 @@ public testrole PREHOOK: query: create role TESTRoLE PREHOOK: type: CREATEROLE -FAILED: Execution Error, return code 1 from org.apache.hadoop.hive.ql.exec.DDLTask. Role testrole already exists. +FAILED: Execution Error, return code 1 from org.apache.hadoop.hive.ql.exec.DDLTask. Error create role: Got exception: org.apache.hadoop.hive.metastore.api.InvalidObjectException Role testrole already exists. http://git-wip-us.apache.org/repos/asf/hive/blob/946f619e/ql/src/test/results/clientnegative/authorization_fail_1.q.out -- diff --git a/ql/src/test/results/clientnegative/authorization_fail_1.q.out b/ql/src/test/results/clientnegative/authorization_fail_1.q.out index 5c78f2a..fc52cb3 100644 --- a/ql/src/test/results/clientnegative/authorization_fail_1.q.out +++ b/ql/src/test/results/clientnegative/authorization_fail_1.q.out @@ -15,4 +15,4 @@ POSTHOOK: Output: default@authorization_fail_1 PREHOOK: query: grant Create on table authorization_fail_1 to user hive_test_user PREHOOK: type: GRANT_PRIVILEGE PREHOOK: Output: default@authorization_fail_1 -FAILED: Execution Error, return code 1 from org.apache.hadoop.hive.ql.exec.DDLTask. org.apache.hadoop.hive.ql.metadata.HiveException: java.lang.RuntimeException: InvalidObjectException(message:CREATE is already granted on table [default,authorization_fail_1] by hive_test_user) +FAILED: Execution Error, return code 1 from org.apache.hadoop.hive.ql.exec.DDLTask. org.apache.hadoop.hive.ql.metadata.HiveException: MetaException(message:Got exception: org.apache.hadoop.hive.metastore.api.InvalidObjectException CREATE is already granted on table [default,authorization_fail_1] by hive_test_user) http://git-wip-us.apache.org/repos/asf/hive/blob/946f619e/ql/src/test/results/clientnegative/authorization_grant_table_dup.q.out -- diff --git a/ql/src/test/results/clientnegative/authorization_grant_table_dup.q.out b/ql/src/test/results/clientnegative/authorization_grant_table_dup.q.out index 9449474..795dc83 100644 --- a/ql/src/test/results/clientnegative/authorization_grant_table_dup.q.out +++ b/ql/src/test/results/clientnegative/authorization_grant_table_dup.q.out @@ -22,4 +22,4 @@ default tauth_gdup user1 USER UPDATE true-1 user1 PREHOOK: query: GRANT INSERT ON tauth_gdup TO USER user1 PREHOOK: type: GRANT_PRIVILEGE PREHOOK: Output: default@tauth_gdup -FAILED: Execution Error, return code 1 from org.apache.hadoop.hive.ql.exec.DDLTask. Error granting privileges: InvalidObjectException(message:INSERT is already granted on table [default,tauth_gdup] by user1) +FAILED: Execution Error, return code 1 from org.apache.hadoop.hive.ql.exec.DDLTask. Error granting privileges: Got exception: org.apache.hadoop.hive.metastore.api.InvalidObjectException INSERT is already granted on table [default,tauth_gdup] by user1 http://git-wip-us.apache.org/repos/asf/hive/blob/946f619e/ql/src/test/results/clientnegative/auth
hive git commit: HIVE-18857: Store default value text instead of default value expression in metastore(Vineet Garg, reviewed by Ashutosh Chauhan)
Repository: hive Updated Branches: refs/heads/master 55fb0a196 -> 2e92451a6 HIVE-18857: Store default value text instead of default value expression in metastore(Vineet Garg, reviewed by Ashutosh Chauhan) Project: http://git-wip-us.apache.org/repos/asf/hive/repo Commit: http://git-wip-us.apache.org/repos/asf/hive/commit/2e92451a Tree: http://git-wip-us.apache.org/repos/asf/hive/tree/2e92451a Diff: http://git-wip-us.apache.org/repos/asf/hive/diff/2e92451a Branch: refs/heads/master Commit: 2e92451a61557248e47453e7917d81c428379bfa Parents: 55fb0a1 Author: Vineet Garg Authored: Mon Apr 9 19:31:14 2018 -0700 Committer: Vineet Garg Committed: Mon Apr 9 19:31:14 2018 -0700 -- .../hadoop/hive/ql/parse/BaseSemanticAnalyzer.java | 15 +-- .../hadoop/hive/ql/parse/DDLSemanticAnalyzer.java| 3 ++- 2 files changed, 11 insertions(+), 7 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hive/blob/2e92451a/ql/src/java/org/apache/hadoop/hive/ql/parse/BaseSemanticAnalyzer.java -- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/parse/BaseSemanticAnalyzer.java b/ql/src/java/org/apache/hadoop/hive/ql/parse/BaseSemanticAnalyzer.java index 5301b2a..d940cdd 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/parse/BaseSemanticAnalyzer.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/parse/BaseSemanticAnalyzer.java @@ -765,10 +765,11 @@ public abstract class BaseSemanticAnalyzer { } protected static void processDefaultConstraints(String catName, String databaseName, String tableName, - ASTNode child, List columnNames, List defaultConstraints, final ASTNode typeChild) + ASTNode child, List columnNames, List defaultConstraints, final ASTNode typeChild, + final TokenRewriteStream tokenRewriteStream) throws SemanticException { List defaultInfos = new ArrayList(); -generateConstraintInfos(child, columnNames, defaultInfos, typeChild, null); +generateConstraintInfos(child, columnNames, defaultInfos, typeChild, tokenRewriteStream); constraintInfosToDefaultConstraints(catName, databaseName, tableName, defaultInfos, defaultConstraints); } @@ -930,7 +931,8 @@ public abstract class BaseSemanticAnalyzer { * @return retrieve the default value and return it as string * @throws SemanticException */ - private static String getDefaultValue(ASTNode defaultValueAST, ASTNode typeChild) throws SemanticException{ + private static String getDefaultValue(ASTNode defaultValueAST, ASTNode typeChild, +final TokenRewriteStream tokenStream) throws SemanticException{ // first create expression from defaultValueAST TypeCheckCtx typeCheckCtx = new TypeCheckCtx(null); ExprNodeDesc defaultValExpr = TypeCheckProcFactory @@ -942,7 +944,8 @@ public abstract class BaseSemanticAnalyzer { } //get default value to be be stored in metastore -String defaultValueText = defaultValExpr.getExprString(); +String defaultValueText = tokenStream.toOriginalString(defaultValueAST.getTokenStartIndex(), + defaultValueAST.getTokenStopIndex()); final int DEFAULT_MAX_LEN = 255; if(defaultValueText.length() > DEFAULT_MAX_LEN) { throw new SemanticException( @@ -1026,7 +1029,7 @@ public abstract class BaseSemanticAnalyzer { rely = false; } else if( child.getToken().getType() == HiveParser.TOK_DEFAULT_VALUE){ // try to get default value only if this is DEFAULT constraint -checkOrDefaultValue = getDefaultValue(grandChild, typeChildForDefault); +checkOrDefaultValue = getDefaultValue(grandChild, typeChildForDefault, tokenRewriteStream); } else if(child.getToken().getType() == HiveParser.TOK_CHECK_CONSTRAINT) { checkOrDefaultValue = getCheckExpression(grandChild, tokenRewriteStream); @@ -1259,7 +1262,7 @@ public abstract class BaseSemanticAnalyzer { break; case HiveParser.TOK_DEFAULT_VALUE: processDefaultConstraints(catName, qualifiedTabName[0], qualifiedTabName[1], constraintChild, -ImmutableList.of(col.getName()), defaultConstraints, typeChild); +ImmutableList.of(col.getName()), defaultConstraints, typeChild, tokenRewriteStream); break; case HiveParser.TOK_NOT_NULL: processNotNullConstraints(catName, qualifiedTabName[0], qualifiedTabName[1], constraintChild, http://git-wip-us.apache.org/repos/asf/hive/blob/2e92451a/ql/src/java/org/apache/hadoop/hive/ql/parse/DDLSemanticAnalyzer.java -- diff --git a/ql/src/ja
hive git commit: HIVE-18857: Store default value text instead of default value expression in metastore(Vineet Garg, reviewed by Ashutosh Chauhan)
Repository: hive Updated Branches: refs/heads/branch-3 946f619e7 -> d1a935816 HIVE-18857: Store default value text instead of default value expression in metastore(Vineet Garg, reviewed by Ashutosh Chauhan) Project: http://git-wip-us.apache.org/repos/asf/hive/repo Commit: http://git-wip-us.apache.org/repos/asf/hive/commit/d1a93581 Tree: http://git-wip-us.apache.org/repos/asf/hive/tree/d1a93581 Diff: http://git-wip-us.apache.org/repos/asf/hive/diff/d1a93581 Branch: refs/heads/branch-3 Commit: d1a9358162f32ef489541cc72aef059e13349497 Parents: 946f619 Author: Vineet Garg Authored: Mon Apr 9 19:32:07 2018 -0700 Committer: Vineet Garg Committed: Mon Apr 9 19:32:07 2018 -0700 -- .../hadoop/hive/ql/parse/BaseSemanticAnalyzer.java | 15 +-- .../hadoop/hive/ql/parse/DDLSemanticAnalyzer.java| 3 ++- 2 files changed, 11 insertions(+), 7 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hive/blob/d1a93581/ql/src/java/org/apache/hadoop/hive/ql/parse/BaseSemanticAnalyzer.java -- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/parse/BaseSemanticAnalyzer.java b/ql/src/java/org/apache/hadoop/hive/ql/parse/BaseSemanticAnalyzer.java index 5301b2a..d940cdd 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/parse/BaseSemanticAnalyzer.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/parse/BaseSemanticAnalyzer.java @@ -765,10 +765,11 @@ public abstract class BaseSemanticAnalyzer { } protected static void processDefaultConstraints(String catName, String databaseName, String tableName, - ASTNode child, List columnNames, List defaultConstraints, final ASTNode typeChild) + ASTNode child, List columnNames, List defaultConstraints, final ASTNode typeChild, + final TokenRewriteStream tokenRewriteStream) throws SemanticException { List defaultInfos = new ArrayList(); -generateConstraintInfos(child, columnNames, defaultInfos, typeChild, null); +generateConstraintInfos(child, columnNames, defaultInfos, typeChild, tokenRewriteStream); constraintInfosToDefaultConstraints(catName, databaseName, tableName, defaultInfos, defaultConstraints); } @@ -930,7 +931,8 @@ public abstract class BaseSemanticAnalyzer { * @return retrieve the default value and return it as string * @throws SemanticException */ - private static String getDefaultValue(ASTNode defaultValueAST, ASTNode typeChild) throws SemanticException{ + private static String getDefaultValue(ASTNode defaultValueAST, ASTNode typeChild, +final TokenRewriteStream tokenStream) throws SemanticException{ // first create expression from defaultValueAST TypeCheckCtx typeCheckCtx = new TypeCheckCtx(null); ExprNodeDesc defaultValExpr = TypeCheckProcFactory @@ -942,7 +944,8 @@ public abstract class BaseSemanticAnalyzer { } //get default value to be be stored in metastore -String defaultValueText = defaultValExpr.getExprString(); +String defaultValueText = tokenStream.toOriginalString(defaultValueAST.getTokenStartIndex(), + defaultValueAST.getTokenStopIndex()); final int DEFAULT_MAX_LEN = 255; if(defaultValueText.length() > DEFAULT_MAX_LEN) { throw new SemanticException( @@ -1026,7 +1029,7 @@ public abstract class BaseSemanticAnalyzer { rely = false; } else if( child.getToken().getType() == HiveParser.TOK_DEFAULT_VALUE){ // try to get default value only if this is DEFAULT constraint -checkOrDefaultValue = getDefaultValue(grandChild, typeChildForDefault); +checkOrDefaultValue = getDefaultValue(grandChild, typeChildForDefault, tokenRewriteStream); } else if(child.getToken().getType() == HiveParser.TOK_CHECK_CONSTRAINT) { checkOrDefaultValue = getCheckExpression(grandChild, tokenRewriteStream); @@ -1259,7 +1262,7 @@ public abstract class BaseSemanticAnalyzer { break; case HiveParser.TOK_DEFAULT_VALUE: processDefaultConstraints(catName, qualifiedTabName[0], qualifiedTabName[1], constraintChild, -ImmutableList.of(col.getName()), defaultConstraints, typeChild); +ImmutableList.of(col.getName()), defaultConstraints, typeChild, tokenRewriteStream); break; case HiveParser.TOK_NOT_NULL: processNotNullConstraints(catName, qualifiedTabName[0], qualifiedTabName[1], constraintChild, http://git-wip-us.apache.org/repos/asf/hive/blob/d1a93581/ql/src/java/org/apache/hadoop/hive/ql/parse/DDLSemanticAnalyzer.java -- diff --git a/ql/sr
hive git commit: HIVE-19144 : TestSparkCliDriver:subquery_scalar - golden file needs to be udpated (Vineet Garg via Ashutosh Chauhan)
Repository: hive Updated Branches: refs/heads/master 2e92451a6 -> 65abf418a HIVE-19144 : TestSparkCliDriver:subquery_scalar - golden file needs to be udpated (Vineet Garg via Ashutosh Chauhan) Signed-off-by: Ashutosh Chauhan Project: http://git-wip-us.apache.org/repos/asf/hive/repo Commit: http://git-wip-us.apache.org/repos/asf/hive/commit/65abf418 Tree: http://git-wip-us.apache.org/repos/asf/hive/tree/65abf418 Diff: http://git-wip-us.apache.org/repos/asf/hive/diff/65abf418 Branch: refs/heads/master Commit: 65abf418ab6504b640dc60c1454e7642157e5667 Parents: 2e92451 Author: Vineet Garg Authored: Mon Apr 9 21:26:42 2018 -0700 Committer: Ashutosh Chauhan Committed: Mon Apr 9 21:26:42 2018 -0700 -- .../clientpositive/spark/subquery_scalar.q.out | 222 +-- 1 file changed, 107 insertions(+), 115 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hive/blob/65abf418/ql/src/test/results/clientpositive/spark/subquery_scalar.q.out -- diff --git a/ql/src/test/results/clientpositive/spark/subquery_scalar.q.out b/ql/src/test/results/clientpositive/spark/subquery_scalar.q.out index 7488f2e..d044da9 100644 --- a/ql/src/test/results/clientpositive/spark/subquery_scalar.q.out +++ b/ql/src/test/results/clientpositive/spark/subquery_scalar.q.out @@ -119,14 +119,14 @@ STAGE PLANS: outputColumnNames: p_size Statistics: Num rows: 1 Data size: 32560 Basic stats: COMPLETE Column stats: NONE Group By Operator - aggregations: avg(p_size) + aggregations: sum(p_size), count(p_size) mode: hash - outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 76 Basic stats: COMPLETE Column stats: NONE + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 16 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator sort order: -Statistics: Num rows: 1 Data size: 76 Basic stats: COMPLETE Column stats: NONE -value expressions: _col0 (type: struct) +Statistics: Num rows: 1 Data size: 16 Basic stats: COMPLETE Column stats: NONE +value expressions: _col0 (type: bigint), _col1 (type: bigint) Reducer 2 Reduce Operator Tree: Join Operator @@ -136,17 +136,17 @@ STAGE PLANS: 0 1 outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9 -Statistics: Num rows: 26 Data size: 5149 Basic stats: COMPLETE Column stats: NONE +Statistics: Num rows: 26 Data size: 3589 Basic stats: COMPLETE Column stats: NONE Filter Operator predicate: (UDFToDouble(_col5) > _col9) (type: boolean) - Statistics: Num rows: 8 Data size: 1584 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 8 Data size: 1104 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 -Statistics: Num rows: 8 Data size: 1584 Basic stats: COMPLETE Column stats: NONE +Statistics: Num rows: 8 Data size: 1104 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false - Statistics: Num rows: 8 Data size: 1584 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 8 Data size: 1104 Basic stats: COMPLETE Column stats: NONE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat @@ -154,14 +154,18 @@ STAGE PLANS: Reducer 4 Reduce Operator Tree: Group By Operator -aggregations: avg(VALUE._col0) +aggregations: sum(VALUE._col0), count(VALUE._col1) mode: mergepartial -outputColumnNames: _col0 -Statistics: Num rows: 1 Data size: 76 Basic stats: COMPLETE Column stats: NONE -Reduce Output Operator - sort order: - Statistics: N
hive git commit: HIVE-19145 : Stabilize statsoptimizer.q test
Repository: hive Updated Branches: refs/heads/master 65abf418a -> 91a0cb8fa HIVE-19145 : Stabilize statsoptimizer.q test Signed-off-by: Ashutosh Chauhan Project: http://git-wip-us.apache.org/repos/asf/hive/repo Commit: http://git-wip-us.apache.org/repos/asf/hive/commit/91a0cb8f Tree: http://git-wip-us.apache.org/repos/asf/hive/tree/91a0cb8f Diff: http://git-wip-us.apache.org/repos/asf/hive/diff/91a0cb8f Branch: refs/heads/master Commit: 91a0cb8fa34d8fcc2f1086209e6f6986e54eb95e Parents: 65abf41 Author: Ashutosh Chauhan Authored: Mon Apr 9 21:28:08 2018 -0700 Committer: Ashutosh Chauhan Committed: Mon Apr 9 21:28:08 2018 -0700 -- .../test/queries/clientpositive/statsoptimizer.q | 8 .../results/clientpositive/statsoptimizer.q.out | 18 +- 2 files changed, 13 insertions(+), 13 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hive/blob/91a0cb8f/ql/src/test/queries/clientpositive/statsoptimizer.q -- diff --git a/ql/src/test/queries/clientpositive/statsoptimizer.q b/ql/src/test/queries/clientpositive/statsoptimizer.q index 428b741..a3dbe4e 100644 --- a/ql/src/test/queries/clientpositive/statsoptimizer.q +++ b/ql/src/test/queries/clientpositive/statsoptimizer.q @@ -2,13 +2,13 @@ set hive.cbo.enable=false; set hive.compute.query.using.stats=true; EXPLAIN -SELECT to_date(current_date()) as GROUP_BY_FIELD, count (*) as src_cnt +SELECT round(year(to_date(current_date())),-3) as GROUP_BY_FIELD, count (*) as src_cnt from src WHERE 1=1 -group by to_date(current_date()); +group by round(year(to_date(current_date())),-3); -SELECT to_date(current_date()) as GROUP_BY_FIELD, count (*) as src_cnt +SELECT round(year(to_date(current_date())),-3) as GROUP_BY_FIELD, count (*) as src_cnt from src WHERE 1=1 -group by to_date(current_date()); +group by round(year(to_date(current_date())),-3); http://git-wip-us.apache.org/repos/asf/hive/blob/91a0cb8f/ql/src/test/results/clientpositive/statsoptimizer.q.out -- diff --git a/ql/src/test/results/clientpositive/statsoptimizer.q.out b/ql/src/test/results/clientpositive/statsoptimizer.q.out index 1f97de3..35330d9 100644 --- a/ql/src/test/results/clientpositive/statsoptimizer.q.out +++ b/ql/src/test/results/clientpositive/statsoptimizer.q.out @@ -1,14 +1,14 @@ PREHOOK: query: EXPLAIN -SELECT to_date(current_date()) as GROUP_BY_FIELD, count (*) as src_cnt +SELECT round(year(to_date(current_date())),-3) as GROUP_BY_FIELD, count (*) as src_cnt from src WHERE 1=1 -group by to_date(current_date()) +group by round(year(to_date(current_date())),-3) PREHOOK: type: QUERY POSTHOOK: query: EXPLAIN -SELECT to_date(current_date()) as GROUP_BY_FIELD, count (*) as src_cnt +SELECT round(year(to_date(current_date())),-3) as GROUP_BY_FIELD, count (*) as src_cnt from src WHERE 1=1 -group by to_date(current_date()) +group by round(year(to_date(current_date())),-3) POSTHOOK: type: QUERY STAGE DEPENDENCIES: Stage-0 is a root stage @@ -20,18 +20,18 @@ STAGE PLANS: Processor Tree: ListSink -PREHOOK: query: SELECT to_date(current_date()) as GROUP_BY_FIELD, count (*) as src_cnt +PREHOOK: query: SELECT round(year(to_date(current_date())),-3) as GROUP_BY_FIELD, count (*) as src_cnt from src WHERE 1=1 -group by to_date(current_date()) +group by round(year(to_date(current_date())),-3) PREHOOK: type: QUERY PREHOOK: Input: default@src A masked pattern was here -POSTHOOK: query: SELECT to_date(current_date()) as GROUP_BY_FIELD, count (*) as src_cnt +POSTHOOK: query: SELECT round(year(to_date(current_date())),-3) as GROUP_BY_FIELD, count (*) as src_cnt from src WHERE 1=1 -group by to_date(current_date()) +group by round(year(to_date(current_date())),-3) POSTHOOK: type: QUERY POSTHOOK: Input: default@src A masked pattern was here -2018-04-04 500 +2000 500
hive git commit: HIVE-19146 : Delete dangling q.out
Repository: hive Updated Branches: refs/heads/master 91a0cb8fa -> dcd9b5941 HIVE-19146 : Delete dangling q.out Signed-off-by: Ashutosh Chauhan Project: http://git-wip-us.apache.org/repos/asf/hive/repo Commit: http://git-wip-us.apache.org/repos/asf/hive/commit/dcd9b594 Tree: http://git-wip-us.apache.org/repos/asf/hive/tree/dcd9b594 Diff: http://git-wip-us.apache.org/repos/asf/hive/diff/dcd9b594 Branch: refs/heads/master Commit: dcd9b5941c53d5219eb35f9f682ba2da69288291 Parents: 91a0cb8 Author: Ashutosh Chauhan Authored: Mon Apr 9 21:30:34 2018 -0700 Committer: Ashutosh Chauhan Committed: Mon Apr 9 21:30:34 2018 -0700 -- .../materialized_view_create.q.out | 288 --- 1 file changed, 288 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hive/blob/dcd9b594/ql/src/test/results/clientpositive/materialized_view_create.q.out -- diff --git a/ql/src/test/results/clientpositive/materialized_view_create.q.out b/ql/src/test/results/clientpositive/materialized_view_create.q.out deleted file mode 100644 index 8952304..000 --- a/ql/src/test/results/clientpositive/materialized_view_create.q.out +++ /dev/null @@ -1,288 +0,0 @@ -PREHOOK: query: create table cmv_basetable (a int, b varchar(256), c decimal(10,2)) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@cmv_basetable -POSTHOOK: query: create table cmv_basetable (a int, b varchar(256), c decimal(10,2)) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@cmv_basetable -PREHOOK: query: insert into cmv_basetable values (1, 'alfred', 10.30),(2, 'bob', 3.14),(2, 'bonnie', 172342.2),(3, 'calvin', 978.76),(3, 'charlie', 9.8) -PREHOOK: type: QUERY -PREHOOK: Input: _dummy_database@_dummy_table -PREHOOK: Output: default@cmv_basetable -POSTHOOK: query: insert into cmv_basetable values (1, 'alfred', 10.30),(2, 'bob', 3.14),(2, 'bonnie', 172342.2),(3, 'calvin', 978.76),(3, 'charlie', 9.8) -POSTHOOK: type: QUERY -POSTHOOK: Input: _dummy_database@_dummy_table -POSTHOOK: Output: default@cmv_basetable -POSTHOOK: Lineage: cmv_basetable.a SCRIPT [] -POSTHOOK: Lineage: cmv_basetable.b SCRIPT [] -POSTHOOK: Lineage: cmv_basetable.c SCRIPT [] -PREHOOK: query: create materialized view cmv_mat_view as select a, b, c from cmv_basetable -PREHOOK: type: CREATE_MATERIALIZED_VIEW -PREHOOK: Input: default@cmv_basetable -PREHOOK: Output: database:default -PREHOOK: Output: default@cmv_mat_view -POSTHOOK: query: create materialized view cmv_mat_view as select a, b, c from cmv_basetable -POSTHOOK: type: CREATE_MATERIALIZED_VIEW -POSTHOOK: Input: default@cmv_basetable -POSTHOOK: Output: database:default -POSTHOOK: Output: default@cmv_mat_view -PREHOOK: query: desc formatted cmv_mat_view -PREHOOK: type: DESCTABLE -PREHOOK: Input: default@cmv_mat_view -POSTHOOK: query: desc formatted cmv_mat_view -POSTHOOK: type: DESCTABLE -POSTHOOK: Input: default@cmv_mat_view -# col_name data_type comment -a int -b varchar(256) -c decimal(10,2) - -# Detailed Table Information -Database: default - A masked pattern was here -Retention: 0 - A masked pattern was here -Table Type:MATERIALIZED_VIEW -Table Parameters: - COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"} - numFiles1 - numRows 5 - rawDataSize 1025 - totalSize 497 - A masked pattern was here - -# Storage Information -SerDe Library: org.apache.hadoop.hive.ql.io.orc.OrcSerde -InputFormat: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat -OutputFormat: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat -Compressed:No -Num Buckets: -1 -Bucket Columns:[] -Sort Columns: [] - -# View Information -View Original Text:select a, b, c from cmv_basetable -View Expanded Text:select `cmv_basetable`.`a`, `cmv_basetable`.`b`, `cmv_basetable`.`c` from `default`.`cmv_basetable` -View Rewrite Enabled: No -PREHOOK: query: select * from cmv_mat_view -PREHOOK: type: QUERY -PREHOOK: Input: default@cmv_mat_view - A masked pattern was here -POSTHOOK: query: sel
hive git commit: HIVE-18783: ALTER TABLE post-commit listener does not include the transactional listener responses (Sergio Pena, reviewed by Vihang Karajgaonkar)
Repository: hive Updated Branches: refs/heads/master a263f0831 -> b290468c0 HIVE-18783: ALTER TABLE post-commit listener does not include the transactional listener responses (Sergio Pena, reviewed by Vihang Karajgaonkar) Project: http://git-wip-us.apache.org/repos/asf/hive/repo Commit: http://git-wip-us.apache.org/repos/asf/hive/commit/b290468c Tree: http://git-wip-us.apache.org/repos/asf/hive/tree/b290468c Diff: http://git-wip-us.apache.org/repos/asf/hive/diff/b290468c Branch: refs/heads/master Commit: b290468c0ffe53daa76fad1e2a063d4596ea2ece Parents: a263f08 Author: Sergio Pena Authored: Mon Apr 9 10:04:43 2018 -0500 Committer: Sergio Pena Committed: Mon Apr 9 10:04:43 2018 -0500 -- .../listener/TestDbNotificationListener.java| 1 + .../hadoop/hive/metastore/HiveAlterHandler.java | 59 ++-- .../hadoop/hive/metastore/HiveMetaStore.java| 30 ++ .../hadoop/hive/metastore/IHMSHandler.java | 6 ++ 4 files changed, 65 insertions(+), 31 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hive/blob/b290468c/itests/hcatalog-unit/src/test/java/org/apache/hive/hcatalog/listener/TestDbNotificationListener.java -- diff --git a/itests/hcatalog-unit/src/test/java/org/apache/hive/hcatalog/listener/TestDbNotificationListener.java b/itests/hcatalog-unit/src/test/java/org/apache/hive/hcatalog/listener/TestDbNotificationListener.java index 823312b..70c6a94 100644 --- a/itests/hcatalog-unit/src/test/java/org/apache/hive/hcatalog/listener/TestDbNotificationListener.java +++ b/itests/hcatalog-unit/src/test/java/org/apache/hive/hcatalog/listener/TestDbNotificationListener.java @@ -472,6 +472,7 @@ public class TestDbNotificationListener { assertEquals(TableType.MANAGED_TABLE.toString(), alterTableMessage.getTableType()); // Verify the eventID was passed to the non-transactional listener +MockMetaStoreEventListener.popAndVerifyLastEventId(EventType.ALTER_TABLE, firstEventId + 2); MockMetaStoreEventListener.popAndVerifyLastEventId(EventType.CREATE_TABLE, firstEventId + 1); // When hive.metastore.transactional.event.listeners is set, http://git-wip-us.apache.org/repos/asf/hive/blob/b290468c/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/HiveAlterHandler.java -- diff --git a/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/HiveAlterHandler.java b/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/HiveAlterHandler.java index ed1b8c5..60bed98 100644 --- a/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/HiveAlterHandler.java +++ b/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/HiveAlterHandler.java @@ -55,6 +55,7 @@ import org.apache.hadoop.hive.metastore.api.hive_metastoreConstants; import java.io.IOException; import java.net.URI; import java.util.ArrayList; +import java.util.Collections; import java.util.HashMap; import java.util.Iterator; import java.util.List; @@ -122,15 +123,24 @@ public class HiveAlterHandler implements AlterHandler { boolean success = false; boolean dataWasMoved = false; -Table oldt; +boolean isPartitionedTable = false; + +Table oldt = null; + List transactionalListeners = null; +List listeners = null; +Map txnAlterTableEventResponses = Collections.emptyMap(); +Map txnDropTableEventResponses = Collections.emptyMap(); +Map txnCreateTableEventResponses = Collections.emptyMap(); +Map txnAddPartitionEventResponses = Collections.emptyMap(); + if (handler != null) { transactionalListeners = handler.getTransactionalListeners(); + listeners = handler.getListeners(); } try { boolean rename = false; - boolean isPartitionedTable = false; List parts; // Switching tables between catalogs is not allowed. @@ -337,23 +347,23 @@ public class HiveAlterHandler implements AlterHandler { if (transactionalListeners != null && !transactionalListeners.isEmpty()) { if (oldt.getDbName().equalsIgnoreCase(newt.getDbName())) { - MetaStoreListenerNotifier.notifyEvent(transactionalListeners, + txnAlterTableEventResponses = MetaStoreListenerNotifier.notifyEvent(transactionalListeners, EventMessage.EventType.ALTER_TABLE, new AlterTableEvent(oldt, newt, false, true, handler), environmentContext); } else { - MetaStoreListenerNotifier.notifyEvent(transactionalListeners, + txnDropTableEventResponses = MetaStoreListenerNotifier.notifyEvent(transactionalListeners, EventMessage.EventType.DROP_TABLE, new DropTableEvent(oldt
hive git commit: HIVE-18859 : Incorrect handling of thrift metastore exceptions (Ganesha Shreedhara via Ashutosh Chauhan)
Repository: hive Updated Branches: refs/heads/master b290468c0 -> 959e77257 HIVE-18859 : Incorrect handling of thrift metastore exceptions (Ganesha Shreedhara via Ashutosh Chauhan) Signed-off-by: Ashutosh Chauhan Project: http://git-wip-us.apache.org/repos/asf/hive/repo Commit: http://git-wip-us.apache.org/repos/asf/hive/commit/959e7725 Tree: http://git-wip-us.apache.org/repos/asf/hive/tree/959e7725 Diff: http://git-wip-us.apache.org/repos/asf/hive/diff/959e7725 Branch: refs/heads/master Commit: 959e77257a006b36769cffd9efb01dd16b139474 Parents: b290468 Author: Ganesha Shreedhara Authored: Mon Mar 12 03:06:00 2018 -0700 Committer: Ashutosh Chauhan Committed: Mon Apr 9 08:06:58 2018 -0700 -- .../AbstractTestAuthorizationApiAuthorizer.java | 19 +++-- .../hadoop/hive/metastore/HiveMetaStore.java| 30 2 files changed, 40 insertions(+), 9 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hive/blob/959e7725/itests/hive-unit/src/test/java/org/apache/hadoop/hive/metastore/AbstractTestAuthorizationApiAuthorizer.java -- diff --git a/itests/hive-unit/src/test/java/org/apache/hadoop/hive/metastore/AbstractTestAuthorizationApiAuthorizer.java b/itests/hive-unit/src/test/java/org/apache/hadoop/hive/metastore/AbstractTestAuthorizationApiAuthorizer.java index abd5e32..69692d0 100644 --- a/itests/hive-unit/src/test/java/org/apache/hadoop/hive/metastore/AbstractTestAuthorizationApiAuthorizer.java +++ b/itests/hive-unit/src/test/java/org/apache/hadoop/hive/metastore/AbstractTestAuthorizationApiAuthorizer.java @@ -35,6 +35,7 @@ import org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge; import org.apache.hadoop.hive.metastore.utils.MetaStoreUtils; import org.apache.hadoop.hive.ql.security.authorization.MetaStoreAuthzAPIAuthorizerEmbedOnly; import org.apache.hadoop.hive.ql.security.authorization.AuthorizationPreEventListener; +import org.apache.thrift.TException; import org.junit.Test; /** @@ -91,15 +92,27 @@ public abstract class AbstractTestAuthorizationApiAuthorizer { // authorization checks passed. String exStackString = ExceptionUtils.getStackTrace(e); assertTrue("Verifying this exception came after authorization check", - exStackString.contains("org.apache.hadoop.hive.metastore.ObjectStore")); + exStackString.contains("org.apache.hadoop.hive.metastore.ObjectStore")); // If its not an exception caused by auth check, ignore it } assertFalse("Authz Exception should have been thrown in remote mode", isRemoteMetastoreMode); System.err.println("No auth exception thrown"); } catch (MetaException e) { System.err.println("Caught exception"); - caughtEx = true; - assertTrue(e.getMessage().contains(MetaStoreAuthzAPIAuthorizerEmbedOnly.errMsg)); + String exStackString = ExceptionUtils.getStackTrace(e); + // Check if MetaException has one of InvalidObjectException or NoSuchObjectExcetion or any exception thrown from ObjectStore , which means that the + // authorization checks passed. + if(exStackString.contains("org.apache.hadoop.hive.metastore.api.NoSuchObjectException") || + exStackString.contains("org.apache.hadoop.hive.metastore.api.InvalidObjectException")) { +assertFalse("No Authz exception thrown in embedded mode", isRemoteMetastoreMode); + } else { +caughtEx = true; + assertTrue(e.getMessage().contains(MetaStoreAuthzAPIAuthorizerEmbedOnly.errMsg)); + } +} catch (TException e) { + String exStackString = ExceptionUtils.getStackTrace(e); + assertTrue("Verifying this exception came after authorization check", + exStackString.contains("org.apache.hadoop.hive.metastore.ObjectStore")); } if (!isRemoteMetastoreMode) { assertFalse("No exception should be thrown in embedded mode", caughtEx); http://git-wip-us.apache.org/repos/asf/hive/blob/959e7725/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/HiveMetaStore.java -- diff --git a/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/HiveMetaStore.java b/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/HiveMetaStore.java index 102e5b4..a2fe7d7 100644 --- a/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/HiveMetaStore.java +++ b/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/HiveMetaStore.java @@ -5981,8 +5981,11 @@ public class HiveMetaStore extends ThriftHiveMetastore { ret = ms.grantRole(role, principalName, principalType, grantor, grantorType, grantOption); } catch (MetaException e) { thro
hive git commit: HIVE-18783: ALTER TABLE post-commit listener does not include the transactional listener responses (Sergio Pena, reviewed by Vihang Karajgaonkar)
Repository: hive Updated Branches: refs/heads/branch-2 941f037c2 -> 7cea4d0da HIVE-18783: ALTER TABLE post-commit listener does not include the transactional listener responses (Sergio Pena, reviewed by Vihang Karajgaonkar) Project: http://git-wip-us.apache.org/repos/asf/hive/repo Commit: http://git-wip-us.apache.org/repos/asf/hive/commit/7cea4d0d Tree: http://git-wip-us.apache.org/repos/asf/hive/tree/7cea4d0d Diff: http://git-wip-us.apache.org/repos/asf/hive/diff/7cea4d0d Branch: refs/heads/branch-2 Commit: 7cea4d0da85b6f4a23fefb1ebb43c442b989c58d Parents: 941f037 Author: Sergio Pena Authored: Mon Apr 9 10:26:10 2018 -0500 Committer: Sergio Pena Committed: Mon Apr 9 10:26:10 2018 -0500 -- .../hcatalog/listener/TestDbNotificationListener.java | 1 + .../apache/hadoop/hive/metastore/HiveAlterHandler.java | 13 - .../apache/hadoop/hive/metastore/HiveMetaStore.java| 10 -- 3 files changed, 17 insertions(+), 7 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hive/blob/7cea4d0d/itests/hcatalog-unit/src/test/java/org/apache/hive/hcatalog/listener/TestDbNotificationListener.java -- diff --git a/itests/hcatalog-unit/src/test/java/org/apache/hive/hcatalog/listener/TestDbNotificationListener.java b/itests/hcatalog-unit/src/test/java/org/apache/hive/hcatalog/listener/TestDbNotificationListener.java index 5a40780..8e9eb3d 100644 --- a/itests/hcatalog-unit/src/test/java/org/apache/hive/hcatalog/listener/TestDbNotificationListener.java +++ b/itests/hcatalog-unit/src/test/java/org/apache/hive/hcatalog/listener/TestDbNotificationListener.java @@ -447,6 +447,7 @@ public class TestDbNotificationListener { assertEquals(TableType.MANAGED_TABLE.toString(), alterTableMessage.getTableType()); // Verify the eventID was passed to the non-transactional listener +MockMetaStoreEventListener.popAndVerifyLastEventId(EventType.ALTER_TABLE, firstEventId + 2); MockMetaStoreEventListener.popAndVerifyLastEventId(EventType.CREATE_TABLE, firstEventId + 1); // When hive.metastore.transactional.event.listeners is set, http://git-wip-us.apache.org/repos/asf/hive/blob/7cea4d0d/metastore/src/java/org/apache/hadoop/hive/metastore/HiveAlterHandler.java -- diff --git a/metastore/src/java/org/apache/hadoop/hive/metastore/HiveAlterHandler.java b/metastore/src/java/org/apache/hadoop/hive/metastore/HiveAlterHandler.java index 3e7c59b..83c68a2 100644 --- a/metastore/src/java/org/apache/hadoop/hive/metastore/HiveAlterHandler.java +++ b/metastore/src/java/org/apache/hadoop/hive/metastore/HiveAlterHandler.java @@ -54,6 +54,7 @@ import org.apache.hive.common.util.HiveStringUtils; import java.io.IOException; import java.net.URI; import java.util.ArrayList; +import java.util.Collections; import java.util.HashMap; import java.util.Iterator; import java.util.List; @@ -126,8 +127,12 @@ public class HiveAlterHandler implements AlterHandler { boolean dataWasMoved = false; Table oldt = null; List transactionalListeners = null; +List listeners = null; +Map txnAlterTableEventResponses = Collections.emptyMap(); + if (handler != null) { transactionalListeners = handler.getTransactionalListeners(); + listeners = handler.getListeners(); } try { @@ -309,7 +314,7 @@ public class HiveAlterHandler implements AlterHandler { } if (transactionalListeners != null && !transactionalListeners.isEmpty()) { -MetaStoreListenerNotifier.notifyEvent(transactionalListeners, +txnAlterTableEventResponses = MetaStoreListenerNotifier.notifyEvent(transactionalListeners, EventMessage.EventType.ALTER_TABLE, new AlterTableEvent(oldt, newt, true, handler), environmentContext); @@ -349,6 +354,12 @@ public class HiveAlterHandler implements AlterHandler { } } } + + if (!listeners.isEmpty()) { +MetaStoreListenerNotifier.notifyEvent(listeners, EventMessage.EventType.ALTER_TABLE, +new AlterTableEvent(oldt, newt, success, handler), +environmentContext, txnAlterTableEventResponses, msdb); + } } } http://git-wip-us.apache.org/repos/asf/hive/blob/7cea4d0d/metastore/src/java/org/apache/hadoop/hive/metastore/HiveMetaStore.java -- diff --git a/metastore/src/java/org/apache/hadoop/hive/metastore/HiveMetaStore.java b/metastore/src/java/org/apache/hadoop/hive/metastore/HiveMetaStore.java index d32e89c..aa233dd 100644 --- a/metastore/src/java/org/apache/hadoop/hive/metastore/HiveMetaStor
hive git commit: HIVE-18783: ALTER TABLE post-commit listener does not include the transactional listener responses (Sergio Pena, reviewed by Vihang Karajgaonkar)
Repository: hive Updated Branches: refs/heads/branch-2.3 d3fa8e0c5 -> f78128a71 HIVE-18783: ALTER TABLE post-commit listener does not include the transactional listener responses (Sergio Pena, reviewed by Vihang Karajgaonkar) Project: http://git-wip-us.apache.org/repos/asf/hive/repo Commit: http://git-wip-us.apache.org/repos/asf/hive/commit/f78128a7 Tree: http://git-wip-us.apache.org/repos/asf/hive/tree/f78128a7 Diff: http://git-wip-us.apache.org/repos/asf/hive/diff/f78128a7 Branch: refs/heads/branch-2.3 Commit: f78128a719772c1157ce6f0b9bce52a0cac0893c Parents: d3fa8e0 Author: Sergio Pena Authored: Mon Apr 9 10:26:10 2018 -0500 Committer: Sergio Pena Committed: Mon Apr 9 11:05:50 2018 -0500 -- .../hcatalog/listener/TestDbNotificationListener.java | 1 + .../hadoop/hive/metastore/HiveAlterHandler.java | 14 +- .../apache/hadoop/hive/metastore/HiveMetaStore.java | 10 -- 3 files changed, 18 insertions(+), 7 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hive/blob/f78128a7/itests/hcatalog-unit/src/test/java/org/apache/hive/hcatalog/listener/TestDbNotificationListener.java -- diff --git a/itests/hcatalog-unit/src/test/java/org/apache/hive/hcatalog/listener/TestDbNotificationListener.java b/itests/hcatalog-unit/src/test/java/org/apache/hive/hcatalog/listener/TestDbNotificationListener.java index 976c3c5..786c4c2 100644 --- a/itests/hcatalog-unit/src/test/java/org/apache/hive/hcatalog/listener/TestDbNotificationListener.java +++ b/itests/hcatalog-unit/src/test/java/org/apache/hive/hcatalog/listener/TestDbNotificationListener.java @@ -444,6 +444,7 @@ public class TestDbNotificationListener { assertEquals(table, alterTableMessage.getTableObjAfter()); // Verify the eventID was passed to the non-transactional listener +MockMetaStoreEventListener.popAndVerifyLastEventId(EventType.ALTER_TABLE, firstEventId + 2); MockMetaStoreEventListener.popAndVerifyLastEventId(EventType.CREATE_TABLE, firstEventId + 1); // When hive.metastore.transactional.event.listeners is set, http://git-wip-us.apache.org/repos/asf/hive/blob/f78128a7/metastore/src/java/org/apache/hadoop/hive/metastore/HiveAlterHandler.java -- diff --git a/metastore/src/java/org/apache/hadoop/hive/metastore/HiveAlterHandler.java b/metastore/src/java/org/apache/hadoop/hive/metastore/HiveAlterHandler.java index 15f2597..7730a24 100644 --- a/metastore/src/java/org/apache/hadoop/hive/metastore/HiveAlterHandler.java +++ b/metastore/src/java/org/apache/hadoop/hive/metastore/HiveAlterHandler.java @@ -54,8 +54,10 @@ import org.apache.hive.common.util.HiveStringUtils; import java.io.IOException; import java.net.URI; import java.util.ArrayList; +import java.util.Collections; import java.util.Iterator; import java.util.List; +import java.util.Map; /** * Hive specific implementation of alter @@ -115,8 +117,12 @@ public class HiveAlterHandler implements AlterHandler { boolean rename = false; Table oldt = null; List transactionalListeners = null; +List listeners = null; +Map txnAlterTableEventResponses = Collections.emptyMap(); + if (handler != null) { transactionalListeners = handler.getTransactionalListeners(); + listeners = handler.getListeners(); } try { @@ -270,7 +276,7 @@ public class HiveAlterHandler implements AlterHandler { alterTableUpdateTableColumnStats(msdb, oldt, newt); if (transactionalListeners != null && !transactionalListeners.isEmpty()) { -MetaStoreListenerNotifier.notifyEvent(transactionalListeners, +txnAlterTableEventResponses = MetaStoreListenerNotifier.notifyEvent(transactionalListeners, EventMessage.EventType.ALTER_TABLE, new AlterTableEvent(oldt, newt, true, handler), environmentContext); @@ -305,6 +311,12 @@ public class HiveAlterHandler implements AlterHandler { } } } + + if (!listeners.isEmpty()) { +MetaStoreListenerNotifier.notifyEvent(listeners, EventMessage.EventType.ALTER_TABLE, +new AlterTableEvent(oldt, newt, success, handler), +environmentContext, txnAlterTableEventResponses, msdb); + } } } http://git-wip-us.apache.org/repos/asf/hive/blob/f78128a7/metastore/src/java/org/apache/hadoop/hive/metastore/HiveMetaStore.java -- diff --git a/metastore/src/java/org/apache/hadoop/hive/metastore/HiveMetaStore.java b/metastore/src/java/org/apache/hadoop/hive/metastore/HiveMetaStore.java index befd15e..d799acf 100644 ---
hive git commit: HIVE-19129: Support DEFAULT keyword with MERGE(Vineet Garg, reviewed by Ashutosh Chauhan)
Repository: hive Updated Branches: refs/heads/master 959e77257 -> e7c15d234 HIVE-19129: Support DEFAULT keyword with MERGE(Vineet Garg, reviewed by Ashutosh Chauhan) Project: http://git-wip-us.apache.org/repos/asf/hive/repo Commit: http://git-wip-us.apache.org/repos/asf/hive/commit/e7c15d23 Tree: http://git-wip-us.apache.org/repos/asf/hive/tree/e7c15d23 Diff: http://git-wip-us.apache.org/repos/asf/hive/diff/e7c15d23 Branch: refs/heads/master Commit: e7c15d2348602b6890aff24a088c2b04fb46af8a Parents: 959e772 Author: Vineet Garg Authored: Mon Apr 9 10:51:46 2018 -0700 Committer: Vineet Garg Committed: Mon Apr 9 10:51:46 2018 -0700 -- .../hadoop/hive/ql/parse/SemanticAnalyzer.java | 26 +- .../ql/parse/UpdateDeleteSemanticAnalyzer.java | 1 + .../insert_into_default_keyword.q | 49 ++ .../llap/insert_into_default_keyword.q.out | 787 +++ 4 files changed, 861 insertions(+), 2 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hive/blob/e7c15d23/ql/src/java/org/apache/hadoop/hive/ql/parse/SemanticAnalyzer.java -- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/parse/SemanticAnalyzer.java b/ql/src/java/org/apache/hadoop/hive/ql/parse/SemanticAnalyzer.java index ff0a2e6..3b74aba 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/parse/SemanticAnalyzer.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/parse/SemanticAnalyzer.java @@ -665,13 +665,13 @@ public class SemanticAnalyzer extends BaseSemanticAnalyzer { /** * This method creates a list of default constraints which corresponds to - * given schema (taretSchema) or target table's column schema (if targetSchema is null) + * given schema (targetSchema) or target table's column schema (if targetSchema is null) * @param tbl * @param targetSchema * @return List of default constraints (including NULL if there is no default) * @throws SemanticException */ - private List getDefaultConstraints(Table tbl, List targetSchema) throws SemanticException{ + private static List getDefaultConstraints(Table tbl, List targetSchema) throws SemanticException{ Map colNameToDefaultVal = null; try { DefaultConstraint dc = Hive.get().getEnabledDefaultConstraints(tbl.getDbName(), tbl.getTableName()); @@ -718,6 +718,28 @@ public class SemanticAnalyzer extends BaseSemanticAnalyzer { return newNode; } + public static String replaceDefaultKeywordForMerge(String valueClause,Table targetTable) + throws SemanticException { +List defaultConstraints = null; +String[] values = valueClause.trim().split(","); +StringBuilder newValueClause = new StringBuilder(); +for (int i = 0; i < values.length; i++) { + if (values[i].trim().toLowerCase().equals("`default`")) { +if (defaultConstraints == null) { + defaultConstraints = getDefaultConstraints(targetTable, null); +} +newValueClause.append(defaultConstraints.get(i)); + } + else { +newValueClause.append(values[i]); + } + if(i != values.length-1) { +newValueClause.append(","); + } +} +return newValueClause.toString(); + } + /** * This method replaces ASTNode corresponding to DEFAULT keyword with either DEFAULT constraint * expression if exists or NULL otherwise http://git-wip-us.apache.org/repos/asf/hive/blob/e7c15d23/ql/src/java/org/apache/hadoop/hive/ql/parse/UpdateDeleteSemanticAnalyzer.java -- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/parse/UpdateDeleteSemanticAnalyzer.java b/ql/src/java/org/apache/hadoop/hive/ql/parse/UpdateDeleteSemanticAnalyzer.java index a660747..0effd92 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/parse/UpdateDeleteSemanticAnalyzer.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/parse/UpdateDeleteSemanticAnalyzer.java @@ -1101,6 +1101,7 @@ public class UpdateDeleteSemanticAnalyzer extends SemanticAnalyzer { List partCols = targetTable.getPartCols(); String valuesClause = getMatchedText((ASTNode)getWhenClauseOperation(whenNotMatchedClause).getChild(0)); valuesClause = valuesClause.substring(1, valuesClause.length() - 1);//strip '(' and ')' +valuesClause = SemanticAnalyzer.replaceDefaultKeywordForMerge(valuesClause, targetTable); rewrittenQueryStr.append("INSERT INTO ").append(getFullTableNameForSQL(target)); addPartitionColsToInsert(partCols, rewrittenQueryStr); http://git-wip-us.apache.org/repos/asf/hive/blob/e7c15d23/ql/src/test/queries/clientpositive/insert_into_default_keyword.q -- diff --git a/ql/src/test/queries/clientpositive/insert_into_default_keyword.q b/ql/src/test/queries/c
[2/3] hive git commit: HIVE-19074: Vectorization: Add llap vectorization_div0.q.out Q output file (missing changes) (Matt McCline, reviewed by Teddy Choi)
http://git-wip-us.apache.org/repos/asf/hive/blob/a21302f1/ql/src/test/results/clientpositive/llap/vectorization_div0.q.out -- diff --git a/ql/src/test/results/clientpositive/llap/vectorization_div0.q.out b/ql/src/test/results/clientpositive/llap/vectorization_div0.q.out new file mode 100644 index 000..e00dcd9 --- /dev/null +++ b/ql/src/test/results/clientpositive/llap/vectorization_div0.q.out @@ -0,0 +1,870 @@ +PREHOOK: query: explain vectorization expression +select cint, cint / 0 as cint_div, ctinyint, ctinyint / 0 as ctinyint_div, cbigint, cbigint / 0 as cbigint_div, cdouble, cdouble / 0.0 as cdouble_div +from alltypesorc order by cint, ctinyint, cbigint, cdouble limit 100 +PREHOOK: type: QUERY +POSTHOOK: query: explain vectorization expression +select cint, cint / 0 as cint_div, ctinyint, ctinyint / 0 as ctinyint_div, cbigint, cbigint / 0 as cbigint_div, cdouble, cdouble / 0.0 as cdouble_div +from alltypesorc order by cint, ctinyint, cbigint, cdouble limit 100 +POSTHOOK: type: QUERY +Explain +PLAN VECTORIZATION: + enabled: true + enabledConditionsMet: [hive.vectorized.execution.enabled IS true] + +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 +Tez + A masked pattern was here + Edges: +Reducer 2 <- Map 1 (SIMPLE_EDGE) + A masked pattern was here + Vertices: +Map 1 +Map Operator Tree: +TableScan + alias: alltypesorc + Statistics: Num rows: 12288 Data size: 220184 Basic stats: COMPLETE Column stats: COMPLETE + TableScan Vectorization: + native: true + Select Operator +expressions: cint (type: int), (UDFToDouble(cint) / 0.0D) (type: double), ctinyint (type: tinyint), (UDFToDouble(ctinyint) / 0.0D) (type: double), cbigint (type: bigint), (UDFToDouble(cbigint) / 0.0D) (type: double), cdouble (type: double), (cdouble / 0.0D) (type: double) +outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7 +Select Vectorization: +className: VectorSelectOperator +native: true +projectedOutputColumnNums: [2, 14, 0, 15, 3, 16, 5, 13] +selectExpressions: DoubleColDivideDoubleScalar(col 13:double, val 0.0)(children: CastLongToDouble(col 2:int) -> 13:double) -> 14:double, DoubleColDivideDoubleScalar(col 13:double, val 0.0)(children: CastLongToDouble(col 0:tinyint) -> 13:double) -> 15:double, DoubleColDivideDoubleScalar(col 13:double, val 0.0)(children: CastLongToDouble(col 3:bigint) -> 13:double) -> 16:double, DoubleColDivideDoubleScalar(col 5:double, val 0.0) -> 13:double +Statistics: Num rows: 12288 Data size: 613400 Basic stats: COMPLETE Column stats: COMPLETE +Reduce Output Operator + key expressions: _col0 (type: int), _col2 (type: tinyint), _col4 (type: bigint), _col6 (type: double) + sort order: + Reduce Sink Vectorization: + className: VectorReduceSinkObjectHashOperator + native: true + nativeConditionsMet: hive.vectorized.execution.reducesink.new.enabled IS true, hive.execution.engine tez IN [tez, spark] IS true, No PTF TopN IS true, No DISTINCT columns IS true, BinarySortableSerDe for keys IS true, LazyBinarySerDe for values IS true + Statistics: Num rows: 12288 Data size: 613400 Basic stats: COMPLETE Column stats: COMPLETE + TopN Hash Memory Usage: 0.1 + value expressions: _col1 (type: double), _col3 (type: double), _col5 (type: double), _col7 (type: double) +Execution mode: vectorized, llap +LLAP IO: all inputs +Map Vectorization: +enabled: true +enabledConditionsMet: hive.vectorized.use.vectorized.input.format IS true +inputFormatFeatureSupport: [] +featureSupportInUse: [] +inputFileFormats: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat +allNative: true +usesVectorUDFAdaptor: false +vectorized: true +Reducer 2 +Execution mode: vectorized, llap +Reduce Vectorization: +enabled: true +enableConditionsMet: hive.vectorized.execution.reduce.enabled IS true, hive.execution.engine tez IN [tez, spark] IS true +allNative: false +usesVectorUDFAdaptor: false +vectorized: true +Reduce Operator Tree: + Select Operator +expressions: KEY.reducesinkkey0 (type: int), VALUE._
[3/3] hive git commit: HIVE-19074: Vectorization: Add llap vectorization_div0.q.out Q output file (missing changes) (Matt McCline, reviewed by Teddy Choi)
HIVE-19074: Vectorization: Add llap vectorization_div0.q.out Q output file (missing changes) (Matt McCline, reviewed by Teddy Choi) Project: http://git-wip-us.apache.org/repos/asf/hive/repo Commit: http://git-wip-us.apache.org/repos/asf/hive/commit/a21302f1 Tree: http://git-wip-us.apache.org/repos/asf/hive/tree/a21302f1 Diff: http://git-wip-us.apache.org/repos/asf/hive/diff/a21302f1 Branch: refs/heads/master Commit: a21302f10777047c2a09c0b74de648b4e3b15f2d Parents: e7c15d2 Author: Matt McCline Authored: Mon Apr 9 13:22:42 2018 -0500 Committer: Matt McCline Committed: Mon Apr 9 13:36:02 2018 -0500 -- .../resources/testconfiguration.properties.orig | 1685 -- .../llap/vectorization_div0.q.out | 870 + .../clientpositive/tez/vectorization_div0.q.out | 870 - 3 files changed, 870 insertions(+), 2555 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hive/blob/a21302f1/itests/src/test/resources/testconfiguration.properties.orig -- diff --git a/itests/src/test/resources/testconfiguration.properties.orig b/itests/src/test/resources/testconfiguration.properties.orig deleted file mode 100644 index a42ae80..000 --- a/itests/src/test/resources/testconfiguration.properties.orig +++ /dev/null @@ -1,1685 +0,0 @@ -# Note: the *.shared groups also run on TestCliDriver - -# NOTE: files should be listed in alphabetical order -minimr.query.files=infer_bucket_sort_map_operators.q,\ - infer_bucket_sort_dyn_part.q,\ - infer_bucket_sort_merge.q,\ - infer_bucket_sort_reducers_power_two.q,\ - infer_bucket_sort_num_buckets.q,\ - root_dir_external_table.q,\ - parallel_orderby.q,\ - bucket_num_reducers.q,\ - udf_using.q,\ - index_bitmap3.q,\ - index_bitmap_auto.q,\ - scriptfile1.q,\ - bucket_num_reducers2.q,\ - bucket_num_reducers_acid.q,\ - bucket_num_reducers_acid2.q,\ - scriptfile1_win.q - -# These tests are disabled for minimr -# ql_rewrite_gbtoidx.q,\ -# ql_rewrite_gbtoidx_cbo_1.q,\ -# ql_rewrite_gbtoidx_cbo_2.q,\ -# smb_mapjoin_8.q,\ - - -# Tests that are not enabled for CLI Driver -disabled.query.files=ql_rewrite_gbtoidx.q,\ - ql_rewrite_gbtoidx_cbo_1.q,\ - cbo_rp_subq_in.q,\ - cbo_rp_subq_not_in.q,\ - cbo_rp_subq_exists.q,\ - orc_llap.q,\ - ql_rewrite_gbtoidx_cbo_2.q,\ - rcfile_merge1.q,\ - stats_filemetadata.q,\ - cbo_rp_insert.q,\ - cbo_rp_lineage2.q - -# NOTE: Add tests to minitez only if it is very -# specific to tez and cannot be added to minillap. -minitez.query.files.shared=delete_orig_table.q,\ - orc_merge12.q,\ - orc_vectorization_ppd.q,\ - update_orig_table.q,\ - vector_join_part_col_char.q,\ - vector_non_string_partition.q,\ - vectorization_div0.q,\ - vectorization_limit.q - -# NOTE: Add tests to minitez only if it is very -# specific to tez and cannot be added to minillap. -minitez.query.files=acid_vectorization_original_tez.q,\ - explainuser_3.q,\ - explainanalyze_1.q,\ - explainanalyze_3.q,\ - explainanalyze_4.q,\ - explainanalyze_5.q,\ - hybridgrace_hashjoin_1.q,\ - hybridgrace_hashjoin_2.q,\ - multi_count_distinct.q,\ - tez-tag.q,\ - tez_union_with_udf.q,\ - vectorization_div0.q - - -minillap.shared.query.files=insert_into1.q,\ - insert_into2.q,\ - llapdecider.q,\ - mapreduce1.q,\ - mapreduce2.q,\ - mm_all.q,\ - mm_cttas.q,\ - orc_merge1.q,\ - orc_merge10.q,\ - orc_merge2.q,\ - orc_merge3.q,\ - orc_merge4.q,\ - orc_merge_diff_fs.q,\ - parallel_colstats.q,\ - parquet_types_vectorization.q,\ - parquet_complex_types_vectorization.q,\ - parquet_map_type_vectorization.q,\ - parquet_struct_type_vectorization.q,\ - orc_struct_type_vectorization.q,\ - union_type_chk.q,\ - cte_2.q,\ - cte_4.q,\ - llap_nullscan.q,\ - dynamic_partition_pruning_2.q,\ - tez_union_dynamic_partition.q - -minillaplocal.shared.query.files=alter_merge_2_orc.q,\ - alter_merge_orc.q,\ - alter_merge_stats_orc.q,\ - authorization_view_8.q,\ - auto_join0.q,\ - auto_join1.q,\ - auto_join21.q,\ - auto_join29.q,\ - auto_join30.q,\ - auto_join_filters.q,\ - auto_join_nulls.q,\ - auto_sortmerge_join_1.q,\ - auto_sortmerge_join_10.q,\ - auto_sortmerge_join_11.q,\ - auto_sortmerge_join_12.q,\ - auto_sortmerge_join_13.q,\ - auto_sortmerge_join_14.q,\ - auto_sortmerge_join_15.q,\ - auto_sortmerge_join_16.q,\ - auto_sortmerge_join_2.q,\ - auto_sortmerge_join_3.q,\ - auto_sortmerge_join_4.q,\ - auto_sortmerge_join_5.q,\ - auto_sortmerge_join_6.q,\ - auto_sortmerge_join_7.q,\ - auto_sortmerge_join_8.q,\ - auto_sortmerge_join_9.q,\ - autoColumnStats_1.q,\ - autoColumnStats_10.q,\ - autoColumnStats_2.q,\ - bucket2.q,\ - bucket3.q,\ - bucket4.q,\ - bucket_map_join_tez1.q,\ - bucket_map_join_tez2.q,\ - cbo_gby.q,\ - cbo_gby_empty.q,\ - cbo_join.q,\ - cbo_limit.q,\ - cbo_semijoin.q,\ - cbo_s
[1/3] hive git commit: HIVE-19074: Vectorization: Add llap vectorization_div0.q.out Q output file (missing changes) (Matt McCline, reviewed by Teddy Choi)
Repository: hive Updated Branches: refs/heads/master e7c15d234 -> a21302f10 http://git-wip-us.apache.org/repos/asf/hive/blob/a21302f1/ql/src/test/results/clientpositive/tez/vectorization_div0.q.out -- diff --git a/ql/src/test/results/clientpositive/tez/vectorization_div0.q.out b/ql/src/test/results/clientpositive/tez/vectorization_div0.q.out deleted file mode 100644 index e00dcd9..000 --- a/ql/src/test/results/clientpositive/tez/vectorization_div0.q.out +++ /dev/null @@ -1,870 +0,0 @@ -PREHOOK: query: explain vectorization expression -select cint, cint / 0 as cint_div, ctinyint, ctinyint / 0 as ctinyint_div, cbigint, cbigint / 0 as cbigint_div, cdouble, cdouble / 0.0 as cdouble_div -from alltypesorc order by cint, ctinyint, cbigint, cdouble limit 100 -PREHOOK: type: QUERY -POSTHOOK: query: explain vectorization expression -select cint, cint / 0 as cint_div, ctinyint, ctinyint / 0 as ctinyint_div, cbigint, cbigint / 0 as cbigint_div, cdouble, cdouble / 0.0 as cdouble_div -from alltypesorc order by cint, ctinyint, cbigint, cdouble limit 100 -POSTHOOK: type: QUERY -Explain -PLAN VECTORIZATION: - enabled: true - enabledConditionsMet: [hive.vectorized.execution.enabled IS true] - -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 -Tez - A masked pattern was here - Edges: -Reducer 2 <- Map 1 (SIMPLE_EDGE) - A masked pattern was here - Vertices: -Map 1 -Map Operator Tree: -TableScan - alias: alltypesorc - Statistics: Num rows: 12288 Data size: 220184 Basic stats: COMPLETE Column stats: COMPLETE - TableScan Vectorization: - native: true - Select Operator -expressions: cint (type: int), (UDFToDouble(cint) / 0.0D) (type: double), ctinyint (type: tinyint), (UDFToDouble(ctinyint) / 0.0D) (type: double), cbigint (type: bigint), (UDFToDouble(cbigint) / 0.0D) (type: double), cdouble (type: double), (cdouble / 0.0D) (type: double) -outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7 -Select Vectorization: -className: VectorSelectOperator -native: true -projectedOutputColumnNums: [2, 14, 0, 15, 3, 16, 5, 13] -selectExpressions: DoubleColDivideDoubleScalar(col 13:double, val 0.0)(children: CastLongToDouble(col 2:int) -> 13:double) -> 14:double, DoubleColDivideDoubleScalar(col 13:double, val 0.0)(children: CastLongToDouble(col 0:tinyint) -> 13:double) -> 15:double, DoubleColDivideDoubleScalar(col 13:double, val 0.0)(children: CastLongToDouble(col 3:bigint) -> 13:double) -> 16:double, DoubleColDivideDoubleScalar(col 5:double, val 0.0) -> 13:double -Statistics: Num rows: 12288 Data size: 613400 Basic stats: COMPLETE Column stats: COMPLETE -Reduce Output Operator - key expressions: _col0 (type: int), _col2 (type: tinyint), _col4 (type: bigint), _col6 (type: double) - sort order: - Reduce Sink Vectorization: - className: VectorReduceSinkObjectHashOperator - native: true - nativeConditionsMet: hive.vectorized.execution.reducesink.new.enabled IS true, hive.execution.engine tez IN [tez, spark] IS true, No PTF TopN IS true, No DISTINCT columns IS true, BinarySortableSerDe for keys IS true, LazyBinarySerDe for values IS true - Statistics: Num rows: 12288 Data size: 613400 Basic stats: COMPLETE Column stats: COMPLETE - TopN Hash Memory Usage: 0.1 - value expressions: _col1 (type: double), _col3 (type: double), _col5 (type: double), _col7 (type: double) -Execution mode: vectorized, llap -LLAP IO: all inputs -Map Vectorization: -enabled: true -enabledConditionsMet: hive.vectorized.use.vectorized.input.format IS true -inputFormatFeatureSupport: [] -featureSupportInUse: [] -inputFileFormats: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat -allNative: true -usesVectorUDFAdaptor: false -vectorized: true -Reducer 2 -Execution mode: vectorized, llap -Reduce Vectorization: -enabled: true -enableConditionsMet: hive.vectorized.execution.reduce.enabled IS true, hive.execution.engine tez IN [tez, spark] IS true -allNative: false -usesVectorUDFAdaptor: false -vectorized: true -Reduce Operator Tree: - Selec
hive git commit: HIVE-18781 broke WarehouseInstance (Alan Gates, reviewed by Thejas Nair)
Repository: hive Updated Branches: refs/heads/master a21302f10 -> 7fa2ba8c7 HIVE-18781 broke WarehouseInstance (Alan Gates, reviewed by Thejas Nair) Project: http://git-wip-us.apache.org/repos/asf/hive/repo Commit: http://git-wip-us.apache.org/repos/asf/hive/commit/7fa2ba8c Tree: http://git-wip-us.apache.org/repos/asf/hive/tree/7fa2ba8c Diff: http://git-wip-us.apache.org/repos/asf/hive/diff/7fa2ba8c Branch: refs/heads/master Commit: 7fa2ba8c77b77d176c6d6180e6444b97aebf1503 Parents: a21302f Author: Alan Gates Authored: Mon Apr 9 11:49:55 2018 -0700 Committer: Alan Gates Committed: Mon Apr 9 11:49:55 2018 -0700 -- .../java/org/apache/hadoop/hive/ql/parse/WarehouseInstance.java | 3 +++ 1 file changed, 3 insertions(+) -- http://git-wip-us.apache.org/repos/asf/hive/blob/7fa2ba8c/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/parse/WarehouseInstance.java -- diff --git a/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/parse/WarehouseInstance.java b/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/parse/WarehouseInstance.java index fe4660c..7c8020d 100644 --- a/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/parse/WarehouseInstance.java +++ b/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/parse/WarehouseInstance.java @@ -123,6 +123,9 @@ public class WarehouseInstance implements Closeable { hiveConf.setIntVar(HiveConf.ConfVars.METASTORETHRIFTCONNECTIONRETRIES, 3); hiveConf.set(HiveConf.ConfVars.PREEXECHOOKS.varname, ""); hiveConf.set(HiveConf.ConfVars.POSTEXECHOOKS.varname, ""); +if (!hiveConf.getVar(HiveConf.ConfVars.HIVE_TXN_MANAGER).equals("org.apache.hadoop.hive.ql.lockmgr.DbTxnManager")) { + hiveConf.set(HiveConf.ConfVars.HIVE_SUPPORT_CONCURRENCY.varname, "false"); +} System.setProperty(HiveConf.ConfVars.PREEXECHOOKS.varname, " "); System.setProperty(HiveConf.ConfVars.POSTEXECHOOKS.varname, " ");
hive git commit: HIVE-19119 Fix the TestAppendPartitions tests which are failing in the pre-commit runs (Marta Kuczora via Alan Gates)
Repository: hive Updated Branches: refs/heads/master 7fa2ba8c7 -> d589ee755 HIVE-19119 Fix the TestAppendPartitions tests which are failing in the pre-commit runs (Marta Kuczora via Alan Gates) Project: http://git-wip-us.apache.org/repos/asf/hive/repo Commit: http://git-wip-us.apache.org/repos/asf/hive/commit/d589ee75 Tree: http://git-wip-us.apache.org/repos/asf/hive/tree/d589ee75 Diff: http://git-wip-us.apache.org/repos/asf/hive/diff/d589ee75 Branch: refs/heads/master Commit: d589ee7550abb52864b71acb94e1efab601ea67e Parents: 7fa2ba8 Author: Alan Gates Authored: Mon Apr 9 12:03:02 2018 -0700 Committer: Alan Gates Committed: Mon Apr 9 12:03:02 2018 -0700 -- .../apache/hadoop/hive/metastore/HiveMetaStore.java | 2 +- .../hive/metastore/client/TestAppendPartitions.java | 14 +- 2 files changed, 6 insertions(+), 10 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hive/blob/d589ee75/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/HiveMetaStore.java -- diff --git a/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/HiveMetaStore.java b/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/HiveMetaStore.java index a2fe7d7..c81b8fa 100644 --- a/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/HiveMetaStore.java +++ b/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/HiveMetaStore.java @@ -3094,7 +3094,7 @@ public class HiveMetaStore extends ThriftHiveMetastore { final String tableName, final List part_vals, final EnvironmentContext envContext) throws InvalidObjectException, AlreadyExistsException, MetaException { if (part_vals == null || part_vals.isEmpty()) { -throw new MetaException("The partition values must not be null."); +throw new MetaException("The partition values must not be null or empty."); } String[] parsedDbName = parseDbName(dbName, conf); startPartitionFunction("append_partition", parsedDbName[CAT_NAME], parsedDbName[DB_NAME], tableName, part_vals); http://git-wip-us.apache.org/repos/asf/hive/blob/d589ee75/standalone-metastore/src/test/java/org/apache/hadoop/hive/metastore/client/TestAppendPartitions.java -- diff --git a/standalone-metastore/src/test/java/org/apache/hadoop/hive/metastore/client/TestAppendPartitions.java b/standalone-metastore/src/test/java/org/apache/hadoop/hive/metastore/client/TestAppendPartitions.java index 75b26f2..37ca40c 100644 --- a/standalone-metastore/src/test/java/org/apache/hadoop/hive/metastore/client/TestAppendPartitions.java +++ b/standalone-metastore/src/test/java/org/apache/hadoop/hive/metastore/client/TestAppendPartitions.java @@ -226,22 +226,18 @@ public class TestAppendPartitions extends MetaStoreClientTest { client.appendPartition(tableWithPartitions.getDbName(), null, partitionValues); } - @Test(expected = InvalidObjectException.class) + @Test(expected = MetaException.class) public void testAppendPartitionEmptyPartValues() throws Exception { Table table = tableWithPartitions; client.appendPartition(table.getDbName(), table.getTableName(), new ArrayList<>()); } - @Test + @Test(expected = MetaException.class) public void testAppendPartitionNullPartValues() throws Exception { -try { - Table table = tableWithPartitions; - client.appendPartition(table.getDbName(), table.getTableName(), (List) null); - Assert.fail("Exception should have been thrown."); -} catch (TTransportException | InvalidObjectException e) { - // TODO: NPE should not be thrown -} + +Table table = tableWithPartitions; +client.appendPartition(table.getDbName(), table.getTableName(), (List) null); } @Test
[hive] Git Push Summary
Repository: hive Updated Branches: refs/heads/branch-3.0.0 [created] d589ee755
[hive] Git Push Summary
Repository: hive Updated Branches: refs/heads/branch-3 [created] d589ee755
[hive] Git Push Summary
Repository: hive Updated Branches: refs/heads/branch-3.0.0 [deleted] d589ee755
hive git commit: HIVE-19134: Update copyright NOTICE and fix rat check failures (Vineet Garg, reviewed by Ashutosh Chauhan)
Repository: hive Updated Branches: refs/heads/master d589ee755 -> 2d3a41045 HIVE-19134: Update copyright NOTICE and fix rat check failures (Vineet Garg, reviewed by Ashutosh Chauhan) Project: http://git-wip-us.apache.org/repos/asf/hive/repo Commit: http://git-wip-us.apache.org/repos/asf/hive/commit/2d3a4104 Tree: http://git-wip-us.apache.org/repos/asf/hive/tree/2d3a4104 Diff: http://git-wip-us.apache.org/repos/asf/hive/diff/2d3a4104 Branch: refs/heads/master Commit: 2d3a410451626b32b60ce8edaa0d8d00d951c7d0 Parents: d589ee7 Author: Vineet Garg Authored: Mon Apr 9 14:34:29 2018 -0700 Committer: Vineet Garg Committed: Mon Apr 9 14:34:29 2018 -0700 -- NOTICE| 2 +- .../hadoop/hive/ql/TestTxnAddPartition.java | 18 ++ 2 files changed, 19 insertions(+), 1 deletion(-) -- http://git-wip-us.apache.org/repos/asf/hive/blob/2d3a4104/NOTICE -- diff --git a/NOTICE b/NOTICE index 0235613..d018205 100644 --- a/NOTICE +++ b/NOTICE @@ -1,5 +1,5 @@ Apache Hive -Copyright 2008-2016 The Apache Software Foundation +Copyright 2008-2018 The Apache Software Foundation This product includes software developed by The Apache Software Foundation (http://www.apache.org/). http://git-wip-us.apache.org/repos/asf/hive/blob/2d3a4104/ql/src/test/org/apache/hadoop/hive/ql/TestTxnAddPartition.java -- diff --git a/ql/src/test/org/apache/hadoop/hive/ql/TestTxnAddPartition.java b/ql/src/test/org/apache/hadoop/hive/ql/TestTxnAddPartition.java index 901b5db..c821365 100644 --- a/ql/src/test/org/apache/hadoop/hive/ql/TestTxnAddPartition.java +++ b/ql/src/test/org/apache/hadoop/hive/ql/TestTxnAddPartition.java @@ -1,3 +1,21 @@ +/* + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + package org.apache.hadoop.hive.ql; import org.apache.hadoop.fs.FileStatus;
hive git commit: HIVE-19134: Update copyright NOTICE and fix rat check failures(Vineet Garg, reviewed by Ashutosh Chauhan)
Repository: hive Updated Branches: refs/heads/branch-3 d589ee755 -> 43cb101ea HIVE-19134: Update copyright NOTICE and fix rat check failures(Vineet Garg, reviewed by Ashutosh Chauhan) Project: http://git-wip-us.apache.org/repos/asf/hive/repo Commit: http://git-wip-us.apache.org/repos/asf/hive/commit/43cb101e Tree: http://git-wip-us.apache.org/repos/asf/hive/tree/43cb101e Diff: http://git-wip-us.apache.org/repos/asf/hive/diff/43cb101e Branch: refs/heads/branch-3 Commit: 43cb101eac4dba1405c2e6337a04b0b8c1240e85 Parents: d589ee7 Author: Vineet Garg Authored: Mon Apr 9 14:35:17 2018 -0700 Committer: Vineet Garg Committed: Mon Apr 9 14:35:17 2018 -0700 -- NOTICE| 2 +- .../hadoop/hive/ql/TestTxnAddPartition.java | 18 ++ 2 files changed, 19 insertions(+), 1 deletion(-) -- http://git-wip-us.apache.org/repos/asf/hive/blob/43cb101e/NOTICE -- diff --git a/NOTICE b/NOTICE index 0235613..d018205 100644 --- a/NOTICE +++ b/NOTICE @@ -1,5 +1,5 @@ Apache Hive -Copyright 2008-2016 The Apache Software Foundation +Copyright 2008-2018 The Apache Software Foundation This product includes software developed by The Apache Software Foundation (http://www.apache.org/). http://git-wip-us.apache.org/repos/asf/hive/blob/43cb101e/ql/src/test/org/apache/hadoop/hive/ql/TestTxnAddPartition.java -- diff --git a/ql/src/test/org/apache/hadoop/hive/ql/TestTxnAddPartition.java b/ql/src/test/org/apache/hadoop/hive/ql/TestTxnAddPartition.java index 901b5db..c821365 100644 --- a/ql/src/test/org/apache/hadoop/hive/ql/TestTxnAddPartition.java +++ b/ql/src/test/org/apache/hadoop/hive/ql/TestTxnAddPartition.java @@ -1,3 +1,21 @@ +/* + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + package org.apache.hadoop.hive.ql; import org.apache.hadoop.fs.FileStatus;
[04/13] hive git commit: HIVE-19128 : Update golden files for spark perf tests
http://git-wip-us.apache.org/repos/asf/hive/blob/328d3f93/ql/src/test/results/clientpositive/perf/spark/query78.q.out -- diff --git a/ql/src/test/results/clientpositive/perf/spark/query78.q.out b/ql/src/test/results/clientpositive/perf/spark/query78.q.out index 1467c5f..c2e8577 100644 --- a/ql/src/test/results/clientpositive/perf/spark/query78.q.out +++ b/ql/src/test/results/clientpositive/perf/spark/query78.q.out @@ -120,57 +120,41 @@ STAGE PLANS: Stage: Stage-1 Spark Edges: -Reducer 11 <- Map 10 (PARTITION-LEVEL SORT, 85), Reducer 14 (PARTITION-LEVEL SORT, 85) -Reducer 12 <- Reducer 11 (GROUP, 93) -Reducer 14 <- Map 13 (PARTITION-LEVEL SORT, 164), Map 15 (PARTITION-LEVEL SORT, 164) -Reducer 17 <- Map 10 (PARTITION-LEVEL SORT, 85), Reducer 20 (PARTITION-LEVEL SORT, 85) +Reducer 11 <- Map 10 (PARTITION-LEVEL SORT, 164), Map 14 (PARTITION-LEVEL SORT, 164) +Reducer 12 <- Map 15 (PARTITION-LEVEL SORT, 85), Reducer 11 (PARTITION-LEVEL SORT, 85) +Reducer 13 <- Reducer 12 (GROUP, 93) +Reducer 17 <- Map 16 (PARTITION-LEVEL SORT, 85), Reducer 20 (PARTITION-LEVEL SORT, 85) Reducer 18 <- Reducer 17 (GROUP, 93) -Reducer 2 <- Map 1 (PARTITION-LEVEL SORT, 219), Reducer 8 (PARTITION-LEVEL SORT, 219) +Reducer 2 <- Map 1 (PARTITION-LEVEL SORT, 432), Map 8 (PARTITION-LEVEL SORT, 432) Reducer 20 <- Map 19 (PARTITION-LEVEL SORT, 177), Map 21 (PARTITION-LEVEL SORT, 177) -Reducer 3 <- Reducer 2 (GROUP, 241) -Reducer 4 <- Reducer 12 (PARTITION-LEVEL SORT, 167), Reducer 3 (PARTITION-LEVEL SORT, 167) -Reducer 5 <- Reducer 18 (PARTITION-LEVEL SORT, 91), Reducer 4 (PARTITION-LEVEL SORT, 91) -Reducer 6 <- Reducer 5 (SORT, 1) -Reducer 8 <- Map 7 (PARTITION-LEVEL SORT, 432), Map 9 (PARTITION-LEVEL SORT, 432) +Reducer 3 <- Map 9 (PARTITION-LEVEL SORT, 219), Reducer 2 (PARTITION-LEVEL SORT, 219) +Reducer 4 <- Reducer 3 (GROUP, 241) +Reducer 5 <- Reducer 13 (PARTITION-LEVEL SORT, 167), Reducer 4 (PARTITION-LEVEL SORT, 167) +Reducer 6 <- Reducer 18 (PARTITION-LEVEL SORT, 91), Reducer 5 (PARTITION-LEVEL SORT, 91) +Reducer 7 <- Reducer 6 (SORT, 1) A masked pattern was here Vertices: Map 1 Map Operator Tree: TableScan - alias: date_dim - Statistics: Num rows: 73049 Data size: 81741831 Basic stats: COMPLETE Column stats: NONE + alias: store_sales + Statistics: Num rows: 575995635 Data size: 50814502088 Basic stats: COMPLETE Column stats: NONE Filter Operator -predicate: ((d_year = 2000) and d_date_sk is not null) (type: boolean) -Statistics: Num rows: 36524 Data size: 40870356 Basic stats: COMPLETE Column stats: NONE +predicate: ss_sold_date_sk is not null (type: boolean) +Statistics: Num rows: 575995635 Data size: 50814502088 Basic stats: COMPLETE Column stats: NONE Select Operator - expressions: d_date_sk (type: int) - outputColumnNames: _col0 - Statistics: Num rows: 36524 Data size: 40870356 Basic stats: COMPLETE Column stats: NONE + expressions: ss_sold_date_sk (type: int), ss_item_sk (type: int), ss_customer_sk (type: int), ss_ticket_number (type: int), ss_quantity (type: int), ss_wholesale_cost (type: decimal(7,2)), ss_sales_price (type: decimal(7,2)) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6 + Statistics: Num rows: 575995635 Data size: 50814502088 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator -key expressions: _col0 (type: int) -sort order: + -Map-reduce partition columns: _col0 (type: int) -Statistics: Num rows: 36524 Data size: 40870356 Basic stats: COMPLETE Column stats: NONE +key expressions: _col1 (type: int), _col3 (type: int) +sort order: ++ +Map-reduce partition columns: _col1 (type: int), _col3 (type: int) +Statistics: Num rows: 575995635 Data size: 50814502088 Basic stats: COMPLETE Column stats: NONE +value expressions: _col0 (type: int), _col2 (type: int), _col4 (type: int), _col5 (type: decimal(7,2)), _col6 (type: decimal(7,2)) Map 10 Map Operator Tree: TableScan - alias: date_dim - Statistics: Num rows: 73049 Data size: 81741831 Basic stats: COMPLETE Column stats: NONE - Filter Operator -
[08/13] hive git commit: HIVE-19128 : Update golden files for spark perf tests
http://git-wip-us.apache.org/repos/asf/hive/blob/328d3f93/ql/src/test/results/clientpositive/perf/spark/query54.q.out -- diff --git a/ql/src/test/results/clientpositive/perf/spark/query54.q.out b/ql/src/test/results/clientpositive/perf/spark/query54.q.out index 43132bc..251d7ad 100644 --- a/ql/src/test/results/clientpositive/perf/spark/query54.q.out +++ b/ql/src/test/results/clientpositive/perf/spark/query54.q.out @@ -1,7 +1,7 @@ -Warning: Shuffle Join JOIN[111][tables = [$hdt$_1, $hdt$_2]] in Work 'Reducer 3' is a cross product -Warning: Shuffle Join JOIN[107][tables = [$hdt$_0, $hdt$_1, $hdt$_2, $hdt$_3, $hdt$_4]] in Work 'Reducer 14' is a cross product -Warning: Shuffle Join JOIN[114][tables = [$hdt$_1, $hdt$_2, $hdt$_0]] in Work 'Reducer 4' is a cross product -Warning: Map Join MAPJOIN[144][bigTable=?] in task 'Stage-1:MAPRED' is a cross product +Warning: Shuffle Join JOIN[111][tables = [$hdt$_1, $hdt$_2]] in Work 'Reducer 4' is a cross product +Warning: Shuffle Join JOIN[104][tables = [$hdt$_0, $hdt$_1, $hdt$_2, $hdt$_3]] in Work 'Reducer 14' is a cross product +Warning: Shuffle Join JOIN[114][tables = [$hdt$_1, $hdt$_2, $hdt$_0]] in Work 'Reducer 5' is a cross product +Warning: Map Join MAPJOIN[143][bigTable=?] in task 'Stage-1:MAPRED' is a cross product PREHOOK: query: explain with my_customers as ( select distinct c_customer_sk @@ -122,11 +122,11 @@ STAGE PLANS: Stage: Stage-2 Spark Edges: -Reducer 29 <- Map 28 (GROUP, 2) -Reducer 30 <- Reducer 29 (GROUP, 1) +Reducer 31 <- Map 30 (GROUP, 2) +Reducer 32 <- Reducer 31 (GROUP, 1) A masked pattern was here Vertices: -Map 28 +Map 30 Map Operator Tree: TableScan alias: date_dim @@ -148,7 +148,7 @@ STAGE PLANS: sort order: + Map-reduce partition columns: _col0 (type: int) Statistics: Num rows: 18262 Data size: 20435178 Basic stats: COMPLETE Column stats: NONE -Reducer 29 +Reducer 31 Reduce Operator Tree: Group By Operator keys: KEY._col0 (type: int) @@ -166,7 +166,7 @@ STAGE PLANS: sort order: Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE value expressions: _col0 (type: bigint) -Reducer 30 +Reducer 32 Local Work: Map Reduce Local Work Reduce Operator Tree: @@ -211,23 +211,23 @@ STAGE PLANS: Stage: Stage-1 Spark Edges: -Reducer 10 <- Reducer 9 (GROUP, 1) +Reducer 10 <- Map 1 (GROUP, 2) Reducer 12 <- Map 11 (PARTITION-LEVEL SORT, 398), Map 15 (PARTITION-LEVEL SORT, 398) Reducer 13 <- Reducer 12 (PARTITION-LEVEL SORT, 772), Reducer 17 (PARTITION-LEVEL SORT, 772) -Reducer 14 <- Reducer 13 (PARTITION-LEVEL SORT, 1), Reducer 32 (PARTITION-LEVEL SORT, 1) +Reducer 14 <- Reducer 13 (PARTITION-LEVEL SORT, 1), Reducer 29 (PARTITION-LEVEL SORT, 1) Reducer 17 <- Map 16 (PARTITION-LEVEL SORT, 654), Reducer 23 (PARTITION-LEVEL SORT, 654) Reducer 2 <- Map 1 (GROUP, 2) -Reducer 20 <- Map 19 (PARTITION-LEVEL SORT, 458), Map 24 (PARTITION-LEVEL SORT, 458), Map 25 (PARTITION-LEVEL SORT, 458) -Reducer 21 <- Map 26 (PARTITION-LEVEL SORT, 505), Reducer 20 (PARTITION-LEVEL SORT, 505) +Reducer 20 <- Map 19 (PARTITION-LEVEL SORT, 459), Map 24 (PARTITION-LEVEL SORT, 459), Map 25 (PARTITION-LEVEL SORT, 459) +Reducer 21 <- Map 26 (PARTITION-LEVEL SORT, 504), Reducer 20 (PARTITION-LEVEL SORT, 504) Reducer 22 <- Map 27 (PARTITION-LEVEL SORT, 1009), Reducer 21 (PARTITION-LEVEL SORT, 1009) Reducer 23 <- Reducer 22 (GROUP, 610) -Reducer 3 <- Reducer 10 (PARTITION-LEVEL SORT, 1), Reducer 2 (PARTITION-LEVEL SORT, 1) -Reducer 32 <- Map 31 (GROUP, 2) -Reducer 4 <- Reducer 14 (PARTITION-LEVEL SORT, 1), Reducer 3 (PARTITION-LEVEL SORT, 1) -Reducer 5 <- Reducer 4 (GROUP, 1009) +Reducer 29 <- Map 28 (GROUP, 2) +Reducer 3 <- Reducer 2 (GROUP, 1) +Reducer 4 <- Reducer 10 (PARTITION-LEVEL SORT, 1), Reducer 3 (PARTITION-LEVEL SORT, 1) +Reducer 5 <- Reducer 14 (PARTITION-LEVEL SORT, 1), Reducer 4 (PARTITION-LEVEL SORT, 1) Reducer 6 <- Reducer 5 (GROUP, 1009) -Reducer 7 <- Reducer 6 (SORT, 1) -Reducer 9 <- Map 1 (GROUP, 2) +Reducer 7 <- Reducer 6 (GROUP, 1009) +Reducer 8 <- Reducer 7 (SORT, 1) A masked pattern was here Vertices: Map 1 @@ -330,11 +330,11 @@ STAGE PLANS: outputColumnNames: _col0, _col1, _col2 Statistics: Num rows: 287989836 Data size: 38999608952 Basic stat
[11/13] hive git commit: HIVE-19128 : Update golden files for spark perf tests
http://git-wip-us.apache.org/repos/asf/hive/blob/328d3f93/ql/src/test/results/clientpositive/perf/spark/query30.q.out -- diff --git a/ql/src/test/results/clientpositive/perf/spark/query30.q.out b/ql/src/test/results/clientpositive/perf/spark/query30.q.out index 6385984..399251d 100644 --- a/ql/src/test/results/clientpositive/perf/spark/query30.q.out +++ b/ql/src/test/results/clientpositive/perf/spark/query30.q.out @@ -66,72 +66,37 @@ STAGE PLANS: Stage: Stage-1 Spark Edges: -Reducer 10 <- Reducer 16 (PARTITION-LEVEL SORT, 262), Reducer 9 (PARTITION-LEVEL SORT, 262) -Reducer 14 <- Map 13 (PARTITION-LEVEL SORT, 11), Map 17 (PARTITION-LEVEL SORT, 11) -Reducer 15 <- Map 18 (PARTITION-LEVEL SORT, 329), Reducer 14 (PARTITION-LEVEL SORT, 329) -Reducer 16 <- Reducer 15 (GROUP PARTITION-LEVEL SORT, 349) -Reducer 2 <- Map 1 (PARTITION-LEVEL SORT, 697), Map 5 (PARTITION-LEVEL SORT, 697) -Reducer 3 <- Reducer 10 (PARTITION-LEVEL SORT, 656), Reducer 2 (PARTITION-LEVEL SORT, 656) -Reducer 4 <- Reducer 3 (SORT, 1) -Reducer 7 <- Map 11 (PARTITION-LEVEL SORT, 11), Map 6 (PARTITION-LEVEL SORT, 11) -Reducer 8 <- Map 12 (PARTITION-LEVEL SORT, 329), Reducer 7 (PARTITION-LEVEL SORT, 329) -Reducer 9 <- Reducer 8 (GROUP, 349) +Reducer 11 <- Map 10 (PARTITION-LEVEL SORT, 11), Map 14 (PARTITION-LEVEL SORT, 11) +Reducer 12 <- Map 15 (PARTITION-LEVEL SORT, 329), Reducer 11 (PARTITION-LEVEL SORT, 329) +Reducer 13 <- Reducer 12 (GROUP PARTITION-LEVEL SORT, 349) +Reducer 17 <- Map 16 (PARTITION-LEVEL SORT, 697), Map 18 (PARTITION-LEVEL SORT, 697) +Reducer 2 <- Map 1 (PARTITION-LEVEL SORT, 11), Map 8 (PARTITION-LEVEL SORT, 11) +Reducer 3 <- Map 9 (PARTITION-LEVEL SORT, 329), Reducer 2 (PARTITION-LEVEL SORT, 329) +Reducer 4 <- Reducer 3 (GROUP, 349) +Reducer 5 <- Reducer 13 (PARTITION-LEVEL SORT, 262), Reducer 4 (PARTITION-LEVEL SORT, 262) +Reducer 6 <- Reducer 17 (PARTITION-LEVEL SORT, 656), Reducer 5 (PARTITION-LEVEL SORT, 656) +Reducer 7 <- Reducer 6 (SORT, 1) A masked pattern was here Vertices: Map 1 Map Operator Tree: TableScan - alias: customer - Statistics: Num rows: 8000 Data size: 68801615852 Basic stats: COMPLETE Column stats: NONE - Filter Operator -predicate: (c_current_addr_sk is not null and c_customer_sk is not null) (type: boolean) -Statistics: Num rows: 8000 Data size: 68801615852 Basic stats: COMPLETE Column stats: NONE -Select Operator - expressions: c_customer_sk (type: int), c_customer_id (type: string), c_current_addr_sk (type: int), c_salutation (type: string), c_first_name (type: string), c_last_name (type: string), c_preferred_cust_flag (type: string), c_birth_day (type: int), c_birth_month (type: int), c_birth_year (type: int), c_birth_country (type: string), c_login (type: string), c_email_address (type: string), c_last_review_date (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13 - Statistics: Num rows: 8000 Data size: 68801615852 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator -key expressions: _col2 (type: int) -sort order: + -Map-reduce partition columns: _col2 (type: int) -Statistics: Num rows: 8000 Data size: 68801615852 Basic stats: COMPLETE Column stats: NONE -value expressions: _col0 (type: int), _col1 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: string), _col6 (type: string), _col7 (type: int), _col8 (type: int), _col9 (type: int), _col10 (type: string), _col11 (type: string), _col12 (type: string), _col13 (type: string) -Map 11 -Map Operator Tree: -TableScan - alias: date_dim - Statistics: Num rows: 73049 Data size: 81741831 Basic stats: COMPLETE Column stats: NONE - Filter Operator -predicate: ((d_year = 2002) and d_date_sk is not null) (type: boolean) -Statistics: Num rows: 36524 Data size: 40870356 Basic stats: COMPLETE Column stats: NONE -Select Operator - expressions: d_date_sk (type: int) - outputColumnNames: _col0 - Statistics: Num rows: 36524 Data size: 40870356 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator -key expressions: _col0 (type: i
[02/13] hive git commit: HIVE-19128 : Update golden files for spark perf tests
http://git-wip-us.apache.org/repos/asf/hive/blob/328d3f93/ql/src/test/results/clientpositive/perf/spark/query85.q.out -- diff --git a/ql/src/test/results/clientpositive/perf/spark/query85.q.out b/ql/src/test/results/clientpositive/perf/spark/query85.q.out index d60751c..c82dcf7 100644 --- a/ql/src/test/results/clientpositive/perf/spark/query85.q.out +++ b/ql/src/test/results/clientpositive/perf/spark/query85.q.out @@ -175,7 +175,7 @@ STAGE PLANS: Spark A masked pattern was here Vertices: -Map 13 +Map 15 Map Operator Tree: TableScan alias: reason @@ -212,21 +212,21 @@ STAGE PLANS: Statistics: Num rows: 4602 Data size: 2696178 Basic stats: COMPLETE Column stats: NONE Spark HashTable Sink Operator keys: - 0 _col10 (type: int) - 1 _col0 (type: int) + 0 _col0 (type: int) + 1 _col2 (type: int) Local Work: Map Reduce Local Work Stage: Stage-1 Spark Edges: -Reducer 2 <- Map 1 (PARTITION-LEVEL SORT, 28), Map 9 (PARTITION-LEVEL SORT, 28) -Reducer 3 <- Map 10 (PARTITION-LEVEL SORT, 98), Reducer 2 (PARTITION-LEVEL SORT, 98) -Reducer 4 <- Map 12 (PARTITION-LEVEL SORT, 5), Reducer 3 (PARTITION-LEVEL SORT, 5) -Reducer 5 <- Map 14 (PARTITION-LEVEL SORT, 11), Reducer 4 (PARTITION-LEVEL SORT, 11) -Reducer 6 <- Map 15 (PARTITION-LEVEL SORT, 7), Reducer 5 (PARTITION-LEVEL SORT, 7) -Reducer 7 <- Reducer 6 (GROUP, 7) -Reducer 8 <- Reducer 7 (SORT, 1) +Reducer 13 <- Map 12 (PARTITION-LEVEL SORT, 20), Map 14 (PARTITION-LEVEL SORT, 20) +Reducer 2 <- Map 1 (PARTITION-LEVEL SORT, 90), Map 8 (PARTITION-LEVEL SORT, 90) +Reducer 3 <- Map 9 (PARTITION-LEVEL SORT, 17), Reducer 2 (PARTITION-LEVEL SORT, 17) +Reducer 4 <- Map 10 (PARTITION-LEVEL SORT, 19), Reducer 3 (PARTITION-LEVEL SORT, 19) +Reducer 5 <- Reducer 13 (PARTITION-LEVEL SORT, 35), Reducer 4 (PARTITION-LEVEL SORT, 35) +Reducer 6 <- Reducer 5 (GROUP, 2) +Reducer 7 <- Reducer 6 (SORT, 1) A masked pattern was here Vertices: Map 1 @@ -242,32 +242,61 @@ STAGE PLANS: outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7 Statistics: Num rows: 14398467 Data size: 1325194184 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator -key expressions: _col0 (type: int), _col5 (type: int) -sort order: ++ -Map-reduce partition columns: _col0 (type: int), _col5 (type: int) +key expressions: _col2 (type: int) +sort order: + +Map-reduce partition columns: _col2 (type: int) Statistics: Num rows: 14398467 Data size: 1325194184 Basic stats: COMPLETE Column stats: NONE -value expressions: _col1 (type: int), _col2 (type: int), _col3 (type: int), _col4 (type: int), _col6 (type: decimal(7,2)), _col7 (type: decimal(7,2)) +value expressions: _col0 (type: int), _col1 (type: int), _col3 (type: int), _col4 (type: int), _col5 (type: int), _col6 (type: decimal(7,2)), _col7 (type: decimal(7,2)) Map 10 Map Operator Tree: TableScan - alias: customer_address - Statistics: Num rows: 4000 Data size: 40595195284 Basic stats: COMPLETE Column stats: NONE + alias: cd2 + Statistics: Num rows: 1861800 Data size: 717186159 Basic stats: COMPLETE Column stats: NONE Filter Operator -predicate: ((ca_country = 'United States') and (ca_state) IN ('KY', 'GA', 'NM', 'MT', 'OR', 'IN', 'WI', 'MO', 'WV') and ca_address_sk is not null) (type: boolean) -Statistics: Num rows: 1000 Data size: 10148798821 Basic stats: COMPLETE Column stats: NONE +predicate: (((cd_education_status = '4 yr Degree') or (cd_education_status = 'Primary') or (cd_education_status = 'Advanced Degree')) and ((cd_marital_status = 'M') or (cd_marital_status = 'D') or (cd_marital_status = 'U')) and cd_demo_sk is not null and cd_education_status is not null and cd_marital_status is not null) (type: boolean) +Statistics: Num rows: 1861800 Data size: 717186159 Basic stats: COMPLETE Column stats: NONE Select Operator - expressions: ca_address_sk (type: int), ca_state (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num r
[01/13] hive git commit: HIVE-19128 : Update golden files for spark perf tests
Repository: hive Updated Branches: refs/heads/master 2d3a41045 -> 328d3f935 http://git-wip-us.apache.org/repos/asf/hive/blob/328d3f93/ql/src/test/results/clientpositive/perf/spark/query92.q.out -- diff --git a/ql/src/test/results/clientpositive/perf/spark/query92.q.out b/ql/src/test/results/clientpositive/perf/spark/query92.q.out index e7b8632..70c4c5a 100644 --- a/ql/src/test/results/clientpositive/perf/spark/query92.q.out +++ b/ql/src/test/results/clientpositive/perf/spark/query92.q.out @@ -67,7 +67,7 @@ STAGE PLANS: Spark A masked pattern was here Vertices: -Map 5 +Map 7 Map Operator Tree: TableScan alias: date_dim @@ -90,7 +90,7 @@ STAGE PLANS: Spark A masked pattern was here Vertices: -Map 9 +Map 10 Map Operator Tree: TableScan alias: date_dim @@ -112,11 +112,11 @@ STAGE PLANS: Stage: Stage-1 Spark Edges: -Reducer 2 <- Map 1 (PARTITION-LEVEL SORT, 261), Reducer 8 (PARTITION-LEVEL SORT, 261) -Reducer 3 <- Reducer 2 (GROUP, 1) -Reducer 4 <- Reducer 3 (SORT, 1) -Reducer 7 <- Map 6 (GROUP, 169) -Reducer 8 <- Map 10 (PARTITION-LEVEL SORT, 87), Reducer 7 (PARTITION-LEVEL SORT, 87) +Reducer 2 <- Map 1 (PARTITION-LEVEL SORT, 156), Map 6 (PARTITION-LEVEL SORT, 156) +Reducer 3 <- Reducer 2 (PARTITION-LEVEL SORT, 270), Reducer 9 (PARTITION-LEVEL SORT, 270) +Reducer 4 <- Reducer 3 (GROUP, 1) +Reducer 5 <- Reducer 4 (SORT, 1) +Reducer 9 <- Map 8 (GROUP, 169) A masked pattern was here Vertices: Map 1 @@ -131,25 +131,13 @@ STAGE PLANS: expressions: ws_sold_date_sk (type: int), ws_item_sk (type: int), ws_ext_discount_amt (type: decimal(7,2)) outputColumnNames: _col0, _col1, _col2 Statistics: Num rows: 144002668 Data size: 19580198212 Basic stats: COMPLETE Column stats: NONE - Map Join Operator -condition map: - Inner Join 0 to 1 -keys: - 0 _col0 (type: int) - 1 _col0 (type: int) -outputColumnNames: _col1, _col2 -input vertices: - 1 Map 5 -Statistics: Num rows: 158402938 Data size: 21538218500 Basic stats: COMPLETE Column stats: NONE -Reduce Output Operator - key expressions: _col1 (type: int) - sort order: + - Map-reduce partition columns: _col1 (type: int) - Statistics: Num rows: 158402938 Data size: 21538218500 Basic stats: COMPLETE Column stats: NONE - value expressions: _col2 (type: decimal(7,2)) -Local Work: - Map Reduce Local Work -Map 10 + Reduce Output Operator +key expressions: _col1 (type: int) +sort order: + +Map-reduce partition columns: _col1 (type: int) +Statistics: Num rows: 144002668 Data size: 19580198212 Basic stats: COMPLETE Column stats: NONE +value expressions: _col0 (type: int), _col2 (type: decimal(7,2)) +Map 6 Map Operator Tree: TableScan alias: item @@ -166,7 +154,7 @@ STAGE PLANS: sort order: + Map-reduce partition columns: _col0 (type: int) Statistics: Num rows: 231000 Data size: 331780228 Basic stats: COMPLETE Column stats: NONE -Map 6 +Map 8 Map Operator Tree: TableScan alias: web_sales @@ -186,7 +174,7 @@ STAGE PLANS: 1 _col0 (type: int) outputColumnNames: _col1, _col2 input vertices: - 1 Map 9 + 1 Map 10 Statistics: Num rows: 158402938 Data size: 21538218500 Basic stats: COMPLETE Column stats: NONE Group By Operator aggregations: sum(_col2), count(_col2) @@ -203,22 +191,50 @@ STAGE PLANS: Local Work: Map Reduce Local Work Reducer 2 +Local Work: + Map Reduce Local Work Reduce Operator Tree: Join Operator condition map: Inner Join 0 to 1 keys: 0 _col1 (type: int) - 1 _col2 (type: int) -output
[12/13] hive git commit: HIVE-19128 : Update golden files for spark perf tests
http://git-wip-us.apache.org/repos/asf/hive/blob/328d3f93/ql/src/test/results/clientpositive/perf/spark/query24.q.out -- diff --git a/ql/src/test/results/clientpositive/perf/spark/query24.q.out b/ql/src/test/results/clientpositive/perf/spark/query24.q.out index 1f291c0..13ac1e8 100644 --- a/ql/src/test/results/clientpositive/perf/spark/query24.q.out +++ b/ql/src/test/results/clientpositive/perf/spark/query24.q.out @@ -131,10 +131,10 @@ STAGE PLANS: Stage: Stage-2 Spark Edges: -Reducer 13 <- Map 12 (PARTITION-LEVEL SORT, 975), Map 20 (PARTITION-LEVEL SORT, 975) -Reducer 14 <- Map 21 (PARTITION-LEVEL SORT, 486), Reducer 13 (PARTITION-LEVEL SORT, 486) -Reducer 15 <- Map 22 (PARTITION-LEVEL SORT, 564), Reducer 14 (PARTITION-LEVEL SORT, 564) -Reducer 16 <- Map 23 (PARTITION-LEVEL SORT, 899), Reducer 15 (PARTITION-LEVEL SORT, 899) +Reducer 13 <- Map 12 (PARTITION-LEVEL SORT, 472), Map 20 (PARTITION-LEVEL SORT, 472) +Reducer 14 <- Map 21 (PARTITION-LEVEL SORT, 1009), Reducer 13 (PARTITION-LEVEL SORT, 1009) +Reducer 15 <- Map 22 (PARTITION-LEVEL SORT, 846), Reducer 14 (PARTITION-LEVEL SORT, 846) +Reducer 16 <- Map 23 (PARTITION-LEVEL SORT, 587), Reducer 15 (PARTITION-LEVEL SORT, 587) Reducer 17 <- Reducer 16 (GROUP, 640) Reducer 18 <- Reducer 17 (GROUP, 1) A masked pattern was here @@ -162,16 +162,33 @@ STAGE PLANS: 1 Map 19 Statistics: Num rows: 633595212 Data size: 55895953508 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator - key expressions: _col1 (type: int) - sort order: + - Map-reduce partition columns: _col1 (type: int) + key expressions: _col0 (type: int), _col3 (type: int) + sort order: ++ + Map-reduce partition columns: _col0 (type: int), _col3 (type: int) Statistics: Num rows: 633595212 Data size: 55895953508 Basic stats: COMPLETE Column stats: NONE - value expressions: _col0 (type: int), _col3 (type: int), _col4 (type: decimal(7,2)), _col6 (type: string), _col8 (type: string), _col9 (type: string) + value expressions: _col1 (type: int), _col4 (type: decimal(7,2)), _col6 (type: string), _col8 (type: string), _col9 (type: string) Local Work: Map Reduce Local Work Map 20 Map Operator Tree: TableScan + alias: store_returns + Statistics: Num rows: 57591150 Data size: 4462194832 Basic stats: COMPLETE Column stats: NONE + Filter Operator +predicate: (sr_item_sk is not null and sr_ticket_number is not null) (type: boolean) +Statistics: Num rows: 57591150 Data size: 4462194832 Basic stats: COMPLETE Column stats: NONE +Select Operator + expressions: sr_item_sk (type: int), sr_ticket_number (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 57591150 Data size: 4462194832 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator +key expressions: _col0 (type: int), _col1 (type: int) +sort order: ++ +Map-reduce partition columns: _col0 (type: int), _col1 (type: int) +Statistics: Num rows: 57591150 Data size: 4462194832 Basic stats: COMPLETE Column stats: NONE +Map 21 +Map Operator Tree: +TableScan alias: customer Statistics: Num rows: 8000 Data size: 68801615852 Basic stats: COMPLETE Column stats: NONE Filter Operator @@ -187,44 +204,9 @@ STAGE PLANS: Map-reduce partition columns: _col0 (type: int) Statistics: Num rows: 8000 Data size: 68801615852 Basic stats: COMPLETE Column stats: NONE value expressions: _col1 (type: string), _col2 (type: string), _col3 (type: string) -Map 21 -Map Operator Tree: -TableScan - alias: item - Statistics: Num rows: 462000 Data size: 663560457 Basic stats: COMPLETE Column stats: NONE - Filter Operator -predicate: i_item_sk is not null (type: boolean) -Statistics: Num rows: 462000 Data size: 663560457 Basic stats: COMPLETE Column stats: NONE -Select Operator - expressions: i_item_sk (type: int), i_current_price (type: decimal(7,2)), i_size (type: string),
[13/13] hive git commit: HIVE-19128 : Update golden files for spark perf tests
HIVE-19128 : Update golden files for spark perf tests Signed-off-by: Ashutosh Chauhan Project: http://git-wip-us.apache.org/repos/asf/hive/repo Commit: http://git-wip-us.apache.org/repos/asf/hive/commit/328d3f93 Tree: http://git-wip-us.apache.org/repos/asf/hive/tree/328d3f93 Diff: http://git-wip-us.apache.org/repos/asf/hive/diff/328d3f93 Branch: refs/heads/master Commit: 328d3f93596c82ac79544f111200659e28f5f8fc Parents: 2d3a410 Author: Ashutosh Chauhan Authored: Mon Apr 9 14:57:11 2018 -0700 Committer: Ashutosh Chauhan Committed: Mon Apr 9 14:57:11 2018 -0700 -- .../clientpositive/perf/spark/query11.q.out | 48 +- .../clientpositive/perf/spark/query15.q.out | 164 +++ .../clientpositive/perf/spark/query16.q.out | 70 ++- .../clientpositive/perf/spark/query18.q.out | 216 .../clientpositive/perf/spark/query19.q.out | 218 - .../clientpositive/perf/spark/query21.q.out | 114 ++--- .../clientpositive/perf/spark/query24.q.out | 282 +-- .../clientpositive/perf/spark/query25.q.out | 118 ++--- .../clientpositive/perf/spark/query29.q.out | 280 ++- .../clientpositive/perf/spark/query30.q.out | 262 +- .../clientpositive/perf/spark/query32.q.out | 132 ++--- .../clientpositive/perf/spark/query34.q.out | 34 +- .../clientpositive/perf/spark/query35.q.out | 74 +-- .../clientpositive/perf/spark/query37.q.out | 16 +- .../clientpositive/perf/spark/query4.q.out | 214 .../clientpositive/perf/spark/query40.q.out | 116 ++--- .../clientpositive/perf/spark/query44.q.out | 246 +- .../clientpositive/perf/spark/query45.q.out | 204 .../clientpositive/perf/spark/query46.q.out | 104 ++-- .../clientpositive/perf/spark/query47.q.out | 92 ++-- .../clientpositive/perf/spark/query48.q.out | 94 ++-- .../clientpositive/perf/spark/query5.q.out | 38 +- .../clientpositive/perf/spark/query50.q.out | 196 .../clientpositive/perf/spark/query53.q.out | 105 ++-- .../clientpositive/perf/spark/query54.q.out | 277 +-- .../clientpositive/perf/spark/query57.q.out | 92 ++-- .../clientpositive/perf/spark/query58.q.out | 490 +-- .../clientpositive/perf/spark/query6.q.out | 350 ++--- .../clientpositive/perf/spark/query61.q.out | 40 +- .../clientpositive/perf/spark/query63.q.out | 105 ++-- .../clientpositive/perf/spark/query65.q.out | 100 ++-- .../clientpositive/perf/spark/query66.q.out | 20 +- .../clientpositive/perf/spark/query67.q.out | 137 +++--- .../clientpositive/perf/spark/query68.q.out | 104 ++-- .../clientpositive/perf/spark/query72.q.out | 461 - .../clientpositive/perf/spark/query73.q.out | 34 +- .../clientpositive/perf/spark/query75.q.out | 248 +- .../clientpositive/perf/spark/query76.q.out | 130 ++--- .../clientpositive/perf/spark/query77.q.out | 66 +-- .../clientpositive/perf/spark/query78.q.out | 273 ++- .../clientpositive/perf/spark/query79.q.out | 10 +- .../clientpositive/perf/spark/query8.q.out | 174 +++ .../clientpositive/perf/spark/query80.q.out | 417 .../clientpositive/perf/spark/query81.q.out | 270 +- .../clientpositive/perf/spark/query82.q.out | 16 +- .../clientpositive/perf/spark/query83.q.out | 354 +++--- .../clientpositive/perf/spark/query85.q.out | 309 ++-- .../clientpositive/perf/spark/query88.q.out | 272 +- .../clientpositive/perf/spark/query89.q.out | 26 +- .../clientpositive/perf/spark/query90.q.out | 68 +-- .../clientpositive/perf/spark/query91.q.out | 14 +- .../clientpositive/perf/spark/query92.q.out | 106 ++-- .../clientpositive/perf/spark/query94.q.out | 70 ++- .../clientpositive/perf/spark/query95.q.out | 100 ++-- .../clientpositive/perf/spark/query97.q.out | 54 +- .../clientpositive/perf/spark/query99.q.out | 42 +- 56 files changed, 4313 insertions(+), 4353 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hive/blob/328d3f93/ql/src/test/results/clientpositive/perf/spark/query11.q.out -- diff --git a/ql/src/test/results/clientpositive/perf/spark/query11.q.out b/ql/src/test/results/clientpositive/perf/spark/query11.q.out index e46aa21..227068c 100644 --- a/ql/src/test/results/clientpositive/perf/spark/query11.q.out +++ b/ql/src/test/results/clientpositive/perf/spark/query11.q.out @@ -300,7 +300,7 @@ STAGE PLANS: alias: date_dim Statistics: Num rows: 73049 Data size: 81741831 Basic stats: COMPLETE Column stats: NONE Filter Operator -predicate: ((d_y
[03/13] hive git commit: HIVE-19128 : Update golden files for spark perf tests
http://git-wip-us.apache.org/repos/asf/hive/blob/328d3f93/ql/src/test/results/clientpositive/perf/spark/query81.q.out -- diff --git a/ql/src/test/results/clientpositive/perf/spark/query81.q.out b/ql/src/test/results/clientpositive/perf/spark/query81.q.out index 5b2d5b3..23998d6 100644 --- a/ql/src/test/results/clientpositive/perf/spark/query81.q.out +++ b/ql/src/test/results/clientpositive/perf/spark/query81.q.out @@ -66,72 +66,37 @@ STAGE PLANS: Stage: Stage-1 Spark Edges: -Reducer 10 <- Reducer 16 (PARTITION-LEVEL SORT, 262), Reducer 9 (PARTITION-LEVEL SORT, 262) -Reducer 14 <- Map 13 (PARTITION-LEVEL SORT, 25), Map 17 (PARTITION-LEVEL SORT, 25) -Reducer 15 <- Map 18 (PARTITION-LEVEL SORT, 344), Reducer 14 (PARTITION-LEVEL SORT, 344) -Reducer 16 <- Reducer 15 (GROUP PARTITION-LEVEL SORT, 349) -Reducer 2 <- Map 1 (PARTITION-LEVEL SORT, 697), Map 5 (PARTITION-LEVEL SORT, 697) -Reducer 3 <- Reducer 10 (PARTITION-LEVEL SORT, 656), Reducer 2 (PARTITION-LEVEL SORT, 656) -Reducer 4 <- Reducer 3 (SORT, 1) -Reducer 7 <- Map 11 (PARTITION-LEVEL SORT, 25), Map 6 (PARTITION-LEVEL SORT, 25) -Reducer 8 <- Map 12 (PARTITION-LEVEL SORT, 344), Reducer 7 (PARTITION-LEVEL SORT, 344) -Reducer 9 <- Reducer 8 (GROUP, 349) +Reducer 11 <- Map 10 (PARTITION-LEVEL SORT, 25), Map 14 (PARTITION-LEVEL SORT, 25) +Reducer 12 <- Map 15 (PARTITION-LEVEL SORT, 344), Reducer 11 (PARTITION-LEVEL SORT, 344) +Reducer 13 <- Reducer 12 (GROUP PARTITION-LEVEL SORT, 349) +Reducer 17 <- Map 16 (PARTITION-LEVEL SORT, 697), Map 18 (PARTITION-LEVEL SORT, 697) +Reducer 2 <- Map 1 (PARTITION-LEVEL SORT, 25), Map 8 (PARTITION-LEVEL SORT, 25) +Reducer 3 <- Map 9 (PARTITION-LEVEL SORT, 344), Reducer 2 (PARTITION-LEVEL SORT, 344) +Reducer 4 <- Reducer 3 (GROUP, 349) +Reducer 5 <- Reducer 13 (PARTITION-LEVEL SORT, 262), Reducer 4 (PARTITION-LEVEL SORT, 262) +Reducer 6 <- Reducer 17 (PARTITION-LEVEL SORT, 656), Reducer 5 (PARTITION-LEVEL SORT, 656) +Reducer 7 <- Reducer 6 (SORT, 1) A masked pattern was here Vertices: Map 1 Map Operator Tree: TableScan - alias: customer - Statistics: Num rows: 8000 Data size: 68801615852 Basic stats: COMPLETE Column stats: NONE - Filter Operator -predicate: (c_current_addr_sk is not null and c_customer_sk is not null) (type: boolean) -Statistics: Num rows: 8000 Data size: 68801615852 Basic stats: COMPLETE Column stats: NONE -Select Operator - expressions: c_customer_sk (type: int), c_customer_id (type: string), c_current_addr_sk (type: int), c_salutation (type: string), c_first_name (type: string), c_last_name (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 - Statistics: Num rows: 8000 Data size: 68801615852 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator -key expressions: _col2 (type: int) -sort order: + -Map-reduce partition columns: _col2 (type: int) -Statistics: Num rows: 8000 Data size: 68801615852 Basic stats: COMPLETE Column stats: NONE -value expressions: _col0 (type: int), _col1 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: string) -Map 11 -Map Operator Tree: -TableScan - alias: date_dim - Statistics: Num rows: 73049 Data size: 81741831 Basic stats: COMPLETE Column stats: NONE - Filter Operator -predicate: ((d_year = 1998) and d_date_sk is not null) (type: boolean) -Statistics: Num rows: 36524 Data size: 40870356 Basic stats: COMPLETE Column stats: NONE -Select Operator - expressions: d_date_sk (type: int) - outputColumnNames: _col0 - Statistics: Num rows: 36524 Data size: 40870356 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator -key expressions: _col0 (type: int) -sort order: + -Map-reduce partition columns: _col0 (type: int) -Statistics: Num rows: 36524 Data size: 40870356 Basic stats: COMPLETE Column stats: NONE -Map 12 -Map Operator Tree: -TableScan - alias: customer_address - Statistics: Num rows: 4000 Data size: 40595195284 Basic stats: COMPLETE Column stats: NONE +
[05/13] hive git commit: HIVE-19128 : Update golden files for spark perf tests
http://git-wip-us.apache.org/repos/asf/hive/blob/328d3f93/ql/src/test/results/clientpositive/perf/spark/query75.q.out -- diff --git a/ql/src/test/results/clientpositive/perf/spark/query75.q.out b/ql/src/test/results/clientpositive/perf/spark/query75.q.out index 4b76570..f4fca34 100644 --- a/ql/src/test/results/clientpositive/perf/spark/query75.q.out +++ b/ql/src/test/results/clientpositive/perf/spark/query75.q.out @@ -144,25 +144,25 @@ STAGE PLANS: Stage: Stage-1 Spark Edges: -Reducer 13 <- Map 12 (PARTITION-LEVEL SORT, 398), Map 16 (PARTITION-LEVEL SORT, 398) -Reducer 14 <- Map 10 (PARTITION-LEVEL SORT, 440), Reducer 13 (PARTITION-LEVEL SORT, 440) +Reducer 13 <- Map 12 (PARTITION-LEVEL SORT, 400), Map 16 (PARTITION-LEVEL SORT, 400) +Reducer 14 <- Map 10 (PARTITION-LEVEL SORT, 438), Reducer 13 (PARTITION-LEVEL SORT, 438) Reducer 15 <- Map 18 (PARTITION-LEVEL SORT, 516), Reducer 14 (PARTITION-LEVEL SORT, 516) -Reducer 2 <- Map 1 (PARTITION-LEVEL SORT, 306), Map 16 (PARTITION-LEVEL SORT, 306) -Reducer 20 <- Map 16 (PARTITION-LEVEL SORT, 154), Map 19 (PARTITION-LEVEL SORT, 154) -Reducer 21 <- Map 10 (PARTITION-LEVEL SORT, 171), Reducer 20 (PARTITION-LEVEL SORT, 171) +Reducer 2 <- Map 1 (PARTITION-LEVEL SORT, 308), Map 16 (PARTITION-LEVEL SORT, 308) +Reducer 20 <- Map 16 (PARTITION-LEVEL SORT, 156), Map 19 (PARTITION-LEVEL SORT, 156) +Reducer 21 <- Map 10 (PARTITION-LEVEL SORT, 169), Reducer 20 (PARTITION-LEVEL SORT, 169) Reducer 22 <- Map 25 (PARTITION-LEVEL SORT, 196), Reducer 21 (PARTITION-LEVEL SORT, 196) -Reducer 27 <- Map 26 (PARTITION-LEVEL SORT, 306), Map 32 (PARTITION-LEVEL SORT, 306) -Reducer 28 <- Map 33 (PARTITION-LEVEL SORT, 338), Reducer 27 (PARTITION-LEVEL SORT, 338) +Reducer 27 <- Map 26 (PARTITION-LEVEL SORT, 308), Map 32 (PARTITION-LEVEL SORT, 308) +Reducer 28 <- Map 33 (PARTITION-LEVEL SORT, 336), Reducer 27 (PARTITION-LEVEL SORT, 336) Reducer 29 <- Map 34 (PARTITION-LEVEL SORT, 393), Reducer 28 (PARTITION-LEVEL SORT, 393) -Reducer 3 <- Map 10 (PARTITION-LEVEL SORT, 338), Reducer 2 (PARTITION-LEVEL SORT, 338) +Reducer 3 <- Map 10 (PARTITION-LEVEL SORT, 336), Reducer 2 (PARTITION-LEVEL SORT, 336) Reducer 30 <- Reducer 29 (GROUP, 934), Reducer 38 (GROUP, 934) Reducer 31 <- Reducer 30 (GROUP PARTITION-LEVEL SORT, 671), Reducer 45 (GROUP PARTITION-LEVEL SORT, 671) -Reducer 36 <- Map 32 (PARTITION-LEVEL SORT, 398), Map 35 (PARTITION-LEVEL SORT, 398) -Reducer 37 <- Map 33 (PARTITION-LEVEL SORT, 440), Reducer 36 (PARTITION-LEVEL SORT, 440) +Reducer 36 <- Map 32 (PARTITION-LEVEL SORT, 400), Map 35 (PARTITION-LEVEL SORT, 400) +Reducer 37 <- Map 33 (PARTITION-LEVEL SORT, 438), Reducer 36 (PARTITION-LEVEL SORT, 438) Reducer 38 <- Map 41 (PARTITION-LEVEL SORT, 516), Reducer 37 (PARTITION-LEVEL SORT, 516) Reducer 4 <- Map 11 (PARTITION-LEVEL SORT, 393), Reducer 3 (PARTITION-LEVEL SORT, 393) -Reducer 43 <- Map 32 (PARTITION-LEVEL SORT, 154), Map 42 (PARTITION-LEVEL SORT, 154) -Reducer 44 <- Map 33 (PARTITION-LEVEL SORT, 171), Reducer 43 (PARTITION-LEVEL SORT, 171) +Reducer 43 <- Map 32 (PARTITION-LEVEL SORT, 156), Map 42 (PARTITION-LEVEL SORT, 156) +Reducer 44 <- Map 33 (PARTITION-LEVEL SORT, 169), Reducer 43 (PARTITION-LEVEL SORT, 169) Reducer 45 <- Map 48 (PARTITION-LEVEL SORT, 196), Reducer 44 (PARTITION-LEVEL SORT, 196) Reducer 5 <- Reducer 15 (GROUP, 934), Reducer 4 (GROUP, 934) Reducer 6 <- Reducer 22 (GROUP PARTITION-LEVEL SORT, 671), Reducer 5 (GROUP PARTITION-LEVEL SORT, 671) @@ -183,29 +183,28 @@ STAGE PLANS: outputColumnNames: _col0, _col1, _col2, _col3, _col4 Statistics: Num rows: 287989836 Data size: 38999608952 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator -key expressions: _col0 (type: int) +key expressions: _col1 (type: int) sort order: + -Map-reduce partition columns: _col0 (type: int) +Map-reduce partition columns: _col1 (type: int) Statistics: Num rows: 287989836 Data size: 38999608952 Basic stats: COMPLETE Column stats: NONE -value expressions: _col1 (type: int), _col2 (type: int), _col3 (type: int), _col4 (type: decimal(7,2)) +value expressions: _col0 (type: int), _col2 (type: int), _col3 (type: int), _col4 (type: decimal(7,2)) Map 10 Map Operator Tree: TableScan - alias: item - Statistics: Num rows: 462000 Data size: 663560457 Basic stats:
[10/13] hive git commit: HIVE-19128 : Update golden files for spark perf tests
http://git-wip-us.apache.org/repos/asf/hive/blob/328d3f93/ql/src/test/results/clientpositive/perf/spark/query40.q.out -- diff --git a/ql/src/test/results/clientpositive/perf/spark/query40.q.out b/ql/src/test/results/clientpositive/perf/spark/query40.q.out index f286294..5360385 100644 --- a/ql/src/test/results/clientpositive/perf/spark/query40.q.out +++ b/ql/src/test/results/clientpositive/perf/spark/query40.q.out @@ -54,8 +54,7 @@ limit 100 POSTHOOK: type: QUERY STAGE DEPENDENCIES: Stage-2 is a root stage - Stage-3 depends on stages: Stage-2 - Stage-1 depends on stages: Stage-3 + Stage-1 depends on stages: Stage-2 Stage-0 depends on stages: Stage-1 STAGE PLANS: @@ -63,44 +62,39 @@ STAGE PLANS: Spark A masked pattern was here Vertices: -Map 9 +Map 8 Map Operator Tree: TableScan - alias: warehouse - Statistics: Num rows: 27 Data size: 27802 Basic stats: COMPLETE Column stats: NONE + alias: date_dim + Statistics: Num rows: 73049 Data size: 81741831 Basic stats: COMPLETE Column stats: NONE Filter Operator -predicate: w_warehouse_sk is not null (type: boolean) -Statistics: Num rows: 27 Data size: 27802 Basic stats: COMPLETE Column stats: NONE +predicate: (CAST( d_date AS TIMESTAMP) BETWEEN TIMESTAMP'1998-03-08 23:00:00.0' AND TIMESTAMP'1998-05-08 00:00:00.0' and d_date_sk is not null) (type: boolean) +Statistics: Num rows: 8116 Data size: 9081804 Basic stats: COMPLETE Column stats: NONE Select Operator - expressions: w_warehouse_sk (type: int), w_state (type: string) + expressions: d_date_sk (type: int), d_date (type: string) outputColumnNames: _col0, _col1 - Statistics: Num rows: 27 Data size: 27802 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 8116 Data size: 9081804 Basic stats: COMPLETE Column stats: NONE Spark HashTable Sink Operator keys: - 0 _col1 (type: int) + 0 _col0 (type: int) 1 _col0 (type: int) Local Work: Map Reduce Local Work - - Stage: Stage-3 -Spark - A masked pattern was here - Vertices: -Map 7 +Map 9 Map Operator Tree: TableScan - alias: date_dim - Statistics: Num rows: 73049 Data size: 81741831 Basic stats: COMPLETE Column stats: NONE + alias: warehouse + Statistics: Num rows: 27 Data size: 27802 Basic stats: COMPLETE Column stats: NONE Filter Operator -predicate: (CAST( d_date AS TIMESTAMP) BETWEEN TIMESTAMP'1998-03-08 23:00:00.0' AND TIMESTAMP'1998-05-08 00:00:00.0' and d_date_sk is not null) (type: boolean) -Statistics: Num rows: 8116 Data size: 9081804 Basic stats: COMPLETE Column stats: NONE +predicate: w_warehouse_sk is not null (type: boolean) +Statistics: Num rows: 27 Data size: 27802 Basic stats: COMPLETE Column stats: NONE Select Operator - expressions: d_date_sk (type: int), d_date (type: string) + expressions: w_warehouse_sk (type: int), w_state (type: string) outputColumnNames: _col0, _col1 - Statistics: Num rows: 8116 Data size: 9081804 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 27 Data size: 27802 Basic stats: COMPLETE Column stats: NONE Spark HashTable Sink Operator keys: - 0 _col0 (type: int) + 0 _col1 (type: int) 1 _col0 (type: int) Local Work: Map Reduce Local Work @@ -109,7 +103,7 @@ STAGE PLANS: Spark Edges: Reducer 2 <- Map 1 (PARTITION-LEVEL SORT, 329), Map 6 (PARTITION-LEVEL SORT, 329) -Reducer 3 <- Map 8 (PARTITION-LEVEL SORT, 370), Reducer 2 (PARTITION-LEVEL SORT, 370) +Reducer 3 <- Map 7 (PARTITION-LEVEL SORT, 336), Reducer 2 (PARTITION-LEVEL SORT, 336) Reducer 4 <- Reducer 3 (GROUP, 447) Reducer 5 <- Reducer 4 (SORT, 1) A masked pattern was here @@ -150,7 +144,7 @@ STAGE PLANS: Map-reduce partition columns: _col0 (type: int), _col1 (type: int) Statistics: Num rows: 28798881 Data size: 3057234680 Basic stats: COMPLETE Column stats: NONE value expressions: _col2 (type:
[07/13] hive git commit: HIVE-19128 : Update golden files for spark perf tests
http://git-wip-us.apache.org/repos/asf/hive/blob/328d3f93/ql/src/test/results/clientpositive/perf/spark/query58.q.out -- diff --git a/ql/src/test/results/clientpositive/perf/spark/query58.q.out b/ql/src/test/results/clientpositive/perf/spark/query58.q.out index f06cbef..eb5dffe 100644 --- a/ql/src/test/results/clientpositive/perf/spark/query58.q.out +++ b/ql/src/test/results/clientpositive/perf/spark/query58.q.out @@ -1,6 +1,6 @@ -Warning: Map Join MAPJOIN[183][bigTable=?] in task 'Stage-1:MAPRED' is a cross product -Warning: Map Join MAPJOIN[184][bigTable=?] in task 'Stage-1:MAPRED' is a cross product -Warning: Map Join MAPJOIN[185][bigTable=?] in task 'Stage-1:MAPRED' is a cross product +Warning: Map Join MAPJOIN[180][bigTable=?] in task 'Stage-1:MAPRED' is a cross product +Warning: Map Join MAPJOIN[181][bigTable=?] in task 'Stage-1:MAPRED' is a cross product +Warning: Map Join MAPJOIN[182][bigTable=?] in task 'Stage-1:MAPRED' is a cross product PREHOOK: query: explain with ss_items as (select i_item_id item_id @@ -140,10 +140,10 @@ STAGE PLANS: Stage: Stage-2 Spark Edges: -Reducer 11 <- Map 10 (GROUP, 1) +Reducer 12 <- Map 11 (GROUP, 1) A masked pattern was here Vertices: -Map 10 +Map 11 Map Operator Tree: TableScan alias: date_dim @@ -162,7 +162,7 @@ STAGE PLANS: sort order: Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE value expressions: _col0 (type: bigint) -Reducer 11 +Reducer 12 Local Work: Map Reduce Local Work Reduce Operator Tree: @@ -184,10 +184,10 @@ STAGE PLANS: Stage: Stage-3 Spark Edges: -Reducer 24 <- Map 23 (GROUP, 1) +Reducer 25 <- Map 24 (GROUP, 1) A masked pattern was here Vertices: -Map 23 +Map 24 Map Operator Tree: TableScan alias: date_dim @@ -206,7 +206,7 @@ STAGE PLANS: sort order: Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE value expressions: _col0 (type: bigint) -Reducer 24 +Reducer 25 Local Work: Map Reduce Local Work Reduce Operator Tree: @@ -228,10 +228,10 @@ STAGE PLANS: Stage: Stage-4 Spark Edges: -Reducer 37 <- Map 36 (GROUP, 1) +Reducer 38 <- Map 37 (GROUP, 1) A masked pattern was here Vertices: -Map 36 +Map 37 Map Operator Tree: TableScan alias: date_dim @@ -250,7 +250,7 @@ STAGE PLANS: sort order: Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE value expressions: _col0 (type: bigint) -Reducer 37 +Reducer 38 Local Work: Map Reduce Local Work Reduce Operator Tree: @@ -272,76 +272,47 @@ STAGE PLANS: Stage: Stage-1 Spark Edges: -Reducer 13 <- Map 12 (PARTITION-LEVEL SORT, 2), Map 15 (PARTITION-LEVEL SORT, 2) -Reducer 14 <- Reducer 13 (GROUP, 2) -Reducer 17 <- Map 16 (PARTITION-LEVEL SORT, 403), Map 20 (PARTITION-LEVEL SORT, 403) -Reducer 18 <- Reducer 17 (PARTITION-LEVEL SORT, 438), Reducer 22 (PARTITION-LEVEL SORT, 438) -Reducer 19 <- Reducer 18 (GROUP, 481) -Reducer 2 <- Map 1 (PARTITION-LEVEL SORT, 310), Map 7 (PARTITION-LEVEL SORT, 310) -Reducer 22 <- Map 21 (PARTITION-LEVEL SORT, 2), Reducer 27 (PARTITION-LEVEL SORT, 2) -Reducer 26 <- Map 25 (PARTITION-LEVEL SORT, 2), Map 28 (PARTITION-LEVEL SORT, 2) -Reducer 27 <- Reducer 26 (GROUP, 2) -Reducer 3 <- Reducer 2 (PARTITION-LEVEL SORT, 336), Reducer 9 (PARTITION-LEVEL SORT, 336) -Reducer 30 <- Map 29 (PARTITION-LEVEL SORT, 159), Map 33 (PARTITION-LEVEL SORT, 159) -Reducer 31 <- Reducer 30 (PARTITION-LEVEL SORT, 169), Reducer 35 (PARTITION-LEVEL SORT, 169) -Reducer 32 <- Reducer 31 (GROUP, 186) -Reducer 35 <- Map 34 (PARTITION-LEVEL SORT, 2), Reducer 40 (PARTITION-LEVEL SORT, 2) -Reducer 39 <- Map 38 (PARTITION-LEVEL SORT, 2), Map 41 (PARTITION-LEVEL SORT, 2) -Reducer 4 <- Reducer 3 (GROUP, 369) -Reducer 40 <- Reducer 39 (GROUP, 2) -Reducer 5 <- Reducer 19 (PARTITION-LEVEL SORT, 518), Reducer 32 (PARTITION-LEVEL SORT, 518), Reducer 4 (PARTITION-LEVEL SORT, 518) -Reducer 6 <- Reducer 5 (SORT, 1) -Reducer 9 <- Map 8 (PARTITION-LEVEL SORT, 2), Reducer 14 (PARTITION-LEVEL SORT, 2) +Reducer 10 <- Reducer
[06/13] hive git commit: HIVE-19128 : Update golden files for spark perf tests
http://git-wip-us.apache.org/repos/asf/hive/blob/328d3f93/ql/src/test/results/clientpositive/perf/spark/query65.q.out -- diff --git a/ql/src/test/results/clientpositive/perf/spark/query65.q.out b/ql/src/test/results/clientpositive/perf/spark/query65.q.out index 3b3baef..575fc5c 100644 --- a/ql/src/test/results/clientpositive/perf/spark/query65.q.out +++ b/ql/src/test/results/clientpositive/perf/spark/query65.q.out @@ -57,7 +57,8 @@ POSTHOOK: type: QUERY STAGE DEPENDENCIES: Stage-2 is a root stage Stage-3 depends on stages: Stage-2 - Stage-1 depends on stages: Stage-3 + Stage-4 depends on stages: Stage-3 + Stage-1 depends on stages: Stage-4 Stage-0 depends on stages: Stage-1 STAGE PLANS: @@ -65,6 +66,29 @@ STAGE PLANS: Spark A masked pattern was here Vertices: +Map 11 +Map Operator Tree: +TableScan + alias: store + Statistics: Num rows: 1704 Data size: 3256276 Basic stats: COMPLETE Column stats: NONE + Filter Operator +predicate: s_store_sk is not null (type: boolean) +Statistics: Num rows: 1704 Data size: 3256276 Basic stats: COMPLETE Column stats: NONE +Select Operator + expressions: s_store_sk (type: int), s_store_name (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1704 Data size: 3256276 Basic stats: COMPLETE Column stats: NONE + Spark HashTable Sink Operator +keys: + 0 _col0 (type: int) + 1 _col0 (type: int) +Local Work: + Map Reduce Local Work + + Stage: Stage-3 +Spark + A masked pattern was here + Vertices: Map 6 Map Operator Tree: TableScan @@ -84,7 +108,7 @@ STAGE PLANS: Local Work: Map Reduce Local Work - Stage: Stage-3 + Stage: Stage-4 Spark A masked pattern was here Vertices: @@ -111,8 +135,8 @@ STAGE PLANS: Spark Edges: Reducer 2 <- Map 1 (GROUP, 437) -Reducer 3 <- Map 10 (PARTITION-LEVEL SORT, 328), Reducer 2 (PARTITION-LEVEL SORT, 328), Reducer 8 (PARTITION-LEVEL SORT, 328) -Reducer 4 <- Map 11 (PARTITION-LEVEL SORT, 166), Reducer 3 (PARTITION-LEVEL SORT, 166) +Reducer 3 <- Reducer 2 (PARTITION-LEVEL SORT, 328), Reducer 8 (PARTITION-LEVEL SORT, 328) +Reducer 4 <- Map 10 (PARTITION-LEVEL SORT, 86), Reducer 3 (PARTITION-LEVEL SORT, 86) Reducer 5 <- Reducer 4 (SORT, 1) Reducer 8 <- Map 7 (GROUP PARTITION-LEVEL SORT, 437) A masked pattern was here @@ -156,24 +180,6 @@ STAGE PLANS: Map 10 Map Operator Tree: TableScan - alias: store - Statistics: Num rows: 1704 Data size: 3256276 Basic stats: COMPLETE Column stats: NONE - Filter Operator -predicate: s_store_sk is not null (type: boolean) -Statistics: Num rows: 1704 Data size: 3256276 Basic stats: COMPLETE Column stats: NONE -Select Operator - expressions: s_store_sk (type: int), s_store_name (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1704 Data size: 3256276 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator -key expressions: _col0 (type: int) -sort order: + -Map-reduce partition columns: _col0 (type: int) -Statistics: Num rows: 1704 Data size: 3256276 Basic stats: COMPLETE Column stats: NONE -value expressions: _col1 (type: string) -Map 11 -Map Operator Tree: -TableScan alias: item Statistics: Num rows: 462000 Data size: 663560457 Basic stats: COMPLETE Column stats: NONE Filter Operator @@ -244,23 +250,23 @@ STAGE PLANS: Join Operator condition map: Inner Join 0 to 1 - Inner Join 0 to 2 keys: 0 _col0 (type: int) 1 _col0 (type: int) - 2 _col0 (type: int) -outputColumnNames: _col1, _col2, _col4, _col6 -Statistics: Num rows: 696954748 Data size: 61485550191 Basic stats: COMPLETE Column stats: NONE +outputColumnNames: _col0, _col1, _col2, _col4 +Statistics: Num rows: 348477374 Data size: 30742775095 Basic stats: COMPLETE Column stats: NONE Filter Operator predica
[09/13] hive git commit: HIVE-19128 : Update golden files for spark perf tests
http://git-wip-us.apache.org/repos/asf/hive/blob/328d3f93/ql/src/test/results/clientpositive/perf/spark/query47.q.out -- diff --git a/ql/src/test/results/clientpositive/perf/spark/query47.q.out b/ql/src/test/results/clientpositive/perf/spark/query47.q.out index 5175f80..bc97fdf 100644 --- a/ql/src/test/results/clientpositive/perf/spark/query47.q.out +++ b/ql/src/test/results/clientpositive/perf/spark/query47.q.out @@ -397,47 +397,47 @@ STAGE PLANS: Statistics: Num rows: 766650239 Data size: 67634106676 Basic stats: COMPLETE Column stats: NONE Group By Operator aggregations: sum(_col3) -keys: _col5 (type: int), _col6 (type: int), _col8 (type: string), _col9 (type: string), _col11 (type: string), _col12 (type: string) +keys: _col11 (type: string), _col12 (type: string), _col8 (type: string), _col9 (type: string), _col5 (type: int), _col6 (type: int) mode: hash outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6 Statistics: Num rows: 766650239 Data size: 67634106676 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator - key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: string) + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: int), _col5 (type: int) sort order: ++ - Map-reduce partition columns: _col0 (type: int), _col1 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: string) + Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: int), _col5 (type: int) Statistics: Num rows: 766650239 Data size: 67634106676 Basic stats: COMPLETE Column stats: NONE value expressions: _col6 (type: decimal(17,2)) Reducer 14 Reduce Operator Tree: Group By Operator aggregations: sum(VALUE._col0) -keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: string), KEY._col3 (type: string), KEY._col4 (type: string), KEY._col5 (type: string) +keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: string), KEY._col3 (type: string), KEY._col4 (type: int), KEY._col5 (type: int) mode: mergepartial outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6 Statistics: Num rows: 383325119 Data size: 33817053293 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator - key expressions: _col3 (type: string), _col2 (type: string), _col4 (type: string), _col5 (type: string), _col0 (type: int) + key expressions: _col3 (type: string), _col2 (type: string), _col0 (type: string), _col1 (type: string), _col4 (type: int) sort order: + - Map-reduce partition columns: _col3 (type: string), _col2 (type: string), _col4 (type: string), _col5 (type: string), _col0 (type: int) + Map-reduce partition columns: _col3 (type: string), _col2 (type: string), _col0 (type: string), _col1 (type: string), _col4 (type: int) Statistics: Num rows: 383325119 Data size: 33817053293 Basic stats: COMPLETE Column stats: NONE - value expressions: _col1 (type: int), _col6 (type: decimal(17,2)) + value expressions: _col5 (type: int), _col6 (type: decimal(17,2)) Reducer 15 Reduce Operator Tree: Select Operator -expressions: KEY.reducesinkkey4 (type: int), VALUE._col0 (type: int), KEY.reducesinkkey1 (type: string), KEY.reducesinkkey0 (type: string), KEY.reducesinkkey2 (type: string), KEY.reducesinkkey3 (type: string), VALUE._col1 (type: decimal(17,2)) +expressions: KEY.reducesinkkey2 (type: string), KEY.reducesinkkey3 (type: string), KEY.reducesinkkey1 (type: string), KEY.reducesinkkey0 (type: string), KEY.reducesinkkey4 (type: int), VALUE._col0 (type: int), VALUE._col1 (type: decimal(17,2)) outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6 Statistics: Num rows: 383325119 Data size: 33817053293 Basic stats: COMPLETE Column stats: NONE PTF Operator Function definitions: Input definition input alias: ptf_0 -output shape: _col0: int, _col1: int, _col2: string, _col3: string, _col4: string
[06/13] hive git commit: HIVE-19128 : Update golden files for spark perf tests
http://git-wip-us.apache.org/repos/asf/hive/blob/9ce42cba/ql/src/test/results/clientpositive/perf/spark/query65.q.out -- diff --git a/ql/src/test/results/clientpositive/perf/spark/query65.q.out b/ql/src/test/results/clientpositive/perf/spark/query65.q.out index 3b3baef..575fc5c 100644 --- a/ql/src/test/results/clientpositive/perf/spark/query65.q.out +++ b/ql/src/test/results/clientpositive/perf/spark/query65.q.out @@ -57,7 +57,8 @@ POSTHOOK: type: QUERY STAGE DEPENDENCIES: Stage-2 is a root stage Stage-3 depends on stages: Stage-2 - Stage-1 depends on stages: Stage-3 + Stage-4 depends on stages: Stage-3 + Stage-1 depends on stages: Stage-4 Stage-0 depends on stages: Stage-1 STAGE PLANS: @@ -65,6 +66,29 @@ STAGE PLANS: Spark A masked pattern was here Vertices: +Map 11 +Map Operator Tree: +TableScan + alias: store + Statistics: Num rows: 1704 Data size: 3256276 Basic stats: COMPLETE Column stats: NONE + Filter Operator +predicate: s_store_sk is not null (type: boolean) +Statistics: Num rows: 1704 Data size: 3256276 Basic stats: COMPLETE Column stats: NONE +Select Operator + expressions: s_store_sk (type: int), s_store_name (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1704 Data size: 3256276 Basic stats: COMPLETE Column stats: NONE + Spark HashTable Sink Operator +keys: + 0 _col0 (type: int) + 1 _col0 (type: int) +Local Work: + Map Reduce Local Work + + Stage: Stage-3 +Spark + A masked pattern was here + Vertices: Map 6 Map Operator Tree: TableScan @@ -84,7 +108,7 @@ STAGE PLANS: Local Work: Map Reduce Local Work - Stage: Stage-3 + Stage: Stage-4 Spark A masked pattern was here Vertices: @@ -111,8 +135,8 @@ STAGE PLANS: Spark Edges: Reducer 2 <- Map 1 (GROUP, 437) -Reducer 3 <- Map 10 (PARTITION-LEVEL SORT, 328), Reducer 2 (PARTITION-LEVEL SORT, 328), Reducer 8 (PARTITION-LEVEL SORT, 328) -Reducer 4 <- Map 11 (PARTITION-LEVEL SORT, 166), Reducer 3 (PARTITION-LEVEL SORT, 166) +Reducer 3 <- Reducer 2 (PARTITION-LEVEL SORT, 328), Reducer 8 (PARTITION-LEVEL SORT, 328) +Reducer 4 <- Map 10 (PARTITION-LEVEL SORT, 86), Reducer 3 (PARTITION-LEVEL SORT, 86) Reducer 5 <- Reducer 4 (SORT, 1) Reducer 8 <- Map 7 (GROUP PARTITION-LEVEL SORT, 437) A masked pattern was here @@ -156,24 +180,6 @@ STAGE PLANS: Map 10 Map Operator Tree: TableScan - alias: store - Statistics: Num rows: 1704 Data size: 3256276 Basic stats: COMPLETE Column stats: NONE - Filter Operator -predicate: s_store_sk is not null (type: boolean) -Statistics: Num rows: 1704 Data size: 3256276 Basic stats: COMPLETE Column stats: NONE -Select Operator - expressions: s_store_sk (type: int), s_store_name (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1704 Data size: 3256276 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator -key expressions: _col0 (type: int) -sort order: + -Map-reduce partition columns: _col0 (type: int) -Statistics: Num rows: 1704 Data size: 3256276 Basic stats: COMPLETE Column stats: NONE -value expressions: _col1 (type: string) -Map 11 -Map Operator Tree: -TableScan alias: item Statistics: Num rows: 462000 Data size: 663560457 Basic stats: COMPLETE Column stats: NONE Filter Operator @@ -244,23 +250,23 @@ STAGE PLANS: Join Operator condition map: Inner Join 0 to 1 - Inner Join 0 to 2 keys: 0 _col0 (type: int) 1 _col0 (type: int) - 2 _col0 (type: int) -outputColumnNames: _col1, _col2, _col4, _col6 -Statistics: Num rows: 696954748 Data size: 61485550191 Basic stats: COMPLETE Column stats: NONE +outputColumnNames: _col0, _col1, _col2, _col4 +Statistics: Num rows: 348477374 Data size: 30742775095 Basic stats: COMPLETE Column stats: NONE Filter Operator predica
[02/13] hive git commit: HIVE-19128 : Update golden files for spark perf tests
http://git-wip-us.apache.org/repos/asf/hive/blob/9ce42cba/ql/src/test/results/clientpositive/perf/spark/query85.q.out -- diff --git a/ql/src/test/results/clientpositive/perf/spark/query85.q.out b/ql/src/test/results/clientpositive/perf/spark/query85.q.out index d60751c..c82dcf7 100644 --- a/ql/src/test/results/clientpositive/perf/spark/query85.q.out +++ b/ql/src/test/results/clientpositive/perf/spark/query85.q.out @@ -175,7 +175,7 @@ STAGE PLANS: Spark A masked pattern was here Vertices: -Map 13 +Map 15 Map Operator Tree: TableScan alias: reason @@ -212,21 +212,21 @@ STAGE PLANS: Statistics: Num rows: 4602 Data size: 2696178 Basic stats: COMPLETE Column stats: NONE Spark HashTable Sink Operator keys: - 0 _col10 (type: int) - 1 _col0 (type: int) + 0 _col0 (type: int) + 1 _col2 (type: int) Local Work: Map Reduce Local Work Stage: Stage-1 Spark Edges: -Reducer 2 <- Map 1 (PARTITION-LEVEL SORT, 28), Map 9 (PARTITION-LEVEL SORT, 28) -Reducer 3 <- Map 10 (PARTITION-LEVEL SORT, 98), Reducer 2 (PARTITION-LEVEL SORT, 98) -Reducer 4 <- Map 12 (PARTITION-LEVEL SORT, 5), Reducer 3 (PARTITION-LEVEL SORT, 5) -Reducer 5 <- Map 14 (PARTITION-LEVEL SORT, 11), Reducer 4 (PARTITION-LEVEL SORT, 11) -Reducer 6 <- Map 15 (PARTITION-LEVEL SORT, 7), Reducer 5 (PARTITION-LEVEL SORT, 7) -Reducer 7 <- Reducer 6 (GROUP, 7) -Reducer 8 <- Reducer 7 (SORT, 1) +Reducer 13 <- Map 12 (PARTITION-LEVEL SORT, 20), Map 14 (PARTITION-LEVEL SORT, 20) +Reducer 2 <- Map 1 (PARTITION-LEVEL SORT, 90), Map 8 (PARTITION-LEVEL SORT, 90) +Reducer 3 <- Map 9 (PARTITION-LEVEL SORT, 17), Reducer 2 (PARTITION-LEVEL SORT, 17) +Reducer 4 <- Map 10 (PARTITION-LEVEL SORT, 19), Reducer 3 (PARTITION-LEVEL SORT, 19) +Reducer 5 <- Reducer 13 (PARTITION-LEVEL SORT, 35), Reducer 4 (PARTITION-LEVEL SORT, 35) +Reducer 6 <- Reducer 5 (GROUP, 2) +Reducer 7 <- Reducer 6 (SORT, 1) A masked pattern was here Vertices: Map 1 @@ -242,32 +242,61 @@ STAGE PLANS: outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7 Statistics: Num rows: 14398467 Data size: 1325194184 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator -key expressions: _col0 (type: int), _col5 (type: int) -sort order: ++ -Map-reduce partition columns: _col0 (type: int), _col5 (type: int) +key expressions: _col2 (type: int) +sort order: + +Map-reduce partition columns: _col2 (type: int) Statistics: Num rows: 14398467 Data size: 1325194184 Basic stats: COMPLETE Column stats: NONE -value expressions: _col1 (type: int), _col2 (type: int), _col3 (type: int), _col4 (type: int), _col6 (type: decimal(7,2)), _col7 (type: decimal(7,2)) +value expressions: _col0 (type: int), _col1 (type: int), _col3 (type: int), _col4 (type: int), _col5 (type: int), _col6 (type: decimal(7,2)), _col7 (type: decimal(7,2)) Map 10 Map Operator Tree: TableScan - alias: customer_address - Statistics: Num rows: 4000 Data size: 40595195284 Basic stats: COMPLETE Column stats: NONE + alias: cd2 + Statistics: Num rows: 1861800 Data size: 717186159 Basic stats: COMPLETE Column stats: NONE Filter Operator -predicate: ((ca_country = 'United States') and (ca_state) IN ('KY', 'GA', 'NM', 'MT', 'OR', 'IN', 'WI', 'MO', 'WV') and ca_address_sk is not null) (type: boolean) -Statistics: Num rows: 1000 Data size: 10148798821 Basic stats: COMPLETE Column stats: NONE +predicate: (((cd_education_status = '4 yr Degree') or (cd_education_status = 'Primary') or (cd_education_status = 'Advanced Degree')) and ((cd_marital_status = 'M') or (cd_marital_status = 'D') or (cd_marital_status = 'U')) and cd_demo_sk is not null and cd_education_status is not null and cd_marital_status is not null) (type: boolean) +Statistics: Num rows: 1861800 Data size: 717186159 Basic stats: COMPLETE Column stats: NONE Select Operator - expressions: ca_address_sk (type: int), ca_state (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num r
[09/13] hive git commit: HIVE-19128 : Update golden files for spark perf tests
http://git-wip-us.apache.org/repos/asf/hive/blob/9ce42cba/ql/src/test/results/clientpositive/perf/spark/query47.q.out -- diff --git a/ql/src/test/results/clientpositive/perf/spark/query47.q.out b/ql/src/test/results/clientpositive/perf/spark/query47.q.out index 5175f80..bc97fdf 100644 --- a/ql/src/test/results/clientpositive/perf/spark/query47.q.out +++ b/ql/src/test/results/clientpositive/perf/spark/query47.q.out @@ -397,47 +397,47 @@ STAGE PLANS: Statistics: Num rows: 766650239 Data size: 67634106676 Basic stats: COMPLETE Column stats: NONE Group By Operator aggregations: sum(_col3) -keys: _col5 (type: int), _col6 (type: int), _col8 (type: string), _col9 (type: string), _col11 (type: string), _col12 (type: string) +keys: _col11 (type: string), _col12 (type: string), _col8 (type: string), _col9 (type: string), _col5 (type: int), _col6 (type: int) mode: hash outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6 Statistics: Num rows: 766650239 Data size: 67634106676 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator - key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: string) + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: int), _col5 (type: int) sort order: ++ - Map-reduce partition columns: _col0 (type: int), _col1 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: string) + Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: int), _col5 (type: int) Statistics: Num rows: 766650239 Data size: 67634106676 Basic stats: COMPLETE Column stats: NONE value expressions: _col6 (type: decimal(17,2)) Reducer 14 Reduce Operator Tree: Group By Operator aggregations: sum(VALUE._col0) -keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: string), KEY._col3 (type: string), KEY._col4 (type: string), KEY._col5 (type: string) +keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: string), KEY._col3 (type: string), KEY._col4 (type: int), KEY._col5 (type: int) mode: mergepartial outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6 Statistics: Num rows: 383325119 Data size: 33817053293 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator - key expressions: _col3 (type: string), _col2 (type: string), _col4 (type: string), _col5 (type: string), _col0 (type: int) + key expressions: _col3 (type: string), _col2 (type: string), _col0 (type: string), _col1 (type: string), _col4 (type: int) sort order: + - Map-reduce partition columns: _col3 (type: string), _col2 (type: string), _col4 (type: string), _col5 (type: string), _col0 (type: int) + Map-reduce partition columns: _col3 (type: string), _col2 (type: string), _col0 (type: string), _col1 (type: string), _col4 (type: int) Statistics: Num rows: 383325119 Data size: 33817053293 Basic stats: COMPLETE Column stats: NONE - value expressions: _col1 (type: int), _col6 (type: decimal(17,2)) + value expressions: _col5 (type: int), _col6 (type: decimal(17,2)) Reducer 15 Reduce Operator Tree: Select Operator -expressions: KEY.reducesinkkey4 (type: int), VALUE._col0 (type: int), KEY.reducesinkkey1 (type: string), KEY.reducesinkkey0 (type: string), KEY.reducesinkkey2 (type: string), KEY.reducesinkkey3 (type: string), VALUE._col1 (type: decimal(17,2)) +expressions: KEY.reducesinkkey2 (type: string), KEY.reducesinkkey3 (type: string), KEY.reducesinkkey1 (type: string), KEY.reducesinkkey0 (type: string), KEY.reducesinkkey4 (type: int), VALUE._col0 (type: int), VALUE._col1 (type: decimal(17,2)) outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6 Statistics: Num rows: 383325119 Data size: 33817053293 Basic stats: COMPLETE Column stats: NONE PTF Operator Function definitions: Input definition input alias: ptf_0 -output shape: _col0: int, _col1: int, _col2: string, _col3: string, _col4: string
[04/13] hive git commit: HIVE-19128 : Update golden files for spark perf tests
http://git-wip-us.apache.org/repos/asf/hive/blob/9ce42cba/ql/src/test/results/clientpositive/perf/spark/query78.q.out -- diff --git a/ql/src/test/results/clientpositive/perf/spark/query78.q.out b/ql/src/test/results/clientpositive/perf/spark/query78.q.out index 1467c5f..c2e8577 100644 --- a/ql/src/test/results/clientpositive/perf/spark/query78.q.out +++ b/ql/src/test/results/clientpositive/perf/spark/query78.q.out @@ -120,57 +120,41 @@ STAGE PLANS: Stage: Stage-1 Spark Edges: -Reducer 11 <- Map 10 (PARTITION-LEVEL SORT, 85), Reducer 14 (PARTITION-LEVEL SORT, 85) -Reducer 12 <- Reducer 11 (GROUP, 93) -Reducer 14 <- Map 13 (PARTITION-LEVEL SORT, 164), Map 15 (PARTITION-LEVEL SORT, 164) -Reducer 17 <- Map 10 (PARTITION-LEVEL SORT, 85), Reducer 20 (PARTITION-LEVEL SORT, 85) +Reducer 11 <- Map 10 (PARTITION-LEVEL SORT, 164), Map 14 (PARTITION-LEVEL SORT, 164) +Reducer 12 <- Map 15 (PARTITION-LEVEL SORT, 85), Reducer 11 (PARTITION-LEVEL SORT, 85) +Reducer 13 <- Reducer 12 (GROUP, 93) +Reducer 17 <- Map 16 (PARTITION-LEVEL SORT, 85), Reducer 20 (PARTITION-LEVEL SORT, 85) Reducer 18 <- Reducer 17 (GROUP, 93) -Reducer 2 <- Map 1 (PARTITION-LEVEL SORT, 219), Reducer 8 (PARTITION-LEVEL SORT, 219) +Reducer 2 <- Map 1 (PARTITION-LEVEL SORT, 432), Map 8 (PARTITION-LEVEL SORT, 432) Reducer 20 <- Map 19 (PARTITION-LEVEL SORT, 177), Map 21 (PARTITION-LEVEL SORT, 177) -Reducer 3 <- Reducer 2 (GROUP, 241) -Reducer 4 <- Reducer 12 (PARTITION-LEVEL SORT, 167), Reducer 3 (PARTITION-LEVEL SORT, 167) -Reducer 5 <- Reducer 18 (PARTITION-LEVEL SORT, 91), Reducer 4 (PARTITION-LEVEL SORT, 91) -Reducer 6 <- Reducer 5 (SORT, 1) -Reducer 8 <- Map 7 (PARTITION-LEVEL SORT, 432), Map 9 (PARTITION-LEVEL SORT, 432) +Reducer 3 <- Map 9 (PARTITION-LEVEL SORT, 219), Reducer 2 (PARTITION-LEVEL SORT, 219) +Reducer 4 <- Reducer 3 (GROUP, 241) +Reducer 5 <- Reducer 13 (PARTITION-LEVEL SORT, 167), Reducer 4 (PARTITION-LEVEL SORT, 167) +Reducer 6 <- Reducer 18 (PARTITION-LEVEL SORT, 91), Reducer 5 (PARTITION-LEVEL SORT, 91) +Reducer 7 <- Reducer 6 (SORT, 1) A masked pattern was here Vertices: Map 1 Map Operator Tree: TableScan - alias: date_dim - Statistics: Num rows: 73049 Data size: 81741831 Basic stats: COMPLETE Column stats: NONE + alias: store_sales + Statistics: Num rows: 575995635 Data size: 50814502088 Basic stats: COMPLETE Column stats: NONE Filter Operator -predicate: ((d_year = 2000) and d_date_sk is not null) (type: boolean) -Statistics: Num rows: 36524 Data size: 40870356 Basic stats: COMPLETE Column stats: NONE +predicate: ss_sold_date_sk is not null (type: boolean) +Statistics: Num rows: 575995635 Data size: 50814502088 Basic stats: COMPLETE Column stats: NONE Select Operator - expressions: d_date_sk (type: int) - outputColumnNames: _col0 - Statistics: Num rows: 36524 Data size: 40870356 Basic stats: COMPLETE Column stats: NONE + expressions: ss_sold_date_sk (type: int), ss_item_sk (type: int), ss_customer_sk (type: int), ss_ticket_number (type: int), ss_quantity (type: int), ss_wholesale_cost (type: decimal(7,2)), ss_sales_price (type: decimal(7,2)) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6 + Statistics: Num rows: 575995635 Data size: 50814502088 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator -key expressions: _col0 (type: int) -sort order: + -Map-reduce partition columns: _col0 (type: int) -Statistics: Num rows: 36524 Data size: 40870356 Basic stats: COMPLETE Column stats: NONE +key expressions: _col1 (type: int), _col3 (type: int) +sort order: ++ +Map-reduce partition columns: _col1 (type: int), _col3 (type: int) +Statistics: Num rows: 575995635 Data size: 50814502088 Basic stats: COMPLETE Column stats: NONE +value expressions: _col0 (type: int), _col2 (type: int), _col4 (type: int), _col5 (type: decimal(7,2)), _col6 (type: decimal(7,2)) Map 10 Map Operator Tree: TableScan - alias: date_dim - Statistics: Num rows: 73049 Data size: 81741831 Basic stats: COMPLETE Column stats: NONE - Filter Operator -
[05/13] hive git commit: HIVE-19128 : Update golden files for spark perf tests
http://git-wip-us.apache.org/repos/asf/hive/blob/9ce42cba/ql/src/test/results/clientpositive/perf/spark/query75.q.out -- diff --git a/ql/src/test/results/clientpositive/perf/spark/query75.q.out b/ql/src/test/results/clientpositive/perf/spark/query75.q.out index 4b76570..f4fca34 100644 --- a/ql/src/test/results/clientpositive/perf/spark/query75.q.out +++ b/ql/src/test/results/clientpositive/perf/spark/query75.q.out @@ -144,25 +144,25 @@ STAGE PLANS: Stage: Stage-1 Spark Edges: -Reducer 13 <- Map 12 (PARTITION-LEVEL SORT, 398), Map 16 (PARTITION-LEVEL SORT, 398) -Reducer 14 <- Map 10 (PARTITION-LEVEL SORT, 440), Reducer 13 (PARTITION-LEVEL SORT, 440) +Reducer 13 <- Map 12 (PARTITION-LEVEL SORT, 400), Map 16 (PARTITION-LEVEL SORT, 400) +Reducer 14 <- Map 10 (PARTITION-LEVEL SORT, 438), Reducer 13 (PARTITION-LEVEL SORT, 438) Reducer 15 <- Map 18 (PARTITION-LEVEL SORT, 516), Reducer 14 (PARTITION-LEVEL SORT, 516) -Reducer 2 <- Map 1 (PARTITION-LEVEL SORT, 306), Map 16 (PARTITION-LEVEL SORT, 306) -Reducer 20 <- Map 16 (PARTITION-LEVEL SORT, 154), Map 19 (PARTITION-LEVEL SORT, 154) -Reducer 21 <- Map 10 (PARTITION-LEVEL SORT, 171), Reducer 20 (PARTITION-LEVEL SORT, 171) +Reducer 2 <- Map 1 (PARTITION-LEVEL SORT, 308), Map 16 (PARTITION-LEVEL SORT, 308) +Reducer 20 <- Map 16 (PARTITION-LEVEL SORT, 156), Map 19 (PARTITION-LEVEL SORT, 156) +Reducer 21 <- Map 10 (PARTITION-LEVEL SORT, 169), Reducer 20 (PARTITION-LEVEL SORT, 169) Reducer 22 <- Map 25 (PARTITION-LEVEL SORT, 196), Reducer 21 (PARTITION-LEVEL SORT, 196) -Reducer 27 <- Map 26 (PARTITION-LEVEL SORT, 306), Map 32 (PARTITION-LEVEL SORT, 306) -Reducer 28 <- Map 33 (PARTITION-LEVEL SORT, 338), Reducer 27 (PARTITION-LEVEL SORT, 338) +Reducer 27 <- Map 26 (PARTITION-LEVEL SORT, 308), Map 32 (PARTITION-LEVEL SORT, 308) +Reducer 28 <- Map 33 (PARTITION-LEVEL SORT, 336), Reducer 27 (PARTITION-LEVEL SORT, 336) Reducer 29 <- Map 34 (PARTITION-LEVEL SORT, 393), Reducer 28 (PARTITION-LEVEL SORT, 393) -Reducer 3 <- Map 10 (PARTITION-LEVEL SORT, 338), Reducer 2 (PARTITION-LEVEL SORT, 338) +Reducer 3 <- Map 10 (PARTITION-LEVEL SORT, 336), Reducer 2 (PARTITION-LEVEL SORT, 336) Reducer 30 <- Reducer 29 (GROUP, 934), Reducer 38 (GROUP, 934) Reducer 31 <- Reducer 30 (GROUP PARTITION-LEVEL SORT, 671), Reducer 45 (GROUP PARTITION-LEVEL SORT, 671) -Reducer 36 <- Map 32 (PARTITION-LEVEL SORT, 398), Map 35 (PARTITION-LEVEL SORT, 398) -Reducer 37 <- Map 33 (PARTITION-LEVEL SORT, 440), Reducer 36 (PARTITION-LEVEL SORT, 440) +Reducer 36 <- Map 32 (PARTITION-LEVEL SORT, 400), Map 35 (PARTITION-LEVEL SORT, 400) +Reducer 37 <- Map 33 (PARTITION-LEVEL SORT, 438), Reducer 36 (PARTITION-LEVEL SORT, 438) Reducer 38 <- Map 41 (PARTITION-LEVEL SORT, 516), Reducer 37 (PARTITION-LEVEL SORT, 516) Reducer 4 <- Map 11 (PARTITION-LEVEL SORT, 393), Reducer 3 (PARTITION-LEVEL SORT, 393) -Reducer 43 <- Map 32 (PARTITION-LEVEL SORT, 154), Map 42 (PARTITION-LEVEL SORT, 154) -Reducer 44 <- Map 33 (PARTITION-LEVEL SORT, 171), Reducer 43 (PARTITION-LEVEL SORT, 171) +Reducer 43 <- Map 32 (PARTITION-LEVEL SORT, 156), Map 42 (PARTITION-LEVEL SORT, 156) +Reducer 44 <- Map 33 (PARTITION-LEVEL SORT, 169), Reducer 43 (PARTITION-LEVEL SORT, 169) Reducer 45 <- Map 48 (PARTITION-LEVEL SORT, 196), Reducer 44 (PARTITION-LEVEL SORT, 196) Reducer 5 <- Reducer 15 (GROUP, 934), Reducer 4 (GROUP, 934) Reducer 6 <- Reducer 22 (GROUP PARTITION-LEVEL SORT, 671), Reducer 5 (GROUP PARTITION-LEVEL SORT, 671) @@ -183,29 +183,28 @@ STAGE PLANS: outputColumnNames: _col0, _col1, _col2, _col3, _col4 Statistics: Num rows: 287989836 Data size: 38999608952 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator -key expressions: _col0 (type: int) +key expressions: _col1 (type: int) sort order: + -Map-reduce partition columns: _col0 (type: int) +Map-reduce partition columns: _col1 (type: int) Statistics: Num rows: 287989836 Data size: 38999608952 Basic stats: COMPLETE Column stats: NONE -value expressions: _col1 (type: int), _col2 (type: int), _col3 (type: int), _col4 (type: decimal(7,2)) +value expressions: _col0 (type: int), _col2 (type: int), _col3 (type: int), _col4 (type: decimal(7,2)) Map 10 Map Operator Tree: TableScan - alias: item - Statistics: Num rows: 462000 Data size: 663560457 Basic stats:
[10/13] hive git commit: HIVE-19128 : Update golden files for spark perf tests
http://git-wip-us.apache.org/repos/asf/hive/blob/9ce42cba/ql/src/test/results/clientpositive/perf/spark/query40.q.out -- diff --git a/ql/src/test/results/clientpositive/perf/spark/query40.q.out b/ql/src/test/results/clientpositive/perf/spark/query40.q.out index f286294..5360385 100644 --- a/ql/src/test/results/clientpositive/perf/spark/query40.q.out +++ b/ql/src/test/results/clientpositive/perf/spark/query40.q.out @@ -54,8 +54,7 @@ limit 100 POSTHOOK: type: QUERY STAGE DEPENDENCIES: Stage-2 is a root stage - Stage-3 depends on stages: Stage-2 - Stage-1 depends on stages: Stage-3 + Stage-1 depends on stages: Stage-2 Stage-0 depends on stages: Stage-1 STAGE PLANS: @@ -63,44 +62,39 @@ STAGE PLANS: Spark A masked pattern was here Vertices: -Map 9 +Map 8 Map Operator Tree: TableScan - alias: warehouse - Statistics: Num rows: 27 Data size: 27802 Basic stats: COMPLETE Column stats: NONE + alias: date_dim + Statistics: Num rows: 73049 Data size: 81741831 Basic stats: COMPLETE Column stats: NONE Filter Operator -predicate: w_warehouse_sk is not null (type: boolean) -Statistics: Num rows: 27 Data size: 27802 Basic stats: COMPLETE Column stats: NONE +predicate: (CAST( d_date AS TIMESTAMP) BETWEEN TIMESTAMP'1998-03-08 23:00:00.0' AND TIMESTAMP'1998-05-08 00:00:00.0' and d_date_sk is not null) (type: boolean) +Statistics: Num rows: 8116 Data size: 9081804 Basic stats: COMPLETE Column stats: NONE Select Operator - expressions: w_warehouse_sk (type: int), w_state (type: string) + expressions: d_date_sk (type: int), d_date (type: string) outputColumnNames: _col0, _col1 - Statistics: Num rows: 27 Data size: 27802 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 8116 Data size: 9081804 Basic stats: COMPLETE Column stats: NONE Spark HashTable Sink Operator keys: - 0 _col1 (type: int) + 0 _col0 (type: int) 1 _col0 (type: int) Local Work: Map Reduce Local Work - - Stage: Stage-3 -Spark - A masked pattern was here - Vertices: -Map 7 +Map 9 Map Operator Tree: TableScan - alias: date_dim - Statistics: Num rows: 73049 Data size: 81741831 Basic stats: COMPLETE Column stats: NONE + alias: warehouse + Statistics: Num rows: 27 Data size: 27802 Basic stats: COMPLETE Column stats: NONE Filter Operator -predicate: (CAST( d_date AS TIMESTAMP) BETWEEN TIMESTAMP'1998-03-08 23:00:00.0' AND TIMESTAMP'1998-05-08 00:00:00.0' and d_date_sk is not null) (type: boolean) -Statistics: Num rows: 8116 Data size: 9081804 Basic stats: COMPLETE Column stats: NONE +predicate: w_warehouse_sk is not null (type: boolean) +Statistics: Num rows: 27 Data size: 27802 Basic stats: COMPLETE Column stats: NONE Select Operator - expressions: d_date_sk (type: int), d_date (type: string) + expressions: w_warehouse_sk (type: int), w_state (type: string) outputColumnNames: _col0, _col1 - Statistics: Num rows: 8116 Data size: 9081804 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 27 Data size: 27802 Basic stats: COMPLETE Column stats: NONE Spark HashTable Sink Operator keys: - 0 _col0 (type: int) + 0 _col1 (type: int) 1 _col0 (type: int) Local Work: Map Reduce Local Work @@ -109,7 +103,7 @@ STAGE PLANS: Spark Edges: Reducer 2 <- Map 1 (PARTITION-LEVEL SORT, 329), Map 6 (PARTITION-LEVEL SORT, 329) -Reducer 3 <- Map 8 (PARTITION-LEVEL SORT, 370), Reducer 2 (PARTITION-LEVEL SORT, 370) +Reducer 3 <- Map 7 (PARTITION-LEVEL SORT, 336), Reducer 2 (PARTITION-LEVEL SORT, 336) Reducer 4 <- Reducer 3 (GROUP, 447) Reducer 5 <- Reducer 4 (SORT, 1) A masked pattern was here @@ -150,7 +144,7 @@ STAGE PLANS: Map-reduce partition columns: _col0 (type: int), _col1 (type: int) Statistics: Num rows: 28798881 Data size: 3057234680 Basic stats: COMPLETE Column stats: NONE value expressions: _col2 (type:
[12/13] hive git commit: HIVE-19128 : Update golden files for spark perf tests
http://git-wip-us.apache.org/repos/asf/hive/blob/9ce42cba/ql/src/test/results/clientpositive/perf/spark/query24.q.out -- diff --git a/ql/src/test/results/clientpositive/perf/spark/query24.q.out b/ql/src/test/results/clientpositive/perf/spark/query24.q.out index 1f291c0..13ac1e8 100644 --- a/ql/src/test/results/clientpositive/perf/spark/query24.q.out +++ b/ql/src/test/results/clientpositive/perf/spark/query24.q.out @@ -131,10 +131,10 @@ STAGE PLANS: Stage: Stage-2 Spark Edges: -Reducer 13 <- Map 12 (PARTITION-LEVEL SORT, 975), Map 20 (PARTITION-LEVEL SORT, 975) -Reducer 14 <- Map 21 (PARTITION-LEVEL SORT, 486), Reducer 13 (PARTITION-LEVEL SORT, 486) -Reducer 15 <- Map 22 (PARTITION-LEVEL SORT, 564), Reducer 14 (PARTITION-LEVEL SORT, 564) -Reducer 16 <- Map 23 (PARTITION-LEVEL SORT, 899), Reducer 15 (PARTITION-LEVEL SORT, 899) +Reducer 13 <- Map 12 (PARTITION-LEVEL SORT, 472), Map 20 (PARTITION-LEVEL SORT, 472) +Reducer 14 <- Map 21 (PARTITION-LEVEL SORT, 1009), Reducer 13 (PARTITION-LEVEL SORT, 1009) +Reducer 15 <- Map 22 (PARTITION-LEVEL SORT, 846), Reducer 14 (PARTITION-LEVEL SORT, 846) +Reducer 16 <- Map 23 (PARTITION-LEVEL SORT, 587), Reducer 15 (PARTITION-LEVEL SORT, 587) Reducer 17 <- Reducer 16 (GROUP, 640) Reducer 18 <- Reducer 17 (GROUP, 1) A masked pattern was here @@ -162,16 +162,33 @@ STAGE PLANS: 1 Map 19 Statistics: Num rows: 633595212 Data size: 55895953508 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator - key expressions: _col1 (type: int) - sort order: + - Map-reduce partition columns: _col1 (type: int) + key expressions: _col0 (type: int), _col3 (type: int) + sort order: ++ + Map-reduce partition columns: _col0 (type: int), _col3 (type: int) Statistics: Num rows: 633595212 Data size: 55895953508 Basic stats: COMPLETE Column stats: NONE - value expressions: _col0 (type: int), _col3 (type: int), _col4 (type: decimal(7,2)), _col6 (type: string), _col8 (type: string), _col9 (type: string) + value expressions: _col1 (type: int), _col4 (type: decimal(7,2)), _col6 (type: string), _col8 (type: string), _col9 (type: string) Local Work: Map Reduce Local Work Map 20 Map Operator Tree: TableScan + alias: store_returns + Statistics: Num rows: 57591150 Data size: 4462194832 Basic stats: COMPLETE Column stats: NONE + Filter Operator +predicate: (sr_item_sk is not null and sr_ticket_number is not null) (type: boolean) +Statistics: Num rows: 57591150 Data size: 4462194832 Basic stats: COMPLETE Column stats: NONE +Select Operator + expressions: sr_item_sk (type: int), sr_ticket_number (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 57591150 Data size: 4462194832 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator +key expressions: _col0 (type: int), _col1 (type: int) +sort order: ++ +Map-reduce partition columns: _col0 (type: int), _col1 (type: int) +Statistics: Num rows: 57591150 Data size: 4462194832 Basic stats: COMPLETE Column stats: NONE +Map 21 +Map Operator Tree: +TableScan alias: customer Statistics: Num rows: 8000 Data size: 68801615852 Basic stats: COMPLETE Column stats: NONE Filter Operator @@ -187,44 +204,9 @@ STAGE PLANS: Map-reduce partition columns: _col0 (type: int) Statistics: Num rows: 8000 Data size: 68801615852 Basic stats: COMPLETE Column stats: NONE value expressions: _col1 (type: string), _col2 (type: string), _col3 (type: string) -Map 21 -Map Operator Tree: -TableScan - alias: item - Statistics: Num rows: 462000 Data size: 663560457 Basic stats: COMPLETE Column stats: NONE - Filter Operator -predicate: i_item_sk is not null (type: boolean) -Statistics: Num rows: 462000 Data size: 663560457 Basic stats: COMPLETE Column stats: NONE -Select Operator - expressions: i_item_sk (type: int), i_current_price (type: decimal(7,2)), i_size (type: string),
[07/13] hive git commit: HIVE-19128 : Update golden files for spark perf tests
http://git-wip-us.apache.org/repos/asf/hive/blob/9ce42cba/ql/src/test/results/clientpositive/perf/spark/query58.q.out -- diff --git a/ql/src/test/results/clientpositive/perf/spark/query58.q.out b/ql/src/test/results/clientpositive/perf/spark/query58.q.out index f06cbef..eb5dffe 100644 --- a/ql/src/test/results/clientpositive/perf/spark/query58.q.out +++ b/ql/src/test/results/clientpositive/perf/spark/query58.q.out @@ -1,6 +1,6 @@ -Warning: Map Join MAPJOIN[183][bigTable=?] in task 'Stage-1:MAPRED' is a cross product -Warning: Map Join MAPJOIN[184][bigTable=?] in task 'Stage-1:MAPRED' is a cross product -Warning: Map Join MAPJOIN[185][bigTable=?] in task 'Stage-1:MAPRED' is a cross product +Warning: Map Join MAPJOIN[180][bigTable=?] in task 'Stage-1:MAPRED' is a cross product +Warning: Map Join MAPJOIN[181][bigTable=?] in task 'Stage-1:MAPRED' is a cross product +Warning: Map Join MAPJOIN[182][bigTable=?] in task 'Stage-1:MAPRED' is a cross product PREHOOK: query: explain with ss_items as (select i_item_id item_id @@ -140,10 +140,10 @@ STAGE PLANS: Stage: Stage-2 Spark Edges: -Reducer 11 <- Map 10 (GROUP, 1) +Reducer 12 <- Map 11 (GROUP, 1) A masked pattern was here Vertices: -Map 10 +Map 11 Map Operator Tree: TableScan alias: date_dim @@ -162,7 +162,7 @@ STAGE PLANS: sort order: Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE value expressions: _col0 (type: bigint) -Reducer 11 +Reducer 12 Local Work: Map Reduce Local Work Reduce Operator Tree: @@ -184,10 +184,10 @@ STAGE PLANS: Stage: Stage-3 Spark Edges: -Reducer 24 <- Map 23 (GROUP, 1) +Reducer 25 <- Map 24 (GROUP, 1) A masked pattern was here Vertices: -Map 23 +Map 24 Map Operator Tree: TableScan alias: date_dim @@ -206,7 +206,7 @@ STAGE PLANS: sort order: Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE value expressions: _col0 (type: bigint) -Reducer 24 +Reducer 25 Local Work: Map Reduce Local Work Reduce Operator Tree: @@ -228,10 +228,10 @@ STAGE PLANS: Stage: Stage-4 Spark Edges: -Reducer 37 <- Map 36 (GROUP, 1) +Reducer 38 <- Map 37 (GROUP, 1) A masked pattern was here Vertices: -Map 36 +Map 37 Map Operator Tree: TableScan alias: date_dim @@ -250,7 +250,7 @@ STAGE PLANS: sort order: Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE value expressions: _col0 (type: bigint) -Reducer 37 +Reducer 38 Local Work: Map Reduce Local Work Reduce Operator Tree: @@ -272,76 +272,47 @@ STAGE PLANS: Stage: Stage-1 Spark Edges: -Reducer 13 <- Map 12 (PARTITION-LEVEL SORT, 2), Map 15 (PARTITION-LEVEL SORT, 2) -Reducer 14 <- Reducer 13 (GROUP, 2) -Reducer 17 <- Map 16 (PARTITION-LEVEL SORT, 403), Map 20 (PARTITION-LEVEL SORT, 403) -Reducer 18 <- Reducer 17 (PARTITION-LEVEL SORT, 438), Reducer 22 (PARTITION-LEVEL SORT, 438) -Reducer 19 <- Reducer 18 (GROUP, 481) -Reducer 2 <- Map 1 (PARTITION-LEVEL SORT, 310), Map 7 (PARTITION-LEVEL SORT, 310) -Reducer 22 <- Map 21 (PARTITION-LEVEL SORT, 2), Reducer 27 (PARTITION-LEVEL SORT, 2) -Reducer 26 <- Map 25 (PARTITION-LEVEL SORT, 2), Map 28 (PARTITION-LEVEL SORT, 2) -Reducer 27 <- Reducer 26 (GROUP, 2) -Reducer 3 <- Reducer 2 (PARTITION-LEVEL SORT, 336), Reducer 9 (PARTITION-LEVEL SORT, 336) -Reducer 30 <- Map 29 (PARTITION-LEVEL SORT, 159), Map 33 (PARTITION-LEVEL SORT, 159) -Reducer 31 <- Reducer 30 (PARTITION-LEVEL SORT, 169), Reducer 35 (PARTITION-LEVEL SORT, 169) -Reducer 32 <- Reducer 31 (GROUP, 186) -Reducer 35 <- Map 34 (PARTITION-LEVEL SORT, 2), Reducer 40 (PARTITION-LEVEL SORT, 2) -Reducer 39 <- Map 38 (PARTITION-LEVEL SORT, 2), Map 41 (PARTITION-LEVEL SORT, 2) -Reducer 4 <- Reducer 3 (GROUP, 369) -Reducer 40 <- Reducer 39 (GROUP, 2) -Reducer 5 <- Reducer 19 (PARTITION-LEVEL SORT, 518), Reducer 32 (PARTITION-LEVEL SORT, 518), Reducer 4 (PARTITION-LEVEL SORT, 518) -Reducer 6 <- Reducer 5 (SORT, 1) -Reducer 9 <- Map 8 (PARTITION-LEVEL SORT, 2), Reducer 14 (PARTITION-LEVEL SORT, 2) +Reducer 10 <- Reducer
[13/13] hive git commit: HIVE-19128 : Update golden files for spark perf tests
HIVE-19128 : Update golden files for spark perf tests Signed-off-by: Ashutosh Chauhan Project: http://git-wip-us.apache.org/repos/asf/hive/repo Commit: http://git-wip-us.apache.org/repos/asf/hive/commit/9ce42cba Tree: http://git-wip-us.apache.org/repos/asf/hive/tree/9ce42cba Diff: http://git-wip-us.apache.org/repos/asf/hive/diff/9ce42cba Branch: refs/heads/branch-3 Commit: 9ce42cba8dfa21855463977f0d2006c77dd04c73 Parents: 43cb101 Author: Ashutosh Chauhan Authored: Mon Apr 9 14:57:11 2018 -0700 Committer: Ashutosh Chauhan Committed: Mon Apr 9 15:07:02 2018 -0700 -- .../clientpositive/perf/spark/query11.q.out | 48 +- .../clientpositive/perf/spark/query15.q.out | 164 +++ .../clientpositive/perf/spark/query16.q.out | 70 ++- .../clientpositive/perf/spark/query18.q.out | 216 .../clientpositive/perf/spark/query19.q.out | 218 - .../clientpositive/perf/spark/query21.q.out | 114 ++--- .../clientpositive/perf/spark/query24.q.out | 282 +-- .../clientpositive/perf/spark/query25.q.out | 118 ++--- .../clientpositive/perf/spark/query29.q.out | 280 ++- .../clientpositive/perf/spark/query30.q.out | 262 +- .../clientpositive/perf/spark/query32.q.out | 132 ++--- .../clientpositive/perf/spark/query34.q.out | 34 +- .../clientpositive/perf/spark/query35.q.out | 74 +-- .../clientpositive/perf/spark/query37.q.out | 16 +- .../clientpositive/perf/spark/query4.q.out | 214 .../clientpositive/perf/spark/query40.q.out | 116 ++--- .../clientpositive/perf/spark/query44.q.out | 246 +- .../clientpositive/perf/spark/query45.q.out | 204 .../clientpositive/perf/spark/query46.q.out | 104 ++-- .../clientpositive/perf/spark/query47.q.out | 92 ++-- .../clientpositive/perf/spark/query48.q.out | 94 ++-- .../clientpositive/perf/spark/query5.q.out | 38 +- .../clientpositive/perf/spark/query50.q.out | 196 .../clientpositive/perf/spark/query53.q.out | 105 ++-- .../clientpositive/perf/spark/query54.q.out | 277 +-- .../clientpositive/perf/spark/query57.q.out | 92 ++-- .../clientpositive/perf/spark/query58.q.out | 490 +-- .../clientpositive/perf/spark/query6.q.out | 350 ++--- .../clientpositive/perf/spark/query61.q.out | 40 +- .../clientpositive/perf/spark/query63.q.out | 105 ++-- .../clientpositive/perf/spark/query65.q.out | 100 ++-- .../clientpositive/perf/spark/query66.q.out | 20 +- .../clientpositive/perf/spark/query67.q.out | 137 +++--- .../clientpositive/perf/spark/query68.q.out | 104 ++-- .../clientpositive/perf/spark/query72.q.out | 461 - .../clientpositive/perf/spark/query73.q.out | 34 +- .../clientpositive/perf/spark/query75.q.out | 248 +- .../clientpositive/perf/spark/query76.q.out | 130 ++--- .../clientpositive/perf/spark/query77.q.out | 66 +-- .../clientpositive/perf/spark/query78.q.out | 273 ++- .../clientpositive/perf/spark/query79.q.out | 10 +- .../clientpositive/perf/spark/query8.q.out | 174 +++ .../clientpositive/perf/spark/query80.q.out | 417 .../clientpositive/perf/spark/query81.q.out | 270 +- .../clientpositive/perf/spark/query82.q.out | 16 +- .../clientpositive/perf/spark/query83.q.out | 354 +++--- .../clientpositive/perf/spark/query85.q.out | 309 ++-- .../clientpositive/perf/spark/query88.q.out | 272 +- .../clientpositive/perf/spark/query89.q.out | 26 +- .../clientpositive/perf/spark/query90.q.out | 68 +-- .../clientpositive/perf/spark/query91.q.out | 14 +- .../clientpositive/perf/spark/query92.q.out | 106 ++-- .../clientpositive/perf/spark/query94.q.out | 70 ++- .../clientpositive/perf/spark/query95.q.out | 100 ++-- .../clientpositive/perf/spark/query97.q.out | 54 +- .../clientpositive/perf/spark/query99.q.out | 42 +- 56 files changed, 4313 insertions(+), 4353 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hive/blob/9ce42cba/ql/src/test/results/clientpositive/perf/spark/query11.q.out -- diff --git a/ql/src/test/results/clientpositive/perf/spark/query11.q.out b/ql/src/test/results/clientpositive/perf/spark/query11.q.out index e46aa21..227068c 100644 --- a/ql/src/test/results/clientpositive/perf/spark/query11.q.out +++ b/ql/src/test/results/clientpositive/perf/spark/query11.q.out @@ -300,7 +300,7 @@ STAGE PLANS: alias: date_dim Statistics: Num rows: 73049 Data size: 81741831 Basic stats: COMPLETE Column stats: NONE Filter Operator -predicate: ((d
[03/13] hive git commit: HIVE-19128 : Update golden files for spark perf tests
http://git-wip-us.apache.org/repos/asf/hive/blob/9ce42cba/ql/src/test/results/clientpositive/perf/spark/query81.q.out -- diff --git a/ql/src/test/results/clientpositive/perf/spark/query81.q.out b/ql/src/test/results/clientpositive/perf/spark/query81.q.out index 5b2d5b3..23998d6 100644 --- a/ql/src/test/results/clientpositive/perf/spark/query81.q.out +++ b/ql/src/test/results/clientpositive/perf/spark/query81.q.out @@ -66,72 +66,37 @@ STAGE PLANS: Stage: Stage-1 Spark Edges: -Reducer 10 <- Reducer 16 (PARTITION-LEVEL SORT, 262), Reducer 9 (PARTITION-LEVEL SORT, 262) -Reducer 14 <- Map 13 (PARTITION-LEVEL SORT, 25), Map 17 (PARTITION-LEVEL SORT, 25) -Reducer 15 <- Map 18 (PARTITION-LEVEL SORT, 344), Reducer 14 (PARTITION-LEVEL SORT, 344) -Reducer 16 <- Reducer 15 (GROUP PARTITION-LEVEL SORT, 349) -Reducer 2 <- Map 1 (PARTITION-LEVEL SORT, 697), Map 5 (PARTITION-LEVEL SORT, 697) -Reducer 3 <- Reducer 10 (PARTITION-LEVEL SORT, 656), Reducer 2 (PARTITION-LEVEL SORT, 656) -Reducer 4 <- Reducer 3 (SORT, 1) -Reducer 7 <- Map 11 (PARTITION-LEVEL SORT, 25), Map 6 (PARTITION-LEVEL SORT, 25) -Reducer 8 <- Map 12 (PARTITION-LEVEL SORT, 344), Reducer 7 (PARTITION-LEVEL SORT, 344) -Reducer 9 <- Reducer 8 (GROUP, 349) +Reducer 11 <- Map 10 (PARTITION-LEVEL SORT, 25), Map 14 (PARTITION-LEVEL SORT, 25) +Reducer 12 <- Map 15 (PARTITION-LEVEL SORT, 344), Reducer 11 (PARTITION-LEVEL SORT, 344) +Reducer 13 <- Reducer 12 (GROUP PARTITION-LEVEL SORT, 349) +Reducer 17 <- Map 16 (PARTITION-LEVEL SORT, 697), Map 18 (PARTITION-LEVEL SORT, 697) +Reducer 2 <- Map 1 (PARTITION-LEVEL SORT, 25), Map 8 (PARTITION-LEVEL SORT, 25) +Reducer 3 <- Map 9 (PARTITION-LEVEL SORT, 344), Reducer 2 (PARTITION-LEVEL SORT, 344) +Reducer 4 <- Reducer 3 (GROUP, 349) +Reducer 5 <- Reducer 13 (PARTITION-LEVEL SORT, 262), Reducer 4 (PARTITION-LEVEL SORT, 262) +Reducer 6 <- Reducer 17 (PARTITION-LEVEL SORT, 656), Reducer 5 (PARTITION-LEVEL SORT, 656) +Reducer 7 <- Reducer 6 (SORT, 1) A masked pattern was here Vertices: Map 1 Map Operator Tree: TableScan - alias: customer - Statistics: Num rows: 8000 Data size: 68801615852 Basic stats: COMPLETE Column stats: NONE - Filter Operator -predicate: (c_current_addr_sk is not null and c_customer_sk is not null) (type: boolean) -Statistics: Num rows: 8000 Data size: 68801615852 Basic stats: COMPLETE Column stats: NONE -Select Operator - expressions: c_customer_sk (type: int), c_customer_id (type: string), c_current_addr_sk (type: int), c_salutation (type: string), c_first_name (type: string), c_last_name (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 - Statistics: Num rows: 8000 Data size: 68801615852 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator -key expressions: _col2 (type: int) -sort order: + -Map-reduce partition columns: _col2 (type: int) -Statistics: Num rows: 8000 Data size: 68801615852 Basic stats: COMPLETE Column stats: NONE -value expressions: _col0 (type: int), _col1 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: string) -Map 11 -Map Operator Tree: -TableScan - alias: date_dim - Statistics: Num rows: 73049 Data size: 81741831 Basic stats: COMPLETE Column stats: NONE - Filter Operator -predicate: ((d_year = 1998) and d_date_sk is not null) (type: boolean) -Statistics: Num rows: 36524 Data size: 40870356 Basic stats: COMPLETE Column stats: NONE -Select Operator - expressions: d_date_sk (type: int) - outputColumnNames: _col0 - Statistics: Num rows: 36524 Data size: 40870356 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator -key expressions: _col0 (type: int) -sort order: + -Map-reduce partition columns: _col0 (type: int) -Statistics: Num rows: 36524 Data size: 40870356 Basic stats: COMPLETE Column stats: NONE -Map 12 -Map Operator Tree: -TableScan - alias: customer_address - Statistics: Num rows: 4000 Data size: 40595195284 Basic stats: COMPLETE Column stats: NONE +
[08/13] hive git commit: HIVE-19128 : Update golden files for spark perf tests
http://git-wip-us.apache.org/repos/asf/hive/blob/9ce42cba/ql/src/test/results/clientpositive/perf/spark/query54.q.out -- diff --git a/ql/src/test/results/clientpositive/perf/spark/query54.q.out b/ql/src/test/results/clientpositive/perf/spark/query54.q.out index 43132bc..251d7ad 100644 --- a/ql/src/test/results/clientpositive/perf/spark/query54.q.out +++ b/ql/src/test/results/clientpositive/perf/spark/query54.q.out @@ -1,7 +1,7 @@ -Warning: Shuffle Join JOIN[111][tables = [$hdt$_1, $hdt$_2]] in Work 'Reducer 3' is a cross product -Warning: Shuffle Join JOIN[107][tables = [$hdt$_0, $hdt$_1, $hdt$_2, $hdt$_3, $hdt$_4]] in Work 'Reducer 14' is a cross product -Warning: Shuffle Join JOIN[114][tables = [$hdt$_1, $hdt$_2, $hdt$_0]] in Work 'Reducer 4' is a cross product -Warning: Map Join MAPJOIN[144][bigTable=?] in task 'Stage-1:MAPRED' is a cross product +Warning: Shuffle Join JOIN[111][tables = [$hdt$_1, $hdt$_2]] in Work 'Reducer 4' is a cross product +Warning: Shuffle Join JOIN[104][tables = [$hdt$_0, $hdt$_1, $hdt$_2, $hdt$_3]] in Work 'Reducer 14' is a cross product +Warning: Shuffle Join JOIN[114][tables = [$hdt$_1, $hdt$_2, $hdt$_0]] in Work 'Reducer 5' is a cross product +Warning: Map Join MAPJOIN[143][bigTable=?] in task 'Stage-1:MAPRED' is a cross product PREHOOK: query: explain with my_customers as ( select distinct c_customer_sk @@ -122,11 +122,11 @@ STAGE PLANS: Stage: Stage-2 Spark Edges: -Reducer 29 <- Map 28 (GROUP, 2) -Reducer 30 <- Reducer 29 (GROUP, 1) +Reducer 31 <- Map 30 (GROUP, 2) +Reducer 32 <- Reducer 31 (GROUP, 1) A masked pattern was here Vertices: -Map 28 +Map 30 Map Operator Tree: TableScan alias: date_dim @@ -148,7 +148,7 @@ STAGE PLANS: sort order: + Map-reduce partition columns: _col0 (type: int) Statistics: Num rows: 18262 Data size: 20435178 Basic stats: COMPLETE Column stats: NONE -Reducer 29 +Reducer 31 Reduce Operator Tree: Group By Operator keys: KEY._col0 (type: int) @@ -166,7 +166,7 @@ STAGE PLANS: sort order: Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE value expressions: _col0 (type: bigint) -Reducer 30 +Reducer 32 Local Work: Map Reduce Local Work Reduce Operator Tree: @@ -211,23 +211,23 @@ STAGE PLANS: Stage: Stage-1 Spark Edges: -Reducer 10 <- Reducer 9 (GROUP, 1) +Reducer 10 <- Map 1 (GROUP, 2) Reducer 12 <- Map 11 (PARTITION-LEVEL SORT, 398), Map 15 (PARTITION-LEVEL SORT, 398) Reducer 13 <- Reducer 12 (PARTITION-LEVEL SORT, 772), Reducer 17 (PARTITION-LEVEL SORT, 772) -Reducer 14 <- Reducer 13 (PARTITION-LEVEL SORT, 1), Reducer 32 (PARTITION-LEVEL SORT, 1) +Reducer 14 <- Reducer 13 (PARTITION-LEVEL SORT, 1), Reducer 29 (PARTITION-LEVEL SORT, 1) Reducer 17 <- Map 16 (PARTITION-LEVEL SORT, 654), Reducer 23 (PARTITION-LEVEL SORT, 654) Reducer 2 <- Map 1 (GROUP, 2) -Reducer 20 <- Map 19 (PARTITION-LEVEL SORT, 458), Map 24 (PARTITION-LEVEL SORT, 458), Map 25 (PARTITION-LEVEL SORT, 458) -Reducer 21 <- Map 26 (PARTITION-LEVEL SORT, 505), Reducer 20 (PARTITION-LEVEL SORT, 505) +Reducer 20 <- Map 19 (PARTITION-LEVEL SORT, 459), Map 24 (PARTITION-LEVEL SORT, 459), Map 25 (PARTITION-LEVEL SORT, 459) +Reducer 21 <- Map 26 (PARTITION-LEVEL SORT, 504), Reducer 20 (PARTITION-LEVEL SORT, 504) Reducer 22 <- Map 27 (PARTITION-LEVEL SORT, 1009), Reducer 21 (PARTITION-LEVEL SORT, 1009) Reducer 23 <- Reducer 22 (GROUP, 610) -Reducer 3 <- Reducer 10 (PARTITION-LEVEL SORT, 1), Reducer 2 (PARTITION-LEVEL SORT, 1) -Reducer 32 <- Map 31 (GROUP, 2) -Reducer 4 <- Reducer 14 (PARTITION-LEVEL SORT, 1), Reducer 3 (PARTITION-LEVEL SORT, 1) -Reducer 5 <- Reducer 4 (GROUP, 1009) +Reducer 29 <- Map 28 (GROUP, 2) +Reducer 3 <- Reducer 2 (GROUP, 1) +Reducer 4 <- Reducer 10 (PARTITION-LEVEL SORT, 1), Reducer 3 (PARTITION-LEVEL SORT, 1) +Reducer 5 <- Reducer 14 (PARTITION-LEVEL SORT, 1), Reducer 4 (PARTITION-LEVEL SORT, 1) Reducer 6 <- Reducer 5 (GROUP, 1009) -Reducer 7 <- Reducer 6 (SORT, 1) -Reducer 9 <- Map 1 (GROUP, 2) +Reducer 7 <- Reducer 6 (GROUP, 1009) +Reducer 8 <- Reducer 7 (SORT, 1) A masked pattern was here Vertices: Map 1 @@ -330,11 +330,11 @@ STAGE PLANS: outputColumnNames: _col0, _col1, _col2 Statistics: Num rows: 287989836 Data size: 38999608952 Basic stat
[01/13] hive git commit: HIVE-19128 : Update golden files for spark perf tests
Repository: hive Updated Branches: refs/heads/branch-3 43cb101ea -> 9ce42cba8 http://git-wip-us.apache.org/repos/asf/hive/blob/9ce42cba/ql/src/test/results/clientpositive/perf/spark/query92.q.out -- diff --git a/ql/src/test/results/clientpositive/perf/spark/query92.q.out b/ql/src/test/results/clientpositive/perf/spark/query92.q.out index e7b8632..70c4c5a 100644 --- a/ql/src/test/results/clientpositive/perf/spark/query92.q.out +++ b/ql/src/test/results/clientpositive/perf/spark/query92.q.out @@ -67,7 +67,7 @@ STAGE PLANS: Spark A masked pattern was here Vertices: -Map 5 +Map 7 Map Operator Tree: TableScan alias: date_dim @@ -90,7 +90,7 @@ STAGE PLANS: Spark A masked pattern was here Vertices: -Map 9 +Map 10 Map Operator Tree: TableScan alias: date_dim @@ -112,11 +112,11 @@ STAGE PLANS: Stage: Stage-1 Spark Edges: -Reducer 2 <- Map 1 (PARTITION-LEVEL SORT, 261), Reducer 8 (PARTITION-LEVEL SORT, 261) -Reducer 3 <- Reducer 2 (GROUP, 1) -Reducer 4 <- Reducer 3 (SORT, 1) -Reducer 7 <- Map 6 (GROUP, 169) -Reducer 8 <- Map 10 (PARTITION-LEVEL SORT, 87), Reducer 7 (PARTITION-LEVEL SORT, 87) +Reducer 2 <- Map 1 (PARTITION-LEVEL SORT, 156), Map 6 (PARTITION-LEVEL SORT, 156) +Reducer 3 <- Reducer 2 (PARTITION-LEVEL SORT, 270), Reducer 9 (PARTITION-LEVEL SORT, 270) +Reducer 4 <- Reducer 3 (GROUP, 1) +Reducer 5 <- Reducer 4 (SORT, 1) +Reducer 9 <- Map 8 (GROUP, 169) A masked pattern was here Vertices: Map 1 @@ -131,25 +131,13 @@ STAGE PLANS: expressions: ws_sold_date_sk (type: int), ws_item_sk (type: int), ws_ext_discount_amt (type: decimal(7,2)) outputColumnNames: _col0, _col1, _col2 Statistics: Num rows: 144002668 Data size: 19580198212 Basic stats: COMPLETE Column stats: NONE - Map Join Operator -condition map: - Inner Join 0 to 1 -keys: - 0 _col0 (type: int) - 1 _col0 (type: int) -outputColumnNames: _col1, _col2 -input vertices: - 1 Map 5 -Statistics: Num rows: 158402938 Data size: 21538218500 Basic stats: COMPLETE Column stats: NONE -Reduce Output Operator - key expressions: _col1 (type: int) - sort order: + - Map-reduce partition columns: _col1 (type: int) - Statistics: Num rows: 158402938 Data size: 21538218500 Basic stats: COMPLETE Column stats: NONE - value expressions: _col2 (type: decimal(7,2)) -Local Work: - Map Reduce Local Work -Map 10 + Reduce Output Operator +key expressions: _col1 (type: int) +sort order: + +Map-reduce partition columns: _col1 (type: int) +Statistics: Num rows: 144002668 Data size: 19580198212 Basic stats: COMPLETE Column stats: NONE +value expressions: _col0 (type: int), _col2 (type: decimal(7,2)) +Map 6 Map Operator Tree: TableScan alias: item @@ -166,7 +154,7 @@ STAGE PLANS: sort order: + Map-reduce partition columns: _col0 (type: int) Statistics: Num rows: 231000 Data size: 331780228 Basic stats: COMPLETE Column stats: NONE -Map 6 +Map 8 Map Operator Tree: TableScan alias: web_sales @@ -186,7 +174,7 @@ STAGE PLANS: 1 _col0 (type: int) outputColumnNames: _col1, _col2 input vertices: - 1 Map 9 + 1 Map 10 Statistics: Num rows: 158402938 Data size: 21538218500 Basic stats: COMPLETE Column stats: NONE Group By Operator aggregations: sum(_col2), count(_col2) @@ -203,22 +191,50 @@ STAGE PLANS: Local Work: Map Reduce Local Work Reducer 2 +Local Work: + Map Reduce Local Work Reduce Operator Tree: Join Operator condition map: Inner Join 0 to 1 keys: 0 _col1 (type: int) - 1 _col2 (type: int) -outp
[11/13] hive git commit: HIVE-19128 : Update golden files for spark perf tests
http://git-wip-us.apache.org/repos/asf/hive/blob/9ce42cba/ql/src/test/results/clientpositive/perf/spark/query30.q.out -- diff --git a/ql/src/test/results/clientpositive/perf/spark/query30.q.out b/ql/src/test/results/clientpositive/perf/spark/query30.q.out index 6385984..399251d 100644 --- a/ql/src/test/results/clientpositive/perf/spark/query30.q.out +++ b/ql/src/test/results/clientpositive/perf/spark/query30.q.out @@ -66,72 +66,37 @@ STAGE PLANS: Stage: Stage-1 Spark Edges: -Reducer 10 <- Reducer 16 (PARTITION-LEVEL SORT, 262), Reducer 9 (PARTITION-LEVEL SORT, 262) -Reducer 14 <- Map 13 (PARTITION-LEVEL SORT, 11), Map 17 (PARTITION-LEVEL SORT, 11) -Reducer 15 <- Map 18 (PARTITION-LEVEL SORT, 329), Reducer 14 (PARTITION-LEVEL SORT, 329) -Reducer 16 <- Reducer 15 (GROUP PARTITION-LEVEL SORT, 349) -Reducer 2 <- Map 1 (PARTITION-LEVEL SORT, 697), Map 5 (PARTITION-LEVEL SORT, 697) -Reducer 3 <- Reducer 10 (PARTITION-LEVEL SORT, 656), Reducer 2 (PARTITION-LEVEL SORT, 656) -Reducer 4 <- Reducer 3 (SORT, 1) -Reducer 7 <- Map 11 (PARTITION-LEVEL SORT, 11), Map 6 (PARTITION-LEVEL SORT, 11) -Reducer 8 <- Map 12 (PARTITION-LEVEL SORT, 329), Reducer 7 (PARTITION-LEVEL SORT, 329) -Reducer 9 <- Reducer 8 (GROUP, 349) +Reducer 11 <- Map 10 (PARTITION-LEVEL SORT, 11), Map 14 (PARTITION-LEVEL SORT, 11) +Reducer 12 <- Map 15 (PARTITION-LEVEL SORT, 329), Reducer 11 (PARTITION-LEVEL SORT, 329) +Reducer 13 <- Reducer 12 (GROUP PARTITION-LEVEL SORT, 349) +Reducer 17 <- Map 16 (PARTITION-LEVEL SORT, 697), Map 18 (PARTITION-LEVEL SORT, 697) +Reducer 2 <- Map 1 (PARTITION-LEVEL SORT, 11), Map 8 (PARTITION-LEVEL SORT, 11) +Reducer 3 <- Map 9 (PARTITION-LEVEL SORT, 329), Reducer 2 (PARTITION-LEVEL SORT, 329) +Reducer 4 <- Reducer 3 (GROUP, 349) +Reducer 5 <- Reducer 13 (PARTITION-LEVEL SORT, 262), Reducer 4 (PARTITION-LEVEL SORT, 262) +Reducer 6 <- Reducer 17 (PARTITION-LEVEL SORT, 656), Reducer 5 (PARTITION-LEVEL SORT, 656) +Reducer 7 <- Reducer 6 (SORT, 1) A masked pattern was here Vertices: Map 1 Map Operator Tree: TableScan - alias: customer - Statistics: Num rows: 8000 Data size: 68801615852 Basic stats: COMPLETE Column stats: NONE - Filter Operator -predicate: (c_current_addr_sk is not null and c_customer_sk is not null) (type: boolean) -Statistics: Num rows: 8000 Data size: 68801615852 Basic stats: COMPLETE Column stats: NONE -Select Operator - expressions: c_customer_sk (type: int), c_customer_id (type: string), c_current_addr_sk (type: int), c_salutation (type: string), c_first_name (type: string), c_last_name (type: string), c_preferred_cust_flag (type: string), c_birth_day (type: int), c_birth_month (type: int), c_birth_year (type: int), c_birth_country (type: string), c_login (type: string), c_email_address (type: string), c_last_review_date (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13 - Statistics: Num rows: 8000 Data size: 68801615852 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator -key expressions: _col2 (type: int) -sort order: + -Map-reduce partition columns: _col2 (type: int) -Statistics: Num rows: 8000 Data size: 68801615852 Basic stats: COMPLETE Column stats: NONE -value expressions: _col0 (type: int), _col1 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: string), _col6 (type: string), _col7 (type: int), _col8 (type: int), _col9 (type: int), _col10 (type: string), _col11 (type: string), _col12 (type: string), _col13 (type: string) -Map 11 -Map Operator Tree: -TableScan - alias: date_dim - Statistics: Num rows: 73049 Data size: 81741831 Basic stats: COMPLETE Column stats: NONE - Filter Operator -predicate: ((d_year = 2002) and d_date_sk is not null) (type: boolean) -Statistics: Num rows: 36524 Data size: 40870356 Basic stats: COMPLETE Column stats: NONE -Select Operator - expressions: d_date_sk (type: int) - outputColumnNames: _col0 - Statistics: Num rows: 36524 Data size: 40870356 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator -key expressions: _col0 (type: i
hive git commit: Preparing for 3.1.0 development - Updated pom to 3.1.0-SNAPSHOT
Repository: hive Updated Branches: refs/heads/master 328d3f935 -> 109c594a1 Preparing for 3.1.0 development - Updated pom to 3.1.0-SNAPSHOT Project: http://git-wip-us.apache.org/repos/asf/hive/repo Commit: http://git-wip-us.apache.org/repos/asf/hive/commit/109c594a Tree: http://git-wip-us.apache.org/repos/asf/hive/tree/109c594a Diff: http://git-wip-us.apache.org/repos/asf/hive/diff/109c594a Branch: refs/heads/master Commit: 109c594a1dbe14e27ecb5c6f6bfa3e93f3796f11 Parents: 328d3f9 Author: Vineet Garg Authored: Mon Apr 9 15:17:11 2018 -0700 Committer: Vineet Garg Committed: Mon Apr 9 15:17:11 2018 -0700 -- accumulo-handler/pom.xml | 2 +- beeline/pom.xml | 2 +- classification/pom.xml | 2 +- cli/pom.xml | 2 +- common/pom.xml | 2 +- contrib/pom.xml | 2 +- druid-handler/pom.xml| 2 +- hbase-handler/pom.xml| 2 +- hcatalog/core/pom.xml| 2 +- hcatalog/hcatalog-pig-adapter/pom.xml| 2 +- hcatalog/pom.xml | 2 +- hcatalog/server-extensions/pom.xml | 2 +- hcatalog/streaming/pom.xml | 2 +- hcatalog/webhcat/java-client/pom.xml | 2 +- hcatalog/webhcat/svr/pom.xml | 2 +- hplsql/pom.xml | 2 +- itests/custom-serde/pom.xml | 2 +- itests/custom-udfs/pom.xml | 2 +- itests/custom-udfs/udf-classloader-udf1/pom.xml | 2 +- itests/custom-udfs/udf-classloader-udf2/pom.xml | 2 +- itests/custom-udfs/udf-classloader-util/pom.xml | 2 +- itests/custom-udfs/udf-vectorized-badexample/pom.xml | 2 +- itests/hcatalog-unit/pom.xml | 2 +- itests/hive-blobstore/pom.xml| 2 +- itests/hive-jmh/pom.xml | 2 +- itests/hive-minikdc/pom.xml | 2 +- itests/hive-unit-hadoop2/pom.xml | 2 +- itests/hive-unit/pom.xml | 2 +- itests/pom.xml | 4 ++-- itests/qtest-accumulo/pom.xml| 2 +- itests/qtest-druid/pom.xml | 2 +- itests/qtest-spark/pom.xml | 2 +- itests/qtest/pom.xml | 2 +- itests/test-serde/pom.xml| 2 +- itests/util/pom.xml | 2 +- jdbc-handler/pom.xml | 2 +- jdbc/pom.xml | 2 +- kryo-registrator/pom.xml | 2 +- llap-client/pom.xml | 2 +- llap-common/pom.xml | 2 +- llap-ext-client/pom.xml | 2 +- llap-server/pom.xml | 2 +- llap-tez/pom.xml | 2 +- metastore/pom.xml| 2 +- packaging/pom.xml| 2 +- pom.xml | 2 +- ql/pom.xml | 2 +- serde/pom.xml| 2 +- service-rpc/pom.xml | 2 +- service/pom.xml | 2 +- shims/0.23/pom.xml | 2 +- shims/aggregator/pom.xml | 2 +- shims/common/pom.xml | 2 +- shims/pom.xml| 2 +- shims/scheduler/pom.xml | 2 +- spark-client/pom.xml | 4 ++-- standalone-metastore/pom.xml | 2 +- testutils/pom.xml| 2 +- vector-code-gen/pom.xml | 2 +- 59 files changed, 61 insertions(+), 61 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hive/blob/109c594a/accumulo-handler/pom.xml -- diff --git a/accumulo-handler/pom.xml b/accumulo-handler/pom.xml index edac1b1..b844c4a 100644 --- a/accumulo-handler/pom.xml +++ b/accumulo-handler/pom.xml @@ -19,7 +19,7 @@ org.apache.hive hive -3.0.0-SNAPSHOT +3.1.0-SNAPSHOT ../pom.xml http://git-wip-us.apache.org/repos/asf/hive/blob/109c594a/beeline/pom.xml -- diff --
hive git commit: HIVE-19014: utilize YARN-8028 (queue ACL check) in Hive Tez session pool (Sergey Shelukhin, reviewed by Jason Dere)
Repository: hive Updated Branches: refs/heads/master 109c594a1 -> 76b696c26 HIVE-19014: utilize YARN-8028 (queue ACL check) in Hive Tez session pool (Sergey Shelukhin, reviewed by Jason Dere) Project: http://git-wip-us.apache.org/repos/asf/hive/repo Commit: http://git-wip-us.apache.org/repos/asf/hive/commit/76b696c2 Tree: http://git-wip-us.apache.org/repos/asf/hive/tree/76b696c2 Diff: http://git-wip-us.apache.org/repos/asf/hive/diff/76b696c2 Branch: refs/heads/master Commit: 76b696c266122851e9704b5cf4d6ffd55efe0240 Parents: 109c594 Author: Jason Dere Authored: Mon Apr 9 16:15:09 2018 -0700 Committer: Jason Dere Committed: Mon Apr 9 16:15:09 2018 -0700 -- .../org/apache/hadoop/hive/conf/HiveConf.java | 3 + .../java/org/apache/hadoop/hive/ql/Driver.java | 1 + .../hadoop/hive/ql/exec/FunctionTask.java | 2 +- .../hive/ql/exec/tez/TezSessionPoolManager.java | 60 ++-- .../apache/hadoop/hive/ql/exec/tez/TezTask.java | 13 +- .../hive/ql/exec/tez/YarnQueueHelper.java | 143 +++ .../hive/ql/parse/DDLSemanticAnalyzer.java | 1 + .../ql/udf/generic/GenericUDFLoggedInUser.java | 1 + .../apache/hive/service/server/HiveServer2.java | 9 +- 9 files changed, 214 insertions(+), 19 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hive/blob/76b696c2/common/src/java/org/apache/hadoop/hive/conf/HiveConf.java -- diff --git a/common/src/java/org/apache/hadoop/hive/conf/HiveConf.java b/common/src/java/org/apache/hadoop/hive/conf/HiveConf.java index 0627c35..17b2485 100644 --- a/common/src/java/org/apache/hadoop/hive/conf/HiveConf.java +++ b/common/src/java/org/apache/hadoop/hive/conf/HiveConf.java @@ -3011,6 +3011,9 @@ public class HiveConf extends Configuration { "This flag is used in HiveServer2 to enable a user to use HiveServer2 without\n" + "turning on Tez for HiveServer2. The user could potentially want to run queries\n" + "over Tez without the pool of sessions."), +HIVE_SERVER2_TEZ_QUEUE_ACCESS_CHECK("hive.server2.tez.queue.access.check", false, +"Whether to check user access to explicitly specified YARN queues. " + + "yarn.resourcemanager.webapp.address must be configured to use this."), HIVE_SERVER2_TEZ_SESSION_LIFETIME("hive.server2.tez.session.lifetime", "162h", new TimeValidator(TimeUnit.HOURS), "The lifetime of the Tez sessions launched by HS2 when default sessions are enabled.\n" + http://git-wip-us.apache.org/repos/asf/hive/blob/76b696c2/ql/src/java/org/apache/hadoop/hive/ql/Driver.java -- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/Driver.java b/ql/src/java/org/apache/hadoop/hive/ql/Driver.java index 79db006..a88453c 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/Driver.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/Driver.java @@ -2045,6 +2045,7 @@ public class Driver implements IDriver { SessionState ss = SessionState.get(); + // TODO: should this use getUserFromAuthenticator? hookContext = new PrivateHookContext(plan, queryState, ctx.getPathToCS(), SessionState.get().getUserName(), ss.getUserIpAddress(), InetAddress.getLocalHost().getHostAddress(), operationId, ss.getSessionId(), Thread.currentThread().getName(), ss.isHiveServerQuery(), perfLogger, queryInfo, ctx); http://git-wip-us.apache.org/repos/asf/hive/blob/76b696c2/ql/src/java/org/apache/hadoop/hive/ql/exec/FunctionTask.java -- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/exec/FunctionTask.java b/ql/src/java/org/apache/hadoop/hive/ql/exec/FunctionTask.java index 1de333e..a0a90a9 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/exec/FunctionTask.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/exec/FunctionTask.java @@ -185,7 +185,7 @@ public class FunctionTask extends Task { funcName, dbName, className, -SessionState.get().getUserName(), +SessionState.get().getUserName(), // TODO: should this use getUserFromAuthenticator? PrincipalType.USER, (int) (System.currentTimeMillis() / 1000), org.apache.hadoop.hive.metastore.api.FunctionType.JAVA, http://git-wip-us.apache.org/repos/asf/hive/blob/76b696c2/ql/src/java/org/apache/hadoop/hive/ql/exec/tez/TezSessionPoolManager.java -- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/exec/tez/TezSessionPoolManager.java b/ql/src/java/org/apache/hadoop/hive/ql/exec/tez/TezSessionPoolManager.java index a051f90..2633390 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/exec/tez/TezSessionPoolManager.java +++ b/ql/src/jav
hive git commit: HIVE-19014: utilize YARN-8028 (queue ACL check) in Hive Tez session pool (Sergey Shelukhin, reviewed by Jason Dere)
Repository: hive Updated Branches: refs/heads/branch-3 9ce42cba8 -> 1cd74b451 HIVE-19014: utilize YARN-8028 (queue ACL check) in Hive Tez session pool (Sergey Shelukhin, reviewed by Jason Dere) Project: http://git-wip-us.apache.org/repos/asf/hive/repo Commit: http://git-wip-us.apache.org/repos/asf/hive/commit/1cd74b45 Tree: http://git-wip-us.apache.org/repos/asf/hive/tree/1cd74b45 Diff: http://git-wip-us.apache.org/repos/asf/hive/diff/1cd74b45 Branch: refs/heads/branch-3 Commit: 1cd74b451c0aa707ed28b2531f72971e8904fade Parents: 9ce42cb Author: Jason Dere Authored: Mon Apr 9 16:15:09 2018 -0700 Committer: Jason Dere Committed: Mon Apr 9 16:27:49 2018 -0700 -- .../org/apache/hadoop/hive/conf/HiveConf.java | 3 + .../java/org/apache/hadoop/hive/ql/Driver.java | 1 + .../hadoop/hive/ql/exec/FunctionTask.java | 2 +- .../hive/ql/exec/tez/TezSessionPoolManager.java | 60 ++-- .../apache/hadoop/hive/ql/exec/tez/TezTask.java | 13 +- .../hive/ql/exec/tez/YarnQueueHelper.java | 143 +++ .../hive/ql/parse/DDLSemanticAnalyzer.java | 1 + .../ql/udf/generic/GenericUDFLoggedInUser.java | 1 + .../apache/hive/service/server/HiveServer2.java | 9 +- 9 files changed, 214 insertions(+), 19 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hive/blob/1cd74b45/common/src/java/org/apache/hadoop/hive/conf/HiveConf.java -- diff --git a/common/src/java/org/apache/hadoop/hive/conf/HiveConf.java b/common/src/java/org/apache/hadoop/hive/conf/HiveConf.java index 0627c35..17b2485 100644 --- a/common/src/java/org/apache/hadoop/hive/conf/HiveConf.java +++ b/common/src/java/org/apache/hadoop/hive/conf/HiveConf.java @@ -3011,6 +3011,9 @@ public class HiveConf extends Configuration { "This flag is used in HiveServer2 to enable a user to use HiveServer2 without\n" + "turning on Tez for HiveServer2. The user could potentially want to run queries\n" + "over Tez without the pool of sessions."), +HIVE_SERVER2_TEZ_QUEUE_ACCESS_CHECK("hive.server2.tez.queue.access.check", false, +"Whether to check user access to explicitly specified YARN queues. " + + "yarn.resourcemanager.webapp.address must be configured to use this."), HIVE_SERVER2_TEZ_SESSION_LIFETIME("hive.server2.tez.session.lifetime", "162h", new TimeValidator(TimeUnit.HOURS), "The lifetime of the Tez sessions launched by HS2 when default sessions are enabled.\n" + http://git-wip-us.apache.org/repos/asf/hive/blob/1cd74b45/ql/src/java/org/apache/hadoop/hive/ql/Driver.java -- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/Driver.java b/ql/src/java/org/apache/hadoop/hive/ql/Driver.java index 79db006..a88453c 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/Driver.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/Driver.java @@ -2045,6 +2045,7 @@ public class Driver implements IDriver { SessionState ss = SessionState.get(); + // TODO: should this use getUserFromAuthenticator? hookContext = new PrivateHookContext(plan, queryState, ctx.getPathToCS(), SessionState.get().getUserName(), ss.getUserIpAddress(), InetAddress.getLocalHost().getHostAddress(), operationId, ss.getSessionId(), Thread.currentThread().getName(), ss.isHiveServerQuery(), perfLogger, queryInfo, ctx); http://git-wip-us.apache.org/repos/asf/hive/blob/1cd74b45/ql/src/java/org/apache/hadoop/hive/ql/exec/FunctionTask.java -- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/exec/FunctionTask.java b/ql/src/java/org/apache/hadoop/hive/ql/exec/FunctionTask.java index 1de333e..a0a90a9 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/exec/FunctionTask.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/exec/FunctionTask.java @@ -185,7 +185,7 @@ public class FunctionTask extends Task { funcName, dbName, className, -SessionState.get().getUserName(), +SessionState.get().getUserName(), // TODO: should this use getUserFromAuthenticator? PrincipalType.USER, (int) (System.currentTimeMillis() / 1000), org.apache.hadoop.hive.metastore.api.FunctionType.JAVA, http://git-wip-us.apache.org/repos/asf/hive/blob/1cd74b45/ql/src/java/org/apache/hadoop/hive/ql/exec/tez/TezSessionPoolManager.java -- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/exec/tez/TezSessionPoolManager.java b/ql/src/java/org/apache/hadoop/hive/ql/exec/tez/TezSessionPoolManager.java index a051f90..2633390 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/exec/tez/TezSessionPoolManager.java +++ b/ql/src