Re: Cloning Hive Trunk: HadoopArchives Dependency Not Found

2015-06-24 Thread Nishant Kelkar
Hmm, so I just ran the same command with 'mvn clean compile -Phadoop-1' and
'mvn clean compile -Phadoop-2' and they seem to work.

Thanks @all! I guess this wasn't an issue. :)

Thanks and Regards,
Nishant Kelkar

On Wed, Jun 24, 2015 at 1:59 AM, Nishant Kelkar nishant@gmail.com
wrote:

 Hi all,

 So I am a beginner to Hive, pardon me if I'm doing something wrong. I
 recently cloned the Hive dev trunk (master branch) via Git. I then went
 into my IDE (IntelliJ) and started looking around. At first, it resolved a
 lot of dependencies. However, after my IDE finished pulling from different
 sources, I saw that it was still missing the *HadoopArchives* dependency.

 Later on, I was trying to build the hive/ql sub-project with 'mvn clean
 compile', but I'm getting the following error:
 *---BEGIN---*

 [ERROR] Failed to execute goal
 org.apache.maven.plugins:maven-compiler-plugin:3.1:compile
 (default-compile) on project hive-exec: Compilation failure: Compilation
 failure:

 [ERROR]
 /Users/nishantkelkar/hive/ql/src/java/org/apache/hadoop/hive/ql/exec/DDLTask.java:[213,31]
 cannot find symbol

 [ERROR] symbol:   class HadoopArchives

 [ERROR] location: package org.apache.hadoop.tools

 [ERROR]
 /Users/nishantkelkar/hive/ql/src/java/org/apache/hadoop/hive/ql/hooks/ATSHook.java:[42,41]
 cannot find symbol

 [ERROR] symbol:   class TimelineClient

 [ERROR] location: package org.apache.hadoop.yarn.client.api

 [ERROR]
 /Users/nishantkelkar/hive/ql/src/java/org/apache/hadoop/hive/ql/hooks/ATSHook.java:[57,18]
 cannot find symbol

 [ERROR] symbol:   class TimelineClient

 [ERROR] location: class org.apache.hadoop.hive.ql.hooks.ATSHook

 [ERROR]
 /Users/nishantkelkar/hive/ql/src/java/org/apache/hadoop/hive/ql/exec/DDLTask.java:[1351,9]
 cannot find symbol

 [ERROR] symbol:   class HadoopArchives

 [ERROR] location: class org.apache.hadoop.hive.ql.exec.DDLTask

 [ERROR]
 /Users/nishantkelkar/hive/ql/src/java/org/apache/hadoop/hive/ql/exec/DDLTask.java:[1351,34]
 cannot find symbol

 [ERROR] symbol:   class HadoopArchives

 [ERROR] location: class org.apache.hadoop.hive.ql.exec.DDLTask

 [ERROR]
 /Users/nishantkelkar/hive/ql/src/java/org/apache/hadoop/hive/ql/hooks/ATSHook.java:[72,26]
 cannot find symbol

 [ERROR] symbol:   variable TimelineClient

 [ERROR] location: class org.apache.hadoop.hive.ql.hooks.ATSHook

 [ERROR] - [Help 1]
 *---END---*
 I looked at the pom.xml, and the following is highlighted in red:

 dependency
   groupIdorg.apache.hadoop/groupId
   artifactIdhadoop-archives/artifactId
   version${hadoop-23.version}/version
   optionaltrue/optional
 /dependency


 This basically means that IntelliJ/Maven was not able to find the
 dependency under the Maven repo. I saw the ticket HIVE-9048
 https://issues.apache.org/jira/browse/HIVE-9048 and HadoopArchives
 should be a sub-dependency of the hadoop-tools dependency right?

 Here's a copy of the relevant mirrors that I'm using in my settings.xml:

 mirror

   idConjars/id

   mirrorOf*/mirrorOf

   nameUS Conjars/name

   urlhttp://conjars.org/repo/url

 /mirror

 mirror

   idUS/id

   mirrorOfcentral/mirrorOf

   nameUS Central/name

   urlhttp://repo.maven.apache.org/maven2/url

 /mirror


 Does anyone have an idea why this could be happening? I was kind of
 expecting just cloning from Git, and running 'mvn clean compile' and seeing
 everything pass and build with success.

 Best Regards,

 Nishant Kelkar



Re: Cloning Hive Trunk: HadoopArchives Dependency Not Found

2015-06-24 Thread Nishant Kelkar
Hi all,

So I am a beginner to Hive, pardon me if I'm doing something wrong. I
recently cloned the Hive dev trunk (master branch) via Git. I then went
into my IDE (IntelliJ) and started looking around. At first, it resolved a
lot of dependencies. However, after my IDE finished pulling from different
sources, I saw that it was still missing the *HadoopArchives* dependency.

Later on, I was trying to build the hive/ql sub-project with 'mvn clean
compile', but I'm getting the following error:
*---BEGIN---*

[ERROR] Failed to execute goal
org.apache.maven.plugins:maven-compiler-plugin:3.1:compile
(default-compile) on project hive-exec: Compilation failure: Compilation
failure:

[ERROR]
/Users/nishantkelkar/hive/ql/src/java/org/apache/hadoop/hive/ql/exec/DDLTask.java:[213,31]
cannot find symbol

[ERROR] symbol:   class HadoopArchives

[ERROR] location: package org.apache.hadoop.tools

[ERROR]
/Users/nishantkelkar/hive/ql/src/java/org/apache/hadoop/hive/ql/hooks/ATSHook.java:[42,41]
cannot find symbol

[ERROR] symbol:   class TimelineClient

[ERROR] location: package org.apache.hadoop.yarn.client.api

[ERROR]
/Users/nishantkelkar/hive/ql/src/java/org/apache/hadoop/hive/ql/hooks/ATSHook.java:[57,18]
cannot find symbol

[ERROR] symbol:   class TimelineClient

[ERROR] location: class org.apache.hadoop.hive.ql.hooks.ATSHook

[ERROR]
/Users/nishantkelkar/hive/ql/src/java/org/apache/hadoop/hive/ql/exec/DDLTask.java:[1351,9]
cannot find symbol

[ERROR] symbol:   class HadoopArchives

[ERROR] location: class org.apache.hadoop.hive.ql.exec.DDLTask

[ERROR]
/Users/nishantkelkar/hive/ql/src/java/org/apache/hadoop/hive/ql/exec/DDLTask.java:[1351,34]
cannot find symbol

[ERROR] symbol:   class HadoopArchives

[ERROR] location: class org.apache.hadoop.hive.ql.exec.DDLTask

[ERROR]
/Users/nishantkelkar/hive/ql/src/java/org/apache/hadoop/hive/ql/hooks/ATSHook.java:[72,26]
cannot find symbol

[ERROR] symbol:   variable TimelineClient

[ERROR] location: class org.apache.hadoop.hive.ql.hooks.ATSHook

[ERROR] - [Help 1]
*---END---*
I looked at the pom.xml, and the following is highlighted in red:

dependency
  groupIdorg.apache.hadoop/groupId
  artifactIdhadoop-archives/artifactId
  version${hadoop-23.version}/version
  optionaltrue/optional
/dependency


This basically means that IntelliJ/Maven was not able to find the
dependency under the Maven repo. I saw the ticket HIVE-9048
https://issues.apache.org/jira/browse/HIVE-9048 and HadoopArchives should
be a sub-dependency of the hadoop-tools dependency right?

Here's a copy of the relevant mirrors that I'm using in my settings.xml:

mirror

  idConjars/id

  mirrorOf*/mirrorOf

  nameUS Conjars/name

  urlhttp://conjars.org/repo/url

/mirror

mirror

  idUS/id

  mirrorOfcentral/mirrorOf

  nameUS Central/name

  urlhttp://repo.maven.apache.org/maven2/url

/mirror


Does anyone have an idea why this could be happening? I was kind of
expecting just cloning from Git, and running 'mvn clean compile' and seeing
everything pass and build with success.

Best Regards,

Nishant Kelkar


Re: Review Request 34757: HIVE-10844: Combine equivalent Works for HoS[Spark Branch]

2015-06-24 Thread chengxiang li


 On 六月 23, 2015, 1:31 p.m., Xuefu Zhang wrote:
  ql/src/test/results/clientpositive/spark/groupby10.q.out, line 60
  https://reviews.apache.org/r/34757/diff/3-4/?file=988071#file988071line60
 
  Interesting. How come we got more stages now?

Not sure, introduced by latest merge from trunk.


- chengxiang


---
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/34757/#review88966
---


On 六月 23, 2015, 7:24 a.m., chengxiang li wrote:
 
 ---
 This is an automatically generated e-mail. To reply, visit:
 https://reviews.apache.org/r/34757/
 ---
 
 (Updated 六月 23, 2015, 7:24 a.m.)
 
 
 Review request for hive and Xuefu Zhang.
 
 
 Bugs: HIVE-10844
 https://issues.apache.org/jira/browse/HIVE-10844
 
 
 Repository: hive-git
 
 
 Description
 ---
 
 Some Hive queries(like TPCDS Q39) may share the same subquery, which 
 translated into sperate, but equivalent Works in SparkWork, combining these 
 equivalent Works into a single one would help to benifit from following 
 dynamic RDD caching optimization.
 
 
 Diffs
 -
 
   
 ql/src/java/org/apache/hadoop/hive/ql/optimizer/OperatorComparatorFactory.java
  PRE-CREATION 
   
 ql/src/java/org/apache/hadoop/hive/ql/optimizer/spark/CombineEquivalentWorkResolver.java
  PRE-CREATION 
   ql/src/java/org/apache/hadoop/hive/ql/parse/spark/SparkCompiler.java 
 19aae70 
   ql/src/java/org/apache/hadoop/hive/ql/plan/JoinCondDesc.java b307b16 
   ql/src/test/results/clientpositive/spark/auto_join30.q.out 7b5c5e7 
   ql/src/test/results/clientpositive/spark/auto_smb_mapjoin_14.q.out 8a43d78 
   ql/src/test/results/clientpositive/spark/groupby10.q.out dd9d9fe 
   ql/src/test/results/clientpositive/spark/groupby7_map.q.out abd6459 
   ql/src/test/results/clientpositive/spark/groupby7_map_skew.q.out 5e69b31 
   ql/src/test/results/clientpositive/spark/groupby7_noskew.q.out 3418b99 
   
 ql/src/test/results/clientpositive/spark/groupby7_noskew_multi_single_reducer.q.out
  2cb126d 
   ql/src/test/results/clientpositive/spark/groupby8.q.out c249b61 
   ql/src/test/results/clientpositive/spark/groupby8_map_skew.q.out 2fb1d73 
   ql/src/test/results/clientpositive/spark/insert_into3.q.out 7df5ba8 
   ql/src/test/results/clientpositive/spark/join22.q.out b1e5b67 
   ql/src/test/results/clientpositive/spark/skewjoinopt11.q.out 8a278ef 
   ql/src/test/results/clientpositive/spark/union10.q.out 5e8fe38 
   ql/src/test/results/clientpositive/spark/union11.q.out 20c27c7 
   ql/src/test/results/clientpositive/spark/union20.q.out 6f0dca6 
   ql/src/test/results/clientpositive/spark/union28.q.out 98582df 
   ql/src/test/results/clientpositive/spark/union3.q.out 834b6d4 
   ql/src/test/results/clientpositive/spark/union30.q.out 3409623 
   ql/src/test/results/clientpositive/spark/union4.q.out c121ef0 
   ql/src/test/results/clientpositive/spark/union5.q.out afee988 
   ql/src/test/results/clientpositive/spark/union_remove_1.q.out ba0e293 
   ql/src/test/results/clientpositive/spark/union_remove_15.q.out 26cfbab 
   ql/src/test/results/clientpositive/spark/union_remove_16.q.out 7a7aaf2 
   ql/src/test/results/clientpositive/spark/union_remove_18.q.out a5e15c5 
   ql/src/test/results/clientpositive/spark/union_remove_19.q.out ad44400 
   ql/src/test/results/clientpositive/spark/union_remove_20.q.out 1d67177 
   ql/src/test/results/clientpositive/spark/union_remove_21.q.out 9f5b070 
   ql/src/test/results/clientpositive/spark/union_remove_22.q.out 2e01432 
   ql/src/test/results/clientpositive/spark/union_remove_24.q.out 2659798 
   ql/src/test/results/clientpositive/spark/union_remove_25.q.out 0a94684 
   ql/src/test/results/clientpositive/spark/union_remove_4.q.out 6c3d596 
   ql/src/test/results/clientpositive/spark/union_remove_6.q.out cd36189 
   ql/src/test/results/clientpositive/spark/union_remove_6_subq.q.out c981ae4 
   ql/src/test/results/clientpositive/spark/union_remove_7.q.out 084fbd6 
   ql/src/test/results/clientpositive/spark/union_top_level.q.out dede1ef 
 
 Diff: https://reviews.apache.org/r/34757/diff/
 
 
 Testing
 ---
 
 
 Thanks,
 
 chengxiang li
 




Re: Review Request 35107: HIVE-6791 Support variable substition for Beeline shell command

2015-06-24 Thread cheng xu

---
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/35107/
---

(Updated June 24, 2015, 3:39 p.m.)


Review request for hive, chinna and Xuefu Zhang.


Changes
---

Update patch addressing Xuefu's comments


Bugs: HIVE-6791
https://issues.apache.org/jira/browse/HIVE-6791


Repository: hive-git


Description
---

Summary:
1) move the beeline-cli convertor to the place where cli is executed(class 
**Commands**)
2) support substitution for source command
3) add some unit test for substitution
4) add one way to get the configuration from HS2


Diffs (updated)
-

  beeline/src/java/org/apache/hive/beeline/BeeLine.java b7d2f2e 
  beeline/src/java/org/apache/hive/beeline/Commands.java a42baa3 
  beeline/src/test/org/apache/hive/beeline/cli/TestHiveCli.java 6cbb030 
  cli/src/java/org/apache/hadoop/hive/cli/CliDriver.java d62fd5c 
  common/src/java/org/apache/hadoop/hive/conf/HiveVariableSource.java 
PRE-CREATION 
  common/src/java/org/apache/hadoop/hive/conf/VariableSubstitution.java 
PRE-CREATION 
  common/src/test/org/apache/hadoop/hive/conf/TestVariableSubstitution.java 
PRE-CREATION 
  ql/src/java/org/apache/hadoop/hive/ql/Driver.java 338e755 
  ql/src/java/org/apache/hadoop/hive/ql/parse/ColumnStatsSemanticAnalyzer.java 
a5f0a7f 
  ql/src/java/org/apache/hadoop/hive/ql/parse/VariableSubstitution.java e8b1d96 
  ql/src/java/org/apache/hadoop/hive/ql/processors/AddResourceProcessor.java 
0558c53 
  ql/src/java/org/apache/hadoop/hive/ql/processors/CompileProcessor.java 
25ce168 
  ql/src/java/org/apache/hadoop/hive/ql/processors/DeleteResourceProcessor.java 
9052c82 
  ql/src/java/org/apache/hadoop/hive/ql/processors/DfsProcessor.java cc0414d 
  ql/src/java/org/apache/hadoop/hive/ql/processors/SetProcessor.java bc9254c 
  service/src/java/org/apache/hive/service/cli/operation/SQLOperation.java 
33ee16b 

Diff: https://reviews.apache.org/r/35107/diff/


Testing
---

Unit test passed


Thanks,

cheng xu



[jira] [Created] (HIVE-11090) ordering issues with windows unit test runs

2015-06-24 Thread Matt McCline (JIRA)
Matt McCline created HIVE-11090:
---

 Summary: ordering issues with windows unit test runs
 Key: HIVE-11090
 URL: https://issues.apache.org/jira/browse/HIVE-11090
 Project: Hive
  Issue Type: Bug
  Components: Hive
Affects Versions: 1.2.1
Reporter: Matt McCline
Assignee: Matt McCline






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


Re: Review Request 34757: HIVE-10844: Combine equivalent Works for HoS[Spark Branch]

2015-06-24 Thread chengxiang li


 On 六月 19, 2015, 1:47 p.m., Xuefu Zhang wrote:
  ql/src/java/org/apache/hadoop/hive/ql/optimizer/spark/CombineEquivalentWorkResolver.java,
   line 98
  https://reviews.apache.org/r/34757/diff/3/?file=988066#file988066line98
 
  I think the recursion should go on even if there is only one child for 
  a given work. For examle, if we have:
  
   w1
   |
   w2
   |
   w3
   / \
  w4  w5
  
  Even if each of w1 and w2 has only one child, it's still possible that 
  we can combine w4 and w5.

created HIVE-11082 to track this.


 On 六月 19, 2015, 1:47 p.m., Xuefu Zhang wrote:
  ql/src/java/org/apache/hadoop/hive/ql/optimizer/spark/CombineEquivalentWorkResolver.java,
   line 207
  https://reviews.apache.org/r/34757/diff/3/?file=988066#file988066line207
 
  Could you explain the reason here?

add comments in latest patch. While combine multi equivalent works into single 
one, we need to update all the references to the replaced works. leave works 
output should be read by further SparkWork/FetchWork, we does not able to 
update work reference across SparkWork, so combine leave works may lead to 
error.


- chengxiang


---
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/34757/#review88537
---


On 六月 23, 2015, 7:24 a.m., chengxiang li wrote:
 
 ---
 This is an automatically generated e-mail. To reply, visit:
 https://reviews.apache.org/r/34757/
 ---
 
 (Updated 六月 23, 2015, 7:24 a.m.)
 
 
 Review request for hive and Xuefu Zhang.
 
 
 Bugs: HIVE-10844
 https://issues.apache.org/jira/browse/HIVE-10844
 
 
 Repository: hive-git
 
 
 Description
 ---
 
 Some Hive queries(like TPCDS Q39) may share the same subquery, which 
 translated into sperate, but equivalent Works in SparkWork, combining these 
 equivalent Works into a single one would help to benifit from following 
 dynamic RDD caching optimization.
 
 
 Diffs
 -
 
   
 ql/src/java/org/apache/hadoop/hive/ql/optimizer/OperatorComparatorFactory.java
  PRE-CREATION 
   
 ql/src/java/org/apache/hadoop/hive/ql/optimizer/spark/CombineEquivalentWorkResolver.java
  PRE-CREATION 
   ql/src/java/org/apache/hadoop/hive/ql/parse/spark/SparkCompiler.java 
 19aae70 
   ql/src/java/org/apache/hadoop/hive/ql/plan/JoinCondDesc.java b307b16 
   ql/src/test/results/clientpositive/spark/auto_join30.q.out 7b5c5e7 
   ql/src/test/results/clientpositive/spark/auto_smb_mapjoin_14.q.out 8a43d78 
   ql/src/test/results/clientpositive/spark/groupby10.q.out dd9d9fe 
   ql/src/test/results/clientpositive/spark/groupby7_map.q.out abd6459 
   ql/src/test/results/clientpositive/spark/groupby7_map_skew.q.out 5e69b31 
   ql/src/test/results/clientpositive/spark/groupby7_noskew.q.out 3418b99 
   
 ql/src/test/results/clientpositive/spark/groupby7_noskew_multi_single_reducer.q.out
  2cb126d 
   ql/src/test/results/clientpositive/spark/groupby8.q.out c249b61 
   ql/src/test/results/clientpositive/spark/groupby8_map_skew.q.out 2fb1d73 
   ql/src/test/results/clientpositive/spark/insert_into3.q.out 7df5ba8 
   ql/src/test/results/clientpositive/spark/join22.q.out b1e5b67 
   ql/src/test/results/clientpositive/spark/skewjoinopt11.q.out 8a278ef 
   ql/src/test/results/clientpositive/spark/union10.q.out 5e8fe38 
   ql/src/test/results/clientpositive/spark/union11.q.out 20c27c7 
   ql/src/test/results/clientpositive/spark/union20.q.out 6f0dca6 
   ql/src/test/results/clientpositive/spark/union28.q.out 98582df 
   ql/src/test/results/clientpositive/spark/union3.q.out 834b6d4 
   ql/src/test/results/clientpositive/spark/union30.q.out 3409623 
   ql/src/test/results/clientpositive/spark/union4.q.out c121ef0 
   ql/src/test/results/clientpositive/spark/union5.q.out afee988 
   ql/src/test/results/clientpositive/spark/union_remove_1.q.out ba0e293 
   ql/src/test/results/clientpositive/spark/union_remove_15.q.out 26cfbab 
   ql/src/test/results/clientpositive/spark/union_remove_16.q.out 7a7aaf2 
   ql/src/test/results/clientpositive/spark/union_remove_18.q.out a5e15c5 
   ql/src/test/results/clientpositive/spark/union_remove_19.q.out ad44400 
   ql/src/test/results/clientpositive/spark/union_remove_20.q.out 1d67177 
   ql/src/test/results/clientpositive/spark/union_remove_21.q.out 9f5b070 
   ql/src/test/results/clientpositive/spark/union_remove_22.q.out 2e01432 
   ql/src/test/results/clientpositive/spark/union_remove_24.q.out 2659798 
   ql/src/test/results/clientpositive/spark/union_remove_25.q.out 0a94684 
   ql/src/test/results/clientpositive/spark/union_remove_4.q.out 6c3d596 
   ql/src/test/results/clientpositive/spark/union_remove_6.q.out cd36189 
   ql/src/test/results/clientpositive/spark/union_remove_6_subq.q.out c981ae4 
   

[jira] [Created] (HIVE-11091) Unable to load data into hive table using Load data local inapth command from unix named pipe

2015-06-24 Thread Manoranjan Sahoo (JIRA)
Manoranjan Sahoo created HIVE-11091:
---

 Summary: Unable to load data into hive table using Load data 
local inapth command from unix named pipe
 Key: HIVE-11091
 URL: https://issues.apache.org/jira/browse/HIVE-11091
 Project: Hive
  Issue Type: Bug
  Components: Hive
Affects Versions: 0.14.0
 Environment: Unix,MacOS
Reporter: Manoranjan Sahoo
Priority: Blocker


Unable to load data into hive table from unix named pipe in Hive 0.14.0 

Please find below the execution details in env ( Hadoop2.6.0 + Hive 0.14.0):

$ mkfifo /tmp/test.txt
$ hive
hive create table test(id bigint,name string);
OK
Time taken: 1.018 seconds
hive LOAD DATA LOCAL INPATH '/tmp/test.txt' OVERWRITE INTO TABLE test;
Loading data to table default.test
Failed with exception addFiles: filesystem error in check phase
FAILED: Execution Error, return code 1 from 
org.apache.hadoop.hive.ql.exec.MoveTask

But in Hadoop 1.3 and hive 0.11.0  it works fine:
hive LOAD DATA LOCAL INPATH '/tmp/test.txt' OVERWRITE INTO TABLE test;
Copying data from file:/tmp/test.txt
Copying file: file:/tmp/test.txt



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[GitHub] hive pull request: Update from original

2015-06-24 Thread gunnartangring
GitHub user gunnartangring opened a pull request:

https://github.com/apache/hive/pull/42

Update from original



You can merge this pull request into a Git repository by running:

$ git pull https://github.com/gunnartangring/hive trunk

Alternatively you can review and apply these changes as the patch at:

https://github.com/apache/hive/pull/42.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

This closes #42


commit 474f12672485cc41b041d50c1e131c20a3175b7f
Author: Jason Dere jd...@apache.org
Date:   2015-02-09T23:06:17Z

HIVE-9520: Create NEXT_DAY UDF (Alexander Pivovarov via Jason Dere)


git-svn-id: https://svn.apache.org/repos/asf/hive/trunk@1658600 
13f79535-47bb-0310-9956-ffa450edef68

commit 3edf89802b95259e90d6c20fc56f5bf29426b2a6
Author: Ashutosh Chauhan hashut...@apache.org
Date:   2015-02-10T00:46:21Z

HIVE-9228 : Problem with subquery using windowing functions (Navis via 
Ashutosh Chauhan)

git-svn-id: https://svn.apache.org/repos/asf/hive/trunk@1658606 
13f79535-47bb-0310-9956-ffa450edef68

commit efa823f4b80d2ca3005da4479cc0a67f96597644
Author: Navis Ryu na...@apache.org
Date:   2015-02-10T06:47:09Z

HIVE-9507 : Make LATERAL VIEW inline(expression) mytable tolerant to 
nulls (Navis reviewed by Ashutosh Chauhan)

git-svn-id: https://svn.apache.org/repos/asf/hive/trunk@1658625 
13f79535-47bb-0310-9956-ffa450edef68

commit 1e39f8e23b0fda4507ece1729974a73d994aff9d
Author: Brock Noland br...@apache.org
Date:   2015-02-10T17:49:19Z

HIVE-9454 - Test failures due to new Calcite version (Laljo via Brock)

git-svn-id: https://svn.apache.org/repos/asf/hive/trunk@1658766 
13f79535-47bb-0310-9956-ffa450edef68

commit 3e7a258f910c78110405c3564d1ab3122281ae27
Author: Xuefu Zhang xu...@apache.org
Date:   2015-02-10T19:38:40Z

HIVE-9574: Lazy computing in HiveBaseFunctionResultList may hurt 
performance [Spark Branch] (Jimmy via Xuefu)

git-svn-id: https://svn.apache.org/repos/asf/hive/branches/spark@1658786 
13f79535-47bb-0310-9956-ffa450edef68

commit 02b20e97133d8b44ebb73ac42b20bce0be19d715
Author: Xuefu Zhang xu...@apache.org
Date:   2015-02-10T19:46:42Z

HIVE-9627: Add cbo_gby_empty.q.out for Spark [Spark Branch] (Jimmy via 
Xuefu)

git-svn-id: https://svn.apache.org/repos/asf/hive/branches/spark@1658788 
13f79535-47bb-0310-9956-ffa450edef68

commit c6d850dcc5bcedb6b17aff4cb804585a79ae2505
Author: Xuefu Zhang xu...@apache.org
Date:   2015-02-11T05:27:29Z

HIVE-9646: Beeline doesn't show Spark job progress info [Spark Branch] 
(Jimmy via Xuefu)

git-svn-id: https://svn.apache.org/repos/asf/hive/branches/spark@1658877 
13f79535-47bb-0310-9956-ffa450edef68

commit f7248105508426c71b655c5dedeea69eec8fb4e5
Author: Jason Dere jd...@apache.org
Date:   2015-02-11T10:22:41Z

HIVE-9594: Add qtests for LAST_DAY udf (Alexander Pivovarov via Jason Dere)


git-svn-id: https://svn.apache.org/repos/asf/hive/trunk@1658914 
13f79535-47bb-0310-9956-ffa450edef68

commit e1b0674e5bca7773fc8656ec95541aa95dfbad4a
Author: Brock Noland br...@apache.org
Date:   2015-02-11T15:48:33Z

HIVE-9612 - Turn off DEBUG logging for Lazy Objects for tests (Sergio via 
Brock/Szehon)

git-svn-id: https://svn.apache.org/repos/asf/hive/trunk@1658991 
13f79535-47bb-0310-9956-ffa450edef68

commit 3004101d1aac5b0e39f2777174aa981e4930e5d0
Author: Brock Noland br...@apache.org
Date:   2015-02-11T18:10:09Z

HIVE-9651 - FileSinkOperator does not pass in conf to serde.initialize 
(Brock via Szehon)

git-svn-id: https://svn.apache.org/repos/asf/hive/trunk@1659027 
13f79535-47bb-0310-9956-ffa450edef68

commit 4df3c6becb7002a27e10374c4a000529c4f266a4
Author: Prasanth J prasan...@apache.org
Date:   2015-02-11T20:46:00Z

HIVE-9652: Tez in place updates should detect redirection of STDERR 
(Prasanth Jayachandran reviewed by Gopal V)

git-svn-id: https://svn.apache.org/repos/asf/hive/trunk@1659065 
13f79535-47bb-0310-9956-ffa450edef68

commit 5d175fcda6c5dad410d98ce963d1e16ef0f06cd1
Author: Vaibhav Gumashta vgumas...@apache.org
Date:   2015-02-11T21:29:28Z

HIVE-9621: HiveServer2 http mode - embedded jetty should use 
SynchronousQueue (Vaibhav Gumashta reviewed by Thejas Nair)

git-svn-id: https://svn.apache.org/repos/asf/hive/trunk@1659098 
13f79535-47bb-0310-9956-ffa450edef68

commit 09ee4803d015c51e864ee923fd60af50e972074c
Author: Owen O'Malley omal...@apache.org
Date:   2015-02-11T23:00:45Z

HIVE-9593. ORC Reader should ignore unknown metadata streams (omalley)


git-svn-id: https://svn.apache.org/repos/asf/hive/trunk@1659106 
13f79535-47bb-0310-9956-ffa450edef68

commit 4bdf3acdb0126adde9a016d458fa83232412133c
Author: Brock Noland br...@apache.org
Date:   2015-02-12T04:52:38Z

HIVE-9648 - Null check key provider before doing set (Brock via Chao)

git-svn-id: 

[GitHub] hive pull request: Update from original

2015-06-24 Thread gunnartangring
Github user gunnartangring closed the pull request at:

https://github.com/apache/hive/pull/42


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[jira] [Created] (HIVE-11093) ALTER TABLE for avro backed tables

2015-06-24 Thread Julian Keppel (JIRA)
Julian Keppel created HIVE-11093:


 Summary: ALTER TABLE for avro backed tables
 Key: HIVE-11093
 URL: https://issues.apache.org/jira/browse/HIVE-11093
 Project: Hive
  Issue Type: Bug
  Components: Hive
Affects Versions: 1.1.0
 Environment: Cloudera CDH 5.4.0
Reporter: Julian Keppel


I use Hive Version 1.1.0 in Cloudera CDH 5.4.0.

I have created an external table:

CREATE EXTERNAL TABLE *tableA*
ROW FORMAT SERDE 'org.apache.hadoop.hive.serde2.avro.AvroSerDe'
STORED AS
INPUTFORMAT
'org.apache.hadoop.hive.ql.io.avro.AvroContainerInputFormat'
OUTPUTFORMAT
'org.apache.hadoop.hive.ql.io.avro.AvroContainerOutputFormat'
LOCATION 'location'
TBLPROPERTIES ('avro.schema.url'=schema_location');

Now I wanted to create a managed table with exactly the same columns except
of one additional column (a generated hash key). The only idea I had was to
create the table with: CREATE TABLE *tableB* LIKE *tableA*;

And then add the additional column with: ALTER TABLE *tableB* ADD COLUMNS
(new_column INT);

The statements run without any errors or exceptions (even in the log files
under /var/log/hive) but the new column doesn't appear.

With a CSV table it worked properly. So I think this has to do with the fact 
that its a avro backed table. But according to this ticket 
(https://issues.apache.org/jira/browse/HIVE-7446) the feature was added in 
version 0.14.0. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HIVE-11092) First delta of an ORC ACID table contains incorrect schema

2015-06-24 Thread Elliot West (JIRA)
Elliot West created HIVE-11092:
--

 Summary: First delta of an ORC ACID table contains incorrect schema
 Key: HIVE-11092
 URL: https://issues.apache.org/jira/browse/HIVE-11092
 Project: Hive
  Issue Type: Bug
  Components: Hive
Reporter: Elliot West
Assignee: Elliot West
Priority: Minor


I've been reading ORC ACID data that backs transactional tables from a process 
external to Hive. Initially I tried to use 'schema on read' but found some 
inconsistencies in the schema returned from the initial delta file and 
subsequent delta and base files. To reproduce the issue by example:

{code}
CREATE TABLE base_table ( id int, message string )
  PARTITIONED BY ( continent string, country string )
  CLUSTERED BY (id) INTO 1 BUCKETS
  STORED AS ORC
  TBLPROPERTIES ('transactional' = 'true');
  
INSERT INTO TABLE base_table PARTITION (continent = 'Asia', country = 'India')
VALUES (1, 'x'), (2, 'y'), (3, 'z');

UPDATE base_table SET message = 'updated' WHERE id = 1;
{code}

Now examining the raw data with the {{orcfiledump}} utility (edited for 
brevity):
{code}
cd hive/warehouse/base_table/continent=Asia/country=India/

hive --orcfiledump delta_001_001/bucket_0
Type: 
structoperation:int,originalTransaction:bigint,bucket:int,rowId:bigint,currentTransaction:bigint,row:struct_col0:int,_col1:string


hive --orcfiledump delta_002_002/bucket_0
Type: 
structoperation:int,originalTransaction:bigint,bucket:int,rowId:bigint,currentTransaction:bigint,row:structid:int,message:string

{code}

The row schema for the first delta that resulted from the inserts has its field 
names erased: {{row:struct_col0:int,_col1:string}}, whereas the delta for the 
update reports the correct schema: {{row:structid:int,message:string}}. I 
have also checked this with my own reader code so am confident that 
{{FileDump}} is not at fault.

I believe that the row field names, and hence schema, should be consistent 
across all ORC files in the ACID data set. This will enable schema on read with 
field access by name (not index), which is currently not possible. Therefore 
I'd like to get this issue resolved.

I'm happy to work on this, however after working through {{OrcRecordUpdater}} 
and {{FileSinkOperator}} and related tests I've failed to reproduce or isolate 
the issue at a smaller scale. I'd be grateful for some suggestions on where to 
look next.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HIVE-11094) Beeline redirecting all output to ErrorStream

2015-06-24 Thread Jesus Camacho Rodriguez (JIRA)
Jesus Camacho Rodriguez created HIVE-11094:
--

 Summary: Beeline redirecting all output to ErrorStream
 Key: HIVE-11094
 URL: https://issues.apache.org/jira/browse/HIVE-11094
 Project: Hive
  Issue Type: Bug
Reporter: Jesus Camacho Rodriguez
Assignee: Jesus Camacho Rodriguez


Beeline is sending all output to ErrorStream, instead of using OutputStream for 
info or debug information.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HIVE-11099) Add support for running negative q-tests [Spark Branch]

2015-06-24 Thread Mohit Sabharwal (JIRA)
Mohit Sabharwal created HIVE-11099:
--

 Summary: Add support for running negative q-tests [Spark Branch]
 Key: HIVE-11099
 URL: https://issues.apache.org/jira/browse/HIVE-11099
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Reporter: Mohit Sabharwal
Assignee: Mohit Sabharwal


Add support for TestSparkNegativeCliDriver TestMiniSparkOnYarnNegativeCliDriver 
to negative q-tests



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HIVE-11098) Load aux jars without requiring restart

2015-06-24 Thread JIRA
Andrés Cordero created HIVE-11098:
-

 Summary: Load aux jars without requiring restart
 Key: HIVE-11098
 URL: https://issues.apache.org/jira/browse/HIVE-11098
 Project: Hive
  Issue Type: New Feature
  Components: UDF
Affects Versions: 0.13.0
Reporter: Andrés Cordero


Currently, loading auxiliary jars (e.g. for UDFs) requires a restart, even for 
files under the path.
Replacing an auxiliary jar also requires a restart to pick up the new contents.

Would be useful if changes were picked up automatically based on filesystem 
notifications, or provide a manual method of forcing a reload.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HIVE-11102) ReaderImpl: getColumnIndicesFromNames does not work for ACID tables

2015-06-24 Thread Gopal V (JIRA)
Gopal V created HIVE-11102:
--

 Summary: ReaderImpl: getColumnIndicesFromNames does not work for 
ACID tables
 Key: HIVE-11102
 URL: https://issues.apache.org/jira/browse/HIVE-11102
 Project: Hive
  Issue Type: Bug
  Components: File Formats
Affects Versions: 1.3.0, 1.2.1, 2.0.0
Reporter: Gopal V
Assignee: Gopal V


ORC reader impl does not estimate the size of ACID data files correctly.

{code}
Caused by: java.lang.IndexOutOfBoundsException: Index: 0
at java.util.Collections$EmptyList.get(Collections.java:3212)
at 
org.apache.hadoop.hive.ql.io.orc.OrcProto$Type.getSubtypes(OrcProto.java:12240)
at 
org.apache.hadoop.hive.ql.io.orc.ReaderImpl.getColumnIndicesFromNames(ReaderImpl.java:651)
at 
org.apache.hadoop.hive.ql.io.orc.ReaderImpl.getRawDataSizeOfColumns(ReaderImpl.java:634)
at 
org.apache.hadoop.hive.ql.io.orc.OrcInputFormat$SplitGenerator.populateAndCacheStripeDetails(OrcInputFormat.java:938)
at 
org.apache.hadoop.hive.ql.io.orc.OrcInputFormat$SplitGenerator.call(OrcInputFormat.java:847)
at 
org.apache.hadoop.hive.ql.io.orc.OrcInputFormat$SplitGenerator.call(OrcInputFormat.java:713)
at java.util.concurrent.FutureTask.run(FutureTask.java:262)
at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
at java.lang.Thread.run(Thread.java:744)
{code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HIVE-11101) Vectorization decimal precision issue in vectorization_short_regress.q

2015-06-24 Thread Matt McCline (JIRA)
Matt McCline created HIVE-11101:
---

 Summary: Vectorization decimal precision issue in 
vectorization_short_regress.q
 Key: HIVE-11101
 URL: https://issues.apache.org/jira/browse/HIVE-11101
 Project: Hive
  Issue Type: Bug
  Components: Hive
Reporter: Matt McCline
Assignee: Matt McCline


Noticed one query result line in vectorization_short_regress.q is different 
when that test is run without vectorization.

It is a decimal precision issue??

{code}
1785c1797
 1969-12-31 16:00:04.063   04XP4DrTCblC788515601.0 79.553  
-1452617198 15601   -407009.58195572987 -15858  -511684.9   
-15601.0158740.1750002  -6432.15344526  -79.553 NULL
-15601.0-2.43391201E8
---
 1969-12-31 16:00:04.063   04XP4DrTCblC788515601.0 79.553  
 -1452617198 15601   -407009.58195572987 -15858  -511684.9   
 -15601.0158740.1750002  -6432.0 -79.553 NULL-15601.0  
   -2.43391201E8
1886a1899
{code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


Re: Review Request 35107: HIVE-6791 Support variable substition for Beeline shell command

2015-06-24 Thread Xuefu Zhang

---
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/35107/#review89251
---



beeline/src/java/org/apache/hive/beeline/Commands.java (line 717)
https://reviews.apache.org/r/35107/#comment141839

Nit: method naming. substitution() - substituteVariables() or substitue()



beeline/src/java/org/apache/hive/beeline/Commands.java (line 791)
https://reviews.apache.org/r/35107/#comment141854

Do we need to close statement when done?



beeline/src/java/org/apache/hive/beeline/Commands.java (line 814)
https://reviews.apache.org/r/35107/#comment141842

Does this mean that env and sys variables are not being substituted for 
shell command?



beeline/src/java/org/apache/hive/beeline/Commands.java (line 861)
https://reviews.apache.org/r/35107/#comment141843

Don't we need to close reader object?



ql/src/java/org/apache/hadoop/hive/ql/processors/SetProcessor.java (line 145)
https://reviews.apache.org/r/35107/#comment141846

Could we keep @Override at a separate line? Same for other places.


- Xuefu Zhang


On June 24, 2015, 7:39 a.m., cheng xu wrote:
 
 ---
 This is an automatically generated e-mail. To reply, visit:
 https://reviews.apache.org/r/35107/
 ---
 
 (Updated June 24, 2015, 7:39 a.m.)
 
 
 Review request for hive, chinna and Xuefu Zhang.
 
 
 Bugs: HIVE-6791
 https://issues.apache.org/jira/browse/HIVE-6791
 
 
 Repository: hive-git
 
 
 Description
 ---
 
 Summary:
 1) move the beeline-cli convertor to the place where cli is executed(class 
 **Commands**)
 2) support substitution for source command
 3) add some unit test for substitution
 4) add one way to get the configuration from HS2
 
 
 Diffs
 -
 
   beeline/src/java/org/apache/hive/beeline/BeeLine.java b7d2f2e 
   beeline/src/java/org/apache/hive/beeline/Commands.java a42baa3 
   beeline/src/test/org/apache/hive/beeline/cli/TestHiveCli.java 6cbb030 
   cli/src/java/org/apache/hadoop/hive/cli/CliDriver.java d62fd5c 
   common/src/java/org/apache/hadoop/hive/conf/HiveVariableSource.java 
 PRE-CREATION 
   common/src/java/org/apache/hadoop/hive/conf/VariableSubstitution.java 
 PRE-CREATION 
   common/src/test/org/apache/hadoop/hive/conf/TestVariableSubstitution.java 
 PRE-CREATION 
   ql/src/java/org/apache/hadoop/hive/ql/Driver.java 338e755 
   
 ql/src/java/org/apache/hadoop/hive/ql/parse/ColumnStatsSemanticAnalyzer.java 
 a5f0a7f 
   ql/src/java/org/apache/hadoop/hive/ql/parse/VariableSubstitution.java 
 e8b1d96 
   ql/src/java/org/apache/hadoop/hive/ql/processors/AddResourceProcessor.java 
 0558c53 
   ql/src/java/org/apache/hadoop/hive/ql/processors/CompileProcessor.java 
 25ce168 
   
 ql/src/java/org/apache/hadoop/hive/ql/processors/DeleteResourceProcessor.java 
 9052c82 
   ql/src/java/org/apache/hadoop/hive/ql/processors/DfsProcessor.java cc0414d 
   ql/src/java/org/apache/hadoop/hive/ql/processors/SetProcessor.java bc9254c 
   service/src/java/org/apache/hive/service/cli/operation/SQLOperation.java 
 33ee16b 
 
 Diff: https://reviews.apache.org/r/35107/diff/
 
 
 Testing
 ---
 
 Unit test passed
 
 
 Thanks,
 
 cheng xu
 




[jira] [Created] (HIVE-11100) Beeline should escape semi-colon in queries

2015-06-24 Thread Chaoyu Tang (JIRA)
Chaoyu Tang created HIVE-11100:
--

 Summary: Beeline should escape semi-colon in queries
 Key: HIVE-11100
 URL: https://issues.apache.org/jira/browse/HIVE-11100
 Project: Hive
  Issue Type: Improvement
  Components: Beeline
Affects Versions: 1.2.0
Reporter: Chaoyu Tang
Assignee: Chaoyu Tang
Priority: Minor


Beeline should escape the semicolon in queries. for example, the query like 
followings:
CREATE TABLE beeline_tb (c1 int, c2 string) ROW FORMAT DELIMITED FIELDS 
TERMINATED BY ';' LINES TERMINATED BY '\n';
or 
CREATE TABLE beeline_tb (c1 int, c2 string) ROW FORMAT DELIMITED FIELDS 
TERMINATED BY '\;' LINES TERMINATED BY '\n';
both failed.
But the 2nd query with semicolon escaped with \ works in CLI.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


Review Request 35846: HIVE-11099: Add support for running negative q-tests [Spark Branch]

2015-06-24 Thread Mohit Sabharwal

---
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/35846/
---

Review request for hive.


Bugs: HIVE-11099
https://issues.apache.org/jira/browse/HIVE-11099


Repository: hive-git


Description
---

HIVE-11099: Add support for running negative q-tests [Spark Branch]

Add support for TestSparkNegativeCliDriver and 
TestMiniSparkOnYarnNegativeCliDriver for negative q-tests

Adding empty spark.query.negative.files and 
miniSparkOnYarn.query.negative.files lists in 
testconfiguration.properties. These lists can be populated as 
we enable negative q-tests. HIVE-11032 is a follow-up
patch which will enable some of these tests


Diffs
-

  itests/qtest-spark/pom.xml dcb76230584857294e0d7346da8e7353084ea3ae 
  itests/src/test/resources/testconfiguration.properties 
7b7559a9590803f2528a24180a962013881cba1b 

Diff: https://reviews.apache.org/r/35846/diff/


Testing
---

Tested with a negative q-test:
mvn test -Dmodule=ql -Dtest=TestSparkNegativeCliDriver 
-Dqfile=groupby2_map_skew_multi_distinct.q -Phadoop-2


Thanks,

Mohit Sabharwal



[jira] [Created] (HIVE-11095) SerDeUtils another bug ,when Text is reused

2015-06-24 Thread xiaowei wang (JIRA)
xiaowei wang created HIVE-11095:
---

 Summary: SerDeUtils  another bug ,when Text is reused
 Key: HIVE-11095
 URL: https://issues.apache.org/jira/browse/HIVE-11095
 Project: Hive
  Issue Type: Bug
  Components: API, CLI
Affects Versions: 1.2.0, 1.0.0, 0.14.0
 Environment: Hadoop 2.3.0-cdh5.0.0
Hive 0.14
Reporter: xiaowei wang
Assignee: xiaowei wang
Priority: Critical
 Fix For: 1.2.0


the method transformTextFromUTF8 have a bug, 
When i query data from a lzo table , I found in results : the length of the 
current row is always largr than the previous row, and sometimes,the current 
row contains the contents of the previous row。 For example ,i execute a sql 
,select * from web_searchhub where logdate=2015061003, the result of sql see 
blow.Notice that ,the second row content contains the first row content.
INFO [03:00:05.589] HttpFrontServer::FrontSH 
msgRecv:Remote=/10.13.193.68:42098,session=3151,thread=254 2015061003
INFO [03:00:05.594] 18941e66-9962-44ad-81bc-3519f47ba274 
session=901,thread=223ession=3151,thread=254 2015061003
The content of origin lzo file content see below ,just 2 rows.
INFO [03:00:05.635] b88e0473-7530-494c-82d8-e2d2ebd2666c_forweb 
session=3148,thread=285
INFO [03:00:05.635] HttpFrontServer::FrontSH 
msgRecv:Remote=/10.13.193.68:42095,session=3148,thread=285
I think this error is caused by the Text reuse,and I found the solutions .
Addicational, table create sql is : 
CREATE EXTERNAL TABLE `web_searchhub`(
`line` string)
PARTITIONED BY (
`logdate` string)
ROW FORMAT DELIMITED
FIELDS TERMINATED BY '
U'
WITH SERDEPROPERTIES (
'serialization.encoding'='GBK')
STORED AS INPUTFORMAT com.hadoop.mapred.DeprecatedLzoTextInputFormat
OUTPUTFORMAT org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat;
LOCATION
'viewfs://nsX/user/hive/warehouse/raw.db/web/web_searchhub' ;



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


Re: Review Request 34757: HIVE-10844: Combine equivalent Works for HoS[Spark Branch]

2015-06-24 Thread Xuefu Zhang


 On June 23, 2015, 1:31 p.m., Xuefu Zhang wrote:
  ql/src/test/results/clientpositive/spark/groupby10.q.out, line 60
  https://reviews.apache.org/r/34757/diff/3-4/?file=988071#file988071line60
 
  Interesting. How come we got more stages now?
 
 chengxiang li wrote:
 Not sure, introduced by latest merge from trunk.

Okay. Got it. Thanks.


- Xuefu


---
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/34757/#review88966
---


On June 23, 2015, 7:24 a.m., chengxiang li wrote:
 
 ---
 This is an automatically generated e-mail. To reply, visit:
 https://reviews.apache.org/r/34757/
 ---
 
 (Updated June 23, 2015, 7:24 a.m.)
 
 
 Review request for hive and Xuefu Zhang.
 
 
 Bugs: HIVE-10844
 https://issues.apache.org/jira/browse/HIVE-10844
 
 
 Repository: hive-git
 
 
 Description
 ---
 
 Some Hive queries(like TPCDS Q39) may share the same subquery, which 
 translated into sperate, but equivalent Works in SparkWork, combining these 
 equivalent Works into a single one would help to benifit from following 
 dynamic RDD caching optimization.
 
 
 Diffs
 -
 
   
 ql/src/java/org/apache/hadoop/hive/ql/optimizer/OperatorComparatorFactory.java
  PRE-CREATION 
   
 ql/src/java/org/apache/hadoop/hive/ql/optimizer/spark/CombineEquivalentWorkResolver.java
  PRE-CREATION 
   ql/src/java/org/apache/hadoop/hive/ql/parse/spark/SparkCompiler.java 
 19aae70 
   ql/src/java/org/apache/hadoop/hive/ql/plan/JoinCondDesc.java b307b16 
   ql/src/test/results/clientpositive/spark/auto_join30.q.out 7b5c5e7 
   ql/src/test/results/clientpositive/spark/auto_smb_mapjoin_14.q.out 8a43d78 
   ql/src/test/results/clientpositive/spark/groupby10.q.out dd9d9fe 
   ql/src/test/results/clientpositive/spark/groupby7_map.q.out abd6459 
   ql/src/test/results/clientpositive/spark/groupby7_map_skew.q.out 5e69b31 
   ql/src/test/results/clientpositive/spark/groupby7_noskew.q.out 3418b99 
   
 ql/src/test/results/clientpositive/spark/groupby7_noskew_multi_single_reducer.q.out
  2cb126d 
   ql/src/test/results/clientpositive/spark/groupby8.q.out c249b61 
   ql/src/test/results/clientpositive/spark/groupby8_map_skew.q.out 2fb1d73 
   ql/src/test/results/clientpositive/spark/insert_into3.q.out 7df5ba8 
   ql/src/test/results/clientpositive/spark/join22.q.out b1e5b67 
   ql/src/test/results/clientpositive/spark/skewjoinopt11.q.out 8a278ef 
   ql/src/test/results/clientpositive/spark/union10.q.out 5e8fe38 
   ql/src/test/results/clientpositive/spark/union11.q.out 20c27c7 
   ql/src/test/results/clientpositive/spark/union20.q.out 6f0dca6 
   ql/src/test/results/clientpositive/spark/union28.q.out 98582df 
   ql/src/test/results/clientpositive/spark/union3.q.out 834b6d4 
   ql/src/test/results/clientpositive/spark/union30.q.out 3409623 
   ql/src/test/results/clientpositive/spark/union4.q.out c121ef0 
   ql/src/test/results/clientpositive/spark/union5.q.out afee988 
   ql/src/test/results/clientpositive/spark/union_remove_1.q.out ba0e293 
   ql/src/test/results/clientpositive/spark/union_remove_15.q.out 26cfbab 
   ql/src/test/results/clientpositive/spark/union_remove_16.q.out 7a7aaf2 
   ql/src/test/results/clientpositive/spark/union_remove_18.q.out a5e15c5 
   ql/src/test/results/clientpositive/spark/union_remove_19.q.out ad44400 
   ql/src/test/results/clientpositive/spark/union_remove_20.q.out 1d67177 
   ql/src/test/results/clientpositive/spark/union_remove_21.q.out 9f5b070 
   ql/src/test/results/clientpositive/spark/union_remove_22.q.out 2e01432 
   ql/src/test/results/clientpositive/spark/union_remove_24.q.out 2659798 
   ql/src/test/results/clientpositive/spark/union_remove_25.q.out 0a94684 
   ql/src/test/results/clientpositive/spark/union_remove_4.q.out 6c3d596 
   ql/src/test/results/clientpositive/spark/union_remove_6.q.out cd36189 
   ql/src/test/results/clientpositive/spark/union_remove_6_subq.q.out c981ae4 
   ql/src/test/results/clientpositive/spark/union_remove_7.q.out 084fbd6 
   ql/src/test/results/clientpositive/spark/union_top_level.q.out dede1ef 
 
 Diff: https://reviews.apache.org/r/34757/diff/
 
 
 Testing
 ---
 
 
 Thanks,
 
 chengxiang li
 




Re: Review Request 35792: HIVE-10438 - Architecture for ResultSet Compression via external plugin

2015-06-24 Thread Xuefu Zhang

---
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/35792/#review89285
---



common/src/java/org/apache/hadoop/hive/conf/HiveConf.java (line 2202)
https://reviews.apache.org/r/35792/#comment141873

1. Remove leading spaces or tabs.
2. The default should be false



conf/hive-site.xml (line 22)
https://reviews.apache.org/r/35792/#comment141875

We probably don't want to add anything here, as these properties have 
default values which are sufficient.



service/src/java/org/apache/hive/service/cli/ColumnCompressor.java (line 15)
https://reviews.apache.org/r/35792/#comment141880

Documentation for these public APIs?



service/src/java/org/apache/hive/service/cli/ColumnCompressorService.java (line 
11)
https://reviews.apache.org/r/35792/#comment141878

1. Why hashtable? We should use HashMap or ConcurrentHashMap rather than 
HashTable.
2.  will break people building Hive with Java 6.



service/src/java/org/apache/hive/service/cli/ColumnCompressorService.java (line 
19)
https://reviews.apache.org/r/35792/#comment141876

Remove trailing spaces. Same for other occurrences.


First round of review comments. Most cosmatic. More will come.

- Xuefu Zhang


On June 23, 2015, 8:39 p.m., Rohit Dholakia wrote:
 
 ---
 This is an automatically generated e-mail. To reply, visit:
 https://reviews.apache.org/r/35792/
 ---
 
 (Updated June 23, 2015, 8:39 p.m.)
 
 
 Review request for hive, Vaibhav Gumashta and Xuefu Zhang.
 
 
 Repository: hive-git
 
 
 Description
 ---
 
 This patch enables ResultSet compression for Hive using external plugins. The 
 patch proposes a plugin architecture that enables using external plugins to 
 compress ResultSets on-the-fly.
 
 
 Diffs
 -
 
   common/src/java/org/apache/hadoop/hive/conf/HiveConf.java 27f68df 
   conf/hive-site.xml dab494e 
   service/if/TCLIService.thrift baf583f 
   service/src/gen/thrift/gen-cpp/TCLIService_types.h 7bceabd 
   service/src/gen/thrift/gen-cpp/TCLIService_types.cpp 86eeea3 
   
 service/src/gen/thrift/gen-javabean/org/apache/hive/service/cli/thrift/TEnColumn.java
  PRE-CREATION 
   
 service/src/gen/thrift/gen-javabean/org/apache/hive/service/cli/thrift/TExecuteStatementReq.java
  4f157ad 
   
 service/src/gen/thrift/gen-javabean/org/apache/hive/service/cli/thrift/TGetTablesReq.java
  c973fcc 
   
 service/src/gen/thrift/gen-javabean/org/apache/hive/service/cli/thrift/TOpenSessionReq.java
  c048161 
   
 service/src/gen/thrift/gen-javabean/org/apache/hive/service/cli/thrift/TOpenSessionResp.java
  351f78b 
   
 service/src/gen/thrift/gen-javabean/org/apache/hive/service/cli/thrift/TProtocolVersion.java
  a4279d2 
   
 service/src/gen/thrift/gen-javabean/org/apache/hive/service/cli/thrift/TRowSet.java
  d16c8a4 
   
 service/src/gen/thrift/gen-javabean/org/apache/hive/service/cli/thrift/TStatus.java
  24a746e 
   service/src/gen/thrift/gen-py/TCLIService/ttypes.py 068727c 
   service/src/gen/thrift/gen-rb/t_c_l_i_service_types.rb b482533 
   service/src/java/org/apache/hive/service/cli/Column.java 2e21f18 
   service/src/java/org/apache/hive/service/cli/ColumnBasedSet.java 47a582e 
   service/src/java/org/apache/hive/service/cli/ColumnCompressor.java 
 PRE-CREATION 
   service/src/java/org/apache/hive/service/cli/ColumnCompressorService.java 
 PRE-CREATION 
   service/src/java/org/apache/hive/service/cli/EncodedColumnBasedSet.java 
 PRE-CREATION 
   service/src/java/org/apache/hive/service/cli/RowSetFactory.java e8f68ea 
   service/src/java/org/apache/hive/service/cli/thrift/ThriftCLIService.java 
 dfb7faa 
 
 Diff: https://reviews.apache.org/r/35792/diff/
 
 
 Testing
 ---
 
 Testing has been done using a docker container-based query submitter that has 
 an integer decompressor as part of it. Using the integer compressor (also 
 provided) and the decompressor, the end-to-end functionality can be observed.
 
 
 File Attachments
 
 
 Patch file
   
 https://reviews.apache.org/media/uploaded/files/2015/06/23/16aa08f8-2393-460a-83ef-72464fc537db__HIVE-10438.patch
 
 
 Thanks,
 
 Rohit Dholakia
 




[jira] [Created] (HIVE-11103) Add banker's rounding BROUND UDF

2015-06-24 Thread Alexander Pivovarov (JIRA)
Alexander Pivovarov created HIVE-11103:
--

 Summary: Add banker's rounding BROUND UDF
 Key: HIVE-11103
 URL: https://issues.apache.org/jira/browse/HIVE-11103
 Project: Hive
  Issue Type: New Feature
  Components: UDF
Reporter: Alexander Pivovarov
Assignee: Alexander Pivovarov


Banker's rounding: the value is rounded to the nearest even number. Also known 
as Gaussian rounding, and, in German, mathematische Rundung.

Example
{code}
  2 digits2 digits
UnroundedStandard roundingGaussian rounding
  54.1754  54.18  54.18
 343.2050 343.21 343.20
+106.2038+106.20+106.20 
=======
 503.5842 503.59 503.58
{code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HIVE-11104) Select operator doesn't propagate constants appearing in expressions

2015-06-24 Thread Ashutosh Chauhan (JIRA)
Ashutosh Chauhan created HIVE-11104:
---

 Summary: Select operator doesn't propagate constants appearing in 
expressions
 Key: HIVE-11104
 URL: https://issues.apache.org/jira/browse/HIVE-11104
 Project: Hive
  Issue Type: Bug
  Components: Logical Optimizer
Affects Versions: 1.2.0, 1.0.0, 0.14.0, 1.1.0, 1.2.1
Reporter: Ashutosh Chauhan
Assignee: Ashutosh Chauhan






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


Re: Review Request 34757: HIVE-10844: Combine equivalent Works for HoS[Spark Branch]

2015-06-24 Thread Xuefu Zhang

---
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/34757/#review89155
---

Ship it!


- Xuefu Zhang


On June 23, 2015, 7:24 a.m., chengxiang li wrote:
 
 ---
 This is an automatically generated e-mail. To reply, visit:
 https://reviews.apache.org/r/34757/
 ---
 
 (Updated June 23, 2015, 7:24 a.m.)
 
 
 Review request for hive and Xuefu Zhang.
 
 
 Bugs: HIVE-10844
 https://issues.apache.org/jira/browse/HIVE-10844
 
 
 Repository: hive-git
 
 
 Description
 ---
 
 Some Hive queries(like TPCDS Q39) may share the same subquery, which 
 translated into sperate, but equivalent Works in SparkWork, combining these 
 equivalent Works into a single one would help to benifit from following 
 dynamic RDD caching optimization.
 
 
 Diffs
 -
 
   
 ql/src/java/org/apache/hadoop/hive/ql/optimizer/OperatorComparatorFactory.java
  PRE-CREATION 
   
 ql/src/java/org/apache/hadoop/hive/ql/optimizer/spark/CombineEquivalentWorkResolver.java
  PRE-CREATION 
   ql/src/java/org/apache/hadoop/hive/ql/parse/spark/SparkCompiler.java 
 19aae70 
   ql/src/java/org/apache/hadoop/hive/ql/plan/JoinCondDesc.java b307b16 
   ql/src/test/results/clientpositive/spark/auto_join30.q.out 7b5c5e7 
   ql/src/test/results/clientpositive/spark/auto_smb_mapjoin_14.q.out 8a43d78 
   ql/src/test/results/clientpositive/spark/groupby10.q.out dd9d9fe 
   ql/src/test/results/clientpositive/spark/groupby7_map.q.out abd6459 
   ql/src/test/results/clientpositive/spark/groupby7_map_skew.q.out 5e69b31 
   ql/src/test/results/clientpositive/spark/groupby7_noskew.q.out 3418b99 
   
 ql/src/test/results/clientpositive/spark/groupby7_noskew_multi_single_reducer.q.out
  2cb126d 
   ql/src/test/results/clientpositive/spark/groupby8.q.out c249b61 
   ql/src/test/results/clientpositive/spark/groupby8_map_skew.q.out 2fb1d73 
   ql/src/test/results/clientpositive/spark/insert_into3.q.out 7df5ba8 
   ql/src/test/results/clientpositive/spark/join22.q.out b1e5b67 
   ql/src/test/results/clientpositive/spark/skewjoinopt11.q.out 8a278ef 
   ql/src/test/results/clientpositive/spark/union10.q.out 5e8fe38 
   ql/src/test/results/clientpositive/spark/union11.q.out 20c27c7 
   ql/src/test/results/clientpositive/spark/union20.q.out 6f0dca6 
   ql/src/test/results/clientpositive/spark/union28.q.out 98582df 
   ql/src/test/results/clientpositive/spark/union3.q.out 834b6d4 
   ql/src/test/results/clientpositive/spark/union30.q.out 3409623 
   ql/src/test/results/clientpositive/spark/union4.q.out c121ef0 
   ql/src/test/results/clientpositive/spark/union5.q.out afee988 
   ql/src/test/results/clientpositive/spark/union_remove_1.q.out ba0e293 
   ql/src/test/results/clientpositive/spark/union_remove_15.q.out 26cfbab 
   ql/src/test/results/clientpositive/spark/union_remove_16.q.out 7a7aaf2 
   ql/src/test/results/clientpositive/spark/union_remove_18.q.out a5e15c5 
   ql/src/test/results/clientpositive/spark/union_remove_19.q.out ad44400 
   ql/src/test/results/clientpositive/spark/union_remove_20.q.out 1d67177 
   ql/src/test/results/clientpositive/spark/union_remove_21.q.out 9f5b070 
   ql/src/test/results/clientpositive/spark/union_remove_22.q.out 2e01432 
   ql/src/test/results/clientpositive/spark/union_remove_24.q.out 2659798 
   ql/src/test/results/clientpositive/spark/union_remove_25.q.out 0a94684 
   ql/src/test/results/clientpositive/spark/union_remove_4.q.out 6c3d596 
   ql/src/test/results/clientpositive/spark/union_remove_6.q.out cd36189 
   ql/src/test/results/clientpositive/spark/union_remove_6_subq.q.out c981ae4 
   ql/src/test/results/clientpositive/spark/union_remove_7.q.out 084fbd6 
   ql/src/test/results/clientpositive/spark/union_top_level.q.out dede1ef 
 
 Diff: https://reviews.apache.org/r/34757/diff/
 
 
 Testing
 ---
 
 
 Thanks,
 
 chengxiang li
 




Re: Review Request 35107: HIVE-6791 Support variable substition for Beeline shell command

2015-06-24 Thread cheng xu

---
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/35107/
---

(Updated June 25, 2015, 1:54 p.m.)


Review request for hive, chinna and Xuefu Zhang.


Changes
---

code clean


Bugs: HIVE-6791
https://issues.apache.org/jira/browse/HIVE-6791


Repository: hive-git


Description
---

Summary:
1) move the beeline-cli convertor to the place where cli is executed(class 
**Commands**)
2) support substitution for source command
3) add some unit test for substitution
4) add one way to get the configuration from HS2


Diffs (updated)
-

  beeline/src/java/org/apache/hive/beeline/BeeLine.java b7d2f2e 
  beeline/src/java/org/apache/hive/beeline/Commands.java a42baa3 
  beeline/src/test/org/apache/hive/beeline/cli/TestHiveCli.java 6cbb030 
  cli/src/java/org/apache/hadoop/hive/cli/CliDriver.java d62fd5c 
  common/src/java/org/apache/hadoop/hive/conf/HiveVariableSource.java 
PRE-CREATION 
  common/src/java/org/apache/hadoop/hive/conf/VariableSubstitution.java 
PRE-CREATION 
  common/src/test/org/apache/hadoop/hive/conf/TestVariableSubstitution.java 
PRE-CREATION 
  ql/src/java/org/apache/hadoop/hive/ql/Driver.java 338e755 
  ql/src/java/org/apache/hadoop/hive/ql/parse/ColumnStatsSemanticAnalyzer.java 
a5f0a7f 
  ql/src/java/org/apache/hadoop/hive/ql/parse/VariableSubstitution.java e8b1d96 
  ql/src/java/org/apache/hadoop/hive/ql/processors/AddResourceProcessor.java 
0558c53 
  ql/src/java/org/apache/hadoop/hive/ql/processors/CompileProcessor.java 
25ce168 
  ql/src/java/org/apache/hadoop/hive/ql/processors/DeleteResourceProcessor.java 
9052c82 
  ql/src/java/org/apache/hadoop/hive/ql/processors/DfsProcessor.java cc0414d 
  ql/src/java/org/apache/hadoop/hive/ql/processors/SetProcessor.java bc9254c 
  service/src/java/org/apache/hive/service/cli/operation/SQLOperation.java 
33ee16b 

Diff: https://reviews.apache.org/r/35107/diff/


Testing
---

Unit test passed


Thanks,

cheng xu



Re: Review Request 35107: HIVE-6791 Support variable substition for Beeline shell command

2015-06-24 Thread cheng xu


 On June 25, 2015, 6:21 a.m., Xuefu Zhang wrote:
  ql/src/java/org/apache/hadoop/hive/ql/processors/SetProcessor.java, line 146
  https://reviews.apache.org/r/35107/diff/3/?file=991115#file991115line146
 
  Could we keep @Override at a separate line? Same for other places.

I am using the code-style file(eclipse-styles.xml) under the dev-support 
folder. Seems annotation before class/method is not wrapped. Anyway, I just 
update all the places in this patch. Thank you for figuring this out.


 On June 25, 2015, 6:21 a.m., Xuefu Zhang wrote:
  beeline/src/java/org/apache/hive/beeline/Commands.java, line 820
  https://reviews.apache.org/r/35107/diff/3/?file=991102#file991102line820
 
  Does this mean that env and sys variables are not being substituted for 
  shell command?

No, this method is only used for retrieving hive configurations. For env and 
sys variables, they are subsituted by VariableSubstitution.


- cheng


---
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/35107/#review89251
---


On June 25, 2015, 1:54 p.m., cheng xu wrote:
 
 ---
 This is an automatically generated e-mail. To reply, visit:
 https://reviews.apache.org/r/35107/
 ---
 
 (Updated June 25, 2015, 1:54 p.m.)
 
 
 Review request for hive, chinna and Xuefu Zhang.
 
 
 Bugs: HIVE-6791
 https://issues.apache.org/jira/browse/HIVE-6791
 
 
 Repository: hive-git
 
 
 Description
 ---
 
 Summary:
 1) move the beeline-cli convertor to the place where cli is executed(class 
 **Commands**)
 2) support substitution for source command
 3) add some unit test for substitution
 4) add one way to get the configuration from HS2
 
 
 Diffs
 -
 
   beeline/src/java/org/apache/hive/beeline/BeeLine.java b7d2f2e 
   beeline/src/java/org/apache/hive/beeline/Commands.java a42baa3 
   beeline/src/test/org/apache/hive/beeline/cli/TestHiveCli.java 6cbb030 
   cli/src/java/org/apache/hadoop/hive/cli/CliDriver.java d62fd5c 
   common/src/java/org/apache/hadoop/hive/conf/HiveVariableSource.java 
 PRE-CREATION 
   common/src/java/org/apache/hadoop/hive/conf/VariableSubstitution.java 
 PRE-CREATION 
   common/src/test/org/apache/hadoop/hive/conf/TestVariableSubstitution.java 
 PRE-CREATION 
   ql/src/java/org/apache/hadoop/hive/ql/Driver.java 338e755 
   
 ql/src/java/org/apache/hadoop/hive/ql/parse/ColumnStatsSemanticAnalyzer.java 
 a5f0a7f 
   ql/src/java/org/apache/hadoop/hive/ql/parse/VariableSubstitution.java 
 e8b1d96 
   ql/src/java/org/apache/hadoop/hive/ql/processors/AddResourceProcessor.java 
 0558c53 
   ql/src/java/org/apache/hadoop/hive/ql/processors/CompileProcessor.java 
 25ce168 
   
 ql/src/java/org/apache/hadoop/hive/ql/processors/DeleteResourceProcessor.java 
 9052c82 
   ql/src/java/org/apache/hadoop/hive/ql/processors/DfsProcessor.java cc0414d 
   ql/src/java/org/apache/hadoop/hive/ql/processors/SetProcessor.java bc9254c 
   service/src/java/org/apache/hive/service/cli/operation/SQLOperation.java 
 33ee16b 
 
 Diff: https://reviews.apache.org/r/35107/diff/
 
 
 Testing
 ---
 
 Unit test passed
 
 
 Thanks,
 
 cheng xu
 




[jira] [Created] (HIVE-11096) Bump the parquet version to 1.7.0

2015-06-24 Thread JIRA
Sergio Peña created HIVE-11096:
--

 Summary: Bump the parquet version to 1.7.0
 Key: HIVE-11096
 URL: https://issues.apache.org/jira/browse/HIVE-11096
 Project: Hive
  Issue Type: Task
Reporter: Sergio Peña
Assignee: Ferdinand Xu
Priority: Minor


Parquet has moved officially as an Apache project since parquet 1.7.0.

This new version does not have any bugfixes nor improvements from its last 
1.6.0 version, but all imports were changed to be org.apache.parquet, and the 
pom.xml must use org.apache.parquet instead of com.twitter.

This ticket should address those import and pom changes only.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


Re: Review Request 35803: HIVE-10895 ObjectStore does not close Query objects in some calls, causing a potential leak in some metastore db resources

2015-06-24 Thread Aihua Xu

---
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/35803/
---

(Updated June 24, 2015, 3:13 p.m.)


Review request for hive.


Repository: hive-git


Description
---

HIVE-10895 ObjectStore does not close Query objects in some calls, causing a 
potential leak in some metastore db resources


Diffs (updated)
-

  metastore/src/java/org/apache/hadoop/hive/metastore/ObjectStore.java 417ecc8 
  metastore/src/test/org/apache/hadoop/hive/metastore/TestObjectStore.java 
PRE-CREATION 

Diff: https://reviews.apache.org/r/35803/diff/


Testing (updated)
---

Testing has been done.


Thanks,

Aihua Xu



[jira] [Created] (HIVE-11097) HiveInputFormat uses String.startsWith to compare splitPath and PathToAliases

2015-06-24 Thread Wan Chang (JIRA)
Wan Chang created HIVE-11097:


 Summary: HiveInputFormat uses String.startsWith to compare 
splitPath and PathToAliases
 Key: HIVE-11097
 URL: https://issues.apache.org/jira/browse/HIVE-11097
 Project: Hive
  Issue Type: Bug
  Components: File Formats
Affects Versions: 1.2.0, 1.0.0, 0.13.1, 0.14.0, 0.13.0
 Environment: Hive 0.13.1, Hive 2.0.0, hadoop 2.4.1
Reporter: Wan Chang
Priority: Critical


Say we have a sql as
{code}
create table if not exists test_orc_src (a int, b int, c int) stored as orc;
create table if not exists test_orc_src2 (a int, b int, d int) stored as orc;
insert overwrite table test_orc_src select 1,2,3 from src limit 1;
insert overwrite table test_orc_src2 select 1,2,4 from src limit 1;
set hive.auto.convert.join = false;
set hive.execution.engine=mr;
select
  tb.c
from test.test_orc_src tb
join (select * from test.test_orc_src2) tm
on tb.a = tm.a
where tb.b = 2
{code}
The correct result is 3 but it produced no result.

I find that in HiveInputFormat.pushProjectionsAndFilters
{code}
match = splitPath.startsWith(key) || splitPathWithNoSchema.startsWith(key);
{code}
It uses startsWith to combine aliases with path, so tm will match two alias in 
this case.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


Re: Review Request 34726: HIVE-10533

2015-06-24 Thread Jesús Camacho Rodríguez

---
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/34726/
---

(Updated June 24, 2015, 5:57 p.m.)


Review request for hive, Ashutosh Chauhan and John Pullokkaran.


Bugs: HIVE-10533
https://issues.apache.org/jira/browse/HIVE-10533


Repository: hive-git


Description
---

CBO (Calcite Return Path): Join to MultiJoin support for outer joins


Diffs (updated)
-

  ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/HiveCalciteUtil.java 
199a358ba16e857f3b4bf9b0c07d1919eeb709d4 
  ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/HiveRelOptUtil.java 
PRE-CREATION 
  
ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/reloperators/HiveMultiJoin.java
 PRE-CREATION 
  
ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/rules/HiveInsertExchange4JoinRule.java
 11c3d23ea7439b713b66a09a9911e47e847f2b18 
  
ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/rules/HiveJoinProjectTransposeRule.java
 PRE-CREATION 
  
ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/rules/HiveJoinToMultiJoinRule.java
 532d7d3b56377946f6a9ad883d7b7dbf1325a8c7 
  
ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/rules/HiveProjectMergeRule.java
 8b90a15b5517a0b83c22b7a9bbbd90398c45184a 
  
ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/rules/HiveRelFieldTrimmer.java
 3d1a3091bd0d9a99366da565f9c136aff0983bb7 
  
ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/translator/HiveOpConverter.java
 a75d029ccf5b966347e5bf4b8c38695e89f80171 
  ql/src/java/org/apache/hadoop/hive/ql/parse/CalcitePlanner.java 
e821b1de5af947768dca916627911ed5d3433a7f 
  ql/src/java/org/apache/hadoop/hive/ql/plan/JoinDesc.java 
37012b49873495b60ea3f6cceadece1941b06b04 
  ql/src/test/queries/clientpositive/cbo_rp_join0.q PRE-CREATION 
  ql/src/test/queries/clientpositive/cbo_rp_join1.q PRE-CREATION 
  ql/src/test/results/clientpositive/cbo_rp_join0.q.out PRE-CREATION 
  ql/src/test/results/clientpositive/cbo_rp_join1.q.out PRE-CREATION 

Diff: https://reviews.apache.org/r/34726/diff/


Testing
---


Thanks,

Jesús Camacho Rodríguez