[jira] [Updated] (HIVE-7653) Hive AvroSerDe does not support circular references in Schema

2015-02-19 Thread Szehon Ho (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-7653?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Szehon Ho updated HIVE-7653:

Resolution: Fixed
Status: Resolved  (was: Patch Available)

Yea the test failures dont look related.  Committed to trunk, thanks for the 
contribution.

 Hive AvroSerDe does not support circular references in Schema
 -

 Key: HIVE-7653
 URL: https://issues.apache.org/jira/browse/HIVE-7653
 Project: Hive
  Issue Type: Bug
Affects Versions: 0.13.1
Reporter: Sachin Goyal
Assignee: Sachin Goyal
 Attachments: HIVE-7653.1.patch, HIVE-7653.2.patch, HIVE-7653.3.patch, 
 HIVE-7653.4.patch, HIVE-7653.5.patch, HIVE-7653.6.patch, HIVE-7653.6.patch


 Avro allows nullable circular references but Hive AvroSerDe does not.
 Example of circular references (passing in Avro but failing in AvroSerDe):
 {code}
 class AvroCycleParent {
   AvroCycleChild child;
   public AvroCycleChild getChild () {return child;}
   public void setChild (AvroCycleChild child) {this.child = child;}
 }
 class AvroCycleChild {
   AvroCycleParent parent;
   public AvroCycleParent getParent () {return parent;}
   public void setParent (AvroCycleParent parent) {this.parent = parent;}
 }
 {code}
 Due to this discrepancy, Hive is unable to read Avro records having 
 circular-references. For some third-party code with such references, it 
 becomes very hard to directly serialize it with Avro and use in Hive.
 I have a patch for this with a unit-test and I will submit it shortly.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-3454) Problem with CAST(BIGINT as TIMESTAMP)

2015-02-19 Thread Jason Dere (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-3454?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14327096#comment-14327096
 ] 

Jason Dere commented on HIVE-3454:
--

If this config setting is initialized once in a static block, then for 
hiveserver2 all subsequent sessions would be stuck with the initial setting 
regardless of the config settings of the session, right? During the MR jobs, 
would we then see the sec/msec behavior flip to use the session's config 
settings since the static variable is being initialized for the first time in 
MR task?

 Problem with CAST(BIGINT as TIMESTAMP)
 --

 Key: HIVE-3454
 URL: https://issues.apache.org/jira/browse/HIVE-3454
 Project: Hive
  Issue Type: Bug
  Components: Types, UDF
Affects Versions: 0.8.0, 0.8.1, 0.9.0, 0.10.0, 0.11.0, 0.12.0, 0.13.0, 
 0.13.1
Reporter: Ryan Harris
Assignee: Aihua Xu
  Labels: newbie, newdev, patch
 Attachments: HIVE-3454.1.patch.txt, HIVE-3454.2.patch, 
 HIVE-3454.3.patch, HIVE-3454.patch


 Ran into an issue while working with timestamp conversion.
 CAST(unix_timestamp() as TIMESTAMP) should create a timestamp for the current 
 time from the BIGINT returned by unix_timestamp()
 Instead, however, a 1970-01-16 timestamp is returned.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Resolved] (HIVE-9613) Left join query plan outputs wrong column when using subquery

2015-02-19 Thread Li Xin (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9613?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Li Xin resolved HIVE-9613.
--
   Resolution: Fixed
Fix Version/s: 1.2.0

seems an issue in ColumnPruner and  fixed in HIVE-9327 

 Left join query plan outputs  wrong column when using subquery
 --

 Key: HIVE-9613
 URL: https://issues.apache.org/jira/browse/HIVE-9613
 Project: Hive
  Issue Type: Bug
  Components: Parser, Query Planning
Affects Versions: 0.14.0, 1.0.0
 Environment: apache hadoop 2.5.1 
Reporter: Li Xin
 Fix For: 1.2.0

 Attachments: test.sql


 I have a query that outputs a column with wrong contents when using 
 subquery,and the contents of that column is equal to another column,not its 
 own.
 I have three tables,as follows:
 table 1: _hivetemp.category_city_rank_:
 ||category||city||rank||
 |jinrongfuwu|shanghai|1|
 |ktvjiuba|shanghai|2|
 table 2:_hivetemp.category_match_:
 ||src_category_en||src_category_cn||dst_category_en||dst_category_cn||
 |danbaobaoxiantouzi|投资担保|担保/贷款|jinrongfuwu|
 |zpwentiyingshi|娱乐/休闲|KTV/酒吧|ktvjiuba|
 table 3:_hivetemp.city_match_:
 ||src_city_name_en||dst_city_name_en||city_name_cn||
 |sh|shanghai|上海|
 And the query is :
 {code}
 select
 a.category,
 a.city,
 a.rank,
 b.src_category_en,
 c.src_city_name_en
 from
 hivetemp.category_city_rank a
 left outer join
 (select
 src_category_en,
 dst_category_en
 from
 hivetemp.category_match) b
 on  a.category = b.dst_category_en
 left outer join
 (select
 src_city_name_en,
 dst_city_name_en
 from
 hivetemp.city_match) c
 on  a.city = c.dst_city_name_en
 {code}
 which shoud output the results as follows,and i test it in hive 0.13:
 ||category||city||rank||src_category_en||src_city_name_en||
 |jinrongfuwu|shanghai|1|danbaobaoxiantouzi|sh|
 |ktvjiuba|shanghai|2|zpwentiyingshi|sh|
 but int hive0.14,the results in the column *src_category_en* is wrong,and is 
 just the *city* contents:
 ||category||city||rank||src_category_en||src_city_name_en||
 |jinrongfuwu|shanghai|1|shanghai|sh|
 |ktvjiuba|shanghai|2|shanghai|sh|
 Using explain to examine the execution plan,i can see the first subquery just 
 outputs one column of *dst_category_en*,and *src_category_en* is just missing.
 {quote}
b:category_match
   TableScan
 alias: category_match
 Statistics: Num rows: 131 Data size: 13149 Basic stats: COMPLETE 
 Column stats: NONE
 Select Operator
   expressions: dst_category_en (type: string)
   outputColumnNames: _col1
   Statistics: Num rows: 131 Data size: 13149 Basic stats: 
 COMPLETE Column stats: NONE
 {quote}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9613) Left join query plan outputs wrong column when using subquery

2015-02-19 Thread Li Xin (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9613?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14327144#comment-14327144
 ] 

Li Xin commented on HIVE-9613:
--

Hi,[~csun], I just tested the upstream trunk and the issue is gone.
Thank you for your time and next time i will do some test in the upstream trunk 
first.
I will close this one as fixed~

 Left join query plan outputs  wrong column when using subquery
 --

 Key: HIVE-9613
 URL: https://issues.apache.org/jira/browse/HIVE-9613
 Project: Hive
  Issue Type: Bug
  Components: Parser, Query Planning
Affects Versions: 0.14.0, 1.0.0
 Environment: apache hadoop 2.5.1 
Reporter: Li Xin
 Fix For: 1.2.0

 Attachments: test.sql


 I have a query that outputs a column with wrong contents when using 
 subquery,and the contents of that column is equal to another column,not its 
 own.
 I have three tables,as follows:
 table 1: _hivetemp.category_city_rank_:
 ||category||city||rank||
 |jinrongfuwu|shanghai|1|
 |ktvjiuba|shanghai|2|
 table 2:_hivetemp.category_match_:
 ||src_category_en||src_category_cn||dst_category_en||dst_category_cn||
 |danbaobaoxiantouzi|投资担保|担保/贷款|jinrongfuwu|
 |zpwentiyingshi|娱乐/休闲|KTV/酒吧|ktvjiuba|
 table 3:_hivetemp.city_match_:
 ||src_city_name_en||dst_city_name_en||city_name_cn||
 |sh|shanghai|上海|
 And the query is :
 {code}
 select
 a.category,
 a.city,
 a.rank,
 b.src_category_en,
 c.src_city_name_en
 from
 hivetemp.category_city_rank a
 left outer join
 (select
 src_category_en,
 dst_category_en
 from
 hivetemp.category_match) b
 on  a.category = b.dst_category_en
 left outer join
 (select
 src_city_name_en,
 dst_city_name_en
 from
 hivetemp.city_match) c
 on  a.city = c.dst_city_name_en
 {code}
 which shoud output the results as follows,and i test it in hive 0.13:
 ||category||city||rank||src_category_en||src_city_name_en||
 |jinrongfuwu|shanghai|1|danbaobaoxiantouzi|sh|
 |ktvjiuba|shanghai|2|zpwentiyingshi|sh|
 but int hive0.14,the results in the column *src_category_en* is wrong,and is 
 just the *city* contents:
 ||category||city||rank||src_category_en||src_city_name_en||
 |jinrongfuwu|shanghai|1|shanghai|sh|
 |ktvjiuba|shanghai|2|shanghai|sh|
 Using explain to examine the execution plan,i can see the first subquery just 
 outputs one column of *dst_category_en*,and *src_category_en* is just missing.
 {quote}
b:category_match
   TableScan
 alias: category_match
 Statistics: Num rows: 131 Data size: 13149 Basic stats: COMPLETE 
 Column stats: NONE
 Select Operator
   expressions: dst_category_en (type: string)
   outputColumnNames: _col1
   Statistics: Num rows: 131 Data size: 13149 Basic stats: 
 COMPLETE Column stats: NONE
 {quote}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9641) Fill out remaining partition functions in HBaseStore

2015-02-19 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9641?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14328003#comment-14328003
 ] 

Hive QA commented on HIVE-9641:
---



{color:red}Overall{color}: -1 no tests executed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12699721/HIVE-9641.patch

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/2823/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/2823/console
Test logs: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-2823/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Tests exited with: NonZeroExitCodeException
Command 'bash /data/hive-ptest/working/scratch/source-prep.sh' failed with exit 
status 1 and output '+ [[ -n /usr/java/jdk1.7.0_45-cloudera ]]
+ export JAVA_HOME=/usr/java/jdk1.7.0_45-cloudera
+ JAVA_HOME=/usr/java/jdk1.7.0_45-cloudera
+ export 
PATH=/usr/java/jdk1.7.0_45-cloudera/bin/:/usr/local/apache-maven-3.0.5/bin:/usr/java/jdk1.7.0_45-cloudera/bin:/usr/local/apache-ant-1.9.1/bin:/usr/local/bin:/bin:/usr/bin:/usr/local/sbin:/usr/sbin:/sbin:/home/hiveptest/bin
+ 
PATH=/usr/java/jdk1.7.0_45-cloudera/bin/:/usr/local/apache-maven-3.0.5/bin:/usr/java/jdk1.7.0_45-cloudera/bin:/usr/local/apache-ant-1.9.1/bin:/usr/local/bin:/bin:/usr/bin:/usr/local/sbin:/usr/sbin:/sbin:/home/hiveptest/bin
+ export 'ANT_OPTS=-Xmx1g -XX:MaxPermSize=256m '
+ ANT_OPTS='-Xmx1g -XX:MaxPermSize=256m '
+ export 'M2_OPTS=-Xmx1g -XX:MaxPermSize=256m -Dhttp.proxyHost=localhost 
-Dhttp.proxyPort=3128'
+ M2_OPTS='-Xmx1g -XX:MaxPermSize=256m -Dhttp.proxyHost=localhost 
-Dhttp.proxyPort=3128'
+ cd /data/hive-ptest/working/
+ tee /data/hive-ptest/logs/PreCommit-HIVE-TRUNK-Build-2823/source-prep.txt
+ [[ false == \t\r\u\e ]]
+ mkdir -p maven ivy
+ [[ svn = \s\v\n ]]
+ [[ -n '' ]]
+ [[ -d apache-svn-trunk-source ]]
+ [[ ! -d apache-svn-trunk-source/.svn ]]
+ [[ ! -d apache-svn-trunk-source ]]
+ cd apache-svn-trunk-source
+ svn revert -R .
++ awk '{print $2}'
++ egrep -v '^X|^Performing status on external'
++ svn status --no-ignore
+ rm -rf target datanucleus.log ant/target shims/target shims/0.20S/target 
shims/0.23/target shims/aggregator/target shims/common/target 
shims/scheduler/target packaging/target hbase-handler/target testutils/target 
jdbc/target metastore/target itests/target itests/thirdparty 
itests/hcatalog-unit/target itests/test-serde/target itests/qtest/target 
itests/hive-unit-hadoop2/target itests/hive-minikdc/target 
itests/hive-jmh/target itests/hive-unit/target itests/custom-serde/target 
itests/util/target itests/qtest-spark/target hcatalog/target 
hcatalog/core/target hcatalog/streaming/target 
hcatalog/server-extensions/target hcatalog/hcatalog-pig-adapter/target 
hcatalog/webhcat/svr/target hcatalog/webhcat/java-client/target 
accumulo-handler/target hwi/target common/target common/src/gen 
spark-client/target service/target contrib/target serde/target beeline/target 
odbc/target cli/target ql/dependency-reduced-pom.xml ql/target
+ svn update

Fetching external item into 'hcatalog/src/test/e2e/harness'
External at revision 1660985.

At revision 1660985.
+ patchCommandPath=/data/hive-ptest/working/scratch/smart-apply-patch.sh
+ patchFilePath=/data/hive-ptest/working/scratch/build.patch
+ [[ -f /data/hive-ptest/working/scratch/build.patch ]]
+ chmod +x /data/hive-ptest/working/scratch/smart-apply-patch.sh
+ /data/hive-ptest/working/scratch/smart-apply-patch.sh 
/data/hive-ptest/working/scratch/build.patch
The patch does not appear to apply with p0, p1, or p2
+ exit 1
'
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12699721 - PreCommit-HIVE-TRUNK-Build

 Fill out remaining partition functions in HBaseStore
 

 Key: HIVE-9641
 URL: https://issues.apache.org/jira/browse/HIVE-9641
 Project: Hive
  Issue Type: Sub-task
  Components: Metastore
Reporter: Alan Gates
Assignee: Alan Gates
 Attachments: HIVE-9641.patch


 A number of the listPartition and getPartition methods are not implemented.  
 The rest need to be implemented.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9641) Fill out remaining partition functions in HBaseStore

2015-02-19 Thread Alan Gates (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9641?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alan Gates updated HIVE-9641:
-
Status: Patch Available  (was: Open)

 Fill out remaining partition functions in HBaseStore
 

 Key: HIVE-9641
 URL: https://issues.apache.org/jira/browse/HIVE-9641
 Project: Hive
  Issue Type: Sub-task
  Components: Metastore
Reporter: Alan Gates
Assignee: Alan Gates
 Attachments: HIVE-9641.patch


 A number of the listPartition and getPartition methods are not implemented.  
 The rest need to be implemented.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


Re: Fix version for hbase-metastore branch

2015-02-19 Thread Ashutosh Chauhan
This is what we have been doing for cbo work. e.g.
https://issues.apache.org/jira/browse/HIVE-9581

On Thu, Feb 19, 2015 at 11:17 AM, Thejas Nair thejas.n...@gmail.com wrote:

 I agree, using a label for fix version makes sense in this case. I believe
 that is what had been done for hive-on-spark and hive-on-tez.


 On Thu, Feb 19, 2015 at 10:56 AM, Alan Gates alanfga...@gmail.com wrote:

  I've been marking JIRAs on this branch as fixed in 1.2, since that's the
  next version.  But that seems wrong as I doubt this code will be in by
  1.2.  What's the usual practice here?  It seems it would make sense to
 make
  a label for this branch and mark them as fixed with that label and then
  when we actually release this in a version we can update all the JIRAs
 with
  that label.
 
  Alan.
 



Re: Fix version for hbase-metastore branch

2015-02-19 Thread Thejas Nair
Looks like hive-on-tez and hive-on-spark didn't update the fix version
after merge to trunk. But I think updating the fix version after merge
makes sense.


On Thu, Feb 19, 2015 at 11:17 AM, Thejas Nair thejas.n...@gmail.com wrote:

 I agree, using a label for fix version makes sense in this case. I believe
 that is what had been done for hive-on-spark and hive-on-tez.


 On Thu, Feb 19, 2015 at 10:56 AM, Alan Gates alanfga...@gmail.com wrote:

 I've been marking JIRAs on this branch as fixed in 1.2, since that's the
 next version.  But that seems wrong as I doubt this code will be in by
 1.2.  What's the usual practice here?  It seems it would make sense to make
 a label for this branch and mark them as fixed with that label and then
 when we actually release this in a version we can update all the JIRAs with
 that label.

 Alan.





[jira] [Commented] (HIVE-7653) Hive AvroSerDe does not support circular references in Schema

2015-02-19 Thread Sachin Goyal (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-7653?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14328050#comment-14328050
 ] 

Sachin Goyal commented on HIVE-7653:


Thank you [~szehon]!
Appreciate your time on reviewing this.

 Hive AvroSerDe does not support circular references in Schema
 -

 Key: HIVE-7653
 URL: https://issues.apache.org/jira/browse/HIVE-7653
 Project: Hive
  Issue Type: Bug
Affects Versions: 0.13.1
Reporter: Sachin Goyal
Assignee: Sachin Goyal
 Attachments: HIVE-7653.1.patch, HIVE-7653.2.patch, HIVE-7653.3.patch, 
 HIVE-7653.4.patch, HIVE-7653.5.patch, HIVE-7653.6.patch, HIVE-7653.6.patch


 Avro allows nullable circular references but Hive AvroSerDe does not.
 Example of circular references (passing in Avro but failing in AvroSerDe):
 {code}
 class AvroCycleParent {
   AvroCycleChild child;
   public AvroCycleChild getChild () {return child;}
   public void setChild (AvroCycleChild child) {this.child = child;}
 }
 class AvroCycleChild {
   AvroCycleParent parent;
   public AvroCycleParent getParent () {return parent;}
   public void setParent (AvroCycleParent parent) {this.parent = parent;}
 }
 {code}
 Due to this discrepancy, Hive is unable to read Avro records having 
 circular-references. For some third-party code with such references, it 
 becomes very hard to directly serialize it with Avro and use in Hive.
 I have a patch for this with a unit-test and I will submit it shortly.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9641) Fill out remaining partition functions in HBaseStore

2015-02-19 Thread Alan Gates (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9641?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alan Gates updated HIVE-9641:
-
Attachment: HIVE-9641.patch

 Fill out remaining partition functions in HBaseStore
 

 Key: HIVE-9641
 URL: https://issues.apache.org/jira/browse/HIVE-9641
 Project: Hive
  Issue Type: Sub-task
  Components: Metastore
Reporter: Alan Gates
Assignee: Alan Gates
 Attachments: HIVE-9641.patch


 A number of the listPartition and getPartition methods are not implemented.  
 The rest need to be implemented.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9718) Insert into dynamic partitions with same column structure in the distibute by clause barfs

2015-02-19 Thread Pavan Srinivas (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9718?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Pavan Srinivas updated HIVE-9718:
-
Attachment: HIVE-9718-0.14.patch
HIVE-9718-1.0.patch
HIVE-9718.patch

Patches for trunk, 0.14 release, 1.0 release

 Insert into dynamic partitions with same column structure in the distibute 
 by clause barfs
 

 Key: HIVE-9718
 URL: https://issues.apache.org/jira/browse/HIVE-9718
 Project: Hive
  Issue Type: Bug
Affects Versions: 0.14.0, 1.0.0
Reporter: Pavan Srinivas
Priority: Critical
 Attachments: HIVE-9718-0.14.patch, HIVE-9718-1.0.patch, 
 HIVE-9718.patch, nation.tbl, patch.txt


 Sample reproducible query: 
 {code}
 SET hive.exec.dynamic.partition.mode=nonstrict;
 SET hive.exec.dynamic.partition=true;
  insert overwrite table nation_new_p partition (some)
 select n_name as name1, n_name as name2, n_name as name3 from nation 
 distribute by name3;
 {code}
 Note: Make sure there is data in the source table to reproduce the issue. 
 During the optimizations done for Jira: 
 https://issues.apache.org/jira/browse/HIVE-4867, an optimization of 
 deduplication of columns is done. But, when one of the columns is used as 
 part of partitioned/distribute by, its not taken care of.  
 The above query produces exception as follows:
 {code}
 Diagnostic Messages for this Task:
 java.lang.RuntimeException: org.apache.hadoop.hive.ql.metadata.HiveException: 
 Hive Runtime Error while processing row 
 {n_nationkey:0,n_name:ALGERIA,n_regionkey:0,n_comment: haggle. 
 carefully final deposits detect slyly agai}
   at org.apache.hadoop.hive.ql.exec.mr.ExecMapper.map(ExecMapper.java:185)
   at org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:54)
   at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:450)
   at org.apache.hadoop.mapred.MapTask.run(MapTask.java:343)
   at 
 org.apache.hadoop.mapred.LocalContainerLauncher$EventHandler.runSubtask(LocalContainerLauncher.java:370)
   at 
 org.apache.hadoop.mapred.LocalContainerLauncher$EventHandler.runTask(LocalContainerLauncher.java:295)
   at 
 org.apache.hadoop.mapred.LocalContainerLauncher$EventHandler.access$200(LocalContainerLauncher.java:181)
   at 
 org.apache.hadoop.mapred.LocalContainerLauncher$EventHandler$1.run(LocalContainerLauncher.java:224)
   at 
 java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471)
   at java.util.concurrent.FutureTask.run(FutureTask.java:262)
   at 
 java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
   at 
 java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
   at java.lang.Thread.run(Thread.java:744)
 Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime 
 Error while processing row 
 {n_nationkey:0,n_name:ALGERIA,n_regionkey:0,n_comment: haggle. 
 carefully final deposits detect slyly agai}
   at 
 org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:503)
   at org.apache.hadoop.hive.ql.exec.mr.ExecMapper.map(ExecMapper.java:176)
   ... 12 more
 Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: 
 java.lang.RuntimeException: cannot find field _col2 from [0:_col0]
   at 
 org.apache.hadoop.hive.ql.exec.ReduceSinkOperator.processOp(ReduceSinkOperator.java:397)
   at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:815)
   at 
 org.apache.hadoop.hive.ql.exec.SelectOperator.processOp(SelectOperator.java:84)
   at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:815)
   at 
 org.apache.hadoop.hive.ql.exec.TableScanOperator.processOp(TableScanOperator.java:95)
   at 
 org.apache.hadoop.hive.ql.exec.MapOperator$MapOpCtx.forward(MapOperator.java:157)
   at 
 org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:493)
   ... 13 more
 Caused by: java.lang.RuntimeException: cannot find field _col2 from [0:_col0]
   at 
 org.apache.hadoop.hive.serde2.objectinspector.ObjectInspectorUtils.getStandardStructFieldRef(ObjectInspectorUtils.java:410)
   at 
 org.apache.hadoop.hive.serde2.objectinspector.StandardStructObjectInspector.getStructFieldRef(StandardStructObjectInspector.java:147)
   at 
 org.apache.hadoop.hive.ql.exec.ExprNodeColumnEvaluator.initialize(ExprNodeColumnEvaluator.java:55)
   at 
 org.apache.hadoop.hive.ql.exec.Operator.initEvaluators(Operator.java:954)
   at 
 org.apache.hadoop.hive.ql.exec.ReduceSinkOperator.processOp(ReduceSinkOperator.java:325)
   ... 19 more
 {code}
 Tables used are: 
 {code}
 CREATE EXTERNAL TABLE `nation`(
   `n_nationkey` int,
   `n_name` string,
   `n_regionkey` int,
   `n_comment` string)
 ROW FORMAT DELIMITED
   

Re: Fix version for hbase-metastore branch

2015-02-19 Thread Thejas Nair
I agree, using a label for fix version makes sense in this case. I believe
that is what had been done for hive-on-spark and hive-on-tez.


On Thu, Feb 19, 2015 at 10:56 AM, Alan Gates alanfga...@gmail.com wrote:

 I've been marking JIRAs on this branch as fixed in 1.2, since that's the
 next version.  But that seems wrong as I doubt this code will be in by
 1.2.  What's the usual practice here?  It seems it would make sense to make
 a label for this branch and mark them as fixed with that label and then
 when we actually release this in a version we can update all the JIRAs with
 that label.

 Alan.



[jira] [Commented] (HIVE-7100) Users of hive should be able to specify skipTrash when dropping tables.

2015-02-19 Thread david serafini (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-7100?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14328026#comment-14328026
 ] 

david serafini commented on HIVE-7100:
--

Looks ok to me.

 Users of hive should be able to specify skipTrash when dropping tables.
 ---

 Key: HIVE-7100
 URL: https://issues.apache.org/jira/browse/HIVE-7100
 Project: Hive
  Issue Type: Improvement
  Components: Query Processor
Affects Versions: 0.13.0
Reporter: Ravi Prakash
Assignee: david serafini
  Labels: TODOC14
 Fix For: 0.14.0

 Attachments: HIVE-7100.1.patch, HIVE-7100.10.patch, 
 HIVE-7100.11.patch, HIVE-7100.2.patch, HIVE-7100.3.patch, HIVE-7100.4.patch, 
 HIVE-7100.5.patch, HIVE-7100.8.patch, HIVE-7100.9.patch, HIVE-7100.patch


 Users of our clusters are often running up against their quota limits because 
 of Hive tables. When they drop tables, they have to then manually delete the 
 files from HDFS using skipTrash. This is cumbersome and unnecessary. We 
 should enable users to skipTrash directly when dropping tables.
 We should also be able to provide this functionality without polluting SQL 
 syntax.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-6977) Delete Hiveserver1

2015-02-19 Thread Lefty Leverenz (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6977?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14327240#comment-14327240
 ] 

Lefty Leverenz commented on HIVE-6977:
--

Was the commit to version 1.0.0 reverted, or should this issue have fix version 
1.0.0 as well as 1.1.0?

 Delete Hiveserver1
 --

 Key: HIVE-6977
 URL: https://issues.apache.org/jira/browse/HIVE-6977
 Project: Hive
  Issue Type: Task
  Components: JDBC, Server Infrastructure
Reporter: Ashutosh Chauhan
Assignee: Ashutosh Chauhan
  Labels: TODOC15
 Fix For: 1.1.0

 Attachments: HIVE-6977.1.patch, HIVE-6977.patch


 See mailing list discussion.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HIVE-9724) Hive source pointing to http rather than https using maven

2015-02-19 Thread srinivas (JIRA)
srinivas created HIVE-9724:
--

 Summary: Hive source pointing to http rather than https using maven
 Key: HIVE-9724
 URL: https://issues.apache.org/jira/browse/HIVE-9724
 Project: Hive
  Issue Type: Bug
Reporter: srinivas


Refereed from the below link

https://cwiki.apache.org/confluence/display/Hive/HiveDeveloperFAQ#HiveDeveloperFAQ-HowdoIimportintoEclipse
?

Generating tarball resulted in below error.

#mvn clean package -DskipTests -Phadoop-1 -Pdist

ERROR::

Failed to execute goal on project hive-shims-common: Could not resolve
dependencies for project
org.apache.hive.shims:hive-shims-common:jar:1.2.0-SNAPSHOT: Could not
transfer artifact jline:jline:jar:2.12 from/to central (
http://repo.maven.apache.org/maven2): Access denied to:
http://repo.maven.apache.org/maven2/jline/jline/2.12/jline-2.12.jar,
ReasonPhrase:Forbidden. - [Help 1]
[ERROR]


Please advice



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HIVE-9723) CBO (Calcite Return Path): Translate Union to Hive Op [CBO branch]

2015-02-19 Thread Jesus Camacho Rodriguez (JIRA)
Jesus Camacho Rodriguez created HIVE-9723:
-

 Summary: CBO (Calcite Return Path): Translate Union to Hive Op 
[CBO branch]
 Key: HIVE-9723
 URL: https://issues.apache.org/jira/browse/HIVE-9723
 Project: Hive
  Issue Type: Sub-task
Reporter: Jesus Camacho Rodriguez
Assignee: Jesus Camacho Rodriguez






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Work started] (HIVE-9581) CBO (Calcite Return Path): Translate Join to Hive Op [CBO branch]

2015-02-19 Thread Jesus Camacho Rodriguez (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9581?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Work on HIVE-9581 started by Jesus Camacho Rodriguez.
-
 CBO (Calcite Return Path): Translate Join to Hive Op [CBO branch]
 -

 Key: HIVE-9581
 URL: https://issues.apache.org/jira/browse/HIVE-9581
 Project: Hive
  Issue Type: Sub-task
  Components: CBO
Affects Versions: cbo-branch
Reporter: Jesus Camacho Rodriguez
Assignee: Jesus Camacho Rodriguez
 Fix For: 1.2.0






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HIVE-9722) CBO (Calcite Return Path): Translate Sort/Limit to Hive Op [CBO branch]

2015-02-19 Thread Jesus Camacho Rodriguez (JIRA)
Jesus Camacho Rodriguez created HIVE-9722:
-

 Summary: CBO (Calcite Return Path): Translate Sort/Limit to Hive 
Op [CBO branch]
 Key: HIVE-9722
 URL: https://issues.apache.org/jira/browse/HIVE-9722
 Project: Hive
  Issue Type: Sub-task
Reporter: Jesus Camacho Rodriguez
Assignee: Jesus Camacho Rodriguez






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-7100) Users of hive should be able to specify skipTrash when dropping tables.

2015-02-19 Thread Lefty Leverenz (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-7100?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14327190#comment-14327190
 ] 

Lefty Leverenz commented on HIVE-7100:
--

Doc note:  This is documented in the wiki for DROP TABLE and ALTER TABLE DROP 
PARTITION.

* [Drop Table | 
https://cwiki.apache.org/confluence/display/Hive/LanguageManual+DDL#LanguageManualDDL-DropTable]
* [Drop Partitions | 
https://cwiki.apache.org/confluence/display/Hive/LanguageManual+DDL#LanguageManualDDL-DropPartitions]

I included DROP PARTITION based on deletePartitionData() in the patch -- please 
review my changes and correct anything that's not right.  If it's okay, the 
TODOC14 label can be removed.

 Users of hive should be able to specify skipTrash when dropping tables.
 ---

 Key: HIVE-7100
 URL: https://issues.apache.org/jira/browse/HIVE-7100
 Project: Hive
  Issue Type: Improvement
  Components: Query Processor
Affects Versions: 0.13.0
Reporter: Ravi Prakash
Assignee: david serafini
  Labels: TODOC14
 Fix For: 0.14.0

 Attachments: HIVE-7100.1.patch, HIVE-7100.10.patch, 
 HIVE-7100.11.patch, HIVE-7100.2.patch, HIVE-7100.3.patch, HIVE-7100.4.patch, 
 HIVE-7100.5.patch, HIVE-7100.8.patch, HIVE-7100.9.patch, HIVE-7100.patch


 Users of our clusters are often running up against their quota limits because 
 of Hive tables. When they drop tables, they have to then manually delete the 
 files from HDFS using skipTrash. This is cumbersome and unnecessary. We 
 should enable users to skipTrash directly when dropping tables.
 We should also be able to provide this functionality without polluting SQL 
 syntax.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9388) HiveServer2 fails to reconnect to MetaStore after MetaStore restart

2015-02-19 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9388?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14327225#comment-14327225
 ] 

Hive QA commented on HIVE-9388:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12699629/HIVE-9388.2.patch

{color:red}ERROR:{color} -1 due to 1 failed/errored test(s), 7561 tests executed
*Failed tests:*
{noformat}
org.apache.hive.spark.client.TestSparkClient.testJobSubmission
{noformat}

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/2822/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/2822/console
Test logs: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-2822/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 1 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12699629 - PreCommit-HIVE-TRUNK-Build

 HiveServer2 fails to reconnect to MetaStore after MetaStore restart
 ---

 Key: HIVE-9388
 URL: https://issues.apache.org/jira/browse/HIVE-9388
 Project: Hive
  Issue Type: Bug
  Components: HiveServer2
Affects Versions: 0.12.0, 0.14.0, 0.13.1
Reporter: Piotr Ackermann
 Attachments: HIVE-9388.2.patch, HIVE-9388.patch


 How to reproduce:
 # Use Hue to connect to HiveServer2
 # Restart Metastore
 # Try to execute any query in Hue
 HiveServer2 report error:
 {quote}
 ERROR hive.log: Got exception: 
 org.apache.thrift.transport.TTransportException null
 org.apache.thrift.transport.TTransportException
 at 
 org.apache.thrift.transport.TIOStreamTransport.read(TIOStreamTransport.java:132)
 at org.apache.thrift.transport.TTransport.readAll(TTransport.java:84)
 at 
 org.apache.thrift.transport.TSaslTransport.readLength(TSaslTransport.java:355)
 at 
 org.apache.thrift.transport.TSaslTransport.readFrame(TSaslTransport.java:432)
 at org.apache.thrift.transport.TSaslTransport.read(TSaslTransport.java:414)
 at 
 org.apache.thrift.transport.TSaslClientTransport.read(TSaslClientTransport.java:37)
 at org.apache.thrift.transport.TTransport.readAll(TTransport.java:84)
 at 
 org.apache.hadoop.hive.thrift.TFilterTransport.readAll(TFilterTransport.java:62)
 at 
 org.apache.thrift.protocol.TBinaryProtocol.readAll(TBinaryProtocol.java:378)
 at 
 org.apache.thrift.protocol.TBinaryProtocol.readI32(TBinaryProtocol.java:297)
 at 
 org.apache.thrift.protocol.TBinaryProtocol.readMessageBegin(TBinaryProtocol.java:204)
 at org.apache.thrift.TServiceClient.receiveBase(TServiceClient.java:69)
 at 
 org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.recv_get_databases(ThriftHiveMetastore.java:600)
 at 
 org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.get_databases(ThriftHiveMetastore.java:587)
 at 
 org.apache.hadoop.hive.metastore.HiveMetaStoreClient.getDatabases(HiveMetaStoreClient.java:837)
 at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
 at 
 sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
 at 
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
 at java.lang.reflect.Method.invoke(Method.java:606)
 at 
 org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.invoke(RetryingMetaStoreClient.java:90)
 at com.sun.proxy.$Proxy10.getDatabases(Unknown Source)
 at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
 at 
 sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
 at 
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
 at java.lang.reflect.Method.invoke(Method.java:606)
 at 
 org.apache.hadoop.hive.metastore.HiveMetaStoreClient$SynchronizedHandler.invoke(HiveMetaStoreClient.java:1681)
 at com.sun.proxy.$Proxy10.getDatabases(Unknown Source)
 at 
 org.apache.hive.service.cli.operation.GetSchemasOperation.run(GetSchemasOperation.java:62)
 at 
 org.apache.hive.service.cli.session.HiveSessionImpl.runOperationWithLogCapture(HiveSessionImpl.java:715)
 at 
 org.apache.hive.service.cli.session.HiveSessionImpl.getSchemas(HiveSessionImpl.java:438)
 at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
 at 
 sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
 at 
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
 at java.lang.reflect.Method.invoke(Method.java:606)
 at 
 org.apache.hive.service.cli.session.HiveSessionProxy.invoke(HiveSessionProxy.java:79)
 at 
 

[jira] [Updated] (HIVE-9581) CBO (Calcite Return Path): Translate Join to Hive Op [CBO branch]

2015-02-19 Thread Jesus Camacho Rodriguez (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9581?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jesus Camacho Rodriguez updated HIVE-9581:
--
Summary: CBO (Calcite Return Path): Translate Join to Hive Op [CBO branch]  
(was: CBO (Calcite Return Path): Translate Join to Hive Op)

 CBO (Calcite Return Path): Translate Join to Hive Op [CBO branch]
 -

 Key: HIVE-9581
 URL: https://issues.apache.org/jira/browse/HIVE-9581
 Project: Hive
  Issue Type: Sub-task
  Components: CBO
Reporter: Jesus Camacho Rodriguez
Assignee: Jesus Camacho Rodriguez
 Fix For: 1.2.0






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


Re: [VOTE] Apache Hive 1.1.0 Release Candidate 3

2015-02-19 Thread Szehon Ho
+1

1.  Verified signature for bin and src
2.  Built src with hadoop2
3.  Ran few queries from beeline with src
4.  Ran few queries from beeline with bin
5.  Verified no SNAPSHOT deps

Thanks
Szehon

On Wed, Feb 18, 2015 at 10:03 PM, Xuefu Zhang xzh...@cloudera.com wrote:

 +1

 1. downloaded the src tarball and built w/ -Phadoop-1/2
 2. verified no binary (jars) in the src tarball

 On Wed, Feb 18, 2015 at 8:56 PM, Brock Noland br...@cloudera.com wrote:

  +1
 
  verified sigs, hashes, created tables, ran MR on YARN jobs
 
  On Wed, Feb 18, 2015 at 8:54 PM, Brock Noland br...@cloudera.com
 wrote:
   Apache Hive 1.1.0 Release Candidate 3 is available here:
   http://people.apache.org/~brock/apache-hive-1.1.0-rc3/
  
   Maven artifacts are available here:
   https://repository.apache.org/content/repositories/orgapachehive-1026/
  
   Source tag for RC3 is at:
   http://svn.apache.org/repos/asf/hive/tags/release-1.1.0-rc3/
  
   My key is located here: https://people.apache.org/keys/group/hive.asc
  
   Voting will conclude in 72 hours
 



[jira] [Updated] (HIVE-9581) CBO (Calcite Return Path): Translate Join to Hive Op [CBO branch]

2015-02-19 Thread Jesus Camacho Rodriguez (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9581?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jesus Camacho Rodriguez updated HIVE-9581:
--
Affects Version/s: cbo-branch

 CBO (Calcite Return Path): Translate Join to Hive Op [CBO branch]
 -

 Key: HIVE-9581
 URL: https://issues.apache.org/jira/browse/HIVE-9581
 Project: Hive
  Issue Type: Sub-task
  Components: CBO
Affects Versions: cbo-branch
Reporter: Jesus Camacho Rodriguez
Assignee: Jesus Camacho Rodriguez
 Fix For: 1.2.0






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Work started] (HIVE-9723) CBO (Calcite Return Path): Translate Union to Hive Op [CBO branch]

2015-02-19 Thread Jesus Camacho Rodriguez (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9723?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Work on HIVE-9723 started by Jesus Camacho Rodriguez.
-
 CBO (Calcite Return Path): Translate Union to Hive Op [CBO branch]
 --

 Key: HIVE-9723
 URL: https://issues.apache.org/jira/browse/HIVE-9723
 Project: Hive
  Issue Type: Sub-task
  Components: CBO
Affects Versions: cbo-branch
Reporter: Jesus Camacho Rodriguez
Assignee: Jesus Camacho Rodriguez
 Fix For: 1.2.0






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9722) CBO (Calcite Return Path): Translate Sort/Limit to Hive Op [CBO branch]

2015-02-19 Thread Jesus Camacho Rodriguez (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9722?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jesus Camacho Rodriguez updated HIVE-9722:
--
Affects Version/s: cbo-branch

 CBO (Calcite Return Path): Translate Sort/Limit to Hive Op [CBO branch]
 ---

 Key: HIVE-9722
 URL: https://issues.apache.org/jira/browse/HIVE-9722
 Project: Hive
  Issue Type: Sub-task
  Components: CBO
Affects Versions: cbo-branch
Reporter: Jesus Camacho Rodriguez
Assignee: Jesus Camacho Rodriguez
 Fix For: 1.2.0






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Work started] (HIVE-9722) CBO (Calcite Return Path): Translate Sort/Limit to Hive Op [CBO branch]

2015-02-19 Thread Jesus Camacho Rodriguez (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9722?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Work on HIVE-9722 started by Jesus Camacho Rodriguez.
-
 CBO (Calcite Return Path): Translate Sort/Limit to Hive Op [CBO branch]
 ---

 Key: HIVE-9722
 URL: https://issues.apache.org/jira/browse/HIVE-9722
 Project: Hive
  Issue Type: Sub-task
  Components: CBO
Affects Versions: cbo-branch
Reporter: Jesus Camacho Rodriguez
Assignee: Jesus Camacho Rodriguez
 Fix For: 1.2.0






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9723) CBO (Calcite Return Path): Translate Union to Hive Op [CBO branch]

2015-02-19 Thread Jesus Camacho Rodriguez (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9723?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jesus Camacho Rodriguez updated HIVE-9723:
--
Affects Version/s: cbo-branch

 CBO (Calcite Return Path): Translate Union to Hive Op [CBO branch]
 --

 Key: HIVE-9723
 URL: https://issues.apache.org/jira/browse/HIVE-9723
 Project: Hive
  Issue Type: Sub-task
  Components: CBO
Affects Versions: cbo-branch
Reporter: Jesus Camacho Rodriguez
Assignee: Jesus Camacho Rodriguez
 Fix For: 1.2.0






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


JIRA change log for 1.0.0 only has 9 bugs

2015-02-19 Thread Lefty Leverenz
The release 1.0.0 JIRA change log
https://issues.apache.org/jira/secure/ReleaseNote.jspa?version=12329278styleName=TextprojectId=12310843
linked
to the downloads http://hive.apache.org/downloads.html page only has 9
bugs but the downloaded release notes file has more than 50.

Spot checking shows that Brock removed Fix Version 1.0.0 from the missing
bugs when he changed 0.15.0 to 1.1.0.  For example, HIVE-6977
https://issues.apache.org/jira/browse/HIVE-6977 (Delete Hiveserver1) only
says it's fixed in 1.1.0 although the patch is also in 1.0.0.

Brock, can you revert those changes easily or do we need to go through the
list manually?

-- Lefty


[jira] [Updated] (HIVE-9581) CBO (Calcite Return Path): Translate Join to Hive Op [CBO branch]

2015-02-19 Thread Jesus Camacho Rodriguez (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9581?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jesus Camacho Rodriguez updated HIVE-9581:
--
Attachment: (was: HIVE-9581.cbo.patch)

 CBO (Calcite Return Path): Translate Join to Hive Op [CBO branch]
 -

 Key: HIVE-9581
 URL: https://issues.apache.org/jira/browse/HIVE-9581
 Project: Hive
  Issue Type: Sub-task
  Components: CBO
Affects Versions: cbo-branch
Reporter: Jesus Camacho Rodriguez
Assignee: Jesus Camacho Rodriguez
 Fix For: 1.2.0

 Attachments: HIVE-9581.cbo.patch






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9718) Insert into dynamic partitions with same column structure in the distibute by clause barfs

2015-02-19 Thread Pavan Srinivas (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9718?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14327372#comment-14327372
 ] 

Pavan Srinivas commented on HIVE-9718:
--

Thanks for the comment. Yes, this optimization helps in preparing a better 
plan. Having said that, this bug breaks backward compatibility(this query works 
on Hive-0.13.1 and Hive-0.11). Ideal way is to fix this and make it compatible. 

 Insert into dynamic partitions with same column structure in the distibute 
 by clause barfs
 

 Key: HIVE-9718
 URL: https://issues.apache.org/jira/browse/HIVE-9718
 Project: Hive
  Issue Type: Bug
Affects Versions: 0.14.0, 1.0.0
Reporter: Pavan Srinivas
Priority: Critical
 Attachments: nation.tbl, patch.txt


 Sample reproducible query: 
 {code}
 SET hive.exec.dynamic.partition.mode=nonstrict;
 SET hive.exec.dynamic.partition=true;
  insert overwrite table nation_new_p partition (some)
 select n_name as name1, n_name as name2, n_name as name3 from nation 
 distribute by name3;
 {code}
 Note: Make sure there is data in the source table to reproduce the issue. 
 During the optimizations done for Jira: 
 https://issues.apache.org/jira/browse/HIVE-4867, an optimization of 
 deduplication of columns is done. But, when one of the columns is used as 
 part of partitioned/distribute by, its not taken care of.  
 The above query produces exception as follows:
 {code}
 Diagnostic Messages for this Task:
 java.lang.RuntimeException: org.apache.hadoop.hive.ql.metadata.HiveException: 
 Hive Runtime Error while processing row 
 {n_nationkey:0,n_name:ALGERIA,n_regionkey:0,n_comment: haggle. 
 carefully final deposits detect slyly agai}
   at org.apache.hadoop.hive.ql.exec.mr.ExecMapper.map(ExecMapper.java:185)
   at org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:54)
   at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:450)
   at org.apache.hadoop.mapred.MapTask.run(MapTask.java:343)
   at 
 org.apache.hadoop.mapred.LocalContainerLauncher$EventHandler.runSubtask(LocalContainerLauncher.java:370)
   at 
 org.apache.hadoop.mapred.LocalContainerLauncher$EventHandler.runTask(LocalContainerLauncher.java:295)
   at 
 org.apache.hadoop.mapred.LocalContainerLauncher$EventHandler.access$200(LocalContainerLauncher.java:181)
   at 
 org.apache.hadoop.mapred.LocalContainerLauncher$EventHandler$1.run(LocalContainerLauncher.java:224)
   at 
 java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471)
   at java.util.concurrent.FutureTask.run(FutureTask.java:262)
   at 
 java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
   at 
 java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
   at java.lang.Thread.run(Thread.java:744)
 Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime 
 Error while processing row 
 {n_nationkey:0,n_name:ALGERIA,n_regionkey:0,n_comment: haggle. 
 carefully final deposits detect slyly agai}
   at 
 org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:503)
   at org.apache.hadoop.hive.ql.exec.mr.ExecMapper.map(ExecMapper.java:176)
   ... 12 more
 Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: 
 java.lang.RuntimeException: cannot find field _col2 from [0:_col0]
   at 
 org.apache.hadoop.hive.ql.exec.ReduceSinkOperator.processOp(ReduceSinkOperator.java:397)
   at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:815)
   at 
 org.apache.hadoop.hive.ql.exec.SelectOperator.processOp(SelectOperator.java:84)
   at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:815)
   at 
 org.apache.hadoop.hive.ql.exec.TableScanOperator.processOp(TableScanOperator.java:95)
   at 
 org.apache.hadoop.hive.ql.exec.MapOperator$MapOpCtx.forward(MapOperator.java:157)
   at 
 org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:493)
   ... 13 more
 Caused by: java.lang.RuntimeException: cannot find field _col2 from [0:_col0]
   at 
 org.apache.hadoop.hive.serde2.objectinspector.ObjectInspectorUtils.getStandardStructFieldRef(ObjectInspectorUtils.java:410)
   at 
 org.apache.hadoop.hive.serde2.objectinspector.StandardStructObjectInspector.getStructFieldRef(StandardStructObjectInspector.java:147)
   at 
 org.apache.hadoop.hive.ql.exec.ExprNodeColumnEvaluator.initialize(ExprNodeColumnEvaluator.java:55)
   at 
 org.apache.hadoop.hive.ql.exec.Operator.initEvaluators(Operator.java:954)
   at 
 org.apache.hadoop.hive.ql.exec.ReduceSinkOperator.processOp(ReduceSinkOperator.java:325)
   ... 19 more
 {code}
 Tables used are: 
 {code}
 CREATE EXTERNAL TABLE `nation`(
   `n_nationkey` int,
   `n_name` 

[jira] [Commented] (HIVE-9718) Insert into dynamic partitions with same column structure in the distibute by clause barfs

2015-02-19 Thread Pavan Srinivas (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9718?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14327377#comment-14327377
 ] 

Pavan Srinivas commented on HIVE-9718:
--

Thanks for the comment. Yes, this optimization helps in preparing a better 
plan. Having said that, this bug breaks backward compatibility(this query works 
on Hive-0.13.1 and Hive-0.11). Ideal way is to fix this and make it compatible. 

 Insert into dynamic partitions with same column structure in the distibute 
 by clause barfs
 

 Key: HIVE-9718
 URL: https://issues.apache.org/jira/browse/HIVE-9718
 Project: Hive
  Issue Type: Bug
Affects Versions: 0.14.0, 1.0.0
Reporter: Pavan Srinivas
Priority: Critical
 Attachments: nation.tbl, patch.txt


 Sample reproducible query: 
 {code}
 SET hive.exec.dynamic.partition.mode=nonstrict;
 SET hive.exec.dynamic.partition=true;
  insert overwrite table nation_new_p partition (some)
 select n_name as name1, n_name as name2, n_name as name3 from nation 
 distribute by name3;
 {code}
 Note: Make sure there is data in the source table to reproduce the issue. 
 During the optimizations done for Jira: 
 https://issues.apache.org/jira/browse/HIVE-4867, an optimization of 
 deduplication of columns is done. But, when one of the columns is used as 
 part of partitioned/distribute by, its not taken care of.  
 The above query produces exception as follows:
 {code}
 Diagnostic Messages for this Task:
 java.lang.RuntimeException: org.apache.hadoop.hive.ql.metadata.HiveException: 
 Hive Runtime Error while processing row 
 {n_nationkey:0,n_name:ALGERIA,n_regionkey:0,n_comment: haggle. 
 carefully final deposits detect slyly agai}
   at org.apache.hadoop.hive.ql.exec.mr.ExecMapper.map(ExecMapper.java:185)
   at org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:54)
   at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:450)
   at org.apache.hadoop.mapred.MapTask.run(MapTask.java:343)
   at 
 org.apache.hadoop.mapred.LocalContainerLauncher$EventHandler.runSubtask(LocalContainerLauncher.java:370)
   at 
 org.apache.hadoop.mapred.LocalContainerLauncher$EventHandler.runTask(LocalContainerLauncher.java:295)
   at 
 org.apache.hadoop.mapred.LocalContainerLauncher$EventHandler.access$200(LocalContainerLauncher.java:181)
   at 
 org.apache.hadoop.mapred.LocalContainerLauncher$EventHandler$1.run(LocalContainerLauncher.java:224)
   at 
 java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471)
   at java.util.concurrent.FutureTask.run(FutureTask.java:262)
   at 
 java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
   at 
 java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
   at java.lang.Thread.run(Thread.java:744)
 Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime 
 Error while processing row 
 {n_nationkey:0,n_name:ALGERIA,n_regionkey:0,n_comment: haggle. 
 carefully final deposits detect slyly agai}
   at 
 org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:503)
   at org.apache.hadoop.hive.ql.exec.mr.ExecMapper.map(ExecMapper.java:176)
   ... 12 more
 Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: 
 java.lang.RuntimeException: cannot find field _col2 from [0:_col0]
   at 
 org.apache.hadoop.hive.ql.exec.ReduceSinkOperator.processOp(ReduceSinkOperator.java:397)
   at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:815)
   at 
 org.apache.hadoop.hive.ql.exec.SelectOperator.processOp(SelectOperator.java:84)
   at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:815)
   at 
 org.apache.hadoop.hive.ql.exec.TableScanOperator.processOp(TableScanOperator.java:95)
   at 
 org.apache.hadoop.hive.ql.exec.MapOperator$MapOpCtx.forward(MapOperator.java:157)
   at 
 org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:493)
   ... 13 more
 Caused by: java.lang.RuntimeException: cannot find field _col2 from [0:_col0]
   at 
 org.apache.hadoop.hive.serde2.objectinspector.ObjectInspectorUtils.getStandardStructFieldRef(ObjectInspectorUtils.java:410)
   at 
 org.apache.hadoop.hive.serde2.objectinspector.StandardStructObjectInspector.getStructFieldRef(StandardStructObjectInspector.java:147)
   at 
 org.apache.hadoop.hive.ql.exec.ExprNodeColumnEvaluator.initialize(ExprNodeColumnEvaluator.java:55)
   at 
 org.apache.hadoop.hive.ql.exec.Operator.initEvaluators(Operator.java:954)
   at 
 org.apache.hadoop.hive.ql.exec.ReduceSinkOperator.processOp(ReduceSinkOperator.java:325)
   ... 19 more
 {code}
 Tables used are: 
 {code}
 CREATE EXTERNAL TABLE `nation`(
   `n_nationkey` int,
   `n_name` 

[jira] [Commented] (HIVE-9718) Insert into dynamic partitions with same column structure in the distibute by clause barfs

2015-02-19 Thread Pavan Srinivas (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9718?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14327371#comment-14327371
 ] 

Pavan Srinivas commented on HIVE-9718:
--

Thanks for the comment. Yes, this optimization helps in preparing a better 
plan. Having said that, this bug breaks backward compatibility(this query works 
on Hive-0.13.1 and Hive-0.11). Ideal way is to fix this and make it compatible. 

 Insert into dynamic partitions with same column structure in the distibute 
 by clause barfs
 

 Key: HIVE-9718
 URL: https://issues.apache.org/jira/browse/HIVE-9718
 Project: Hive
  Issue Type: Bug
Affects Versions: 0.14.0, 1.0.0
Reporter: Pavan Srinivas
Priority: Critical
 Attachments: nation.tbl, patch.txt


 Sample reproducible query: 
 {code}
 SET hive.exec.dynamic.partition.mode=nonstrict;
 SET hive.exec.dynamic.partition=true;
  insert overwrite table nation_new_p partition (some)
 select n_name as name1, n_name as name2, n_name as name3 from nation 
 distribute by name3;
 {code}
 Note: Make sure there is data in the source table to reproduce the issue. 
 During the optimizations done for Jira: 
 https://issues.apache.org/jira/browse/HIVE-4867, an optimization of 
 deduplication of columns is done. But, when one of the columns is used as 
 part of partitioned/distribute by, its not taken care of.  
 The above query produces exception as follows:
 {code}
 Diagnostic Messages for this Task:
 java.lang.RuntimeException: org.apache.hadoop.hive.ql.metadata.HiveException: 
 Hive Runtime Error while processing row 
 {n_nationkey:0,n_name:ALGERIA,n_regionkey:0,n_comment: haggle. 
 carefully final deposits detect slyly agai}
   at org.apache.hadoop.hive.ql.exec.mr.ExecMapper.map(ExecMapper.java:185)
   at org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:54)
   at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:450)
   at org.apache.hadoop.mapred.MapTask.run(MapTask.java:343)
   at 
 org.apache.hadoop.mapred.LocalContainerLauncher$EventHandler.runSubtask(LocalContainerLauncher.java:370)
   at 
 org.apache.hadoop.mapred.LocalContainerLauncher$EventHandler.runTask(LocalContainerLauncher.java:295)
   at 
 org.apache.hadoop.mapred.LocalContainerLauncher$EventHandler.access$200(LocalContainerLauncher.java:181)
   at 
 org.apache.hadoop.mapred.LocalContainerLauncher$EventHandler$1.run(LocalContainerLauncher.java:224)
   at 
 java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471)
   at java.util.concurrent.FutureTask.run(FutureTask.java:262)
   at 
 java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
   at 
 java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
   at java.lang.Thread.run(Thread.java:744)
 Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime 
 Error while processing row 
 {n_nationkey:0,n_name:ALGERIA,n_regionkey:0,n_comment: haggle. 
 carefully final deposits detect slyly agai}
   at 
 org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:503)
   at org.apache.hadoop.hive.ql.exec.mr.ExecMapper.map(ExecMapper.java:176)
   ... 12 more
 Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: 
 java.lang.RuntimeException: cannot find field _col2 from [0:_col0]
   at 
 org.apache.hadoop.hive.ql.exec.ReduceSinkOperator.processOp(ReduceSinkOperator.java:397)
   at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:815)
   at 
 org.apache.hadoop.hive.ql.exec.SelectOperator.processOp(SelectOperator.java:84)
   at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:815)
   at 
 org.apache.hadoop.hive.ql.exec.TableScanOperator.processOp(TableScanOperator.java:95)
   at 
 org.apache.hadoop.hive.ql.exec.MapOperator$MapOpCtx.forward(MapOperator.java:157)
   at 
 org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:493)
   ... 13 more
 Caused by: java.lang.RuntimeException: cannot find field _col2 from [0:_col0]
   at 
 org.apache.hadoop.hive.serde2.objectinspector.ObjectInspectorUtils.getStandardStructFieldRef(ObjectInspectorUtils.java:410)
   at 
 org.apache.hadoop.hive.serde2.objectinspector.StandardStructObjectInspector.getStructFieldRef(StandardStructObjectInspector.java:147)
   at 
 org.apache.hadoop.hive.ql.exec.ExprNodeColumnEvaluator.initialize(ExprNodeColumnEvaluator.java:55)
   at 
 org.apache.hadoop.hive.ql.exec.Operator.initEvaluators(Operator.java:954)
   at 
 org.apache.hadoop.hive.ql.exec.ReduceSinkOperator.processOp(ReduceSinkOperator.java:325)
   ... 19 more
 {code}
 Tables used are: 
 {code}
 CREATE EXTERNAL TABLE `nation`(
   `n_nationkey` int,
   `n_name` 

[jira] [Commented] (HIVE-9718) Insert into dynamic partitions with same column structure in the distibute by clause barfs

2015-02-19 Thread Pavan Srinivas (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9718?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14327374#comment-14327374
 ] 

Pavan Srinivas commented on HIVE-9718:
--

Thanks for the comment. Yes, this optimization helps in preparing a better 
plan. Having said that, this bug breaks backward compatibility(this query works 
on Hive-0.13.1 and Hive-0.11). Ideal way is to fix this and make it compatible. 

 Insert into dynamic partitions with same column structure in the distibute 
 by clause barfs
 

 Key: HIVE-9718
 URL: https://issues.apache.org/jira/browse/HIVE-9718
 Project: Hive
  Issue Type: Bug
Affects Versions: 0.14.0, 1.0.0
Reporter: Pavan Srinivas
Priority: Critical
 Attachments: nation.tbl, patch.txt


 Sample reproducible query: 
 {code}
 SET hive.exec.dynamic.partition.mode=nonstrict;
 SET hive.exec.dynamic.partition=true;
  insert overwrite table nation_new_p partition (some)
 select n_name as name1, n_name as name2, n_name as name3 from nation 
 distribute by name3;
 {code}
 Note: Make sure there is data in the source table to reproduce the issue. 
 During the optimizations done for Jira: 
 https://issues.apache.org/jira/browse/HIVE-4867, an optimization of 
 deduplication of columns is done. But, when one of the columns is used as 
 part of partitioned/distribute by, its not taken care of.  
 The above query produces exception as follows:
 {code}
 Diagnostic Messages for this Task:
 java.lang.RuntimeException: org.apache.hadoop.hive.ql.metadata.HiveException: 
 Hive Runtime Error while processing row 
 {n_nationkey:0,n_name:ALGERIA,n_regionkey:0,n_comment: haggle. 
 carefully final deposits detect slyly agai}
   at org.apache.hadoop.hive.ql.exec.mr.ExecMapper.map(ExecMapper.java:185)
   at org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:54)
   at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:450)
   at org.apache.hadoop.mapred.MapTask.run(MapTask.java:343)
   at 
 org.apache.hadoop.mapred.LocalContainerLauncher$EventHandler.runSubtask(LocalContainerLauncher.java:370)
   at 
 org.apache.hadoop.mapred.LocalContainerLauncher$EventHandler.runTask(LocalContainerLauncher.java:295)
   at 
 org.apache.hadoop.mapred.LocalContainerLauncher$EventHandler.access$200(LocalContainerLauncher.java:181)
   at 
 org.apache.hadoop.mapred.LocalContainerLauncher$EventHandler$1.run(LocalContainerLauncher.java:224)
   at 
 java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471)
   at java.util.concurrent.FutureTask.run(FutureTask.java:262)
   at 
 java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
   at 
 java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
   at java.lang.Thread.run(Thread.java:744)
 Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime 
 Error while processing row 
 {n_nationkey:0,n_name:ALGERIA,n_regionkey:0,n_comment: haggle. 
 carefully final deposits detect slyly agai}
   at 
 org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:503)
   at org.apache.hadoop.hive.ql.exec.mr.ExecMapper.map(ExecMapper.java:176)
   ... 12 more
 Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: 
 java.lang.RuntimeException: cannot find field _col2 from [0:_col0]
   at 
 org.apache.hadoop.hive.ql.exec.ReduceSinkOperator.processOp(ReduceSinkOperator.java:397)
   at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:815)
   at 
 org.apache.hadoop.hive.ql.exec.SelectOperator.processOp(SelectOperator.java:84)
   at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:815)
   at 
 org.apache.hadoop.hive.ql.exec.TableScanOperator.processOp(TableScanOperator.java:95)
   at 
 org.apache.hadoop.hive.ql.exec.MapOperator$MapOpCtx.forward(MapOperator.java:157)
   at 
 org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:493)
   ... 13 more
 Caused by: java.lang.RuntimeException: cannot find field _col2 from [0:_col0]
   at 
 org.apache.hadoop.hive.serde2.objectinspector.ObjectInspectorUtils.getStandardStructFieldRef(ObjectInspectorUtils.java:410)
   at 
 org.apache.hadoop.hive.serde2.objectinspector.StandardStructObjectInspector.getStructFieldRef(StandardStructObjectInspector.java:147)
   at 
 org.apache.hadoop.hive.ql.exec.ExprNodeColumnEvaluator.initialize(ExprNodeColumnEvaluator.java:55)
   at 
 org.apache.hadoop.hive.ql.exec.Operator.initEvaluators(Operator.java:954)
   at 
 org.apache.hadoop.hive.ql.exec.ReduceSinkOperator.processOp(ReduceSinkOperator.java:325)
   ... 19 more
 {code}
 Tables used are: 
 {code}
 CREATE EXTERNAL TABLE `nation`(
   `n_nationkey` int,
   `n_name` 

[jira] [Updated] (HIVE-9388) HiveServer2 fails to reconnect to MetaStore after MetaStore restart

2015-02-19 Thread Mariusz Strzelecki (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9388?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mariusz Strzelecki updated HIVE-9388:
-
Affects Version/s: 1.0.0

 HiveServer2 fails to reconnect to MetaStore after MetaStore restart
 ---

 Key: HIVE-9388
 URL: https://issues.apache.org/jira/browse/HIVE-9388
 Project: Hive
  Issue Type: Bug
  Components: HiveServer2
Affects Versions: 0.12.0, 0.14.0, 0.13.1, 1.0.0
Reporter: Piotr Ackermann
 Attachments: HIVE-9388.2.patch, HIVE-9388.patch


 How to reproduce:
 # Use Hue to connect to HiveServer2
 # Restart Metastore
 # Try to execute any query in Hue
 HiveServer2 report error:
 {quote}
 ERROR hive.log: Got exception: 
 org.apache.thrift.transport.TTransportException null
 org.apache.thrift.transport.TTransportException
 at 
 org.apache.thrift.transport.TIOStreamTransport.read(TIOStreamTransport.java:132)
 at org.apache.thrift.transport.TTransport.readAll(TTransport.java:84)
 at 
 org.apache.thrift.transport.TSaslTransport.readLength(TSaslTransport.java:355)
 at 
 org.apache.thrift.transport.TSaslTransport.readFrame(TSaslTransport.java:432)
 at org.apache.thrift.transport.TSaslTransport.read(TSaslTransport.java:414)
 at 
 org.apache.thrift.transport.TSaslClientTransport.read(TSaslClientTransport.java:37)
 at org.apache.thrift.transport.TTransport.readAll(TTransport.java:84)
 at 
 org.apache.hadoop.hive.thrift.TFilterTransport.readAll(TFilterTransport.java:62)
 at 
 org.apache.thrift.protocol.TBinaryProtocol.readAll(TBinaryProtocol.java:378)
 at 
 org.apache.thrift.protocol.TBinaryProtocol.readI32(TBinaryProtocol.java:297)
 at 
 org.apache.thrift.protocol.TBinaryProtocol.readMessageBegin(TBinaryProtocol.java:204)
 at org.apache.thrift.TServiceClient.receiveBase(TServiceClient.java:69)
 at 
 org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.recv_get_databases(ThriftHiveMetastore.java:600)
 at 
 org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.get_databases(ThriftHiveMetastore.java:587)
 at 
 org.apache.hadoop.hive.metastore.HiveMetaStoreClient.getDatabases(HiveMetaStoreClient.java:837)
 at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
 at 
 sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
 at 
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
 at java.lang.reflect.Method.invoke(Method.java:606)
 at 
 org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.invoke(RetryingMetaStoreClient.java:90)
 at com.sun.proxy.$Proxy10.getDatabases(Unknown Source)
 at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
 at 
 sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
 at 
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
 at java.lang.reflect.Method.invoke(Method.java:606)
 at 
 org.apache.hadoop.hive.metastore.HiveMetaStoreClient$SynchronizedHandler.invoke(HiveMetaStoreClient.java:1681)
 at com.sun.proxy.$Proxy10.getDatabases(Unknown Source)
 at 
 org.apache.hive.service.cli.operation.GetSchemasOperation.run(GetSchemasOperation.java:62)
 at 
 org.apache.hive.service.cli.session.HiveSessionImpl.runOperationWithLogCapture(HiveSessionImpl.java:715)
 at 
 org.apache.hive.service.cli.session.HiveSessionImpl.getSchemas(HiveSessionImpl.java:438)
 at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
 at 
 sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
 at 
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
 at java.lang.reflect.Method.invoke(Method.java:606)
 at 
 org.apache.hive.service.cli.session.HiveSessionProxy.invoke(HiveSessionProxy.java:79)
 at 
 org.apache.hive.service.cli.session.HiveSessionProxy.access$000(HiveSessionProxy.java:37)
 at 
 org.apache.hive.service.cli.session.HiveSessionProxy$1.run(HiveSessionProxy.java:64)
 at java.security.AccessController.doPrivileged(Native Method)
 at javax.security.auth.Subject.doAs(Subject.java:415)
 at 
 org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1642)
 at 
 org.apache.hadoop.hive.shims.HadoopShimsSecure.doAs(HadoopShimsSecure.java:502)
 at 
 org.apache.hive.service.cli.session.HiveSessionProxy.invoke(HiveSessionProxy.java:60)
 at com.sun.proxy.$Proxy19.getSchemas(Unknown Source)
 at org.apache.hive.service.cli.CLIService.getSchemas(CLIService.java:277)
 at 
 org.apache.hive.service.cli.thrift.ThriftCLIService.GetSchemas(ThriftCLIService.java:436)
 at 
 org.apache.hive.service.cli.thrift.TCLIService$Processor$GetSchemas.getResult(TCLIService.java:1433)
 at 
 org.apache.hive.service.cli.thrift.TCLIService$Processor$GetSchemas.getResult(TCLIService.java:1418)
 at 

[jira] [Updated] (HIVE-9581) CBO (Calcite Return Path): Translate Join to Hive Op [CBO branch]

2015-02-19 Thread Jesus Camacho Rodriguez (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9581?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jesus Camacho Rodriguez updated HIVE-9581:
--
Attachment: HIVE-9581.cbo.patch

[~ashutoshc] [~jpullokkaran]

 CBO (Calcite Return Path): Translate Join to Hive Op [CBO branch]
 -

 Key: HIVE-9581
 URL: https://issues.apache.org/jira/browse/HIVE-9581
 Project: Hive
  Issue Type: Sub-task
  Components: CBO
Affects Versions: cbo-branch
Reporter: Jesus Camacho Rodriguez
Assignee: Jesus Camacho Rodriguez
 Fix For: 1.2.0

 Attachments: HIVE-9581.cbo.patch






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9388) HiveServer2 fails to reconnect to MetaStore after MetaStore restart

2015-02-19 Thread Mariusz Strzelecki (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9388?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14327383#comment-14327383
 ] 

Mariusz Strzelecki commented on HIVE-9388:
--

Unit test failure seems to be unrelated.

Bug can be also reproduced without hue:
1. Start hive metastore
2. Start hive cli
3. Restart hive metastore
4. Run show databases; in cli - an exception is thrown and client is not 
reconnected

Tested on 1.0.0 release.


 HiveServer2 fails to reconnect to MetaStore after MetaStore restart
 ---

 Key: HIVE-9388
 URL: https://issues.apache.org/jira/browse/HIVE-9388
 Project: Hive
  Issue Type: Bug
  Components: HiveServer2
Affects Versions: 0.12.0, 0.14.0, 0.13.1
Reporter: Piotr Ackermann
 Attachments: HIVE-9388.2.patch, HIVE-9388.patch


 How to reproduce:
 # Use Hue to connect to HiveServer2
 # Restart Metastore
 # Try to execute any query in Hue
 HiveServer2 report error:
 {quote}
 ERROR hive.log: Got exception: 
 org.apache.thrift.transport.TTransportException null
 org.apache.thrift.transport.TTransportException
 at 
 org.apache.thrift.transport.TIOStreamTransport.read(TIOStreamTransport.java:132)
 at org.apache.thrift.transport.TTransport.readAll(TTransport.java:84)
 at 
 org.apache.thrift.transport.TSaslTransport.readLength(TSaslTransport.java:355)
 at 
 org.apache.thrift.transport.TSaslTransport.readFrame(TSaslTransport.java:432)
 at org.apache.thrift.transport.TSaslTransport.read(TSaslTransport.java:414)
 at 
 org.apache.thrift.transport.TSaslClientTransport.read(TSaslClientTransport.java:37)
 at org.apache.thrift.transport.TTransport.readAll(TTransport.java:84)
 at 
 org.apache.hadoop.hive.thrift.TFilterTransport.readAll(TFilterTransport.java:62)
 at 
 org.apache.thrift.protocol.TBinaryProtocol.readAll(TBinaryProtocol.java:378)
 at 
 org.apache.thrift.protocol.TBinaryProtocol.readI32(TBinaryProtocol.java:297)
 at 
 org.apache.thrift.protocol.TBinaryProtocol.readMessageBegin(TBinaryProtocol.java:204)
 at org.apache.thrift.TServiceClient.receiveBase(TServiceClient.java:69)
 at 
 org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.recv_get_databases(ThriftHiveMetastore.java:600)
 at 
 org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.get_databases(ThriftHiveMetastore.java:587)
 at 
 org.apache.hadoop.hive.metastore.HiveMetaStoreClient.getDatabases(HiveMetaStoreClient.java:837)
 at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
 at 
 sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
 at 
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
 at java.lang.reflect.Method.invoke(Method.java:606)
 at 
 org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.invoke(RetryingMetaStoreClient.java:90)
 at com.sun.proxy.$Proxy10.getDatabases(Unknown Source)
 at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
 at 
 sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
 at 
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
 at java.lang.reflect.Method.invoke(Method.java:606)
 at 
 org.apache.hadoop.hive.metastore.HiveMetaStoreClient$SynchronizedHandler.invoke(HiveMetaStoreClient.java:1681)
 at com.sun.proxy.$Proxy10.getDatabases(Unknown Source)
 at 
 org.apache.hive.service.cli.operation.GetSchemasOperation.run(GetSchemasOperation.java:62)
 at 
 org.apache.hive.service.cli.session.HiveSessionImpl.runOperationWithLogCapture(HiveSessionImpl.java:715)
 at 
 org.apache.hive.service.cli.session.HiveSessionImpl.getSchemas(HiveSessionImpl.java:438)
 at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
 at 
 sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
 at 
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
 at java.lang.reflect.Method.invoke(Method.java:606)
 at 
 org.apache.hive.service.cli.session.HiveSessionProxy.invoke(HiveSessionProxy.java:79)
 at 
 org.apache.hive.service.cli.session.HiveSessionProxy.access$000(HiveSessionProxy.java:37)
 at 
 org.apache.hive.service.cli.session.HiveSessionProxy$1.run(HiveSessionProxy.java:64)
 at java.security.AccessController.doPrivileged(Native Method)
 at javax.security.auth.Subject.doAs(Subject.java:415)
 at 
 org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1642)
 at 
 org.apache.hadoop.hive.shims.HadoopShimsSecure.doAs(HadoopShimsSecure.java:502)
 at 
 org.apache.hive.service.cli.session.HiveSessionProxy.invoke(HiveSessionProxy.java:60)
 at com.sun.proxy.$Proxy19.getSchemas(Unknown Source)
 at org.apache.hive.service.cli.CLIService.getSchemas(CLIService.java:277)
 at 
 

[jira] [Updated] (HIVE-6977) Delete Hiveserver1

2015-02-19 Thread Thejas M Nair (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6977?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Thejas M Nair updated HIVE-6977:

Fix Version/s: 1.0.0

 Delete Hiveserver1
 --

 Key: HIVE-6977
 URL: https://issues.apache.org/jira/browse/HIVE-6977
 Project: Hive
  Issue Type: Task
  Components: JDBC, Server Infrastructure
Reporter: Ashutosh Chauhan
Assignee: Ashutosh Chauhan
  Labels: TODOC15
 Fix For: 1.0.0, 1.1.0

 Attachments: HIVE-6977.1.patch, HIVE-6977.patch


 See mailing list discussion.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9718) Insert into dynamic partitions with same column structure in the distibute by clause barfs

2015-02-19 Thread Pavan Srinivas (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9718?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14327375#comment-14327375
 ] 

Pavan Srinivas commented on HIVE-9718:
--

Thanks for the comment. Yes, this optimization helps in preparing a better 
plan. Having said that, this bug breaks backward compatibility(this query works 
on Hive-0.13.1 and Hive-0.11). Ideal way is to fix this and make it compatible. 

 Insert into dynamic partitions with same column structure in the distibute 
 by clause barfs
 

 Key: HIVE-9718
 URL: https://issues.apache.org/jira/browse/HIVE-9718
 Project: Hive
  Issue Type: Bug
Affects Versions: 0.14.0, 1.0.0
Reporter: Pavan Srinivas
Priority: Critical
 Attachments: nation.tbl, patch.txt


 Sample reproducible query: 
 {code}
 SET hive.exec.dynamic.partition.mode=nonstrict;
 SET hive.exec.dynamic.partition=true;
  insert overwrite table nation_new_p partition (some)
 select n_name as name1, n_name as name2, n_name as name3 from nation 
 distribute by name3;
 {code}
 Note: Make sure there is data in the source table to reproduce the issue. 
 During the optimizations done for Jira: 
 https://issues.apache.org/jira/browse/HIVE-4867, an optimization of 
 deduplication of columns is done. But, when one of the columns is used as 
 part of partitioned/distribute by, its not taken care of.  
 The above query produces exception as follows:
 {code}
 Diagnostic Messages for this Task:
 java.lang.RuntimeException: org.apache.hadoop.hive.ql.metadata.HiveException: 
 Hive Runtime Error while processing row 
 {n_nationkey:0,n_name:ALGERIA,n_regionkey:0,n_comment: haggle. 
 carefully final deposits detect slyly agai}
   at org.apache.hadoop.hive.ql.exec.mr.ExecMapper.map(ExecMapper.java:185)
   at org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:54)
   at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:450)
   at org.apache.hadoop.mapred.MapTask.run(MapTask.java:343)
   at 
 org.apache.hadoop.mapred.LocalContainerLauncher$EventHandler.runSubtask(LocalContainerLauncher.java:370)
   at 
 org.apache.hadoop.mapred.LocalContainerLauncher$EventHandler.runTask(LocalContainerLauncher.java:295)
   at 
 org.apache.hadoop.mapred.LocalContainerLauncher$EventHandler.access$200(LocalContainerLauncher.java:181)
   at 
 org.apache.hadoop.mapred.LocalContainerLauncher$EventHandler$1.run(LocalContainerLauncher.java:224)
   at 
 java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471)
   at java.util.concurrent.FutureTask.run(FutureTask.java:262)
   at 
 java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
   at 
 java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
   at java.lang.Thread.run(Thread.java:744)
 Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime 
 Error while processing row 
 {n_nationkey:0,n_name:ALGERIA,n_regionkey:0,n_comment: haggle. 
 carefully final deposits detect slyly agai}
   at 
 org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:503)
   at org.apache.hadoop.hive.ql.exec.mr.ExecMapper.map(ExecMapper.java:176)
   ... 12 more
 Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: 
 java.lang.RuntimeException: cannot find field _col2 from [0:_col0]
   at 
 org.apache.hadoop.hive.ql.exec.ReduceSinkOperator.processOp(ReduceSinkOperator.java:397)
   at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:815)
   at 
 org.apache.hadoop.hive.ql.exec.SelectOperator.processOp(SelectOperator.java:84)
   at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:815)
   at 
 org.apache.hadoop.hive.ql.exec.TableScanOperator.processOp(TableScanOperator.java:95)
   at 
 org.apache.hadoop.hive.ql.exec.MapOperator$MapOpCtx.forward(MapOperator.java:157)
   at 
 org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:493)
   ... 13 more
 Caused by: java.lang.RuntimeException: cannot find field _col2 from [0:_col0]
   at 
 org.apache.hadoop.hive.serde2.objectinspector.ObjectInspectorUtils.getStandardStructFieldRef(ObjectInspectorUtils.java:410)
   at 
 org.apache.hadoop.hive.serde2.objectinspector.StandardStructObjectInspector.getStructFieldRef(StandardStructObjectInspector.java:147)
   at 
 org.apache.hadoop.hive.ql.exec.ExprNodeColumnEvaluator.initialize(ExprNodeColumnEvaluator.java:55)
   at 
 org.apache.hadoop.hive.ql.exec.Operator.initEvaluators(Operator.java:954)
   at 
 org.apache.hadoop.hive.ql.exec.ReduceSinkOperator.processOp(ReduceSinkOperator.java:325)
   ... 19 more
 {code}
 Tables used are: 
 {code}
 CREATE EXTERNAL TABLE `nation`(
   `n_nationkey` int,
   `n_name` 

[jira] [Commented] (HIVE-9718) Insert into dynamic partitions with same column structure in the distibute by clause barfs

2015-02-19 Thread Pavan Srinivas (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9718?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14327376#comment-14327376
 ] 

Pavan Srinivas commented on HIVE-9718:
--

Thanks for the comment. Yes, this optimization helps in preparing a better 
plan. Having said that, this bug breaks backward compatibility(this query works 
on Hive-0.13.1 and Hive-0.11). Ideal way is to fix this and make it compatible. 

 Insert into dynamic partitions with same column structure in the distibute 
 by clause barfs
 

 Key: HIVE-9718
 URL: https://issues.apache.org/jira/browse/HIVE-9718
 Project: Hive
  Issue Type: Bug
Affects Versions: 0.14.0, 1.0.0
Reporter: Pavan Srinivas
Priority: Critical
 Attachments: nation.tbl, patch.txt


 Sample reproducible query: 
 {code}
 SET hive.exec.dynamic.partition.mode=nonstrict;
 SET hive.exec.dynamic.partition=true;
  insert overwrite table nation_new_p partition (some)
 select n_name as name1, n_name as name2, n_name as name3 from nation 
 distribute by name3;
 {code}
 Note: Make sure there is data in the source table to reproduce the issue. 
 During the optimizations done for Jira: 
 https://issues.apache.org/jira/browse/HIVE-4867, an optimization of 
 deduplication of columns is done. But, when one of the columns is used as 
 part of partitioned/distribute by, its not taken care of.  
 The above query produces exception as follows:
 {code}
 Diagnostic Messages for this Task:
 java.lang.RuntimeException: org.apache.hadoop.hive.ql.metadata.HiveException: 
 Hive Runtime Error while processing row 
 {n_nationkey:0,n_name:ALGERIA,n_regionkey:0,n_comment: haggle. 
 carefully final deposits detect slyly agai}
   at org.apache.hadoop.hive.ql.exec.mr.ExecMapper.map(ExecMapper.java:185)
   at org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:54)
   at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:450)
   at org.apache.hadoop.mapred.MapTask.run(MapTask.java:343)
   at 
 org.apache.hadoop.mapred.LocalContainerLauncher$EventHandler.runSubtask(LocalContainerLauncher.java:370)
   at 
 org.apache.hadoop.mapred.LocalContainerLauncher$EventHandler.runTask(LocalContainerLauncher.java:295)
   at 
 org.apache.hadoop.mapred.LocalContainerLauncher$EventHandler.access$200(LocalContainerLauncher.java:181)
   at 
 org.apache.hadoop.mapred.LocalContainerLauncher$EventHandler$1.run(LocalContainerLauncher.java:224)
   at 
 java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471)
   at java.util.concurrent.FutureTask.run(FutureTask.java:262)
   at 
 java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
   at 
 java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
   at java.lang.Thread.run(Thread.java:744)
 Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime 
 Error while processing row 
 {n_nationkey:0,n_name:ALGERIA,n_regionkey:0,n_comment: haggle. 
 carefully final deposits detect slyly agai}
   at 
 org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:503)
   at org.apache.hadoop.hive.ql.exec.mr.ExecMapper.map(ExecMapper.java:176)
   ... 12 more
 Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: 
 java.lang.RuntimeException: cannot find field _col2 from [0:_col0]
   at 
 org.apache.hadoop.hive.ql.exec.ReduceSinkOperator.processOp(ReduceSinkOperator.java:397)
   at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:815)
   at 
 org.apache.hadoop.hive.ql.exec.SelectOperator.processOp(SelectOperator.java:84)
   at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:815)
   at 
 org.apache.hadoop.hive.ql.exec.TableScanOperator.processOp(TableScanOperator.java:95)
   at 
 org.apache.hadoop.hive.ql.exec.MapOperator$MapOpCtx.forward(MapOperator.java:157)
   at 
 org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:493)
   ... 13 more
 Caused by: java.lang.RuntimeException: cannot find field _col2 from [0:_col0]
   at 
 org.apache.hadoop.hive.serde2.objectinspector.ObjectInspectorUtils.getStandardStructFieldRef(ObjectInspectorUtils.java:410)
   at 
 org.apache.hadoop.hive.serde2.objectinspector.StandardStructObjectInspector.getStructFieldRef(StandardStructObjectInspector.java:147)
   at 
 org.apache.hadoop.hive.ql.exec.ExprNodeColumnEvaluator.initialize(ExprNodeColumnEvaluator.java:55)
   at 
 org.apache.hadoop.hive.ql.exec.Operator.initEvaluators(Operator.java:954)
   at 
 org.apache.hadoop.hive.ql.exec.ReduceSinkOperator.processOp(ReduceSinkOperator.java:325)
   ... 19 more
 {code}
 Tables used are: 
 {code}
 CREATE EXTERNAL TABLE `nation`(
   `n_nationkey` int,
   `n_name` 

[jira] [Commented] (HIVE-9718) Insert into dynamic partitions with same column structure in the distibute by clause barfs

2015-02-19 Thread Pavan Srinivas (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9718?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14327373#comment-14327373
 ] 

Pavan Srinivas commented on HIVE-9718:
--

Thanks for the comment. Yes, this optimization helps in preparing a better 
plan. Having said that, this bug breaks backward compatibility(this query works 
on Hive-0.13.1 and Hive-0.11). Ideal way is to fix this and make it compatible. 

 Insert into dynamic partitions with same column structure in the distibute 
 by clause barfs
 

 Key: HIVE-9718
 URL: https://issues.apache.org/jira/browse/HIVE-9718
 Project: Hive
  Issue Type: Bug
Affects Versions: 0.14.0, 1.0.0
Reporter: Pavan Srinivas
Priority: Critical
 Attachments: nation.tbl, patch.txt


 Sample reproducible query: 
 {code}
 SET hive.exec.dynamic.partition.mode=nonstrict;
 SET hive.exec.dynamic.partition=true;
  insert overwrite table nation_new_p partition (some)
 select n_name as name1, n_name as name2, n_name as name3 from nation 
 distribute by name3;
 {code}
 Note: Make sure there is data in the source table to reproduce the issue. 
 During the optimizations done for Jira: 
 https://issues.apache.org/jira/browse/HIVE-4867, an optimization of 
 deduplication of columns is done. But, when one of the columns is used as 
 part of partitioned/distribute by, its not taken care of.  
 The above query produces exception as follows:
 {code}
 Diagnostic Messages for this Task:
 java.lang.RuntimeException: org.apache.hadoop.hive.ql.metadata.HiveException: 
 Hive Runtime Error while processing row 
 {n_nationkey:0,n_name:ALGERIA,n_regionkey:0,n_comment: haggle. 
 carefully final deposits detect slyly agai}
   at org.apache.hadoop.hive.ql.exec.mr.ExecMapper.map(ExecMapper.java:185)
   at org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:54)
   at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:450)
   at org.apache.hadoop.mapred.MapTask.run(MapTask.java:343)
   at 
 org.apache.hadoop.mapred.LocalContainerLauncher$EventHandler.runSubtask(LocalContainerLauncher.java:370)
   at 
 org.apache.hadoop.mapred.LocalContainerLauncher$EventHandler.runTask(LocalContainerLauncher.java:295)
   at 
 org.apache.hadoop.mapred.LocalContainerLauncher$EventHandler.access$200(LocalContainerLauncher.java:181)
   at 
 org.apache.hadoop.mapred.LocalContainerLauncher$EventHandler$1.run(LocalContainerLauncher.java:224)
   at 
 java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471)
   at java.util.concurrent.FutureTask.run(FutureTask.java:262)
   at 
 java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
   at 
 java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
   at java.lang.Thread.run(Thread.java:744)
 Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime 
 Error while processing row 
 {n_nationkey:0,n_name:ALGERIA,n_regionkey:0,n_comment: haggle. 
 carefully final deposits detect slyly agai}
   at 
 org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:503)
   at org.apache.hadoop.hive.ql.exec.mr.ExecMapper.map(ExecMapper.java:176)
   ... 12 more
 Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: 
 java.lang.RuntimeException: cannot find field _col2 from [0:_col0]
   at 
 org.apache.hadoop.hive.ql.exec.ReduceSinkOperator.processOp(ReduceSinkOperator.java:397)
   at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:815)
   at 
 org.apache.hadoop.hive.ql.exec.SelectOperator.processOp(SelectOperator.java:84)
   at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:815)
   at 
 org.apache.hadoop.hive.ql.exec.TableScanOperator.processOp(TableScanOperator.java:95)
   at 
 org.apache.hadoop.hive.ql.exec.MapOperator$MapOpCtx.forward(MapOperator.java:157)
   at 
 org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:493)
   ... 13 more
 Caused by: java.lang.RuntimeException: cannot find field _col2 from [0:_col0]
   at 
 org.apache.hadoop.hive.serde2.objectinspector.ObjectInspectorUtils.getStandardStructFieldRef(ObjectInspectorUtils.java:410)
   at 
 org.apache.hadoop.hive.serde2.objectinspector.StandardStructObjectInspector.getStructFieldRef(StandardStructObjectInspector.java:147)
   at 
 org.apache.hadoop.hive.ql.exec.ExprNodeColumnEvaluator.initialize(ExprNodeColumnEvaluator.java:55)
   at 
 org.apache.hadoop.hive.ql.exec.Operator.initEvaluators(Operator.java:954)
   at 
 org.apache.hadoop.hive.ql.exec.ReduceSinkOperator.processOp(ReduceSinkOperator.java:325)
   ... 19 more
 {code}
 Tables used are: 
 {code}
 CREATE EXTERNAL TABLE `nation`(
   `n_nationkey` int,
   `n_name` 

[jira] [Commented] (HIVE-9718) Insert into dynamic partitions with same column structure in the distibute by clause barfs

2015-02-19 Thread Pavan Srinivas (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9718?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14327370#comment-14327370
 ] 

Pavan Srinivas commented on HIVE-9718:
--

Thanks for the comment. Yes, this optimization helps in preparing a better 
plan. Having said that, this bug breaks backward compatibility(this query works 
on Hive-0.13.1 and Hive-0.11). Ideal way is to fix this and make it compatible. 

 Insert into dynamic partitions with same column structure in the distibute 
 by clause barfs
 

 Key: HIVE-9718
 URL: https://issues.apache.org/jira/browse/HIVE-9718
 Project: Hive
  Issue Type: Bug
Affects Versions: 0.14.0, 1.0.0
Reporter: Pavan Srinivas
Priority: Critical
 Attachments: nation.tbl, patch.txt


 Sample reproducible query: 
 {code}
 SET hive.exec.dynamic.partition.mode=nonstrict;
 SET hive.exec.dynamic.partition=true;
  insert overwrite table nation_new_p partition (some)
 select n_name as name1, n_name as name2, n_name as name3 from nation 
 distribute by name3;
 {code}
 Note: Make sure there is data in the source table to reproduce the issue. 
 During the optimizations done for Jira: 
 https://issues.apache.org/jira/browse/HIVE-4867, an optimization of 
 deduplication of columns is done. But, when one of the columns is used as 
 part of partitioned/distribute by, its not taken care of.  
 The above query produces exception as follows:
 {code}
 Diagnostic Messages for this Task:
 java.lang.RuntimeException: org.apache.hadoop.hive.ql.metadata.HiveException: 
 Hive Runtime Error while processing row 
 {n_nationkey:0,n_name:ALGERIA,n_regionkey:0,n_comment: haggle. 
 carefully final deposits detect slyly agai}
   at org.apache.hadoop.hive.ql.exec.mr.ExecMapper.map(ExecMapper.java:185)
   at org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:54)
   at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:450)
   at org.apache.hadoop.mapred.MapTask.run(MapTask.java:343)
   at 
 org.apache.hadoop.mapred.LocalContainerLauncher$EventHandler.runSubtask(LocalContainerLauncher.java:370)
   at 
 org.apache.hadoop.mapred.LocalContainerLauncher$EventHandler.runTask(LocalContainerLauncher.java:295)
   at 
 org.apache.hadoop.mapred.LocalContainerLauncher$EventHandler.access$200(LocalContainerLauncher.java:181)
   at 
 org.apache.hadoop.mapred.LocalContainerLauncher$EventHandler$1.run(LocalContainerLauncher.java:224)
   at 
 java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471)
   at java.util.concurrent.FutureTask.run(FutureTask.java:262)
   at 
 java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
   at 
 java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
   at java.lang.Thread.run(Thread.java:744)
 Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime 
 Error while processing row 
 {n_nationkey:0,n_name:ALGERIA,n_regionkey:0,n_comment: haggle. 
 carefully final deposits detect slyly agai}
   at 
 org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:503)
   at org.apache.hadoop.hive.ql.exec.mr.ExecMapper.map(ExecMapper.java:176)
   ... 12 more
 Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: 
 java.lang.RuntimeException: cannot find field _col2 from [0:_col0]
   at 
 org.apache.hadoop.hive.ql.exec.ReduceSinkOperator.processOp(ReduceSinkOperator.java:397)
   at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:815)
   at 
 org.apache.hadoop.hive.ql.exec.SelectOperator.processOp(SelectOperator.java:84)
   at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:815)
   at 
 org.apache.hadoop.hive.ql.exec.TableScanOperator.processOp(TableScanOperator.java:95)
   at 
 org.apache.hadoop.hive.ql.exec.MapOperator$MapOpCtx.forward(MapOperator.java:157)
   at 
 org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:493)
   ... 13 more
 Caused by: java.lang.RuntimeException: cannot find field _col2 from [0:_col0]
   at 
 org.apache.hadoop.hive.serde2.objectinspector.ObjectInspectorUtils.getStandardStructFieldRef(ObjectInspectorUtils.java:410)
   at 
 org.apache.hadoop.hive.serde2.objectinspector.StandardStructObjectInspector.getStructFieldRef(StandardStructObjectInspector.java:147)
   at 
 org.apache.hadoop.hive.ql.exec.ExprNodeColumnEvaluator.initialize(ExprNodeColumnEvaluator.java:55)
   at 
 org.apache.hadoop.hive.ql.exec.Operator.initEvaluators(Operator.java:954)
   at 
 org.apache.hadoop.hive.ql.exec.ReduceSinkOperator.processOp(ReduceSinkOperator.java:325)
   ... 19 more
 {code}
 Tables used are: 
 {code}
 CREATE EXTERNAL TABLE `nation`(
   `n_nationkey` int,
   `n_name` 

[jira] [Commented] (HIVE-6977) Delete Hiveserver1

2015-02-19 Thread Thejas M Nair (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6977?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14327439#comment-14327439
 ] 

Thejas M Nair commented on HIVE-6977:
-

This is also part of 1.0.0 release.



 Delete Hiveserver1
 --

 Key: HIVE-6977
 URL: https://issues.apache.org/jira/browse/HIVE-6977
 Project: Hive
  Issue Type: Task
  Components: JDBC, Server Infrastructure
Reporter: Ashutosh Chauhan
Assignee: Ashutosh Chauhan
  Labels: TODOC15
 Fix For: 1.0.0, 1.1.0

 Attachments: HIVE-6977.1.patch, HIVE-6977.patch


 See mailing list discussion.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9581) CBO (Calcite Return Path): Translate Join to Hive Op [CBO branch]

2015-02-19 Thread Jesus Camacho Rodriguez (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9581?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jesus Camacho Rodriguez updated HIVE-9581:
--
Attachment: HIVE-9581.cbo.patch

 CBO (Calcite Return Path): Translate Join to Hive Op [CBO branch]
 -

 Key: HIVE-9581
 URL: https://issues.apache.org/jira/browse/HIVE-9581
 Project: Hive
  Issue Type: Sub-task
  Components: CBO
Affects Versions: cbo-branch
Reporter: Jesus Camacho Rodriguez
Assignee: Jesus Camacho Rodriguez
 Fix For: 1.2.0

 Attachments: HIVE-9581.cbo.patch






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9718) Insert into dynamic partitions with same column structure in the distibute by clause barfs

2015-02-19 Thread Pavan Srinivas (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9718?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14327384#comment-14327384
 ] 

Pavan Srinivas commented on HIVE-9718:
--

Apologies for the spam. Some constancy issue with Jira, made me to add the same 
comment again and again. 

 Insert into dynamic partitions with same column structure in the distibute 
 by clause barfs
 

 Key: HIVE-9718
 URL: https://issues.apache.org/jira/browse/HIVE-9718
 Project: Hive
  Issue Type: Bug
Affects Versions: 0.14.0, 1.0.0
Reporter: Pavan Srinivas
Priority: Critical
 Attachments: nation.tbl, patch.txt


 Sample reproducible query: 
 {code}
 SET hive.exec.dynamic.partition.mode=nonstrict;
 SET hive.exec.dynamic.partition=true;
  insert overwrite table nation_new_p partition (some)
 select n_name as name1, n_name as name2, n_name as name3 from nation 
 distribute by name3;
 {code}
 Note: Make sure there is data in the source table to reproduce the issue. 
 During the optimizations done for Jira: 
 https://issues.apache.org/jira/browse/HIVE-4867, an optimization of 
 deduplication of columns is done. But, when one of the columns is used as 
 part of partitioned/distribute by, its not taken care of.  
 The above query produces exception as follows:
 {code}
 Diagnostic Messages for this Task:
 java.lang.RuntimeException: org.apache.hadoop.hive.ql.metadata.HiveException: 
 Hive Runtime Error while processing row 
 {n_nationkey:0,n_name:ALGERIA,n_regionkey:0,n_comment: haggle. 
 carefully final deposits detect slyly agai}
   at org.apache.hadoop.hive.ql.exec.mr.ExecMapper.map(ExecMapper.java:185)
   at org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:54)
   at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:450)
   at org.apache.hadoop.mapred.MapTask.run(MapTask.java:343)
   at 
 org.apache.hadoop.mapred.LocalContainerLauncher$EventHandler.runSubtask(LocalContainerLauncher.java:370)
   at 
 org.apache.hadoop.mapred.LocalContainerLauncher$EventHandler.runTask(LocalContainerLauncher.java:295)
   at 
 org.apache.hadoop.mapred.LocalContainerLauncher$EventHandler.access$200(LocalContainerLauncher.java:181)
   at 
 org.apache.hadoop.mapred.LocalContainerLauncher$EventHandler$1.run(LocalContainerLauncher.java:224)
   at 
 java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471)
   at java.util.concurrent.FutureTask.run(FutureTask.java:262)
   at 
 java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
   at 
 java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
   at java.lang.Thread.run(Thread.java:744)
 Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime 
 Error while processing row 
 {n_nationkey:0,n_name:ALGERIA,n_regionkey:0,n_comment: haggle. 
 carefully final deposits detect slyly agai}
   at 
 org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:503)
   at org.apache.hadoop.hive.ql.exec.mr.ExecMapper.map(ExecMapper.java:176)
   ... 12 more
 Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: 
 java.lang.RuntimeException: cannot find field _col2 from [0:_col0]
   at 
 org.apache.hadoop.hive.ql.exec.ReduceSinkOperator.processOp(ReduceSinkOperator.java:397)
   at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:815)
   at 
 org.apache.hadoop.hive.ql.exec.SelectOperator.processOp(SelectOperator.java:84)
   at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:815)
   at 
 org.apache.hadoop.hive.ql.exec.TableScanOperator.processOp(TableScanOperator.java:95)
   at 
 org.apache.hadoop.hive.ql.exec.MapOperator$MapOpCtx.forward(MapOperator.java:157)
   at 
 org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:493)
   ... 13 more
 Caused by: java.lang.RuntimeException: cannot find field _col2 from [0:_col0]
   at 
 org.apache.hadoop.hive.serde2.objectinspector.ObjectInspectorUtils.getStandardStructFieldRef(ObjectInspectorUtils.java:410)
   at 
 org.apache.hadoop.hive.serde2.objectinspector.StandardStructObjectInspector.getStructFieldRef(StandardStructObjectInspector.java:147)
   at 
 org.apache.hadoop.hive.ql.exec.ExprNodeColumnEvaluator.initialize(ExprNodeColumnEvaluator.java:55)
   at 
 org.apache.hadoop.hive.ql.exec.Operator.initEvaluators(Operator.java:954)
   at 
 org.apache.hadoop.hive.ql.exec.ReduceSinkOperator.processOp(ReduceSinkOperator.java:325)
   ... 19 more
 {code}
 Tables used are: 
 {code}
 CREATE EXTERNAL TABLE `nation`(
   `n_nationkey` int,
   `n_name` string,
   `n_regionkey` int,
   `n_comment` string)
 ROW FORMAT DELIMITED
   FIELDS TERMINATED BY '|'
 STORED AS INPUTFORMAT
   

[jira] [Commented] (HIVE-8807) Obsolete default values in webhcat-default.xml

2015-02-19 Thread Thejas M Nair (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8807?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14327441#comment-14327441
 ] 

Thejas M Nair commented on HIVE-8807:
-

We can add instructions to update this as part of release instructions - 
https://cwiki.apache.org/confluence/display/Hive/HowToRelease


 Obsolete default values in webhcat-default.xml
 --

 Key: HIVE-8807
 URL: https://issues.apache.org/jira/browse/HIVE-8807
 Project: Hive
  Issue Type: Bug
  Components: WebHCat
Affects Versions: 0.12.0, 0.13.0, 0.14.0
Reporter: Lefty Leverenz
Assignee: Eugene Koifman
 Fix For: 1.0.0

 Attachments: HIVE8807.patch


 The defaults for templeton.pig.path  templeton.hive.path are 0.11 in 
 webhcat-default.xml but they ought to match current release numbers.
 The Pig version is 0.12.0 for Hive 0.14 RC0 (as shown in pom.xml).
 no precommit tests



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9537) string expressions on a fixed length character do not preserve trailing spaces

2015-02-19 Thread N Campbell (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9537?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14327408#comment-14327408
 ] 

N Campbell commented on HIVE-9537:
--

I wouldn't suggest that use MySQL and Postgres alone is ideal to cross check 
SQL semantics. 

Should the Apache implementation for CHAR(n) stay this way going forward I 
recommend you get the documentation improved. 

Having seen companies migrate applications across vendors where this sort of 
thing burns them. Many may not know what ISO-SQL 20xx states (let alone care) 
and will presume that their source vendor(s) have followed the specification 
etc. 

Given the server side string operations etc, they may as well stay with VARCHAR 
and know that trailing spaces are not preserved vs the 'appears' to have spaces 
when projected but not in other cases. 

See discussion: 
http://www.postgresql.org/docs/9.1/static/datatype-character.html

See discussion of MySQL JDBC: 
https://dev.mysql.com/doc/connector-j/en/connector-j-reference-configuration-properties.html
 re padCharsWithSpace

If you apply a fold (upper/lower) the type of the result is supposed to be 
from the character value expression

Length should be the number of characters in the string value expression - 
where number of  characters is based upon the semantics of the character set of 
the string value expression

If both operands are fixed length character strings, concatenation result is a 
fixed length character string with a length equal to the sum of the lengths of 
the operands where the length cannot exceed the maximum allowed for a fixed 
length character string.

Cheers.

 string expressions on a fixed length character do not preserve trailing spaces
 --

 Key: HIVE-9537
 URL: https://issues.apache.org/jira/browse/HIVE-9537
 Project: Hive
  Issue Type: Bug
  Components: SQL
Reporter: N Campbell
Assignee: Aihua Xu

 When a string expression such as upper or lower is applied to a fixed length 
 column the trailing spaces of the fixed length character are not preserved.
 {code:sql}
 CREATE TABLE  if not exists TCHAR ( 
 RNUM int, 
 CCHAR char(32)
 )
 ROW FORMAT DELIMITED 
 FIELDS TERMINATED BY '|' 
 LINES TERMINATED BY '\n' 
 STORED AS TEXTFILE;
 {code}
 {{cchar}} as a {{char(32)}}.
 {code:sql}
 select cchar, concat(cchar, cchar), concat(lower(cchar), cchar), 
 concat(upper(cchar), cchar) 
 from tchar;
 {code}
 0|\N
 1|
 2| 
 3|BB
 4|EE
 5|FF



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9388) HiveServer2 fails to reconnect to MetaStore after MetaStore restart

2015-02-19 Thread Mariusz Strzelecki (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9388?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mariusz Strzelecki updated HIVE-9388:
-
Attachment: HIVE-9388.2.patch

Patch prepared for current trunk.

 HiveServer2 fails to reconnect to MetaStore after MetaStore restart
 ---

 Key: HIVE-9388
 URL: https://issues.apache.org/jira/browse/HIVE-9388
 Project: Hive
  Issue Type: Bug
  Components: HiveServer2
Affects Versions: 0.12.0, 0.14.0, 0.13.1
Reporter: Piotr Ackermann
 Attachments: HIVE-9388.2.patch, HIVE-9388.patch


 How to reproduce:
 # Use Hue to connect to HiveServer2
 # Restart Metastore
 # Try to execute any query in Hue
 HiveServer2 report error:
 {quote}
 ERROR hive.log: Got exception: 
 org.apache.thrift.transport.TTransportException null
 org.apache.thrift.transport.TTransportException
 at 
 org.apache.thrift.transport.TIOStreamTransport.read(TIOStreamTransport.java:132)
 at org.apache.thrift.transport.TTransport.readAll(TTransport.java:84)
 at 
 org.apache.thrift.transport.TSaslTransport.readLength(TSaslTransport.java:355)
 at 
 org.apache.thrift.transport.TSaslTransport.readFrame(TSaslTransport.java:432)
 at org.apache.thrift.transport.TSaslTransport.read(TSaslTransport.java:414)
 at 
 org.apache.thrift.transport.TSaslClientTransport.read(TSaslClientTransport.java:37)
 at org.apache.thrift.transport.TTransport.readAll(TTransport.java:84)
 at 
 org.apache.hadoop.hive.thrift.TFilterTransport.readAll(TFilterTransport.java:62)
 at 
 org.apache.thrift.protocol.TBinaryProtocol.readAll(TBinaryProtocol.java:378)
 at 
 org.apache.thrift.protocol.TBinaryProtocol.readI32(TBinaryProtocol.java:297)
 at 
 org.apache.thrift.protocol.TBinaryProtocol.readMessageBegin(TBinaryProtocol.java:204)
 at org.apache.thrift.TServiceClient.receiveBase(TServiceClient.java:69)
 at 
 org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.recv_get_databases(ThriftHiveMetastore.java:600)
 at 
 org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.get_databases(ThriftHiveMetastore.java:587)
 at 
 org.apache.hadoop.hive.metastore.HiveMetaStoreClient.getDatabases(HiveMetaStoreClient.java:837)
 at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
 at 
 sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
 at 
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
 at java.lang.reflect.Method.invoke(Method.java:606)
 at 
 org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.invoke(RetryingMetaStoreClient.java:90)
 at com.sun.proxy.$Proxy10.getDatabases(Unknown Source)
 at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
 at 
 sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
 at 
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
 at java.lang.reflect.Method.invoke(Method.java:606)
 at 
 org.apache.hadoop.hive.metastore.HiveMetaStoreClient$SynchronizedHandler.invoke(HiveMetaStoreClient.java:1681)
 at com.sun.proxy.$Proxy10.getDatabases(Unknown Source)
 at 
 org.apache.hive.service.cli.operation.GetSchemasOperation.run(GetSchemasOperation.java:62)
 at 
 org.apache.hive.service.cli.session.HiveSessionImpl.runOperationWithLogCapture(HiveSessionImpl.java:715)
 at 
 org.apache.hive.service.cli.session.HiveSessionImpl.getSchemas(HiveSessionImpl.java:438)
 at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
 at 
 sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
 at 
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
 at java.lang.reflect.Method.invoke(Method.java:606)
 at 
 org.apache.hive.service.cli.session.HiveSessionProxy.invoke(HiveSessionProxy.java:79)
 at 
 org.apache.hive.service.cli.session.HiveSessionProxy.access$000(HiveSessionProxy.java:37)
 at 
 org.apache.hive.service.cli.session.HiveSessionProxy$1.run(HiveSessionProxy.java:64)
 at java.security.AccessController.doPrivileged(Native Method)
 at javax.security.auth.Subject.doAs(Subject.java:415)
 at 
 org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1642)
 at 
 org.apache.hadoop.hive.shims.HadoopShimsSecure.doAs(HadoopShimsSecure.java:502)
 at 
 org.apache.hive.service.cli.session.HiveSessionProxy.invoke(HiveSessionProxy.java:60)
 at com.sun.proxy.$Proxy19.getSchemas(Unknown Source)
 at org.apache.hive.service.cli.CLIService.getSchemas(CLIService.java:277)
 at 
 org.apache.hive.service.cli.thrift.ThriftCLIService.GetSchemas(ThriftCLIService.java:436)
 at 
 org.apache.hive.service.cli.thrift.TCLIService$Processor$GetSchemas.getResult(TCLIService.java:1433)
 at 
 org.apache.hive.service.cli.thrift.TCLIService$Processor$GetSchemas.getResult(TCLIService.java:1418)
 at 

[jira] [Created] (HIVE-9725) Need to add indices, privileges, and functions to HBaseImport [hbase-metastore branch]

2015-02-19 Thread Alan Gates (JIRA)
Alan Gates created HIVE-9725:


 Summary: Need to add indices, privileges, and functions to 
HBaseImport [hbase-metastore branch]
 Key: HIVE-9725
 URL: https://issues.apache.org/jira/browse/HIVE-9725
 Project: Hive
  Issue Type: Sub-task
  Components: Metastore
Reporter: Alan Gates
Assignee: Alan Gates


HBaseImport doesn't include these objects because they weren't supported in the 
metastore yet when it was created.  These need to be added.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


Fix version for hbase-metastore branch

2015-02-19 Thread Alan Gates
I've been marking JIRAs on this branch as fixed in 1.2, since that's the 
next version.  But that seems wrong as I doubt this code will be in by 
1.2.  What's the usual practice here?  It seems it would make sense to 
make a label for this branch and mark them as fixed with that label and 
then when we actually release this in a version we can update all the 
JIRAs with that label.


Alan.


[jira] [Resolved] (HIVE-9581) CBO (Calcite Return Path): Translate Join to Hive Op [CBO branch]

2015-02-19 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9581?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan resolved HIVE-9581.

   Resolution: Fixed
Fix Version/s: (was: 1.2.0)
   cbo-branch

Committed to branch. Thanks, Jesus!

 CBO (Calcite Return Path): Translate Join to Hive Op [CBO branch]
 -

 Key: HIVE-9581
 URL: https://issues.apache.org/jira/browse/HIVE-9581
 Project: Hive
  Issue Type: Sub-task
  Components: CBO
Affects Versions: cbo-branch
Reporter: Jesus Camacho Rodriguez
Assignee: Jesus Camacho Rodriguez
 Fix For: cbo-branch

 Attachments: HIVE-9581.cbo.patch






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9606) Need a tool to export metadata from RDBMS based metastore into HBase

2015-02-19 Thread Alan Gates (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9606?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alan Gates updated HIVE-9606:
-
   Resolution: Fixed
Fix Version/s: 1.2.0
   Status: Resolved  (was: Patch Available)

Patch 2 checked in, with added comments suggested by Thejas.

 Need a tool to export metadata from RDBMS based metastore into HBase
 

 Key: HIVE-9606
 URL: https://issues.apache.org/jira/browse/HIVE-9606
 Project: Hive
  Issue Type: Sub-task
  Components: Metastore
Reporter: Alan Gates
Assignee: Alan Gates
 Fix For: 1.2.0

 Attachments: HIVE-9606.2.patch, HIVE-9606.patch


 For testing (and eventually for end user use) we need a tool that can take 
 data from an existing RDBMS based metastore and create the corresponding 
 objects in an HBase based metastore.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HIVE-9726) Upgrade to spark 1.3

2015-02-19 Thread Brock Noland (JIRA)
Brock Noland created HIVE-9726:
--

 Summary: Upgrade to spark 1.3
 Key: HIVE-9726
 URL: https://issues.apache.org/jira/browse/HIVE-9726
 Project: Hive
  Issue Type: Sub-task
Reporter: Brock Noland






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9726) Upgrade to spark 1.3 [Spark Branch]

2015-02-19 Thread Brock Noland (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9726?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Brock Noland updated HIVE-9726:
---
Affects Version/s: spark-branch
  Summary: Upgrade to spark 1.3 [Spark Branch]  (was: Upgrade to 
spark 1.3)

 Upgrade to spark 1.3 [Spark Branch]
 ---

 Key: HIVE-9726
 URL: https://issues.apache.org/jira/browse/HIVE-9726
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Affects Versions: spark-branch
Reporter: Brock Noland





--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9671) Support Impersonation [Spark Branch]

2015-02-19 Thread Brock Noland (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9671?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Brock Noland updated HIVE-9671:
---
Attachment: HIVE-9671.1-spark.patch

 Support Impersonation [Spark Branch]
 

 Key: HIVE-9671
 URL: https://issues.apache.org/jira/browse/HIVE-9671
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Reporter: Brock Noland
 Attachments: HIVE-9671.1-spark.patch


 SPARK-5493 in 1.3 implemented proxy user authentication. We need to implement 
 using this option in spark client.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9671) Support Impersonation [Spark Branch]

2015-02-19 Thread Brock Noland (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9671?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Brock Noland updated HIVE-9671:
---
Assignee: Brock Noland
  Status: Patch Available  (was: Open)

 Support Impersonation [Spark Branch]
 

 Key: HIVE-9671
 URL: https://issues.apache.org/jira/browse/HIVE-9671
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Reporter: Brock Noland
Assignee: Brock Noland
 Attachments: HIVE-9671.1-spark.patch


 SPARK-5493 in 1.3 implemented proxy user authentication. We need to implement 
 using this option in spark client.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HIVE-9727) GroupingID translation from Calcite

2015-02-19 Thread Jesus Camacho Rodriguez (JIRA)
Jesus Camacho Rodriguez created HIVE-9727:
-

 Summary: GroupingID translation from Calcite
 Key: HIVE-9727
 URL: https://issues.apache.org/jira/browse/HIVE-9727
 Project: Hive
  Issue Type: Bug
Reporter: Jesus Camacho Rodriguez
Assignee: Jesus Camacho Rodriguez


The translation from Calcite back to Hive might produce wrong results while 
interacting with other Calcite optimization rules.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9727) GroupingID translation from Calcite

2015-02-19 Thread Jesus Camacho Rodriguez (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9727?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jesus Camacho Rodriguez updated HIVE-9727:
--
Attachment: HIVE-9727.patch

[~jpullokkaran]

 GroupingID translation from Calcite
 ---

 Key: HIVE-9727
 URL: https://issues.apache.org/jira/browse/HIVE-9727
 Project: Hive
  Issue Type: Bug
Reporter: Jesus Camacho Rodriguez
Assignee: Jesus Camacho Rodriguez
 Attachments: HIVE-9727.patch


 The translation from Calcite back to Hive might produce wrong results while 
 interacting with other Calcite optimization rules.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Work started] (HIVE-9727) GroupingID translation from Calcite

2015-02-19 Thread Jesus Camacho Rodriguez (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9727?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Work on HIVE-9727 started by Jesus Camacho Rodriguez.
-
 GroupingID translation from Calcite
 ---

 Key: HIVE-9727
 URL: https://issues.apache.org/jira/browse/HIVE-9727
 Project: Hive
  Issue Type: Bug
Reporter: Jesus Camacho Rodriguez
Assignee: Jesus Camacho Rodriguez

 The translation from Calcite back to Hive might produce wrong results while 
 interacting with other Calcite optimization rules.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HIVE-9728) LLAP: add heap mode to allocator (for q files, YARN w/o direct buffer accounting support)

2015-02-19 Thread Sergey Shelukhin (JIRA)
Sergey Shelukhin created HIVE-9728:
--

 Summary: LLAP: add heap mode to allocator (for q files, YARN w/o 
direct buffer accounting support)
 Key: HIVE-9728
 URL: https://issues.apache.org/jira/browse/HIVE-9728
 Project: Hive
  Issue Type: Sub-task
Reporter: Sergey Shelukhin
Assignee: Sergey Shelukhin






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9727) GroupingID translation from Calcite

2015-02-19 Thread Jesus Camacho Rodriguez (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9727?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jesus Camacho Rodriguez updated HIVE-9727:
--
Status: Patch Available  (was: In Progress)

 GroupingID translation from Calcite
 ---

 Key: HIVE-9727
 URL: https://issues.apache.org/jira/browse/HIVE-9727
 Project: Hive
  Issue Type: Bug
Reporter: Jesus Camacho Rodriguez
Assignee: Jesus Camacho Rodriguez
 Attachments: HIVE-9727.patch


 The translation from Calcite back to Hive might produce wrong results while 
 interacting with other Calcite optimization rules.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


Review Request 31209: HIVE-9727

2015-02-19 Thread Jesús Camacho Rodríguez

---
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/31209/
---

Review request for hive and John Pullokkaran.


Bugs: HIVE-9727
https://issues.apache.org/jira/browse/HIVE-9727


Repository: hive-git


Description
---

GroupingID translation from Calcite


Diffs
-

  
ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/reloperators/HiveGroupingID.java
 345b64af8514466c84e9899e9c019b679b761ba6 
  
ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/translator/ASTConverter.java
 ea5918110fa1255f105c646c08e7d307afb3f94b 

Diff: https://reviews.apache.org/r/31209/diff/


Testing
---

Existing tests (groupby*.q)


Thanks,

Jesús Camacho Rodríguez



[jira] [Created] (HIVE-9729) LLAP: design and implement proper metadata cache

2015-02-19 Thread Sergey Shelukhin (JIRA)
Sergey Shelukhin created HIVE-9729:
--

 Summary: LLAP: design and implement proper metadata cache
 Key: HIVE-9729
 URL: https://issues.apache.org/jira/browse/HIVE-9729
 Project: Hive
  Issue Type: Sub-task
Reporter: Sergey Shelukhin


Simple approach: add external priorities to data cache, read metadata parts of 
orc file into it. Advantage: simple; consistent management (no need to 
coordinate sizes and eviction between data and metadata caches, etc); 
disadvantage - have to decode every time.
Maybe add decoded metadata cache on top - fixed size, small and opportunistic? 
Or some other approach.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HIVE-9730) LLAP: make sure logging is never called when not needed

2015-02-19 Thread Sergey Shelukhin (JIRA)
Sergey Shelukhin created HIVE-9730:
--

 Summary: LLAP: make sure logging is never called when not needed
 Key: HIVE-9730
 URL: https://issues.apache.org/jira/browse/HIVE-9730
 Project: Hive
  Issue Type: Sub-task
Reporter: Sergey Shelukhin
Assignee: Sergey Shelukhin


log4j logging has really inefficient serialization



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HIVE-9731) WebHCat MapReduce Streaming Job does not allow StreamXmlRecordReader to be specified

2015-02-19 Thread Hari Sankar Sivarama Subramaniyan (JIRA)
Hari Sankar Sivarama Subramaniyan created HIVE-9731:
---

 Summary: WebHCat MapReduce Streaming Job does not allow 
StreamXmlRecordReader to be specified
 Key: HIVE-9731
 URL: https://issues.apache.org/jira/browse/HIVE-9731
 Project: Hive
  Issue Type: Bug
  Components: WebHCat
Reporter: Hari Sankar Sivarama Subramaniyan
Assignee: Hari Sankar Sivarama Subramaniyan


Hadoop Streaming allows -inputreader parameter to specify use of 
StreamXmlRecorderReader
(example)
hadoop jar hadoop-streaming-2.5.1.jar \
-inputreader StreamXmlRecord,begin=BEGIN_STRING,end=END_STRING \
(rest of the command)
WebHCat's StreamingDelegator does not include -inputreader as a valid option 
when submitting jobs to 
http://www.myserver.com/templeton/v1/mapreduce/streaming endpoint. If 
-inputreader is specified and passed to templeton server (perhaps via CURL 
operation) , it will get truncated and not passed as parameter from 
TempletonControllerJob to Hadoop.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9731) WebHCat MapReduce Streaming Job does not allow StreamXmlRecordReader to be specified

2015-02-19 Thread Hari Sankar Sivarama Subramaniyan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9731?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Hari Sankar Sivarama Subramaniyan updated HIVE-9731:

Status: Patch Available  (was: Open)

 WebHCat MapReduce Streaming Job does not allow StreamXmlRecordReader to be 
 specified
 

 Key: HIVE-9731
 URL: https://issues.apache.org/jira/browse/HIVE-9731
 Project: Hive
  Issue Type: Bug
  Components: WebHCat
Reporter: Hari Sankar Sivarama Subramaniyan
Assignee: Hari Sankar Sivarama Subramaniyan
 Attachments: HIVE-9731.1.patch


 Hadoop Streaming allows -inputreader parameter to specify use of 
 StreamXmlRecorderReader
 (example)
 hadoop jar hadoop-streaming-2.5.1.jar \
 -inputreader StreamXmlRecord,begin=BEGIN_STRING,end=END_STRING \
 (rest of the command)
 WebHCat's StreamingDelegator does not include -inputreader as a valid option 
 when submitting jobs to 
 http://www.myserver.com/templeton/v1/mapreduce/streaming endpoint. If 
 -inputreader is specified and passed to templeton server (perhaps via CURL 
 operation) , it will get truncated and not passed as parameter from 
 TempletonControllerJob to Hadoop.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9731) WebHCat MapReduce Streaming Job does not allow StreamXmlRecordReader to be specified

2015-02-19 Thread Hari Sankar Sivarama Subramaniyan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9731?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Hari Sankar Sivarama Subramaniyan updated HIVE-9731:

Attachment: HIVE-9731.1.patch

[~ekoifman] / [~thejas] Can you please review the change.

Thanks
Hari

 WebHCat MapReduce Streaming Job does not allow StreamXmlRecordReader to be 
 specified
 

 Key: HIVE-9731
 URL: https://issues.apache.org/jira/browse/HIVE-9731
 Project: Hive
  Issue Type: Bug
  Components: WebHCat
Reporter: Hari Sankar Sivarama Subramaniyan
Assignee: Hari Sankar Sivarama Subramaniyan
 Attachments: HIVE-9731.1.patch


 Hadoop Streaming allows -inputreader parameter to specify use of 
 StreamXmlRecorderReader
 (example)
 hadoop jar hadoop-streaming-2.5.1.jar \
 -inputreader StreamXmlRecord,begin=BEGIN_STRING,end=END_STRING \
 (rest of the command)
 WebHCat's StreamingDelegator does not include -inputreader as a valid option 
 when submitting jobs to 
 http://www.myserver.com/templeton/v1/mapreduce/streaming endpoint. If 
 -inputreader is specified and passed to templeton server (perhaps via CURL 
 operation) , it will get truncated and not passed as parameter from 
 TempletonControllerJob to Hadoop.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


Re: setting up a branch for testing

2015-02-19 Thread Sergey Shelukhin
Can you elaborate on how many machines are needed, minimum (is it 1+, just 
determined by the throughput of QA runs that we want; or is there some fixed 
requirement too), and what is the setup/process to make them work with HiveQA 
(in general so we'd know how we can provide machines)?

Thanks!

From: Szehon sze...@cloudera.com
Sent: Friday, January 23, 2015 7:38 PM
To: dev@hive.apache.org
Subject: Re: setting up a branch for testing

Yea but for precommit testing it would need a cluster setup that runs an 
instance of Ptest server.  We only have spark branch setup for that other than 
trunk, we have one cluster running spark and another running trunk.

Setup is doable (actually just need to setup master) but takes some steps, and 
physical machines.

Thanks
Szehon

 On Jan 23, 2015, at 6:04 PM, Sergey Shelukhin ser...@hortonworks.com wrote:

 Hi.
 Hive dev doc mentions that patches can be tested by HiveQA against the
 branch by supplying the branch name in the patch name.
 However, as far as I understand this requires some setup for each specific
 branch.

 Is it possible to set up llap branch for HiveQA testing?

 --
 CONFIDENTIALITY NOTICE
 NOTICE: This message is intended for the use of the individual or entity to
 which it is addressed and may contain information that is confidential,
 privileged and exempt from disclosure under applicable law. If the reader
 of this message is not the intended recipient, you are hereby notified that
 any printing, copying, dissemination, distribution, disclosure or
 forwarding of this communication is strictly prohibited. If you have
 received this communication in error, please contact the sender immediately
 and delete it from your system. Thank You.


[jira] [Commented] (HIVE-9653) LLAP: create a reasonable q file test for ORC IO

2015-02-19 Thread Sergey Shelukhin (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9653?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14328222#comment-14328222
 ] 

Sergey Shelukhin commented on HIVE-9653:


I added some test. We need to make it more reasonable :)

 LLAP: create a reasonable q file test for ORC IO
 

 Key: HIVE-9653
 URL: https://issues.apache.org/jira/browse/HIVE-9653
 Project: Hive
  Issue Type: Sub-task
Reporter: Sergey Shelukhin
Assignee: Sergey Shelukhin

 multiple RGs, RG filters, variety of queries



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9730) LLAP: make sure logging is never called when not needed

2015-02-19 Thread Gopal V (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9730?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gopal V updated HIVE-9730:
--
Attachment: log4j-llap.png

 LLAP: make sure logging is never called when not needed
 ---

 Key: HIVE-9730
 URL: https://issues.apache.org/jira/browse/HIVE-9730
 Project: Hive
  Issue Type: Sub-task
Reporter: Sergey Shelukhin
Assignee: Sergey Shelukhin
 Attachments: log4j-llap.png


 log4j logging has really inefficient serialization



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9730) LLAP: make sure logging is never called when not needed

2015-02-19 Thread Gopal V (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9730?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gopal V updated HIVE-9730:
--
Description: 
log4j logging has really inefficient serialization

!log4j-llap.png!

  was:log4j logging has really inefficient serialization


 LLAP: make sure logging is never called when not needed
 ---

 Key: HIVE-9730
 URL: https://issues.apache.org/jira/browse/HIVE-9730
 Project: Hive
  Issue Type: Sub-task
Reporter: Sergey Shelukhin
Assignee: Sergey Shelukhin
 Attachments: log4j-llap.png


 log4j logging has really inefficient serialization
 !log4j-llap.png!



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9086) Add language support to PURGE data while dropping partitions.

2015-02-19 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9086?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14328533#comment-14328533
 ] 

Hive QA commented on HIVE-9086:
---



{color:green}Overall{color}: +1 all checks pass

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12686790/HIVE-9086.1.patch

{color:green}SUCCESS:{color} +1 7566 tests passed

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/2828/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/2828/console
Test logs: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-2828/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12686790 - PreCommit-HIVE-TRUNK-Build

 Add language support to PURGE data while dropping partitions.
 -

 Key: HIVE-9086
 URL: https://issues.apache.org/jira/browse/HIVE-9086
 Project: Hive
  Issue Type: Bug
  Components: Metastore
Affects Versions: 0.15.0
Reporter: Mithun Radhakrishnan
Assignee: Mithun Radhakrishnan
 Attachments: HIVE-9086.1.patch


 HIVE-9083 adds metastore-support to skip-trash while dropping partitions. 
 This patch includes language support to do the same.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9556) create UDF to calculate the Levenshtein distance between two strings

2015-02-19 Thread Lefty Leverenz (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9556?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Lefty Leverenz updated HIVE-9556:
-
Labels:   (was: TODOC1.2)

 create UDF to calculate the Levenshtein distance between two strings
 

 Key: HIVE-9556
 URL: https://issues.apache.org/jira/browse/HIVE-9556
 Project: Hive
  Issue Type: Improvement
  Components: UDF
Reporter: Alexander Pivovarov
Assignee: Alexander Pivovarov
 Fix For: 1.2.0

 Attachments: HIVE-9556.1.patch, HIVE-9556.2.patch, HIVE-9556.3.patch


 Levenshtein distance is a string metric for measuring the difference between 
 two sequences. Informally, the Levenshtein distance between two words is the 
 minimum number of single-character edits (i.e. insertions, deletions or 
 substitutions) required to change one word into the other. It is named after 
 Vladimir Levenshtein, who considered this distance in 1965.
 Example:
 The Levenshtein distance between kitten and sitting is 3
 1. kitten → sitten (substitution of s for k)
 2. sitten → sittin (substitution of i for e)
 3. sittin → sitting (insertion of g at the end).
 {code}
 select levenshtein('kitten', 'sitting');
 3
 {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9556) create UDF to calculate the Levenshtein distance between two strings

2015-02-19 Thread Lefty Leverenz (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9556?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14328585#comment-14328585
 ] 

Lefty Leverenz commented on HIVE-9556:
--

Thanks for the doc, [~apivovarov].  I removed the TODOC1.2 label.

 create UDF to calculate the Levenshtein distance between two strings
 

 Key: HIVE-9556
 URL: https://issues.apache.org/jira/browse/HIVE-9556
 Project: Hive
  Issue Type: Improvement
  Components: UDF
Reporter: Alexander Pivovarov
Assignee: Alexander Pivovarov
 Fix For: 1.2.0

 Attachments: HIVE-9556.1.patch, HIVE-9556.2.patch, HIVE-9556.3.patch


 Levenshtein distance is a string metric for measuring the difference between 
 two sequences. Informally, the Levenshtein distance between two words is the 
 minimum number of single-character edits (i.e. insertions, deletions or 
 substitutions) required to change one word into the other. It is named after 
 Vladimir Levenshtein, who considered this distance in 1965.
 Example:
 The Levenshtein distance between kitten and sitting is 3
 1. kitten → sitten (substitution of s for k)
 2. sitten → sittin (substitution of i for e)
 3. sittin → sitting (insertion of g at the end).
 {code}
 select levenshtein('kitten', 'sitting');
 3
 {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9277) Hybrid Hybrid Grace Hash Join

2015-02-19 Thread Wei Zheng (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9277?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Wei Zheng updated HIVE-9277:

Attachment: HIVE-9277.01.patch

Uploading 1st patch for testing

 Hybrid Hybrid Grace Hash Join
 -

 Key: HIVE-9277
 URL: https://issues.apache.org/jira/browse/HIVE-9277
 Project: Hive
  Issue Type: New Feature
  Components: Physical Optimizer
Reporter: Wei Zheng
Assignee: Wei Zheng
  Labels: join
 Attachments: HIVE-9277.01.patch, 
 High-leveldesignforHybridHybridGraceHashJoinv1.0.pdf


 We are proposing an enhanced hash join algorithm called “hybrid hybrid grace 
 hash join”. We can benefit from this feature as illustrated below:
 o The query will not fail even if the estimated memory requirement is 
 slightly wrong
 o Expensive garbage collection overhead can be avoided when hash table grows
 o Join execution using a Map join operator even though the small table 
 doesn't fit in memory as spilling some data from the build and probe sides 
 will still be cheaper than having to shuffle the large fact table
 The design was based on Hadoop’s parallel processing capability and 
 significant amount of memory available.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9277) Hybrid Hybrid Grace Hash Join

2015-02-19 Thread Wei Zheng (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9277?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Wei Zheng updated HIVE-9277:

Status: Patch Available  (was: Open)

 Hybrid Hybrid Grace Hash Join
 -

 Key: HIVE-9277
 URL: https://issues.apache.org/jira/browse/HIVE-9277
 Project: Hive
  Issue Type: New Feature
  Components: Physical Optimizer
Reporter: Wei Zheng
Assignee: Wei Zheng
  Labels: join
 Attachments: HIVE-9277.01.patch, 
 High-leveldesignforHybridHybridGraceHashJoinv1.0.pdf


 We are proposing an enhanced hash join algorithm called “hybrid hybrid grace 
 hash join”. We can benefit from this feature as illustrated below:
 o The query will not fail even if the estimated memory requirement is 
 slightly wrong
 o Expensive garbage collection overhead can be avoided when hash table grows
 o Join execution using a Map join operator even though the small table 
 doesn't fit in memory as spilling some data from the build and probe sides 
 will still be cheaper than having to shuffle the large fact table
 The design was based on Hadoop’s parallel processing capability and 
 significant amount of memory available.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HIVE-9738) create SOUNDEX udf

2015-02-19 Thread Alexander Pivovarov (JIRA)
Alexander Pivovarov created HIVE-9738:
-

 Summary: create SOUNDEX udf
 Key: HIVE-9738
 URL: https://issues.apache.org/jira/browse/HIVE-9738
 Project: Hive
  Issue Type: Improvement
  Components: UDF
Reporter: Alexander Pivovarov
Assignee: Alexander Pivovarov






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9728) LLAP: add heap mode to allocator (for q files, YARN w/o direct buffer accounting support)

2015-02-19 Thread Lefty Leverenz (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9728?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14328618#comment-14328618
 ] 

Lefty Leverenz commented on HIVE-9728:
--

Doc note:  This adds *hive.llap.io.cache.direct* to HiveConf.java in the LLAP 
branch, so it will need to be documented when the branch gets merged to trunk.

 LLAP: add heap mode to allocator (for q files, YARN w/o direct buffer 
 accounting support)
 -

 Key: HIVE-9728
 URL: https://issues.apache.org/jira/browse/HIVE-9728
 Project: Hive
  Issue Type: Sub-task
Reporter: Sergey Shelukhin
Assignee: Sergey Shelukhin





--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9726) Upgrade to spark 1.3 [Spark Branch]

2015-02-19 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9726?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14328536#comment-14328536
 ] 

Hive QA commented on HIVE-9726:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12699788/HIVE-9671.1-spark.patch

{color:red}ERROR:{color} -1 due to 3 failed/errored test(s), 7510 tests executed
*Failed tests:*
{noformat}
TestMiniSparkOnYarnCliDriver - did not produce a TEST-*.xml file
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_mapjoin_memcheck
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_udaf_percentile_approx_23
{noformat}

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-SPARK-Build/739/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-SPARK-Build/739/console
Test logs: 
http://ec2-50-18-27-0.us-west-1.compute.amazonaws.com/logs/PreCommit-HIVE-SPARK-Build-739/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 3 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12699788 - PreCommit-HIVE-SPARK-Build

 Upgrade to spark 1.3 [Spark Branch]
 ---

 Key: HIVE-9726
 URL: https://issues.apache.org/jira/browse/HIVE-9726
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Affects Versions: spark-branch
Reporter: Brock Noland
Assignee: Brock Noland
 Attachments: HIVE-9671.1-spark.patch






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9277) Hybrid Hybrid Grace Hash Join

2015-02-19 Thread Wei Zheng (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9277?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Wei Zheng updated HIVE-9277:

Attachment: HIVE-9277.02.patch

Uploading 2nd patch for testing

 Hybrid Hybrid Grace Hash Join
 -

 Key: HIVE-9277
 URL: https://issues.apache.org/jira/browse/HIVE-9277
 Project: Hive
  Issue Type: New Feature
  Components: Physical Optimizer
Reporter: Wei Zheng
Assignee: Wei Zheng
  Labels: join
 Attachments: HIVE-9277.01.patch, HIVE-9277.02.patch, 
 High-leveldesignforHybridHybridGraceHashJoinv1.0.pdf


 We are proposing an enhanced hash join algorithm called “hybrid hybrid grace 
 hash join”. We can benefit from this feature as illustrated below:
 o The query will not fail even if the estimated memory requirement is 
 slightly wrong
 o Expensive garbage collection overhead can be avoided when hash table grows
 o Join execution using a Map join operator even though the small table 
 doesn't fit in memory as spilling some data from the build and probe sides 
 will still be cheaper than having to shuffle the large fact table
 The design was based on Hadoop’s parallel processing capability and 
 significant amount of memory available.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


Re: Fix version for hbase-metastore branch

2015-02-19 Thread Lefty Leverenz
Also, what should we use for a documentation label?  (HIVE-9606
https://issues.apache.org/jira/browse/HIVE-9606 needs one.)

TODOC labels are proliferating for all the releases and branches, but I
don't think a generic TODOC label would be helpful.  So what would be a
good abbreviation for the hbase-metastore branch?  Maybe TODOC-HBMETA?

-- Lefty

On Thu, Feb 19, 2015 at 7:12 PM, Alan Gates alanfga...@gmail.com wrote:

 Could someone with admin permissions on our JIRA add an
 hbase-metastore-branch label?  I'll take care of changing all the fix
 versions for the few JIRA's we've already committed.  Thanks.

 Alan.

   Ashutosh Chauhan hashut...@apache.org
  February 19, 2015 at 11:22
 This is what we have been doing for cbo work. e.g.
 https://issues.apache.org/jira/browse/HIVE-9581


   Thejas Nair thejas.n...@gmail.com
  February 19, 2015 at 11:17
 I agree, using a label for fix version makes sense in this case. I believe
 that is what had been done for hive-on-spark and hive-on-tez.



   Alan Gates alanfga...@gmail.com
  February 19, 2015 at 10:56
 I've been marking JIRAs on this branch as fixed in 1.2, since that's the
 next version.  But that seems wrong as I doubt this code will be in by
 1.2.  What's the usual practice here?  It seems it would make sense to make
 a label for this branch and mark them as fixed with that label and then
 when we actually release this in a version we can update all the JIRAs with
 that label.

 Alan.




Review Request 31222: HIVE-9738 create SOUNDEX udf

2015-02-19 Thread Alexander Pivovarov

---
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/31222/
---

Review request for hive, Alan Gates and Jason Dere.


Bugs: HIVE-9738
https://issues.apache.org/jira/browse/HIVE-9738


Repository: hive-git


Description
---

HIVE-9738 create SOUNDEX udf


Diffs
-

  ql/src/java/org/apache/hadoop/hive/ql/exec/FunctionRegistry.java 
bfeb33c9edf79071bf106a3780e1c6bc21be3753 
  ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFSoundex.java 
PRE-CREATION 
  ql/src/test/org/apache/hadoop/hive/ql/udf/generic/TestGenericUDFSoundex.java 
PRE-CREATION 
  ql/src/test/queries/clientpositive/udf_soundex.q PRE-CREATION 
  ql/src/test/results/clientpositive/show_functions.q.out 
d4b06507f2356cc58ff7312ff78261406f3ceb4f 
  ql/src/test/results/clientpositive/udf_soundex.q.out PRE-CREATION 

Diff: https://reviews.apache.org/r/31222/diff/


Testing
---


Thanks,

Alexander Pivovarov



[jira] [Updated] (HIVE-9738) create SOUNDEX udf

2015-02-19 Thread Alexander Pivovarov (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9738?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alexander Pivovarov updated HIVE-9738:
--
Attachment: HIVE-9738.1.patch

patch #1

 create SOUNDEX udf
 --

 Key: HIVE-9738
 URL: https://issues.apache.org/jira/browse/HIVE-9738
 Project: Hive
  Issue Type: Improvement
  Components: UDF
Reporter: Alexander Pivovarov
Assignee: Alexander Pivovarov
 Attachments: HIVE-9738.1.patch


 Soundex is an encoding used to relate similar names, but can also be used as 
 a general purpose scheme to find word with similar phonemes.
 The American Soundex System
 The soundex code consist of the first letter of the name followed by three 
 digits. These three digits are determined by dropping the letters a, e, i, o, 
 u, h, w and y and adding three digits from the remaining letters of the name 
 according to the table below. There are only two additional rules. (1) If two 
 or more consecutive letters have the same code, they are coded as one letter. 
 (2) If there are an insufficient numbers of letters to make the three digits, 
 the remaining digits are set to zero.
 Soundex Table
  1 b,f,p,v
  2 c,g,j,k,q,s,x,z
  3 d, t
  4 l
  5 m, n
  6 r
 Examples:
 Miller M460
 Peterson P362
 Peters P362
 Auerbach A612
 Uhrbach U612
 Moskowitz M232
 Moskovitz M213
 Implementation:
 http://commons.apache.org/proper/commons-codec/apidocs/org/apache/commons/codec/language/Soundex.html



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Work started] (HIVE-9738) create SOUNDEX udf

2015-02-19 Thread Alexander Pivovarov (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9738?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Work on HIVE-9738 started by Alexander Pivovarov.
-
 create SOUNDEX udf
 --

 Key: HIVE-9738
 URL: https://issues.apache.org/jira/browse/HIVE-9738
 Project: Hive
  Issue Type: Improvement
  Components: UDF
Reporter: Alexander Pivovarov
Assignee: Alexander Pivovarov
 Attachments: HIVE-9738.1.patch


 Soundex is an encoding used to relate similar names, but can also be used as 
 a general purpose scheme to find word with similar phonemes.
 The American Soundex System
 The soundex code consist of the first letter of the name followed by three 
 digits. These three digits are determined by dropping the letters a, e, i, o, 
 u, h, w and y and adding three digits from the remaining letters of the name 
 according to the table below. There are only two additional rules. (1) If two 
 or more consecutive letters have the same code, they are coded as one letter. 
 (2) If there are an insufficient numbers of letters to make the three digits, 
 the remaining digits are set to zero.
 Soundex Table
  1 b,f,p,v
  2 c,g,j,k,q,s,x,z
  3 d, t
  4 l
  5 m, n
  6 r
 Examples:
 Miller M460
 Peterson P362
 Peters P362
 Auerbach A612
 Uhrbach U612
 Moskowitz M232
 Moskovitz M213
 Implementation:
 http://commons.apache.org/proper/commons-codec/apidocs/org/apache/commons/codec/language/Soundex.html



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9738) create SOUNDEX udf

2015-02-19 Thread Alexander Pivovarov (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9738?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alexander Pivovarov updated HIVE-9738:
--
Status: Patch Available  (was: In Progress)

 create SOUNDEX udf
 --

 Key: HIVE-9738
 URL: https://issues.apache.org/jira/browse/HIVE-9738
 Project: Hive
  Issue Type: Improvement
  Components: UDF
Reporter: Alexander Pivovarov
Assignee: Alexander Pivovarov
 Attachments: HIVE-9738.1.patch


 Soundex is an encoding used to relate similar names, but can also be used as 
 a general purpose scheme to find word with similar phonemes.
 The American Soundex System
 The soundex code consist of the first letter of the name followed by three 
 digits. These three digits are determined by dropping the letters a, e, i, o, 
 u, h, w and y and adding three digits from the remaining letters of the name 
 according to the table below. There are only two additional rules. (1) If two 
 or more consecutive letters have the same code, they are coded as one letter. 
 (2) If there are an insufficient numbers of letters to make the three digits, 
 the remaining digits are set to zero.
 Soundex Table
  1 b,f,p,v
  2 c,g,j,k,q,s,x,z
  3 d, t
  4 l
  5 m, n
  6 r
 Examples:
 Miller M460
 Peterson P362
 Peters P362
 Auerbach A612
 Uhrbach U612
 Moskowitz M232
 Moskovitz M213
 Implementation:
 http://commons.apache.org/proper/commons-codec/apidocs/org/apache/commons/codec/language/Soundex.html



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Resolved] (HIVE-9728) LLAP: add heap mode to allocator (for q files, YARN w/o direct buffer accounting support)

2015-02-19 Thread Sergey Shelukhin (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9728?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sergey Shelukhin resolved HIVE-9728.

Resolution: Fixed

 LLAP: add heap mode to allocator (for q files, YARN w/o direct buffer 
 accounting support)
 -

 Key: HIVE-9728
 URL: https://issues.apache.org/jira/browse/HIVE-9728
 Project: Hive
  Issue Type: Sub-task
Reporter: Sergey Shelukhin
Assignee: Sergey Shelukhin





--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HIVE-9737) Issue come while creating the table in hbase using java Impla API

2015-02-19 Thread Mohit Sharma (JIRA)
Mohit Sharma created HIVE-9737:
--

 Summary: Issue come while creating the table in hbase using java 
Impla API
 Key: HIVE-9737
 URL: https://issues.apache.org/jira/browse/HIVE-9737
 Project: Hive
  Issue Type: Bug
  Components: Database/Schema
 Environment: Linux OS, Using Java Impala API with thrift, Cloudera 
Hbase database
Reporter: Mohit Sharma


I am trying to create hbase table using this query
CREATE TABLE foo4(rowkey STRING, a STRING, b STRING) STORED BY 
'org.apache.hadoop.hive.hbase.HBaseStorageHandler' WITH SERDEPROPERTIES 
('hbase.columns.mapping' = ':key,f:c1,f:c2') TBLPROPERTIES ('hbase.table.name' 
= 'bar4');

with the help of java Impala api and I am sharing you code link
https://github.com/pauldeschacht/impala-java-client

When I am trying to create table in habse I am facing this issue

AnalysisException: Syntax error in line 2:
STORED BY 'org.apache.hadoop.hive.hbase.HBaseStorageHandler'
   ^
Encountered: BY
Expected: AS

CAUSED BY: Exception: Syntax error,HY000,0,false

Please help me what I do?



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9086) Add language support to PURGE data while dropping partitions.

2015-02-19 Thread Lefty Leverenz (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9086?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14328606#comment-14328606
 ] 

Lefty Leverenz commented on HIVE-9086:
--

Doc note:  Uh oh, I documented this prematurely (for HIVE-7100).  But did I get 
it right, except for the jira attibution and release number?

* [DDL -- Drop Partitions | 
https://cwiki.apache.org/confluence/display/Hive/LanguageManual+DDL#LanguageManualDDL-DropPartitions]

 Add language support to PURGE data while dropping partitions.
 -

 Key: HIVE-9086
 URL: https://issues.apache.org/jira/browse/HIVE-9086
 Project: Hive
  Issue Type: Bug
  Components: Metastore
Affects Versions: 0.15.0
Reporter: Mithun Radhakrishnan
Assignee: Mithun Radhakrishnan
 Attachments: HIVE-9086.1.patch


 HIVE-9083 adds metastore-support to skip-trash while dropping partitions. 
 This patch includes language support to do the same.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9738) create SOUNDEX udf

2015-02-19 Thread Alexander Pivovarov (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9738?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alexander Pivovarov updated HIVE-9738:
--
Description: 
Soundex is an encoding used to relate similar names, but can also be used as a 
general purpose scheme to find word with similar phonemes.

The American Soundex System
The soundex code consist of the first letter of the name followed by three 
digits. These three digits are determined by dropping the letters a, e, i, o, 
u, h, w and y and adding three digits from the remaining letters of the name 
according to the table below. There are only two additional rules. (1) If two 
or more consecutive letters have the same code, they are coded as one letter. 
(2) If there are an insufficient numbers of letters to make the three digits, 
the remaining digits are set to zero.

Soundex Table
 1 b,f,p,v
 2 c,g,j,k,q,s,x,z
 3 d, t
 4 l
 5 m, n
 6 r

Examples:
Miller M460
Peterson P362
Peters P362
Auerbach A612
Uhrbach U612
Moskowitz M232
Moskovitz M213

Implementation:
http://commons.apache.org/proper/commons-codec/apidocs/org/apache/commons/codec/language/Soundex.html


  was:
Soundex is an encoding used to relate similar names, but can also be used as a 
general purpose scheme to find word with similar phonemes.

The American Soundex System
The soundex code consist of the first letter of the name followed by three 
digits. These three digits are determined by dropping the letters a, e, i, o, 
u, h, w and y and adding three digits from the remaining letters of the name 
according to the table below. There are only two additional rules. (1) If two 
or more consecutive letters have the same code, they are coded as one letter. 
(2) If there are an insufficient numbers of letters to make the three digits, 
the remaining digits are set to zero.

Soundex Table
 1 b,f,p,v
 2 c,g,j,k,q,s,x,z
 3 d, t
 4 l
 5 m, n
 6 r

Examples:
Miller M460
Peterson P362
Peters P362
Auerbach A612
Uhrbach U612
Moskowitz M232
Moskovitz M213




 create SOUNDEX udf
 --

 Key: HIVE-9738
 URL: https://issues.apache.org/jira/browse/HIVE-9738
 Project: Hive
  Issue Type: Improvement
  Components: UDF
Reporter: Alexander Pivovarov
Assignee: Alexander Pivovarov

 Soundex is an encoding used to relate similar names, but can also be used as 
 a general purpose scheme to find word with similar phonemes.
 The American Soundex System
 The soundex code consist of the first letter of the name followed by three 
 digits. These three digits are determined by dropping the letters a, e, i, o, 
 u, h, w and y and adding three digits from the remaining letters of the name 
 according to the table below. There are only two additional rules. (1) If two 
 or more consecutive letters have the same code, they are coded as one letter. 
 (2) If there are an insufficient numbers of letters to make the three digits, 
 the remaining digits are set to zero.
 Soundex Table
  1 b,f,p,v
  2 c,g,j,k,q,s,x,z
  3 d, t
  4 l
  5 m, n
  6 r
 Examples:
 Miller M460
 Peterson P362
 Peters P362
 Auerbach A612
 Uhrbach U612
 Moskowitz M232
 Moskovitz M213
 Implementation:
 http://commons.apache.org/proper/commons-codec/apidocs/org/apache/commons/codec/language/Soundex.html



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9738) create SOUNDEX udf

2015-02-19 Thread Alexander Pivovarov (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9738?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alexander Pivovarov updated HIVE-9738:
--
Description: 
Soundex is an encoding used to relate similar names, but can also be used as a 
general purpose scheme to find word with similar phonemes.

The American Soundex System
The soundex code consist of the first letter of the name followed by three 
digits. These three digits are determined by dropping the letters a, e, i, o, 
u, h, w and y and adding three digits from the remaining letters of the name 
according to the table below. There are only two additional rules. (1) If two 
or more consecutive letters have the same code, they are coded as one letter. 
(2) If there are an insufficient numbers of letters to make the three digits, 
the remaining digits are set to zero.

Soundex Table
 1 b,f,p,v
 2 c,g,j,k,q,s,x,z
 3 d, t
 4 l
 5 m, n
 6 r

Examples:
Miller M460
Peterson P362
Peters P362
Auerbach A612
Uhrbach U612
Moskowitz M232
Moskovitz M213



 create SOUNDEX udf
 --

 Key: HIVE-9738
 URL: https://issues.apache.org/jira/browse/HIVE-9738
 Project: Hive
  Issue Type: Improvement
  Components: UDF
Reporter: Alexander Pivovarov
Assignee: Alexander Pivovarov

 Soundex is an encoding used to relate similar names, but can also be used as 
 a general purpose scheme to find word with similar phonemes.
 The American Soundex System
 The soundex code consist of the first letter of the name followed by three 
 digits. These three digits are determined by dropping the letters a, e, i, o, 
 u, h, w and y and adding three digits from the remaining letters of the name 
 according to the table below. There are only two additional rules. (1) If two 
 or more consecutive letters have the same code, they are coded as one letter. 
 (2) If there are an insufficient numbers of letters to make the three digits, 
 the remaining digits are set to zero.
 Soundex Table
  1 b,f,p,v
  2 c,g,j,k,q,s,x,z
  3 d, t
  4 l
  5 m, n
  6 r
 Examples:
 Miller M460
 Peterson P362
 Peters P362
 Auerbach A612
 Uhrbach U612
 Moskowitz M232
 Moskovitz M213



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-6617) Reduce ambiguity in grammar

2015-02-19 Thread Pengcheng Xiong (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6617?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Pengcheng Xiong updated HIVE-6617:
--
Status: Open  (was: Patch Available)

 Reduce ambiguity in grammar
 ---

 Key: HIVE-6617
 URL: https://issues.apache.org/jira/browse/HIVE-6617
 Project: Hive
  Issue Type: Task
Reporter: Ashutosh Chauhan
Assignee: Pengcheng Xiong
 Attachments: HIVE-6617.01.patch, HIVE-6617.02.patch, 
 HIVE-6617.03.patch, HIVE-6617.04.patch, HIVE-6617.05.patch, 
 HIVE-6617.06.patch, HIVE-6617.07.patch, HIVE-6617.08.patch, 
 HIVE-6617.09.patch, HIVE-6617.10.patch, HIVE-6617.11.patch, 
 HIVE-6617.12.patch, HIVE-6617.13.patch, HIVE-6617.14.patch, 
 HIVE-6617.15.patch, HIVE-6617.16.patch, HIVE-6617.17.patch


 CLEAR LIBRARY CACHE
 As of today, antlr reports 214 warnings. Need to bring down this number, 
 ideally to 0.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-6617) Reduce ambiguity in grammar

2015-02-19 Thread Pengcheng Xiong (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6617?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Pengcheng Xiong updated HIVE-6617:
--
Attachment: HIVE-6617.17.patch

address [~jpullokkaran]'s comments, (1) add back keywords under a special 
category; (2) add negative/positive tests for these keywords

 Reduce ambiguity in grammar
 ---

 Key: HIVE-6617
 URL: https://issues.apache.org/jira/browse/HIVE-6617
 Project: Hive
  Issue Type: Task
Reporter: Ashutosh Chauhan
Assignee: Pengcheng Xiong
 Attachments: HIVE-6617.01.patch, HIVE-6617.02.patch, 
 HIVE-6617.03.patch, HIVE-6617.04.patch, HIVE-6617.05.patch, 
 HIVE-6617.06.patch, HIVE-6617.07.patch, HIVE-6617.08.patch, 
 HIVE-6617.09.patch, HIVE-6617.10.patch, HIVE-6617.11.patch, 
 HIVE-6617.12.patch, HIVE-6617.13.patch, HIVE-6617.14.patch, 
 HIVE-6617.15.patch, HIVE-6617.16.patch, HIVE-6617.17.patch


 CLEAR LIBRARY CACHE
 As of today, antlr reports 214 warnings. Need to bring down this number, 
 ideally to 0.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-6617) Reduce ambiguity in grammar

2015-02-19 Thread Pengcheng Xiong (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6617?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Pengcheng Xiong updated HIVE-6617:
--
Status: Patch Available  (was: Open)

 Reduce ambiguity in grammar
 ---

 Key: HIVE-6617
 URL: https://issues.apache.org/jira/browse/HIVE-6617
 Project: Hive
  Issue Type: Task
Reporter: Ashutosh Chauhan
Assignee: Pengcheng Xiong
 Attachments: HIVE-6617.01.patch, HIVE-6617.02.patch, 
 HIVE-6617.03.patch, HIVE-6617.04.patch, HIVE-6617.05.patch, 
 HIVE-6617.06.patch, HIVE-6617.07.patch, HIVE-6617.08.patch, 
 HIVE-6617.09.patch, HIVE-6617.10.patch, HIVE-6617.11.patch, 
 HIVE-6617.12.patch, HIVE-6617.13.patch, HIVE-6617.14.patch, 
 HIVE-6617.15.patch, HIVE-6617.16.patch, HIVE-6617.17.patch


 CLEAR LIBRARY CACHE
 As of today, antlr reports 214 warnings. Need to bring down this number, 
 ideally to 0.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Resolved] (HIVE-9719) Up calcite version on cbo branch

2015-02-19 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9719?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan resolved HIVE-9719.

   Resolution: Fixed
Fix Version/s: cbo-branch

Committed to cbo branch.

 Up calcite version on cbo branch
 

 Key: HIVE-9719
 URL: https://issues.apache.org/jira/browse/HIVE-9719
 Project: Hive
  Issue Type: Task
  Components: CBO
Affects Versions: cbo-branch
Reporter: Ashutosh Chauhan
Assignee: Ashutosh Chauhan
 Fix For: cbo-branch

 Attachments: HIVE-9719.cbo.patch


 CALCITE-594 is now checked in calcite master.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HIVE-9734) Correlating expression cannot contain unqualified column references

2015-02-19 Thread N Campbell (JIRA)
N Campbell created HIVE-9734:


 Summary: Correlating expression cannot contain unqualified column 
references
 Key: HIVE-9734
 URL: https://issues.apache.org/jira/browse/HIVE-9734
 Project: Hive
  Issue Type: Bug
  Components: SQL
Affects Versions: 0.14.0
Reporter: N Campbell
Priority: Minor


The following valid ISO-SQL constructs will not parse

Unsupported SubQuery Expression 'c1': Correlating expression cannot contain 
unqualified column references

The statement will parse if the predicate is changed to where tjoin1.c1 in ( 
)

select tjoin1.rnum, tjoin1.c1, tjoin1.c2 from tjoin1 where c1 in ( select c1 
from tjoin1) 

select tjoin1.rnum, tjoin1.c1, tjoin1.c2 from tjoin1 where c1 in ( select 
tjoin1.c1 from tjoin1) 

select tjoin1.rnum, tjoin1.c1, tjoin1.c2 from tjoin1 where c1 in ( select x.c1 
from tjoin1 x) 




--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9733) Fix test failures on cbo branch

2015-02-19 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9733?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-9733:
---
   Resolution: Fixed
Fix Version/s: cbo-branch
   Status: Resolved  (was: Patch Available)

Committed to branch.

 Fix test failures on cbo branch
 ---

 Key: HIVE-9733
 URL: https://issues.apache.org/jira/browse/HIVE-9733
 Project: Hive
  Issue Type: Bug
  Components: CBO, Query Planning
Affects Versions: cbo-branch
Reporter: Ashutosh Chauhan
Assignee: Ashutosh Chauhan
 Fix For: cbo-branch

 Attachments: HIVE-9733.cbo.patch


 A constant changed their home class.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


Re: Review Request 31072: HIVE-9678 create timediff UDF

2015-02-19 Thread Alexander Pivovarov

---
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/31072/
---

(Updated Feb. 20, 2015, 1:04 a.m.)


Review request for hive, Jason Dere and Thejas Nair.


Changes
---

remove timezone support and add nanos precision support


Bugs: HIVE-9678
https://issues.apache.org/jira/browse/HIVE-9678


Repository: hive-git


Description
---

HIVE-9678 create timediff UDF


Diffs (updated)
-

  ql/src/java/org/apache/hadoop/hive/ql/exec/FunctionRegistry.java 
909f6ae6ec2aa24942286f15ffa7293378ca53fd 
  ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFTimeDiff.java 
PRE-CREATION 
  ql/src/test/org/apache/hadoop/hive/ql/udf/generic/TestGenericUDFTimeDiff.java 
PRE-CREATION 
  ql/src/test/queries/clientpositive/udf_timediff.q PRE-CREATION 
  ql/src/test/results/clientpositive/show_functions.q.out 
a58f453ddcca8a99319b8b465af0db3cc8f7302f 
  ql/src/test/results/clientpositive/udf_timediff.q.out PRE-CREATION 

Diff: https://reviews.apache.org/r/31072/diff/


Testing
---


Thanks,

Alexander Pivovarov



[jira] [Commented] (HIVE-9678) create timediff UDF

2015-02-19 Thread Alexander Pivovarov (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9678?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14328409#comment-14328409
 ] 

Alexander Pivovarov commented on HIVE-9678:
---

patch #3

 create timediff UDF
 ---

 Key: HIVE-9678
 URL: https://issues.apache.org/jira/browse/HIVE-9678
 Project: Hive
  Issue Type: Improvement
  Components: UDF
Reporter: Alexander Pivovarov
Assignee: Alexander Pivovarov
 Attachments: HIVE-9678.1.patch, HIVE-9678.2.patch, HIVE-9678.3.patch


 MySQL has very useful function timediff. We should have it in Hive
 {code}
 select timediff('2015-02-12 05:09:07', '2015-02-12 01:18:20');
 OK
 03:50:47
 {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9674) *DropPartitionEvent should handle partition-sets.

2015-02-19 Thread Mithun Radhakrishnan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9674?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mithun Radhakrishnan updated HIVE-9674:
---
Attachment: (was: HIVE-9674.1.patch)

 *DropPartitionEvent should handle partition-sets.
 -

 Key: HIVE-9674
 URL: https://issues.apache.org/jira/browse/HIVE-9674
 Project: Hive
  Issue Type: Bug
  Components: Metastore
Affects Versions: 0.14.0
Reporter: Mithun Radhakrishnan
Assignee: Mithun Radhakrishnan

 Dropping a set of N partitions from a table currently results in N 
 DropPartitionEvents (and N PreDropPartitionEvents) being fired serially. This 
 is wasteful, especially so for large N. It also makes it impossible to even 
 try to run authorization-checks on all partitions in a batch.
 Taking the cue from HIVE-9609, we should compose an {{IterablePartition}} 
 in the event, and expose them via an {{Iterator}}.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9678) create timediff UDF

2015-02-19 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9678?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14328480#comment-14328480
 ] 

Hive QA commented on HIVE-9678:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12699787/HIVE-9678.3.patch

{color:red}ERROR:{color} -1 due to 1 failed/errored test(s), 7566 tests executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_udaf_percentile_approx_23
{noformat}

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/2826/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/2826/console
Test logs: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-2826/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 1 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12699787 - PreCommit-HIVE-TRUNK-Build

 create timediff UDF
 ---

 Key: HIVE-9678
 URL: https://issues.apache.org/jira/browse/HIVE-9678
 Project: Hive
  Issue Type: Improvement
  Components: UDF
Reporter: Alexander Pivovarov
Assignee: Alexander Pivovarov
 Attachments: HIVE-9678.1.patch, HIVE-9678.2.patch, HIVE-9678.3.patch


 MySQL has very useful function timediff. We should have it in Hive
 {code}
 select timediff('2015-02-12 05:09:07', '2015-02-12 01:18:20');
 OK
 03:50:47
 {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9718) Insert into dynamic partitions with same column structure in the distibute by clause barfs

2015-02-19 Thread Ashutosh Chauhan (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9718?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14327713#comment-14327713
 ] 

Ashutosh Chauhan commented on HIVE-9718:


You need to name your patch file per 
https://cwiki.apache.org/confluence/display/Hive/Hive+PreCommit+Patch+Testing 
for automated QA to run.

 Insert into dynamic partitions with same column structure in the distibute 
 by clause barfs
 

 Key: HIVE-9718
 URL: https://issues.apache.org/jira/browse/HIVE-9718
 Project: Hive
  Issue Type: Bug
Affects Versions: 0.14.0, 1.0.0
Reporter: Pavan Srinivas
Priority: Critical
 Attachments: nation.tbl, patch.txt


 Sample reproducible query: 
 {code}
 SET hive.exec.dynamic.partition.mode=nonstrict;
 SET hive.exec.dynamic.partition=true;
  insert overwrite table nation_new_p partition (some)
 select n_name as name1, n_name as name2, n_name as name3 from nation 
 distribute by name3;
 {code}
 Note: Make sure there is data in the source table to reproduce the issue. 
 During the optimizations done for Jira: 
 https://issues.apache.org/jira/browse/HIVE-4867, an optimization of 
 deduplication of columns is done. But, when one of the columns is used as 
 part of partitioned/distribute by, its not taken care of.  
 The above query produces exception as follows:
 {code}
 Diagnostic Messages for this Task:
 java.lang.RuntimeException: org.apache.hadoop.hive.ql.metadata.HiveException: 
 Hive Runtime Error while processing row 
 {n_nationkey:0,n_name:ALGERIA,n_regionkey:0,n_comment: haggle. 
 carefully final deposits detect slyly agai}
   at org.apache.hadoop.hive.ql.exec.mr.ExecMapper.map(ExecMapper.java:185)
   at org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:54)
   at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:450)
   at org.apache.hadoop.mapred.MapTask.run(MapTask.java:343)
   at 
 org.apache.hadoop.mapred.LocalContainerLauncher$EventHandler.runSubtask(LocalContainerLauncher.java:370)
   at 
 org.apache.hadoop.mapred.LocalContainerLauncher$EventHandler.runTask(LocalContainerLauncher.java:295)
   at 
 org.apache.hadoop.mapred.LocalContainerLauncher$EventHandler.access$200(LocalContainerLauncher.java:181)
   at 
 org.apache.hadoop.mapred.LocalContainerLauncher$EventHandler$1.run(LocalContainerLauncher.java:224)
   at 
 java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471)
   at java.util.concurrent.FutureTask.run(FutureTask.java:262)
   at 
 java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
   at 
 java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
   at java.lang.Thread.run(Thread.java:744)
 Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime 
 Error while processing row 
 {n_nationkey:0,n_name:ALGERIA,n_regionkey:0,n_comment: haggle. 
 carefully final deposits detect slyly agai}
   at 
 org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:503)
   at org.apache.hadoop.hive.ql.exec.mr.ExecMapper.map(ExecMapper.java:176)
   ... 12 more
 Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: 
 java.lang.RuntimeException: cannot find field _col2 from [0:_col0]
   at 
 org.apache.hadoop.hive.ql.exec.ReduceSinkOperator.processOp(ReduceSinkOperator.java:397)
   at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:815)
   at 
 org.apache.hadoop.hive.ql.exec.SelectOperator.processOp(SelectOperator.java:84)
   at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:815)
   at 
 org.apache.hadoop.hive.ql.exec.TableScanOperator.processOp(TableScanOperator.java:95)
   at 
 org.apache.hadoop.hive.ql.exec.MapOperator$MapOpCtx.forward(MapOperator.java:157)
   at 
 org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:493)
   ... 13 more
 Caused by: java.lang.RuntimeException: cannot find field _col2 from [0:_col0]
   at 
 org.apache.hadoop.hive.serde2.objectinspector.ObjectInspectorUtils.getStandardStructFieldRef(ObjectInspectorUtils.java:410)
   at 
 org.apache.hadoop.hive.serde2.objectinspector.StandardStructObjectInspector.getStructFieldRef(StandardStructObjectInspector.java:147)
   at 
 org.apache.hadoop.hive.ql.exec.ExprNodeColumnEvaluator.initialize(ExprNodeColumnEvaluator.java:55)
   at 
 org.apache.hadoop.hive.ql.exec.Operator.initEvaluators(Operator.java:954)
   at 
 org.apache.hadoop.hive.ql.exec.ReduceSinkOperator.processOp(ReduceSinkOperator.java:325)
   ... 19 more
 {code}
 Tables used are: 
 {code}
 CREATE EXTERNAL TABLE `nation`(
   `n_nationkey` int,
   `n_name` string,
   `n_regionkey` int,
   `n_comment` string)
 ROW FORMAT DELIMITED
   FIELDS 

  1   2   >