[jira] [Commented] (HIVE-3251) Hive doesn't remove scrach directories while killing running MR job

2012-07-19 Thread Namit Jain (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-3251?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13418109#comment-13418109
 ] 

Namit Jain commented on HIVE-3251:
--

+1

 Hive doesn't remove scrach directories while killing running MR job
 ---

 Key: HIVE-3251
 URL: https://issues.apache.org/jira/browse/HIVE-3251
 Project: Hive
  Issue Type: Bug
  Components: Server Infrastructure
Reporter: Gang Tim Liu
Assignee: Gang Tim Liu
 Attachments: HIVE-3151.patch


 While killing running MR job, hive doesn't clean up scratch directory 
 (mapred.cache.files). So that, afterwards, scratch directory is left there in 
 hdfs. HDFS name node doesn't know it and try to do lease recovery. while such 
 instances happen more, it will eventually crash namenode.
 The fix is to leverage hdfs clean up functionality. While creating scratch 
 dirs, hive registers it to hdfs cleanup hook. While killing happens, hdfs 
 will clean them up.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HIVE-3221) HiveConf.getPositionFromInternalName does not support more than sinle digit column numbers

2012-07-19 Thread Ashutosh Chauhan (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-3221?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13418119#comment-13418119
 ] 

Ashutosh Chauhan commented on HIVE-3221:


+1. Looks good. Will commit if tests pass.

 HiveConf.getPositionFromInternalName does not support more than sinle digit 
 column numbers
 --

 Key: HIVE-3221
 URL: https://issues.apache.org/jira/browse/HIVE-3221
 Project: Hive
  Issue Type: Bug
Reporter: Sushanth Sowmyan
Assignee: Sushanth Sowmyan
 Attachments: HIVE-3221.patch


 For positions above 9, HiveConf.getPositionFromInternalName only looks at the 
 last digit, and thus, causes collisions.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HIVE-3205) Bucketed mapjoin on partitioned table which has no partition throws NPE

2012-07-19 Thread Namit Jain (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-3205?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13418133#comment-13418133
 ] 

Namit Jain commented on HIVE-3205:
--

very minor comments - mostly looks good

 Bucketed mapjoin on partitioned table which has no partition throws NPE
 ---

 Key: HIVE-3205
 URL: https://issues.apache.org/jira/browse/HIVE-3205
 Project: Hive
  Issue Type: Bug
  Components: Query Processor
Affects Versions: 0.10.0
 Environment: ubuntu 10.04
Reporter: Navis
Assignee: Navis
Priority: Minor

 {code}
 create table hive_test_smb_bucket1 (key int, value string) partitioned by (ds 
 string) clustered by (key) sorted by (key) into 2 buckets;
 create table hive_test_smb_bucket2 (key int, value string) partitioned by (ds 
 string) clustered by (key) sorted by (key) into 2 buckets;
 set hive.optimize.bucketmapjoin = true;
 set hive.input.format = 
 org.apache.hadoop.hive.ql.io.BucketizedHiveInputFormat;
 explain
 SELECT /* + MAPJOIN(b) */ b.key as k1, b.value, b.ds, a.key as k2
 FROM hive_test_smb_bucket1 a JOIN
 hive_test_smb_bucket2 b
 ON a.key = b.key WHERE a.ds = '2010-10-15' and b.ds='2010-10-15' and  b.key 
 IS NOT NULL;
 {code}
 throws NPE
 {noformat}
 2012-06-28 08:59:13,459 ERROR ql.Driver (SessionState.java:printError(400)) - 
 FAILED: NullPointerException null
 java.lang.NullPointerException
   at 
 org.apache.hadoop.hive.ql.optimizer.BucketMapJoinOptimizer$BucketMapjoinOptProc.process(BucketMapJoinOptimizer.java:269)
   at 
 org.apache.hadoop.hive.ql.lib.DefaultRuleDispatcher.dispatch(DefaultRuleDispatcher.java:89)
   at 
 org.apache.hadoop.hive.ql.lib.DefaultGraphWalker.dispatch(DefaultGraphWalker.java:88)
   at 
 org.apache.hadoop.hive.ql.lib.DefaultGraphWalker.walk(DefaultGraphWalker.java:125)
   at 
 org.apache.hadoop.hive.ql.lib.DefaultGraphWalker.startWalking(DefaultGraphWalker.java:102)
   at 
 org.apache.hadoop.hive.ql.optimizer.BucketMapJoinOptimizer.transform(BucketMapJoinOptimizer.java:100)
   at 
 org.apache.hadoop.hive.ql.optimizer.Optimizer.optimize(Optimizer.java:87)
   at 
 org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.analyzeInternal(SemanticAnalyzer.java:7564)
   at 
 org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer.analyze(BaseSemanticAnalyzer.java:245)
   at 
 org.apache.hadoop.hive.ql.parse.ExplainSemanticAnalyzer.analyzeInternal(ExplainSemanticAnalyzer.java:50)
   at 
 org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer.analyze(BaseSemanticAnalyzer.java:245)
   at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:430)
   at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:335)
   at org.apache.hadoop.hive.ql.Driver.run(Driver.java:902)
   at 
 org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:258)
   at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:215)
   at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:406)
   at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:744)
   at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:607)
   at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
   at 
 sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
   at 
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
   at java.lang.reflect.Method.invoke(Method.java:597)
   at org.apache.hadoop.util.RunJar.main(RunJar.java:186)
 {noformat}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Updated] (HIVE-3205) Bucketed mapjoin on partitioned table which has no partition throws NPE

2012-07-19 Thread Namit Jain (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-3205?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Namit Jain updated HIVE-3205:
-

Status: Open  (was: Patch Available)

 Bucketed mapjoin on partitioned table which has no partition throws NPE
 ---

 Key: HIVE-3205
 URL: https://issues.apache.org/jira/browse/HIVE-3205
 Project: Hive
  Issue Type: Bug
  Components: Query Processor
Affects Versions: 0.10.0
 Environment: ubuntu 10.04
Reporter: Navis
Assignee: Navis
Priority: Minor

 {code}
 create table hive_test_smb_bucket1 (key int, value string) partitioned by (ds 
 string) clustered by (key) sorted by (key) into 2 buckets;
 create table hive_test_smb_bucket2 (key int, value string) partitioned by (ds 
 string) clustered by (key) sorted by (key) into 2 buckets;
 set hive.optimize.bucketmapjoin = true;
 set hive.input.format = 
 org.apache.hadoop.hive.ql.io.BucketizedHiveInputFormat;
 explain
 SELECT /* + MAPJOIN(b) */ b.key as k1, b.value, b.ds, a.key as k2
 FROM hive_test_smb_bucket1 a JOIN
 hive_test_smb_bucket2 b
 ON a.key = b.key WHERE a.ds = '2010-10-15' and b.ds='2010-10-15' and  b.key 
 IS NOT NULL;
 {code}
 throws NPE
 {noformat}
 2012-06-28 08:59:13,459 ERROR ql.Driver (SessionState.java:printError(400)) - 
 FAILED: NullPointerException null
 java.lang.NullPointerException
   at 
 org.apache.hadoop.hive.ql.optimizer.BucketMapJoinOptimizer$BucketMapjoinOptProc.process(BucketMapJoinOptimizer.java:269)
   at 
 org.apache.hadoop.hive.ql.lib.DefaultRuleDispatcher.dispatch(DefaultRuleDispatcher.java:89)
   at 
 org.apache.hadoop.hive.ql.lib.DefaultGraphWalker.dispatch(DefaultGraphWalker.java:88)
   at 
 org.apache.hadoop.hive.ql.lib.DefaultGraphWalker.walk(DefaultGraphWalker.java:125)
   at 
 org.apache.hadoop.hive.ql.lib.DefaultGraphWalker.startWalking(DefaultGraphWalker.java:102)
   at 
 org.apache.hadoop.hive.ql.optimizer.BucketMapJoinOptimizer.transform(BucketMapJoinOptimizer.java:100)
   at 
 org.apache.hadoop.hive.ql.optimizer.Optimizer.optimize(Optimizer.java:87)
   at 
 org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.analyzeInternal(SemanticAnalyzer.java:7564)
   at 
 org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer.analyze(BaseSemanticAnalyzer.java:245)
   at 
 org.apache.hadoop.hive.ql.parse.ExplainSemanticAnalyzer.analyzeInternal(ExplainSemanticAnalyzer.java:50)
   at 
 org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer.analyze(BaseSemanticAnalyzer.java:245)
   at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:430)
   at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:335)
   at org.apache.hadoop.hive.ql.Driver.run(Driver.java:902)
   at 
 org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:258)
   at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:215)
   at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:406)
   at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:744)
   at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:607)
   at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
   at 
 sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
   at 
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
   at java.lang.reflect.Method.invoke(Method.java:597)
   at org.apache.hadoop.util.RunJar.main(RunJar.java:186)
 {noformat}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Updated] (HIVE-3070) Filter on outer join condition removed while merging join tree

2012-07-19 Thread Navis (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-3070?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Navis updated HIVE-3070:


Affects Version/s: 0.10.0
   0.9.0
 Assignee: Navis
  Summary: Filter on outer join condition removed while merging 
join tree  (was: outer join with filter over multiple tables on the same column 
have bug)

 Filter on outer join condition removed while merging join tree
 --

 Key: HIVE-3070
 URL: https://issues.apache.org/jira/browse/HIVE-3070
 Project: Hive
  Issue Type: Bug
  Components: SQL
Affects Versions: 0.8.1, 0.9.0, 0.10.0
Reporter: Zhang Xinyu
Assignee: Navis

 should the result of query A: 
 select s.aa, s.bb, c.key keyc from (select a.key aa, b.key bb from src a left 
 outer join src b on a.key=b.key) s left outer join src c on s.bb=c.key and 
 s.bb10 where s.aa20;
 be the same as query B:
 select a.key keya, b.key keyb, c.key keyc from src a left outer join src b on 
 a.key=b.key left outer join src c on b.key=c.key and b.key10 where a.key20;
 ?
 Currently, the result is different, query B gets wrong result!
 In SemanticAnalyzer.java, mergeJoins():
 ArrayListArrayListASTNode filters = target.getFilters();
 for (int i = 0; i  nodeRightAliases.length; i++) {
   filters.add(node.getFilters().get(i + 1));
 }
 filters in node.getFilters().get(0) are lost.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Updated] (HIVE-3070) Filter on outer join condition removed while merging join tree

2012-07-19 Thread Navis (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-3070?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Navis updated HIVE-3070:


Status: Patch Available  (was: Open)

https://reviews.facebook.net/D4239

 Filter on outer join condition removed while merging join tree
 --

 Key: HIVE-3070
 URL: https://issues.apache.org/jira/browse/HIVE-3070
 Project: Hive
  Issue Type: Bug
  Components: SQL
Affects Versions: 0.9.0, 0.8.1, 0.10.0
Reporter: Zhang Xinyu
Assignee: Navis

 should the result of query A: 
 select s.aa, s.bb, c.key keyc from (select a.key aa, b.key bb from src a left 
 outer join src b on a.key=b.key) s left outer join src c on s.bb=c.key and 
 s.bb10 where s.aa20;
 be the same as query B:
 select a.key keya, b.key keyb, c.key keyc from src a left outer join src b on 
 a.key=b.key left outer join src c on b.key=c.key and b.key10 where a.key20;
 ?
 Currently, the result is different, query B gets wrong result!
 In SemanticAnalyzer.java, mergeJoins():
 ArrayListArrayListASTNode filters = target.getFilters();
 for (int i = 0; i  nodeRightAliases.length; i++) {
   filters.add(node.getFilters().get(i + 1));
 }
 filters in node.getFilters().get(0) are lost.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Updated] (HIVE-3205) Bucketed mapjoin on partitioned table which has no partition throws NPE

2012-07-19 Thread Navis (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-3205?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Navis updated HIVE-3205:


Status: Patch Available  (was: Open)

 Bucketed mapjoin on partitioned table which has no partition throws NPE
 ---

 Key: HIVE-3205
 URL: https://issues.apache.org/jira/browse/HIVE-3205
 Project: Hive
  Issue Type: Bug
  Components: Query Processor
Affects Versions: 0.10.0
 Environment: ubuntu 10.04
Reporter: Navis
Assignee: Navis
Priority: Minor

 {code}
 create table hive_test_smb_bucket1 (key int, value string) partitioned by (ds 
 string) clustered by (key) sorted by (key) into 2 buckets;
 create table hive_test_smb_bucket2 (key int, value string) partitioned by (ds 
 string) clustered by (key) sorted by (key) into 2 buckets;
 set hive.optimize.bucketmapjoin = true;
 set hive.input.format = 
 org.apache.hadoop.hive.ql.io.BucketizedHiveInputFormat;
 explain
 SELECT /* + MAPJOIN(b) */ b.key as k1, b.value, b.ds, a.key as k2
 FROM hive_test_smb_bucket1 a JOIN
 hive_test_smb_bucket2 b
 ON a.key = b.key WHERE a.ds = '2010-10-15' and b.ds='2010-10-15' and  b.key 
 IS NOT NULL;
 {code}
 throws NPE
 {noformat}
 2012-06-28 08:59:13,459 ERROR ql.Driver (SessionState.java:printError(400)) - 
 FAILED: NullPointerException null
 java.lang.NullPointerException
   at 
 org.apache.hadoop.hive.ql.optimizer.BucketMapJoinOptimizer$BucketMapjoinOptProc.process(BucketMapJoinOptimizer.java:269)
   at 
 org.apache.hadoop.hive.ql.lib.DefaultRuleDispatcher.dispatch(DefaultRuleDispatcher.java:89)
   at 
 org.apache.hadoop.hive.ql.lib.DefaultGraphWalker.dispatch(DefaultGraphWalker.java:88)
   at 
 org.apache.hadoop.hive.ql.lib.DefaultGraphWalker.walk(DefaultGraphWalker.java:125)
   at 
 org.apache.hadoop.hive.ql.lib.DefaultGraphWalker.startWalking(DefaultGraphWalker.java:102)
   at 
 org.apache.hadoop.hive.ql.optimizer.BucketMapJoinOptimizer.transform(BucketMapJoinOptimizer.java:100)
   at 
 org.apache.hadoop.hive.ql.optimizer.Optimizer.optimize(Optimizer.java:87)
   at 
 org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.analyzeInternal(SemanticAnalyzer.java:7564)
   at 
 org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer.analyze(BaseSemanticAnalyzer.java:245)
   at 
 org.apache.hadoop.hive.ql.parse.ExplainSemanticAnalyzer.analyzeInternal(ExplainSemanticAnalyzer.java:50)
   at 
 org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer.analyze(BaseSemanticAnalyzer.java:245)
   at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:430)
   at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:335)
   at org.apache.hadoop.hive.ql.Driver.run(Driver.java:902)
   at 
 org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:258)
   at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:215)
   at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:406)
   at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:744)
   at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:607)
   at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
   at 
 sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
   at 
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
   at java.lang.reflect.Method.invoke(Method.java:597)
   at org.apache.hadoop.util.RunJar.main(RunJar.java:186)
 {noformat}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HIVE-3205) Bucketed mapjoin on partitioned table which has no partition throws NPE

2012-07-19 Thread Namit Jain (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-3205?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13418178#comment-13418178
 ] 

Namit Jain commented on HIVE-3205:
--

+1

 Bucketed mapjoin on partitioned table which has no partition throws NPE
 ---

 Key: HIVE-3205
 URL: https://issues.apache.org/jira/browse/HIVE-3205
 Project: Hive
  Issue Type: Bug
  Components: Query Processor
Affects Versions: 0.10.0
 Environment: ubuntu 10.04
Reporter: Navis
Assignee: Navis
Priority: Minor

 {code}
 create table hive_test_smb_bucket1 (key int, value string) partitioned by (ds 
 string) clustered by (key) sorted by (key) into 2 buckets;
 create table hive_test_smb_bucket2 (key int, value string) partitioned by (ds 
 string) clustered by (key) sorted by (key) into 2 buckets;
 set hive.optimize.bucketmapjoin = true;
 set hive.input.format = 
 org.apache.hadoop.hive.ql.io.BucketizedHiveInputFormat;
 explain
 SELECT /* + MAPJOIN(b) */ b.key as k1, b.value, b.ds, a.key as k2
 FROM hive_test_smb_bucket1 a JOIN
 hive_test_smb_bucket2 b
 ON a.key = b.key WHERE a.ds = '2010-10-15' and b.ds='2010-10-15' and  b.key 
 IS NOT NULL;
 {code}
 throws NPE
 {noformat}
 2012-06-28 08:59:13,459 ERROR ql.Driver (SessionState.java:printError(400)) - 
 FAILED: NullPointerException null
 java.lang.NullPointerException
   at 
 org.apache.hadoop.hive.ql.optimizer.BucketMapJoinOptimizer$BucketMapjoinOptProc.process(BucketMapJoinOptimizer.java:269)
   at 
 org.apache.hadoop.hive.ql.lib.DefaultRuleDispatcher.dispatch(DefaultRuleDispatcher.java:89)
   at 
 org.apache.hadoop.hive.ql.lib.DefaultGraphWalker.dispatch(DefaultGraphWalker.java:88)
   at 
 org.apache.hadoop.hive.ql.lib.DefaultGraphWalker.walk(DefaultGraphWalker.java:125)
   at 
 org.apache.hadoop.hive.ql.lib.DefaultGraphWalker.startWalking(DefaultGraphWalker.java:102)
   at 
 org.apache.hadoop.hive.ql.optimizer.BucketMapJoinOptimizer.transform(BucketMapJoinOptimizer.java:100)
   at 
 org.apache.hadoop.hive.ql.optimizer.Optimizer.optimize(Optimizer.java:87)
   at 
 org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.analyzeInternal(SemanticAnalyzer.java:7564)
   at 
 org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer.analyze(BaseSemanticAnalyzer.java:245)
   at 
 org.apache.hadoop.hive.ql.parse.ExplainSemanticAnalyzer.analyzeInternal(ExplainSemanticAnalyzer.java:50)
   at 
 org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer.analyze(BaseSemanticAnalyzer.java:245)
   at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:430)
   at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:335)
   at org.apache.hadoop.hive.ql.Driver.run(Driver.java:902)
   at 
 org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:258)
   at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:215)
   at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:406)
   at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:744)
   at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:607)
   at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
   at 
 sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
   at 
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
   at java.lang.reflect.Method.invoke(Method.java:597)
   at org.apache.hadoop.util.RunJar.main(RunJar.java:186)
 {noformat}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Updated] (HIVE-3271) Previledge can be granted by any user(not owner) to any user(even to the same user)

2012-07-19 Thread Unnikrishnan V T (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-3271?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Unnikrishnan V T updated HIVE-3271:
---


I have created two users user 'unni' and user 'sachin'. user unni created a 
table 'test3' so that user sachin cannot view that table. But user sachin is 
able to grant all permission to the table test3.
I have set
1)hive.security.authorization.enabled=true(in hive-site.xml)
2)dfs.permissions=true(in hdfs-site.xml)
3)dfs.permissions.supergroup=supergroup(in hdfs-site.xml)
user unni is in supergroup group.
The user sachin is even able to revoke all permissions from the owner of the 
table user unni. User sachin is not a member of supergroup.

 Previledge can be granted by any user(not owner) to any user(even to the same 
 user)
 ---

 Key: HIVE-3271
 URL: https://issues.apache.org/jira/browse/HIVE-3271
 Project: Hive
  Issue Type: Bug
  Components: Authorization
Affects Versions: 0.8.1
Reporter: Unnikrishnan V T
Priority: Critical
 Attachments: Screenshot.png


 I have created two users user 'unni' and user 'sachin'. user unni created a 
 table 'test3' so that user sachin cannot view that table. But user sachin is 
 able to grant all permission to the table test3.
 I have set 
 1)hive.security.authorization.enabled=true(in hive-site.xml)
 2)dfs.permissions=true(in hdfs-site.xml)
 3)dfs.permissions.supergroup=supergroup(in hdfs-site.xml)
 User sachin and user unni are in supergroup group.
 The user sachin is even able to revoke all permissions from the owner of the 
 table user unni.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HIVE-3246) java primitive type for binary datatype should be byte[]

2012-07-19 Thread Edward Capriolo (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-3246?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13418208#comment-13418208
 ] 

Edward Capriolo commented on HIVE-3246:
---

We have to do a release note it people who write binary UDFs are going to be 
effected by the change.

https://github.com/edwardcapriolo/hive_cassandra_udfs

It seems like if you wrote a UDF that returns binary you may be effected 
depending on how you wrote it.

 java primitive type for binary datatype should be byte[]
 

 Key: HIVE-3246
 URL: https://issues.apache.org/jira/browse/HIVE-3246
 Project: Hive
  Issue Type: Bug
Affects Versions: 0.9.0
Reporter: Thejas M Nair
Assignee: Thejas M Nair
 Attachments: HIVE-3246.1.patch, HIVE-3246.2.patch


 PrimitiveObjectInspector.getPrimitiveJavaObject is supposed to return a java 
 object. But in case of binary datatype, it returns ByteArrayRef (not java 
 standard type). The suitable java object for it would be byte[]. 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HIVE-3273) Add avro jars into hive execution classpath

2012-07-19 Thread Edward Capriolo (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-3273?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13418210#comment-13418210
 ] 

Edward Capriolo commented on HIVE-3273:
---

I can see this happening but when exactly do we get this error?

 Add avro jars into hive execution classpath
 ---

 Key: HIVE-3273
 URL: https://issues.apache.org/jira/browse/HIVE-3273
 Project: Hive
  Issue Type: Bug
Reporter: Zhenxiao Luo
Assignee: Zhenxiao Luo
 Attachments: HIVE-3273.1.patch.txt


 avro*.jar should be added to hive execution classpath

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HIVE-3098) Memory leak from large number of FileSystem instances in FileSystem.CACHE. (Must cache UGIs.)

2012-07-19 Thread Daryn Sharp (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-3098?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13418294#comment-13418294
 ] 

Daryn Sharp commented on HIVE-3098:
---

It's probably safer to use the iterator's remove instead of a direct remove 
from the map.  Ie. change {{trashedUGIs.add(ugiCache.remove(entry.getKey())}} 
to {{entry.remove(); trashedUGIs.add(entry.getKey)}}.

How does this protect against requests using trashed UGIs having their 
filesystems closed before the request is complete?  Is it assuming all requests 
will complete within the time interval between which the ugi is trashed and 
then purged?  If true, it seems fragile esp. when considering rpc retries, HA 
failover delays, etc.

Has anyone benchmarked cached ugis to ensure this isn't a premature 
optimization?

 Memory leak from large number of FileSystem instances in FileSystem.CACHE. 
 (Must cache UGIs.)
 -

 Key: HIVE-3098
 URL: https://issues.apache.org/jira/browse/HIVE-3098
 Project: Hive
  Issue Type: Bug
  Components: Shims
Affects Versions: 0.9.0
 Environment: Running with Hadoop 20.205.0.3+ / 1.0.x with security 
 turned on.
Reporter: Mithun Radhakrishnan
Assignee: Mithun Radhakrishnan
 Attachments: Hive_3098.patch


 The problem manifested from stress-testing HCatalog 0.4.1 (as part of testing 
 the Oracle backend).
 The HCatalog server ran out of memory (-Xmx2048m) when pounded by 60-threads, 
 in under 24 hours. The heap-dump indicates that hadoop::FileSystem.CACHE had 
 100 instances of FileSystem, whose combined retained-mem consumed the 
 entire heap.
 It boiled down to hadoop::UserGroupInformation::equals() being implemented 
 such that the Subject member is compared for equality (==), and not 
 equivalence (.equals()). This causes equivalent UGI instances to compare as 
 unequal, and causes a new FileSystem instance to be created and cached.
 The UGI.equals() is so implemented, incidentally, as a fix for yet another 
 problem (HADOOP-6670); so it is unlikely that that implementation can be 
 modified.
 The solution for this is to check for UGI equivalence in HCatalog (i.e. in 
 the Hive metastore), using an cache for UGI instances in the shims.
 I have a patch to fix this. I'll upload it shortly. I just ran an overnight 
 test to confirm that the memory-leak has been arrested.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




Build failed in Jenkins: Hive-0.9.1-SNAPSHOT-h0.21-keepgoing=false #80

2012-07-19 Thread Apache Jenkins Server
See 
https://builds.apache.org/job/Hive-0.9.1-SNAPSHOT-h0.21-keepgoing=false/80/

--
[...truncated 10160 lines...]
 [echo] Project: odbc
 [copy] Warning: 
https://builds.apache.org/job/Hive-0.9.1-SNAPSHOT-h0.21-keepgoing=false/ws/hive/odbc/src/conf
 does not exist.

ivy-resolve-test:
 [echo] Project: odbc

ivy-retrieve-test:
 [echo] Project: odbc

compile-test:
 [echo] Project: odbc

create-dirs:
 [echo] Project: serde
 [copy] Warning: 
https://builds.apache.org/job/Hive-0.9.1-SNAPSHOT-h0.21-keepgoing=false/ws/hive/serde/src/test/resources
 does not exist.

init:
 [echo] Project: serde

ivy-init-settings:
 [echo] Project: serde

ivy-resolve:
 [echo] Project: serde
[ivy:resolve] :: loading settings :: file = 
https://builds.apache.org/job/Hive-0.9.1-SNAPSHOT-h0.21-keepgoing=false/ws/hive/ivy/ivysettings.xml
[ivy:report] Processing 
https://builds.apache.org/job/Hive-0.9.1-SNAPSHOT-h0.21-keepgoing=false/80/artifact/hive/build/ivy/resolution-cache/org.apache.hive-hive-serde-default.xml
 to 
https://builds.apache.org/job/Hive-0.9.1-SNAPSHOT-h0.21-keepgoing=false/80/artifact/hive/build/ivy/report/org.apache.hive-hive-serde-default.html

ivy-retrieve:
 [echo] Project: serde

dynamic-serde:

compile:
 [echo] Project: serde

ivy-resolve-test:
 [echo] Project: serde

ivy-retrieve-test:
 [echo] Project: serde

compile-test:
 [echo] Project: serde
[javac] Compiling 26 source files to 
https://builds.apache.org/job/Hive-0.9.1-SNAPSHOT-h0.21-keepgoing=false/80/artifact/hive/build/serde/test/classes
[javac] Note: Some input files use or override a deprecated API.
[javac] Note: Recompile with -Xlint:deprecation for details.
[javac] Note: Some input files use unchecked or unsafe operations.
[javac] Note: Recompile with -Xlint:unchecked for details.

create-dirs:
 [echo] Project: service
 [copy] Warning: 
https://builds.apache.org/job/Hive-0.9.1-SNAPSHOT-h0.21-keepgoing=false/ws/hive/service/src/test/resources
 does not exist.

init:
 [echo] Project: service

ivy-init-settings:
 [echo] Project: service

ivy-resolve:
 [echo] Project: service
[ivy:resolve] :: loading settings :: file = 
https://builds.apache.org/job/Hive-0.9.1-SNAPSHOT-h0.21-keepgoing=false/ws/hive/ivy/ivysettings.xml
[ivy:report] Processing 
https://builds.apache.org/job/Hive-0.9.1-SNAPSHOT-h0.21-keepgoing=false/80/artifact/hive/build/ivy/resolution-cache/org.apache.hive-hive-service-default.xml
 to 
https://builds.apache.org/job/Hive-0.9.1-SNAPSHOT-h0.21-keepgoing=false/80/artifact/hive/build/ivy/report/org.apache.hive-hive-service-default.html

ivy-retrieve:
 [echo] Project: service

compile:
 [echo] Project: service

ivy-resolve-test:
 [echo] Project: service

ivy-retrieve-test:
 [echo] Project: service

compile-test:
 [echo] Project: service
[javac] Compiling 2 source files to 
https://builds.apache.org/job/Hive-0.9.1-SNAPSHOT-h0.21-keepgoing=false/80/artifact/hive/build/service/test/classes

test:
 [echo] Project: hive

test-shims:
 [echo] Project: hive

test-conditions:
 [echo] Project: shims

gen-test:
 [echo] Project: shims

create-dirs:
 [echo] Project: shims
 [copy] Warning: 
https://builds.apache.org/job/Hive-0.9.1-SNAPSHOT-h0.21-keepgoing=false/ws/hive/shims/src/test/resources
 does not exist.

init:
 [echo] Project: shims

ivy-init-settings:
 [echo] Project: shims

ivy-resolve:
 [echo] Project: shims
[ivy:resolve] :: loading settings :: file = 
https://builds.apache.org/job/Hive-0.9.1-SNAPSHOT-h0.21-keepgoing=false/ws/hive/ivy/ivysettings.xml
[ivy:report] Processing 
https://builds.apache.org/job/Hive-0.9.1-SNAPSHOT-h0.21-keepgoing=false/80/artifact/hive/build/ivy/resolution-cache/org.apache.hive-hive-shims-default.xml
 to 
https://builds.apache.org/job/Hive-0.9.1-SNAPSHOT-h0.21-keepgoing=false/80/artifact/hive/build/ivy/report/org.apache.hive-hive-shims-default.html

ivy-retrieve:
 [echo] Project: shims

compile:
 [echo] Project: shims
 [echo] Building shims 0.20

build_shims:
 [echo] Project: shims
 [echo] Compiling 
https://builds.apache.org/job/Hive-0.9.1-SNAPSHOT-h0.21-keepgoing=false/ws/hive/shims/src/common/java;/home/hudson/hudson-slave/workspace/Hive-0.9.1-SNAPSHOT-h0.21-keepgoing=false/hive/shims/src/0.20/java
 against hadoop 0.20.2 
(https://builds.apache.org/job/Hive-0.9.1-SNAPSHOT-h0.21-keepgoing=false/80/artifact/hive/build/hadoopcore/hadoop-0.20.2)

ivy-init-settings:
 [echo] Project: shims

ivy-resolve-hadoop-shim:
 [echo] Project: shims
[ivy:resolve] :: loading settings :: file = 
https://builds.apache.org/job/Hive-0.9.1-SNAPSHOT-h0.21-keepgoing=false/ws/hive/ivy/ivysettings.xml

ivy-retrieve-hadoop-shim:
 [echo] Project: shims
 [echo] Building shims 0.20S

build_shims:
 [echo] Project: shims
 [echo] Compiling 

[jira] [Commented] (HIVE-3098) Memory leak from large number of FileSystem instances in FileSystem.CACHE. (Must cache UGIs.)

2012-07-19 Thread Mithun Radhakrishnan (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-3098?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13418441#comment-13418441
 ] 

Mithun Radhakrishnan commented on HIVE-3098:


Hey, All. First off, I'm considering a rewrite, essentially identical but 
implemented via Guava caching. This will save on lines-of-code.

@Ashutosh: Sure. I'll put a new patch on Phabricator, as soon as I've 
incorporated Rohini's comments. The code could stand to be refactored.

@Daryn: Thank you for reviewing.
1. If the concern is over ConcurrentModificationExceptions, that's not a 
problem since ugiCache is a ConcurrentHashMap. :]
2. Yes, that was the assumption, but you have a point. Perhaps a better way to 
deal with this is through WeakReferences (or some such).
3. I'm not sure I follow your concern about premature-optimization. We're 
caching to avoid the creation of multiple FS instances. We've seen from tests 
that this solves that problem. Would you please elaborate?

 Memory leak from large number of FileSystem instances in FileSystem.CACHE. 
 (Must cache UGIs.)
 -

 Key: HIVE-3098
 URL: https://issues.apache.org/jira/browse/HIVE-3098
 Project: Hive
  Issue Type: Bug
  Components: Shims
Affects Versions: 0.9.0
 Environment: Running with Hadoop 20.205.0.3+ / 1.0.x with security 
 turned on.
Reporter: Mithun Radhakrishnan
Assignee: Mithun Radhakrishnan
 Attachments: Hive_3098.patch


 The problem manifested from stress-testing HCatalog 0.4.1 (as part of testing 
 the Oracle backend).
 The HCatalog server ran out of memory (-Xmx2048m) when pounded by 60-threads, 
 in under 24 hours. The heap-dump indicates that hadoop::FileSystem.CACHE had 
 100 instances of FileSystem, whose combined retained-mem consumed the 
 entire heap.
 It boiled down to hadoop::UserGroupInformation::equals() being implemented 
 such that the Subject member is compared for equality (==), and not 
 equivalence (.equals()). This causes equivalent UGI instances to compare as 
 unequal, and causes a new FileSystem instance to be created and cached.
 The UGI.equals() is so implemented, incidentally, as a fix for yet another 
 problem (HADOOP-6670); so it is unlikely that that implementation can be 
 modified.
 The solution for this is to check for UGI equivalence in HCatalog (i.e. in 
 the Hive metastore), using an cache for UGI instances in the shims.
 I have a patch to fix this. I'll upload it shortly. I just ran an overnight 
 test to confirm that the memory-leak has been arrested.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Resolved] (HIVE-3251) Hive doesn't remove scrach directories while killing running MR job

2012-07-19 Thread Namit Jain (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-3251?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Namit Jain resolved HIVE-3251.
--

  Resolution: Fixed
Hadoop Flags: Reviewed

Committed. Thanks Tim

 Hive doesn't remove scrach directories while killing running MR job
 ---

 Key: HIVE-3251
 URL: https://issues.apache.org/jira/browse/HIVE-3251
 Project: Hive
  Issue Type: Bug
  Components: Server Infrastructure
Reporter: Gang Tim Liu
Assignee: Gang Tim Liu
 Attachments: HIVE-3151.patch


 While killing running MR job, hive doesn't clean up scratch directory 
 (mapred.cache.files). So that, afterwards, scratch directory is left there in 
 hdfs. HDFS name node doesn't know it and try to do lease recovery. while such 
 instances happen more, it will eventually crash namenode.
 The fix is to leverage hdfs clean up functionality. While creating scratch 
 dirs, hive registers it to hdfs cleanup hook. While killing happens, hdfs 
 will clean them up.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HIVE-3246) java primitive type for binary datatype should be byte[]

2012-07-19 Thread Ashutosh Chauhan (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-3246?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13418447#comment-13418447
 ] 

Ashutosh Chauhan commented on HIVE-3246:


Yeah. I will add the release notes while resolving this ticket.

 java primitive type for binary datatype should be byte[]
 

 Key: HIVE-3246
 URL: https://issues.apache.org/jira/browse/HIVE-3246
 Project: Hive
  Issue Type: Bug
Affects Versions: 0.9.0
Reporter: Thejas M Nair
Assignee: Thejas M Nair
 Attachments: HIVE-3246.1.patch, HIVE-3246.2.patch


 PrimitiveObjectInspector.getPrimitiveJavaObject is supposed to return a java 
 object. But in case of binary datatype, it returns ByteArrayRef (not java 
 standard type). The suitable java object for it would be byte[]. 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Updated] (HIVE-3246) java primitive type for binary datatype should be byte[]

2012-07-19 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-3246?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-3246:
---

   Resolution: Fixed
Fix Version/s: 0.10.0
 Release Note: This fix changes the internal representation of binary type 
within Hive. UDFs which were using binary type and making use of the fact that 
java representation of binary data in Hive is ByteArrayRef needs to be updated 
to reflect that now that representation is byte[] instead. Also note that this 
doesn't change the format for on-disk data.  (was: Makes the java primitive 
type for binary datatype - byte[]
)
 Hadoop Flags: Incompatible change
   Status: Resolved  (was: Patch Available)

Committed to trunk. Thanks, Thejas!

 java primitive type for binary datatype should be byte[]
 

 Key: HIVE-3246
 URL: https://issues.apache.org/jira/browse/HIVE-3246
 Project: Hive
  Issue Type: Bug
Affects Versions: 0.9.0
Reporter: Thejas M Nair
Assignee: Thejas M Nair
 Fix For: 0.10.0

 Attachments: HIVE-3246.1.patch, HIVE-3246.2.patch


 PrimitiveObjectInspector.getPrimitiveJavaObject is supposed to return a java 
 object. But in case of binary datatype, it returns ByteArrayRef (not java 
 standard type). The suitable java object for it would be byte[]. 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HIVE-3273) Add avro jars into hive execution classpath

2012-07-19 Thread Zhenxiao Luo (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-3273?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13418478#comment-13418478
 ] 

Zhenxiao Luo commented on HIVE-3273:


@Edward:

I found this error when I am backporting HIVE-895 to an old version of 
Hive(CDH3), where I did not put avro jars into hive execution classpath, and 
kept having the following exception:

java.lang.ClassNotFoundException: org.apache.avro.io.DatumReader

We do not have this error on current trunk, since Hadoop classpath is not 
setting correctly in build-common.xml test target. I think this should be 
fixed in HIVE-3274.

I think fixing this would emphasis that it is not enough to put avro jars in 
the hive client classpath. Avro jars should be put in hive execution(the 
mapreduce job) classpath. This would prevent future bugs.

What do you think about it?

Thanks,
Zhenxiao

 Add avro jars into hive execution classpath
 ---

 Key: HIVE-3273
 URL: https://issues.apache.org/jira/browse/HIVE-3273
 Project: Hive
  Issue Type: Bug
Reporter: Zhenxiao Luo
Assignee: Zhenxiao Luo
 Attachments: HIVE-3273.1.patch.txt


 avro*.jar should be added to hive execution classpath

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HIVE-3098) Memory leak from large number of FileSystem instances in FileSystem.CACHE. (Must cache UGIs.)

2012-07-19 Thread Daryn Sharp (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-3098?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13418485#comment-13418485
 ] 

Daryn Sharp commented on HIVE-3098:
---

I think the large discussion has obscured the fact that there's nothing wrong 
with the fs api.  The leak is a result of hive misusing the fs api.  Simply 
calling {{FileSystem.closeAllForUGI(remoteUser)}} at the end of a request will 
stop hive from leaking.  My guess, and I could be wrong, is that trying to 
cache UGIs will provide a negligible performance increase.  Hence why I 
question if this is premature optimization.

 Memory leak from large number of FileSystem instances in FileSystem.CACHE. 
 (Must cache UGIs.)
 -

 Key: HIVE-3098
 URL: https://issues.apache.org/jira/browse/HIVE-3098
 Project: Hive
  Issue Type: Bug
  Components: Shims
Affects Versions: 0.9.0
 Environment: Running with Hadoop 20.205.0.3+ / 1.0.x with security 
 turned on.
Reporter: Mithun Radhakrishnan
Assignee: Mithun Radhakrishnan
 Attachments: Hive_3098.patch


 The problem manifested from stress-testing HCatalog 0.4.1 (as part of testing 
 the Oracle backend).
 The HCatalog server ran out of memory (-Xmx2048m) when pounded by 60-threads, 
 in under 24 hours. The heap-dump indicates that hadoop::FileSystem.CACHE had 
 100 instances of FileSystem, whose combined retained-mem consumed the 
 entire heap.
 It boiled down to hadoop::UserGroupInformation::equals() being implemented 
 such that the Subject member is compared for equality (==), and not 
 equivalence (.equals()). This causes equivalent UGI instances to compare as 
 unequal, and causes a new FileSystem instance to be created and cached.
 The UGI.equals() is so implemented, incidentally, as a fix for yet another 
 problem (HADOOP-6670); so it is unlikely that that implementation can be 
 modified.
 The solution for this is to check for UGI equivalence in HCatalog (i.e. in 
 the Hive metastore), using an cache for UGI instances in the shims.
 I have a patch to fix this. I'll upload it shortly. I just ran an overnight 
 test to confirm that the memory-leak has been arrested.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HIVE-3251) Hive doesn't remove scrach directories while killing running MR job

2012-07-19 Thread Gang Tim Liu (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-3251?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13418515#comment-13418515
 ] 

Gang Tim Liu commented on HIVE-3251:


Namit, thank you very much for review and commit.

 Hive doesn't remove scrach directories while killing running MR job
 ---

 Key: HIVE-3251
 URL: https://issues.apache.org/jira/browse/HIVE-3251
 Project: Hive
  Issue Type: Bug
  Components: Server Infrastructure
Reporter: Gang Tim Liu
Assignee: Gang Tim Liu
 Attachments: HIVE-3151.patch


 While killing running MR job, hive doesn't clean up scratch directory 
 (mapred.cache.files). So that, afterwards, scratch directory is left there in 
 hdfs. HDFS name node doesn't know it and try to do lease recovery. while such 
 instances happen more, it will eventually crash namenode.
 The fix is to leverage hdfs clean up functionality. While creating scratch 
 dirs, hive registers it to hdfs cleanup hook. While killing happens, hdfs 
 will clean them up.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HIVE-1653) Ability to enforce correct stats

2012-07-19 Thread Kevin Wilfong (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-1653?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13418545#comment-13418545
 ] 

Kevin Wilfong commented on HIVE-1653:
-

Commented on the diff.

 Ability to enforce correct stats
 

 Key: HIVE-1653
 URL: https://issues.apache.org/jira/browse/HIVE-1653
 Project: Hive
  Issue Type: Improvement
  Components: Query Processor
Reporter: Namit Jain
Assignee: Namit Jain

 This is a follow-up for https://issues.apache.org/jira/browse/HIVE-1361.
 If one of the mappers/reducers cannot publish stats, it may lead to wrong 
 aggregated stats.
 There should be a way to avoid this - at the least, a configuration variable 
 which fails the 
 task if stats cannot be published

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Updated] (HIVE-1653) Ability to enforce correct stats

2012-07-19 Thread Kevin Wilfong (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-1653?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kevin Wilfong updated HIVE-1653:


Status: Open  (was: Patch Available)

 Ability to enforce correct stats
 

 Key: HIVE-1653
 URL: https://issues.apache.org/jira/browse/HIVE-1653
 Project: Hive
  Issue Type: Improvement
  Components: Query Processor
Reporter: Namit Jain
Assignee: Namit Jain

 This is a follow-up for https://issues.apache.org/jira/browse/HIVE-1361.
 If one of the mappers/reducers cannot publish stats, it may lead to wrong 
 aggregated stats.
 There should be a way to avoid this - at the least, a configuration variable 
 which fails the 
 task if stats cannot be published

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HIVE-3261) alter the number of buckets for a non-empty partitioned table should not be allowed

2012-07-19 Thread Kevin Wilfong (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-3261?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13418550#comment-13418550
 ] 

Kevin Wilfong commented on HIVE-3261:
-

This seems ok for now, but could you file a follow up task to use partition 
metadata instead of table metadata wherever bucketing information is used, and 
mention that the error message added by this JIRA can be removed once that is 
done.

One small comment on the diff.

 alter the number of buckets for a non-empty partitioned table should not be 
 allowed
 ---

 Key: HIVE-3261
 URL: https://issues.apache.org/jira/browse/HIVE-3261
 Project: Hive
  Issue Type: Bug
Reporter: Namit Jain
Assignee: Namit Jain

 This is dangerous since the code uses the table metadata everywhere to get 
 the number of buckets

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Updated] (HIVE-3261) alter the number of buckets for a non-empty partitioned table should not be allowed

2012-07-19 Thread Carl Steinbach (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-3261?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Carl Steinbach updated HIVE-3261:
-

Status: Open  (was: Patch Available)

@Namit: I also added a small comment on phabricator. Thanks.

 alter the number of buckets for a non-empty partitioned table should not be 
 allowed
 ---

 Key: HIVE-3261
 URL: https://issues.apache.org/jira/browse/HIVE-3261
 Project: Hive
  Issue Type: Bug
Reporter: Namit Jain
Assignee: Namit Jain

 This is dangerous since the code uses the table metadata everywhere to get 
 the number of buckets

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Updated] (HIVE-3261) alter the number of buckets for a non-empty partitioned table should not be allowed

2012-07-19 Thread Carl Steinbach (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-3261?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Carl Steinbach updated HIVE-3261:
-

Component/s: Query Processor
 Labels: bucketing  (was: )

 alter the number of buckets for a non-empty partitioned table should not be 
 allowed
 ---

 Key: HIVE-3261
 URL: https://issues.apache.org/jira/browse/HIVE-3261
 Project: Hive
  Issue Type: Bug
  Components: Query Processor
Reporter: Namit Jain
Assignee: Namit Jain
  Labels: bucketing

 This is dangerous since the code uses the table metadata everywhere to get 
 the number of buckets

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Updated] (HIVE-3271) Previledge can be granted by any user(not owner) to any user(even to the same user)

2012-07-19 Thread Carl Steinbach (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-3271?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Carl Steinbach updated HIVE-3271:
-

Component/s: Security

 Previledge can be granted by any user(not owner) to any user(even to the same 
 user)
 ---

 Key: HIVE-3271
 URL: https://issues.apache.org/jira/browse/HIVE-3271
 Project: Hive
  Issue Type: Bug
  Components: Authorization, Security
Affects Versions: 0.8.1
Reporter: Unnikrishnan V T
Priority: Critical
 Attachments: Screenshot.png


 I have created two users user 'unni' and user 'sachin'. user unni created a 
 table 'test3' so that user sachin cannot view that table. But user sachin is 
 able to grant all permission to the table test3.
 I have set 
 1)hive.security.authorization.enabled=true(in hive-site.xml)
 2)dfs.permissions=true(in hdfs-site.xml)
 3)dfs.permissions.supergroup=supergroup(in hdfs-site.xml)
 User sachin and user unni are in supergroup group.
 The user sachin is even able to revoke all permissions from the owner of the 
 table user unni.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Updated] (HIVE-3271) Privilege can be granted by any user(not owner) to any user(even to the same user)

2012-07-19 Thread Carl Steinbach (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-3271?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Carl Steinbach updated HIVE-3271:
-

Summary: Privilege can be granted by any user(not owner) to any user(even 
to the same user)  (was: Previledge can be granted by any user(not owner) to 
any user(even to the same user))

 Privilege can be granted by any user(not owner) to any user(even to the same 
 user)
 --

 Key: HIVE-3271
 URL: https://issues.apache.org/jira/browse/HIVE-3271
 Project: Hive
  Issue Type: Bug
  Components: Authorization, Security
Affects Versions: 0.8.1
Reporter: Unnikrishnan V T
Priority: Critical
 Attachments: Screenshot.png


 I have created two users user 'unni' and user 'sachin'. user unni created a 
 table 'test3' so that user sachin cannot view that table. But user sachin is 
 able to grant all permission to the table test3.
 I have set 
 1)hive.security.authorization.enabled=true(in hive-site.xml)
 2)dfs.permissions=true(in hdfs-site.xml)
 3)dfs.permissions.supergroup=supergroup(in hdfs-site.xml)
 User sachin and user unni are in supergroup group.
 The user sachin is even able to revoke all permissions from the owner of the 
 table user unni.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




Jenkins build is back to normal : Hive-0.9.1-SNAPSHOT-h0.21 #80

2012-07-19 Thread Apache Jenkins Server
See https://builds.apache.org/job/Hive-0.9.1-SNAPSHOT-h0.21/80/



[jira] [Updated] (HIVE-1948) Have audit logging in the Metastore

2012-07-19 Thread Carl Steinbach (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-1948?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Carl Steinbach updated HIVE-1948:
-

Component/s: Security
 Logging

 Have audit logging in the Metastore
 ---

 Key: HIVE-1948
 URL: https://issues.apache.org/jira/browse/HIVE-1948
 Project: Hive
  Issue Type: Improvement
  Components: Logging, Metastore, Security
Reporter: Devaraj Das
Assignee: Devaraj Das
 Fix For: 0.7.0

 Attachments: audit-log-2.patch, audit-log-3.patch, audit-log.1.patch, 
 audit-log.patch


 It would be good to have audit logging in the metastore, similar to Hadoop's 
 NameNode audit logging. This would allow administrators to dig into details 
 about which user performed metadata operations (like create/drop 
 tables/partitions) and from where (IP address).

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Created] (HIVE-3277) Enable Metastore audit logging for non-secure connections

2012-07-19 Thread Carl Steinbach (JIRA)
Carl Steinbach created HIVE-3277:


 Summary: Enable Metastore audit logging for non-secure connections
 Key: HIVE-3277
 URL: https://issues.apache.org/jira/browse/HIVE-3277
 Project: Hive
  Issue Type: Improvement
  Components: Logging, Metastore, Security
Reporter: Carl Steinbach




--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HIVE-3277) Enable Metastore audit logging for non-secure connections

2012-07-19 Thread Carl Steinbach (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-3277?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13418664#comment-13418664
 ] 

Carl Steinbach commented on HIVE-3277:
--

Audit logging for the metastore was added in HIVE-1948, but it only gets 
invoked if the client connection is secure. We should also enable this for 
non-secure connections.

 Enable Metastore audit logging for non-secure connections
 -

 Key: HIVE-3277
 URL: https://issues.apache.org/jira/browse/HIVE-3277
 Project: Hive
  Issue Type: Improvement
  Components: Logging, Metastore, Security
Reporter: Carl Steinbach



--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Created] (HIVE-3278) Create wrappers around Thrift Metastore objects for improved readability

2012-07-19 Thread Carl Steinbach (JIRA)
Carl Steinbach created HIVE-3278:


 Summary: Create wrappers around Thrift Metastore objects for 
improved readability
 Key: HIVE-3278
 URL: https://issues.apache.org/jira/browse/HIVE-3278
 Project: Hive
  Issue Type: Improvement
  Components: Metastore
Reporter: Carl Steinbach
Assignee: Carl Steinbach




--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




Hive-trunk-h0.21 - Build # 1552 - Still Failing

2012-07-19 Thread Apache Jenkins Server
Changes for Build #1549

Changes for Build #1550
[namit] HIVE-3230 Make logging of plan progress in HadoopJobExecHelper 
configurable
(Kevin Wilfong via namit)


Changes for Build #1551

Changes for Build #1552



1 tests failed.
REGRESSION:  
org.apache.hadoop.hive.cli.TestNegativeMinimrCliDriver.testNegativeCliDriver_minimr_broken_pipe

Error Message:
Unexpected exception See build/ql/tmp/hive.log, or try ant test ... 
-Dtest.silent=false to get more logs.

Stack Trace:
junit.framework.AssertionFailedError: Unexpected exception
See build/ql/tmp/hive.log, or try ant test ... -Dtest.silent=false to get 
more logs.
at junit.framework.Assert.fail(Assert.java:50)
at 
org.apache.hadoop.hive.cli.TestNegativeMinimrCliDriver.testNegativeCliDriver_minimr_broken_pipe(TestNegativeMinimrCliDriver.java:212)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:597)
at junit.framework.TestCase.runTest(TestCase.java:168)
at junit.framework.TestCase.runBare(TestCase.java:134)
at junit.framework.TestResult$1.protect(TestResult.java:110)
at junit.framework.TestResult.runProtected(TestResult.java:128)
at junit.framework.TestResult.run(TestResult.java:113)
at junit.framework.TestCase.run(TestCase.java:124)
at junit.framework.TestSuite.runTest(TestSuite.java:243)
at junit.framework.TestSuite.run(TestSuite.java:238)
at 
org.apache.tools.ant.taskdefs.optional.junit.JUnitTestRunner.run(JUnitTestRunner.java:422)
at 
org.apache.tools.ant.taskdefs.optional.junit.JUnitTestRunner.launch(JUnitTestRunner.java:931)
at 
org.apache.tools.ant.taskdefs.optional.junit.JUnitTestRunner.main(JUnitTestRunner.java:785)




The Apache Jenkins build system has built Hive-trunk-h0.21 (build 
#$BUILD_NUMBER)

Status: Still Failing

Check console output at https://builds.apache.org/job/Hive-trunk-h0.21/1552/ to 
view the results.

[jira] [Created] (HIVE-3279) Table schema not being copied to Partitions with no columns

2012-07-19 Thread Travis Crawford (JIRA)
Travis Crawford created HIVE-3279:
-

 Summary: Table schema not being copied to Partitions with no 
columns
 Key: HIVE-3279
 URL: https://issues.apache.org/jira/browse/HIVE-3279
 Project: Hive
  Issue Type: Bug
Reporter: Travis Crawford
Assignee: Travis Crawford


Hive has a feature where {{Partition}}'s without any defined columns use the 
{{Table}} schema. This happens in 
{{[Partition.initialize|https://github.com/apache/hive/blob/trunk/ql/src/java/org/apache/hadoop/hive/ql/metadata/Partition.java#L167}}

{code}
// set default if columns are not set
if (tPartition.getSd().getCols() == null) {
  if (table.getCols() != null) {
tPartition.getSd().setCols(table.getCols());
  }
}
{code}

There's an issue though, because 
{{[Table.getEmptyTable|https://github.com/apache/hive/blob/trunk/ql/src/java/org/apache/hadoop/hive/ql/metadata/Table.java#L121]}}
 initializes cols to an empty array, which of course is not null, causing the 
above feature to not work as expected.

I'm not sure of the fix - is there a case where cols can indeed be null? I 
think the best thing to do here is:

{code}
-if (tPartition.getSd().getCols() == null) {
+if (tPartition.getSd().getCols() == null || 
tPartition.getSd().getCols().size() == 0) {
{code}

Thoughts?

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Updated] (HIVE-3279) Table schema not being copied to Partitions with no columns

2012-07-19 Thread Travis Crawford (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-3279?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Travis Crawford updated HIVE-3279:
--

Description: 
Hive has a feature where {{Partition}}'s without any defined columns use the 
{{Table}} schema. This happens in 
{{[Partition.initialize|https://github.com/apache/hive/blob/trunk/ql/src/java/org/apache/hadoop/hive/ql/metadata/Partition.java#L167]}}

{code}
// set default if columns are not set
if (tPartition.getSd().getCols() == null) {
  if (table.getCols() != null) {
tPartition.getSd().setCols(table.getCols());
  }
}
{code}

There's an issue though, because 
{{[Table.getEmptyTable|https://github.com/apache/hive/blob/trunk/ql/src/java/org/apache/hadoop/hive/ql/metadata/Table.java#L121]}}
 initializes cols to an empty array, which of course is not null, causing the 
above feature to not work as expected.

I'm not sure of the fix - is there a case where cols can indeed be null? I 
think the best thing to do here is:

{code}
-if (tPartition.getSd().getCols() == null) {
+if (tPartition.getSd().getCols() == null || 
tPartition.getSd().getCols().size() == 0) {
{code}

Thoughts?

  was:
Hive has a feature where {{Partition}}'s without any defined columns use the 
{{Table}} schema. This happens in 
{{[Partition.initialize|https://github.com/apache/hive/blob/trunk/ql/src/java/org/apache/hadoop/hive/ql/metadata/Partition.java#L167}}

{code}
// set default if columns are not set
if (tPartition.getSd().getCols() == null) {
  if (table.getCols() != null) {
tPartition.getSd().setCols(table.getCols());
  }
}
{code}

There's an issue though, because 
{{[Table.getEmptyTable|https://github.com/apache/hive/blob/trunk/ql/src/java/org/apache/hadoop/hive/ql/metadata/Table.java#L121]}}
 initializes cols to an empty array, which of course is not null, causing the 
above feature to not work as expected.

I'm not sure of the fix - is there a case where cols can indeed be null? I 
think the best thing to do here is:

{code}
-if (tPartition.getSd().getCols() == null) {
+if (tPartition.getSd().getCols() == null || 
tPartition.getSd().getCols().size() == 0) {
{code}

Thoughts?


 Table schema not being copied to Partitions with no columns
 ---

 Key: HIVE-3279
 URL: https://issues.apache.org/jira/browse/HIVE-3279
 Project: Hive
  Issue Type: Bug
Reporter: Travis Crawford
Assignee: Travis Crawford

 Hive has a feature where {{Partition}}'s without any defined columns use the 
 {{Table}} schema. This happens in 
 {{[Partition.initialize|https://github.com/apache/hive/blob/trunk/ql/src/java/org/apache/hadoop/hive/ql/metadata/Partition.java#L167]}}
 {code}
 // set default if columns are not set
 if (tPartition.getSd().getCols() == null) {
   if (table.getCols() != null) {
 tPartition.getSd().setCols(table.getCols());
   }
 }
 {code}
 There's an issue though, because 
 {{[Table.getEmptyTable|https://github.com/apache/hive/blob/trunk/ql/src/java/org/apache/hadoop/hive/ql/metadata/Table.java#L121]}}
  initializes cols to an empty array, which of course is not null, causing the 
 above feature to not work as expected.
 I'm not sure of the fix - is there a case where cols can indeed be null? I 
 think the best thing to do here is:
 {code}
 -if (tPartition.getSd().getCols() == null) {
 +if (tPartition.getSd().getCols() == null || 
 tPartition.getSd().getCols().size() == 0) {
 {code}
 Thoughts?

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Created] (HIVE-3280) Make HiveMetaStoreClient a public API

2012-07-19 Thread Carl Steinbach (JIRA)
Carl Steinbach created HIVE-3280:


 Summary: Make HiveMetaStoreClient a public API
 Key: HIVE-3280
 URL: https://issues.apache.org/jira/browse/HIVE-3280
 Project: Hive
  Issue Type: Improvement
  Components: Metastore
Reporter: Carl Steinbach




--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HIVE-3280) Make HiveMetaStoreClient a public API

2012-07-19 Thread Carl Steinbach (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-3280?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13418693#comment-13418693
 ] 

Carl Steinbach commented on HIVE-3280:
--

At the moment HiveMetaStoreClient is treated as a de-facto public API. We 
should make this official and provide better documentation.

 Make HiveMetaStoreClient a public API
 -

 Key: HIVE-3280
 URL: https://issues.apache.org/jira/browse/HIVE-3280
 Project: Hive
  Issue Type: Improvement
  Components: Metastore
Reporter: Carl Steinbach
  Labels: api-addition



--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




Re: Is HiveMetastoreClient a public interface?

2012-07-19 Thread Carl Steinbach
Sorry for the delay in responding.


 The java API is an ask from one of our users for Hcatalog. Given that
 HiveMetaStoreClient is a thin wrapper around public apis, then change
 shouldn't be drastic? It seems the cleaner way to go about it is to create
 an hcat client wrapper around HiveMetaStoreClient and have HCatClient as
 public?


The problem with this proposal is that until HiveMetaStoreClient is a
public API, HCatalog has just as little right to integrate with it directly
as any other third-party. I think the best solution is to invest the time
required to make HiveMetaStoreClient a public API, or to wrap
HiveMetaStoreClient with HiveMetaStoreClient2 and make that a public API
(we'll probably need to take the latter approach since HiveMetastoreClient
currently exposes raw Thrift types).

I filed HIVE-3280 to track this task.

Thanks.

Carl


[jira] [Updated] (HIVE-3126) Generate build the velocity based Hive tests on windows by fixing the path issues

2012-07-19 Thread Carl Steinbach (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-3126?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Carl Steinbach updated HIVE-3126:
-

Status: Open  (was: Patch Available)

@Kanna: I responded to your comments. Thanks.

 Generate  build the velocity based Hive tests on windows by fixing the path 
 issues
 ---

 Key: HIVE-3126
 URL: https://issues.apache.org/jira/browse/HIVE-3126
 Project: Hive
  Issue Type: Bug
  Components: Tests
Affects Versions: 0.9.0, 0.10.0, 0.9.1
Reporter: Kanna Karanam
Assignee: Kanna Karanam
  Labels: Windows, test
 Fix For: 0.10.0

 Attachments: HIVE-3126.1.patch.txt, HIVE-3126.2.patch.txt, 
 HIVE-3126.3.patch.txt, HIVE-3126.4.patch.txt, HIVE-3126.5.patch.txt, 
 HIVE-3126.6.patch.txt, HIVE-3126.7.patch.txt, HIVE-3126.8.patch.txt


 1)Escape the backward slash in Canonical Path if unit test runs on windows.
 2)Diff comparison – 
  a.   Ignore the extra spacing on windows
  b.   Ignore the different line endings on windows  Unix
  c.   Convert the file paths to windows specific. (Handle spaces 
 etc..)
 3)Set the right file scheme  class path separators while invoking the junit 
 task from 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HIVE-3278) Create wrappers around Thrift Metastore objects for improved readability

2012-07-19 Thread Travis Crawford (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-3278?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13418733#comment-13418733
 ] 

Travis Crawford commented on HIVE-3278:
---

Hey [~cwsteinbach] - can you clarify what you mean by wrappers? There are 
wrappers already (for example 
https://github.com/apache/hive/blob/trunk/ql/src/java/org/apache/hadoop/hive/ql/metadata/Table.java)
 and I'm curious what the distinction will be.

FWIW, having two sets of identically named classes (stuff under metastore.api  
ql.metadata) has already caused some confusion. Typically this confusion has 
been people using the metastore.api versions not realizing ql.metadata adds 
some business logic.

 Create wrappers around Thrift Metastore objects for improved readability
 

 Key: HIVE-3278
 URL: https://issues.apache.org/jira/browse/HIVE-3278
 Project: Hive
  Issue Type: Improvement
  Components: Metastore
Reporter: Carl Steinbach
Assignee: Carl Steinbach



--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Created] (HIVE-3281) Alter table ignores table/partition not existing in semantic analysis

2012-07-19 Thread Kevin Wilfong (JIRA)
Kevin Wilfong created HIVE-3281:
---

 Summary: Alter table ignores table/partition not existing in 
semantic analysis
 Key: HIVE-3281
 URL: https://issues.apache.org/jira/browse/HIVE-3281
 Project: Hive
  Issue Type: Bug
  Components: Logging, Query Processor
Affects Versions: 0.10.0
Reporter: Kevin Wilfong
Assignee: Kevin Wilfong
Priority: Minor


In the method addInputsOutputsAlterTable in the DDLSemanticAnalyzer we get the 
table and partition being altered from the metastore.  Failures to get these 
are quietly ignored, even though they resurface during query execution.  We can 
throw a semantic exception here instead of ignoring them and save some 
time/resources.  Also, by doing it here we save on retrieving the 
table/partition from the metastore since we're doing it here already.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Created] (HIVE-3282) Convert runtime exceptions to semantic exceptions for missing partitions/tables in show/describe statements

2012-07-19 Thread Sambavi Muthukrishnan (JIRA)
Sambavi Muthukrishnan created HIVE-3282:
---

 Summary: Convert runtime exceptions to semantic exceptions for 
missing partitions/tables in show/describe statements
 Key: HIVE-3282
 URL: https://issues.apache.org/jira/browse/HIVE-3282
 Project: Hive
  Issue Type: Bug
  Components: Query Processor
Affects Versions: 0.9.0
Reporter: Sambavi Muthukrishnan
Assignee: Sambavi Muthukrishnan
Priority: Minor
 Fix For: 0.9.1




--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Updated] (HIVE-3282) Convert runtime exceptions to semantic exceptions for missing partitions/tables in show/describe statements

2012-07-19 Thread Sambavi Muthukrishnan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-3282?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sambavi Muthukrishnan updated HIVE-3282:


   Description: 
The SHOW PARTITIONS command in Hive does not check for valid table and 
partition names during query compilation. Calling this command with 
non-existent table causes a run-time exception.

The DESC command also does not check for this in semantic analysis.

hive desc xxxyyy;
OK
Table xxxyyy does not exist
Time taken: 1.403 seconds

hive show partitions xxxyyy;
Table xxxyyy does not exist
FAILED: Execution Error, return code 1 from 
org.apache.hadoop.hive.ql.exec.DDLTask

Remaining Estimate: 24h
 Original Estimate: 24h

 Convert runtime exceptions to semantic exceptions for missing 
 partitions/tables in show/describe statements
 ---

 Key: HIVE-3282
 URL: https://issues.apache.org/jira/browse/HIVE-3282
 Project: Hive
  Issue Type: Bug
  Components: Query Processor
Affects Versions: 0.9.0
Reporter: Sambavi Muthukrishnan
Assignee: Sambavi Muthukrishnan
Priority: Minor
 Fix For: 0.9.1

   Original Estimate: 24h
  Remaining Estimate: 24h

 The SHOW PARTITIONS command in Hive does not check for valid table and 
 partition names during query compilation. Calling this command with 
 non-existent table causes a run-time exception.
 The DESC command also does not check for this in semantic analysis.
 hive desc xxxyyy;
 OK
 Table xxxyyy does not exist
 Time taken: 1.403 seconds
 hive show partitions xxxyyy;
 Table xxxyyy does not exist
 FAILED: Execution Error, return code 1 from 
 org.apache.hadoop.hive.ql.exec.DDLTask

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HIVE-3126) Generate build the velocity based Hive tests on windows by fixing the path issues

2012-07-19 Thread Kanna Karanam (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-3126?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13418816#comment-13418816
 ] 

Kanna Karanam commented on HIVE-3126:
-

@Carl – I added some follow up comments. can you please look into them?

 Generate  build the velocity based Hive tests on windows by fixing the path 
 issues
 ---

 Key: HIVE-3126
 URL: https://issues.apache.org/jira/browse/HIVE-3126
 Project: Hive
  Issue Type: Bug
  Components: Tests
Affects Versions: 0.9.0, 0.10.0, 0.9.1
Reporter: Kanna Karanam
Assignee: Kanna Karanam
  Labels: Windows, test
 Fix For: 0.10.0

 Attachments: HIVE-3126.1.patch.txt, HIVE-3126.2.patch.txt, 
 HIVE-3126.3.patch.txt, HIVE-3126.4.patch.txt, HIVE-3126.5.patch.txt, 
 HIVE-3126.6.patch.txt, HIVE-3126.7.patch.txt, HIVE-3126.8.patch.txt


 1)Escape the backward slash in Canonical Path if unit test runs on windows.
 2)Diff comparison – 
  a.   Ignore the extra spacing on windows
  b.   Ignore the different line endings on windows  Unix
  c.   Convert the file paths to windows specific. (Handle spaces 
 etc..)
 3)Set the right file scheme  class path separators while invoking the junit 
 task from 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HIVE-3246) java primitive type for binary datatype should be byte[]

2012-07-19 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-3246?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13418835#comment-13418835
 ] 

Hudson commented on HIVE-3246:
--

Integrated in Hive-trunk-h0.21 #1553 (See 
[https://builds.apache.org/job/Hive-trunk-h0.21/1553/])
HIVE-3246 : java primitive type for binary datatype should be byte[] 
(Thejas Nair via Ashutosh Chauhan) (Revision 1363427)

 Result = FAILURE
hashutosh : 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1363427
Files : 
* /hive/trunk/serde/src/java/org/apache/hadoop/hive/serde2/lazy/LazyUtils.java
* 
/hive/trunk/serde/src/java/org/apache/hadoop/hive/serde2/lazy/objectinspector/primitive/LazyBinaryObjectInspector.java
* 
/hive/trunk/serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/primitive/BinaryObjectInspector.java
* 
/hive/trunk/serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/primitive/JavaBinaryObjectInspector.java
* 
/hive/trunk/serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/primitive/PrimitiveObjectInspectorConverter.java
* 
/hive/trunk/serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/primitive/PrimitiveObjectInspectorUtils.java
* 
/hive/trunk/serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/primitive/SettableBinaryObjectInspector.java
* 
/hive/trunk/serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/primitive/WritableBinaryObjectInspector.java
* 
/hive/trunk/serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/primitive/WritableConstantBinaryObjectInspector.java
* /hive/trunk/serde/src/test/org/apache/hadoop/hive/serde2/TestStatsSerde.java
* 
/hive/trunk/serde/src/test/org/apache/hadoop/hive/serde2/binarysortable/MyTestClass.java
* 
/hive/trunk/serde/src/test/org/apache/hadoop/hive/serde2/binarysortable/TestBinarySortableSerDe.java
* 
/hive/trunk/serde/src/test/org/apache/hadoop/hive/serde2/columnar/TestLazyBinaryColumnarSerDe.java
* 
/hive/trunk/serde/src/test/org/apache/hadoop/hive/serde2/lazybinary/MyTestClassBigger.java
* 
/hive/trunk/serde/src/test/org/apache/hadoop/hive/serde2/lazybinary/TestLazyBinarySerDe.java
* 
/hive/trunk/serde/src/test/org/apache/hadoop/hive/serde2/objectinspector/TestStandardObjectInspectors.java


 java primitive type for binary datatype should be byte[]
 

 Key: HIVE-3246
 URL: https://issues.apache.org/jira/browse/HIVE-3246
 Project: Hive
  Issue Type: Bug
Affects Versions: 0.9.0
Reporter: Thejas M Nair
Assignee: Thejas M Nair
 Fix For: 0.10.0

 Attachments: HIVE-3246.1.patch, HIVE-3246.2.patch


 PrimitiveObjectInspector.getPrimitiveJavaObject is supposed to return a java 
 object. But in case of binary datatype, it returns ByteArrayRef (not java 
 standard type). The suitable java object for it would be byte[]. 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HIVE-3251) Hive doesn't remove scrach directories while killing running MR job

2012-07-19 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-3251?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13418836#comment-13418836
 ] 

Hudson commented on HIVE-3251:
--

Integrated in Hive-trunk-h0.21 #1553 (See 
[https://builds.apache.org/job/Hive-trunk-h0.21/1553/])
HIVE-3251 Hive doesn't remove scrach directories while killing running MR 
job
(Gang Tim Liu via namit) (Revision 1363424)

 Result = FAILURE
namit : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1363424
Files : 
* /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/Context.java
* /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/Driver.java
* /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/Utilities.java


 Hive doesn't remove scrach directories while killing running MR job
 ---

 Key: HIVE-3251
 URL: https://issues.apache.org/jira/browse/HIVE-3251
 Project: Hive
  Issue Type: Bug
  Components: Server Infrastructure
Reporter: Gang Tim Liu
Assignee: Gang Tim Liu
 Attachments: HIVE-3151.patch


 While killing running MR job, hive doesn't clean up scratch directory 
 (mapred.cache.files). So that, afterwards, scratch directory is left there in 
 hdfs. HDFS name node doesn't know it and try to do lease recovery. while such 
 instances happen more, it will eventually crash namenode.
 The fix is to leverage hdfs clean up functionality. While creating scratch 
 dirs, hive registers it to hdfs cleanup hook. While killing happens, hdfs 
 will clean them up.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Updated] (HIVE-3267) escaped columns in cluster/distribute/order/sort by are not working

2012-07-19 Thread Kevin Wilfong (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-3267?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kevin Wilfong updated HIVE-3267:


  Component/s: Query Processor
Affects Version/s: 0.10.0
Fix Version/s: 0.10.0

 escaped columns in cluster/distribute/order/sort by are not working
 ---

 Key: HIVE-3267
 URL: https://issues.apache.org/jira/browse/HIVE-3267
 Project: Hive
  Issue Type: Bug
  Components: Query Processor
Affects Versions: 0.10.0
Reporter: Namit Jain
Assignee: Namit Jain
 Fix For: 0.10.0


 The following query:
 select `key`, value from src cluster by `key`, value;
 fails

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Updated] (HIVE-3267) escaped columns in cluster/distribute/order/sort by are not working

2012-07-19 Thread Kevin Wilfong (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-3267?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kevin Wilfong updated HIVE-3267:


Resolution: Fixed
Status: Resolved  (was: Patch Available)

Committed, thanks Namit.

 escaped columns in cluster/distribute/order/sort by are not working
 ---

 Key: HIVE-3267
 URL: https://issues.apache.org/jira/browse/HIVE-3267
 Project: Hive
  Issue Type: Bug
  Components: Query Processor
Affects Versions: 0.10.0
Reporter: Namit Jain
Assignee: Namit Jain
 Fix For: 0.10.0


 The following query:
 select `key`, value from src cluster by `key`, value;
 fails

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Updated] (HIVE-3205) Bucketed mapjoin on partitioned table which has no partition throws NPE

2012-07-19 Thread Namit Jain (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-3205?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Namit Jain updated HIVE-3205:
-

  Resolution: Fixed
Hadoop Flags: Reviewed
  Status: Resolved  (was: Patch Available)

Committed. Thanks Navis

 Bucketed mapjoin on partitioned table which has no partition throws NPE
 ---

 Key: HIVE-3205
 URL: https://issues.apache.org/jira/browse/HIVE-3205
 Project: Hive
  Issue Type: Bug
  Components: Query Processor
Affects Versions: 0.10.0
 Environment: ubuntu 10.04
Reporter: Navis
Assignee: Navis
Priority: Minor

 {code}
 create table hive_test_smb_bucket1 (key int, value string) partitioned by (ds 
 string) clustered by (key) sorted by (key) into 2 buckets;
 create table hive_test_smb_bucket2 (key int, value string) partitioned by (ds 
 string) clustered by (key) sorted by (key) into 2 buckets;
 set hive.optimize.bucketmapjoin = true;
 set hive.input.format = 
 org.apache.hadoop.hive.ql.io.BucketizedHiveInputFormat;
 explain
 SELECT /* + MAPJOIN(b) */ b.key as k1, b.value, b.ds, a.key as k2
 FROM hive_test_smb_bucket1 a JOIN
 hive_test_smb_bucket2 b
 ON a.key = b.key WHERE a.ds = '2010-10-15' and b.ds='2010-10-15' and  b.key 
 IS NOT NULL;
 {code}
 throws NPE
 {noformat}
 2012-06-28 08:59:13,459 ERROR ql.Driver (SessionState.java:printError(400)) - 
 FAILED: NullPointerException null
 java.lang.NullPointerException
   at 
 org.apache.hadoop.hive.ql.optimizer.BucketMapJoinOptimizer$BucketMapjoinOptProc.process(BucketMapJoinOptimizer.java:269)
   at 
 org.apache.hadoop.hive.ql.lib.DefaultRuleDispatcher.dispatch(DefaultRuleDispatcher.java:89)
   at 
 org.apache.hadoop.hive.ql.lib.DefaultGraphWalker.dispatch(DefaultGraphWalker.java:88)
   at 
 org.apache.hadoop.hive.ql.lib.DefaultGraphWalker.walk(DefaultGraphWalker.java:125)
   at 
 org.apache.hadoop.hive.ql.lib.DefaultGraphWalker.startWalking(DefaultGraphWalker.java:102)
   at 
 org.apache.hadoop.hive.ql.optimizer.BucketMapJoinOptimizer.transform(BucketMapJoinOptimizer.java:100)
   at 
 org.apache.hadoop.hive.ql.optimizer.Optimizer.optimize(Optimizer.java:87)
   at 
 org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.analyzeInternal(SemanticAnalyzer.java:7564)
   at 
 org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer.analyze(BaseSemanticAnalyzer.java:245)
   at 
 org.apache.hadoop.hive.ql.parse.ExplainSemanticAnalyzer.analyzeInternal(ExplainSemanticAnalyzer.java:50)
   at 
 org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer.analyze(BaseSemanticAnalyzer.java:245)
   at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:430)
   at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:335)
   at org.apache.hadoop.hive.ql.Driver.run(Driver.java:902)
   at 
 org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:258)
   at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:215)
   at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:406)
   at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:744)
   at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:607)
   at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
   at 
 sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
   at 
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
   at java.lang.reflect.Method.invoke(Method.java:597)
   at org.apache.hadoop.util.RunJar.main(RunJar.java:186)
 {noformat}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Created] (HIVE-3283) bucket information should be used from the partition instead of the table

2012-07-19 Thread Namit Jain (JIRA)
Namit Jain created HIVE-3283:


 Summary: bucket information should be used from the partition 
instead of the table
 Key: HIVE-3283
 URL: https://issues.apache.org/jira/browse/HIVE-3283
 Project: Hive
  Issue Type: Bug
Reporter: Namit Jain


Current Hive uses the number of buckets from the table object.
Ideally, the number of buckets from the partition should be used

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Updated] (HIVE-3283) bucket information should be used from the partition instead of the table

2012-07-19 Thread Namit Jain (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-3283?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Namit Jain updated HIVE-3283:
-

Description: 
Currently Hive uses the number of buckets from the table object.
Ideally, the number of buckets from the partition should be used

  was:
Current Hive uses the number of buckets from the table object.
Ideally, the number of buckets from the partition should be used


 bucket information should be used from the partition instead of the table
 -

 Key: HIVE-3283
 URL: https://issues.apache.org/jira/browse/HIVE-3283
 Project: Hive
  Issue Type: Bug
Reporter: Namit Jain

 Currently Hive uses the number of buckets from the table object.
 Ideally, the number of buckets from the partition should be used

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HIVE-3261) alter the number of buckets for a non-empty partitioned table should not be allowed

2012-07-19 Thread Namit Jain (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-3261?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13418914#comment-13418914
 ] 

Namit Jain commented on HIVE-3261:
--

comments addressed

 alter the number of buckets for a non-empty partitioned table should not be 
 allowed
 ---

 Key: HIVE-3261
 URL: https://issues.apache.org/jira/browse/HIVE-3261
 Project: Hive
  Issue Type: Bug
  Components: Query Processor
Reporter: Namit Jain
Assignee: Namit Jain
  Labels: bucketing

 This is dangerous since the code uses the table metadata everywhere to get 
 the number of buckets

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Updated] (HIVE-3261) alter the number of buckets for a non-empty partitioned table should not be allowed

2012-07-19 Thread Namit Jain (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-3261?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Namit Jain updated HIVE-3261:
-

Status: Patch Available  (was: Open)

 alter the number of buckets for a non-empty partitioned table should not be 
 allowed
 ---

 Key: HIVE-3261
 URL: https://issues.apache.org/jira/browse/HIVE-3261
 Project: Hive
  Issue Type: Bug
  Components: Query Processor
Reporter: Namit Jain
Assignee: Namit Jain
  Labels: bucketing

 This is dangerous since the code uses the table metadata everywhere to get 
 the number of buckets

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Created] (HIVE-3284) create ExecutionException(s)

2012-07-19 Thread Namit Jain (JIRA)
Namit Jain created HIVE-3284:


 Summary: create ExecutionException(s)
 Key: HIVE-3284
 URL: https://issues.apache.org/jira/browse/HIVE-3284
 Project: Hive
  Issue Type: Bug
Reporter: Namit Jain


Hive currently contains SemanticExceptions for all compile time exceptions.
It might be a good idea to create ExecutionExceptions to get the list
of all execution exceptions.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Updated] (HIVE-1653) Ability to enforce correct stats

2012-07-19 Thread Namit Jain (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-1653?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Namit Jain updated HIVE-1653:
-

Status: Patch Available  (was: Open)

comments

 Ability to enforce correct stats
 

 Key: HIVE-1653
 URL: https://issues.apache.org/jira/browse/HIVE-1653
 Project: Hive
  Issue Type: Improvement
  Components: Query Processor
Reporter: Namit Jain
Assignee: Namit Jain

 This is a follow-up for https://issues.apache.org/jira/browse/HIVE-1361.
 If one of the mappers/reducers cannot publish stats, it may lead to wrong 
 aggregated stats.
 There should be a way to avoid this - at the least, a configuration variable 
 which fails the 
 task if stats cannot be published

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira