[jira] [Commented] (HIVE-9453) Initial patch [hbase-metastore branch]

2015-07-17 Thread Lefty Leverenz (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9453?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14632301#comment-14632301
 ] 

Lefty Leverenz commented on HIVE-9453:
--

Since it isn't the 1.2 release after all, I'm linking this issue to HIVE-9752 
(Documentation for HBase metastore).

Also, the new tool to dump out the contents of the metastore should be 
documented (Alan's comment, 30/Jan/15).

> Initial patch [hbase-metastore branch]
> --
>
> Key: HIVE-9453
> URL: https://issues.apache.org/jira/browse/HIVE-9453
> Project: Hive
>  Issue Type: Sub-task
>  Components: Metastore
>Affects Versions: hbase-metastore-branch
>Reporter: Alan Gates
>Assignee: Alan Gates
> Fix For: hbase-metastore-branch
>
> Attachments: HIVE-9453-reviewcomments.pdf, HIVE-9453.2.patch, 
> HIVE-9453.3.patch, HIVE-9453.patch
>
>
> This initial patch has several important features:
> # HBaseStore, a new implementation of RawStore that stores the data in HBase.
> # Subclasses of the thrift metastore objects to remove the massive 
> duplication of data where every partition contains a nearly identical storage 
> descriptor.
> # Caches for catalog objects and statistics so that repeated metastore calls 
> don't result in repeated calls against HBase.
> Currently this "works" to the point that "load table" and "select" work.  I 
> have not tested any other statements, and I suspect most fail.  There is no 
> security, no authorization, and a no a lot of other things.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-11306) Add a bloom-1 filter for Hybrid MapJoin spills

2015-07-17 Thread Gopal V (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-11306?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gopal V updated HIVE-11306:
---
Attachment: HIVE-11306.1.patch

> Add a bloom-1 filter for Hybrid MapJoin spills
> --
>
> Key: HIVE-11306
> URL: https://issues.apache.org/jira/browse/HIVE-11306
> Project: Hive
>  Issue Type: Improvement
>  Components: Hive
>Affects Versions: 1.3.0, 2.0.0
>Reporter: Gopal V
>Assignee: Gopal V
> Attachments: HIVE-11306.1.patch
>
>
> HIVE-9277 implemented Spillable joins for Tez, which suffers from a 
> corner-case performance issue when joining wide small tables against a narrow 
> big table (like a user info table join events stream).
> The fact that the wide table is spilled causes extra IO, even though the nDV 
> of the join key might be in the thousands.
> A cheap bloom-1 filter would add a massive performance gain for such queries, 
> massively cutting down on the spill IO costs for the big-table spills.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9152) Dynamic Partition Pruning [Spark Branch]

2015-07-17 Thread Lefty Leverenz (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9152?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14632293#comment-14632293
 ] 

Lefty Leverenz commented on HIVE-9152:
--

Doc note:  This adds two configuration parameters 
(*hive.spark.dynamic.partition.pruning* and 
*hive.spark.dynamic.partition.pruning.max.data.size*) to HiveConf.java, so they 
should be documented in Hive on Spark: Getting Started and in the Spark section 
of Configuration Properties.

* doc now or only after merge branch with master? -- [Hive on Spark: Getting 
Started -- Configuring Hive | 
https://cwiki.apache.org/confluence/display/Hive/Hive+on+Spark%3A+Getting+Started#HiveonSpark:GettingStarted-ConfiguringHive]
* wait until merge with master -- [Configuration Properties -- Spark | 
https://cwiki.apache.org/confluence/display/Hive/Configuration+Properties#ConfigurationProperties-Spark]

Hive on Spark: Getting Started might also need general information about 
dynamic partition pruning.

> Dynamic Partition Pruning [Spark Branch]
> 
>
> Key: HIVE-9152
> URL: https://issues.apache.org/jira/browse/HIVE-9152
> Project: Hive
>  Issue Type: Sub-task
>  Components: Spark
>Affects Versions: spark-branch
>Reporter: Brock Noland
>Assignee: Chao Sun
>  Labels: TODOC-SPARK
> Fix For: spark-branch
>
> Attachments: HIVE-9152.1-spark.patch, HIVE-9152.10-spark.patch, 
> HIVE-9152.11-spark.patch, HIVE-9152.12-spark.patch, HIVE-9152.2-spark.patch, 
> HIVE-9152.3-spark.patch, HIVE-9152.4-spark.patch, HIVE-9152.5-spark.patch, 
> HIVE-9152.6-spark.patch, HIVE-9152.8-spark.patch, HIVE-9152.9-spark.patch
>
>
> Tez implemented dynamic partition pruning in HIVE-7826. This is a nice 
> optimization and we should implement the same in HOS.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-11305) LLAP: Hybrid Map-join cache returns invalid data

2015-07-17 Thread Gopal V (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-11305?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14632290#comment-14632290
 ] 

Gopal V commented on HIVE-11305:


Running the same tests with {{./dist/hive/bin/hive --hiveconf 
hive.execution.mode=container -f q55-test.sql}} works.

> LLAP: Hybrid Map-join cache returns invalid data 
> -
>
> Key: HIVE-11305
> URL: https://issues.apache.org/jira/browse/HIVE-11305
> Project: Hive
>  Issue Type: Sub-task
>Affects Versions: llap
> Environment: TPC-DS 200 scale data
>Reporter: Gopal V
>Assignee: Sergey Shelukhin
>Priority: Critical
> Fix For: llap
>
> Attachments: q55-test.sql
>
>
> Start a 1-node LLAP cluster with 16 executors and run attached test-case on 
> the single node instance.
> {code}
> Caused by: java.lang.ClassCastException: 
> org.apache.hadoop.hive.ql.exec.persistence.HybridHashTableContainer cannot be 
> cast to 
> org.apache.hadoop.hive.ql.exec.vector.mapjoin.hashtable.VectorMapJoinTableContainer
> at 
> org.apache.hadoop.hive.ql.exec.vector.mapjoin.VectorMapJoinCommonOperator.loadHashTable(VectorMapJoinCommonOperator.java:648)
> at 
> org.apache.hadoop.hive.ql.exec.MapJoinOperator.cleanUpInputFileChangedOp(MapJoinOperator.java:314)
> at 
> org.apache.hadoop.hive.ql.exec.Operator.cleanUpInputFileChanged(Operator.java:1104)
> at 
> org.apache.hadoop.hive.ql.exec.Operator.cleanUpInputFileChanged(Operator.java:1108)
> at 
> org.apache.hadoop.hive.ql.exec.Operator.cleanUpInputFileChanged(Operator.java:1108)
> at 
> org.apache.hadoop.hive.ql.exec.Operator.cleanUpInputFileChanged(Operator.java:1108)
> at 
> org.apache.hadoop.hive.ql.exec.Operator.cleanUpInputFileChanged(Operator.java:1108)
> at 
> org.apache.hadoop.hive.ql.exec.vector.VectorMapOperator.process(VectorMapOperator.java:37)
> at 
> org.apache.hadoop.hive.ql.exec.tez.MapRecordSource.processRow(MapRecordSource.java:86)
> ... 17 more
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-11305) LLAP: Hybrid Map-join cache returns invalid data

2015-07-17 Thread Gopal V (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-11305?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gopal V updated HIVE-11305:
---
Attachment: q55-test.sql

> LLAP: Hybrid Map-join cache returns invalid data 
> -
>
> Key: HIVE-11305
> URL: https://issues.apache.org/jira/browse/HIVE-11305
> Project: Hive
>  Issue Type: Sub-task
>Affects Versions: llap
> Environment: TPC-DS 200 scale data
>Reporter: Gopal V
>Assignee: Sergey Shelukhin
>Priority: Critical
> Fix For: llap
>
> Attachments: q55-test.sql
>
>
> Start a 1-node LLAP cluster with 16 executors and run attached test-case on 
> the single node instance.
> {code}
> Caused by: java.lang.ClassCastException: 
> org.apache.hadoop.hive.ql.exec.persistence.HybridHashTableContainer cannot be 
> cast to 
> org.apache.hadoop.hive.ql.exec.vector.mapjoin.hashtable.VectorMapJoinTableContainer
> at 
> org.apache.hadoop.hive.ql.exec.vector.mapjoin.VectorMapJoinCommonOperator.loadHashTable(VectorMapJoinCommonOperator.java:648)
> at 
> org.apache.hadoop.hive.ql.exec.MapJoinOperator.cleanUpInputFileChangedOp(MapJoinOperator.java:314)
> at 
> org.apache.hadoop.hive.ql.exec.Operator.cleanUpInputFileChanged(Operator.java:1104)
> at 
> org.apache.hadoop.hive.ql.exec.Operator.cleanUpInputFileChanged(Operator.java:1108)
> at 
> org.apache.hadoop.hive.ql.exec.Operator.cleanUpInputFileChanged(Operator.java:1108)
> at 
> org.apache.hadoop.hive.ql.exec.Operator.cleanUpInputFileChanged(Operator.java:1108)
> at 
> org.apache.hadoop.hive.ql.exec.Operator.cleanUpInputFileChanged(Operator.java:1108)
> at 
> org.apache.hadoop.hive.ql.exec.vector.VectorMapOperator.process(VectorMapOperator.java:37)
> at 
> org.apache.hadoop.hive.ql.exec.tez.MapRecordSource.processRow(MapRecordSource.java:86)
> ... 17 more
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9152) Dynamic Partition Pruning [Spark Branch]

2015-07-17 Thread Lefty Leverenz (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9152?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Lefty Leverenz updated HIVE-9152:
-
Labels: TODOC-SPARK  (was: )

> Dynamic Partition Pruning [Spark Branch]
> 
>
> Key: HIVE-9152
> URL: https://issues.apache.org/jira/browse/HIVE-9152
> Project: Hive
>  Issue Type: Sub-task
>  Components: Spark
>Affects Versions: spark-branch
>Reporter: Brock Noland
>Assignee: Chao Sun
>  Labels: TODOC-SPARK
> Fix For: spark-branch
>
> Attachments: HIVE-9152.1-spark.patch, HIVE-9152.10-spark.patch, 
> HIVE-9152.11-spark.patch, HIVE-9152.12-spark.patch, HIVE-9152.2-spark.patch, 
> HIVE-9152.3-spark.patch, HIVE-9152.4-spark.patch, HIVE-9152.5-spark.patch, 
> HIVE-9152.6-spark.patch, HIVE-9152.8-spark.patch, HIVE-9152.9-spark.patch
>
>
> Tez implemented dynamic partition pruning in HIVE-7826. This is a nice 
> optimization and we should implement the same in HOS.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-11077) Add support in parser and wire up to txn manager

2015-07-17 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-11077?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14632284#comment-14632284
 ] 

Hive QA commented on HIVE-11077:




{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12745919/HIVE-11077.6.patch

{color:red}ERROR:{color} -1 due to 1 failed/errored test(s), 9242 tests executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_cbo_rp_join0
{noformat}

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/4647/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/4647/console
Test logs: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-4647/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 1 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12745919 - PreCommit-HIVE-TRUNK-Build

> Add support in parser and wire up to txn manager
> 
>
> Key: HIVE-11077
> URL: https://issues.apache.org/jira/browse/HIVE-11077
> Project: Hive
>  Issue Type: Sub-task
>  Components: SQL, Transactions
>Affects Versions: 1.2.0
>Reporter: Eugene Koifman
>Assignee: Eugene Koifman
> Attachments: HIVE-11077.3.patch, HIVE-11077.5.patch, 
> HIVE-11077.6.patch, HIVE-11077.patch
>
>




--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Resolved] (HIVE-10930) LLAP: Set java.io.tmpdir correctly for LLAP Slider instance

2015-07-17 Thread Gopal V (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-10930?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gopal V resolved HIVE-10930.

  Resolution: Fixed
Release Note: HIVE-10930: LLAP: Set java.io.tmpdir correctly for LLAP 
Slider instance (Gopal V)

> LLAP: Set java.io.tmpdir correctly for LLAP Slider instance
> ---
>
> Key: HIVE-10930
> URL: https://issues.apache.org/jira/browse/HIVE-10930
> Project: Hive
>  Issue Type: Sub-task
>Affects Versions: llap
>Reporter: Gopal V
>Assignee: Gopal V
> Fix For: llap
>
> Attachments: HIVE-10930.patch
>
>
> LLAP's Hybrid Grace Hash is IO bound writing to /tmp.
> Use the yarn local dirs instead of "/tmp" so that createTempFile works 
> correctly.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-10930) LLAP: Set java.io.tmpdir correctly for LLAP Slider instance

2015-07-17 Thread Gopal V (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-10930?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gopal V updated HIVE-10930:
---
Attachment: HIVE-10930.patch

> LLAP: Set java.io.tmpdir correctly for LLAP Slider instance
> ---
>
> Key: HIVE-10930
> URL: https://issues.apache.org/jira/browse/HIVE-10930
> Project: Hive
>  Issue Type: Sub-task
>Affects Versions: llap
>Reporter: Gopal V
>Assignee: Gopal V
> Fix For: llap
>
> Attachments: HIVE-10930.patch
>
>
> LLAP's Hybrid Grace Hash is IO bound writing to /tmp.
> Use the yarn local dirs instead of "/tmp" so that createTempFile works 
> correctly.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-11303) Getting Tez LimitExceededException after dag execution on large query

2015-07-17 Thread Gopal V (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-11303?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14632255#comment-14632255
 ] 

Gopal V commented on HIVE-11303:


is this triggered by the {{hive.tez.exec.print.summary}} ?

> Getting Tez LimitExceededException after dag execution on large query
> -
>
> Key: HIVE-11303
> URL: https://issues.apache.org/jira/browse/HIVE-11303
> Project: Hive
>  Issue Type: Bug
>  Components: Tez
>Reporter: Jason Dere
>
> {noformat}
> 2015-07-17 18:18:11,830 INFO  [main]: counters.Limits 
> (Limits.java:ensureInitialized(59)) - Counter limits initialized with 
> parameters:  GROUP_NAME_MAX=256, MAX_GROUPS=500, COUNTER_NAME_MAX=64, 
> MAX_COUNTERS=1200
> 2015-07-17 18:18:11,841 ERROR [main]: exec.Task (TezTask.java:execute(189)) - 
> Failed to execute tez graph.
> org.apache.tez.common.counters.LimitExceededException: Too many counters: 
> 1201 max=1200
> at org.apache.tez.common.counters.Limits.checkCounters(Limits.java:87)
> at org.apache.tez.common.counters.Limits.incrCounters(Limits.java:94)
> at 
> org.apache.tez.common.counters.AbstractCounterGroup.addCounter(AbstractCounterGroup.java:76)
> at 
> org.apache.tez.common.counters.AbstractCounterGroup.addCounterImpl(AbstractCounterGroup.java:93)
> at 
> org.apache.tez.common.counters.AbstractCounterGroup.findCounter(AbstractCounterGroup.java:104)
> at 
> org.apache.tez.dag.api.DagTypeConverters.convertTezCountersFromProto(DagTypeConverters.java:567)
> at 
> org.apache.tez.dag.api.client.DAGStatus.getDAGCounters(DAGStatus.java:148)
> at 
> org.apache.hadoop.hive.ql.exec.tez.TezTask.execute(TezTask.java:175)
> at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:160)
> at 
> org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:89)
> at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:1673)
> at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:1432)
> at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1213)
> at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1064)
> at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1054)
> at 
> org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:213)
> at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:165)
> at 
> org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:376)
> at 
> org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:311)
> at 
> org.apache.hadoop.hive.cli.CliDriver.processReader(CliDriver.java:409)
> at 
> org.apache.hadoop.hive.cli.CliDriver.processFile(CliDriver.java:425)
> at 
> org.apache.hadoop.hive.cli.CliDriver.executeDriver(CliDriver.java:714)
> at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:681)
> at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:621)
> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
> at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> at java.lang.reflect.Method.invoke(Method.java:497)
> at org.apache.hadoop.util.RunJar.run(RunJar.java:221)
> at org.apache.hadoop.util.RunJar.main(RunJar.java:136)
> {noformat}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-10799) Refactor the SearchArgumentFactory to remove the dependence on ExprNodeGenericFuncDesc

2015-07-17 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-10799?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14632252#comment-14632252
 ] 

Hive QA commented on HIVE-10799:




{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12745912/HIVE-10799.patch

{color:red}ERROR:{color} -1 due to 1 failed/errored test(s), 9239 tests executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_cbo_rp_join0
{noformat}

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/4646/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/4646/console
Test logs: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-4646/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 1 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12745912 - PreCommit-HIVE-TRUNK-Build

> Refactor the SearchArgumentFactory to remove the dependence on 
> ExprNodeGenericFuncDesc
> --
>
> Key: HIVE-10799
> URL: https://issues.apache.org/jira/browse/HIVE-10799
> Project: Hive
>  Issue Type: Sub-task
>Reporter: Owen O'Malley
>Assignee: Owen O'Malley
> Attachments: HIVE-10799.patch, HIVE-10799.patch, HIVE-10799.patch, 
> HIVE-10799.patch
>
>
> SearchArgumentFactory and SearchArgumentImpl are high level and shouldn't 
> depend on the internals of Hive's AST model.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-11210) Remove dependency on HiveConf from Orc reader & writer

2015-07-17 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-11210?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=1463#comment-1463
 ] 

Hive QA commented on HIVE-11210:




{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12745907/HIVE-11210.patch

{color:red}ERROR:{color} -1 due to 2 failed/errored test(s), 9225 tests executed
*Failed tests:*
{noformat}
TestCustomAuthentication - did not produce a TEST-*.xml file
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_cbo_rp_join0
{noformat}

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/4645/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/4645/console
Test logs: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-4645/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 2 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12745907 - PreCommit-HIVE-TRUNK-Build

> Remove dependency on HiveConf from Orc reader & writer
> --
>
> Key: HIVE-11210
> URL: https://issues.apache.org/jira/browse/HIVE-11210
> Project: Hive
>  Issue Type: Sub-task
>Reporter: Owen O'Malley
>Assignee: Owen O'Malley
> Attachments: HIVE-11210.patch, HIVE-11210.patch
>
>
> Currently the ORC reader and writer get their default values from HiveConf. I 
> propose that we make the reader and writer have their own programatic 
> defaults and the OrcInputFormat and OrcOutputFormat can use the version in 
> HiveConf.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-11300) HBase metastore: Support token and master key methods

2015-07-17 Thread Alan Gates (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-11300?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alan Gates updated HIVE-11300:
--
Attachment: HIVE-11300.patch

> HBase metastore: Support token and master key methods
> -
>
> Key: HIVE-11300
> URL: https://issues.apache.org/jira/browse/HIVE-11300
> Project: Hive
>  Issue Type: Improvement
>  Components: Metastore
>Affects Versions: hbase-metastore-branch
>Reporter: Alan Gates
>Assignee: Alan Gates
> Attachments: HIVE-11300.patch
>
>
> The methods addToken, removeToken, getToken, getAllTokenIdentifiers, 
> addMasterKey, updateMasterKey, removeMasterKey, and getMasterKeys() need to 
> be implemented.  They are all used in security.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-10937) LLAP: make ObjectCache for plans work properly in the daemon

2015-07-17 Thread Sergey Shelukhin (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-10937?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sergey Shelukhin updated HIVE-10937:

Attachment: HIVE-10937.03.patch

Rebased the patch and extended cleanup code for it

> LLAP: make ObjectCache for plans work properly in the daemon
> 
>
> Key: HIVE-10937
> URL: https://issues.apache.org/jira/browse/HIVE-10937
> Project: Hive
>  Issue Type: Sub-task
>Reporter: Sergey Shelukhin
>Assignee: Sergey Shelukhin
> Fix For: llap
>
> Attachments: HIVE-10937.01.patch, HIVE-10937.02.patch, 
> HIVE-10937.03.patch, HIVE-10937.patch
>
>
> There's perf hit otherwise, esp. when stupid planner creates 1009 reducers of 
> 4Mb each.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-11303) Getting Tez LimitExceededException after dag execution on large query

2015-07-17 Thread Jason Dere (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-11303?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14632212#comment-14632212
 ] 

Jason Dere commented on HIVE-11303:
---

Thing is, I've actually set tez.counters.max=16000 on the Hive command line. I 
can also see in the Tez logs, that the counters are set to 16000 during the dag 
execution.

The log messages in the Jira description (including the line where 
MAX_COUNTERS=1200) is from my hive.log.  So it seems like when we're trying to 
fetch the counters after execution, that tez.counters.max is somehow not set in 
the DagClient's configuration.  I suspect the TezSessionState's conf is coming 
from the SessionState's configuration, which might be getting initialized with 
a blank config from SessionState.getSessionConf().

> Getting Tez LimitExceededException after dag execution on large query
> -
>
> Key: HIVE-11303
> URL: https://issues.apache.org/jira/browse/HIVE-11303
> Project: Hive
>  Issue Type: Bug
>  Components: Tez
>Reporter: Jason Dere
>
> {noformat}
> 2015-07-17 18:18:11,830 INFO  [main]: counters.Limits 
> (Limits.java:ensureInitialized(59)) - Counter limits initialized with 
> parameters:  GROUP_NAME_MAX=256, MAX_GROUPS=500, COUNTER_NAME_MAX=64, 
> MAX_COUNTERS=1200
> 2015-07-17 18:18:11,841 ERROR [main]: exec.Task (TezTask.java:execute(189)) - 
> Failed to execute tez graph.
> org.apache.tez.common.counters.LimitExceededException: Too many counters: 
> 1201 max=1200
> at org.apache.tez.common.counters.Limits.checkCounters(Limits.java:87)
> at org.apache.tez.common.counters.Limits.incrCounters(Limits.java:94)
> at 
> org.apache.tez.common.counters.AbstractCounterGroup.addCounter(AbstractCounterGroup.java:76)
> at 
> org.apache.tez.common.counters.AbstractCounterGroup.addCounterImpl(AbstractCounterGroup.java:93)
> at 
> org.apache.tez.common.counters.AbstractCounterGroup.findCounter(AbstractCounterGroup.java:104)
> at 
> org.apache.tez.dag.api.DagTypeConverters.convertTezCountersFromProto(DagTypeConverters.java:567)
> at 
> org.apache.tez.dag.api.client.DAGStatus.getDAGCounters(DAGStatus.java:148)
> at 
> org.apache.hadoop.hive.ql.exec.tez.TezTask.execute(TezTask.java:175)
> at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:160)
> at 
> org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:89)
> at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:1673)
> at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:1432)
> at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1213)
> at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1064)
> at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1054)
> at 
> org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:213)
> at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:165)
> at 
> org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:376)
> at 
> org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:311)
> at 
> org.apache.hadoop.hive.cli.CliDriver.processReader(CliDriver.java:409)
> at 
> org.apache.hadoop.hive.cli.CliDriver.processFile(CliDriver.java:425)
> at 
> org.apache.hadoop.hive.cli.CliDriver.executeDriver(CliDriver.java:714)
> at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:681)
> at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:621)
> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
> at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> at java.lang.reflect.Method.invoke(Method.java:497)
> at org.apache.hadoop.util.RunJar.run(RunJar.java:221)
> at org.apache.hadoop.util.RunJar.main(RunJar.java:136)
> {noformat}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-11244) Beeline prompt info improvement for cluster mode

2015-07-17 Thread Nemon Lou (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-11244?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Nemon Lou updated HIVE-11244:
-
Attachment: HIVE-11244.2.patch

Code style fixed.
I do see the failed test with this patch ,but after revert it and patch 
again,then fail to reproduce.

> Beeline prompt info improvement for cluster mode
> 
>
> Key: HIVE-11244
> URL: https://issues.apache.org/jira/browse/HIVE-11244
> Project: Hive
>  Issue Type: Improvement
>  Components: Beeline
>Affects Versions: 1.2.0
>Reporter: Nemon Lou
>Assignee: Nemon Lou
>Priority: Minor
>  Labels: TODOC2.0
> Attachments: HIVE-11244.1.patch, HIVE-11244.2.patch, 
> HIVE-11244.patch, HIVE-11244.patch, HIVE-11244.patch
>
>
> Currently Beeline prompt info for Cluster mode is like this:
> {noformat}
> 0: jdbc:hive2://192.168.115.1:24002,192.168.1>
> {noformat}
> Using the very HiveServer2's IP that this beeline connect to is more helpful 
> for users.
> Like this:
> {noformat}
> 0: jdbc:hive2://192.168.115.1:24002>
> {noformat}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-11077) Add support in parser and wire up to txn manager

2015-07-17 Thread Eugene Koifman (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-11077?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Eugene Koifman updated HIVE-11077:
--
Attachment: HIVE-11077.6.patch

> Add support in parser and wire up to txn manager
> 
>
> Key: HIVE-11077
> URL: https://issues.apache.org/jira/browse/HIVE-11077
> Project: Hive
>  Issue Type: Sub-task
>  Components: SQL, Transactions
>Affects Versions: 1.2.0
>Reporter: Eugene Koifman
>Assignee: Eugene Koifman
> Attachments: HIVE-11077.3.patch, HIVE-11077.5.patch, 
> HIVE-11077.6.patch, HIVE-11077.patch
>
>




--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-11301) thrift metastore issue when getting stats results in disconnect

2015-07-17 Thread Sergey Shelukhin (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-11301?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14632175#comment-14632175
 ] 

Sergey Shelukhin commented on HIVE-11301:
-

(I know what the problem is, I don't have cycles to fix it right now :)

> thrift metastore issue when getting stats results in disconnect
> ---
>
> Key: HIVE-11301
> URL: https://issues.apache.org/jira/browse/HIVE-11301
> Project: Hive
>  Issue Type: Bug
>  Components: Metastore
>Reporter: Sergey Shelukhin
>
> On metastore side it looks like this:
> {noformat}
> 2015-07-17 20:32:27,795 ERROR [pool-3-thread-150]: server.TThreadPoolServer 
> (TThreadPoolServer.java:run(294)) - Thrift error occurred during processing 
> of message.
> org.apache.thrift.protocol.TProtocolException: Required field 'colStats' is 
> unset! Struct:AggrStats(colStats:null, partsFound:0)
> at 
> org.apache.hadoop.hive.metastore.api.AggrStats.validate(AggrStats.java:389)
> at 
> org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$get_aggr_stats_for_result.validate(ThriftHiveMetastore.java)
> at 
> org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$get_aggr_stats_for_result$get_aggr_stats_for_resultStandardScheme.write(ThriftHiveMetastore.java)
> at 
> org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$get_aggr_stats_for_result$get_aggr_stats_for_resultStandardScheme.write(ThriftHiveMetastore.java)
> at 
> org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$get_aggr_stats_for_result.write(ThriftHiveMetastore.java)
> at org.apache.thrift.ProcessFunction.process(ProcessFunction.java:53)
> at 
> org.apache.hadoop.hive.metastore.TUGIBasedProcessor$1.run(TUGIBasedProcessor.java:110)
> at 
> org.apache.hadoop.hive.metastore.TUGIBasedProcessor$1.run(TUGIBasedProcessor.java:106)
> at java.security.AccessController.doPrivileged(Native Method)
> at javax.security.auth.Subject.doAs(Subject.java:422)
> at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1657)
> at 
> org.apache.hadoop.hive.metastore.TUGIBasedProcessor.process(TUGIBasedProcessor.java:118)
> at 
> org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPoolServer.java:285)
> at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
> at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
> at java.lang.Thread.run(Thread.java:745)
> {noformat}
> and then
> {noformat}
> 2015-07-17 20:32:27,796 WARN  [pool-3-thread-150]: 
> transport.TIOStreamTransport (TIOStreamTransport.java:close(112)) - Error 
> closing output stream.
> java.net.SocketException: Socket closed
> at 
> java.net.SocketOutputStream.socketWrite(SocketOutputStream.java:116)
> at java.net.SocketOutputStream.write(SocketOutputStream.java:153)
> at 
> java.io.BufferedOutputStream.flushBuffer(BufferedOutputStream.java:82)
> at java.io.BufferedOutputStream.flush(BufferedOutputStream.java:140)
> at java.io.FilterOutputStream.close(FilterOutputStream.java:158)
> at 
> org.apache.thrift.transport.TIOStreamTransport.close(TIOStreamTransport.java:110)
> at org.apache.thrift.transport.TSocket.close(TSocket.java:196)
> at 
> org.apache.hadoop.hive.thrift.TFilterTransport.close(TFilterTransport.java:52)
> at 
> org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPoolServer.java:304)
> at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
> at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
> at java.lang.Thread.run(Thread.java:745)
> {noformat}
> Which on client manifests as
> {noformat}
> 2015-07-17 20:32:27,796 WARN  [main()]: metastore.RetryingMetaStoreClient 
> (RetryingMetaStoreClient.java:invoke(187)) - MetaStoreClient lost connection. 
> Attempting to reconnect.
> org.apache.thrift.transport.TTransportException
> at 
> org.apache.thrift.transport.TIOStreamTransport.read(TIOStreamTransport.java:132)
> at org.apache.thrift.transport.TTransport.readAll(TTransport.java:86)
> at 
> org.apache.thrift.protocol.TBinaryProtocol.readAll(TBinaryProtocol.java:429)
> at 
> org.apache.thrift.protocol.TBinaryProtocol.readI32(TBinaryProtocol.java:318)
> at 
> org.apache.thrift.protocol.TBinaryProtocol.readMessageBegin(TBinaryProtocol.java:219)
> at 
> org.apache.thrift.TServiceClient.receiveBase(TServiceClient.java:69)
> at 
> org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.recv_get_aggr_stats_for(ThriftHiveMetastore.java:3029)
> at 
> org.apache.hadoop.

[jira] [Commented] (HIVE-11299) Hive query planning allocating lots of memory and hitting OOMs

2015-07-17 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-11299?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14632170#comment-14632170
 ] 

Hive QA commented on HIVE-11299:




{color:red}Overall{color}: -1 no tests executed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12745898/HIVE-11299.patch

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/4644/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/4644/console
Test logs: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-4644/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Tests exited with: NonZeroExitCodeException
Command 'bash /data/hive-ptest/working/scratch/source-prep.sh' failed with exit 
status 1 and output '+ [[ -n /usr/java/jdk1.7.0_45-cloudera ]]
+ export JAVA_HOME=/usr/java/jdk1.7.0_45-cloudera
+ JAVA_HOME=/usr/java/jdk1.7.0_45-cloudera
+ export 
PATH=/usr/java/jdk1.7.0_45-cloudera/bin/:/usr/local/apache-maven-3.0.5/bin:/usr/java/jdk1.7.0_45-cloudera/bin:/usr/local/apache-ant-1.9.1/bin:/usr/local/bin:/bin:/usr/bin:/usr/local/sbin:/usr/sbin:/sbin:/home/hiveptest/bin
+ 
PATH=/usr/java/jdk1.7.0_45-cloudera/bin/:/usr/local/apache-maven-3.0.5/bin:/usr/java/jdk1.7.0_45-cloudera/bin:/usr/local/apache-ant-1.9.1/bin:/usr/local/bin:/bin:/usr/bin:/usr/local/sbin:/usr/sbin:/sbin:/home/hiveptest/bin
+ export 'ANT_OPTS=-Xmx1g -XX:MaxPermSize=256m '
+ ANT_OPTS='-Xmx1g -XX:MaxPermSize=256m '
+ export 'M2_OPTS=-Xmx1g -XX:MaxPermSize=256m -Dhttp.proxyHost=localhost 
-Dhttp.proxyPort=3128'
+ M2_OPTS='-Xmx1g -XX:MaxPermSize=256m -Dhttp.proxyHost=localhost 
-Dhttp.proxyPort=3128'
+ cd /data/hive-ptest/working/
+ tee /data/hive-ptest/logs/PreCommit-HIVE-TRUNK-Build-4644/source-prep.txt
+ [[ false == \t\r\u\e ]]
+ mkdir -p maven ivy
+ [[ git = \s\v\n ]]
+ [[ git = \g\i\t ]]
+ [[ -z master ]]
+ [[ -d apache-github-source-source ]]
+ [[ ! -d apache-github-source-source/.git ]]
+ [[ ! -d apache-github-source-source ]]
+ cd apache-github-source-source
+ git fetch origin
+ git reset --hard HEAD
HEAD is now at 7b17df1 HIVE-11291 : Avoid allocation storm while doing rule 
matching on operator/expression trees (Ashutosh Chauhan via Hari Sankar)
+ git clean -f -d
+ git checkout master
Already on 'master'
+ git reset --hard origin/master
HEAD is now at 7b17df1 HIVE-11291 : Avoid allocation storm while doing rule 
matching on operator/expression trees (Ashutosh Chauhan via Hari Sankar)
+ git merge --ff-only origin/master
Already up-to-date.
+ git gc
+ patchCommandPath=/data/hive-ptest/working/scratch/smart-apply-patch.sh
+ patchFilePath=/data/hive-ptest/working/scratch/build.patch
+ [[ -f /data/hive-ptest/working/scratch/build.patch ]]
+ chmod +x /data/hive-ptest/working/scratch/smart-apply-patch.sh
+ /data/hive-ptest/working/scratch/smart-apply-patch.sh 
/data/hive-ptest/working/scratch/build.patch
The patch does not appear to apply with p0, p1, or p2
+ exit 1
'
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12745898 - PreCommit-HIVE-TRUNK-Build

> Hive query planning allocating lots of memory and hitting OOMs 
> ---
>
> Key: HIVE-11299
> URL: https://issues.apache.org/jira/browse/HIVE-11299
> Project: Hive
>  Issue Type: Bug
>  Components: Hive
>Affects Versions: 0.14.0
>Reporter: Mostafa Mokhtar
>Assignee: Mostafa Mokhtar
> Attachments: HIVE-11299.patch
>
>
> Plan generation for queries with lots of disjunct filters spends lots of time 
> compiling. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-11209) Clean up dependencies in HiveDecimalWritable

2015-07-17 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-11209?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14632168#comment-14632168
 ] 

Hive QA commented on HIVE-11209:




{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12745893/HIVE-11209.patch

{color:red}ERROR:{color} -1 due to 1 failed/errored test(s), 9226 tests executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_cbo_rp_join0
{noformat}

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/4643/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/4643/console
Test logs: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-4643/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 1 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12745893 - PreCommit-HIVE-TRUNK-Build

> Clean up dependencies in HiveDecimalWritable
> 
>
> Key: HIVE-11209
> URL: https://issues.apache.org/jira/browse/HIVE-11209
> Project: Hive
>  Issue Type: Sub-task
>Reporter: Owen O'Malley
>Assignee: Owen O'Malley
> Fix For: 2.0.0
>
> Attachments: HIVE-11209.patch, HIVE-11209.patch, HIVE-11209.patch, 
> HIVE-11209.patch
>
>
> Currently HiveDecimalWritable depends on:
> * org.apache.hadoop.hive.serde2.ByteStream
> * org.apache.hadoop.hive.serde2.lazybinary.LazyBinaryUtils
> * org.apache.hadoop.hive.serde2.typeinfo.HiveDecimalUtils
> since we need HiveDecimalWritable for the decimal VectorizedColumnBatch, 
> breaking these dependencies will improve things.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-11295) LLAP: clean up ORC dependencies on object pools

2015-07-17 Thread Sergey Shelukhin (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-11295?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14632162#comment-14632162
 ] 

Sergey Shelukhin commented on HIVE-11295:
-

Btw this officially makes Hive Enterprise software, cause it adds what amounts 
to a FactoryFactory. Maybe we should keep the interfaces, but remove the pools 
for objects that don't cross thread boundaries, and re-add them if we see 
problems again. Maybe the problems were only with objects that do cross thread 
boundaries. Maybe we also need pool for LlapDataBuffer but that would have to 
tie in with refcounting.

> LLAP: clean up ORC dependencies on object pools
> ---
>
> Key: HIVE-11295
> URL: https://issues.apache.org/jira/browse/HIVE-11295
> Project: Hive
>  Issue Type: Sub-task
>Reporter: Sergey Shelukhin
>Assignee: Sergey Shelukhin
> Attachments: HIVE-11259.patch
>
>
> Before there's storage handler module, we can clean some things up



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-11295) LLAP: clean up ORC dependencies on object pools

2015-07-17 Thread Sergey Shelukhin (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-11295?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14632136#comment-14632136
 ] 

Sergey Shelukhin commented on HIVE-11295:
-

After the other one would be better.

> LLAP: clean up ORC dependencies on object pools
> ---
>
> Key: HIVE-11295
> URL: https://issues.apache.org/jira/browse/HIVE-11295
> Project: Hive
>  Issue Type: Sub-task
>Reporter: Sergey Shelukhin
>Assignee: Sergey Shelukhin
> Attachments: HIVE-11259.patch
>
>
> Before there's storage handler module, we can clean some things up



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-10884) Enable some beeline tests and turn on HIVE-4239 by default

2015-07-17 Thread Sergey Shelukhin (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-10884?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14632135#comment-14632135
 ] 

Sergey Shelukhin commented on HIVE-10884:
-

Looks like it timed out without any output saved...
{noformat}
2015-07-16 10:21:04,139  INFO HostExecutor.executeTests:176 Starting isolated 
execution on 54.80.208.185
2015-07-16 11:57:56,933  INFO LocalCommand.:35 Starting timeout 1h rsync 
-vaPe "ssh -i /home/hiveptest/.ssh/hive-ptest-user-key" --timeout 600 
hiveptest@107.22.49.60:/home/hiveptest/107.22.49.60-hiveptest-2/logs/ 
/data/hive-ptest/logs/PreCommit-HIVE-TRUNK-Build-4628/failed/TestBeeLineDriver
20
{noformat}
Would the above directory be preserved somewhere?

> Enable some beeline tests and turn on HIVE-4239 by default
> --
>
> Key: HIVE-10884
> URL: https://issues.apache.org/jira/browse/HIVE-10884
> Project: Hive
>  Issue Type: Bug
>Reporter: Sergey Shelukhin
>Assignee: Sergey Shelukhin
> Attachments: HIVE-10884.01.patch, HIVE-10884.02.patch, 
> HIVE-10884.03.patch, HIVE-10884.04.patch, HIVE-10884.05.patch, 
> HIVE-10884.06.patch, HIVE-10884.07.patch, HIVE-10884.07.patch, 
> HIVE-10884.patch
>
>
> See comments in HIVE-4239.
> Beeline tests with parallelism need to be enabled to turn compilation 
> parallelism on by default.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-10799) Refactor the SearchArgumentFactory to remove the dependence on ExprNodeGenericFuncDesc

2015-07-17 Thread Owen O'Malley (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-10799?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Owen O'Malley updated HIVE-10799:
-
Attachment: HIVE-10799.patch

Fixed a couple of problems.

> Refactor the SearchArgumentFactory to remove the dependence on 
> ExprNodeGenericFuncDesc
> --
>
> Key: HIVE-10799
> URL: https://issues.apache.org/jira/browse/HIVE-10799
> Project: Hive
>  Issue Type: Sub-task
>Reporter: Owen O'Malley
>Assignee: Owen O'Malley
> Attachments: HIVE-10799.patch, HIVE-10799.patch, HIVE-10799.patch, 
> HIVE-10799.patch
>
>
> SearchArgumentFactory and SearchArgumentImpl are high level and shouldn't 
> depend on the internals of Hive's AST model.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-11288) Avro SerDe InstanceCache returns incorrect schema

2015-07-17 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-11288?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14632084#comment-14632084
 ] 

Hive QA commented on HIVE-11288:




{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12745875/HIVE-11288.2.patch

{color:red}ERROR:{color} -1 due to 1 failed/errored test(s), 9226 tests executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_cbo_rp_join0
{noformat}

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/4642/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/4642/console
Test logs: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-4642/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 1 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12745875 - PreCommit-HIVE-TRUNK-Build

> Avro SerDe InstanceCache returns incorrect schema
> -
>
> Key: HIVE-11288
> URL: https://issues.apache.org/jira/browse/HIVE-11288
> Project: Hive
>  Issue Type: Bug
>Reporter: Greg Phillips
> Attachments: HIVE-11288.2.patch, HIVE-11288.patch
>
>
> To reproduce this error, take two fields in an avro schema document matching 
> the following:
> "type" :  { "type": "array", "items": [ "null",  { "type": "map", "values": [ 
> "null", "string" ] } ]  }
> "type" : { "type": "map", "values": [ "null" , { "type": "array", "items": [ 
> "null" , "string"] } ] }
> After creating two tables in hive with these schemas, the describe statement 
> on each of them will only return the schema for the first one loaded.  This 
> is due to a hashCode() collision in the InstanceCache.  
> A patch will be included in this ticket shortly which removes the hashCode 
> call from the InstanceCache's internal HashMap, and instead provides the 
> entire schema object.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-10953) Get partial stats instead of complete stats in some queries

2015-07-17 Thread Alan Gates (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-10953?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alan Gates updated HIVE-10953:
--
Affects Version/s: hbase-metastore-branch

> Get partial stats instead of complete stats in some queries
> ---
>
> Key: HIVE-10953
> URL: https://issues.apache.org/jira/browse/HIVE-10953
> Project: Hive
>  Issue Type: Sub-task
>  Components: Metastore
>Affects Versions: hbase-metastore-branch
>Reporter: Daniel Dai
>Assignee: Vaibhav Gumashta
> Fix For: hbase-metastore-branch
>
>
> In ppd_constant_where.q, the result is different than benchmark:
> Result:
> Statistics: Num rows: 0 Data size: 11624 Basic stats: PARTIAL Column stats: 
> NONE
> Benchmark:
> Statistics: Num rows: 1000 Data size: 10624 Basic stats: COMPLETE Column 
> stats: NONE
> This might cause quite a few failures so we need to investigate it first.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-10950) Unit test against HBase Metastore

2015-07-17 Thread Alan Gates (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-10950?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alan Gates updated HIVE-10950:
--
Affects Version/s: hbase-metastore-branch

> Unit test against HBase Metastore
> -
>
> Key: HIVE-10950
> URL: https://issues.apache.org/jira/browse/HIVE-10950
> Project: Hive
>  Issue Type: Sub-task
>  Components: Metastore
>Affects Versions: hbase-metastore-branch
>Reporter: Daniel Dai
>Assignee: Daniel Dai
> Fix For: hbase-metastore-branch
>
> Attachments: HIVE-10950-1.patch, HIVE-10950-2.patch
>
>
> We need to run the entire Hive UT against HBase Metastore and make sure they 
> pass.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-10289) Support filter on non-first partition key and non-string partition key

2015-07-17 Thread Alan Gates (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-10289?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alan Gates updated HIVE-10289:
--
Affects Version/s: hbase-metastore-branch

> Support filter on non-first partition key and non-string partition key
> --
>
> Key: HIVE-10289
> URL: https://issues.apache.org/jira/browse/HIVE-10289
> Project: Hive
>  Issue Type: Sub-task
>  Components: Metastore
>Affects Versions: hbase-metastore-branch
>Reporter: Daniel Dai
>Assignee: Daniel Dai
> Attachments: HIVE-10289.1.patch
>
>
> Currently, partition filtering only handles the first partition key and the 
> type for this partition key must be string. In order to break this 
> limitation, several improvements are required:
> 1. Change serialization format for partition key. Currently partition keys 
> are serialized into delimited string, which sorted on string order not with 
> regard to the actual type of the partition key. We use BinarySortableSerDe 
> for this purpose.
> 2. For filter condition not on the initial partition keys, push it into HBase 
> RowFilter. RowFilter will deserialize the partition key and evaluate the 
> filter condition.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-10169) get metatool to work with hbase metastore

2015-07-17 Thread Alan Gates (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-10169?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alan Gates updated HIVE-10169:
--
Affects Version/s: hbase-metastore-branch

> get metatool to work with hbase metastore
> -
>
> Key: HIVE-10169
> URL: https://issues.apache.org/jira/browse/HIVE-10169
> Project: Hive
>  Issue Type: Sub-task
>  Components: Metastore
>Affects Versions: hbase-metastore-branch
>Reporter: Thejas M Nair
>
> The metatool is used for  enabling namenode HA, and it uses ObjectStore 
> directly. 
> There needs to be a way to support equivalent functionality with hbase 
> metastore.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Resolved] (HIVE-9905) Investigate ways to improve NDV calculations during stats aggregation [hbase-metastore branch]

2015-07-17 Thread Alan Gates (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9905?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alan Gates resolved HIVE-9905.
--
Resolution: Duplicate

This isn't HBase metastore specific.  Besides it looks like HIVE-9689 will do 
what we want, so closing this as a dup of that.

> Investigate ways to improve NDV calculations during stats aggregation 
> [hbase-metastore branch]
> --
>
> Key: HIVE-9905
> URL: https://issues.apache.org/jira/browse/HIVE-9905
> Project: Hive
>  Issue Type: Sub-task
>  Components: Metastore
>Reporter: Vaibhav Gumashta
>Assignee: Vaibhav Gumashta
>




--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9904) Perform stats aggregation in HBase co-processor [hbase-metastore branch]

2015-07-17 Thread Alan Gates (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9904?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alan Gates updated HIVE-9904:
-
Affects Version/s: hbase-metastore-branch

> Perform stats aggregation in HBase co-processor [hbase-metastore branch] 
> -
>
> Key: HIVE-9904
> URL: https://issues.apache.org/jira/browse/HIVE-9904
> Project: Hive
>  Issue Type: Sub-task
>  Components: Metastore
>Affects Versions: hbase-metastore-branch
>Reporter: Vaibhav Gumashta
>Assignee: Vaibhav Gumashta
>




--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-11210) Remove dependency on HiveConf from Orc reader & writer

2015-07-17 Thread Owen O'Malley (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-11210?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Owen O'Malley updated HIVE-11210:
-
Attachment: HIVE-11210.patch

Sorry, I managed to get the wrong version of the patch.

> Remove dependency on HiveConf from Orc reader & writer
> --
>
> Key: HIVE-11210
> URL: https://issues.apache.org/jira/browse/HIVE-11210
> Project: Hive
>  Issue Type: Sub-task
>Reporter: Owen O'Malley
>Assignee: Owen O'Malley
> Attachments: HIVE-11210.patch, HIVE-11210.patch
>
>
> Currently the ORC reader and writer get their default values from HiveConf. I 
> propose that we make the reader and writer have their own programatic 
> defaults and the OrcInputFormat and OrcOutputFormat can use the version in 
> HiveConf.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9578) Add support for getDatabases and alterDatabase calls [hbase-metastore branch]

2015-07-17 Thread Alan Gates (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9578?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alan Gates updated HIVE-9578:
-
Affects Version/s: hbase-metastore-branch

> Add support for getDatabases and alterDatabase calls [hbase-metastore branch]
> -
>
> Key: HIVE-9578
> URL: https://issues.apache.org/jira/browse/HIVE-9578
> Project: Hive
>  Issue Type: Sub-task
>  Components: Metastore
>Affects Versions: hbase-metastore-branch
>Reporter: Alan Gates
>Assignee: Alan Gates
> Fix For: hbase-metastore-branch
>
> Attachments: HIVE-9578.2.patch, HIVE-9578.patch
>
>
> The initial patch only supporting getting a single database, add database, 
> and drop database.  Support needs to be added for alter database, getting all 
> the databases, and getting database names by pattern.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9453) Initial patch [hbase-metastore branch]

2015-07-17 Thread Alan Gates (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9453?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alan Gates updated HIVE-9453:
-
Affects Version/s: hbase-metastore-branch

> Initial patch [hbase-metastore branch]
> --
>
> Key: HIVE-9453
> URL: https://issues.apache.org/jira/browse/HIVE-9453
> Project: Hive
>  Issue Type: Sub-task
>  Components: Metastore
>Affects Versions: hbase-metastore-branch
>Reporter: Alan Gates
>Assignee: Alan Gates
> Fix For: hbase-metastore-branch
>
> Attachments: HIVE-9453-reviewcomments.pdf, HIVE-9453.2.patch, 
> HIVE-9453.3.patch, HIVE-9453.patch
>
>
> This initial patch has several important features:
> # HBaseStore, a new implementation of RawStore that stores the data in HBase.
> # Subclasses of the thrift metastore objects to remove the massive 
> duplication of data where every partition contains a nearly identical storage 
> descriptor.
> # Caches for catalog objects and statistics so that repeated metastore calls 
> don't result in repeated calls against HBase.
> Currently this "works" to the point that "load table" and "select" work.  I 
> have not tested any other statements, and I suspect most fail.  There is no 
> security, no authorization, and a no a lot of other things.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-11210) Remove dependency on HiveConf from Orc reader & writer

2015-07-17 Thread Owen O'Malley (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-11210?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Owen O'Malley updated HIVE-11210:
-
Attachment: (was: HIVE-11210.patch)

> Remove dependency on HiveConf from Orc reader & writer
> --
>
> Key: HIVE-11210
> URL: https://issues.apache.org/jira/browse/HIVE-11210
> Project: Hive
>  Issue Type: Sub-task
>Reporter: Owen O'Malley
>Assignee: Owen O'Malley
> Attachments: HIVE-11210.patch
>
>
> Currently the ORC reader and writer get their default values from HiveConf. I 
> propose that we make the reader and writer have their own programatic 
> defaults and the OrcInputFormat and OrcOutputFormat can use the version in 
> HiveConf.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9452) Use HBase to store Hive metadata

2015-07-17 Thread Alan Gates (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9452?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alan Gates updated HIVE-9452:
-
Affects Version/s: hbase-metastore-branch

> Use HBase to store Hive metadata
> 
>
> Key: HIVE-9452
> URL: https://issues.apache.org/jira/browse/HIVE-9452
> Project: Hive
>  Issue Type: Improvement
>  Components: Metastore
>Affects Versions: hbase-metastore-branch
>Reporter: Alan Gates
>Assignee: Alan Gates
> Attachments: HBaseMetastoreApproach.pdf
>
>
> This is an umbrella JIRA for a project to explore using HBase to store the 
> Hive data catalog (ie the metastore).  This project has several goals:
> # The current metastore implementation is slow when tables have thousands or 
> more partitions.  With Tez and Spark engines we are pushing Hive to a point 
> where queries only take a few seconds to run.  But planning the query can 
> take as long as running it.  Much of this time is spent in metadata 
> operations.
> # Due to scale limitations we have never allowed tasks to communicate 
> directly with the metastore.  However, with the development of LLAP this 
> requirement will have to be relaxed.  If we can relax this there are other 
> use cases that could benefit from this.  
> # Eating our own dogfood.  Rather than using external systems to store our 
> metadata there are benefits to using other components in the Hadoop system.
> The proposal is to create a new branch and work on the prototype there.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Comment Edited] (HIVE-11274) Support executing script file from hdfs in beeline

2015-07-17 Thread Xuefu Zhang (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-11274?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14632022#comment-14632022
 ] 

Xuefu Zhang edited comment on HIVE-11274 at 7/17/15 10:38 PM:
--

[~jeffzhang] does Ferdinand's comment make sense to you? I'm closing this one 
and feel free to reopen if otherwise.


was (Author: xuefuz):
[~jeffzhang] does Ferdinand's comments make sense to you? I'm closing this one 
and feel free to reopen if otherwise.

> Support executing script file from hdfs in beeline
> --
>
> Key: HIVE-11274
> URL: https://issues.apache.org/jira/browse/HIVE-11274
> Project: Hive
>  Issue Type: Improvement
>Reporter: Jeff Zhang
>Assignee: Ferdinand Xu
>
> It is supported for cli to execute hive script on hdfs, it should be 
> straightforward to support it in beeline too.
> Beeline.java
> {code}
>   private int executeFile(String fileName) {
> FileInputStream initStream = null;
> try {
>   initStream = new FileInputStream(fileName);
>   return execute(getConsoleReader(initStream), true);
> } catch (Throwable t) {
>   handleException(t);
>   return ERRNO_OTHER;
> } finally {
>   IOUtils.closeStream(initStream);
>   consoleReader = null;
>   output("");   // dummy new line
> }
>   }
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Comment Edited] (HIVE-11274) Support executing script file from hdfs in beeline

2015-07-17 Thread Xuefu Zhang (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-11274?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14632022#comment-14632022
 ] 

Xuefu Zhang edited comment on HIVE-11274 at 7/17/15 10:38 PM:
--

[~jeffzhang] does Ferdinand's comments make sense to you? I'm closing this one 
and feel free to reopen if otherwise.


was (Author: xuefuz):
[~jeffzhang] doesn't Ferdinand's comments makes sense to you? I'm closing this 
one and feel free to reopen if otherwise.

> Support executing script file from hdfs in beeline
> --
>
> Key: HIVE-11274
> URL: https://issues.apache.org/jira/browse/HIVE-11274
> Project: Hive
>  Issue Type: Improvement
>Reporter: Jeff Zhang
>Assignee: Ferdinand Xu
>
> It is supported for cli to execute hive script on hdfs, it should be 
> straightforward to support it in beeline too.
> Beeline.java
> {code}
>   private int executeFile(String fileName) {
> FileInputStream initStream = null;
> try {
>   initStream = new FileInputStream(fileName);
>   return execute(getConsoleReader(initStream), true);
> } catch (Throwable t) {
>   handleException(t);
>   return ERRNO_OTHER;
> } finally {
>   IOUtils.closeStream(initStream);
>   consoleReader = null;
>   output("");   // dummy new line
> }
>   }
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-11294) Use HBase to cache aggregated stats

2015-07-17 Thread ASF GitHub Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-11294?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14632048#comment-14632048
 ] 

ASF GitHub Bot commented on HIVE-11294:
---

GitHub user alanfgates opened a pull request:

https://github.com/apache/hive/pull/46

Hbase metastore

HIVE-11294 Cache aggregated stats in hbase

You can merge this pull request into a Git repository by running:

$ git pull https://github.com/alanfgates/hive hbase-metastore

Alternatively you can review and apply these changes as the patch at:

https://github.com/apache/hive/pull/46.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

This closes #46


commit 58ca965ca58925e4846e14cf37d64638edb3e353
Author: Alan Gates 
Date:   2015-01-23T19:05:19Z

Creating branch for hbase metastore development.


git-svn-id: 
https://svn.apache.org/repos/asf/hive/branches/hbase-metastore@1654331 
13f79535-47bb-0310-9956-ffa450edef68

commit e81704eeaf061e7152c63379b828debdfdfe9a8e
Author: Alan Gates 
Date:   2015-02-04T20:00:49Z

HIVE-9453 Initial patch [hbase-metastore branch] (Alan Gates, reviewed by 
Thejas Nair)


git-svn-id: 
https://svn.apache.org/repos/asf/hive/branches/hbase-metastore@1657394 
13f79535-47bb-0310-9956-ffa450edef68

commit 120047c308a011f17f94258779e4c47fcff191a5
Author: Alan Gates 
Date:   2015-02-10T21:41:19Z

HIVE-9578 Add support for getDatabases and alterDatabase calls 
[hbase-metastore branch]


git-svn-id: 
https://svn.apache.org/repos/asf/hive/branches/hbase-metastore@1658821 
13f79535-47bb-0310-9956-ffa450edef68

commit c7dcf4a85d5cef4151d7b72115b4a1027af488ac
Author: Alan Gates 
Date:   2015-02-10T22:47:53Z

HIVE-9579 Support all get tables [hbase-metastore branch]


git-svn-id: 
https://svn.apache.org/repos/asf/hive/branches/hbase-metastore@1658829 
13f79535-47bb-0310-9956-ffa450edef68

commit cd82601f6e5c4aea5db72b74629b6aada23cf7c2
Author: Alan Gates 
Date:   2015-02-19T18:51:39Z

HIVE-9606 Need a tool to export metadata from RDBMS based metastore into 
HBase [hbase-metastore branch] (Alan Gates, reviewed by Thejas Nair)


git-svn-id: 
https://svn.apache.org/repos/asf/hive/branches/hbase-metastore@1660974 
13f79535-47bb-0310-9956-ffa450edef68

commit 21ba0a93dcadcf8726ae7a18e81f758921552697
Author: Alan Gates 
Date:   2015-03-02T18:17:01Z

HIVE-9641 Fill out remaining partition functions in HBaseStore (Alan Gates 
reviewed by Vaibhav Gumashta)


git-svn-id: 
https://svn.apache.org/repos/asf/hive/branches/hbase-metastore@1663386 
13f79535-47bb-0310-9956-ffa450edef68

commit bb0010caa8734af34ecac7b5b054afd95b49b0e0
Author: Alan Gates 
Date:   2015-03-05T03:46:28Z

HIVE-9677 Implement privileges call in HBaseStore (Alan Gates)


git-svn-id: 
https://svn.apache.org/repos/asf/hive/branches/hbase-metastore@1664220 
13f79535-47bb-0310-9956-ffa450edef68

commit 408d467bffa0923714b8262427f1d1d72ba9932d
Author: Alan Gates 
Date:   2015-03-05T04:09:31Z

HIVE-9763 Remove M* classes from RawStore interface (Alan Gates)


git-svn-id: 
https://svn.apache.org/repos/asf/hive/branches/hbase-metastore@1664226 
13f79535-47bb-0310-9956-ffa450edef68

commit be575b84a8d5b4764550499c60bdccb7b0d1f82f
Author: Alan Gates 
Date:   2015-03-05T04:35:53Z

HIVE-9802 Refactor HBaseReadWrite to allow different implementations 
underneath, also includes code from HIVE-9783 Move serialization of objects in 
HBase to protocol buffers (Alan Gates)


git-svn-id: 
https://svn.apache.org/repos/asf/hive/branches/hbase-metastore@1664227 
13f79535-47bb-0310-9956-ffa450edef68

commit 646468fcdfd612524961751ba3575a26a1fb3fcb
Author: Alan Gates 
Date:   2015-03-12T19:41:56Z

HIVE-9885 HBase metastore protobufs need to check hasX() functions for 
values that are not auto set in serialization [hbase-metastore branch] (Alan 
Gates)


git-svn-id: 
https://svn.apache.org/repos/asf/hive/branches/hbase-metastore@1666280 
13f79535-47bb-0310-9956-ffa450edef68

commit d268b81bcd2fbaae6fdbfe2f980f644ce5e3d92d
Author: Alan Gates 
Date:   2015-03-12T20:01:30Z

HIVE-9874 Partition storage descriptors being set from table sd without 
copying [hbase-metastore branch] (Alan Gates)


git-svn-id: 
https://svn.apache.org/repos/asf/hive/branches/hbase-metastore@1666285 
13f79535-47bb-0310-9956-ffa450edef68

commit 24f224867b70bacc91a877242aee1b0b84efd127
Author: Alan Gates 
Date:   2015-03-17T15:50:31Z

HIVE-9942 Implement functions methods in HBaseStore (Alan Gates)


git-svn-id: 
https://svn.apache.org/repos/asf/hive/branches/hbase-metastore@1667338 
13f79535-47bb-0310-9956-ffa450edef68

commit 35ac8b0257614523064c1b3647a91e123b78e999
Author: Alan Gates 
Date:   2015-03-18T17:41:50Z

HIVE-10008 Ne

[jira] [Commented] (HIVE-11274) Support executing script file from hdfs in beeline

2015-07-17 Thread Jeff Zhang (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-11274?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14632043#comment-14632043
 ] 

Jeff Zhang commented on HIVE-11274:
---

Make sense. 

> Support executing script file from hdfs in beeline
> --
>
> Key: HIVE-11274
> URL: https://issues.apache.org/jira/browse/HIVE-11274
> Project: Hive
>  Issue Type: Improvement
>Reporter: Jeff Zhang
>Assignee: Ferdinand Xu
>
> It is supported for cli to execute hive script on hdfs, it should be 
> straightforward to support it in beeline too.
> Beeline.java
> {code}
>   private int executeFile(String fileName) {
> FileInputStream initStream = null;
> try {
>   initStream = new FileInputStream(fileName);
>   return execute(getConsoleReader(initStream), true);
> } catch (Throwable t) {
>   handleException(t);
>   return ERRNO_OTHER;
> } finally {
>   IOUtils.closeStream(initStream);
>   consoleReader = null;
>   output("");   // dummy new line
> }
>   }
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Resolved] (HIVE-11274) Support executing script file from hdfs in beeline

2015-07-17 Thread Xuefu Zhang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-11274?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Xuefu Zhang resolved HIVE-11274.

Resolution: Won't Fix

[~jeffzhang] doesn't Ferdinand's comments makes sense to you? I'm closing this 
one and feel free to reopen if otherwise.

> Support executing script file from hdfs in beeline
> --
>
> Key: HIVE-11274
> URL: https://issues.apache.org/jira/browse/HIVE-11274
> Project: Hive
>  Issue Type: Improvement
>Reporter: Jeff Zhang
>Assignee: Ferdinand Xu
>
> It is supported for cli to execute hive script on hdfs, it should be 
> straightforward to support it in beeline too.
> Beeline.java
> {code}
>   private int executeFile(String fileName) {
> FileInputStream initStream = null;
> try {
>   initStream = new FileInputStream(fileName);
>   return execute(getConsoleReader(initStream), true);
> } catch (Throwable t) {
>   handleException(t);
>   return ERRNO_OTHER;
> } finally {
>   IOUtils.closeStream(initStream);
>   consoleReader = null;
>   output("");   // dummy new line
> }
>   }
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-11299) Hive query planning allocating lots of memory and hitting OOMs

2015-07-17 Thread Mostafa Mokhtar (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-11299?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mostafa Mokhtar updated HIVE-11299:
---
Attachment: HIVE-11299.patch

> Hive query planning allocating lots of memory and hitting OOMs 
> ---
>
> Key: HIVE-11299
> URL: https://issues.apache.org/jira/browse/HIVE-11299
> Project: Hive
>  Issue Type: Bug
>  Components: Hive
>Affects Versions: 0.14.0
>Reporter: Mostafa Mokhtar
>Assignee: Mostafa Mokhtar
> Attachments: HIVE-11299.patch
>
>
> Plan generation for queries with lots of disjunct filters spends lots of time 
> compiling. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-11299) Hive query planning allocating lots of memory and hitting OOMs

2015-07-17 Thread Mostafa Mokhtar (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-11299?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mostafa Mokhtar updated HIVE-11299:
---
Attachment: (was: HIVE-11299.patch)

> Hive query planning allocating lots of memory and hitting OOMs 
> ---
>
> Key: HIVE-11299
> URL: https://issues.apache.org/jira/browse/HIVE-11299
> Project: Hive
>  Issue Type: Bug
>  Components: Hive
>Affects Versions: 0.14.0
>Reporter: Mostafa Mokhtar
>Assignee: Mostafa Mokhtar
>
> Plan generation for queries with lots of disjunct filters spends lots of time 
> compiling. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-11299) Hive query planning allocating lots of memory and hitting OOMs

2015-07-17 Thread Mostafa Mokhtar (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-11299?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mostafa Mokhtar updated HIVE-11299:
---
Attachment: HIVE-11299.patch

> Hive query planning allocating lots of memory and hitting OOMs 
> ---
>
> Key: HIVE-11299
> URL: https://issues.apache.org/jira/browse/HIVE-11299
> Project: Hive
>  Issue Type: Bug
>  Components: Hive
>Affects Versions: 0.14.0
>Reporter: Mostafa Mokhtar
>Assignee: Mostafa Mokhtar
> Attachments: HIVE-11299.patch
>
>
> Plan generation for queries with lots of disjunct filters spends lots of time 
> compiling. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-11296) Merge from master to spark branch [Spark Branch]

2015-07-17 Thread Xuefu Zhang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-11296?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Xuefu Zhang updated HIVE-11296:
---
Component/s: (was: spark-branch)
 Spark
Summary: Merge from master to spark branch [Spark Branch]  (was: Merge 
from master to spark branch)

> Merge from master to spark branch [Spark Branch]
> 
>
> Key: HIVE-11296
> URL: https://issues.apache.org/jira/browse/HIVE-11296
> Project: Hive
>  Issue Type: Bug
>  Components: Spark
>Reporter: Chao Sun
>Assignee: Chao Sun
>




--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-11209) Clean up dependencies in HiveDecimalWritable

2015-07-17 Thread Owen O'Malley (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-11209?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Owen O'Malley updated HIVE-11209:
-
Attachment: HIVE-11209.patch

Found and fixed a screwup in the code that I moved over. I also fixed some of 
the debugging code that got exercised when the unit tests failed by adding 
INTERVAL_YEAR_MONTH and INTERVAL_DAY_TIME support to the buildJSONString.

> Clean up dependencies in HiveDecimalWritable
> 
>
> Key: HIVE-11209
> URL: https://issues.apache.org/jira/browse/HIVE-11209
> Project: Hive
>  Issue Type: Sub-task
>Reporter: Owen O'Malley
>Assignee: Owen O'Malley
> Fix For: 2.0.0
>
> Attachments: HIVE-11209.patch, HIVE-11209.patch, HIVE-11209.patch, 
> HIVE-11209.patch
>
>
> Currently HiveDecimalWritable depends on:
> * org.apache.hadoop.hive.serde2.ByteStream
> * org.apache.hadoop.hive.serde2.lazybinary.LazyBinaryUtils
> * org.apache.hadoop.hive.serde2.typeinfo.HiveDecimalUtils
> since we need HiveDecimalWritable for the decimal VectorizedColumnBatch, 
> breaking these dependencies will improve things.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-11294) Use HBase to cache aggregated stats

2015-07-17 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-11294?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14631945#comment-14631945
 ] 

Hive QA commented on HIVE-11294:




{color:red}Overall{color}: -1 no tests executed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12745874/HIVE-11294.patch

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/4641/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/4641/console
Test logs: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-4641/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Tests exited with: NonZeroExitCodeException
Command 'bash /data/hive-ptest/working/scratch/source-prep.sh' failed with exit 
status 1 and output '+ [[ -n /usr/java/jdk1.7.0_45-cloudera ]]
+ export JAVA_HOME=/usr/java/jdk1.7.0_45-cloudera
+ JAVA_HOME=/usr/java/jdk1.7.0_45-cloudera
+ export 
PATH=/usr/java/jdk1.7.0_45-cloudera/bin/:/usr/local/apache-maven-3.0.5/bin:/usr/java/jdk1.7.0_45-cloudera/bin:/usr/local/apache-ant-1.9.1/bin:/usr/local/bin:/bin:/usr/bin:/usr/local/sbin:/usr/sbin:/sbin:/home/hiveptest/bin
+ 
PATH=/usr/java/jdk1.7.0_45-cloudera/bin/:/usr/local/apache-maven-3.0.5/bin:/usr/java/jdk1.7.0_45-cloudera/bin:/usr/local/apache-ant-1.9.1/bin:/usr/local/bin:/bin:/usr/bin:/usr/local/sbin:/usr/sbin:/sbin:/home/hiveptest/bin
+ export 'ANT_OPTS=-Xmx1g -XX:MaxPermSize=256m '
+ ANT_OPTS='-Xmx1g -XX:MaxPermSize=256m '
+ export 'M2_OPTS=-Xmx1g -XX:MaxPermSize=256m -Dhttp.proxyHost=localhost 
-Dhttp.proxyPort=3128'
+ M2_OPTS='-Xmx1g -XX:MaxPermSize=256m -Dhttp.proxyHost=localhost 
-Dhttp.proxyPort=3128'
+ cd /data/hive-ptest/working/
+ tee /data/hive-ptest/logs/PreCommit-HIVE-TRUNK-Build-4641/source-prep.txt
+ [[ false == \t\r\u\e ]]
+ mkdir -p maven ivy
+ [[ git = \s\v\n ]]
+ [[ git = \g\i\t ]]
+ [[ -z master ]]
+ [[ -d apache-github-source-source ]]
+ [[ ! -d apache-github-source-source/.git ]]
+ [[ ! -d apache-github-source-source ]]
+ cd apache-github-source-source
+ git fetch origin
>From https://github.com/apache/hive
   7fe23aa..8646c12  master -> origin/master
   dddffe1..4221699  spark  -> origin/spark
+ git reset --hard HEAD
HEAD is now at 7fe23aa HIVE-11282: CBO (Calcite Return Path): Inferring Hive 
type char/varchar of length zero which is not allowed (Jesus Camacho Rodriguez, 
reviewed by Ashutosh Chauhan)
+ git clean -f -d
+ git checkout master
Already on 'master'
Your branch is behind 'origin/master' by 2 commits, and can be fast-forwarded.
+ git reset --hard origin/master
HEAD is now at 8646c12 HIVE-11285 : ObjectInspector for partition columns in 
FetchOperator in SMBJoin causes exception (Pengcheng Xiong via Ashutosh Chauhan)
+ git merge --ff-only origin/master
Already up-to-date.
+ git gc
+ patchCommandPath=/data/hive-ptest/working/scratch/smart-apply-patch.sh
+ patchFilePath=/data/hive-ptest/working/scratch/build.patch
+ [[ -f /data/hive-ptest/working/scratch/build.patch ]]
+ chmod +x /data/hive-ptest/working/scratch/smart-apply-patch.sh
+ /data/hive-ptest/working/scratch/smart-apply-patch.sh 
/data/hive-ptest/working/scratch/build.patch
The patch does not appear to apply with p0, p1, or p2
+ exit 1
'
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12745874 - PreCommit-HIVE-TRUNK-Build

> Use HBase to cache aggregated stats
> ---
>
> Key: HIVE-11294
> URL: https://issues.apache.org/jira/browse/HIVE-11294
> Project: Hive
>  Issue Type: Improvement
>  Components: Metastore
>Affects Versions: hbase-metastore-branch
>Reporter: Alan Gates
>Assignee: Alan Gates
> Fix For: hbase-metastore-branch
>
> Attachments: HIVE-11294.patch
>
>
> Currently stats are cached only in the memory of the client.  Given that 
> HBase can easily manage the scale of caching aggregated stats we should be 
> using it to do so.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-11291) Avoid allocation storm while doing rule matching on operator/expression trees

2015-07-17 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-11291?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14631937#comment-14631937
 ] 

Hive QA commented on HIVE-11291:




{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12745848/HIVE-11291.patch

{color:red}ERROR:{color} -1 due to 2 failed/errored test(s), 9223 tests executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_cbo_rp_join0
org.apache.hive.hcatalog.hbase.TestPigHBaseStorageHandler.org.apache.hive.hcatalog.hbase.TestPigHBaseStorageHandler
{noformat}

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/4640/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/4640/console
Test logs: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-4640/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 2 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12745848 - PreCommit-HIVE-TRUNK-Build

> Avoid allocation storm while doing rule matching on operator/expression trees
> -
>
> Key: HIVE-11291
> URL: https://issues.apache.org/jira/browse/HIVE-11291
> Project: Hive
>  Issue Type: Improvement
>Reporter: Ashutosh Chauhan
>Assignee: Ashutosh Chauhan
> Attachments: HIVE-11291.patch
>
>
> RuleRegExMatch repeatedly allocates string while trying to find a matching 
> pattern. This results in huge GC churn for large trees.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-11284) Fix cbo_rp_join0 failure on master

2015-07-17 Thread Ashutosh Chauhan (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-11284?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14631933#comment-14631933
 ] 

Ashutosh Chauhan commented on HIVE-11284:
-

Cool. +1

> Fix cbo_rp_join0 failure on master
> --
>
> Key: HIVE-11284
> URL: https://issues.apache.org/jira/browse/HIVE-11284
> Project: Hive
>  Issue Type: Bug
>  Components: CBO, Query Planning
>Affects Versions: 2.0.0
>Reporter: Ashutosh Chauhan
>Assignee: Jesus Camacho Rodriguez
> Attachments: HIVE-11284.patch
>
>
> It first failed in this build : 
> http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/4621/



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-11284) Fix cbo_rp_join0 failure on master

2015-07-17 Thread Jesus Camacho Rodriguez (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-11284?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14631924#comment-14631924
 ] 

Jesus Camacho Rodriguez commented on HIVE-11284:


Sure.

==
Background:
When RT is on, HiveJoinToMultiJoinRule is responsible for creating MultiJoin 
operators out of Join operators (if possible). But the plan might be the 
following:
{noformat}
Join(...)
Project(...)
Join(...)
...
Filter(...)
...
{noformat} 
Thus, first we need to pull up the Project operator(s) through Join(s) using 
HiveJoinProjectTransposeRule, to obtain:
{noformat}
Project(...)
Join(...)
Join(...)
...
Filter(...)
...
{noformat} 
Then, HiveJoinToMultiJoinRule will find both Join operators and merge them. 
Finally, the Project operator(s) will be pushed down again throw the tree (this 
is done by HiveRelTrimmer).

==
Problem:
HIVE-11223 enabled the return path for the cbo_rp_join0.q test.
At the same time, HIVE-11252 changed the application of the rules.
- Before HIVE-11252, HiveJoinToMultiJoinRule was applied first, then 
HiveJoinProjectTransposeRule + HiveJoinToMultiJoinRule. This was merging the 
operators correctly, because no projects were found in between the Joins.
- After HIVE-11252, only HiveJoinToMultiJoinRule + HiveJoinProjectTransposeRule 
are applied (this is correct). The problem is that 1) 
HiveJoinProjectTransposeRule.LEFT and HiveJoinProjectTransposeRule.RIGHT were 
not enabled, and 2) even if they were enabled, they were not working as 
expected (more on this below). That's why the test failed.

==
Solution:
HiveJoinProjectTransposeRule.LEFT and HiveJoinProjectTransposeRule.RIGHT get 
enabled in the patch. In addition, JoinProjectTransposeRule was only working on 
inner joins, but we need it to work also on outer joins in Hive. That's why I 
had to bring so much code from the Calcite superclass 
(JoinProjectTransposeRule) to the Hive subclass (HiveJoinProjectTransposeRule) 
to solve the issue. I plan to create a Calcite patch so we can use 
JoinProjectTransposeRule with all kind of Joins. 

> Fix cbo_rp_join0 failure on master
> --
>
> Key: HIVE-11284
> URL: https://issues.apache.org/jira/browse/HIVE-11284
> Project: Hive
>  Issue Type: Bug
>  Components: CBO, Query Planning
>Affects Versions: 2.0.0
>Reporter: Ashutosh Chauhan
>Assignee: Jesus Camacho Rodriguez
> Attachments: HIVE-11284.patch
>
>
> It first failed in this build : 
> http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/4621/



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8678) Pig fails to correctly load DATE fields using HCatalog

2015-07-17 Thread Sushanth Sowmyan (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8678?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14631906#comment-14631906
 ] 

Sushanth Sowmyan commented on HIVE-8678:


Something seems weird here - looking at the code, it looks like the current 
code, where it simply casts to Date should be the right way to do this, since 
it should have called .getPrimitiveJavaObject() on the PrimitiveObjectInspector 
to get this object, and DateObjectInspector.getPrimitiveJavaObject() should 
have returned a Date. However, clearly, from your stack trace, you're getting a 
string. I'll dig into this and update as I find more.

> Pig fails to correctly load DATE fields using HCatalog
> --
>
> Key: HIVE-8678
> URL: https://issues.apache.org/jira/browse/HIVE-8678
> Project: Hive
>  Issue Type: Bug
>  Components: HCatalog
>Affects Versions: 0.13.1
>Reporter: Michael McLellan
>Assignee: Sushanth Sowmyan
>
> Using:
> Hadoop 2.5.0-cdh5.2.0 
> Pig 0.12.0-cdh5.2.0
> Hive 0.13.1-cdh5.2.0
> When using pig -useHCatalog to load a Hive table that has a DATE field, when 
> trying to DUMP the field, the following error occurs:
> {code}
> 2014-10-30 22:58:05,469 [main] ERROR 
> org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.Launcher - 
> org.apache.pig.backend.executionengine.ExecException: ERROR 6018: Error 
> converting read value to tuple
> at 
> org.apache.hive.hcatalog.pig.HCatBaseLoader.getNext(HCatBaseLoader.java:76)
> at org.apache.hive.hcatalog.pig.HCatLoader.getNext(HCatLoader.java:58)
> at 
> org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.PigRecordReader.nextKeyValue(PigRecordReader.java:211)
> at 
> org.apache.hadoop.mapred.MapTask$NewTrackingRecordReader.nextKeyValue(MapTask.java:553)
> at 
> org.apache.hadoop.mapreduce.task.MapContextImpl.nextKeyValue(MapContextImpl.java:80)
> at 
> org.apache.hadoop.mapreduce.lib.map.WrappedMapper$Context.nextKeyValue(WrappedMapper.java:91)
> at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:144)
> at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:784)
> at org.apache.hadoop.mapred.MapTask.run(MapTask.java:341)
> at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:168)
> at java.security.AccessController.doPrivileged(Native Method)
> at javax.security.auth.Subject.doAs(Subject.java:415)
> at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1614)
> at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:163)
> Caused by: java.lang.ClassCastException: java.lang.String cannot be cast to 
> java.sql.Date
> at 
> org.apache.hive.hcatalog.pig.PigHCatUtil.extractPigObject(PigHCatUtil.java:420)
> at 
> org.apache.hive.hcatalog.pig.PigHCatUtil.transformToTuple(PigHCatUtil.java:457)
> at 
> org.apache.hive.hcatalog.pig.PigHCatUtil.transformToTuple(PigHCatUtil.java:375)
> at 
> org.apache.hive.hcatalog.pig.HCatBaseLoader.getNext(HCatBaseLoader.java:64)
> 2014-10-30 22:58:05,469 [main] ERROR 
> org.apache.pig.tools.pigstats.SimplePigStats - ERROR 6018: Error converting 
> read value to tuple
> {code}
> It seems to be occuring here: 
> https://github.com/apache/hive/blob/trunk/hcatalog/hcatalog-pig-adapter/src/main/java/org/apache/hive/hcatalog/pig/PigHCatUtil.java#L433
> and that it should be:
> {code}Date d = Date.valueOf(o);{code} 
> instead of 
> {code}Date d = (Date) o;{code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Assigned] (HIVE-8678) Pig fails to correctly load DATE fields using HCatalog

2015-07-17 Thread Sushanth Sowmyan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-8678?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sushanth Sowmyan reassigned HIVE-8678:
--

Assignee: Sushanth Sowmyan

> Pig fails to correctly load DATE fields using HCatalog
> --
>
> Key: HIVE-8678
> URL: https://issues.apache.org/jira/browse/HIVE-8678
> Project: Hive
>  Issue Type: Bug
>  Components: HCatalog
>Affects Versions: 0.13.1
>Reporter: Michael McLellan
>Assignee: Sushanth Sowmyan
>
> Using:
> Hadoop 2.5.0-cdh5.2.0 
> Pig 0.12.0-cdh5.2.0
> Hive 0.13.1-cdh5.2.0
> When using pig -useHCatalog to load a Hive table that has a DATE field, when 
> trying to DUMP the field, the following error occurs:
> {code}
> 2014-10-30 22:58:05,469 [main] ERROR 
> org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.Launcher - 
> org.apache.pig.backend.executionengine.ExecException: ERROR 6018: Error 
> converting read value to tuple
> at 
> org.apache.hive.hcatalog.pig.HCatBaseLoader.getNext(HCatBaseLoader.java:76)
> at org.apache.hive.hcatalog.pig.HCatLoader.getNext(HCatLoader.java:58)
> at 
> org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.PigRecordReader.nextKeyValue(PigRecordReader.java:211)
> at 
> org.apache.hadoop.mapred.MapTask$NewTrackingRecordReader.nextKeyValue(MapTask.java:553)
> at 
> org.apache.hadoop.mapreduce.task.MapContextImpl.nextKeyValue(MapContextImpl.java:80)
> at 
> org.apache.hadoop.mapreduce.lib.map.WrappedMapper$Context.nextKeyValue(WrappedMapper.java:91)
> at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:144)
> at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:784)
> at org.apache.hadoop.mapred.MapTask.run(MapTask.java:341)
> at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:168)
> at java.security.AccessController.doPrivileged(Native Method)
> at javax.security.auth.Subject.doAs(Subject.java:415)
> at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1614)
> at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:163)
> Caused by: java.lang.ClassCastException: java.lang.String cannot be cast to 
> java.sql.Date
> at 
> org.apache.hive.hcatalog.pig.PigHCatUtil.extractPigObject(PigHCatUtil.java:420)
> at 
> org.apache.hive.hcatalog.pig.PigHCatUtil.transformToTuple(PigHCatUtil.java:457)
> at 
> org.apache.hive.hcatalog.pig.PigHCatUtil.transformToTuple(PigHCatUtil.java:375)
> at 
> org.apache.hive.hcatalog.pig.HCatBaseLoader.getNext(HCatBaseLoader.java:64)
> 2014-10-30 22:58:05,469 [main] ERROR 
> org.apache.pig.tools.pigstats.SimplePigStats - ERROR 6018: Error converting 
> read value to tuple
> {code}
> It seems to be occuring here: 
> https://github.com/apache/hive/blob/trunk/hcatalog/hcatalog-pig-adapter/src/main/java/org/apache/hive/hcatalog/pig/PigHCatUtil.java#L433
> and that it should be:
> {code}Date d = Date.valueOf(o);{code} 
> instead of 
> {code}Date d = (Date) o;{code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-11077) Add support in parser and wire up to txn manager

2015-07-17 Thread Eugene Koifman (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-11077?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14631904#comment-14631904
 ] 

Eugene Koifman commented on HIVE-11077:
---

This is not quite ready for review.  Hopefully by Monday.

> Add support in parser and wire up to txn manager
> 
>
> Key: HIVE-11077
> URL: https://issues.apache.org/jira/browse/HIVE-11077
> Project: Hive
>  Issue Type: Sub-task
>  Components: SQL, Transactions
>Affects Versions: 1.2.0
>Reporter: Eugene Koifman
>Assignee: Eugene Koifman
> Attachments: HIVE-11077.3.patch, HIVE-11077.5.patch, HIVE-11077.patch
>
>




--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-11295) LLAP: clean up ORC dependencies on object pools

2015-07-17 Thread Sergey Shelukhin (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-11295?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sergey Shelukhin updated HIVE-11295:

Attachment: HIVE-11259.patch

this patch depends on the other patch.
[~prasanth_j] can you review?

> LLAP: clean up ORC dependencies on object pools
> ---
>
> Key: HIVE-11295
> URL: https://issues.apache.org/jira/browse/HIVE-11295
> Project: Hive
>  Issue Type: Sub-task
>Reporter: Sergey Shelukhin
>Assignee: Sergey Shelukhin
> Attachments: HIVE-11259.patch
>
>
> Before there's storage handler module, we can clean some things up



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-11281) Memoize toStringTree() for ASTNode

2015-07-17 Thread Gunther Hagleitner (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-11281?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14631874#comment-14631874
 ] 

Gunther Hagleitner commented on HIVE-11281:
---

How many places do we have that transform nodes? Can we make the nodes 
immutable?

> Memoize toStringTree() for ASTNode
> --
>
> Key: HIVE-11281
> URL: https://issues.apache.org/jira/browse/HIVE-11281
> Project: Hive
>  Issue Type: Bug
>  Components: Query Planning
>Affects Versions: 1.0.0, 1.2.0, 1.1.0
>Reporter: Ashutosh Chauhan
>Assignee: Ashutosh Chauhan
> Attachments: HIVE-11281.2.patch, HIVE-11281.3.patch, HIVE-11281.patch
>
>
> Currently it defaults to BaseTree::toStringTree() which always computes it 
> recursively. For very large trees it becomes expensive. On one particular 
> test query compilation time went down from 55 seconds to 2 seconds.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-11284) Fix cbo_rp_join0 failure on master

2015-07-17 Thread Ashutosh Chauhan (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-11284?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14631813#comment-14631813
 ] 

Ashutosh Chauhan commented on HIVE-11284:
-

Can you briefly explain what was the issue and how is patch fixing it?

> Fix cbo_rp_join0 failure on master
> --
>
> Key: HIVE-11284
> URL: https://issues.apache.org/jira/browse/HIVE-11284
> Project: Hive
>  Issue Type: Bug
>  Components: CBO, Query Planning
>Affects Versions: 2.0.0
>Reporter: Ashutosh Chauhan
>Assignee: Jesus Camacho Rodriguez
> Attachments: HIVE-11284.patch
>
>
> It first failed in this build : 
> http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/4621/



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-11077) Add support in parser and wire up to txn manager

2015-07-17 Thread Alan Gates (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-11077?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14631809#comment-14631809
 ] 

Alan Gates commented on HIVE-11077:
---

This is a big patch.  Could you create a review board entry?

> Add support in parser and wire up to txn manager
> 
>
> Key: HIVE-11077
> URL: https://issues.apache.org/jira/browse/HIVE-11077
> Project: Hive
>  Issue Type: Sub-task
>  Components: SQL, Transactions
>Affects Versions: 1.2.0
>Reporter: Eugene Koifman
>Assignee: Eugene Koifman
> Attachments: HIVE-11077.3.patch, HIVE-11077.5.patch, HIVE-11077.patch
>
>




--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-11210) Remove dependency on HiveConf from Orc reader & writer

2015-07-17 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-11210?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14631808#comment-14631808
 ] 

Hive QA commented on HIVE-11210:




{color:red}Overall{color}: -1 no tests executed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12745845/HIVE-11210.patch

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/4639/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/4639/console
Test logs: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-4639/

Messages:
{noformat}
 This message was trimmed, see log for full details 
[INFO] 
[INFO] --- maven-resources-plugin:2.6:testResources (default-testResources) @ 
spark-client ---
[INFO] Using 'UTF-8' encoding to copy filtered resources.
[INFO] Copying 1 resource
[INFO] Copying 3 resources
[INFO] 
[INFO] --- maven-antrun-plugin:1.7:run (setup-test-dirs) @ spark-client ---
[INFO] Executing tasks

main:
[mkdir] Created dir: 
/data/hive-ptest/working/apache-github-source-source/spark-client/target/tmp
[mkdir] Created dir: 
/data/hive-ptest/working/apache-github-source-source/spark-client/target/warehouse
[mkdir] Created dir: 
/data/hive-ptest/working/apache-github-source-source/spark-client/target/tmp/conf
 [copy] Copying 11 files to 
/data/hive-ptest/working/apache-github-source-source/spark-client/target/tmp/conf
[INFO] Executed tasks
[INFO] 
[INFO] --- maven-compiler-plugin:3.1:testCompile (default-testCompile) @ 
spark-client ---
[INFO] Compiling 5 source files to 
/data/hive-ptest/working/apache-github-source-source/spark-client/target/test-classes
[INFO] 
[INFO] --- maven-dependency-plugin:2.8:copy (copy-guava-14) @ spark-client ---
[INFO] Configured Artifact: com.google.guava:guava:14.0.1:jar
[INFO] Copying guava-14.0.1.jar to 
/data/hive-ptest/working/apache-github-source-source/spark-client/target/dependency/guava-14.0.1.jar
[INFO] 
[INFO] --- maven-surefire-plugin:2.16:test (default-test) @ spark-client ---
[INFO] Tests are skipped.
[INFO] 
[INFO] --- maven-jar-plugin:2.2:jar (default-jar) @ spark-client ---
[INFO] Building jar: 
/data/hive-ptest/working/apache-github-source-source/spark-client/target/spark-client-2.0.0-SNAPSHOT.jar
[INFO] 
[INFO] --- maven-site-plugin:3.3:attach-descriptor (attach-descriptor) @ 
spark-client ---
[INFO] 
[INFO] --- maven-install-plugin:2.4:install (default-install) @ spark-client ---
[INFO] Installing 
/data/hive-ptest/working/apache-github-source-source/spark-client/target/spark-client-2.0.0-SNAPSHOT.jar
 to 
/home/hiveptest/.m2/repository/org/apache/hive/spark-client/2.0.0-SNAPSHOT/spark-client-2.0.0-SNAPSHOT.jar
[INFO] Installing 
/data/hive-ptest/working/apache-github-source-source/spark-client/pom.xml to 
/home/hiveptest/.m2/repository/org/apache/hive/spark-client/2.0.0-SNAPSHOT/spark-client-2.0.0-SNAPSHOT.pom
[INFO] 
[INFO] 
[INFO] Building Hive Query Language 2.0.0-SNAPSHOT
[INFO] 
[INFO] 
[INFO] --- maven-clean-plugin:2.5:clean (default-clean) @ hive-exec ---
[INFO] Deleting /data/hive-ptest/working/apache-github-source-source/ql/target
[INFO] Deleting /data/hive-ptest/working/apache-github-source-source/ql 
(includes = [datanucleus.log, derby.log], excludes = [])
[INFO] 
[INFO] --- maven-enforcer-plugin:1.3.1:enforce (enforce-no-snapshots) @ 
hive-exec ---
[INFO] 
[INFO] --- maven-antrun-plugin:1.7:run (generate-sources) @ hive-exec ---
[INFO] Executing tasks

main:
[mkdir] Created dir: 
/data/hive-ptest/working/apache-github-source-source/ql/target/generated-sources/java/org/apache/hadoop/hive/ql/exec/vector/expressions/gen
[mkdir] Created dir: 
/data/hive-ptest/working/apache-github-source-source/ql/target/generated-sources/java/org/apache/hadoop/hive/ql/exec/vector/expressions/aggregates/gen
[mkdir] Created dir: 
/data/hive-ptest/working/apache-github-source-source/ql/target/generated-test-sources/java/org/apache/hadoop/hive/ql/exec/vector/expressions/gen
Generating vector expression code
Generating vector expression test code
[INFO] Executed tasks
[INFO] 
[INFO] --- build-helper-maven-plugin:1.8:add-source (add-source) @ hive-exec ---
[INFO] Source directory: 
/data/hive-ptest/working/apache-github-source-source/ql/src/gen/protobuf/gen-java
 added.
[INFO] Source directory: 
/data/hive-ptest/working/apache-github-source-source/ql/src/gen/thrift/gen-javabean
 added.
[INFO] Source directory: 
/data/hive-ptest/working/apache-github-source-source/ql/target/generated-sources/java
 added.
[INFO] 
[INFO] --- antlr3-maven-plugin:3.4:antlr (default) @ hive-exec ---
[IN

[jira] [Commented] (HIVE-11284) Fix cbo_rp_join0 failure on master

2015-07-17 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-11284?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14631802#comment-14631802
 ] 

Hive QA commented on HIVE-11284:




{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12745846/HIVE-11284.patch

{color:red}ERROR:{color} -1 due to 1 failed/errored test(s), 9225 tests executed
*Failed tests:*
{noformat}
org.apache.hive.hcatalog.streaming.TestStreaming.testTransactionBatchCommit_Json
{noformat}

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/4638/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/4638/console
Test logs: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-4638/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 1 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12745846 - PreCommit-HIVE-TRUNK-Build

> Fix cbo_rp_join0 failure on master
> --
>
> Key: HIVE-11284
> URL: https://issues.apache.org/jira/browse/HIVE-11284
> Project: Hive
>  Issue Type: Bug
>  Components: CBO, Query Planning
>Affects Versions: 2.0.0
>Reporter: Ashutosh Chauhan
>Assignee: Jesus Camacho Rodriguez
> Attachments: HIVE-11284.patch
>
>
> It first failed in this build : 
> http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/4621/



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-11288) Avro SerDe InstanceCache returns incorrect schema

2015-07-17 Thread Greg Phillips (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-11288?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Greg Phillips updated HIVE-11288:
-
Attachment: HIVE-11288.2.patch

Implemented hashCode and equals methods for test class 'Foo' to allow it to 
work properly within a HashMap.

> Avro SerDe InstanceCache returns incorrect schema
> -
>
> Key: HIVE-11288
> URL: https://issues.apache.org/jira/browse/HIVE-11288
> Project: Hive
>  Issue Type: Bug
>Reporter: Greg Phillips
> Attachments: HIVE-11288.2.patch, HIVE-11288.patch
>
>
> To reproduce this error, take two fields in an avro schema document matching 
> the following:
> "type" :  { "type": "array", "items": [ "null",  { "type": "map", "values": [ 
> "null", "string" ] } ]  }
> "type" : { "type": "map", "values": [ "null" , { "type": "array", "items": [ 
> "null" , "string"] } ] }
> After creating two tables in hive with these schemas, the describe statement 
> on each of them will only return the schema for the first one loaded.  This 
> is due to a hashCode() collision in the InstanceCache.  
> A patch will be included in this ticket shortly which removes the hashCode 
> call from the InstanceCache's internal HashMap, and instead provides the 
> entire schema object.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-11294) Use HBase to cache aggregated stats

2015-07-17 Thread Alan Gates (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-11294?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alan Gates updated HIVE-11294:
--
Attachment: HIVE-11294.patch

This patch adds caching of the aggregates stats to HBase.  It also 
fundamentally changes how cached entries are matched.  Now only exact matches 
are taken, rather than partial matches as was done in the past.  The key for 
entries in the cache is an md5 sum of the dbname, tablename, and sorted list of 
partition names.  This allows for reasonable key sizes and fast lookup.

A limited number of entries are still kept in memory (10K by default) for a 
limited time (1 min by default).  This is to reduce back and forth to HBase.

Entries in HBase are kept in the cache for 1 week or until a partition's stats 
are updated or the partition is dropped.  Determining when an aggregate needs 
to be dropped is not straight forward.  Since the key is an md5 sum we cannot 
determine from the key if an entry contains the partition that was updated or 
dropped.  To deal with this each entry also contains a bloom filter of all the 
partition names.  When a partition is updated or dropped it is added a queue.  
Every 5 seconds a separate thread takes all of the entries from the queue and 
does a full scan of the cache.  It uses the bloom filters to determine if any 
of the entries in the queue match one of the partitions in the aggregate.  If 
so, it drops the aggregate entry.  Given that this is done by a bloom filter 
there will be some false positives (entries that get dropped that shouldn't) 
but the error rate was chosen to be very low (0.1%).  This makes the bloom 
filter larger but the motivation in choosing the bloom filter was to minimize 
processing time rather than to save space.

All of this means there will be lag between when a partition is dropped or 
updated and when the aggregate is dropped.  It will be < 5 seconds if the drop 
was done on the same HS2 instance, or <65 seconds if done on another instance.  
Given that these are statistics I think that's acceptable.

Ideally we would not drop an aggregate as soon as a single partition is dropped 
or updated.  Instead we should be tracking the number of invalidated partitions 
and only drop the aggregate once it reaches a threshold like 5%.  Doing this 
would require implementing the invalidation logic as a co-processor rather than 
as a filter, which is why I didn't do it this way to begin with.

> Use HBase to cache aggregated stats
> ---
>
> Key: HIVE-11294
> URL: https://issues.apache.org/jira/browse/HIVE-11294
> Project: Hive
>  Issue Type: Improvement
>  Components: Metastore
>Affects Versions: hbase-metastore-branch
>Reporter: Alan Gates
>Assignee: Alan Gates
> Attachments: HIVE-11294.patch
>
>
> Currently stats are cached only in the memory of the client.  Given that 
> HBase can easily manage the scale of caching aggregated stats we should be 
> using it to do so.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-11284) Fix cbo_rp_join0 failure on master

2015-07-17 Thread Jesus Camacho Rodriguez (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-11284?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14631778#comment-14631778
 ] 

Jesus Camacho Rodriguez commented on HIVE-11284:


[~ashutoshc], actually it was the interaction of HIVE-11252 and HIVE-11223, and 
I didn't realize as the QA run on both at the same time. This patch solves the 
issue.

> Fix cbo_rp_join0 failure on master
> --
>
> Key: HIVE-11284
> URL: https://issues.apache.org/jira/browse/HIVE-11284
> Project: Hive
>  Issue Type: Bug
>  Components: CBO, Query Planning
>Affects Versions: 2.0.0
>Reporter: Ashutosh Chauhan
>Assignee: Jesus Camacho Rodriguez
> Attachments: HIVE-11284.patch
>
>
> It first failed in this build : 
> http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/4621/



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-11233) Support of binary sortable key and values for HBaseStorageHandler

2015-07-17 Thread Svetozar Ivanov (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-11233?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Svetozar Ivanov updated HIVE-11233:
---
Attachment: HIVE-11233-branch-1.2.patch

> Support of binary sortable key and values for HBaseStorageHandler
> -
>
> Key: HIVE-11233
> URL: https://issues.apache.org/jira/browse/HIVE-11233
> Project: Hive
>  Issue Type: New Feature
>  Components: HBase Handler
>Affects Versions: 1.2.1
>Reporter: Svetozar Ivanov
>Assignee: Svetozar Ivanov
>  Labels: Binary, Hbase, Numeric, Phoenix, Sortable
> Attachments: HIVE-11233-branch-1.2.patch, HIVE-11233.1-1.2.patch, 
> HIVE-11233.1.patch, HIVE-11233.patch
>
>
> Currently HBaseStorageHandler doesn't provide mechanism for storage of binary 
> sortable key and values. It is necessary when given HBase table is used for 
> persistence by Apache Hive and Apache Phoenix. In that way all byte arrays 
> read or written by Hive will be compatible with binary sortable format used 
> by Phoenix.
> It turns out the major difference is in all numeric data types accordingly 
> officially provided documentation - 
> https://phoenix.apache.org/language/datatypes.html.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-11233) Support of binary sortable key and values for HBaseStorageHandler

2015-07-17 Thread Svetozar Ivanov (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-11233?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Svetozar Ivanov updated HIVE-11233:
---
Attachment: HIVE-11233.1.patch

> Support of binary sortable key and values for HBaseStorageHandler
> -
>
> Key: HIVE-11233
> URL: https://issues.apache.org/jira/browse/HIVE-11233
> Project: Hive
>  Issue Type: New Feature
>  Components: HBase Handler
>Affects Versions: 1.2.1
>Reporter: Svetozar Ivanov
>Assignee: Svetozar Ivanov
>  Labels: Binary, Hbase, Numeric, Phoenix, Sortable
> Attachments: HIVE-11233.1-1.2.patch, HIVE-11233.1.patch, 
> HIVE-11233.patch
>
>
> Currently HBaseStorageHandler doesn't provide mechanism for storage of binary 
> sortable key and values. It is necessary when given HBase table is used for 
> persistence by Apache Hive and Apache Phoenix. In that way all byte arrays 
> read or written by Hive will be compatible with binary sortable format used 
> by Phoenix.
> It turns out the major difference is in all numeric data types accordingly 
> officially provided documentation - 
> https://phoenix.apache.org/language/datatypes.html.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-11233) Support of binary sortable key and values for HBaseStorageHandler

2015-07-17 Thread Svetozar Ivanov (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-11233?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Svetozar Ivanov updated HIVE-11233:
---
Attachment: HIVE-11233.1-1.2.patch

> Support of binary sortable key and values for HBaseStorageHandler
> -
>
> Key: HIVE-11233
> URL: https://issues.apache.org/jira/browse/HIVE-11233
> Project: Hive
>  Issue Type: New Feature
>  Components: HBase Handler
>Affects Versions: 1.2.1
>Reporter: Svetozar Ivanov
>Assignee: Svetozar Ivanov
>  Labels: Binary, Hbase, Numeric, Phoenix, Sortable
> Attachments: HIVE-11233.1-1.2.patch, HIVE-11233.patch
>
>
> Currently HBaseStorageHandler doesn't provide mechanism for storage of binary 
> sortable key and values. It is necessary when given HBase table is used for 
> persistence by Apache Hive and Apache Phoenix. In that way all byte arrays 
> read or written by Hive will be compatible with binary sortable format used 
> by Phoenix.
> It turns out the major difference is in all numeric data types accordingly 
> officially provided documentation - 
> https://phoenix.apache.org/language/datatypes.html.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-11288) Avro SerDe InstanceCache returns incorrect schema

2015-07-17 Thread Greg Phillips (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-11288?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14631774#comment-14631774
 ] 

Greg Phillips commented on HIVE-11288:
--

The test is wrong.  It is relying on the java Object.equals() which requires 
the exact same object to return true.
This is similar to the original problem with the InstanceCache.  It is relying 
on the hashCode to determine equality which breaks the hashCode contract.  I 
will provide the patch for the unit test shortly.

> Avro SerDe InstanceCache returns incorrect schema
> -
>
> Key: HIVE-11288
> URL: https://issues.apache.org/jira/browse/HIVE-11288
> Project: Hive
>  Issue Type: Bug
>Reporter: Greg Phillips
> Attachments: HIVE-11288.patch
>
>
> To reproduce this error, take two fields in an avro schema document matching 
> the following:
> "type" :  { "type": "array", "items": [ "null",  { "type": "map", "values": [ 
> "null", "string" ] } ]  }
> "type" : { "type": "map", "values": [ "null" , { "type": "array", "items": [ 
> "null" , "string"] } ] }
> After creating two tables in hive with these schemas, the describe statement 
> on each of them will only return the schema for the first one loaded.  This 
> is due to a hashCode() collision in the InstanceCache.  
> A patch will be included in this ticket shortly which removes the hashCode 
> call from the InstanceCache's internal HashMap, and instead provides the 
> entire schema object.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-11233) Support of binary sortable key and values for HBaseStorageHandler

2015-07-17 Thread Svetozar Ivanov (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-11233?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Svetozar Ivanov updated HIVE-11233:
---
Attachment: (was: HIVE-11233.1-branch-1.2.patch)

> Support of binary sortable key and values for HBaseStorageHandler
> -
>
> Key: HIVE-11233
> URL: https://issues.apache.org/jira/browse/HIVE-11233
> Project: Hive
>  Issue Type: New Feature
>  Components: HBase Handler
>Affects Versions: 1.2.1
>Reporter: Svetozar Ivanov
>Assignee: Svetozar Ivanov
>  Labels: Binary, Hbase, Numeric, Phoenix, Sortable
> Attachments: HIVE-11233.patch
>
>
> Currently HBaseStorageHandler doesn't provide mechanism for storage of binary 
> sortable key and values. It is necessary when given HBase table is used for 
> persistence by Apache Hive and Apache Phoenix. In that way all byte arrays 
> read or written by Hive will be compatible with binary sortable format used 
> by Phoenix.
> It turns out the major difference is in all numeric data types accordingly 
> officially provided documentation - 
> https://phoenix.apache.org/language/datatypes.html.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-11233) Support of binary sortable key and values for HBaseStorageHandler

2015-07-17 Thread Svetozar Ivanov (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-11233?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Svetozar Ivanov updated HIVE-11233:
---
Attachment: (was: HIVE-11233.2-branch-1.2.patch)

> Support of binary sortable key and values for HBaseStorageHandler
> -
>
> Key: HIVE-11233
> URL: https://issues.apache.org/jira/browse/HIVE-11233
> Project: Hive
>  Issue Type: New Feature
>  Components: HBase Handler
>Affects Versions: 1.2.1
>Reporter: Svetozar Ivanov
>Assignee: Svetozar Ivanov
>  Labels: Binary, Hbase, Numeric, Phoenix, Sortable
> Attachments: HIVE-11233.patch
>
>
> Currently HBaseStorageHandler doesn't provide mechanism for storage of binary 
> sortable key and values. It is necessary when given HBase table is used for 
> persistence by Apache Hive and Apache Phoenix. In that way all byte arrays 
> read or written by Hive will be compatible with binary sortable format used 
> by Phoenix.
> It turns out the major difference is in all numeric data types accordingly 
> officially provided documentation - 
> https://phoenix.apache.org/language/datatypes.html.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-11282) CBO (Calcite Return Path): Inferring Hive type char/varchar of length zero which is not allowed

2015-07-17 Thread Jesus Camacho Rodriguez (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-11282?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14631755#comment-14631755
 ] 

Jesus Camacho Rodriguez commented on HIVE-11282:


Fail is unrelated (HIVE-11284).

> CBO (Calcite Return Path): Inferring Hive type char/varchar of length zero 
> which is not allowed
> ---
>
> Key: HIVE-11282
> URL: https://issues.apache.org/jira/browse/HIVE-11282
> Project: Hive
>  Issue Type: Sub-task
>  Components: CBO
>Reporter: Jesus Camacho Rodriguez
>Assignee: Jesus Camacho Rodriguez
> Fix For: 2.0.0
>
> Attachments: HIVE-11282.patch
>
>
> When RT is on, we try to infer the Hive type from the Calcite type for the 
> value '’ e.g. in udf3.q, and we end up with char (length=0) as a result. The 
> min length of char/varchar in Hive is 1, thus an Exception is thrown.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-11269) Intermittent bug with kryo serializers

2015-07-17 Thread Soundararajan Velu (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-11269?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14631680#comment-14631680
 ] 

Soundararajan Velu commented on HIVE-11269:
---

Cool the later two issues were due to non serializable fields in the UDF and 
UDAF, made them transient and it seems to proceed.

the first problem still remains, I will provide you the sample table and data 
to reproduce this problem,


> Intermittent bug with kryo serializers
> --
>
> Key: HIVE-11269
> URL: https://issues.apache.org/jira/browse/HIVE-11269
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 1.3.0
>Reporter: Soundararajan Velu
>
> getting kryo serialization exception when running large queries, this works 
> fine on hive 0.14, 
> and works fine on 1.3.0 with the following flags,
> set hive.plan.serialization.format=kryo;
> set hive.exec.parallel=false;
> set hive.limit.optimize.enable=false;
> set hive.optimize.metadataonly=false;
> set hive.optimize.reducededuplication=false;
> set hive.optimize.sort.dynamic.partition=false;
> set hive.stats.fetch.partition.stats=false;
> set hive.vectorized.execution.enabled=false;
> set hive.vectorized.execution.reduce.enabled=false;
> set hive.cbo.enable=false;
> set hive.compute.query.using.stats=false;
> set hive.multigroupby.singlereducer=false;
> set hive.optimize.ppd=false;
> set hive.optimize.skewjoin.compiletime=false;
> set hive.optimize.skewjoin=false;
> set hive.optimize.union.remove=false;
> set hive.mapred.mode=nonstrict;
> set hive.auto.convert.join.noconditionaltask=false;
> set hive.optimize.sort.dynamic.partition=false;
> set hive.rpc.query.plan=true;
> StackTrace
>   colExprMap (org.apache.hadoop.hive.ql.exec.SelectOperator)
>   childOperators (org.apache.hadoop.hive.ql.exec.JoinOperator)
>   reducer (org.apache.hadoop.hive.ql.plan.ReduceWork)
>   at 
> org.apache.hadoop.hive.ql.exec.Utilities.getBaseWork(Utilities.java:462)
>   at 
> org.apache.hadoop.hive.ql.exec.Utilities.getReduceWork(Utilities.java:309)
>   at 
> org.apache.hadoop.hive.ql.exec.mr.ExecReducer.configure(ExecReducer.java:112)
>   ... 14 more
>   Caused by: org.apache.hive.com.esotericsoftware.kryo.KryoException: 
> java.lang.NullPointerException
>   Serialization trace:
>   chidren (org.apache.hadoop.hive.ql.plan.ExprNodeGenericFuncDesc)
>   chidren (org.apache.hadoop.hive.ql.plan.ExprNodeGenericFuncDesc)
>   chidren (org.apache.hadoop.hive.ql.plan.ExprNodeGenericFuncDesc)
>   chidren (org.apache.hadoop.hive.ql.plan.ExprNodeGenericFuncDesc)
>   chidren (org.apache.hadoop.hive.ql.plan.ExprNodeGenericFuncDesc)
>   chidren (org.apache.hadoop.hive.ql.plan.ExprNodeGenericFuncDesc)
>   chidren (org.apache.hadoop.hive.ql.plan.ExprNodeGenericFuncDesc)
>   chidren (org.apache.hadoop.hive.ql.plan.ExprNodeGenericFuncDesc)
>   chidren (org.apache.hadoop.hive.ql.plan.ExprNodeGenericFuncDesc)
>   chidren (org.apache.hadoop.hive.ql.plan.ExprNodeGenericFuncDesc)
>   chidren (org.apache.hadoop.hive.ql.plan.ExprNodeGenericFuncDesc)
>   chidren (org.apache.hadoop.hive.ql.plan.ExprNodeGenericFuncDesc)
>   chidren (org.apache.hadoop.hive.ql.plan.ExprNodeGenericFuncDesc)
>   colExprMap (org.apache.hadoop.hive.ql.exec.SelectOperator)
>   childOperators (org.apache.hadoop.hive.ql.exec.JoinOperator)
>   reducer (org.apache.hadoop.hive.ql.plan.ReduceWork)
>   at 
> org.apache.hive.com.esotericsoftware.kryo.serializers.ObjectField.read(ObjectField.java:125)
>   at 
> org.apache.hive.com.esotericsoftware.kryo.serializers.FieldSerializer.read(FieldSerializer.java:507)
>   at 
> org.apache.hive.com.esotericsoftware.kryo.Kryo.readClassAndObject(Kryo.java:776)
>   at 
> org.apache.hive.com.esotericsoftware.kryo.serializers.CollectionSerializer.read(CollectionSerializer.java:112)
>   at 
> org.apache.hive.com.esotericsoftware.kryo.serializers.CollectionSerializer.read(CollectionSerializer.java:18)
>   at 
> org.apache.hive.com.esotericsoftware.kryo.Kryo.readObject(Kryo.java:694)
>   at 
> org.apache.hive.com.esotericsoftware.kryo.serializers.ObjectField.read(ObjectField.java:106)
>   at 
> org.apache.hive.com.esotericsoftware.kryo.serializers.FieldSerializer.read(FieldSerializer.java:507)
>   at 
> org.apache.hive.com.esotericsoftware.kryo.Kryo.readClassAndObject(Kryo.java:776)
>   at 
> org.apache.hive.com.esotericsoftware.kryo.serializers.CollectionSerializer.read(CollectionSerializer.java:112)
>   at 
> org.apache.hive.com.esotericsoftware.kryo.serializers.CollectionSerializer.read(CollectionSerializer.java:18)
>   at 
> org.apache.hive.com.esotericsoftware.kryo.Kryo.readObject(Kryo.java:694)
>   at 
> org.apache.hive.com.esotericsoftware.kryo.serializers.ObjectField.read(ObjectField.java:106)
>   at 
> org.apache.hive.com.esotericsoftware.kryo.serializers.FieldS

[jira] [Commented] (HIVE-11291) Avoid allocation storm while doing rule matching on operator/expression trees

2015-07-17 Thread Hari Sankar Sivarama Subramaniyan (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-11291?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14631675#comment-14631675
 ] 

Hari Sankar Sivarama Subramaniyan commented on HIVE-11291:
--

[~ashutoshc] The patch looks good to me, +1 pending tests. I believe this idea 
has to be incorporated as part of HIVE-11141  

Thanks
Hari

> Avoid allocation storm while doing rule matching on operator/expression trees
> -
>
> Key: HIVE-11291
> URL: https://issues.apache.org/jira/browse/HIVE-11291
> Project: Hive
>  Issue Type: Improvement
>Reporter: Ashutosh Chauhan
>Assignee: Ashutosh Chauhan
> Attachments: HIVE-11291.patch
>
>
> RuleRegExMatch repeatedly allocates string while trying to find a matching 
> pattern. This results in huge GC churn for large trees.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-11291) Avoid allocation storm while doing rule matching on operator/expression trees

2015-07-17 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-11291?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-11291:

Attachment: HIVE-11291.patch

In most cases, we are trying to do exact matches, regex matching is not 
required.
[~hsubramaniyan] Can you take a look?

> Avoid allocation storm while doing rule matching on operator/expression trees
> -
>
> Key: HIVE-11291
> URL: https://issues.apache.org/jira/browse/HIVE-11291
> Project: Hive
>  Issue Type: Improvement
>Reporter: Ashutosh Chauhan
>Assignee: Ashutosh Chauhan
> Attachments: HIVE-11291.patch
>
>
> RuleRegExMatch repeatedly allocates string while trying to find a matching 
> pattern. This results in huge GC churn for large trees.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-11284) Fix cbo_rp_join0 failure on master

2015-07-17 Thread Jesus Camacho Rodriguez (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-11284?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jesus Camacho Rodriguez updated HIVE-11284:
---
Attachment: HIVE-11284.patch

> Fix cbo_rp_join0 failure on master
> --
>
> Key: HIVE-11284
> URL: https://issues.apache.org/jira/browse/HIVE-11284
> Project: Hive
>  Issue Type: Bug
>  Components: CBO, Query Planning
>Affects Versions: 2.0.0
>Reporter: Ashutosh Chauhan
>Assignee: Jesus Camacho Rodriguez
> Attachments: HIVE-11284.patch
>
>
> It first failed in this build : 
> http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/4621/



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-11269) Intermittent bug with kryo serializers

2015-07-17 Thread Xuefu Zhang (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-11269?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14631644#comment-14631644
 ] 

Xuefu Zhang commented on HIVE-11269:


You seemed hitting a new problem. In either case it's best to provide a repro 
case. Otherwise, it's hard to find the problem and so the fix.

> Intermittent bug with kryo serializers
> --
>
> Key: HIVE-11269
> URL: https://issues.apache.org/jira/browse/HIVE-11269
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 1.3.0
>Reporter: Soundararajan Velu
>
> getting kryo serialization exception when running large queries, this works 
> fine on hive 0.14, 
> and works fine on 1.3.0 with the following flags,
> set hive.plan.serialization.format=kryo;
> set hive.exec.parallel=false;
> set hive.limit.optimize.enable=false;
> set hive.optimize.metadataonly=false;
> set hive.optimize.reducededuplication=false;
> set hive.optimize.sort.dynamic.partition=false;
> set hive.stats.fetch.partition.stats=false;
> set hive.vectorized.execution.enabled=false;
> set hive.vectorized.execution.reduce.enabled=false;
> set hive.cbo.enable=false;
> set hive.compute.query.using.stats=false;
> set hive.multigroupby.singlereducer=false;
> set hive.optimize.ppd=false;
> set hive.optimize.skewjoin.compiletime=false;
> set hive.optimize.skewjoin=false;
> set hive.optimize.union.remove=false;
> set hive.mapred.mode=nonstrict;
> set hive.auto.convert.join.noconditionaltask=false;
> set hive.optimize.sort.dynamic.partition=false;
> set hive.rpc.query.plan=true;
> StackTrace
>   colExprMap (org.apache.hadoop.hive.ql.exec.SelectOperator)
>   childOperators (org.apache.hadoop.hive.ql.exec.JoinOperator)
>   reducer (org.apache.hadoop.hive.ql.plan.ReduceWork)
>   at 
> org.apache.hadoop.hive.ql.exec.Utilities.getBaseWork(Utilities.java:462)
>   at 
> org.apache.hadoop.hive.ql.exec.Utilities.getReduceWork(Utilities.java:309)
>   at 
> org.apache.hadoop.hive.ql.exec.mr.ExecReducer.configure(ExecReducer.java:112)
>   ... 14 more
>   Caused by: org.apache.hive.com.esotericsoftware.kryo.KryoException: 
> java.lang.NullPointerException
>   Serialization trace:
>   chidren (org.apache.hadoop.hive.ql.plan.ExprNodeGenericFuncDesc)
>   chidren (org.apache.hadoop.hive.ql.plan.ExprNodeGenericFuncDesc)
>   chidren (org.apache.hadoop.hive.ql.plan.ExprNodeGenericFuncDesc)
>   chidren (org.apache.hadoop.hive.ql.plan.ExprNodeGenericFuncDesc)
>   chidren (org.apache.hadoop.hive.ql.plan.ExprNodeGenericFuncDesc)
>   chidren (org.apache.hadoop.hive.ql.plan.ExprNodeGenericFuncDesc)
>   chidren (org.apache.hadoop.hive.ql.plan.ExprNodeGenericFuncDesc)
>   chidren (org.apache.hadoop.hive.ql.plan.ExprNodeGenericFuncDesc)
>   chidren (org.apache.hadoop.hive.ql.plan.ExprNodeGenericFuncDesc)
>   chidren (org.apache.hadoop.hive.ql.plan.ExprNodeGenericFuncDesc)
>   chidren (org.apache.hadoop.hive.ql.plan.ExprNodeGenericFuncDesc)
>   chidren (org.apache.hadoop.hive.ql.plan.ExprNodeGenericFuncDesc)
>   chidren (org.apache.hadoop.hive.ql.plan.ExprNodeGenericFuncDesc)
>   colExprMap (org.apache.hadoop.hive.ql.exec.SelectOperator)
>   childOperators (org.apache.hadoop.hive.ql.exec.JoinOperator)
>   reducer (org.apache.hadoop.hive.ql.plan.ReduceWork)
>   at 
> org.apache.hive.com.esotericsoftware.kryo.serializers.ObjectField.read(ObjectField.java:125)
>   at 
> org.apache.hive.com.esotericsoftware.kryo.serializers.FieldSerializer.read(FieldSerializer.java:507)
>   at 
> org.apache.hive.com.esotericsoftware.kryo.Kryo.readClassAndObject(Kryo.java:776)
>   at 
> org.apache.hive.com.esotericsoftware.kryo.serializers.CollectionSerializer.read(CollectionSerializer.java:112)
>   at 
> org.apache.hive.com.esotericsoftware.kryo.serializers.CollectionSerializer.read(CollectionSerializer.java:18)
>   at 
> org.apache.hive.com.esotericsoftware.kryo.Kryo.readObject(Kryo.java:694)
>   at 
> org.apache.hive.com.esotericsoftware.kryo.serializers.ObjectField.read(ObjectField.java:106)
>   at 
> org.apache.hive.com.esotericsoftware.kryo.serializers.FieldSerializer.read(FieldSerializer.java:507)
>   at 
> org.apache.hive.com.esotericsoftware.kryo.Kryo.readClassAndObject(Kryo.java:776)
>   at 
> org.apache.hive.com.esotericsoftware.kryo.serializers.CollectionSerializer.read(CollectionSerializer.java:112)
>   at 
> org.apache.hive.com.esotericsoftware.kryo.serializers.CollectionSerializer.read(CollectionSerializer.java:18)
>   at 
> org.apache.hive.com.esotericsoftware.kryo.Kryo.readObject(Kryo.java:694)
>   at 
> org.apache.hive.com.esotericsoftware.kryo.serializers.ObjectField.read(ObjectField.java:106)
>   at 
> org.apache.hive.com.esotericsoftware.kryo.serializers.FieldSerializer.read(FieldSerializer.java:507)
>   at 
> org.apache.hive.com.esotericsoftware.kryo.Kryo.readCla

[jira] [Updated] (HIVE-11210) Remove dependency on HiveConf from Orc reader & writer

2015-07-17 Thread Owen O'Malley (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-11210?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Owen O'Malley updated HIVE-11210:
-
Attachment: HIVE-11210.patch

Removed the use of Configuration.getDouble, which messes up the jenkins build.

> Remove dependency on HiveConf from Orc reader & writer
> --
>
> Key: HIVE-11210
> URL: https://issues.apache.org/jira/browse/HIVE-11210
> Project: Hive
>  Issue Type: Sub-task
>Reporter: Owen O'Malley
>Assignee: Owen O'Malley
> Attachments: HIVE-11210.patch, HIVE-11210.patch
>
>
> Currently the ORC reader and writer get their default values from HiveConf. I 
> propose that we make the reader and writer have their own programatic 
> defaults and the OrcInputFormat and OrcOutputFormat can use the version in 
> HiveConf.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-11210) Remove dependency on HiveConf from Orc reader & writer

2015-07-17 Thread Owen O'Malley (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-11210?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Owen O'Malley updated HIVE-11210:
-
Attachment: (was: HIVE-11210.patch)

> Remove dependency on HiveConf from Orc reader & writer
> --
>
> Key: HIVE-11210
> URL: https://issues.apache.org/jira/browse/HIVE-11210
> Project: Hive
>  Issue Type: Sub-task
>Reporter: Owen O'Malley
>Assignee: Owen O'Malley
> Attachments: HIVE-11210.patch
>
>
> Currently the ORC reader and writer get their default values from HiveConf. I 
> propose that we make the reader and writer have their own programatic 
> defaults and the OrcInputFormat and OrcOutputFormat can use the version in 
> HiveConf.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-11210) Remove dependency on HiveConf from Orc reader & writer

2015-07-17 Thread Owen O'Malley (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-11210?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Owen O'Malley updated HIVE-11210:
-
Attachment: (was: HIVE-11210.patch)

> Remove dependency on HiveConf from Orc reader & writer
> --
>
> Key: HIVE-11210
> URL: https://issues.apache.org/jira/browse/HIVE-11210
> Project: Hive
>  Issue Type: Sub-task
>Reporter: Owen O'Malley
>Assignee: Owen O'Malley
> Attachments: HIVE-11210.patch, HIVE-11210.patch
>
>
> Currently the ORC reader and writer get their default values from HiveConf. I 
> propose that we make the reader and writer have their own programatic 
> defaults and the OrcInputFormat and OrcOutputFormat can use the version in 
> HiveConf.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-11280) Support executing script file from hdfs in new CLI [Beeline-CLI branch]

2015-07-17 Thread Xuefu Zhang (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-11280?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14631637#comment-14631637
 ] 

Xuefu Zhang commented on HIVE-11280:


+1

> Support executing script file from hdfs in new CLI [Beeline-CLI branch]
> ---
>
> Key: HIVE-11280
> URL: https://issues.apache.org/jira/browse/HIVE-11280
> Project: Hive
>  Issue Type: Sub-task
>  Components: Beeline, CLI
>Reporter: Ferdinand Xu
>Assignee: Ferdinand Xu
> Attachments: HIVE-11280.1-beeline-cli.patch
>
>
> In HIVE-7136, old CLI is able to read hive scripts from any of the supported 
> file systems in hadoop eco-system. We need to support it in new CLI as well.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-10171) Create a storage-api module

2015-07-17 Thread Owen O'Malley (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-10171?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Owen O'Malley updated HIVE-10171:
-
Fix Version/s: 2.0.0

> Create a storage-api module
> ---
>
> Key: HIVE-10171
> URL: https://issues.apache.org/jira/browse/HIVE-10171
> Project: Hive
>  Issue Type: Bug
>Reporter: Owen O'Malley
>Assignee: Owen O'Malley
> Fix For: 2.0.0
>
>
> To support high performance file formats, I'd like to propose that we move 
> the minimal set of classes that are required to integrate with Hive in to a 
> new module named "storage-api". This module will include VectorizedRowBatch, 
> the various ColumnVector classes, and the SARG classes. It will form the 
> start of an API that high performance storage formats can use to integrate 
> with Hive. Both ORC and Parquet can use the new API to support vectorization 
> and SARGs without performance destroying shims.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Assigned] (HIVE-11253) Move SearchArgument and VectorizedRowBatch classes to storage-api.

2015-07-17 Thread Owen O'Malley (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-11253?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Owen O'Malley reassigned HIVE-11253:


Assignee: Owen O'Malley

> Move SearchArgument and VectorizedRowBatch classes to storage-api.
> --
>
> Key: HIVE-11253
> URL: https://issues.apache.org/jira/browse/HIVE-11253
> Project: Hive
>  Issue Type: Sub-task
>Reporter: Owen O'Malley
>Assignee: Owen O'Malley
>




--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-11269) Intermittent bug with kryo serializers

2015-07-17 Thread Soundararajan Velu (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-11269?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14631576#comment-14631576
 ] 

Soundararajan Velu commented on HIVE-11269:
---

during runtime I get the below exception when I run the same query on MR, on 
Tez I get it on the console itself as described in the previous commment,

Launching Job 10 out of 11
Number of reduce tasks not specified. Estimated from input data size: 300
In order to change the average load for a reducer (in bytes):
  set hive.exec.reducers.bytes.per.reducer=
In order to limit the maximum number of reducers:
  set hive.exec.reducers.max=
In order to set a constant number of reducers:
  set mapreduce.job.reduces=
java.lang.RuntimeException: Error caching reduce.xml: 
org.apache.hive.com.esotericsoftware.kryo.KryoException: 
java.util.ConcurrentModificationException
Serialization trace:
classes (sun.misc.Launcher$AppClassLoader)
classloader (java.security.ProtectionDomain)
context (java.security.AccessControlContext)
acc (java.net.URLClassLoader)
contextClassLoader (java.lang.Thread)
thread (java.util.concurrent.ThreadPoolExecutor$Worker)
workers (java.util.concurrent.ThreadPoolExecutor)
heartBeatThreadPool (net.rubyeye.xmemcached.impl.MemcachedHandler)
handler (net.rubyeye.xmemcached.impl.MemcachedConnector)
connector (net.rubyeye.xmemcached.XMemcachedClient)
memClient (org.apache.hadoop.hive.ql.udf.generic.MecGet)
genericUDF (org.apache.hadoop.hive.ql.plan.ExprNodeGenericFuncDesc)
colExprMap (org.apache.hadoop.hive.ql.exec.SelectOperator)
childOperators (org.apache.hadoop.hive.ql.exec.JoinOperator)
reducer (org.apache.hadoop.hive.ql.plan.ReduceWork)
at 
org.apache.hadoop.hive.ql.exec.Utilities.setBaseWork(Utilities.java:748)
at 
org.apache.hadoop.hive.ql.exec.Utilities.setReduceWork(Utilities.java:687)
at 
org.apache.hadoop.hive.ql.exec.Utilities.setMapRedWork(Utilities.java:678)
at 
org.apache.hadoop.hive.ql.exec.mr.ExecDriver.execute(ExecDriver.java:375)
at 
org.apache.hadoop.hive.ql.exec.mr.MapRedTask.execute(MapRedTask.java:137)
at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:160)
at 
org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:89)
at org.apache.hadoop.hive.ql.exec.TaskRunner.run(TaskRunner.java:75)
Caused by: org.apache.hive.com.esotericsoftware.kryo.KryoException: 
java.util.ConcurrentModificationException
Serialization trace:
classes (sun.misc.Launcher$AppClassLoader)
classloader (java.security.ProtectionDomain)
context (java.security.AccessControlContext)
acc (java.net.URLClassLoader)
contextClassLoader (java.lang.Thread)
thread (java.util.concurrent.ThreadPoolExecutor$Worker)
workers (java.util.concurrent.ThreadPoolExecutor)
heartBeatThreadPool (net.rubyeye.xmemcached.impl.MemcachedHandler)
handler (net.rubyeye.xmemcached.impl.MemcachedConnector)
connector (net.rubyeye.xmemcached.XMemcachedClient)
memClient (org.apache.hadoop.hive.ql.udf.generic.MecGet)
genericUDF (org.apache.hadoop.hive.ql.plan.ExprNodeGenericFuncDesc)
colExprMap (org.apache.hadoop.hive.ql.exec.SelectOperator)
childOperators (org.apache.hadoop.hive.ql.exec.JoinOperator)
reducer (org.apache.hadoop.hive.ql.plan.ReduceWork)
at 
org.apache.hive.com.esotericsoftware.kryo.serializers.ObjectField.write(ObjectField.java:82)
at 
org.apache.hive.com.esotericsoftware.kryo.serializers.FieldSerializer.write(FieldSerializer.java:474)
at 
org.apache.hive.com.esotericsoftware.kryo.Kryo.writeObject(Kryo.java:538)
at 
org.apache.hive.com.esotericsoftware.kryo.serializers.ObjectField.write(ObjectField.java:61)
at 
org.apache.hive.com.esotericsoftware.kryo.serializers.FieldSerializer.write(FieldSerializer.java:474)
at 
org.apache.hive.com.esotericsoftware.kryo.Kryo.writeClassAndObject(Kryo.java:614)
at 
org.apache.hive.com.esotericsoftware.kryo.serializers.DefaultArraySerializers$ObjectArraySerializer.write(DefaultArraySerializers.java:337)
at 
org.apache.hive.com.esotericsoftware.kryo.serializers.DefaultArraySerializers$ObjectArraySerializer.write(DefaultArraySerializers.java:276)
at 
org.apache.hive.com.esotericsoftware.kryo.Kryo.writeObject(Kryo.java:538)
at 
org.apache.hive.com.esotericsoftware.kryo.serializers.ObjectField.write(ObjectField.java:61)
at 
org.apache.hive.com.esotericsoftware.kryo.serializers.FieldSerializer.write(FieldSerializer.java:474)
at 
org.apache.hive.com.esotericsoftware.kryo.Kryo.writeObjectOrNull(Kryo.java:592)
at 
org.apache.hive.com.esotericsoftware.kryo.serializers.ObjectField.write(ObjectField.java:68)
at 
org.apache.hive.com.esotericsoftware.kryo.serializers.FieldSerializer.write(FieldSerializer.java:474)
at 
org.apache.hive.com.esotericsoftware.kryo.Kryo.writeObject(Kryo.java:538)
at 
org.apache.hive.com.esotericsoftware.kryo.ser

[jira] [Commented] (HIVE-11254) Process result sets returned by a stored procedure

2015-07-17 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-11254?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14631546#comment-14631546
 ] 

Hive QA commented on HIVE-11254:




{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12745815/HIVE-11254.2.patch

{color:red}ERROR:{color} -1 due to 8 failed/errored test(s), 9227 tests executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_cbo_rp_join0
org.apache.hive.hcatalog.streaming.TestStreaming.testTransactionBatchCommit_Json
org.apache.hive.hplsql.TestHplsqlDb.testCreateProcedureReturnCursor
org.apache.hive.hplsql.TestHplsqlDb.testCreateProcedureReturnCursor2
org.apache.hive.hplsql.TestHplsqlLocal.testException2
org.apache.hive.hplsql.TestHplsqlLocal.testException3
org.apache.hive.hplsql.TestHplsqlLocal.testException4
org.apache.hive.hplsql.TestHplsqlLocal.testException5
{noformat}

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/4637/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/4637/console
Test logs: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-4637/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 8 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12745815 - PreCommit-HIVE-TRUNK-Build

> Process result sets returned by a stored procedure
> --
>
> Key: HIVE-11254
> URL: https://issues.apache.org/jira/browse/HIVE-11254
> Project: Hive
>  Issue Type: Improvement
>  Components: hpl/sql
>Reporter: Dmitry Tolpeko
>Assignee: Dmitry Tolpeko
> Attachments: HIVE-11254.1.patch, HIVE-11254.2.patch
>
>
> Stored procedure can return one or more result sets. A caller should be able 
> to process them.
>  



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-11269) Intermittent bug with kryo serializers

2015-07-17 Thread Soundararajan Velu (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-11269?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14631534#comment-14631534
 ] 

Soundararajan Velu commented on HIVE-11269:
---

Thanks Xuefu, I cloned 1.3.0 release 2 weeks back, was this fixed post that, if 
possible can you please provide the issue id that fixes this bug,

Now I am getting the below exception while using custom udaf and udfs,

Caused by: java.lang.Exception: XMLEncoder: discarding statement 
ArrayList.add(FileSinkOperator);
... 67 more
Caused by: java.lang.RuntimeException: Cannot serialize object
at 
org.apache.hadoop.hive.ql.exec.Utilities$2.exceptionThrown(Utilities.java:1034)
at 
java.beans.DefaultPersistenceDelegate.initBean(DefaultPersistenceDelegate.java:259)
at 
java.beans.DefaultPersistenceDelegate.initialize(DefaultPersistenceDelegate.java:403)
at 
java.beans.PersistenceDelegate.writeObject(PersistenceDelegate.java:118)
at java.beans.Encoder.writeObject(Encoder.java:74)
at java.beans.XMLEncoder.writeObject(XMLEncoder.java:327)
at java.beans.Encoder.writeExpression(Encoder.java:330)
at java.beans.XMLEncoder.writeExpression(XMLEncoder.java:454)
at 
java.beans.PersistenceDelegate.writeObject(PersistenceDelegate.java:115)
at java.beans.Encoder.writeObject(Encoder.java:74)
at java.beans.XMLEncoder.writeObject(XMLEncoder.java:327)
at java.beans.Encoder.writeObject1(Encoder.java:258)
at java.beans.Encoder.cloneStatement(Encoder.java:271)
at java.beans.Encoder.writeStatement(Encoder.java:301)
at java.beans.XMLEncoder.writeStatement(XMLEncoder.java:400)
... 66 more
Caused by: java.lang.RuntimeException: Cannot serialize object
at 
org.apache.hadoop.hive.ql.exec.Utilities$2.exceptionThrown(Utilities.java:1034)
at 
java.beans.DefaultPersistenceDelegate.initBean(DefaultPersistenceDelegate.java:259)
at 
java.beans.DefaultPersistenceDelegate.initialize(DefaultPersistenceDelegate.java:403)
at 
java.beans.PersistenceDelegate.writeObject(PersistenceDelegate.java:118)
at java.beans.Encoder.writeObject(Encoder.java:74)
at java.beans.XMLEncoder.writeObject(XMLEncoder.java:327)
at java.beans.Encoder.writeExpression(Encoder.java:330)
at java.beans.XMLEncoder.writeExpression(XMLEncoder.java:454)
at 
java.beans.PersistenceDelegate.writeObject(PersistenceDelegate.java:115)
at java.beans.Encoder.writeObject(Encoder.java:74)
at java.beans.XMLEncoder.writeObject(XMLEncoder.java:327)
at java.beans.Encoder.writeExpression(Encoder.java:330)
at java.beans.XMLEncoder.writeExpression(XMLEncoder.java:454)
at 
java.beans.DefaultPersistenceDelegate.doProperty(DefaultPersistenceDelegate.java:194)
at 
java.beans.DefaultPersistenceDelegate.initBean(DefaultPersistenceDelegate.java:256)
... 79 more
Caused by: java.lang.RuntimeException: Cannot serialize object
at 
org.apache.hadoop.hive.ql.exec.Utilities$2.exceptionThrown(Utilities.java:1034)
at java.beans.Encoder.getValue(Encoder.java:108)
at java.beans.Encoder.get(Encoder.java:252)
at 
java.beans.PersistenceDelegate.writeObject(PersistenceDelegate.java:112)
at java.beans.Encoder.writeObject(Encoder.java:74)
at java.beans.XMLEncoder.writeObject(XMLEncoder.java:327)
at java.beans.Encoder.writeExpression(Encoder.java:330)
at java.beans.XMLEncoder.writeExpression(XMLEncoder.java:454)
at 
java.beans.PersistenceDelegate.writeObject(PersistenceDelegate.java:115)
at java.beans.Encoder.writeObject(Encoder.java:74)
at java.beans.XMLEncoder.writeObject(XMLEncoder.java:327)
at java.beans.Encoder.writeExpression(Encoder.java:330)
at java.beans.XMLEncoder.writeExpression(XMLEncoder.java:454)
at 
java.beans.DefaultPersistenceDelegate.doProperty(DefaultPersistenceDelegate.java:194)
at 
java.beans.DefaultPersistenceDelegate.initBean(DefaultPersistenceDelegate.java:256)
... 92 more
Caused by: java.lang.InstantiationException: 
org.apache.hadoop.hive.ql.plan.OpTraits
at java.lang.Class.newInstance(Class.java:359)
at sun.reflect.GeneratedMethodAccessor41.invoke(Unknown Source)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:606)
at sun.reflect.misc.Trampoline.invoke(MethodUtil.java:75)
at sun.reflect.GeneratedMethodAccessor35.invoke(Unknown Source)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:606)
at sun.reflect.misc.MethodUtil.invoke(MethodUtil.java:279)
at java.beans.Statement.invokeInternal(Statement.java:292)

[jira] [Commented] (HIVE-11288) Avro SerDe InstanceCache returns incorrect schema

2015-07-17 Thread Chaoyu Tang (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-11288?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14631508#comment-14631508
 ] 

Chaoyu Tang commented on HIVE-11288:


[~gphillips] Could you take a look at the failed test 
org.apache.hadoop.hive.serde2.avro.TestInstanceCache.instanceCachesOnlyCreateOneInstance?
 is it related to patch?

> Avro SerDe InstanceCache returns incorrect schema
> -
>
> Key: HIVE-11288
> URL: https://issues.apache.org/jira/browse/HIVE-11288
> Project: Hive
>  Issue Type: Bug
>Reporter: Greg Phillips
> Attachments: HIVE-11288.patch
>
>
> To reproduce this error, take two fields in an avro schema document matching 
> the following:
> "type" :  { "type": "array", "items": [ "null",  { "type": "map", "values": [ 
> "null", "string" ] } ]  }
> "type" : { "type": "map", "values": [ "null" , { "type": "array", "items": [ 
> "null" , "string"] } ] }
> After creating two tables in hive with these schemas, the describe statement 
> on each of them will only return the schema for the first one loaded.  This 
> is due to a hashCode() collision in the InstanceCache.  
> A patch will be included in this ticket shortly which removes the hashCode 
> call from the InstanceCache's internal HashMap, and instead provides the 
> entire schema object.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Resolved] (HIVE-11273) LLAP: Register for finishable state change notifications when adding a task instead of when scheduling it

2015-07-17 Thread Siddharth Seth (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-11273?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Siddharth Seth resolved HIVE-11273.
---
Resolution: Fixed

> LLAP: Register for finishable state change notifications when adding a task 
> instead of when scheduling it
> -
>
> Key: HIVE-11273
> URL: https://issues.apache.org/jira/browse/HIVE-11273
> Project: Hive
>  Issue Type: Sub-task
>Reporter: Siddharth Seth
>Assignee: Siddharth Seth
> Fix For: llap
>
> Attachments: HIVE-11273.1.txt
>
>
> Registering when trying to execute is far too late. The task won't be 
> considered for execution (queue may not be re-oredered) without the 
> notification coming in.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Resolved] (HIVE-11263) LLAP: TaskExecutorService state is not cleaned up

2015-07-17 Thread Siddharth Seth (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-11263?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Siddharth Seth resolved HIVE-11263.
---
   Resolution: Fixed
Fix Version/s: llap

> LLAP: TaskExecutorService state is not cleaned up
> -
>
> Key: HIVE-11263
> URL: https://issues.apache.org/jira/browse/HIVE-11263
> Project: Hive
>  Issue Type: Sub-task
>Reporter: Sergey Shelukhin
>Assignee: Siddharth Seth
> Fix For: llap
>
> Attachments: HIVE-11263.1.txt, HIVE-11263.2.txt
>
>
> See TaskExecutorService::getExecutorsStatus, this is used to report on 
> queue/etc. status in JMX.
> Currently, it reports 100s of bogus tasks in queue:
> {noformat}
> "ExecutorsStatus" : [ "attempt_1435700346116_1888_1_04_000205_22 
> (sershe_20150714174105_0d013941-1f0e-4f74-9387-a2f29279a185:3/Map 1, in 
> queue)", "attempt_1435700346116_1889_1_05_000101_1 
> (sershe_20150714174104_b0b9f300-667e-4370-bb01-f9cb7da331e0:4/Map 1, in 
> queue)", "attempt_1435700346116_1889_1_05_000191_3 
> (sershe_20150714174104_b0b9f300-667e-4370-bb01-f9cb7da331e0:4/Map 1, in 
> queue)", "attempt_1435700346116_1887_7_00_000202_3 
> (sershe_20150714174737_bea682d1-fa0f-4281-a1cb-439d85bb2016:22/Map 5, in 
> queue)", "attempt_1435700346116_1886_1_04_73_15 
> (sershe_20150714174108_f9483d76-8fd9-4f82-96ee-17231b6f9b2c:1/Reducer 2, in 
> queue)", "attempt_1435700346116_1887_8_04_000166_15 
> (sershe_20150714174900_710d7d69-3d66-45e9-865b-cd0f87bb0d98:27/Map 1, in 
> queue)", "attempt_1435700346116_1888_11_05_000140_3 
> (sershe_20150714174903_50359459-5342-4d1b-852c-622a3fa92a27:28/Map 3, in 
> queue)", "attempt_1435700346116_1886_1_04_42_29 
> (sershe_20150714174108_f9483d76-8fd9-4f82-96ee-17231b6f9b2c:1/Reducer 2, in 
> queue)", "attempt_1435700346116_1888_2_03_000169_12 
> (sershe_20150714174310_97ce1d4b-8029-4ef6-a823-46e29f09718a:5/Map 1, in 
> queue)", "attempt_1435700346116_1887_1_04_000197_18 
> (sershe_20150714174107_8fcfe954-4eeb-46e5-bad5-42a47327b26c:2/Map 1, in 
> queue)", "attempt_1435700346116_1887_1_04_000218_21 
> (sershe_20150714174107_8fcfe954-4eeb-46e5-bad5-42a47327b26c:2/Map 1, in 
> queue)", "attempt_1435700346116_1886_7_09_84_1 
> (sershe_20150714174841_462b9bdb-c017-47c2-9fa7-7edfbfc09e60:24/Map 1, in 
> queue)", "attempt_1435700346116_1887_5_04_78_0 
> (sershe_20150714174509_9a5cd476-b3c8-4679-af8e-1188922713a2:14/Map 3, in 
> queue)", "attempt_1435700346116_1887_7_04_000162_6 
> (sershe_20150714174737_bea682d1-fa0f-4281-a1cb-439d85bb2016:22/Map 3, in 
> queue)", "attempt_1435700346116_1887_7_04_000180_0 
> (sershe_20150714174737_bea682d1-fa0f-4281-a1cb-439d85bb2016:22/Map 3, in 
> queue)", "attempt_1435700346116_1886_3_04_000144_0 
> (sershe_20150714174435_fe3077dd-a97f-4582-995b-5f723170b02f:12/Reducer 2, in 
> queue)", "attempt_1435700346116_1887_5_00_000153_1 
> (sershe_20150714174509_9a5cd476-b3c8-4679-af8e-1188922713a2:14/Map 5, in 
> queue)", "attempt_1435700346116_1887_7_04_000141_7 
> (sershe_20150714174737_bea682d1-fa0f-4281-a1cb-439d85bb2016:22/Map 3, in 
> queue)", "attempt_1435700346116_1887_1_04_24_7 
> (sershe_20150714174107_8fcfe954-4eeb-46e5-bad5-42a47327b26c:2/Map 1, in 
> queue)", "attempt_1435700346116_1887_5_04_000130_1 
> (sershe_20150714174509_9a5cd476-b3c8-4679-af8e-1188922713a2:14/Map 3, in 
> queue)", "attempt_1435700346116_1888_1_04_000200_1 
> (sershe_20150714174105_0d013941-1f0e-4f74-9387-a2f29279a185:3/Map 1, in 
> queue)", "attempt_1435700346116_1886_15_04_000180_0 
> (sershe_20150714175411_bda950b7-8aa5-417f-84f6-dd646247dca8:43/Map 1, in 
> queue)", "attempt_1435700346116_1887_7_00_000205_1 
> (sershe_20150714174737_bea682d1-fa0f-4281-a1cb-439d85bb2016:22/Map 5, in 
> queue)", "attempt_1435700346116_1888_4_04_000183_4 
> (sershe_20150714174407_f0924540-f69f-45c2-831a-9d2d1f66a124:10/Map 1, in 
> queue)", "attempt_1435700346116_1887_1_04_81_6 
> (sershe_20150714174107_8fcfe954-4eeb-46e5-bad5-42a47327b26c:2/Map 1, in 
> queue)", "attempt_1435700346116_1888_1_04_80_4 
> (sershe_20150714174105_0d013941-1f0e-4f74-9387-a2f29279a185:3/Map 1, in 
> queue)", "attempt_1435700346116_1887_7_04_05_3 
> (sershe_20150714174737_bea682d1-fa0f-4281-a1cb-439d85bb2016:22/Map 3, in 
> queue)", "attempt_1435700346116_1887_7_00_000169_2 
> (sershe_20150714174737_bea682d1-fa0f-4281-a1cb-439d85bb2016:22/Map 5, in 
> queue)", "attempt_1435700346116_1888_8_04_37_2 
> (sershe_20150714174731_261f2d52-8c47-4db6-8f17-8098efe144a2:20/Reducer 3, in 
> queue)", "attempt_1435700346116_1887_9_00_96_6 
> (sershe_20150714175015_cc1b6647-8479-4c5f-918c-00935bff7232:30/Map 5, in 
> queue)", "attempt_1435700346116_1888_11_01_01_2 
> (sershe_20150714174903_50359459-5342-4d1b-852c-622a3fa92a27:28/Map 7, in 
> queue)", "attempt_1435700346116_1889_1_05_000206_8 
> (sershe_20150714174104_b0

[jira] [Commented] (HIVE-11273) LLAP: Register for finishable state change notifications when adding a task instead of when scheduling it

2015-07-17 Thread Siddharth Seth (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-11273?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14631487#comment-14631487
 ] 

Siddharth Seth commented on HIVE-11273:
---

Thanks for the reviews on this and 11263. 11263 actually depends upon (and 
applies on) this patch, where most of the test plumbing has been added. I'll go 
ahead and commit both of them separately since they're separate issues and have 
been reviewed. 

> LLAP: Register for finishable state change notifications when adding a task 
> instead of when scheduling it
> -
>
> Key: HIVE-11273
> URL: https://issues.apache.org/jira/browse/HIVE-11273
> Project: Hive
>  Issue Type: Sub-task
>Reporter: Siddharth Seth
>Assignee: Siddharth Seth
> Fix For: llap
>
> Attachments: HIVE-11273.1.txt
>
>
> Registering when trying to execute is far too late. The task won't be 
> considered for execution (queue may not be re-oredered) without the 
> notification coming in.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-11280) Support executing script file from hdfs in new CLI [Beeline-CLI branch]

2015-07-17 Thread Ferdinand Xu (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-11280?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14631452#comment-14631452
 ] 

Ferdinand Xu commented on HIVE-11280:
-

Hi [~xuefuz], do you have interest reviewing this? 

> Support executing script file from hdfs in new CLI [Beeline-CLI branch]
> ---
>
> Key: HIVE-11280
> URL: https://issues.apache.org/jira/browse/HIVE-11280
> Project: Hive
>  Issue Type: Sub-task
>  Components: Beeline, CLI
>Reporter: Ferdinand Xu
>Assignee: Ferdinand Xu
> Attachments: HIVE-11280.1-beeline-cli.patch
>
>
> In HIVE-7136, old CLI is able to read hive scripts from any of the supported 
> file systems in hadoop eco-system. We need to support it in new CLI as well.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-11288) Avro SerDe InstanceCache returns incorrect schema

2015-07-17 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-11288?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14631433#comment-14631433
 ] 

Hive QA commented on HIVE-11288:




{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12745809/HIVE-11288.patch

{color:red}ERROR:{color} -1 due to 2 failed/errored test(s), 9225 tests executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_cbo_rp_join0
org.apache.hadoop.hive.serde2.avro.TestInstanceCache.instanceCachesOnlyCreateOneInstance
{noformat}

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/4636/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/4636/console
Test logs: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-4636/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 2 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12745809 - PreCommit-HIVE-TRUNK-Build

> Avro SerDe InstanceCache returns incorrect schema
> -
>
> Key: HIVE-11288
> URL: https://issues.apache.org/jira/browse/HIVE-11288
> Project: Hive
>  Issue Type: Bug
>Reporter: Greg Phillips
> Attachments: HIVE-11288.patch
>
>
> To reproduce this error, take two fields in an avro schema document matching 
> the following:
> "type" :  { "type": "array", "items": [ "null",  { "type": "map", "values": [ 
> "null", "string" ] } ]  }
> "type" : { "type": "map", "values": [ "null" , { "type": "array", "items": [ 
> "null" , "string"] } ] }
> After creating two tables in hive with these schemas, the describe statement 
> on each of them will only return the schema for the first one loaded.  This 
> is due to a hashCode() collision in the InstanceCache.  
> A patch will be included in this ticket shortly which removes the hashCode 
> call from the InstanceCache's internal HashMap, and instead provides the 
> entire schema object.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-11254) Process result sets returned by a stored procedure

2015-07-17 Thread Dmitry Tolpeko (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-11254?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Dmitry Tolpeko updated HIVE-11254:
--
Attachment: HIVE-11254.2.patch

Patch 2 created

> Process result sets returned by a stored procedure
> --
>
> Key: HIVE-11254
> URL: https://issues.apache.org/jira/browse/HIVE-11254
> Project: Hive
>  Issue Type: Improvement
>  Components: hpl/sql
>Reporter: Dmitry Tolpeko
>Assignee: Dmitry Tolpeko
> Attachments: HIVE-11254.1.patch, HIVE-11254.2.patch
>
>
> Stored procedure can return one or more result sets. A caller should be able 
> to process them.
>  



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-11276) Optimization around job submission and adding jars [Spark Branch]

2015-07-17 Thread Xuefu Zhang (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-11276?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14631311#comment-14631311
 ] 

Xuefu Zhang commented on HIVE-11276:


Thanks. I will create a JIRA for that.

> Optimization around job submission and adding jars [Spark Branch]
> -
>
> Key: HIVE-11276
> URL: https://issues.apache.org/jira/browse/HIVE-11276
> Project: Hive
>  Issue Type: Sub-task
>  Components: Spark
>Affects Versions: 1.1.0
>Reporter: Xuefu Zhang
>Assignee: Chengxiang Li
>
> It seems that Hive on Spark has some room for performance improvement on job 
> submission. Specifically, we are calling refreshLocalResources() for every 
> job submission despite there is are no changes in the jar list. Since Hive on 
> Spark is reusing the containers in the whole user session, we might be able 
> to optimize that.
> We do need to take into consideration the case of dynamic allocation, in 
> which new executors might be added.
> This task is some R&D in this area.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-11254) Process result sets returned by a stored procedure

2015-07-17 Thread Dmitry Tolpeko (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-11254?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14631310#comment-14631310
 ] 

Dmitry Tolpeko commented on HIVE-11254:
---

Got an error in tests:
 java.sql.SQLException: Could not open client transport with JDBC Uri: 
jdbc:hive2://localhost:1: java.net.ConnectException: Connection refused

I will modify the connection URL to use embedded URL jdbc:hive2://

> Process result sets returned by a stored procedure
> --
>
> Key: HIVE-11254
> URL: https://issues.apache.org/jira/browse/HIVE-11254
> Project: Hive
>  Issue Type: Improvement
>  Components: hpl/sql
>Reporter: Dmitry Tolpeko
>Assignee: Dmitry Tolpeko
> Attachments: HIVE-11254.1.patch
>
>
> Stored procedure can return one or more result sets. A caller should be able 
> to process them.
>  



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Resolved] (HIVE-11276) Optimization around job submission and adding jars [Spark Branch]

2015-07-17 Thread Xuefu Zhang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-11276?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Xuefu Zhang resolved HIVE-11276.

Resolution: Not A Problem

> Optimization around job submission and adding jars [Spark Branch]
> -
>
> Key: HIVE-11276
> URL: https://issues.apache.org/jira/browse/HIVE-11276
> Project: Hive
>  Issue Type: Sub-task
>  Components: Spark
>Affects Versions: 1.1.0
>Reporter: Xuefu Zhang
>Assignee: Chengxiang Li
>
> It seems that Hive on Spark has some room for performance improvement on job 
> submission. Specifically, we are calling refreshLocalResources() for every 
> job submission despite there is are no changes in the jar list. Since Hive on 
> Spark is reusing the containers in the whole user session, we might be able 
> to optimize that.
> We do need to take into consideration the case of dynamic allocation, in 
> which new executors might be added.
> This task is some R&D in this area.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-11288) Avro SerDe InstanceCache returns incorrect schema

2015-07-17 Thread Greg Phillips (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-11288?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Greg Phillips updated HIVE-11288:
-
Attachment: HIVE-11288.patch

> Avro SerDe InstanceCache returns incorrect schema
> -
>
> Key: HIVE-11288
> URL: https://issues.apache.org/jira/browse/HIVE-11288
> Project: Hive
>  Issue Type: Bug
>Reporter: Greg Phillips
> Attachments: HIVE-11288.patch
>
>
> To reproduce this error, take two fields in an avro schema document matching 
> the following:
> "type" :  { "type": "array", "items": [ "null",  { "type": "map", "values": [ 
> "null", "string" ] } ]  }
> "type" : { "type": "map", "values": [ "null" , { "type": "array", "items": [ 
> "null" , "string"] } ] }
> After creating two tables in hive with these schemas, the describe statement 
> on each of them will only return the schema for the first one loaded.  This 
> is due to a hashCode() collision in the InstanceCache.  
> A patch will be included in this ticket shortly which removes the hashCode 
> call from the InstanceCache's internal HashMap, and instead provides the 
> entire schema object.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-11254) Process result sets returned by a stored procedure

2015-07-17 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-11254?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14631227#comment-14631227
 ] 

Hive QA commented on HIVE-11254:




{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12745797/HIVE-11254.1.patch

{color:red}ERROR:{color} -1 due to 3 failed/errored test(s), 9227 tests executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_cbo_rp_join0
org.apache.hive.hplsql.TestHplsqlDb.testCreateProcedureReturnCursor
org.apache.hive.hplsql.TestHplsqlDb.testCreateProcedureReturnCursor2
{noformat}

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/4635/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/4635/console
Test logs: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-4635/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 3 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12745797 - PreCommit-HIVE-TRUNK-Build

> Process result sets returned by a stored procedure
> --
>
> Key: HIVE-11254
> URL: https://issues.apache.org/jira/browse/HIVE-11254
> Project: Hive
>  Issue Type: Improvement
>  Components: hpl/sql
>Reporter: Dmitry Tolpeko
>Assignee: Dmitry Tolpeko
> Attachments: HIVE-11254.1.patch
>
>
> Stored procedure can return one or more result sets. A caller should be able 
> to process them.
>  



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-11254) Process result sets returned by a stored procedure

2015-07-17 Thread Dmitry Tolpeko (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-11254?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Dmitry Tolpeko updated HIVE-11254:
--
Attachment: HIVE-11254.1.patch

> Process result sets returned by a stored procedure
> --
>
> Key: HIVE-11254
> URL: https://issues.apache.org/jira/browse/HIVE-11254
> Project: Hive
>  Issue Type: Improvement
>  Components: hpl/sql
>Reporter: Dmitry Tolpeko
>Assignee: Dmitry Tolpeko
> Attachments: HIVE-11254.1.patch
>
>
> Stored procedure can return one or more result sets. A caller should be able 
> to process them.
>  



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-11280) Support executing script file from hdfs in new CLI [Beeline-CLI branch]

2015-07-17 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-11280?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14631107#comment-14631107
 ] 

Hive QA commented on HIVE-11280:




{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12745777/HIVE-11280.1-beeline-cli.patch

{color:red}ERROR:{color} -1 due to 3 failed/errored test(s), 9233 tests executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_cbo_rp_join0
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_sortmerge_join_2
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_sortmerge_join_8
{noformat}

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-BEELINE-Build/10/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-BEELINE-Build/10/console
Test logs: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-BEELINE-Build-10/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 3 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12745777 - PreCommit-HIVE-BEELINE-Build

> Support executing script file from hdfs in new CLI [Beeline-CLI branch]
> ---
>
> Key: HIVE-11280
> URL: https://issues.apache.org/jira/browse/HIVE-11280
> Project: Hive
>  Issue Type: Sub-task
>  Components: Beeline, CLI
>Reporter: Ferdinand Xu
>Assignee: Ferdinand Xu
> Attachments: HIVE-11280.1-beeline-cli.patch
>
>
> In HIVE-7136, old CLI is able to read hive scripts from any of the supported 
> file systems in hadoop eco-system. We need to support it in new CLI as well.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-7193) Hive should support additional LDAP authentication parameters

2015-07-17 Thread Lefty Leverenz (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-7193?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14630992#comment-14630992
 ] 

Lefty Leverenz commented on HIVE-7193:
--

Doc note:  The configuration parameters are documented in the HiveServer2 
section of Configuration Properties, so I removed the TODOC1.3 label.

* [Configuration Properties -- hive.server2.authentication.ldap.groupDNPattern 
| 
https://cwiki.apache.org/confluence/display/Hive/Configuration+Properties#ConfigurationProperties-hive.server2.authentication.ldap.groupDNPattern]
* [Configuration Properties -- hive.server2.authentication.ldap.groupFilter | 
https://cwiki.apache.org/confluence/display/Hive/Configuration+Properties#ConfigurationProperties-hive.server2.authentication.ldap.groupFilter]
* [Configuration Properties -- hive.server2.authentication.ldap.userDNPattern | 
https://cwiki.apache.org/confluence/display/Hive/Configuration+Properties#ConfigurationProperties-hive.server2.authentication.ldap.userDNPattern]
* [Configuration Properties -- hive.server2.authentication.ldap.userFilter | 
https://cwiki.apache.org/confluence/display/Hive/Configuration+Properties#ConfigurationProperties-hive.server2.authentication.ldap.userFilter]
* [Configuration Properties -- hive.server2.authentication.ldap.customLDAPQuery 
| 
https://cwiki.apache.org/confluence/display/Hive/Configuration+Properties#ConfigurationProperties-hive.server2.authentication.ldap.customLDAPQuery]

Setting Up HiveServer2 has a link to User and Group Filter Support ... (see 
link in last comment).

> Hive should support additional LDAP authentication parameters
> -
>
> Key: HIVE-7193
> URL: https://issues.apache.org/jira/browse/HIVE-7193
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 0.10.0
>Reporter: Mala Chikka Kempanna
>Assignee: Naveen Gangam
> Fix For: 1.3.0, 2.0.0
>
> Attachments: HIVE-7193.2.patch, HIVE-7193.3.patch, HIVE-7193.4.patch, 
> HIVE-7193.5.patch, HIVE-7193.6.patch, HIVE-7193.patch, 
> LDAPAuthentication_Design_Doc.docx, LDAPAuthentication_Design_Doc_V2.docx
>
>
> Currently hive has only following authenticator parameters for LDAP 
> authentication for hiveserver2:
> {code:xml}
>  
>   hive.server2.authentication 
>   LDAP 
>  
>  
>   hive.server2.authentication.ldap.url 
>   ldap://our_ldap_address 
>  
> {code}
> We need to include other LDAP properties as part of hive-LDAP authentication 
> like below:
> {noformat}
> a group search base -> dc=domain,dc=com 
> a group search filter -> member={0} 
> a user search base -> dc=domain,dc=com 
> a user search filter -> sAMAAccountName={0} 
> a list of valid user groups -> group1,group2,group3 
> {noformat}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-7193) Hive should support additional LDAP authentication parameters

2015-07-17 Thread Lefty Leverenz (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-7193?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Lefty Leverenz updated HIVE-7193:
-
Labels:   (was: TODOC1.3)

> Hive should support additional LDAP authentication parameters
> -
>
> Key: HIVE-7193
> URL: https://issues.apache.org/jira/browse/HIVE-7193
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 0.10.0
>Reporter: Mala Chikka Kempanna
>Assignee: Naveen Gangam
> Fix For: 1.3.0, 2.0.0
>
> Attachments: HIVE-7193.2.patch, HIVE-7193.3.patch, HIVE-7193.4.patch, 
> HIVE-7193.5.patch, HIVE-7193.6.patch, HIVE-7193.patch, 
> LDAPAuthentication_Design_Doc.docx, LDAPAuthentication_Design_Doc_V2.docx
>
>
> Currently hive has only following authenticator parameters for LDAP 
> authentication for hiveserver2:
> {code:xml}
>  
>   hive.server2.authentication 
>   LDAP 
>  
>  
>   hive.server2.authentication.ldap.url 
>   ldap://our_ldap_address 
>  
> {code}
> We need to include other LDAP properties as part of hive-LDAP authentication 
> like below:
> {noformat}
> a group search base -> dc=domain,dc=com 
> a group search filter -> member={0} 
> a user search base -> dc=domain,dc=com 
> a user search filter -> sAMAAccountName={0} 
> a list of valid user groups -> group1,group2,group3 
> {noformat}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-11285) ObjectInspector for partition columns in FetchOperator in SMBJoin causes exception

2015-07-17 Thread Damien Carol (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-11285?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Damien Carol updated HIVE-11285:

Description: 
STEPS TO REPRODUCE:
{noformat}
*$ cat data.out 
1|One 
2|Two
{noformat}
{code:sql}
hql> 
CREATE TABLE data_table (key INT, value STRING) ROW FORMAT DELIMITED FIELDS 
TERMINATED BY '|'; 
LOAD DATA LOCAL INPATH '${system:user.dir}/data.out' INTO TABLE data_table;
CREATE TABLE smb_table (key INT, value STRING) 
CLUSTERED BY (key) 
SORTED BY (key) INTO 1 BUCKETS 
STORED AS ORC;
CREATE TABLE smb_table_part (key INT, value STRING) 
PARTITIONED BY (p1 DECIMAL) 
CLUSTERED BY (key) 
SORTED BY (key) INTO 1 BUCKETS 
STORED AS ORC;
INSERT OVERWRITE TABLE smb_table SELECT * FROM data_table; 
INSERT OVERWRITE TABLE smb_table_part PARTITION (p1) SELECT key, value, 100 as 
p1 FROM data_table;
SET hive.execution.engine=mr; 
SET hive.enforce.sortmergebucketmapjoin=false; 
SET hive.auto.convert.sortmerge.join=true; 
SET hive.optimize.bucketmapjoin = true; 
SET hive.optimize.bucketmapjoin.sortedmerge = true; 
SET hive.input.format = org.apache.hadoop.hive.ql.io.BucketizedHiveInputFormat;
SELECT s1.key, s2.p1 
FROM smb_table s1 
INNER JOIN smb_table_part s2 
ON s1.key = s2.key 
ORDER BY s1.key;
{code}
ERROR:
{noformat}
2015-07-15 13:39:04,333 WARN main org.apache.hadoop.mapred.YarnChild: Exception 
running child : java.lang.RuntimeException: 
org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime Error while 
processing row
{"key":1,"value":"One"}

at org.apache.hadoop.hive.ql.exec.mr.ExecMapper.map(ExecMapper.java:185) 
at org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:54) 
at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:450) 
at org.apache.hadoop.mapred.MapTask.run(MapTask.java:343) 
at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:163) 
at java.security.AccessController.doPrivileged(Native Method) 
at javax.security.auth.Subject.doAs(Subject.java:415) 
at 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1628)
 
at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:158) 
Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime Error 
while processing row
{"key":1,"value":"One"}

at org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:503) 
at org.apache.hadoop.hive.ql.exec.mr.ExecMapper.map(ExecMapper.java:176) 
... 8 more 
Caused by: java.lang.RuntimeException: Map local work failed 
at 
org.apache.hadoop.hive.ql.exec.SMBMapJoinOperator.fetchOneRow(SMBMapJoinOperator.java:569)
 
at 
org.apache.hadoop.hive.ql.exec.SMBMapJoinOperator.fetchNextGroup(SMBMapJoinOperator.java:429)
 
at 
org.apache.hadoop.hive.ql.exec.SMBMapJoinOperator.processOp(SMBMapJoinOperator.java:260)
 
at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:815) 
at 
org.apache.hadoop.hive.ql.exec.FilterOperator.processOp(FilterOperator.java:120)
 
at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:815) 
at 
org.apache.hadoop.hive.ql.exec.TableScanOperator.processOp(TableScanOperator.java:95)
 
at 
org.apache.hadoop.hive.ql.exec.MapOperator$MapOpCtx.forward(MapOperator.java:157)
 
at org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:493) 
... 9 more 
Caused by: java.lang.ClassCastException: org.apache.hadoop.io.IntWritable 
cannot be cast to java.lang.Integer 
at 
org.apache.hadoop.hive.serde2.objectinspector.primitive.JavaIntObjectInspector.getPrimitiveWritableObject(JavaIntObjectInspector.java:35)
 
at 
org.apache.hadoop.hive.serde2.objectinspector.ObjectInspectorUtils.copyToStandardObject(ObjectInspectorUtils.java:305)
 
at org.apache.hadoop.hive.ql.exec.JoinUtil.computeValues(JoinUtil.java:193) 
at 
org.apache.hadoop.hive.ql.exec.CommonJoinOperator.getFilteredValue(CommonJoinOperator.java:408)
 
at 
org.apache.hadoop.hive.ql.exec.SMBMapJoinOperator.processOp(SMBMapJoinOperator.java:270)
 
at 
org.apache.hadoop.hive.ql.exec.SMBMapJoinOperator.fetchOneRow(SMBMapJoinOperator.java:558)
 
... 17 more
{noformat}

  was:
{code}
STEPS TO REPRODUCE:
*$ cat data.out 
1|One 
2|Two
hql> 
CREATE TABLE data_table (key INT, value STRING) ROW FORMAT DELIMITED FIELDS 
TERMINATED BY '|'; 
LOAD DATA LOCAL INPATH '$
{system:user.dir}
/data.out' INTO TABLE data_table;
CREATE TABLE smb_table (key INT, value STRING) 
CLUSTERED BY (key) 
SORTED BY (key) INTO 1 BUCKETS 
STORED AS ORC;
CREATE TABLE smb_table_part (key INT, value STRING) 
PARTITIONED BY (p1 DECIMAL) 
CLUSTERED BY (key) 
SORTED BY (key) INTO 1 BUCKETS 
STORED AS ORC;
INSERT OVERWRITE TABLE smb_table SELECT * FROM data_table; 
INSERT OVERWRITE TABLE smb_table_part PARTITION (p1) SELECT key, value, 100 as 
p1 FROM data_table;
SET hive.execution.engine=mr; 
SET hive.enforce.sortmergebucketmapjoin=false; 
SET hive.auto.convert.sortmerge.join=true; 
SET hive.optimize.bucketmapjoin = true; 
SET hive.optimize.bucketmapjoin.sortedmerge = true; 
SET hive.input.format = 

  1   2   >