[jira] [Updated] (HIVE-8456) Support Hive Counter to collect spark job metric[Spark Branch]

2014-10-16 Thread Chengxiang Li (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-8456?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chengxiang Li updated HIVE-8456:

Attachment: HIVE-8456.3-spark.patch

set javaSparkContext and hiveConf as transient variables, as they are only used 
in driver side.

 Support Hive Counter to collect spark job metric[Spark Branch]
 --

 Key: HIVE-8456
 URL: https://issues.apache.org/jira/browse/HIVE-8456
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Reporter: Chengxiang Li
Assignee: Chengxiang Li
  Labels: Spark-M3
 Attachments: HIVE-8456.1-spark.patch, HIVE-8456.2-spark.patch, 
 HIVE-8456.3-spark.patch


 Several Hive query metric in Hive operators is collected by Hive Counter, 
 such as CREATEDFILES and DESERIALIZE_ERRORS, Besides, Hive use Counter as an 
 option to collect table stats info.  Spark support Accumulator which is 
 pretty similiar with Hive Counter, we could try to enable Hive Counter based 
 on it.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8456) Support Hive Counter to collect spark job metric[Spark Branch]

2014-10-16 Thread Rui Li (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8456?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14173414#comment-14173414
 ] 

Rui Li commented on HIVE-8456:
--

[~chengxiang li] - thanks for the explanation!

I agree we don't have to identify all the needed counters for now.
For #3, I don't see hive create counters for those enums. So do you mean it's 
an improvement to add counters for them on spark?

 Support Hive Counter to collect spark job metric[Spark Branch]
 --

 Key: HIVE-8456
 URL: https://issues.apache.org/jira/browse/HIVE-8456
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Reporter: Chengxiang Li
Assignee: Chengxiang Li
  Labels: Spark-M3
 Attachments: HIVE-8456.1-spark.patch, HIVE-8456.2-spark.patch, 
 HIVE-8456.3-spark.patch


 Several Hive query metric in Hive operators is collected by Hive Counter, 
 such as CREATEDFILES and DESERIALIZE_ERRORS, Besides, Hive use Counter as an 
 option to collect table stats info.  Spark support Accumulator which is 
 pretty similiar with Hive Counter, we could try to enable Hive Counter based 
 on it.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HIVE-8482) hive HashTableLoader error

2014-10-16 Thread q79969786 (JIRA)
q79969786 created HIVE-8482:
---

 Summary: hive HashTableLoader error
 Key: HIVE-8482
 URL: https://issues.apache.org/jira/browse/HIVE-8482
 Project: Hive
  Issue Type: Bug
Reporter: q79969786


java.lang.RuntimeException: org.apache.hadoop.hive.ql.metadata.HiveException: 
java.io.FileNotFoundException: 
/data/data01/data/tmp/hadoop-hadoop/mapred/local/taskTracker/hduser0401/distcache/4554609282698230126_-1275563112_394094114/CNSH041566.app.paic.com.cn/tmp/hduser0401/hive_2014-10-16_13-42-26_451_2739170706811325303-1/-mr-10024/HashTable-Stage-31/Stage-31.tar.gz/MapJoin-mapfile21--.hashtable
 (No such file or directory)
at org.apache.hadoop.hive.ql.exec.mr.ExecMapper.map(ExecMapper.java:195)
at org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:50)
at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:430)
at org.apache.hadoop.mapred.MapTask.run(MapTask.java:366)
at org.apache.hadoop.mapred.Child$4.run(Child.java:255)
at java.security.AccessController.doPrivileged(Native Method)
at javax.security.auth.Subject.doAs(Subject.java:396)
at 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1190)
at org.apache.hadoop.mapred.Child.main(Child.java:249)
Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: 
java.io.FileNotFoundException: 
/data/data01/data/tmp/hadoop-hadoop/mapred/local/taskTracker/hduser0401/distcache/4554609282698230126_-1275563112_394094114/CNSH041566.app.paic.com.cn/tmp/hduser0401/hive_2014-10-16_13-42-26_451_2739170706811325303-1/-mr-10024/HashTable-Stage-31/Stage-31.tar.gz/MapJoin-mapfile21--.hashtable
 (No such file or directory)
at 
org.apache.hadoop.hive.ql.exec.mr.HashTableLoader.load(HashTableLoader.java:104)
at 
org.apache.hadoop.hive.ql.exec.MapJoinOperator.loadHashTable(MapJoinOperator.java:152)
at 
org.apache.hadoop.hive.ql.exec.MapJoinOperator.cleanUpInputFileChangedOp(MapJoinOperator.java:178)
at 
org.apache.hadoop.hive.ql.exec.Operator.cleanUpInputFileChanged(Operator.java:1029)
at 
org.apache.hadoop.hive.ql.exec.Operator.cleanUpInputFileChanged(Operator.java:1033)
at 
org.apache.hadoop.hive.ql.exec.Operator.cleanUpInputFileChanged(Operator.java:1033)
at 
org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:505)
at org.apache.hadoop.hive.ql.exec.mr.ExecMapper.map(ExecMapper.java:177)
... 8 more
Caused by: java.io.FileNotFoundException: 
/data/data01/data/tmp/hadoop-hadoop/mapred/local/taskTracker/hduser0401/distcache/4554609282698230126_-1275563112_394094114/CNSH041566.app.paic.com.cn/tmp/hduser0401/hive_2014-10-16_13-42-26_451_2739170706811325303-1/-mr-10024/HashTable-Stage-31/Stage-31.tar.gz/MapJoin-mapfile21--.hashtable
 (No such file or directory)
at java.io.FileInputStream.open(Native Method)
at java.io.FileInputStream.init(FileInputStream.java:120)
at java.io.FileInputStream.init(FileInputStream.java:79)
at 
org.apache.hadoop.hive.ql.exec.mr.HashTableLoader.load(HashTableLoader.java:95)
... 15 more


FAILED: Execution Error, return code 2 from 
org.apache.hadoop.hive.ql.exec.mr.MapRedTask



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-2573) Create per-session function registry

2014-10-16 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-2573?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14173431#comment-14173431
 ] 

Hive QA commented on HIVE-2573:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12674959/HIVE-2573.7.patch

{color:red}ERROR:{color} -1 due to 5 failed/errored test(s), 6559 tests executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestHBaseCliDriver.testCliDriver_hbase_custom_key
org.apache.hadoop.hive.cli.TestHBaseCliDriver.testCliDriver_hbase_custom_key2
org.apache.hadoop.hive.cli.TestHBaseCliDriver.testCliDriver_hbase_custom_key3
org.apache.hive.beeline.TestSchemaTool.testSchemaInit
org.apache.hive.beeline.TestSchemaTool.testSchemaUpgrade
{noformat}

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/1292/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/1292/console
Test logs: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-1292/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 5 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12674959
 - PreCommit-HIVE-TRUNK-Build

 Create per-session function registry 
 -

 Key: HIVE-2573
 URL: https://issues.apache.org/jira/browse/HIVE-2573
 Project: Hive
  Issue Type: Improvement
  Components: Server Infrastructure
Reporter: Navis
Assignee: Navis
Priority: Minor
 Attachments: ASF.LICENSE.NOT.GRANTED--HIVE-2573.D3231.1.patch, 
 HIVE-2573.1.patch.txt, HIVE-2573.2.patch.txt, HIVE-2573.3.patch.txt, 
 HIVE-2573.4.patch.txt, HIVE-2573.5.patch, HIVE-2573.6.patch, HIVE-2573.7.patch


 Currently the function registry is shared resource and could be overrided by 
 other users when using HiveServer. If per-session function registry is 
 provided, this situation could be prevented.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8456) Support Hive Counter to collect spark job metric[Spark Branch]

2014-10-16 Thread Chengxiang Li (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8456?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14173437#comment-14173437
 ] 

Chengxiang Li commented on HIVE-8456:
-

I haven't look into the details for these enums, we can register these counters 
while enable the feature depends on these enums. I only submit Spark counter 
implementation here, and left the usage of spark counter in the following 
tickets.

 Support Hive Counter to collect spark job metric[Spark Branch]
 --

 Key: HIVE-8456
 URL: https://issues.apache.org/jira/browse/HIVE-8456
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Reporter: Chengxiang Li
Assignee: Chengxiang Li
  Labels: Spark-M3
 Attachments: HIVE-8456.1-spark.patch, HIVE-8456.2-spark.patch, 
 HIVE-8456.3-spark.patch


 Several Hive query metric in Hive operators is collected by Hive Counter, 
 such as CREATEDFILES and DESERIALIZE_ERRORS, Besides, Hive use Counter as an 
 option to collect table stats info.  Spark support Accumulator which is 
 pretty similiar with Hive Counter, we could try to enable Hive Counter based 
 on it.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8456) Support Hive Counter to collect spark job metric[Spark Branch]

2014-10-16 Thread Rui Li (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8456?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14173443#comment-14173443
 ] 

Rui Li commented on HIVE-8456:
--

I see. That makes sense.

+1 The patch looks good to me.

 Support Hive Counter to collect spark job metric[Spark Branch]
 --

 Key: HIVE-8456
 URL: https://issues.apache.org/jira/browse/HIVE-8456
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Reporter: Chengxiang Li
Assignee: Chengxiang Li
  Labels: Spark-M3
 Attachments: HIVE-8456.1-spark.patch, HIVE-8456.2-spark.patch, 
 HIVE-8456.3-spark.patch


 Several Hive query metric in Hive operators is collected by Hive Counter, 
 such as CREATEDFILES and DESERIALIZE_ERRORS, Besides, Hive use Counter as an 
 option to collect table stats info.  Spark support Accumulator which is 
 pretty similiar with Hive Counter, we could try to enable Hive Counter based 
 on it.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8406) Research on skewed join [Spark Branch]

2014-10-16 Thread Rui Li (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8406?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14173449#comment-14173449
 ] 

Rui Li commented on HIVE-8406:
--

Skew join optimization depends on map join.

 Research on skewed join [Spark Branch]
 --

 Key: HIVE-8406
 URL: https://issues.apache.org/jira/browse/HIVE-8406
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Reporter: Xuefu Zhang
Assignee: Rui Li

 Research on how to handle skewed join for hive on spark. Here is original 
 hive's design doc for skewed join, 
 https://cwiki.apache.org/confluence/display/Hive/Skewed+Join+Optimization.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-8320) Error in MetaException(message:Got exception: org.apache.thrift.transport.TTransportException java.net.SocketTimeoutException: Read timed out)

2014-10-16 Thread gavin kim (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-8320?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

gavin kim updated HIVE-8320:

Attachment: (was: 
0001-make-to-synchronize-hiveserver2-session-s-metastore-.patch)

 Error in MetaException(message:Got exception: 
 org.apache.thrift.transport.TTransportException 
 java.net.SocketTimeoutException: Read timed out)
 --

 Key: HIVE-8320
 URL: https://issues.apache.org/jira/browse/HIVE-8320
 Project: Hive
  Issue Type: Improvement
  Components: HiveServer2
Affects Versions: 0.13.1
Reporter: gavin kim
Assignee: gavin kim
Priority: Minor
  Labels: patch
 Fix For: 0.13.1

 Attachments: HIVE-8320.1.patch


 I'm using Hive 13.1 in cdh environment.
 Using hue's beeswax, sometimes hiveserver2 occur MetaException.
 And after that, hive meta data request timed out.
 error log's detail is below.
 2014-09-29 12:05:44,829 ERROR hive.log: Got exception: 
 org.apache.thrift.transport.TTransportException 
 java.net.SocketTimeoutException: Read timed out
 org.apache.thrift.transport.TTransportException: 
 java.net.SocketTimeoutException: Read timed out
 at 
 org.apache.thrift.transport.TIOStreamTransport.read(TIOStreamTransport.java:129)
 at org.apache.thrift.transport.TTransport.readAll(TTransport.java:84)
 at 
 org.apache.thrift.protocol.TBinaryProtocol.readAll(TBinaryProtocol.java:378)
 at 
 org.apache.thrift.protocol.TBinaryProtocol.readI32(TBinaryProtocol.java:297)
 at 
 org.apache.thrift.protocol.TBinaryProtocol.readMessageBegin(TBinaryProtocol.java:204)
 at 
 org.apache.thrift.TServiceClient.receiveBase(TServiceClient.java:69)
 at 
 org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.recv_get_databases(ThriftHiveMetastore.java:600)
 at 
 org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.get_databases(ThriftHiveMetastore.java:587)
 at 
 org.apache.hadoop.hive.metastore.HiveMetaStoreClient.getDatabases(HiveMetaStoreClient.java:826)
 at 
 org.apache.hive.service.cli.operation.GetSchemasOperation.run(GetSchemasOperation.java:62)
 at 
 org.apache.hive.service.cli.session.HiveSessionImpl.runOperationWithLogCapture(HiveSessionImpl.java:562)
 at 
 org.apache.hive.service.cli.session.HiveSessionImpl.getSchemas(HiveSessionImpl.java:315)
 at sun.reflect.GeneratedMethodAccessor9.invoke(Unknown Source)
 at 
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
 at java.lang.reflect.Method.invoke(Method.java:606)
 at 
 org.apache.hive.service.cli.session.HiveSessionProxy.invoke(HiveSessionProxy.java:79)
 at 
 org.apache.hive.service.cli.session.HiveSessionProxy.access$000(HiveSessionProxy.java:37)
 at 
 org.apache.hive.service.cli.session.HiveSessionProxy$1.run(HiveSessionProxy.java:64)
 at java.security.AccessController.doPrivileged(Native Method)
 at javax.security.auth.Subject.doAs(Subject.java:415)
 at 
 org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1548)
 at 
 org.apache.hadoop.hive.shims.HadoopShimsSecure.doAs(HadoopShimsSecure.java:493)
 at 
 org.apache.hive.service.cli.session.HiveSessionProxy.invoke(HiveSessionProxy.java:60)
 at com.sun.proxy.$Proxy13.getSchemas(Unknown Source)
 at 
 org.apache.hive.service.cli.CLIService.getSchemas(CLIService.java:273)
 at 
 org.apache.hive.service.cli.thrift.ThriftCLIService.GetSchemas(ThriftCLIService.java:402)
 at 
 org.apache.hive.service.cli.thrift.TCLIService$Processor$GetSchemas.getResult(TCLIService.java:1429)
 at 
 org.apache.hive.service.cli.thrift.TCLIService$Processor$GetSchemas.getResult(TCLIService.java:1414)
 at org.apache.thrift.ProcessFunction.process(ProcessFunction.java:39)
 at org.apache.thrift.TBaseProcessor.process(TBaseProcessor.java:39)
 at 
 org.apache.hive.service.auth.TSetIpAddressProcessor.process(TSetIpAddressProcessor.java:55)
 at 
 org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPoolServer.java:206)
 at 
 java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
 at 
 java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
 at java.lang.Thread.run(Thread.java:745)
 Caused by: java.net.SocketTimeoutException: Read timed out
 at java.net.SocketInputStream.socketRead0(Native Method)
 at java.net.SocketInputStream.read(SocketInputStream.java:152)
 at java.net.SocketInputStream.read(SocketInputStream.java:122)
 at 

[jira] [Updated] (HIVE-8320) Error in MetaException(message:Got exception: org.apache.thrift.transport.TTransportException java.net.SocketTimeoutException: Read timed out)

2014-10-16 Thread gavin kim (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-8320?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

gavin kim updated HIVE-8320:

Attachment: HIVE-8320.2.patch

I'm done my code.
But, I'm confused how to use review board.

Is it right posted in review board automatically when i submit patch in jira??

 Error in MetaException(message:Got exception: 
 org.apache.thrift.transport.TTransportException 
 java.net.SocketTimeoutException: Read timed out)
 --

 Key: HIVE-8320
 URL: https://issues.apache.org/jira/browse/HIVE-8320
 Project: Hive
  Issue Type: Improvement
  Components: HiveServer2
Affects Versions: 0.13.1
Reporter: gavin kim
Assignee: gavin kim
Priority: Minor
  Labels: patch
 Fix For: 0.13.1

 Attachments: HIVE-8320.1.patch, HIVE-8320.2.patch


 I'm using Hive 13.1 in cdh environment.
 Using hue's beeswax, sometimes hiveserver2 occur MetaException.
 And after that, hive meta data request timed out.
 error log's detail is below.
 2014-09-29 12:05:44,829 ERROR hive.log: Got exception: 
 org.apache.thrift.transport.TTransportException 
 java.net.SocketTimeoutException: Read timed out
 org.apache.thrift.transport.TTransportException: 
 java.net.SocketTimeoutException: Read timed out
 at 
 org.apache.thrift.transport.TIOStreamTransport.read(TIOStreamTransport.java:129)
 at org.apache.thrift.transport.TTransport.readAll(TTransport.java:84)
 at 
 org.apache.thrift.protocol.TBinaryProtocol.readAll(TBinaryProtocol.java:378)
 at 
 org.apache.thrift.protocol.TBinaryProtocol.readI32(TBinaryProtocol.java:297)
 at 
 org.apache.thrift.protocol.TBinaryProtocol.readMessageBegin(TBinaryProtocol.java:204)
 at 
 org.apache.thrift.TServiceClient.receiveBase(TServiceClient.java:69)
 at 
 org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.recv_get_databases(ThriftHiveMetastore.java:600)
 at 
 org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.get_databases(ThriftHiveMetastore.java:587)
 at 
 org.apache.hadoop.hive.metastore.HiveMetaStoreClient.getDatabases(HiveMetaStoreClient.java:826)
 at 
 org.apache.hive.service.cli.operation.GetSchemasOperation.run(GetSchemasOperation.java:62)
 at 
 org.apache.hive.service.cli.session.HiveSessionImpl.runOperationWithLogCapture(HiveSessionImpl.java:562)
 at 
 org.apache.hive.service.cli.session.HiveSessionImpl.getSchemas(HiveSessionImpl.java:315)
 at sun.reflect.GeneratedMethodAccessor9.invoke(Unknown Source)
 at 
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
 at java.lang.reflect.Method.invoke(Method.java:606)
 at 
 org.apache.hive.service.cli.session.HiveSessionProxy.invoke(HiveSessionProxy.java:79)
 at 
 org.apache.hive.service.cli.session.HiveSessionProxy.access$000(HiveSessionProxy.java:37)
 at 
 org.apache.hive.service.cli.session.HiveSessionProxy$1.run(HiveSessionProxy.java:64)
 at java.security.AccessController.doPrivileged(Native Method)
 at javax.security.auth.Subject.doAs(Subject.java:415)
 at 
 org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1548)
 at 
 org.apache.hadoop.hive.shims.HadoopShimsSecure.doAs(HadoopShimsSecure.java:493)
 at 
 org.apache.hive.service.cli.session.HiveSessionProxy.invoke(HiveSessionProxy.java:60)
 at com.sun.proxy.$Proxy13.getSchemas(Unknown Source)
 at 
 org.apache.hive.service.cli.CLIService.getSchemas(CLIService.java:273)
 at 
 org.apache.hive.service.cli.thrift.ThriftCLIService.GetSchemas(ThriftCLIService.java:402)
 at 
 org.apache.hive.service.cli.thrift.TCLIService$Processor$GetSchemas.getResult(TCLIService.java:1429)
 at 
 org.apache.hive.service.cli.thrift.TCLIService$Processor$GetSchemas.getResult(TCLIService.java:1414)
 at org.apache.thrift.ProcessFunction.process(ProcessFunction.java:39)
 at org.apache.thrift.TBaseProcessor.process(TBaseProcessor.java:39)
 at 
 org.apache.hive.service.auth.TSetIpAddressProcessor.process(TSetIpAddressProcessor.java:55)
 at 
 org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPoolServer.java:206)
 at 
 java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
 at 
 java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
 at java.lang.Thread.run(Thread.java:745)
 Caused by: java.net.SocketTimeoutException: Read timed out
 at java.net.SocketInputStream.socketRead0(Native Method)
 at 

[jira] [Commented] (HIVE-8467) Table Copy - Background, incremental data load

2014-10-16 Thread Gunther Hagleitner (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8467?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14173455#comment-14173455
 ] 

Gunther Hagleitner commented on HIVE-8467:
--

Materialized views don't necessarily have to keep the tables in sync, do they? 
Other vendors allow deferred refreshes and for the user to specify integrity 
levels. I.e.: You can still put the onus on the user and you don't necessarily 
have to offer a background sync method (you can choose to additional options 
later.)

As far as other engines go - you have the same problem right? You can expose 
the table copy or view, but the smarts how and when to rewrite queries has to 
be built into each of those, or left to the user. With materialized views, 
other engines will also know how the tables are derived, which seems beneficial 
(well, if they speak SQL at least). For Pig and MR you will likely have to bake 
assumptions into the scripts/code.

Could say more about retention policy, max size and in general how you have 
seen ppl choose which partitions to add to the table copy? Is it typically the 
newest n partition? Or the last month of data? That'd be interesting - to see 
if it can be mapped on materialized views and how hard it'd be for the CBO can 
handle it.


 Table Copy - Background, incremental data load
 --

 Key: HIVE-8467
 URL: https://issues.apache.org/jira/browse/HIVE-8467
 Project: Hive
  Issue Type: New Feature
Reporter: Rajat Venkatesh
 Attachments: Table Copies.pdf


 Traditionally, Hive and other tools in the Hadoop eco-system havent required 
 a load stage. However, with recent developments, Hive is much more performant 
 when data is stored in specific formats like ORC, Parquet, Avro etc. 
 Technologies like Presto, also work much better with certain data formats. At 
 the same time, data is generated or obtained from 3rd parties in non-optimal 
 formats such as CSV, tab-limited or JSON. Many a times, its not an option to 
 change the data format at the source. We've found that users either use 
 sub-optimal formats or spend a large amount of effort creating and 
 maintaining copies. We want to propose a new construct - Table Copy - to help 
 “load” data into an optimal storage format.
 I am going to attach a PDF document with a lot more details especially 
 addressing how is this different from bulk loads in relational DBs or 
 materialized views.
 Looking forward to hear if others see a similar need to formalize conversion 
 of data to different storage formats.  If yes, are the details in the PDF 
 document a good start ?



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-8320) Error in MetaException(message:Got exception: org.apache.thrift.transport.TTransportException java.net.SocketTimeoutException: Read timed out)

2014-10-16 Thread gavin kim (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-8320?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

gavin kim updated HIVE-8320:

Status: Patch Available  (was: Open)

 Error in MetaException(message:Got exception: 
 org.apache.thrift.transport.TTransportException 
 java.net.SocketTimeoutException: Read timed out)
 --

 Key: HIVE-8320
 URL: https://issues.apache.org/jira/browse/HIVE-8320
 Project: Hive
  Issue Type: Improvement
  Components: HiveServer2
Affects Versions: 0.13.1
Reporter: gavin kim
Assignee: gavin kim
Priority: Minor
  Labels: patch
 Fix For: 0.13.1

 Attachments: HIVE-8320.1.patch, HIVE-8320.2.patch


 I'm using Hive 13.1 in cdh environment.
 Using hue's beeswax, sometimes hiveserver2 occur MetaException.
 And after that, hive meta data request timed out.
 error log's detail is below.
 2014-09-29 12:05:44,829 ERROR hive.log: Got exception: 
 org.apache.thrift.transport.TTransportException 
 java.net.SocketTimeoutException: Read timed out
 org.apache.thrift.transport.TTransportException: 
 java.net.SocketTimeoutException: Read timed out
 at 
 org.apache.thrift.transport.TIOStreamTransport.read(TIOStreamTransport.java:129)
 at org.apache.thrift.transport.TTransport.readAll(TTransport.java:84)
 at 
 org.apache.thrift.protocol.TBinaryProtocol.readAll(TBinaryProtocol.java:378)
 at 
 org.apache.thrift.protocol.TBinaryProtocol.readI32(TBinaryProtocol.java:297)
 at 
 org.apache.thrift.protocol.TBinaryProtocol.readMessageBegin(TBinaryProtocol.java:204)
 at 
 org.apache.thrift.TServiceClient.receiveBase(TServiceClient.java:69)
 at 
 org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.recv_get_databases(ThriftHiveMetastore.java:600)
 at 
 org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.get_databases(ThriftHiveMetastore.java:587)
 at 
 org.apache.hadoop.hive.metastore.HiveMetaStoreClient.getDatabases(HiveMetaStoreClient.java:826)
 at 
 org.apache.hive.service.cli.operation.GetSchemasOperation.run(GetSchemasOperation.java:62)
 at 
 org.apache.hive.service.cli.session.HiveSessionImpl.runOperationWithLogCapture(HiveSessionImpl.java:562)
 at 
 org.apache.hive.service.cli.session.HiveSessionImpl.getSchemas(HiveSessionImpl.java:315)
 at sun.reflect.GeneratedMethodAccessor9.invoke(Unknown Source)
 at 
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
 at java.lang.reflect.Method.invoke(Method.java:606)
 at 
 org.apache.hive.service.cli.session.HiveSessionProxy.invoke(HiveSessionProxy.java:79)
 at 
 org.apache.hive.service.cli.session.HiveSessionProxy.access$000(HiveSessionProxy.java:37)
 at 
 org.apache.hive.service.cli.session.HiveSessionProxy$1.run(HiveSessionProxy.java:64)
 at java.security.AccessController.doPrivileged(Native Method)
 at javax.security.auth.Subject.doAs(Subject.java:415)
 at 
 org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1548)
 at 
 org.apache.hadoop.hive.shims.HadoopShimsSecure.doAs(HadoopShimsSecure.java:493)
 at 
 org.apache.hive.service.cli.session.HiveSessionProxy.invoke(HiveSessionProxy.java:60)
 at com.sun.proxy.$Proxy13.getSchemas(Unknown Source)
 at 
 org.apache.hive.service.cli.CLIService.getSchemas(CLIService.java:273)
 at 
 org.apache.hive.service.cli.thrift.ThriftCLIService.GetSchemas(ThriftCLIService.java:402)
 at 
 org.apache.hive.service.cli.thrift.TCLIService$Processor$GetSchemas.getResult(TCLIService.java:1429)
 at 
 org.apache.hive.service.cli.thrift.TCLIService$Processor$GetSchemas.getResult(TCLIService.java:1414)
 at org.apache.thrift.ProcessFunction.process(ProcessFunction.java:39)
 at org.apache.thrift.TBaseProcessor.process(TBaseProcessor.java:39)
 at 
 org.apache.hive.service.auth.TSetIpAddressProcessor.process(TSetIpAddressProcessor.java:55)
 at 
 org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPoolServer.java:206)
 at 
 java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
 at 
 java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
 at java.lang.Thread.run(Thread.java:745)
 Caused by: java.net.SocketTimeoutException: Read timed out
 at java.net.SocketInputStream.socketRead0(Native Method)
 at java.net.SocketInputStream.read(SocketInputStream.java:152)
 at java.net.SocketInputStream.read(SocketInputStream.java:122)
 at 

[jira] [Commented] (HIVE-8320) Error in MetaException(message:Got exception: org.apache.thrift.transport.TTransportException java.net.SocketTimeoutException: Read timed out)

2014-10-16 Thread gavin kim (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8320?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14173463#comment-14173463
 ] 

gavin kim commented on HIVE-8320:
-

I posted patch manullay. 

The url is https://reviews.apache.org/r/26805/

How can i get that there is no problem in code convention??




 Error in MetaException(message:Got exception: 
 org.apache.thrift.transport.TTransportException 
 java.net.SocketTimeoutException: Read timed out)
 --

 Key: HIVE-8320
 URL: https://issues.apache.org/jira/browse/HIVE-8320
 Project: Hive
  Issue Type: Improvement
  Components: HiveServer2
Affects Versions: 0.13.1
Reporter: gavin kim
Assignee: gavin kim
Priority: Minor
  Labels: patch
 Fix For: 0.13.1

 Attachments: HIVE-8320.1.patch, HIVE-8320.2.patch


 I'm using Hive 13.1 in cdh environment.
 Using hue's beeswax, sometimes hiveserver2 occur MetaException.
 And after that, hive meta data request timed out.
 error log's detail is below.
 2014-09-29 12:05:44,829 ERROR hive.log: Got exception: 
 org.apache.thrift.transport.TTransportException 
 java.net.SocketTimeoutException: Read timed out
 org.apache.thrift.transport.TTransportException: 
 java.net.SocketTimeoutException: Read timed out
 at 
 org.apache.thrift.transport.TIOStreamTransport.read(TIOStreamTransport.java:129)
 at org.apache.thrift.transport.TTransport.readAll(TTransport.java:84)
 at 
 org.apache.thrift.protocol.TBinaryProtocol.readAll(TBinaryProtocol.java:378)
 at 
 org.apache.thrift.protocol.TBinaryProtocol.readI32(TBinaryProtocol.java:297)
 at 
 org.apache.thrift.protocol.TBinaryProtocol.readMessageBegin(TBinaryProtocol.java:204)
 at 
 org.apache.thrift.TServiceClient.receiveBase(TServiceClient.java:69)
 at 
 org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.recv_get_databases(ThriftHiveMetastore.java:600)
 at 
 org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.get_databases(ThriftHiveMetastore.java:587)
 at 
 org.apache.hadoop.hive.metastore.HiveMetaStoreClient.getDatabases(HiveMetaStoreClient.java:826)
 at 
 org.apache.hive.service.cli.operation.GetSchemasOperation.run(GetSchemasOperation.java:62)
 at 
 org.apache.hive.service.cli.session.HiveSessionImpl.runOperationWithLogCapture(HiveSessionImpl.java:562)
 at 
 org.apache.hive.service.cli.session.HiveSessionImpl.getSchemas(HiveSessionImpl.java:315)
 at sun.reflect.GeneratedMethodAccessor9.invoke(Unknown Source)
 at 
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
 at java.lang.reflect.Method.invoke(Method.java:606)
 at 
 org.apache.hive.service.cli.session.HiveSessionProxy.invoke(HiveSessionProxy.java:79)
 at 
 org.apache.hive.service.cli.session.HiveSessionProxy.access$000(HiveSessionProxy.java:37)
 at 
 org.apache.hive.service.cli.session.HiveSessionProxy$1.run(HiveSessionProxy.java:64)
 at java.security.AccessController.doPrivileged(Native Method)
 at javax.security.auth.Subject.doAs(Subject.java:415)
 at 
 org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1548)
 at 
 org.apache.hadoop.hive.shims.HadoopShimsSecure.doAs(HadoopShimsSecure.java:493)
 at 
 org.apache.hive.service.cli.session.HiveSessionProxy.invoke(HiveSessionProxy.java:60)
 at com.sun.proxy.$Proxy13.getSchemas(Unknown Source)
 at 
 org.apache.hive.service.cli.CLIService.getSchemas(CLIService.java:273)
 at 
 org.apache.hive.service.cli.thrift.ThriftCLIService.GetSchemas(ThriftCLIService.java:402)
 at 
 org.apache.hive.service.cli.thrift.TCLIService$Processor$GetSchemas.getResult(TCLIService.java:1429)
 at 
 org.apache.hive.service.cli.thrift.TCLIService$Processor$GetSchemas.getResult(TCLIService.java:1414)
 at org.apache.thrift.ProcessFunction.process(ProcessFunction.java:39)
 at org.apache.thrift.TBaseProcessor.process(TBaseProcessor.java:39)
 at 
 org.apache.hive.service.auth.TSetIpAddressProcessor.process(TSetIpAddressProcessor.java:55)
 at 
 org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPoolServer.java:206)
 at 
 java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
 at 
 java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
 at java.lang.Thread.run(Thread.java:745)
 Caused by: java.net.SocketTimeoutException: Read timed out
 at java.net.SocketInputStream.socketRead0(Native Method)
 at 

Review Request 26805: HIVE-8320: edit hiveserver2 session's metastore client to use ThreadLocal client

2014-10-16 Thread Gavin Kim

---
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/26805/
---

Review request for hive and Thejas Nair.


Bugs: HIVE-8320
https://issues.apache.org/jira/browse/HIVE-8320


Repository: hive-git


Description
---

HIVE-8320: edit hiveserver2 session's metastore client to use ThreadLocal client


Diffs
-

  service/src/java/org/apache/hive/service/cli/session/HiveSessionImpl.java 
a9d5902 

Diff: https://reviews.apache.org/r/26805/diff/


Testing
---


Thanks,

Gavin Kim



[jira] [Updated] (HIVE-8465) Fix some minor test fails on trunk

2014-10-16 Thread Navis (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-8465?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Navis updated HIVE-8465:

Description: 
org.apache.hive.beeline.TestSchemaTool.testSchemaInit
org.apache.hive.beeline.TestSchemaTool.testSchemaUpgrade
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_parallel

-- due to HIVE-2828
org.apache.hadoop.hive.cli.TestHBaseCliDriver.testCliDriver_hbase_custom_key
org.apache.hadoop.hive.cli.TestHBaseCliDriver.testCliDriver_hbase_custom_key2
org.apache.hadoop.hive.cli.TestHBaseCliDriver.testCliDriver_hbase_custom_key3


  was:
org.apache.hive.beeline.TestSchemaTool.testSchemaInit
org.apache.hive.beeline.TestSchemaTool.testSchemaUpgrade
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_parallel


 Fix some minor test fails on trunk
 --

 Key: HIVE-8465
 URL: https://issues.apache.org/jira/browse/HIVE-8465
 Project: Hive
  Issue Type: Task
  Components: Tests
Reporter: Navis
Assignee: Jason Dere
Priority: Minor
 Attachments: HIVE-8465.1.patch.txt, HIVE-8465.2.patch


 org.apache.hive.beeline.TestSchemaTool.testSchemaInit
 org.apache.hive.beeline.TestSchemaTool.testSchemaUpgrade
 org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_parallel
 -- due to HIVE-2828
 org.apache.hadoop.hive.cli.TestHBaseCliDriver.testCliDriver_hbase_custom_key
 org.apache.hadoop.hive.cli.TestHBaseCliDriver.testCliDriver_hbase_custom_key2
 org.apache.hadoop.hive.cli.TestHBaseCliDriver.testCliDriver_hbase_custom_key3



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-2828) make timestamp accessible in the hbase KeyValue

2014-10-16 Thread Navis (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-2828?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14173468#comment-14173468
 ] 

Navis commented on HIVE-2828:
-

My bad. Some codes not in the patch is included by mistake, failing 
hbase_custom_key* tests. I'll merge patch for that with HIVE-8465.

 make timestamp accessible in the hbase KeyValue 
 

 Key: HIVE-2828
 URL: https://issues.apache.org/jira/browse/HIVE-2828
 Project: Hive
  Issue Type: Improvement
  Components: HBase Handler
Reporter: Navis
Assignee: Navis
Priority: Trivial
 Fix For: 0.15.0

 Attachments: ASF.LICENSE.NOT.GRANTED--HIVE-2828.D1989.1.patch, 
 ASF.LICENSE.NOT.GRANTED--HIVE-2828.D1989.2.patch, 
 ASF.LICENSE.NOT.GRANTED--HIVE-2828.D1989.3.patch, 
 ASF.LICENSE.NOT.GRANTED--HIVE-2828.D1989.4.patch, 
 ASF.LICENSE.NOT.GRANTED--HIVE-2828.D1989.5.patch, HIVE-2828.6.patch.txt, 
 HIVE-2828.7.patch.txt, HIVE-2828.8.patch.txt, HIVE-2828.9.patch.txt


 Originated from HIVE-2781 and not accepted, but I think this could be helpful 
 to someone.
 By using special column notation ':timestamp' in HBASE_COLUMNS_MAPPING, user 
 might access timestamp value in hbase KeyValue.
 {code}
 CREATE TABLE hbase_table (key int, value string, time timestamp)
   STORED BY 'org.apache.hadoop.hive.hbase.HBaseStorageHandler'
   WITH SERDEPROPERTIES (hbase.columns.mapping = :key,cf:string,:timestamp)
 {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-8320) Error in MetaException(message:Got exception: org.apache.thrift.transport.TTransportException java.net.SocketTimeoutException: Read timed out)

2014-10-16 Thread gavin kim (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-8320?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

gavin kim updated HIVE-8320:

Attachment: 0001-make-to-synchronize-hiveserver2-session-s-metastore-.patch

 Error in MetaException(message:Got exception: 
 org.apache.thrift.transport.TTransportException 
 java.net.SocketTimeoutException: Read timed out)
 --

 Key: HIVE-8320
 URL: https://issues.apache.org/jira/browse/HIVE-8320
 Project: Hive
  Issue Type: Improvement
  Components: HiveServer2
Affects Versions: 0.13.1
Reporter: gavin kim
Assignee: gavin kim
Priority: Minor
  Labels: patch
 Fix For: 0.13.1

 Attachments: 
 0001-make-to-synchronize-hiveserver2-session-s-metastore-.patch, 
 HIVE-8320.1.patch, HIVE-8320.2.patch


 I'm using Hive 13.1 in cdh environment.
 Using hue's beeswax, sometimes hiveserver2 occur MetaException.
 And after that, hive meta data request timed out.
 error log's detail is below.
 2014-09-29 12:05:44,829 ERROR hive.log: Got exception: 
 org.apache.thrift.transport.TTransportException 
 java.net.SocketTimeoutException: Read timed out
 org.apache.thrift.transport.TTransportException: 
 java.net.SocketTimeoutException: Read timed out
 at 
 org.apache.thrift.transport.TIOStreamTransport.read(TIOStreamTransport.java:129)
 at org.apache.thrift.transport.TTransport.readAll(TTransport.java:84)
 at 
 org.apache.thrift.protocol.TBinaryProtocol.readAll(TBinaryProtocol.java:378)
 at 
 org.apache.thrift.protocol.TBinaryProtocol.readI32(TBinaryProtocol.java:297)
 at 
 org.apache.thrift.protocol.TBinaryProtocol.readMessageBegin(TBinaryProtocol.java:204)
 at 
 org.apache.thrift.TServiceClient.receiveBase(TServiceClient.java:69)
 at 
 org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.recv_get_databases(ThriftHiveMetastore.java:600)
 at 
 org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.get_databases(ThriftHiveMetastore.java:587)
 at 
 org.apache.hadoop.hive.metastore.HiveMetaStoreClient.getDatabases(HiveMetaStoreClient.java:826)
 at 
 org.apache.hive.service.cli.operation.GetSchemasOperation.run(GetSchemasOperation.java:62)
 at 
 org.apache.hive.service.cli.session.HiveSessionImpl.runOperationWithLogCapture(HiveSessionImpl.java:562)
 at 
 org.apache.hive.service.cli.session.HiveSessionImpl.getSchemas(HiveSessionImpl.java:315)
 at sun.reflect.GeneratedMethodAccessor9.invoke(Unknown Source)
 at 
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
 at java.lang.reflect.Method.invoke(Method.java:606)
 at 
 org.apache.hive.service.cli.session.HiveSessionProxy.invoke(HiveSessionProxy.java:79)
 at 
 org.apache.hive.service.cli.session.HiveSessionProxy.access$000(HiveSessionProxy.java:37)
 at 
 org.apache.hive.service.cli.session.HiveSessionProxy$1.run(HiveSessionProxy.java:64)
 at java.security.AccessController.doPrivileged(Native Method)
 at javax.security.auth.Subject.doAs(Subject.java:415)
 at 
 org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1548)
 at 
 org.apache.hadoop.hive.shims.HadoopShimsSecure.doAs(HadoopShimsSecure.java:493)
 at 
 org.apache.hive.service.cli.session.HiveSessionProxy.invoke(HiveSessionProxy.java:60)
 at com.sun.proxy.$Proxy13.getSchemas(Unknown Source)
 at 
 org.apache.hive.service.cli.CLIService.getSchemas(CLIService.java:273)
 at 
 org.apache.hive.service.cli.thrift.ThriftCLIService.GetSchemas(ThriftCLIService.java:402)
 at 
 org.apache.hive.service.cli.thrift.TCLIService$Processor$GetSchemas.getResult(TCLIService.java:1429)
 at 
 org.apache.hive.service.cli.thrift.TCLIService$Processor$GetSchemas.getResult(TCLIService.java:1414)
 at org.apache.thrift.ProcessFunction.process(ProcessFunction.java:39)
 at org.apache.thrift.TBaseProcessor.process(TBaseProcessor.java:39)
 at 
 org.apache.hive.service.auth.TSetIpAddressProcessor.process(TSetIpAddressProcessor.java:55)
 at 
 org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPoolServer.java:206)
 at 
 java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
 at 
 java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
 at java.lang.Thread.run(Thread.java:745)
 Caused by: java.net.SocketTimeoutException: Read timed out
 at java.net.SocketInputStream.socketRead0(Native Method)
 at java.net.SocketInputStream.read(SocketInputStream.java:152)
 at 

[jira] [Updated] (HIVE-2828) make timestamp accessible in the hbase KeyValue

2014-10-16 Thread Lefty Leverenz (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-2828?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Lefty Leverenz updated HIVE-2828:
-
Labels: TODOC15  (was: )

 make timestamp accessible in the hbase KeyValue 
 

 Key: HIVE-2828
 URL: https://issues.apache.org/jira/browse/HIVE-2828
 Project: Hive
  Issue Type: Improvement
  Components: HBase Handler
Reporter: Navis
Assignee: Navis
Priority: Trivial
  Labels: TODOC15
 Fix For: 0.15.0

 Attachments: ASF.LICENSE.NOT.GRANTED--HIVE-2828.D1989.1.patch, 
 ASF.LICENSE.NOT.GRANTED--HIVE-2828.D1989.2.patch, 
 ASF.LICENSE.NOT.GRANTED--HIVE-2828.D1989.3.patch, 
 ASF.LICENSE.NOT.GRANTED--HIVE-2828.D1989.4.patch, 
 ASF.LICENSE.NOT.GRANTED--HIVE-2828.D1989.5.patch, HIVE-2828.6.patch.txt, 
 HIVE-2828.7.patch.txt, HIVE-2828.8.patch.txt, HIVE-2828.9.patch.txt


 Originated from HIVE-2781 and not accepted, but I think this could be helpful 
 to someone.
 By using special column notation ':timestamp' in HBASE_COLUMNS_MAPPING, user 
 might access timestamp value in hbase KeyValue.
 {code}
 CREATE TABLE hbase_table (key int, value string, time timestamp)
   STORED BY 'org.apache.hadoop.hive.hbase.HBaseStorageHandler'
   WITH SERDEPROPERTIES (hbase.columns.mapping = :key,cf:string,:timestamp)
 {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-8465) Fix some minor test fails on trunk

2014-10-16 Thread Navis (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-8465?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Navis updated HIVE-8465:

Attachment: HIVE-8465.3.patch.txt

 Fix some minor test fails on trunk
 --

 Key: HIVE-8465
 URL: https://issues.apache.org/jira/browse/HIVE-8465
 Project: Hive
  Issue Type: Task
  Components: Tests
Reporter: Navis
Assignee: Jason Dere
Priority: Minor
 Attachments: HIVE-8465.1.patch.txt, HIVE-8465.2.patch, 
 HIVE-8465.3.patch.txt


 org.apache.hive.beeline.TestSchemaTool.testSchemaInit
 org.apache.hive.beeline.TestSchemaTool.testSchemaUpgrade
 org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_parallel
 -- due to HIVE-2828
 org.apache.hadoop.hive.cli.TestHBaseCliDriver.testCliDriver_hbase_custom_key
 org.apache.hadoop.hive.cli.TestHBaseCliDriver.testCliDriver_hbase_custom_key2
 org.apache.hadoop.hive.cli.TestHBaseCliDriver.testCliDriver_hbase_custom_key3



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8466) nonReserved keywords can not be used as table alias

2014-10-16 Thread Navis (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8466?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14173482#comment-14173482
 ] 

Navis commented on HIVE-8466:
-

TestSchemaTool is not related to this and missing_overwrite just needs an 
update of golden file. But *dynamic_partition_pruning tests are showing a bug 
in trunk, not recognizing alias of join source. 

 nonReserved keywords can not be used as table alias
 ---

 Key: HIVE-8466
 URL: https://issues.apache.org/jira/browse/HIVE-8466
 Project: Hive
  Issue Type: Bug
  Components: SQL
Affects Versions: 0.12.0, 0.13.0, 0.13.1
Reporter: cw
Priority: Minor
 Attachments: HIVE-8466.1.patch


 There is a small mistake in the patch of issue HIVE-2906. See the change of 
 FromClauseParser.g
 -: tabname=tableName (ts=tableSample)? (KW_AS? alias=identifier)?
 -- ^(TOK_TABREF $tabname $ts? $alias?)
 +: tabname=tableName (props=tableProperties)? (ts=tableSample)? (KW_AS? 
 alias=Identifier)?
 +- ^(TOK_TABREF $tabname $props? $ts? $alias?)
 With the 'identifier' changed to 'Identifier' we can not use nonReserved 
 keywords as table alias.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-2828) make timestamp accessible in the hbase KeyValue

2014-10-16 Thread Lefty Leverenz (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-2828?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14173481#comment-14173481
 ] 

Lefty Leverenz commented on HIVE-2828:
--

Doc note:  This should be documented in the HBase Integration wikidoc, with 
version information, and a release note would be nice.

* [HBase Integration | 
https://cwiki.apache.org/confluence/display/Hive/HBaseIntegration]

 make timestamp accessible in the hbase KeyValue 
 

 Key: HIVE-2828
 URL: https://issues.apache.org/jira/browse/HIVE-2828
 Project: Hive
  Issue Type: Improvement
  Components: HBase Handler
Reporter: Navis
Assignee: Navis
Priority: Trivial
  Labels: TODOC15
 Fix For: 0.15.0

 Attachments: ASF.LICENSE.NOT.GRANTED--HIVE-2828.D1989.1.patch, 
 ASF.LICENSE.NOT.GRANTED--HIVE-2828.D1989.2.patch, 
 ASF.LICENSE.NOT.GRANTED--HIVE-2828.D1989.3.patch, 
 ASF.LICENSE.NOT.GRANTED--HIVE-2828.D1989.4.patch, 
 ASF.LICENSE.NOT.GRANTED--HIVE-2828.D1989.5.patch, HIVE-2828.6.patch.txt, 
 HIVE-2828.7.patch.txt, HIVE-2828.8.patch.txt, HIVE-2828.9.patch.txt


 Originated from HIVE-2781 and not accepted, but I think this could be helpful 
 to someone.
 By using special column notation ':timestamp' in HBASE_COLUMNS_MAPPING, user 
 might access timestamp value in hbase KeyValue.
 {code}
 CREATE TABLE hbase_table (key int, value string, time timestamp)
   STORED BY 'org.apache.hadoop.hive.hbase.HBaseStorageHandler'
   WITH SERDEPROPERTIES (hbase.columns.mapping = :key,cf:string,:timestamp)
 {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


Re: Review Request 26481: HIVE-8411: Support partial partition spec for certain ALTER PARTITION statements

2014-10-16 Thread Gunther Hagleitner

---
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/26481/#review56905
---



ql/src/java/org/apache/hadoop/hive/ql/exec/DDLTask.java
https://reviews.apache.org/r/26481/#comment97340

there's a flag for enabling dynamic partitioning. is this checked somewhere 
in this path? ppl use that to avoid someone accidentally changing everything.



ql/src/java/org/apache/hadoop/hive/ql/exec/DDLTask.java
https://reviews.apache.org/r/26481/#comment97339

why does this fail? it seems more consistent to just alter 0 partitions in 
this case. or at least just print a warning instead of failing.



ql/src/java/org/apache/hadoop/hive/ql/exec/DDLTask.java
https://reviews.apache.org/r/26481/#comment97341

LOG.error?



ql/src/test/queries/clientpositive/alter_partition_change_col.q
https://reviews.apache.org/r/26481/#comment97342

does this support multi level partitioning? if so can you specify only some 
of the partitions?



ql/src/test/queries/clientpositive/alter_partition_change_col.q
https://reviews.apache.org/r/26481/#comment97343

add a query to verify the output?


- Gunther Hagleitner


On Oct. 9, 2014, 3:14 a.m., Jason Dere wrote:
 
 ---
 This is an automatically generated e-mail. To reply, visit:
 https://reviews.apache.org/r/26481/
 ---
 
 (Updated Oct. 9, 2014, 3:14 a.m.)
 
 
 Review request for hive, Gunther Hagleitner and Thejas Nair.
 
 
 Bugs: HIVE-8411
 https://issues.apache.org/jira/browse/HIVE-8411
 
 
 Repository: hive-git
 
 
 Description
 ---
 
 Allow partial partition spec for some alter table statements, such as:
 
 alter table alter_partition_change_col1 partition (p1) change column c2 c2 
 decimal(10,0);
 
 
 Diffs
 -
 
   ql/src/java/org/apache/hadoop/hive/ql/exec/DDLTask.java d5374bc 
   ql/src/java/org/apache/hadoop/hive/ql/parse/DDLSemanticAnalyzer.java 
 4e58ad8 
   ql/src/test/queries/clientpositive/alter_partition_change_col.q baabb9f 
   ql/src/test/results/clientpositive/alter_partition_change_col.q.out 7123e40 
 
 Diff: https://reviews.apache.org/r/26481/diff/
 
 
 Testing
 ---
 
 updated qfile test
 
 
 Thanks,
 
 Jason Dere
 




[jira] [Commented] (HIVE-8448) Union All might not work due to the type conversion issue

2014-10-16 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8448?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14173485#comment-14173485
 ] 

Hive QA commented on HIVE-8448:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12675081/HIVE-8448.4.patch

{color:red}ERROR:{color} -1 due to 6 failed/errored test(s), 6560 tests executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestHBaseCliDriver.testCliDriver_hbase_custom_key
org.apache.hadoop.hive.cli.TestHBaseCliDriver.testCliDriver_hbase_custom_key2
org.apache.hadoop.hive.cli.TestHBaseCliDriver.testCliDriver_hbase_custom_key3
org.apache.hive.beeline.TestSchemaTool.testSchemaInit
org.apache.hive.beeline.TestSchemaTool.testSchemaUpgrade
org.apache.hive.minikdc.TestMiniHiveKdc.testLogin
{noformat}

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/1293/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/1293/console
Test logs: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-1293/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 6 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12675081
 - PreCommit-HIVE-TRUNK-Build

 Union All might not work due to the type conversion issue
 -

 Key: HIVE-8448
 URL: https://issues.apache.org/jira/browse/HIVE-8448
 Project: Hive
  Issue Type: Bug
Affects Versions: 0.13.1
Reporter: Chaoyu Tang
Assignee: Yongzhi Chen
Priority: Minor
 Attachments: HIVE-8448.4.patch


 create table t1 (val date);
 insert overwrite table t1 select '2014-10-10' from src limit 1;
 create table t2 (val varchar(10));
 insert overwrite table t2 select '2014-10-10' from src limit 1; 
 ==
 Query:
 select t.val from
 (select val from t1
 union all
 select val from t1
 union all
 select val from t2
 union all
 select val from t1) t;
 ==
 Will throw exception: 
 {code}
 Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: Incompatible 
 types for union operator
   at 
 org.apache.hadoop.hive.ql.exec.UnionOperator.initializeOp(UnionOperator.java:86)
   at org.apache.hadoop.hive.ql.exec.Operator.initialize(Operator.java:380)
   at org.apache.hadoop.hive.ql.exec.Operator.initialize(Operator.java:464)
   at 
 org.apache.hadoop.hive.ql.exec.Operator.initializeChildren(Operator.java:420)
   at 
 org.apache.hadoop.hive.ql.exec.SelectOperator.initializeOp(SelectOperator.java:65)
   at org.apache.hadoop.hive.ql.exec.Operator.initialize(Operator.java:380)
   at org.apache.hadoop.hive.ql.exec.Operator.initialize(Operator.java:464)
   at 
 org.apache.hadoop.hive.ql.exec.Operator.initializeChildren(Operator.java:420)
   at 
 org.apache.hadoop.hive.ql.exec.TableScanOperator.initializeOp(TableScanOperator.java:193)
   at org.apache.hadoop.hive.ql.exec.Operator.initialize(Operator.java:380)
   at 
 org.apache.hadoop.hive.ql.exec.MapOperator.initializeOp(MapOperator.java:443)
   at org.apache.hadoop.hive.ql.exec.Operator.initialize(Operator.java:380)
   at 
 org.apache.hadoop.hive.ql.exec.mr.ExecMapper.configure(ExecMapper.java:133)
   ... 22 more
 {code}
 It was because at this query parse step, getCommonClassForUnionAll is used, 
 but at execution getCommonClass is used. They are not used consistently in 
 union. The later one does not support the implicit conversion from date to 
 string, which is the problem cause.
 The change might be simple to fix this particular union issue but I noticed 
 that there are three versions of getCommonClass: getCommonClass, 
 getCommonClassForComparison, getCommonClassForUnionAll, and wonder if they 
 need to be cleaned and refactored.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8411) Support partial partition spec for certain ALTER PARTITION statements

2014-10-16 Thread Gunther Hagleitner (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8411?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14173484#comment-14173484
 ] 

Gunther Hagleitner commented on HIVE-8411:
--

Some comments on rb. Otherwise +1

 Support partial partition spec for certain ALTER PARTITION statements
 -

 Key: HIVE-8411
 URL: https://issues.apache.org/jira/browse/HIVE-8411
 Project: Hive
  Issue Type: Bug
Reporter: Jason Dere
Assignee: Jason Dere
 Fix For: 0.14.0

 Attachments: HIVE-8411.1.patch


 To help address concerns [~hagleitn] had about having to update many 
 partitions 
 [here|https://issues.apache.org/jira/browse/HIVE-7971?focusedCommentId=14131948page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-14131948]



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8341) Transaction information in config file can grow excessively large

2014-10-16 Thread Gunther Hagleitner (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8341?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14173494#comment-14173494
 ] 

Gunther Hagleitner commented on HIVE-8341:
--

[~alangates] could you comment on the test failures. At least 
testScriptOperatorEvnVarsProcessing seems related. Also, do you have a 
reviewboard entry for this?

Patch looks good though. I think the blacklist is a good approach. Will test 
tomorrow.

 Transaction information in config file can grow excessively large
 -

 Key: HIVE-8341
 URL: https://issues.apache.org/jira/browse/HIVE-8341
 Project: Hive
  Issue Type: Bug
  Components: Transactions
Affects Versions: 0.14.0
Reporter: Alan Gates
Assignee: Alan Gates
Priority: Critical
 Attachments: HIVE-8341.2.patch, HIVE-8341.patch


 In our testing we have seen cases where the transaction list grows very 
 large.  We need a more efficient way of communicating the list.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-8466) nonReserved keywords can not be used as table alias

2014-10-16 Thread Navis (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-8466?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Navis updated HIVE-8466:

Attachment: HIVE-8466.2.patch.txt

 nonReserved keywords can not be used as table alias
 ---

 Key: HIVE-8466
 URL: https://issues.apache.org/jira/browse/HIVE-8466
 Project: Hive
  Issue Type: Bug
  Components: SQL
Affects Versions: 0.12.0, 0.13.0, 0.13.1
Reporter: cw
Priority: Minor
 Attachments: HIVE-8466.1.patch, HIVE-8466.2.patch.txt


 There is a small mistake in the patch of issue HIVE-2906. See the change of 
 FromClauseParser.g
 -: tabname=tableName (ts=tableSample)? (KW_AS? alias=identifier)?
 -- ^(TOK_TABREF $tabname $ts? $alias?)
 +: tabname=tableName (props=tableProperties)? (ts=tableSample)? (KW_AS? 
 alias=Identifier)?
 +- ^(TOK_TABREF $tabname $props? $ts? $alias?)
 With the 'identifier' changed to 'Identifier' we can not use nonReserved 
 keywords as table alias.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-7156) Group-By operator stat-annotation only uses distinct approx to generate rollups

2014-10-16 Thread Lefty Leverenz (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-7156?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Lefty Leverenz updated HIVE-7156:
-
Labels:   (was: TODOC14)

 Group-By operator stat-annotation only uses distinct approx to generate 
 rollups
 ---

 Key: HIVE-7156
 URL: https://issues.apache.org/jira/browse/HIVE-7156
 Project: Hive
  Issue Type: Sub-task
Affects Versions: 0.14.0
Reporter: Gopal V
Assignee: Prasanth J
Priority: Blocker
 Fix For: 0.14.0

 Attachments: HIVE-7156.1.patch, HIVE-7156.2.patch, HIVE-7156.3.patch, 
 HIVE-7156.4.patch, HIVE-7156.5.patch, HIVE-7156.6.patch, HIVE-7156.7.patch, 
 HIVE-7156.8.patch, HIVE-7156.8.patch, HIVE-7156.9.patch, hive-debug.log.bz2


 The stats annotation for a group-by only annotates the reduce-side row-count 
 with the distinct values.
 The map-side gets the row-count as the rows output instead of distinct * 
 parallelism, while the reducer side gets the correct parallelism.
 {code}
 hive explain select distinct L_SHIPDATE from lineitem;
   Vertices:
 Map 1 
 Map Operator Tree:
 TableScan
   alias: lineitem
   Statistics: Num rows: 589709 Data size: 4745677733354 
 Basic stats: COMPLETE Column stats: COMPLETE
   Select Operator
 expressions: l_shipdate (type: string)
 outputColumnNames: l_shipdate
 Statistics: Num rows: 589709 Data size: 4745677733354 
 Basic stats: COMPLETE Column stats: COMPLETE
 Group By Operator
   keys: l_shipdate (type: string)
   mode: hash
   outputColumnNames: _col0
   Statistics: Num rows: 589709 Data size: 
 563999032646 Basic stats: COMPLETE Column stats: COMPLETE
   Reduce Output Operator
 key expressions: _col0 (type: string)
 sort order: +
 Map-reduce partition columns: _col0 (type: string)
 Statistics: Num rows: 589709 Data size: 
 563999032646 Basic stats: COMPLETE Column stats: COMPLETE
 Execution mode: vectorized
 Reducer 2 
 Reduce Operator Tree:
   Group By Operator
 keys: KEY._col0 (type: string)
 mode: mergepartial
 outputColumnNames: _col0
 Statistics: Num rows: 1955 Data size: 183770 Basic stats: 
 COMPLETE Column stats: COMPLETE
 Select Operator
   expressions: _col0 (type: string)
   outputColumnNames: _col0
   Statistics: Num rows: 1955 Data size: 183770 Basic stats: 
 COMPLETE Column stats: COMPLETE
 {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-7156) Group-By operator stat-annotation only uses distinct approx to generate rollups

2014-10-16 Thread Lefty Leverenz (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-7156?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14173506#comment-14173506
 ] 

Lefty Leverenz commented on HIVE-7156:
--

Doc done:  Removal of *hive.stats.map.parallelism* in 0.14 is noted in two 
places in the wiki:

* [Configuration Properties -- hive.stats.map.parallelism | 
https://cwiki.apache.org/confluence/display/Hive/Configuration+Properties#ConfigurationProperties-hive.stats.map.parallelism]
* [Configuration Properties -- Tez (list of related parameters) | 
https://cwiki.apache.org/confluence/display/Hive/Configuration+Properties#ConfigurationProperties-Tez]

 Group-By operator stat-annotation only uses distinct approx to generate 
 rollups
 ---

 Key: HIVE-7156
 URL: https://issues.apache.org/jira/browse/HIVE-7156
 Project: Hive
  Issue Type: Sub-task
Affects Versions: 0.14.0
Reporter: Gopal V
Assignee: Prasanth J
Priority: Blocker
 Fix For: 0.14.0

 Attachments: HIVE-7156.1.patch, HIVE-7156.2.patch, HIVE-7156.3.patch, 
 HIVE-7156.4.patch, HIVE-7156.5.patch, HIVE-7156.6.patch, HIVE-7156.7.patch, 
 HIVE-7156.8.patch, HIVE-7156.8.patch, HIVE-7156.9.patch, hive-debug.log.bz2


 The stats annotation for a group-by only annotates the reduce-side row-count 
 with the distinct values.
 The map-side gets the row-count as the rows output instead of distinct * 
 parallelism, while the reducer side gets the correct parallelism.
 {code}
 hive explain select distinct L_SHIPDATE from lineitem;
   Vertices:
 Map 1 
 Map Operator Tree:
 TableScan
   alias: lineitem
   Statistics: Num rows: 589709 Data size: 4745677733354 
 Basic stats: COMPLETE Column stats: COMPLETE
   Select Operator
 expressions: l_shipdate (type: string)
 outputColumnNames: l_shipdate
 Statistics: Num rows: 589709 Data size: 4745677733354 
 Basic stats: COMPLETE Column stats: COMPLETE
 Group By Operator
   keys: l_shipdate (type: string)
   mode: hash
   outputColumnNames: _col0
   Statistics: Num rows: 589709 Data size: 
 563999032646 Basic stats: COMPLETE Column stats: COMPLETE
   Reduce Output Operator
 key expressions: _col0 (type: string)
 sort order: +
 Map-reduce partition columns: _col0 (type: string)
 Statistics: Num rows: 589709 Data size: 
 563999032646 Basic stats: COMPLETE Column stats: COMPLETE
 Execution mode: vectorized
 Reducer 2 
 Reduce Operator Tree:
   Group By Operator
 keys: KEY._col0 (type: string)
 mode: mergepartial
 outputColumnNames: _col0
 Statistics: Num rows: 1955 Data size: 183770 Basic stats: 
 COMPLETE Column stats: COMPLETE
 Select Operator
   expressions: _col0 (type: string)
   outputColumnNames: _col0
   Statistics: Num rows: 1955 Data size: 183770 Basic stats: 
 COMPLETE Column stats: COMPLETE
 {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-6715) Hive JDBC should include username into open session request for non-sasl connection

2014-10-16 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6715?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14173529#comment-14173529
 ] 

Hive QA commented on HIVE-6715:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12675019/HIVE-6715.3.patch

{color:red}ERROR:{color} -1 due to 5 failed/errored test(s), 6560 tests executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestHBaseCliDriver.testCliDriver_hbase_custom_key
org.apache.hadoop.hive.cli.TestHBaseCliDriver.testCliDriver_hbase_custom_key2
org.apache.hadoop.hive.cli.TestHBaseCliDriver.testCliDriver_hbase_custom_key3
org.apache.hive.beeline.TestSchemaTool.testSchemaInit
org.apache.hive.beeline.TestSchemaTool.testSchemaUpgrade
{noformat}

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/1294/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/1294/console
Test logs: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-1294/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 5 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12675019
 - PreCommit-HIVE-TRUNK-Build

 Hive JDBC should include username into open session request for non-sasl 
 connection
 ---

 Key: HIVE-6715
 URL: https://issues.apache.org/jira/browse/HIVE-6715
 Project: Hive
  Issue Type: Bug
  Components: JDBC
Reporter: Srinath
Assignee: Prasad Mujumdar
Priority: Critical
 Fix For: 0.14.0

 Attachments: HIVE-6715.1.patch, HIVE-6715.2.patch, HIVE-6715.3.patch


 The only parameter from sessVars that's being set in 
 HiveConnection.openSession() is HS2_PROXY_USER. 
 HIVE_AUTH_USER must also be set.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8467) Table Copy - Background, incremental data load

2014-10-16 Thread Rajat Venkatesh (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8467?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14173534#comment-14173534
 ] 

Rajat Venkatesh commented on HIVE-8467:
---

No they dont have to. The databases I know provide both options - sync on user 
input or automatically. I am not confident we can support automatic sync on 
external tables. Since it feels like a big feature gap, I chose a different 
name.

Yes - we also have diffs we would like to contribute in other projects to use 
Table Copy. Since the optimization is at the storage level, its very simple. 
Replace partitions from the table copy when possible.  Directories when it 
comes to Pig or M/R.  If materialized views are chosen, then the optimizers 
have to mature in more or less lock step. 

WRT to retention policy, the common case is to only keep the newest n 
partitions limited by size of the copy. We didnt chose a date range. Sometimes 
the date partition is not the top level one. This is a moving window. If older 
partitions are accessed then it will fall back to reading partitions from the 
Hive Table. 

 Table Copy - Background, incremental data load
 --

 Key: HIVE-8467
 URL: https://issues.apache.org/jira/browse/HIVE-8467
 Project: Hive
  Issue Type: New Feature
Reporter: Rajat Venkatesh
 Attachments: Table Copies.pdf


 Traditionally, Hive and other tools in the Hadoop eco-system havent required 
 a load stage. However, with recent developments, Hive is much more performant 
 when data is stored in specific formats like ORC, Parquet, Avro etc. 
 Technologies like Presto, also work much better with certain data formats. At 
 the same time, data is generated or obtained from 3rd parties in non-optimal 
 formats such as CSV, tab-limited or JSON. Many a times, its not an option to 
 change the data format at the source. We've found that users either use 
 sub-optimal formats or spend a large amount of effort creating and 
 maintaining copies. We want to propose a new construct - Table Copy - to help 
 “load” data into an optimal storage format.
 I am going to attach a PDF document with a lot more details especially 
 addressing how is this different from bulk loads in relational DBs or 
 materialized views.
 Looking forward to hear if others see a similar need to formalize conversion 
 of data to different storage formats.  If yes, are the details in the PDF 
 document a good start ?



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-8482) set hive.auto.convert.join=true has a bug when create table as select with join

2014-10-16 Thread q79969786 (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-8482?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

q79969786 updated HIVE-8482:

  Description: 
014-10-16 16:37:22,899 Stage-31 map = 0%,  reduce = 0%
2014-10-16 16:38:19,081 Stage-31 map = 100%,  reduce = 100%
Ended Job = job_201409101641_249763 with errors
Error during job, obtaining debugging information...
Job Tracking URL: 
http://CNSH041566.app.paic.com.cn:50030/jobdetails.jsp?user.name=hadoopjobid=job_201409101641_249763
Examining task ID: task_201409101641_249763_m_02 (and more) from job 
job_201409101641_249763

Task with the most failures(4): 
-
Task ID:
  task_201409101641_249763_m_00

URL:
  
http://CNSH041566.app.paic.com.cn:50030/taskdetails.jsp?jobid=job_201409101641_249763tipid=task_201409101641_249763_m_00
-
Diagnostic Messages for this Task:
java.lang.RuntimeException: org.apache.hadoop.hive.ql.metadata.HiveException: 
java.io.FileNotFoundException: 
/data/data01/data/tmp/hadoop-hadoop/mapred/local/taskTracker/hduser0401/distcache/-5448139015889594124_-204538_404393650/CNSH041566.app.paic.com.cn/tmp/hduser0401/hive_2014-10-16_16-33-51_335_4584605482989653833-1/-mr-10024/HashTable-Stage-31/Stage-31.tar.gz/MapJoin-mapfile21--.hashtable
 (No such file or directory)
at org.apache.hadoop.hive.ql.exec.mr.ExecMapper.map(ExecMapper.java:195)
at org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:50)
at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:430)
at org.apache.hadoop.mapred.MapTask.run(MapTask.java:366)
at org.apache.hadoop.mapred.Child$4.run(Child.java:255)
at java.security.AccessController.doPrivileged(Native Method)
at javax.security.auth.Subject.doAs(Subject.java:396)
at 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1190)
at org.apache.hadoop.mapred.Child.main(Child.java:249)
Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: 
java.io.FileNotFoundException: 
/data/data01/data/tmp/hadoop-hadoop/mapred/local/taskTracker/hduser0401/distcache/-5448139015889594124_-204538_404393650/CNSH041566.app.paic.com.cn/tmp/hduser0401/hive_2014-10-16_16-33-51_335_4584605482989653833-1/-mr-10024/HashTable-Stage-31/Stage-31.tar.gz/MapJoin-mapfile21--.hashtable
 (No such file or directory)
at 
org.apache.hadoop.hive.ql.exec.mr.HashTableLoader.load(HashTableLoader.java:104)
at 
org.apache.hadoop.hive.ql.exec.MapJoinOperator.loadHashTable(MapJoinOperator.java:152)
at 
org.apache.hadoop.hive.ql.exec.MapJoinOperator.cleanUpInputFileChangedOp(MapJoinOperator.java:178)
at 
org.apache.hadoop.hive.ql.exec.Operator.cleanUpInputFileChanged(Operator.java:1029)
at 
org.apache.hadoop.hive.ql.exec.Operator.cleanUpInputFileChanged(Operator.java:1033)
at 
org.apache.hadoop.hive.ql.exec.Operator.cleanUpInputFileChanged(Operator.java:1033)
at 
org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:505)
at org.apache.hadoop.hive.ql.exec.mr.ExecMapper.map(ExecMapper.java:177)
... 8 more
Caused by: java.io.FileNotFoundException: 
/data/data01/data/tmp/hadoop-hadoop/mapred/local/taskTracker/hduser0401/distcache/-5448139015889594124_-204538_404393650/CNSH041566.app.paic.com.cn/tmp/hduser0401/hive_2014-10-16_16-33-51_335_4584605482989653833-1/-mr-10024/HashTable-Stage-31/Stage-31.tar.gz/MapJoin-mapfile21--.hashtable
 (No such file or directory)
at java.io.FileInputStream.open(Native Method)
at java.io.FileInputStream.init(FileInputStream.java:120)
at java.io.FileInputStream.init(FileInputStream.java:79)
at 
org.apache.hadoop.hive.ql.exec.mr.HashTableLoader.load(HashTableLoader.java:95)
... 15 more


FAILED: Execution Error, return code 2 from 
org.apache.hadoop.hive.ql.exec.mr.MapRedTask
MapReduce Jobs Launched: 

  was:
java.lang.RuntimeException: org.apache.hadoop.hive.ql.metadata.HiveException: 
java.io.FileNotFoundException: 
/data/data01/data/tmp/hadoop-hadoop/mapred/local/taskTracker/hduser0401/distcache/4554609282698230126_-1275563112_394094114/CNSH041566.app.paic.com.cn/tmp/hduser0401/hive_2014-10-16_13-42-26_451_2739170706811325303-1/-mr-10024/HashTable-Stage-31/Stage-31.tar.gz/MapJoin-mapfile21--.hashtable
 (No such file or directory)
at org.apache.hadoop.hive.ql.exec.mr.ExecMapper.map(ExecMapper.java:195)
at org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:50)
at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:430)
at org.apache.hadoop.mapred.MapTask.run(MapTask.java:366)
at org.apache.hadoop.mapred.Child$4.run(Child.java:255)
at java.security.AccessController.doPrivileged(Native Method)
at javax.security.auth.Subject.doAs(Subject.java:396)
at 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1190)
at 

[jira] [Commented] (HIVE-8387) add retry logic to ZooKeeperStorage in WebHCat

2014-10-16 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8387?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14173574#comment-14173574
 ] 

Hive QA commented on HIVE-8387:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12675058/HIVE-8387.patch

{color:red}ERROR:{color} -1 due to 5 failed/errored test(s), 6559 tests executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestHBaseCliDriver.testCliDriver_hbase_custom_key
org.apache.hadoop.hive.cli.TestHBaseCliDriver.testCliDriver_hbase_custom_key2
org.apache.hadoop.hive.cli.TestHBaseCliDriver.testCliDriver_hbase_custom_key3
org.apache.hive.beeline.TestSchemaTool.testSchemaInit
org.apache.hive.beeline.TestSchemaTool.testSchemaUpgrade
{noformat}

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/1295/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/1295/console
Test logs: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-1295/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 5 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12675058
 - PreCommit-HIVE-TRUNK-Build

 add retry logic to ZooKeeperStorage in WebHCat
 --

 Key: HIVE-8387
 URL: https://issues.apache.org/jira/browse/HIVE-8387
 Project: Hive
  Issue Type: Bug
  Components: WebHCat
Affects Versions: 0.13.1
Reporter: Eugene Koifman
Assignee: Eugene Koifman
 Attachments: HIVE-8387.patch


 ZK interactions may run into transient errors that should be retried.  
 Currently there is no retry logic in WebHCat for this.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Assigned] (HIVE-7709) Create SparkReporter [Spark Branch]

2014-10-16 Thread Chengxiang Li (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-7709?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chengxiang Li reassigned HIVE-7709:
---

Assignee: Chengxiang Li

 Create SparkReporter [Spark Branch]
 ---

 Key: HIVE-7709
 URL: https://issues.apache.org/jira/browse/HIVE-7709
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Reporter: Chengxiang Li
Assignee: Chengxiang Li

 Hive operators use Reporter to collect global information,  with Hive on 
 Spark mode, we need a new implementation of Reporter to collect hive operator 
 level information based on spark specified Counter. This task should depends 
 on HIVE-7551.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HIVE-8483) HiveServer2 can`t correct re-establish metastore communication

2014-10-16 Thread Stanilovsky Evgeny (JIRA)
Stanilovsky Evgeny created HIVE-8483:


 Summary: HiveServer2 can`t correct re-establish metastore 
communication
 Key: HIVE-8483
 URL: https://issues.apache.org/jira/browse/HIVE-8483
 Project: Hive
  Issue Type: Improvement
Affects Versions: 0.12.0
 Environment: unbuntu 14.04 java 1.7
Reporter: Stanilovsky Evgeny


with hive-site.xml config like:

property
  namehive.metastore.uris/name
  valuethrift://mashine1:9083,thrift://mashine1:9083/value
/property

log:

2014-10-16 14:04:04,502 INFO  hive.metastore 
(HiveMetaStoreClient.java:open(257)) - Trying to connect to metastore with URI 
thrift://mashine1:9083
2014-10-16 14:04:04,515 INFO  hive.metastore 
(HiveMetaStoreClient.java:open(345)) - Connected to metastore.

if split current connection, for example with command:  iptables -I INPUT -d 
mashine1 -j REJECT, and try to refresh database list through beeline : show 
tables; we have an stack like :

2014-10-07 10:14:33,898 ERROR hive.log 
(MetaStoreUtils.java:logAndThrowMetaException(911)) - Got exception: 
org.apache.thrift.transport.TTransportException java.net.SocketException: 
Broken pipe
org.apache.thrift.transport.TTransportException: java.net.SocketException: 
Broken pipe
at 
org.apache.thrift.transport.TIOStreamTransport.flush(TIOStreamTransport.java:161)
at org.apache.thrift.TServiceClient.sendBase(TServiceClient.java:65)
at 
org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.send_get_databases(ThriftHiveMetastore.java:498)
at 
org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.get_databases(ThriftHiveMetastore.java:490)
at 
org.apache.hadoop.hive.metastore.HiveMetaStoreClient.getDatabases(HiveMetaStoreClient.java:735)
at sun.reflect.GeneratedMethodAccessor15.invoke(Unknown Source)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:606)
at 
org.apache.hadoop.hive.metastore.HiveMetaStoreClient$SynchronizedHandler.invoke(HiveMetaStoreClient.java:1425)
at com.sun.proxy.$Proxy14.getDatabases(Unknown Source)
at 
org.apache.hive.service.cli.operation.GetSchemasOperation.run(GetSchemasOperation.java:73)
at 
org.apache.hive.service.cli.session.HiveSessionImpl.runOperationWithLogCapture(HiveSessionImpl.java:514)
at 
org.apache.hive.service.cli.session.HiveSessionImpl.getSchemas(HiveSessionImpl.java:284)
at sun.reflect.GeneratedMethodAccessor19.invoke(Unknown Source)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:606)
at 
org.apache.hive.service.cli.session.HiveSessionProxy$1.run(HiveSessionProxy.java:62)
at java.security.AccessController.doPrivileged(Native Method)
at javax.security.auth.Subject.doAs(Subject.java:415)
at 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1554)
at 
org.apache.hadoop.hive.shims.HadoopShimsSecure.doAs(HadoopShimsSecure.java:556)
at 
org.apache.hive.service.cli.session.HiveSessionProxy.invoke(HiveSessionProxy.java:57)
at com.sun.proxy.$Proxy13.getSchemas(Unknown Source)
at 
org.apache.hive.service.cli.CLIService.getSchemas(CLIService.java:222)
at 
org.apache.hive.service.cli.thrift.ThriftCLIService.GetSchemas(ThriftCLIService.java:360)
at 
org.apache.hive.service.cli.thrift.TCLIService$Processor$GetSchemas.getResult(TCLIService.java:1433)
at 
org.apache.hive.service.cli.thrift.TCLIService$Processor$GetSchemas.getResult(TCLIService.java:1418)
at org.apache.thrift.ProcessFunction.process(ProcessFunction.java:39)
at org.apache.thrift.TBaseProcessor.process(TBaseProcessor.java:39)
at 
org.apache.hive.service.auth.TSetIpAddressProcessor.process(TSetIpAddressProcessor.java:57)
at 
org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPoolServer.java:244)
at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
at java.lang.Thread.run(Thread.java:744)




--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-8483) HiveServer2 can`t correct re-establish metastore communication

2014-10-16 Thread Stanilovsky Evgeny (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-8483?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Stanilovsky Evgeny updated HIVE-8483:
-
Attachment: patch1.patch

 HiveServer2 can`t correct re-establish metastore communication
 --

 Key: HIVE-8483
 URL: https://issues.apache.org/jira/browse/HIVE-8483
 Project: Hive
  Issue Type: Improvement
Affects Versions: 0.12.0
 Environment: unbuntu 14.04 java 1.7
Reporter: Stanilovsky Evgeny
  Labels: patch
 Attachments: patch1.patch


 with hive-site.xml config like:
 property
   namehive.metastore.uris/name
   valuethrift://mashine1:9083,thrift://mashine1:9083/value
 /property
 log:
 2014-10-16 14:04:04,502 INFO  hive.metastore 
 (HiveMetaStoreClient.java:open(257)) - Trying to connect to metastore with 
 URI thrift://mashine1:9083
 2014-10-16 14:04:04,515 INFO  hive.metastore 
 (HiveMetaStoreClient.java:open(345)) - Connected to metastore.
 if split current connection, for example with command:  iptables -I INPUT -d 
 mashine1 -j REJECT, and try to refresh database list through beeline : show 
 tables; we have an stack like :
 2014-10-07 10:14:33,898 ERROR hive.log 
 (MetaStoreUtils.java:logAndThrowMetaException(911)) - Got exception: 
 org.apache.thrift.transport.TTransportException java.net.SocketException: 
 Broken pipe
 org.apache.thrift.transport.TTransportException: java.net.SocketException: 
 Broken pipe
 at 
 org.apache.thrift.transport.TIOStreamTransport.flush(TIOStreamTransport.java:161)
 at org.apache.thrift.TServiceClient.sendBase(TServiceClient.java:65)
 at 
 org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.send_get_databases(ThriftHiveMetastore.java:498)
 at 
 org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.get_databases(ThriftHiveMetastore.java:490)
 at 
 org.apache.hadoop.hive.metastore.HiveMetaStoreClient.getDatabases(HiveMetaStoreClient.java:735)
 at sun.reflect.GeneratedMethodAccessor15.invoke(Unknown Source)
 at 
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
 at java.lang.reflect.Method.invoke(Method.java:606)
 at 
 org.apache.hadoop.hive.metastore.HiveMetaStoreClient$SynchronizedHandler.invoke(HiveMetaStoreClient.java:1425)
 at com.sun.proxy.$Proxy14.getDatabases(Unknown Source)
 at 
 org.apache.hive.service.cli.operation.GetSchemasOperation.run(GetSchemasOperation.java:73)
 at 
 org.apache.hive.service.cli.session.HiveSessionImpl.runOperationWithLogCapture(HiveSessionImpl.java:514)
 at 
 org.apache.hive.service.cli.session.HiveSessionImpl.getSchemas(HiveSessionImpl.java:284)
 at sun.reflect.GeneratedMethodAccessor19.invoke(Unknown Source)
 at 
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
 at java.lang.reflect.Method.invoke(Method.java:606)
 at 
 org.apache.hive.service.cli.session.HiveSessionProxy$1.run(HiveSessionProxy.java:62)
 at java.security.AccessController.doPrivileged(Native Method)
 at javax.security.auth.Subject.doAs(Subject.java:415)
 at 
 org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1554)
 at 
 org.apache.hadoop.hive.shims.HadoopShimsSecure.doAs(HadoopShimsSecure.java:556)
 at 
 org.apache.hive.service.cli.session.HiveSessionProxy.invoke(HiveSessionProxy.java:57)
 at com.sun.proxy.$Proxy13.getSchemas(Unknown Source)
 at 
 org.apache.hive.service.cli.CLIService.getSchemas(CLIService.java:222)
 at 
 org.apache.hive.service.cli.thrift.ThriftCLIService.GetSchemas(ThriftCLIService.java:360)
 at 
 org.apache.hive.service.cli.thrift.TCLIService$Processor$GetSchemas.getResult(TCLIService.java:1433)
 at 
 org.apache.hive.service.cli.thrift.TCLIService$Processor$GetSchemas.getResult(TCLIService.java:1418)
 at org.apache.thrift.ProcessFunction.process(ProcessFunction.java:39)
 at org.apache.thrift.TBaseProcessor.process(TBaseProcessor.java:39)
 at 
 org.apache.hive.service.auth.TSetIpAddressProcessor.process(TSetIpAddressProcessor.java:57)
 at 
 org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPoolServer.java:244)
 at 
 java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
 at 
 java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
 at java.lang.Thread.run(Thread.java:744)



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-8483) HiveServer2 can`t correct re-establish metastore communication

2014-10-16 Thread Stanilovsky Evgeny (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-8483?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Stanilovsky Evgeny updated HIVE-8483:
-
Description: 
with hive-site.xml config like:

property
  namehive.metastore.uris/name
  valuethrift://mashine1:9083,thrift://mashine1:9083/value
/property

log:

2014-10-16 14:04:04,502 INFO  hive.metastore 
(HiveMetaStoreClient.java:open(257)) - Trying to connect to metastore with URI 
thrift://mashine1:9083
2014-10-16 14:04:04,515 INFO  hive.metastore 
(HiveMetaStoreClient.java:open(345)) - Connected to metastore.

if split current connection, for example with command:  iptables -I INPUT -d 
mashine1 -j REJECT, and try to refresh database list through hue\beeline : show 
tables; we have an stack like :

2014-10-07 10:14:33,898 ERROR hive.log 
(MetaStoreUtils.java:logAndThrowMetaException(911)) - Got exception: 
org.apache.thrift.transport.TTransportException java.net.SocketException: 
Broken pipe
org.apache.thrift.transport.TTransportException: java.net.SocketException: 
Broken pipe
at 
org.apache.thrift.transport.TIOStreamTransport.flush(TIOStreamTransport.java:161)
at org.apache.thrift.TServiceClient.sendBase(TServiceClient.java:65)
at 
org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.send_get_databases(ThriftHiveMetastore.java:498)
at 
org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.get_databases(ThriftHiveMetastore.java:490)
at 
org.apache.hadoop.hive.metastore.HiveMetaStoreClient.getDatabases(HiveMetaStoreClient.java:735)
at sun.reflect.GeneratedMethodAccessor15.invoke(Unknown Source)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:606)
at 
org.apache.hadoop.hive.metastore.HiveMetaStoreClient$SynchronizedHandler.invoke(HiveMetaStoreClient.java:1425)
at com.sun.proxy.$Proxy14.getDatabases(Unknown Source)
at 
org.apache.hive.service.cli.operation.GetSchemasOperation.run(GetSchemasOperation.java:73)
at 
org.apache.hive.service.cli.session.HiveSessionImpl.runOperationWithLogCapture(HiveSessionImpl.java:514)
at 
org.apache.hive.service.cli.session.HiveSessionImpl.getSchemas(HiveSessionImpl.java:284)
at sun.reflect.GeneratedMethodAccessor19.invoke(Unknown Source)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:606)
at 
org.apache.hive.service.cli.session.HiveSessionProxy$1.run(HiveSessionProxy.java:62)
at java.security.AccessController.doPrivileged(Native Method)
at javax.security.auth.Subject.doAs(Subject.java:415)
at 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1554)
at 
org.apache.hadoop.hive.shims.HadoopShimsSecure.doAs(HadoopShimsSecure.java:556)
at 
org.apache.hive.service.cli.session.HiveSessionProxy.invoke(HiveSessionProxy.java:57)
at com.sun.proxy.$Proxy13.getSchemas(Unknown Source)
at 
org.apache.hive.service.cli.CLIService.getSchemas(CLIService.java:222)
at 
org.apache.hive.service.cli.thrift.ThriftCLIService.GetSchemas(ThriftCLIService.java:360)
at 
org.apache.hive.service.cli.thrift.TCLIService$Processor$GetSchemas.getResult(TCLIService.java:1433)
at 
org.apache.hive.service.cli.thrift.TCLIService$Processor$GetSchemas.getResult(TCLIService.java:1418)
at org.apache.thrift.ProcessFunction.process(ProcessFunction.java:39)
at org.apache.thrift.TBaseProcessor.process(TBaseProcessor.java:39)
at 
org.apache.hive.service.auth.TSetIpAddressProcessor.process(TSetIpAddressProcessor.java:57)
at 
org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPoolServer.java:244)
at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
at java.lang.Thread.run(Thread.java:744)


  was:
with hive-site.xml config like:

property
  namehive.metastore.uris/name
  valuethrift://mashine1:9083,thrift://mashine1:9083/value
/property

log:

2014-10-16 14:04:04,502 INFO  hive.metastore 
(HiveMetaStoreClient.java:open(257)) - Trying to connect to metastore with URI 
thrift://mashine1:9083
2014-10-16 14:04:04,515 INFO  hive.metastore 
(HiveMetaStoreClient.java:open(345)) - Connected to metastore.

if split current connection, for example with command:  iptables -I INPUT -d 
mashine1 -j REJECT, and try to refresh database list through beeline : show 
tables; we have an stack like :

2014-10-07 10:14:33,898 ERROR hive.log 
(MetaStoreUtils.java:logAndThrowMetaException(911)) - Got exception: 
org.apache.thrift.transport.TTransportException java.net.SocketException: 
Broken pipe

[jira] [Updated] (HIVE-8483) HiveServer2 can`t correct re-establish metastore communication

2014-10-16 Thread Stanilovsky Evgeny (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-8483?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Stanilovsky Evgeny updated HIVE-8483:
-
Description: 
with hive-site.xml config like:

property
  namehive.metastore.uris/name
  valuethrift://mashine1:9083,thrift://mashine2:9083/value
/property

log:

2014-10-16 14:04:04,502 INFO  hive.metastore 
(HiveMetaStoreClient.java:open(257)) - Trying to connect to metastore with URI 
thrift://mashine1:9083
2014-10-16 14:04:04,515 INFO  hive.metastore 
(HiveMetaStoreClient.java:open(345)) - Connected to metastore.

if split current connection, for example with command:  iptables -I INPUT -d 
mashine1 -j REJECT, and try to refresh database list through hue\beeline : show 
tables; we have an stack like :

2014-10-07 10:14:33,898 ERROR hive.log 
(MetaStoreUtils.java:logAndThrowMetaException(911)) - Got exception: 
org.apache.thrift.transport.TTransportException java.net.SocketException: 
Broken pipe
org.apache.thrift.transport.TTransportException: java.net.SocketException: 
Broken pipe
at 
org.apache.thrift.transport.TIOStreamTransport.flush(TIOStreamTransport.java:161)
at org.apache.thrift.TServiceClient.sendBase(TServiceClient.java:65)
at 
org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.send_get_databases(ThriftHiveMetastore.java:498)
at 
org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.get_databases(ThriftHiveMetastore.java:490)
at 
org.apache.hadoop.hive.metastore.HiveMetaStoreClient.getDatabases(HiveMetaStoreClient.java:735)
at sun.reflect.GeneratedMethodAccessor15.invoke(Unknown Source)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:606)
at 
org.apache.hadoop.hive.metastore.HiveMetaStoreClient$SynchronizedHandler.invoke(HiveMetaStoreClient.java:1425)
at com.sun.proxy.$Proxy14.getDatabases(Unknown Source)
at 
org.apache.hive.service.cli.operation.GetSchemasOperation.run(GetSchemasOperation.java:73)
at 
org.apache.hive.service.cli.session.HiveSessionImpl.runOperationWithLogCapture(HiveSessionImpl.java:514)
at 
org.apache.hive.service.cli.session.HiveSessionImpl.getSchemas(HiveSessionImpl.java:284)
at sun.reflect.GeneratedMethodAccessor19.invoke(Unknown Source)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:606)
at 
org.apache.hive.service.cli.session.HiveSessionProxy$1.run(HiveSessionProxy.java:62)
at java.security.AccessController.doPrivileged(Native Method)
at javax.security.auth.Subject.doAs(Subject.java:415)
at 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1554)
at 
org.apache.hadoop.hive.shims.HadoopShimsSecure.doAs(HadoopShimsSecure.java:556)
at 
org.apache.hive.service.cli.session.HiveSessionProxy.invoke(HiveSessionProxy.java:57)
at com.sun.proxy.$Proxy13.getSchemas(Unknown Source)
at 
org.apache.hive.service.cli.CLIService.getSchemas(CLIService.java:222)
at 
org.apache.hive.service.cli.thrift.ThriftCLIService.GetSchemas(ThriftCLIService.java:360)
at 
org.apache.hive.service.cli.thrift.TCLIService$Processor$GetSchemas.getResult(TCLIService.java:1433)
at 
org.apache.hive.service.cli.thrift.TCLIService$Processor$GetSchemas.getResult(TCLIService.java:1418)
at org.apache.thrift.ProcessFunction.process(ProcessFunction.java:39)
at org.apache.thrift.TBaseProcessor.process(TBaseProcessor.java:39)
at 
org.apache.hive.service.auth.TSetIpAddressProcessor.process(TSetIpAddressProcessor.java:57)
at 
org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPoolServer.java:244)
at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
at java.lang.Thread.run(Thread.java:744)


  was:
with hive-site.xml config like:

property
  namehive.metastore.uris/name
  valuethrift://mashine1:9083,thrift://mashine1:9083/value
/property

log:

2014-10-16 14:04:04,502 INFO  hive.metastore 
(HiveMetaStoreClient.java:open(257)) - Trying to connect to metastore with URI 
thrift://mashine1:9083
2014-10-16 14:04:04,515 INFO  hive.metastore 
(HiveMetaStoreClient.java:open(345)) - Connected to metastore.

if split current connection, for example with command:  iptables -I INPUT -d 
mashine1 -j REJECT, and try to refresh database list through hue\beeline : show 
tables; we have an stack like :

2014-10-07 10:14:33,898 ERROR hive.log 
(MetaStoreUtils.java:logAndThrowMetaException(911)) - Got exception: 
org.apache.thrift.transport.TTransportException java.net.SocketException: 
Broken pipe

[jira] [Updated] (HIVE-8483) HiveServer2 can`t correct re-establish metastore communication

2014-10-16 Thread Stanilovsky Evgeny (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-8483?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Stanilovsky Evgeny updated HIVE-8483:
-
Status: Patch Available  (was: Open)

 HiveServer2 can`t correct re-establish metastore communication
 --

 Key: HIVE-8483
 URL: https://issues.apache.org/jira/browse/HIVE-8483
 Project: Hive
  Issue Type: Improvement
Affects Versions: 0.12.0
 Environment: unbuntu 14.04 java 1.7
Reporter: Stanilovsky Evgeny
  Labels: patch
 Attachments: patch1.patch


 with hive-site.xml config like:
 property
   namehive.metastore.uris/name
   valuethrift://mashine1:9083,thrift://mashine1:9083/value
 /property
 log:
 2014-10-16 14:04:04,502 INFO  hive.metastore 
 (HiveMetaStoreClient.java:open(257)) - Trying to connect to metastore with 
 URI thrift://mashine1:9083
 2014-10-16 14:04:04,515 INFO  hive.metastore 
 (HiveMetaStoreClient.java:open(345)) - Connected to metastore.
 if split current connection, for example with command:  iptables -I INPUT -d 
 mashine1 -j REJECT, and try to refresh database list through hue\beeline : 
 show tables; we have an stack like :
 2014-10-07 10:14:33,898 ERROR hive.log 
 (MetaStoreUtils.java:logAndThrowMetaException(911)) - Got exception: 
 org.apache.thrift.transport.TTransportException java.net.SocketException: 
 Broken pipe
 org.apache.thrift.transport.TTransportException: java.net.SocketException: 
 Broken pipe
 at 
 org.apache.thrift.transport.TIOStreamTransport.flush(TIOStreamTransport.java:161)
 at org.apache.thrift.TServiceClient.sendBase(TServiceClient.java:65)
 at 
 org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.send_get_databases(ThriftHiveMetastore.java:498)
 at 
 org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.get_databases(ThriftHiveMetastore.java:490)
 at 
 org.apache.hadoop.hive.metastore.HiveMetaStoreClient.getDatabases(HiveMetaStoreClient.java:735)
 at sun.reflect.GeneratedMethodAccessor15.invoke(Unknown Source)
 at 
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
 at java.lang.reflect.Method.invoke(Method.java:606)
 at 
 org.apache.hadoop.hive.metastore.HiveMetaStoreClient$SynchronizedHandler.invoke(HiveMetaStoreClient.java:1425)
 at com.sun.proxy.$Proxy14.getDatabases(Unknown Source)
 at 
 org.apache.hive.service.cli.operation.GetSchemasOperation.run(GetSchemasOperation.java:73)
 at 
 org.apache.hive.service.cli.session.HiveSessionImpl.runOperationWithLogCapture(HiveSessionImpl.java:514)
 at 
 org.apache.hive.service.cli.session.HiveSessionImpl.getSchemas(HiveSessionImpl.java:284)
 at sun.reflect.GeneratedMethodAccessor19.invoke(Unknown Source)
 at 
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
 at java.lang.reflect.Method.invoke(Method.java:606)
 at 
 org.apache.hive.service.cli.session.HiveSessionProxy$1.run(HiveSessionProxy.java:62)
 at java.security.AccessController.doPrivileged(Native Method)
 at javax.security.auth.Subject.doAs(Subject.java:415)
 at 
 org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1554)
 at 
 org.apache.hadoop.hive.shims.HadoopShimsSecure.doAs(HadoopShimsSecure.java:556)
 at 
 org.apache.hive.service.cli.session.HiveSessionProxy.invoke(HiveSessionProxy.java:57)
 at com.sun.proxy.$Proxy13.getSchemas(Unknown Source)
 at 
 org.apache.hive.service.cli.CLIService.getSchemas(CLIService.java:222)
 at 
 org.apache.hive.service.cli.thrift.ThriftCLIService.GetSchemas(ThriftCLIService.java:360)
 at 
 org.apache.hive.service.cli.thrift.TCLIService$Processor$GetSchemas.getResult(TCLIService.java:1433)
 at 
 org.apache.hive.service.cli.thrift.TCLIService$Processor$GetSchemas.getResult(TCLIService.java:1418)
 at org.apache.thrift.ProcessFunction.process(ProcessFunction.java:39)
 at org.apache.thrift.TBaseProcessor.process(TBaseProcessor.java:39)
 at 
 org.apache.hive.service.auth.TSetIpAddressProcessor.process(TSetIpAddressProcessor.java:57)
 at 
 org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPoolServer.java:244)
 at 
 java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
 at 
 java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
 at java.lang.Thread.run(Thread.java:744)



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8428) PCR doesnt remove filters involving casts

2014-10-16 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8428?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14173606#comment-14173606
 ] 

Hive QA commented on HIVE-8428:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12675080/HIVE-8428.3.patch

{color:red}ERROR:{color} -1 due to 6 failed/errored test(s), 6559 tests executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_udaf_histogram_numeric
org.apache.hadoop.hive.cli.TestHBaseCliDriver.testCliDriver_hbase_custom_key
org.apache.hadoop.hive.cli.TestHBaseCliDriver.testCliDriver_hbase_custom_key2
org.apache.hadoop.hive.cli.TestHBaseCliDriver.testCliDriver_hbase_custom_key3
org.apache.hive.beeline.TestSchemaTool.testSchemaInit
org.apache.hive.beeline.TestSchemaTool.testSchemaUpgrade
{noformat}

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/1296/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/1296/console
Test logs: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-1296/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 6 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12675080
 - PreCommit-HIVE-TRUNK-Build

 PCR doesnt remove filters involving casts
 -

 Key: HIVE-8428
 URL: https://issues.apache.org/jira/browse/HIVE-8428
 Project: Hive
  Issue Type: Improvement
  Components: Logical Optimizer
Affects Versions: 0.11.0, 0.12.0, 0.13.0
Reporter: Ashutosh Chauhan
Assignee: Ashutosh Chauhan
 Attachments: HIVE-8428.1.patch, HIVE-8428.2.patch, HIVE-8428.3.patch, 
 HIVE-8428.patch


 e.g.,
 select key,value from srcpart where hr = cast(11 as double);



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HIVE-8484) HCatalog throws an exception if Pig job is of type 'fetch'

2014-10-16 Thread Lorand Bendig (JIRA)
Lorand Bendig created HIVE-8484:
---

 Summary: HCatalog throws an exception if Pig job is of type 'fetch'
 Key: HIVE-8484
 URL: https://issues.apache.org/jira/browse/HIVE-8484
 Project: Hive
  Issue Type: Bug
  Components: HCatalog
Affects Versions: 0.14.0
Reporter: Lorand Bendig
 Fix For: 0.14.0


When Pig tries to retrieve result in fetch mode through HCatalog then 
HCatLoader#setLocation(String location, Job job) can't set the outputschema 
because HCatUtil#checkJobContextIfRunningFromBackend(job) always returns false :
{code}
public static boolean checkJobContextIfRunningFromBackend(JobContext j) {
if (j.getConfiguration().get(mapred.task.id, ).equals() 
!(true.equals(j.getConfiguration().get(pig.illustrating {
  return false;
}
return true;
  }
{code}
This is because in fetch mode we don't have a mapred.task.id. A null 
outputschema will raise an exception when HCatBaseLoader#getNext() is called: 
(ERROR 6018: Error converting read value to tuple).



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8471) Cleanup TestSSL

2014-10-16 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8471?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14173633#comment-14173633
 ] 

Hive QA commented on HIVE-8471:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12675083/HIVE-8471.patch

{color:red}ERROR:{color} -1 due to 9 failed/errored test(s), 6559 tests executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestHBaseCliDriver.testCliDriver_hbase_custom_key
org.apache.hadoop.hive.cli.TestHBaseCliDriver.testCliDriver_hbase_custom_key2
org.apache.hadoop.hive.cli.TestHBaseCliDriver.testCliDriver_hbase_custom_key3
org.apache.hive.beeline.TestSchemaTool.testSchemaInit
org.apache.hive.beeline.TestSchemaTool.testSchemaUpgrade
org.apache.hive.jdbc.TestSSL.testSSLConnectionWithProperty
org.apache.hive.jdbc.TestSSL.testSSLConnectionWithURL
org.apache.hive.jdbc.TestSSL.testSSLFetch
org.apache.hive.jdbc.TestSSL.testSSLFetchHttp
{noformat}

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/1297/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/1297/console
Test logs: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-1297/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 9 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12675083
 - PreCommit-HIVE-TRUNK-Build

 Cleanup TestSSL
 ---

 Key: HIVE-8471
 URL: https://issues.apache.org/jira/browse/HIVE-8471
 Project: Hive
  Issue Type: Improvement
Reporter: Brock Noland
Assignee: Brock Noland
Priority: Minor
 Attachments: HIVE-8471.patch


 conf is static and there are some other minor cleanups



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-8484) HCatalog throws an exception if Pig job is of type 'fetch'

2014-10-16 Thread Lorand Bendig (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-8484?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Lorand Bendig updated HIVE-8484:

Attachment: HIVE-8484.patch

As a solution I'd check for the property 'pig.job.converted.fetch' and if set, 
would return true.

 HCatalog throws an exception if Pig job is of type 'fetch'
 --

 Key: HIVE-8484
 URL: https://issues.apache.org/jira/browse/HIVE-8484
 Project: Hive
  Issue Type: Bug
  Components: HCatalog
Affects Versions: 0.14.0
Reporter: Lorand Bendig
 Fix For: 0.14.0

 Attachments: HIVE-8484.patch


 When Pig tries to retrieve result in fetch mode through HCatalog then 
 HCatLoader#setLocation(String location, Job job) can't set the outputschema 
 because HCatUtil#checkJobContextIfRunningFromBackend(job) always returns 
 false :
 {code}
 public static boolean checkJobContextIfRunningFromBackend(JobContext j) {
 if (j.getConfiguration().get(mapred.task.id, ).equals() 
 !(true.equals(j.getConfiguration().get(pig.illustrating {
   return false;
 }
 return true;
   }
 {code}
 This is because in fetch mode we don't have a mapred.task.id. A null 
 outputschema will raise an exception when HCatBaseLoader#getNext() is called: 
 (ERROR 6018: Error converting read value to tuple).



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-8484) HCatalog throws an exception if Pig job is of type 'fetch'

2014-10-16 Thread Lorand Bendig (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-8484?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Lorand Bendig updated HIVE-8484:

Status: Patch Available  (was: Open)

 HCatalog throws an exception if Pig job is of type 'fetch'
 --

 Key: HIVE-8484
 URL: https://issues.apache.org/jira/browse/HIVE-8484
 Project: Hive
  Issue Type: Bug
  Components: HCatalog
Affects Versions: 0.14.0
Reporter: Lorand Bendig
 Fix For: 0.14.0

 Attachments: HIVE-8484.patch


 When Pig tries to retrieve result in fetch mode through HCatalog then 
 HCatLoader#setLocation(String location, Job job) can't set the outputschema 
 because HCatUtil#checkJobContextIfRunningFromBackend(job) always returns 
 false :
 {code}
 public static boolean checkJobContextIfRunningFromBackend(JobContext j) {
 if (j.getConfiguration().get(mapred.task.id, ).equals() 
 !(true.equals(j.getConfiguration().get(pig.illustrating {
   return false;
 }
 return true;
   }
 {code}
 This is because in fetch mode we don't have a mapred.task.id. A null 
 outputschema will raise an exception when HCatBaseLoader#getNext() is called: 
 (ERROR 6018: Error converting read value to tuple).



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8395) CBO: enable by default

2014-10-16 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8395?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14173679#comment-14173679
 ] 

Hive QA commented on HIVE-8395:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12675104/HIVE-8395.06.patch

{color:red}ERROR:{color} -1 due to 102 failed/errored test(s), 6559 tests 
executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestAccumuloCliDriver.testCliDriver_accumulo_predicate_pushdown
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_ambiguous_col
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_annotate_stats_groupby
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_annotate_stats_groupby2
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_annotate_stats_select
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_constprog2
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_correlationoptimizer6
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_correlationoptimizer8
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_filter_join_breaktask2
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_index_auto
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_index_auto_file_format
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_index_auto_mult_tables
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_index_auto_mult_tables_compact
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_index_auto_multiple
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_index_auto_partitioned
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_index_auto_self_join
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_index_auto_unused
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_index_auto_update
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_index_bitmap_auto_partitioned
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_index_bitmap_compression
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_index_compression
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_index_stale
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_index_stale_partitioned
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_input25
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_input_part3
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_input_part6
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_join12
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_join_merge_multi_expressions
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_limit_partition_metadataonly
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_list_bucket_dml_1
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_list_bucket_dml_11
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_list_bucket_dml_12
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_list_bucket_dml_13
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_list_bucket_dml_2
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_list_bucket_dml_3
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_list_bucket_dml_4
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_list_bucket_dml_5
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_list_bucket_dml_6
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_list_bucket_dml_7
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_list_bucket_dml_8
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_list_bucket_dml_9
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_list_bucket_query_multiskew_1
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_list_bucket_query_multiskew_2
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_list_bucket_query_multiskew_3
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_list_bucket_query_oneskew_1
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_list_bucket_query_oneskew_2
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_list_bucket_query_oneskew_3
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_mergejoins
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_partition_wise_fileformat2
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_pcr
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_ppd_gby
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_ppd_gby2
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_ppd_join3
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_ppd_outer_join4
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_ppd_udf_case
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_ppd_union_view
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_print_header
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_select_unquote_and
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_select_unquote_not

[jira] [Created] (HIVE-8485) HMS on Oracle incompatibility

2014-10-16 Thread Ryan Pridgeon (JIRA)
Ryan Pridgeon created HIVE-8485:
---

 Summary: HMS on Oracle incompatibility
 Key: HIVE-8485
 URL: https://issues.apache.org/jira/browse/HIVE-8485
 Project: Hive
  Issue Type: Bug
  Components: Metastore
 Environment: Oracle as metastore DB
Reporter: Ryan Pridgeon


Oracle does not distinguish between empty strings and NULL,which proves 
problematic for DataNucleus.
In the event a user creates a table with some property stored as an empty 
string the table will no longer be accessible.
i.e. TBLPROPERTIES ('serialization.null.format'='')
If they try to select, describe, drop, etc the client prints the following 
exception.
ERROR ql.Driver: FAILED: SemanticException [Error 10001]: Table not found 
table name
The work around for this was to go into the hive metastore on the Oracle 
database and replace NULL with some other string. Users could then drop the 
tables or alter their data to use the new null format they just set.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8476) JavaDoc updates to HiveEndPoint.newConnection() for secure streaming with Kerberos

2014-10-16 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8476?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14173730#comment-14173730
 ] 

Hive QA commented on HIVE-8476:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12675131/HIVE-8476.patch

{color:red}ERROR:{color} -1 due to 5 failed/errored test(s), 6559 tests executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestHBaseCliDriver.testCliDriver_hbase_custom_key
org.apache.hadoop.hive.cli.TestHBaseCliDriver.testCliDriver_hbase_custom_key2
org.apache.hadoop.hive.cli.TestHBaseCliDriver.testCliDriver_hbase_custom_key3
org.apache.hive.beeline.TestSchemaTool.testSchemaInit
org.apache.hive.beeline.TestSchemaTool.testSchemaUpgrade
{noformat}

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/1299/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/1299/console
Test logs: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-1299/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 5 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12675131
 - PreCommit-HIVE-TRUNK-Build

 JavaDoc updates to HiveEndPoint.newConnection() for secure streaming with 
 Kerberos
 --

 Key: HIVE-8476
 URL: https://issues.apache.org/jira/browse/HIVE-8476
 Project: Hive
  Issue Type: Bug
  Components: HCatalog
Affects Versions: 0.13.1
Reporter: Roshan Naik
Assignee: Roshan Naik
 Fix For: 0.14.0

 Attachments: HIVE-8476.patch


 Add additional notes on using kerberos authenticated streaming connection in 
 HiveEndPoint.newConnection() method



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-7898) HCatStorer should ignore namespaces generated by Pig

2014-10-16 Thread Justin Leet (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-7898?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14173733#comment-14173733
 ] 

Justin Leet commented on HIVE-7898:
---

Anybody willing to review this? https://reviews.apache.org/r/25140/


 HCatStorer should ignore namespaces generated by Pig
 

 Key: HIVE-7898
 URL: https://issues.apache.org/jira/browse/HIVE-7898
 Project: Hive
  Issue Type: Improvement
  Components: HCatalog
Affects Versions: 0.13.1
Reporter: Justin Leet
Assignee: Justin Leet
Priority: Minor
 Attachments: HIVE-7898.1.patch


 Currently, Pig aliases must exactly match the names of HCat columns for 
 HCatStorer to be successful.  However, several Pig operations prepend a 
 namespace to the alias in order to differentiate fields (e.g. after a group 
 with field b, you might have A::b).  In this case, even if the fields are in 
 the right order and the alias without namespace matches, the store will fail 
 because it tries to match the long form of the alias, despite the namespace 
 being extraneous information in this case.   Note that multiple aliases can 
 be applied (e.g. A::B::C::d).
 A workaround is possible by doing a 
 FOREACH relation GENERATE field1 AS field1, field2 AS field2, etc.  
 This quickly becomes tedious and bloated for tables with many fields.
 Changing this would normally require care around columns named, for example, 
 `A::b` as has been introduced in Hive 13.  However, a different function call 
 only validates Pig aliases if they follow the old rules for Hive columns.  As 
 such, a direct change (rather than attempting to match either the 
 namespace::alias or just alias) maintains compatibility for now.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8448) Union All might not work due to the type conversion issue

2014-10-16 Thread Yongzhi Chen (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8448?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14173744#comment-14173744
 ] 

Yongzhi Chen commented on HIVE-8448:


[~szehon] and [~jdere], I checked the failure tests. 2 of them are new ones, 
but it related to login so it should not be caused by the change.
The rest of the failures are aged more than 2 days, and do not have union 
operations in the queries. So I think the all the test failures
are not related to the change.
Thanks.

 Union All might not work due to the type conversion issue
 -

 Key: HIVE-8448
 URL: https://issues.apache.org/jira/browse/HIVE-8448
 Project: Hive
  Issue Type: Bug
Affects Versions: 0.13.1
Reporter: Chaoyu Tang
Assignee: Yongzhi Chen
Priority: Minor
 Attachments: HIVE-8448.4.patch


 create table t1 (val date);
 insert overwrite table t1 select '2014-10-10' from src limit 1;
 create table t2 (val varchar(10));
 insert overwrite table t2 select '2014-10-10' from src limit 1; 
 ==
 Query:
 select t.val from
 (select val from t1
 union all
 select val from t1
 union all
 select val from t2
 union all
 select val from t1) t;
 ==
 Will throw exception: 
 {code}
 Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: Incompatible 
 types for union operator
   at 
 org.apache.hadoop.hive.ql.exec.UnionOperator.initializeOp(UnionOperator.java:86)
   at org.apache.hadoop.hive.ql.exec.Operator.initialize(Operator.java:380)
   at org.apache.hadoop.hive.ql.exec.Operator.initialize(Operator.java:464)
   at 
 org.apache.hadoop.hive.ql.exec.Operator.initializeChildren(Operator.java:420)
   at 
 org.apache.hadoop.hive.ql.exec.SelectOperator.initializeOp(SelectOperator.java:65)
   at org.apache.hadoop.hive.ql.exec.Operator.initialize(Operator.java:380)
   at org.apache.hadoop.hive.ql.exec.Operator.initialize(Operator.java:464)
   at 
 org.apache.hadoop.hive.ql.exec.Operator.initializeChildren(Operator.java:420)
   at 
 org.apache.hadoop.hive.ql.exec.TableScanOperator.initializeOp(TableScanOperator.java:193)
   at org.apache.hadoop.hive.ql.exec.Operator.initialize(Operator.java:380)
   at 
 org.apache.hadoop.hive.ql.exec.MapOperator.initializeOp(MapOperator.java:443)
   at org.apache.hadoop.hive.ql.exec.Operator.initialize(Operator.java:380)
   at 
 org.apache.hadoop.hive.ql.exec.mr.ExecMapper.configure(ExecMapper.java:133)
   ... 22 more
 {code}
 It was because at this query parse step, getCommonClassForUnionAll is used, 
 but at execution getCommonClass is used. They are not used consistently in 
 union. The later one does not support the implicit conversion from date to 
 string, which is the problem cause.
 The change might be simple to fix this particular union issue but I noticed 
 that there are three versions of getCommonClass: getCommonClass, 
 getCommonClassForComparison, getCommonClassForUnionAll, and wonder if they 
 need to be cleaned and refactored.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-7205) Wrong results when union all of grouping followed by group by with correlation optimization

2014-10-16 Thread Yin Huai (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-7205?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14173804#comment-14173804
 ] 

Yin Huai commented on HIVE-7205:


[~navis] Can you update the review board? I will take a look. Thank you.

 Wrong results when union all of grouping followed by group by with 
 correlation optimization
 ---

 Key: HIVE-7205
 URL: https://issues.apache.org/jira/browse/HIVE-7205
 Project: Hive
  Issue Type: Bug
Affects Versions: 0.12.0, 0.13.0, 0.13.1
Reporter: dima machlin
Assignee: Navis
Priority: Critical
 Attachments: HIVE-7205.1.patch.txt, HIVE-7205.2.patch.txt, 
 HIVE-7205.3.patch.txt, HIVE-7205.4.patch.txt


 use case :
 table TBL (a string,b string) contains single row : 'a','a'
 the following query :
 {code:sql}
 select b, sum(cc) from (
 select b,count(1) as cc from TBL group by b
 union all
 select a as b,count(1) as cc from TBL group by a
 ) z
 group by b
 {code}
 returns 
 a 1
 a 1
 while set hive.optimize.correlation=true;
 if we change set hive.optimize.correlation=false;
 it returns correct results : a 2
 The plan with correlation optimization :
 {code:sql}
 ABSTRACT SYNTAX TREE:
   (TOK_QUERY (TOK_FROM (TOK_SUBQUERY (TOK_UNION (TOK_QUERY (TOK_FROM 
 (TOK_TABREF (TOK_TABNAME DB TBL))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR 
 TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_TABLE_OR_COL b)) (TOK_SELEXPR 
 (TOK_FUNCTION count 1) cc)) (TOK_GROUPBY (TOK_TABLE_OR_COL b (TOK_QUERY 
 (TOK_FROM (TOK_TABREF (TOK_TABNAME DB TBL))) (TOK_INSERT (TOK_DESTINATION 
 (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_TABLE_OR_COL a) b) 
 (TOK_SELEXPR (TOK_FUNCTION count 1) cc)) (TOK_GROUPBY (TOK_TABLE_OR_COL 
 a) z)) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT 
 (TOK_SELEXPR (TOK_TABLE_OR_COL b)) (TOK_SELEXPR (TOK_FUNCTION sum 
 (TOK_TABLE_OR_COL cc (TOK_GROUPBY (TOK_TABLE_OR_COL b
 STAGE DEPENDENCIES:
   Stage-1 is a root stage
   Stage-0 is a root stage
 STAGE PLANS:
   Stage: Stage-1
 Map Reduce
   Alias - Map Operator Tree:
 null-subquery1:z-subquery1:TBL 
   TableScan
 alias: TBL
 Select Operator
   expressions:
 expr: b
 type: string
   outputColumnNames: b
   Group By Operator
 aggregations:
   expr: count(1)
 bucketGroup: false
 keys:
   expr: b
   type: string
 mode: hash
 outputColumnNames: _col0, _col1
 Reduce Output Operator
   key expressions:
 expr: _col0
 type: string
   sort order: +
   Map-reduce partition columns:
 expr: _col0
 type: string
   tag: 0
   value expressions:
 expr: _col1
 type: bigint
 null-subquery2:z-subquery2:TBL 
   TableScan
 alias: TBL
 Select Operator
   expressions:
 expr: a
 type: string
   outputColumnNames: a
   Group By Operator
 aggregations:
   expr: count(1)
 bucketGroup: false
 keys:
   expr: a
   type: string
 mode: hash
 outputColumnNames: _col0, _col1
 Reduce Output Operator
   key expressions:
 expr: _col0
 type: string
   sort order: +
   Map-reduce partition columns:
 expr: _col0
 type: string
   tag: 1
   value expressions:
 expr: _col1
 type: bigint
   Reduce Operator Tree:
 Demux Operator
   Group By Operator
 aggregations:
   expr: count(VALUE._col0)
 bucketGroup: false
 keys:
   expr: KEY._col0
   type: string
 mode: mergepartial
 outputColumnNames: _col0, _col1
 Select Operator
   expressions:
 expr: _col0
 type: string
 expr: _col1
 type: bigint
   outputColumnNames: _col0, _col1
   Union
 Select Operator
   expressions:
   

[jira] [Updated] (HIVE-8428) PCR doesnt remove filters involving casts

2014-10-16 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-8428?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-8428:
---
Fix Version/s: 0.15.0

 PCR doesnt remove filters involving casts
 -

 Key: HIVE-8428
 URL: https://issues.apache.org/jira/browse/HIVE-8428
 Project: Hive
  Issue Type: Improvement
  Components: Logical Optimizer
Affects Versions: 0.11.0, 0.12.0, 0.13.0
Reporter: Ashutosh Chauhan
Assignee: Ashutosh Chauhan
 Fix For: 0.15.0

 Attachments: HIVE-8428.1.patch, HIVE-8428.2.patch, HIVE-8428.3.patch, 
 HIVE-8428.patch


 e.g.,
 select key,value from srcpart where hr = cast(11 as double);



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-8428) PCR doesnt remove filters involving casts

2014-10-16 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-8428?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-8428:
---
Resolution: Fixed
Status: Resolved  (was: Patch Available)

Cant repro udaf_context_histogram which anyways should not be impacted by this 
change. Committed to trunk.
[~vikram.dixit] This will be good to have in 0.14 as well.

 PCR doesnt remove filters involving casts
 -

 Key: HIVE-8428
 URL: https://issues.apache.org/jira/browse/HIVE-8428
 Project: Hive
  Issue Type: Improvement
  Components: Logical Optimizer
Affects Versions: 0.11.0, 0.12.0, 0.13.0
Reporter: Ashutosh Chauhan
Assignee: Ashutosh Chauhan
 Attachments: HIVE-8428.1.patch, HIVE-8428.2.patch, HIVE-8428.3.patch, 
 HIVE-8428.patch


 e.g.,
 select key,value from srcpart where hr = cast(11 as double);



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8464) Vectorized reducer nested group by query returns wrong results

2014-10-16 Thread Ashutosh Chauhan (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8464?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14173860#comment-14173860
 ] 

Ashutosh Chauhan commented on HIVE-8464:


+1

 Vectorized reducer nested group by query returns wrong results
 --

 Key: HIVE-8464
 URL: https://issues.apache.org/jira/browse/HIVE-8464
 Project: Hive
  Issue Type: Bug
Affects Versions: 0.14.0
Reporter: Matt McCline
Assignee: Matt McCline
Priority: Critical
 Fix For: 0.14.0

 Attachments: HIVE-8464.01.patch


 Query has 2 Group By operators in reduce-stage and VectorGroupByOperator can 
 only handle one Group By operator being fed key batches by reduce-shuffle.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HIVE-8486) TPC-DS Query 96 parallelism is not set correcly

2014-10-16 Thread Brock Noland (JIRA)
Brock Noland created HIVE-8486:
--

 Summary: TPC-DS Query 96 parallelism is not set correcly
 Key: HIVE-8486
 URL: https://issues.apache.org/jira/browse/HIVE-8486
 Project: Hive
  Issue Type: Sub-task
Reporter: Brock Noland
Assignee: Chao


When we run the query on a 20B we only have a parallelism factor of 1.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-8477) Update timestamp in status console [Spark Branch]

2014-10-16 Thread Brock Noland (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-8477?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Brock Noland updated HIVE-8477:
---
Attachment: HIVE-8477-spark.patch

 Update timestamp in status console [Spark Branch]
 -

 Key: HIVE-8477
 URL: https://issues.apache.org/jira/browse/HIVE-8477
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Affects Versions: spark-branch
Reporter: Brock Noland
Assignee: Brock Noland
 Attachments: HIVE-8477-spark.patch, HIVE-8477.patch


 Note our time versus the log time
 {noformat}
 2014-10-15 19:18:07,646 INFO  [main]: status.SparkJobMonitor 
 (SessionState.java:printInfo(820)) - 2014-18-15 07:18:07,646 Stage-0_0: 
 2(+9)/11 
 {noformat}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8456) Support Hive Counter to collect spark job metric[Spark Branch]

2014-10-16 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8456?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14173925#comment-14173925
 ] 

Hive QA commented on HIVE-8456:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12675227/HIVE-8456.3-spark.patch

{color:red}ERROR:{color} -1 due to 2 failed/errored test(s), 6769 tests executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_sample_islocalmode_hook
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_tez_smb_1
{noformat}

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-SPARK-Build/223/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-SPARK-Build/223/console
Test logs: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-SPARK-Build-223/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 2 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12675227

 Support Hive Counter to collect spark job metric[Spark Branch]
 --

 Key: HIVE-8456
 URL: https://issues.apache.org/jira/browse/HIVE-8456
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Reporter: Chengxiang Li
Assignee: Chengxiang Li
  Labels: Spark-M3
 Attachments: HIVE-8456.1-spark.patch, HIVE-8456.2-spark.patch, 
 HIVE-8456.3-spark.patch


 Several Hive query metric in Hive operators is collected by Hive Counter, 
 such as CREATEDFILES and DESERIALIZE_ERRORS, Besides, Hive use Counter as an 
 option to collect table stats info.  Spark support Accumulator which is 
 pretty similiar with Hive Counter, we could try to enable Hive Counter based 
 on it.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-7366) getDatabase using direct sql

2014-10-16 Thread Sushanth Sowmyan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-7366?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sushanth Sowmyan updated HIVE-7366:
---
Resolution: Fixed
Status: Resolved  (was: Patch Available)

Thanks, Sergey, committed to 0.14 and trunk.

 getDatabase using direct sql
 

 Key: HIVE-7366
 URL: https://issues.apache.org/jira/browse/HIVE-7366
 Project: Hive
  Issue Type: Bug
  Components: Metastore
Affects Versions: 0.14.0
Reporter: Sushanth Sowmyan
Assignee: Sushanth Sowmyan
Priority: Critical
 Fix For: 0.14.0

 Attachments: HIVE-7366.2.patch, HIVE-7366.3.patch, HIVE-7366.4.patch, 
 HIVE-7366.patch


 Given that get_database is easily one of the most frequent calls made on the 
 metastore, we should have the ability to bypass datanucleus for that, and use 
 direct SQL instead.
 This was something that I did initially as part of debugging HIVE-7368, but I 
 think that given the frequency of this call, it's useful to have it in 
 mainline direct sql.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8389) Fix CBO when indexes are used

2014-10-16 Thread Pengcheng Xiong (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8389?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14174000#comment-14174000
 ] 

Pengcheng Xiong commented on HIVE-8389:
---

Done. Thanks [~leftylev] 

 Fix CBO when indexes are used
 -

 Key: HIVE-8389
 URL: https://issues.apache.org/jira/browse/HIVE-8389
 Project: Hive
  Issue Type: Bug
Reporter: Pengcheng Xiong
Assignee: Pengcheng Xiong
  Labels: TODOC14
 Fix For: 0.15.0

 Attachments: HIVE-8389.1.patch, HIVE-8389.2.patch, HIVE-8389.3.patch


 Disable CBO when indexes are used



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8484) HCatalog throws an exception if Pig job is of type 'fetch'

2014-10-16 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8484?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14174004#comment-14174004
 ] 

Hive QA commented on HIVE-8484:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12675260/HIVE-8484.patch

{color:red}ERROR:{color} -1 due to 5 failed/errored test(s), 6559 tests executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestHBaseCliDriver.testCliDriver_hbase_custom_key
org.apache.hadoop.hive.cli.TestHBaseCliDriver.testCliDriver_hbase_custom_key2
org.apache.hadoop.hive.cli.TestHBaseCliDriver.testCliDriver_hbase_custom_key3
org.apache.hive.beeline.TestSchemaTool.testSchemaInit
org.apache.hive.beeline.TestSchemaTool.testSchemaUpgrade
{noformat}

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/1300/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/1300/console
Test logs: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-1300/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 5 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12675260
 - PreCommit-HIVE-TRUNK-Build

 HCatalog throws an exception if Pig job is of type 'fetch'
 --

 Key: HIVE-8484
 URL: https://issues.apache.org/jira/browse/HIVE-8484
 Project: Hive
  Issue Type: Bug
  Components: HCatalog
Affects Versions: 0.14.0
Reporter: Lorand Bendig
 Fix For: 0.14.0

 Attachments: HIVE-8484.patch


 When Pig tries to retrieve result in fetch mode through HCatalog then 
 HCatLoader#setLocation(String location, Job job) can't set the outputschema 
 because HCatUtil#checkJobContextIfRunningFromBackend(job) always returns 
 false :
 {code}
 public static boolean checkJobContextIfRunningFromBackend(JobContext j) {
 if (j.getConfiguration().get(mapred.task.id, ).equals() 
 !(true.equals(j.getConfiguration().get(pig.illustrating {
   return false;
 }
 return true;
   }
 {code}
 This is because in fetch mode we don't have a mapred.task.id. A null 
 outputschema will raise an exception when HCatBaseLoader#getNext() is called: 
 (ERROR 6018: Error converting read value to tuple).



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-8448) Union All might not work due to the type conversion issue

2014-10-16 Thread Szehon Ho (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-8448?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Szehon Ho updated HIVE-8448:

   Resolution: Fixed
Fix Version/s: 0.15.0
   Status: Resolved  (was: Patch Available)

Makes sense.  Committed to trunk, thanks Yongzhi for the contribution and Jason 
for additional review !

 Union All might not work due to the type conversion issue
 -

 Key: HIVE-8448
 URL: https://issues.apache.org/jira/browse/HIVE-8448
 Project: Hive
  Issue Type: Bug
Affects Versions: 0.13.1
Reporter: Chaoyu Tang
Assignee: Yongzhi Chen
Priority: Minor
 Fix For: 0.15.0

 Attachments: HIVE-8448.4.patch


 create table t1 (val date);
 insert overwrite table t1 select '2014-10-10' from src limit 1;
 create table t2 (val varchar(10));
 insert overwrite table t2 select '2014-10-10' from src limit 1; 
 ==
 Query:
 select t.val from
 (select val from t1
 union all
 select val from t1
 union all
 select val from t2
 union all
 select val from t1) t;
 ==
 Will throw exception: 
 {code}
 Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: Incompatible 
 types for union operator
   at 
 org.apache.hadoop.hive.ql.exec.UnionOperator.initializeOp(UnionOperator.java:86)
   at org.apache.hadoop.hive.ql.exec.Operator.initialize(Operator.java:380)
   at org.apache.hadoop.hive.ql.exec.Operator.initialize(Operator.java:464)
   at 
 org.apache.hadoop.hive.ql.exec.Operator.initializeChildren(Operator.java:420)
   at 
 org.apache.hadoop.hive.ql.exec.SelectOperator.initializeOp(SelectOperator.java:65)
   at org.apache.hadoop.hive.ql.exec.Operator.initialize(Operator.java:380)
   at org.apache.hadoop.hive.ql.exec.Operator.initialize(Operator.java:464)
   at 
 org.apache.hadoop.hive.ql.exec.Operator.initializeChildren(Operator.java:420)
   at 
 org.apache.hadoop.hive.ql.exec.TableScanOperator.initializeOp(TableScanOperator.java:193)
   at org.apache.hadoop.hive.ql.exec.Operator.initialize(Operator.java:380)
   at 
 org.apache.hadoop.hive.ql.exec.MapOperator.initializeOp(MapOperator.java:443)
   at org.apache.hadoop.hive.ql.exec.Operator.initialize(Operator.java:380)
   at 
 org.apache.hadoop.hive.ql.exec.mr.ExecMapper.configure(ExecMapper.java:133)
   ... 22 more
 {code}
 It was because at this query parse step, getCommonClassForUnionAll is used, 
 but at execution getCommonClass is used. They are not used consistently in 
 union. The later one does not support the implicit conversion from date to 
 string, which is the problem cause.
 The change might be simple to fix this particular union issue but I noticed 
 that there are three versions of getCommonClass: getCommonClass, 
 getCommonClassForComparison, getCommonClassForUnionAll, and wonder if they 
 need to be cleaned and refactored.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8477) Update timestamp in status console [Spark Branch]

2014-10-16 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8477?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14174006#comment-14174006
 ] 

Hive QA commented on HIVE-8477:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12675290/HIVE-8477-spark.patch

{color:red}ERROR:{color} -1 due to 2 failed/errored test(s), 6769 tests executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_sample_islocalmode_hook
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_tez_smb_1
{noformat}

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-SPARK-Build/224/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-SPARK-Build/224/console
Test logs: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-SPARK-Build-224/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 2 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12675290

 Update timestamp in status console [Spark Branch]
 -

 Key: HIVE-8477
 URL: https://issues.apache.org/jira/browse/HIVE-8477
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Affects Versions: spark-branch
Reporter: Brock Noland
Assignee: Brock Noland
 Attachments: HIVE-8477-spark.patch, HIVE-8477.patch


 Note our time versus the log time
 {noformat}
 2014-10-15 19:18:07,646 INFO  [main]: status.SparkJobMonitor 
 (SessionState.java:printInfo(820)) - 2014-18-15 07:18:07,646 Stage-0_0: 
 2(+9)/11 
 {noformat}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8477) Update timestamp in status console [Spark Branch]

2014-10-16 Thread Xuefu Zhang (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8477?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14174012#comment-14174012
 ] 

Xuefu Zhang commented on HIVE-8477:
---

+1

 Update timestamp in status console [Spark Branch]
 -

 Key: HIVE-8477
 URL: https://issues.apache.org/jira/browse/HIVE-8477
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Affects Versions: spark-branch
Reporter: Brock Noland
Assignee: Brock Noland
 Attachments: HIVE-8477-spark.patch, HIVE-8477.patch


 Note our time versus the log time
 {noformat}
 2014-10-15 19:18:07,646 INFO  [main]: status.SparkJobMonitor 
 (SessionState.java:printInfo(820)) - 2014-18-15 07:18:07,646 Stage-0_0: 
 2(+9)/11 
 {noformat}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-8456) Support Hive Counter to collect spark job metric[Spark Branch]

2014-10-16 Thread Szehon Ho (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-8456?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Szehon Ho updated HIVE-8456:

   Resolution: Fixed
Fix Version/s: spark-branch
   Status: Resolved  (was: Patch Available)

Committed to spark, you can begin the work to use these counters.

Thanks Chengxiang for the contribution, and Rui for review!

 Support Hive Counter to collect spark job metric[Spark Branch]
 --

 Key: HIVE-8456
 URL: https://issues.apache.org/jira/browse/HIVE-8456
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Reporter: Chengxiang Li
Assignee: Chengxiang Li
  Labels: Spark-M3
 Fix For: spark-branch

 Attachments: HIVE-8456.1-spark.patch, HIVE-8456.2-spark.patch, 
 HIVE-8456.3-spark.patch


 Several Hive query metric in Hive operators is collected by Hive Counter, 
 such as CREATEDFILES and DESERIALIZE_ERRORS, Besides, Hive use Counter as an 
 option to collect table stats info.  Spark support Accumulator which is 
 pretty similiar with Hive Counter, we could try to enable Hive Counter based 
 on it.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-8389) Fix CBO when indexes are used

2014-10-16 Thread Sergey Shelukhin (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-8389?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sergey Shelukhin updated HIVE-8389:
---
Fix Version/s: (was: 0.15.0)
   0.14.0

 Fix CBO when indexes are used
 -

 Key: HIVE-8389
 URL: https://issues.apache.org/jira/browse/HIVE-8389
 Project: Hive
  Issue Type: Bug
Reporter: Pengcheng Xiong
Assignee: Pengcheng Xiong
  Labels: TODOC14
 Fix For: 0.14.0

 Attachments: HIVE-8389.1.patch, HIVE-8389.2.patch, HIVE-8389.3.patch


 Disable CBO when indexes are used



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8389) Fix CBO when indexes are used

2014-10-16 Thread Sergey Shelukhin (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8389?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14174033#comment-14174033
 ] 

Sergey Shelukhin commented on HIVE-8389:


committed to 14

 Fix CBO when indexes are used
 -

 Key: HIVE-8389
 URL: https://issues.apache.org/jira/browse/HIVE-8389
 Project: Hive
  Issue Type: Bug
Reporter: Pengcheng Xiong
Assignee: Pengcheng Xiong
  Labels: TODOC14
 Fix For: 0.14.0

 Attachments: HIVE-8389.1.patch, HIVE-8389.2.patch, HIVE-8389.3.patch


 Disable CBO when indexes are used



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-8460) ORC SARG literal creation for double from float may lead to wrong evaluation of SARG

2014-10-16 Thread Prasanth J (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-8460?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Prasanth J updated HIVE-8460:
-
Attachment: HIVE-8460.2.patch

 ORC SARG literal creation for double from float may lead to wrong evaluation 
 of SARG
 

 Key: HIVE-8460
 URL: https://issues.apache.org/jira/browse/HIVE-8460
 Project: Hive
  Issue Type: Bug
  Components: Query Processor
Affects Versions: 0.13.0, 0.13.1
Reporter: Ashutosh Chauhan
Assignee: Prasanth J
 Attachments: HIVE-8460.1.patch, HIVE-8460.2.patch


 e.g., expression like d = 0.22 where 0.22 is of float type, gets converted to 
 double d = 0.21986 in SARG creation. This will cause erroneous evaluation 
 of SARG expression.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8460) ORC SARG literal creation for double from float may lead to wrong evaluation of SARG

2014-10-16 Thread Ashutosh Chauhan (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8460?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14174086#comment-14174086
 ] 

Ashutosh Chauhan commented on HIVE-8460:


+1

 ORC SARG literal creation for double from float may lead to wrong evaluation 
 of SARG
 

 Key: HIVE-8460
 URL: https://issues.apache.org/jira/browse/HIVE-8460
 Project: Hive
  Issue Type: Bug
  Components: Query Processor
Affects Versions: 0.13.0, 0.13.1
Reporter: Ashutosh Chauhan
Assignee: Prasanth J
 Attachments: HIVE-8460.1.patch, HIVE-8460.2.patch


 e.g., expression like d = 0.22 where 0.22 is of float type, gets converted to 
 double d = 0.21986 in SARG creation. This will cause erroneous evaluation 
 of SARG expression.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8460) ORC SARG literal creation for double from float may lead to wrong evaluation of SARG

2014-10-16 Thread Ashutosh Chauhan (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8460?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14174090#comment-14174090
 ] 

Ashutosh Chauhan commented on HIVE-8460:


[~vikram.dixit] It will be good to have this in 0.14 as well.

 ORC SARG literal creation for double from float may lead to wrong evaluation 
 of SARG
 

 Key: HIVE-8460
 URL: https://issues.apache.org/jira/browse/HIVE-8460
 Project: Hive
  Issue Type: Bug
  Components: Query Processor
Affects Versions: 0.13.0, 0.13.1
Reporter: Ashutosh Chauhan
Assignee: Prasanth J
 Attachments: HIVE-8460.1.patch, HIVE-8460.2.patch


 e.g., expression like d = 0.22 where 0.22 is of float type, gets converted to 
 double d = 0.21986 in SARG creation. This will cause erroneous evaluation 
 of SARG expression.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8428) PCR doesnt remove filters involving casts

2014-10-16 Thread Vikram Dixit K (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8428?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14174122#comment-14174122
 ] 

Vikram Dixit K commented on HIVE-8428:
--

+1 for 0.14

 PCR doesnt remove filters involving casts
 -

 Key: HIVE-8428
 URL: https://issues.apache.org/jira/browse/HIVE-8428
 Project: Hive
  Issue Type: Improvement
  Components: Logical Optimizer
Affects Versions: 0.11.0, 0.12.0, 0.13.0
Reporter: Ashutosh Chauhan
Assignee: Ashutosh Chauhan
 Fix For: 0.15.0

 Attachments: HIVE-8428.1.patch, HIVE-8428.2.patch, HIVE-8428.3.patch, 
 HIVE-8428.patch


 e.g.,
 select key,value from srcpart where hr = cast(11 as double);



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-8395) CBO: enable by default

2014-10-16 Thread Sergey Shelukhin (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-8395?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sergey Shelukhin updated HIVE-8395:
---
Attachment: HIVE-8395.07.patch

Update with changes to out files for recently fixed issues.

 CBO: enable by default
 --

 Key: HIVE-8395
 URL: https://issues.apache.org/jira/browse/HIVE-8395
 Project: Hive
  Issue Type: Sub-task
  Components: CBO
Reporter: Sergey Shelukhin
Assignee: Sergey Shelukhin
 Fix For: 0.15.0

 Attachments: HIVE-8395.01.patch, HIVE-8395.02.patch, 
 HIVE-8395.03.patch, HIVE-8395.04.patch, HIVE-8395.05.patch, 
 HIVE-8395.06.patch, HIVE-8395.07.patch, HIVE-8395.patch






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HIVE-8487) CBO causes hbase_custom_key tests to fail

2014-10-16 Thread Sergey Shelukhin (JIRA)
Sergey Shelukhin created HIVE-8487:
--

 Summary: CBO causes hbase_custom_key tests to fail
 Key: HIVE-8487
 URL: https://issues.apache.org/jira/browse/HIVE-8487
 Project: Hive
  Issue Type: Bug
Reporter: Sergey Shelukhin
Priority: Critical


{noformat}
FAILED: SemanticException java.lang.NullPointerException
org.apache.hadoop.hive.ql.parse.SemanticException: 
java.lang.NullPointerException
at 
org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genFileSinkPlan(SemanticAnalyzer.java:6291)
at 
org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genPostGroupByBodyPlan(SemanticAnalyzer.java:8899)
at 
org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genBodyPlan(SemanticAnalyzer.java:8790)
at 
org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genPlan(SemanticAnalyzer.java:9609)
at 
org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genPlan(SemanticAnalyzer.java:9503)
at 
org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.analyzeInternal(SemanticAnalyzer.java:9960)
at 
org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer.analyze(BaseSemanticAnalyzer.java:221)
at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:417)
at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:303)
at org.apache.hadoop.hive.ql.Driver.compileInternal(Driver.java:1070)
at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1132)
at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1007)
at org.apache.hadoop.hive.ql.Driver.run(Driver.java:997)
at 
org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:247)
at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:199)
at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:410)
at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:345)
at org.apache.hadoop.hive.ql.QTestUtil.executeClient(QTestUtil.java:832)
at 
org.apache.hadoop.hive.cli.TestHBaseCliDriver.runTest(TestHBaseCliDriver.java:112)
at 
org.apache.hadoop.hive.cli.TestHBaseCliDriver.testCliDriver_hbase_custom_key(TestHBaseCliDriver.java:94)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:597)
at junit.framework.TestCase.runTest(TestCase.java:168)
at junit.framework.TestCase.runBare(TestCase.java:134)
at junit.framework.TestResult$1.protect(TestResult.java:110)
at junit.framework.TestResult.runProtected(TestResult.java:128)
at junit.framework.TestResult.run(TestResult.java:113)
at junit.framework.TestCase.run(TestCase.java:124)
at junit.framework.TestSuite.runTest(TestSuite.java:243)
at junit.framework.TestSuite.run(TestSuite.java:238)
at junit.extensions.TestDecorator.basicRun(TestDecorator.java:24)
at junit.extensions.TestSetup$1.protect(TestSetup.java:23)
at junit.framework.TestResult.runProtected(TestResult.java:128)
at junit.extensions.TestSetup.run(TestSetup.java:27)
at 
org.junit.internal.runners.JUnit38ClassRunner.run(JUnit38ClassRunner.java:83)
at 
org.apache.maven.surefire.junit4.JUnit4Provider.execute(JUnit4Provider.java:264)
at 
org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:153)
at 
org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:124)
at 
org.apache.maven.surefire.booter.ForkedBooter.invokeProviderInSameClassLoader(ForkedBooter.java:200)
at 
org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:153)
at 
org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:103)
Caused by: org.apache.hadoop.hive.serde2.SerDeException: 
java.lang.NullPointerException
at 
org.apache.hadoop.hive.hbase.HBaseSerDeParameters.initKeyFactory(HBaseSerDeParameters.java:178)
at 
org.apache.hadoop.hive.hbase.HBaseSerDeParameters.init(HBaseSerDeParameters.java:103)
at 
org.apache.hadoop.hive.hbase.HBaseSerDe.initialize(HBaseSerDe.java:93)
at 
org.apache.hadoop.hive.serde2.AbstractSerDe.initialize(AbstractSerDe.java:51)
at 
org.apache.hadoop.hive.serde2.SerDeUtils.initializeSerDe(SerDeUtils.java:519)
at 
org.apache.hadoop.hive.ql.plan.TableDesc.getDeserializer(TableDesc.java:84)
at 
org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genFileSinkPlan(SemanticAnalyzer.java:6281)
... 42 more
Caused by: java.lang.NullPointerException
at 
org.apache.hadoop.hive.hbase.HBaseSerDeParameters.createKeyFactory(HBaseSerDeParameters.java:186)
at 

[jira] [Updated] (HIVE-6715) Hive JDBC should include username into open session request for non-sasl connection

2014-10-16 Thread Thejas M Nair (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6715?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Thejas M Nair updated HIVE-6715:

Resolution: Fixed
Status: Resolved  (was: Patch Available)

Patch committed to trunk and 0.14 branch.
Thanks  [~prasadm]!


 Hive JDBC should include username into open session request for non-sasl 
 connection
 ---

 Key: HIVE-6715
 URL: https://issues.apache.org/jira/browse/HIVE-6715
 Project: Hive
  Issue Type: Bug
  Components: JDBC
Reporter: Srinath
Assignee: Prasad Mujumdar
Priority: Critical
 Fix For: 0.14.0

 Attachments: HIVE-6715.1.patch, HIVE-6715.2.patch, HIVE-6715.3.patch


 The only parameter from sessVars that's being set in 
 HiveConnection.openSession() is HS2_PROXY_USER. 
 HIVE_AUTH_USER must also be set.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HIVE-8488) hash() doesn't match between string and char/varchar

2014-10-16 Thread Jason Dere (JIRA)
Jason Dere created HIVE-8488:


 Summary: hash() doesn't match between string and char/varchar
 Key: HIVE-8488
 URL: https://issues.apache.org/jira/browse/HIVE-8488
 Project: Hive
  Issue Type: Bug
  Components: UDF
Reporter: Jason Dere
Assignee: Jason Dere


{noformat}
hive select * from tab1;
OK
val_484 val_484 val_484

hive select hash(c1), hash(c2), hash(c3) from tab1;
OK
230901778   1973712113  1973712113
{noformat}

This may throw off users expecting string/varchar/char types to be fairly 
interchangeable.




--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8488) hash() doesn't match between string and char/varchar

2014-10-16 Thread Jason Dere (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8488?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14174151#comment-14174151
 ] 

Jason Dere commented on HIVE-8488:
--

hash() uses ObjectInspectorUtils.hashCode(), which has special case handling 
for Text so that the hashcode matches String.hashCode(). This special case 
handling should have been done for char/varchar as well.
Note that fixing this will break existing behavior of hash() for char/varchar.

 hash() doesn't match between string and char/varchar
 

 Key: HIVE-8488
 URL: https://issues.apache.org/jira/browse/HIVE-8488
 Project: Hive
  Issue Type: Bug
  Components: UDF
Reporter: Jason Dere
Assignee: Jason Dere

 {noformat}
 hive select * from tab1;
 OK
 val_484   val_484 val_484
 hive select hash(c1), hash(c2), hash(c3) from tab1;
 OK
 230901778 1973712113  1973712113
 {noformat}
 This may throw off users expecting string/varchar/char types to be fairly 
 interchangeable.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-8488) hash() doesn't match between string and char/varchar

2014-10-16 Thread Jason Dere (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-8488?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jason Dere updated HIVE-8488:
-
Attachment: HIVE-8488.1.patch

 hash() doesn't match between string and char/varchar
 

 Key: HIVE-8488
 URL: https://issues.apache.org/jira/browse/HIVE-8488
 Project: Hive
  Issue Type: Bug
  Components: UDF
Reporter: Jason Dere
Assignee: Jason Dere
 Attachments: HIVE-8488.1.patch


 {noformat}
 hive select * from tab1;
 OK
 val_484   val_484 val_484
 hive select hash(c1), hash(c2), hash(c3) from tab1;
 OK
 230901778 1973712113  1973712113
 {noformat}
 This may throw off users expecting string/varchar/char types to be fairly 
 interchangeable.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-8488) hash() doesn't match between string and char/varchar

2014-10-16 Thread Jason Dere (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-8488?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jason Dere updated HIVE-8488:
-
Status: Patch Available  (was: Open)

 hash() doesn't match between string and char/varchar
 

 Key: HIVE-8488
 URL: https://issues.apache.org/jira/browse/HIVE-8488
 Project: Hive
  Issue Type: Bug
  Components: UDF
Reporter: Jason Dere
Assignee: Jason Dere
 Attachments: HIVE-8488.1.patch


 {noformat}
 hive select * from tab1;
 OK
 val_484   val_484 val_484
 hive select hash(c1), hash(c2), hash(c3) from tab1;
 OK
 230901778 1973712113  1973712113
 {noformat}
 This may throw off users expecting string/varchar/char types to be fairly 
 interchangeable.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-8341) Transaction information in config file can grow excessively large

2014-10-16 Thread Alan Gates (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-8341?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alan Gates updated HIVE-8341:
-
Status: Open  (was: Patch Available)

The TestOperators failure is caused by this patch.  The rest I believe are 
unrelated.  I'll put up a new version of the patch that addresses the 
TestOperators failure.

 Transaction information in config file can grow excessively large
 -

 Key: HIVE-8341
 URL: https://issues.apache.org/jira/browse/HIVE-8341
 Project: Hive
  Issue Type: Bug
  Components: Transactions
Affects Versions: 0.14.0
Reporter: Alan Gates
Assignee: Alan Gates
Priority: Critical
 Attachments: HIVE-8341.2.patch, HIVE-8341.patch


 In our testing we have seen cases where the transaction list grows very 
 large.  We need a more efficient way of communicating the list.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-8341) Transaction information in config file can grow excessively large

2014-10-16 Thread Alan Gates (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-8341?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alan Gates updated HIVE-8341:
-
Status: Patch Available  (was: Open)

 Transaction information in config file can grow excessively large
 -

 Key: HIVE-8341
 URL: https://issues.apache.org/jira/browse/HIVE-8341
 Project: Hive
  Issue Type: Bug
  Components: Transactions
Affects Versions: 0.14.0
Reporter: Alan Gates
Assignee: Alan Gates
Priority: Critical
 Attachments: HIVE-8341.2.patch, HIVE-8341.3.patch, HIVE-8341.patch


 In our testing we have seen cases where the transaction list grows very 
 large.  We need a more efficient way of communicating the list.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-8341) Transaction information in config file can grow excessively large

2014-10-16 Thread Alan Gates (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-8341?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alan Gates updated HIVE-8341:
-
Attachment: HIVE-8341.3.patch

 Transaction information in config file can grow excessively large
 -

 Key: HIVE-8341
 URL: https://issues.apache.org/jira/browse/HIVE-8341
 Project: Hive
  Issue Type: Bug
  Components: Transactions
Affects Versions: 0.14.0
Reporter: Alan Gates
Assignee: Alan Gates
Priority: Critical
 Attachments: HIVE-8341.2.patch, HIVE-8341.3.patch, HIVE-8341.patch


 In our testing we have seen cases where the transaction list grows very 
 large.  We need a more efficient way of communicating the list.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-8477) Update timestamp in status console [Spark Branch]

2014-10-16 Thread Brock Noland (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-8477?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Brock Noland updated HIVE-8477:
---
   Resolution: Fixed
Fix Version/s: spark-branch
   Status: Resolved  (was: Patch Available)

Thank you! I have committed to spark.

 Update timestamp in status console [Spark Branch]
 -

 Key: HIVE-8477
 URL: https://issues.apache.org/jira/browse/HIVE-8477
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Affects Versions: spark-branch
Reporter: Brock Noland
Assignee: Brock Noland
 Fix For: spark-branch

 Attachments: HIVE-8477-spark.patch, HIVE-8477.patch


 Note our time versus the log time
 {noformat}
 2014-10-15 19:18:07,646 INFO  [main]: status.SparkJobMonitor 
 (SessionState.java:printInfo(820)) - 2014-18-15 07:18:07,646 Stage-0_0: 
 2(+9)/11 
 {noformat}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-8429) Add records in/out counters

2014-10-16 Thread Gunther Hagleitner (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-8429?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gunther Hagleitner updated HIVE-8429:
-
Fix Version/s: 0.14.0

 Add records in/out counters
 ---

 Key: HIVE-8429
 URL: https://issues.apache.org/jira/browse/HIVE-8429
 Project: Hive
  Issue Type: Bug
Reporter: Gunther Hagleitner
Assignee: Gunther Hagleitner
 Fix For: 0.14.0

 Attachments: HIVE-8429.1.patch, HIVE-8429.2.patch, HIVE-8429.3.patch, 
 HIVE-8429.4.patch


 We don't do counters for input/output records right now. That would help for 
 debugging though (if it can be done with minimal overhead).



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8429) Add records in/out counters

2014-10-16 Thread Gunther Hagleitner (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8429?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14174189#comment-14174189
 ] 

Gunther Hagleitner commented on HIVE-8429:
--

Failures are unrelated. [~gopalv] can you take another look?

 Add records in/out counters
 ---

 Key: HIVE-8429
 URL: https://issues.apache.org/jira/browse/HIVE-8429
 Project: Hive
  Issue Type: Bug
Reporter: Gunther Hagleitner
Assignee: Gunther Hagleitner
 Fix For: 0.14.0

 Attachments: HIVE-8429.1.patch, HIVE-8429.2.patch, HIVE-8429.3.patch, 
 HIVE-8429.4.patch


 We don't do counters for input/output records right now. That would help for 
 debugging though (if it can be done with minimal overhead).



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8465) Fix some minor test fails on trunk

2014-10-16 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8465?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14174193#comment-14174193
 ] 

Hive QA commented on HIVE-8465:
---



{color:red}Overall{color}: -1 no tests executed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12675240/HIVE-8465.3.patch.txt

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/1301/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/1301/console
Test logs: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-1301/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Tests exited with: NonZeroExitCodeException
Command 'bash /data/hive-ptest/working/scratch/source-prep.sh' failed with exit 
status 1 and output '+ [[ -n /usr/java/jdk1.7.0_45-cloudera ]]
+ export JAVA_HOME=/usr/java/jdk1.7.0_45-cloudera
+ JAVA_HOME=/usr/java/jdk1.7.0_45-cloudera
+ export 
PATH=/usr/java/jdk1.7.0_45-cloudera/bin/:/usr/local/apache-maven-3.0.5/bin:/usr/java/jdk1.6.0_34/bin:/usr/local/apache-ant-1.9.1/bin:/usr/local/bin:/bin:/usr/bin:/usr/local/sbin:/usr/sbin:/sbin:/home/hiveptest/bin
+ 
PATH=/usr/java/jdk1.7.0_45-cloudera/bin/:/usr/local/apache-maven-3.0.5/bin:/usr/java/jdk1.6.0_34/bin:/usr/local/apache-ant-1.9.1/bin:/usr/local/bin:/bin:/usr/bin:/usr/local/sbin:/usr/sbin:/sbin:/home/hiveptest/bin
+ export 'ANT_OPTS=-Xmx1g -XX:MaxPermSize=256m '
+ ANT_OPTS='-Xmx1g -XX:MaxPermSize=256m '
+ export 'M2_OPTS=-Xmx1g -XX:MaxPermSize=256m -Dhttp.proxyHost=localhost 
-Dhttp.proxyPort=3128'
+ M2_OPTS='-Xmx1g -XX:MaxPermSize=256m -Dhttp.proxyHost=localhost 
-Dhttp.proxyPort=3128'
+ cd /data/hive-ptest/working/
+ tee /data/hive-ptest/logs/PreCommit-HIVE-TRUNK-Build-1301/source-prep.txt
+ [[ false == \t\r\u\e ]]
+ mkdir -p maven ivy
+ [[ svn = \s\v\n ]]
+ [[ -n '' ]]
+ [[ -d apache-svn-trunk-source ]]
+ [[ ! -d apache-svn-trunk-source/.svn ]]
+ [[ ! -d apache-svn-trunk-source ]]
+ cd apache-svn-trunk-source
+ svn revert -R .
++ awk '{print $2}'
++ egrep -v '^X|^Performing status on external'
++ svn status --no-ignore
+ rm -rf target datanucleus.log ant/target shims/target shims/0.20/target 
shims/0.20S/target shims/0.23/target shims/aggregator/target 
shims/common/target shims/common-secure/target packaging/target 
hbase-handler/target testutils/target jdbc/target metastore/target 
itests/target itests/hcatalog-unit/target itests/test-serde/target 
itests/qtest/target itests/hive-unit/target itests/custom-serde/target 
itests/util/target hcatalog/target hcatalog/core/target 
hcatalog/streaming/target hcatalog/server-extensions/target 
hcatalog/webhcat/svr/target hcatalog/webhcat/java-client/target 
hcatalog/hcatalog-pig-adapter/target accumulo-handler/target hwi/target 
common/target common/src/gen contrib/target service/target serde/target 
beeline/target odbc/target cli/target ql/dependency-reduced-pom.xml ql/target
+ svn update
Ujdbc/src/java/org/apache/hive/jdbc/HiveConnection.java
Aitests/hive-unit/src/test/java/org/apache/hive/jdbc/TestNoSaslAuth.java

Fetching external item into 'hcatalog/src/test/e2e/harness'
Updated external to revision 1632432.

Updated to revision 1632432.
+ patchCommandPath=/data/hive-ptest/working/scratch/smart-apply-patch.sh
+ patchFilePath=/data/hive-ptest/working/scratch/build.patch
+ [[ -f /data/hive-ptest/working/scratch/build.patch ]]
+ chmod +x /data/hive-ptest/working/scratch/smart-apply-patch.sh
+ /data/hive-ptest/working/scratch/smart-apply-patch.sh 
/data/hive-ptest/working/scratch/build.patch
The patch does not appear to apply with p0, p1, or p2
+ exit 1
'
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12675240
 - PreCommit-HIVE-TRUNK-Build

 Fix some minor test fails on trunk
 --

 Key: HIVE-8465
 URL: https://issues.apache.org/jira/browse/HIVE-8465
 Project: Hive
  Issue Type: Task
  Components: Tests
Reporter: Navis
Assignee: Jason Dere
Priority: Minor
 Attachments: HIVE-8465.1.patch.txt, HIVE-8465.2.patch, 
 HIVE-8465.3.patch.txt


 org.apache.hive.beeline.TestSchemaTool.testSchemaInit
 org.apache.hive.beeline.TestSchemaTool.testSchemaUpgrade
 org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_parallel
 -- due to HIVE-2828
 org.apache.hadoop.hive.cli.TestHBaseCliDriver.testCliDriver_hbase_custom_key
 org.apache.hadoop.hive.cli.TestHBaseCliDriver.testCliDriver_hbase_custom_key2
 org.apache.hadoop.hive.cli.TestHBaseCliDriver.testCliDriver_hbase_custom_key3



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-6758) Beeline doesn't work with -e option when started in background

2014-10-16 Thread Mohit Sabharwal (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6758?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14174204#comment-14174204
 ] 

Mohit Sabharwal commented on HIVE-6758:
---

I looked into it a bit, and here's my understanding:
In beeline, for pretty table formatting (which hive cli does not do), we get 
the terminal width (and height) using 
jline.Terminal.setupTerminal().getTerminalWidth(), which [sets terminal 
properties|http://jline.sourcearchive.com/documentation/0.9.94-5/classjline_1_1UnixTerminal_da5160e7ec686c49b8f39e03d986cc2f.html#da5160e7ec686c49b8f39e03d986cc2f]
 by calling stty. 

However, when beeline is backgrounded, kernel's terminal driver sends a SIGTTOU 
when stty is called, which if not caught, defaults to stopping the process. I 
tried ignoring the signal in beeline:
{code}
Signal.handle(new Signal(TTOU), SignalHandler.SIG_IGN);
{code}
This puts beeline process state back to running, though it remains hung and 
doesn't resume normal operation. I didn't see anything helpful using jstack.  
Not sure what's happening here, will need some more digging.

However, since we don't care about pretty table formatting in background mode, 
solution proposed by [~qwertymaniac] and [~brocknoland] also works fine. 
Attached patch does that. It essentially sets the terminal type to 
[UnsupportedTerminal|http://jline.sourcearchive.com/documentation/0.9.94-5/classjline_1_1UnsupportedTerminal.html]
 (as opposed to UnixTerminal) which doesn't call stty. 

Alternatively, instead of setting this system property, we could simply not 
initialize a Terminal object if the process is backgrounded. Both approaches 
result in default terminal height and width values. Am going with beeline 
script-only approach.

Would appreciate comments from anyone more familiar with beeline/hive cli. 
Thanks.

 Beeline doesn't work with -e option when started in background
 --

 Key: HIVE-6758
 URL: https://issues.apache.org/jira/browse/HIVE-6758
 Project: Hive
  Issue Type: Improvement
  Components: CLI
Affects Versions: 0.11.0
Reporter: Johndee Burks
Assignee: Mohit Sabharwal

 In hive CLI you could easily integrate its use into a script and back ground 
 the process like this: 
 hive -e some query 
 Beeline does not run when you do the same even with the -f switch. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-6758) Beeline doesn't work with -e option when started in background

2014-10-16 Thread Mohit Sabharwal (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6758?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mohit Sabharwal updated HIVE-6758:
--
Status: Patch Available  (was: Reopened)

 Beeline doesn't work with -e option when started in background
 --

 Key: HIVE-6758
 URL: https://issues.apache.org/jira/browse/HIVE-6758
 Project: Hive
  Issue Type: Improvement
  Components: CLI
Affects Versions: 0.11.0
Reporter: Johndee Burks
Assignee: Mohit Sabharwal
 Attachments: HIVE-6758.patch


 In hive CLI you could easily integrate its use into a script and back ground 
 the process like this: 
 hive -e some query 
 Beeline does not run when you do the same even with the -f switch. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-6758) Beeline doesn't work with -e option when started in background

2014-10-16 Thread Mohit Sabharwal (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6758?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mohit Sabharwal updated HIVE-6758:
--
Attachment: HIVE-6758.patch

 Beeline doesn't work with -e option when started in background
 --

 Key: HIVE-6758
 URL: https://issues.apache.org/jira/browse/HIVE-6758
 Project: Hive
  Issue Type: Improvement
  Components: CLI
Affects Versions: 0.11.0
Reporter: Johndee Burks
Assignee: Mohit Sabharwal
 Attachments: HIVE-6758.patch


 In hive CLI you could easily integrate its use into a script and back ground 
 the process like this: 
 hive -e some query 
 Beeline does not run when you do the same even with the -f switch. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-7858) Parquet compression should be configurable via table property

2014-10-16 Thread Brock Noland (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-7858?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Brock Noland updated HIVE-7858:
---
Release Note: The property parquet.compression can now be configured as a 
table property.  (was: The property parquet.compression can not be configured 
as a table property.)

 Parquet compression should be configurable via table property
 -

 Key: HIVE-7858
 URL: https://issues.apache.org/jira/browse/HIVE-7858
 Project: Hive
  Issue Type: Improvement
Reporter: Brock Noland
Assignee: Ferdinand Xu
 Fix For: 0.15.0

 Attachments: HIVE-7858.1.patch, HIVE-7858.patch, HIVE-7858.patch


 ORC supports the orc.compress table property:
 https://cwiki.apache.org/confluence/display/Hive/LanguageManual+ORC
 {noformat}
 create table Addresses (
   name string,
   street string,
   city string,
   state string,
   zip int
 ) stored as orc tblproperties (orc.compress=NONE);
 {noformat}
 I think it'd be great to support the same for Parquet.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-7858) Parquet compression should be configurable via table property

2014-10-16 Thread Brock Noland (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-7858?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Brock Noland updated HIVE-7858:
---
Labels: TODOC15  (was: )

We should doc it.. I put a release note The property parquet.compression can 
now be configured as a table property.

I think we should put a item here: 
https://cwiki.apache.org/confluence/display/Hive/Parquet

* Parquet Table Properties *

with this one. We can update it as we add more..

 Parquet compression should be configurable via table property
 -

 Key: HIVE-7858
 URL: https://issues.apache.org/jira/browse/HIVE-7858
 Project: Hive
  Issue Type: Improvement
Reporter: Brock Noland
Assignee: Ferdinand Xu
  Labels: TODOC15
 Fix For: 0.15.0

 Attachments: HIVE-7858.1.patch, HIVE-7858.patch, HIVE-7858.patch


 ORC supports the orc.compress table property:
 https://cwiki.apache.org/confluence/display/Hive/LanguageManual+ORC
 {noformat}
 create table Addresses (
   name string,
   street string,
   city string,
   state string,
   zip int
 ) stored as orc tblproperties (orc.compress=NONE);
 {noformat}
 I think it'd be great to support the same for Parquet.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8486) TPC-DS Query 96 parallelism is not set correcly

2014-10-16 Thread Chao (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8486?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14174223#comment-14174223
 ] 

Chao commented on HIVE-8486:


I just talked with [~szehon] offline, and he pointed me to 
{{ReduceSinkOperator::computeHashCode}}. This function returns:

{noformat}
return bucketNumber  0  ? keyHashCode : keyHashCode * 31 + bucketNumber;
{noformat}

So, if the {{bucketNumber}} is 0, and we set the number of partitions to 31, 
then all keys will go to the same partition. I think this explains why the 
issue only happens only when we set {{mapreduce.job.reduces}} to 31. I also 
verified it locally.

 TPC-DS Query 96 parallelism is not set correcly
 ---

 Key: HIVE-8486
 URL: https://issues.apache.org/jira/browse/HIVE-8486
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Reporter: Brock Noland
Assignee: Chao

 When we run the query on a 20B we only have a parallelism factor of 1.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


Re: Review Request 26481: HIVE-8411: Support partial partition spec for certain ALTER PARTITION statements

2014-10-16 Thread Jason Dere


 On Oct. 16, 2014, 7:32 a.m., Gunther Hagleitner wrote:
  ql/src/java/org/apache/hadoop/hive/ql/exec/DDLTask.java, line 3265
  https://reviews.apache.org/r/26481/diff/1/?file=716405#file716405line3265
 
  there's a flag for enabling dynamic partitioning. is this checked 
  somewhere in this path? ppl use that to avoid someone accidentally changing 
  everything.

Good point, will add that check.


 On Oct. 16, 2014, 7:32 a.m., Gunther Hagleitner wrote:
  ql/src/java/org/apache/hadoop/hive/ql/exec/DDLTask.java, line 3269
  https://reviews.apache.org/r/26481/diff/1/?file=716405#file716405line3269
 
  why does this fail? it seems more consistent to just alter 0 partitions 
  in this case. or at least just print a warning instead of failing.

This is just the existing behavior right? This is if the user specifies a 
non-existent partition, including if they specify all partition values in the 
spec, like (part1='nonexistent', part2='partition'). I think this is sensible 
behavior, but reply if you think otherwise. We could just fail in the event 
that the partition spec specifies values for all of the partition columns, if 
you think that is better.


 On Oct. 16, 2014, 7:32 a.m., Gunther Hagleitner wrote:
  ql/src/java/org/apache/hadoop/hive/ql/exec/DDLTask.java, line 3301
  https://reviews.apache.org/r/26481/diff/1/?file=716405#file716405line3301
 
  LOG.error?

Whoops, will fix.


 On Oct. 16, 2014, 7:32 a.m., Gunther Hagleitner wrote:
  ql/src/test/queries/clientpositive/alter_partition_change_col.q, line 60
  https://reviews.apache.org/r/26481/diff/1/?file=716407#file716407line60
 
  does this support multi level partitioning? if so can you specify only 
  some of the partitions?

You mean like multiple partition columns? I'll change the test to include that.


 On Oct. 16, 2014, 7:32 a.m., Gunther Hagleitner wrote:
  ql/src/test/queries/clientpositive/alter_partition_change_col.q, line 63
  https://reviews.apache.org/r/26481/diff/1/?file=716407#file716407line63
 
  add a query to verify the output?

Will fix.


- Jason


---
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/26481/#review56905
---


On Oct. 9, 2014, 3:14 a.m., Jason Dere wrote:
 
 ---
 This is an automatically generated e-mail. To reply, visit:
 https://reviews.apache.org/r/26481/
 ---
 
 (Updated Oct. 9, 2014, 3:14 a.m.)
 
 
 Review request for hive, Gunther Hagleitner and Thejas Nair.
 
 
 Bugs: HIVE-8411
 https://issues.apache.org/jira/browse/HIVE-8411
 
 
 Repository: hive-git
 
 
 Description
 ---
 
 Allow partial partition spec for some alter table statements, such as:
 
 alter table alter_partition_change_col1 partition (p1) change column c2 c2 
 decimal(10,0);
 
 
 Diffs
 -
 
   ql/src/java/org/apache/hadoop/hive/ql/exec/DDLTask.java d5374bc 
   ql/src/java/org/apache/hadoop/hive/ql/parse/DDLSemanticAnalyzer.java 
 4e58ad8 
   ql/src/test/queries/clientpositive/alter_partition_change_col.q baabb9f 
   ql/src/test/results/clientpositive/alter_partition_change_col.q.out 7123e40 
 
 Diff: https://reviews.apache.org/r/26481/diff/
 
 
 Testing
 ---
 
 updated qfile test
 
 
 Thanks,
 
 Jason Dere
 




[jira] [Commented] (HIVE-8486) TPC-DS Query 96 parallelism is not set correcly

2014-10-16 Thread Brock Noland (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8486?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14174277#comment-14174277
 ] 

Brock Noland commented on HIVE-8486:


Thanks Chao! Two questions:

1) Can we fix this the calculation described above?
2) Why are we estimating a single reducer in the normal case? The join covers 
the entire store_sales table which is 2.2GB. Perhap we need to add some debug 
statements to SetSparkReducerParallelism?

 TPC-DS Query 96 parallelism is not set correcly
 ---

 Key: HIVE-8486
 URL: https://issues.apache.org/jira/browse/HIVE-8486
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Reporter: Brock Noland
Assignee: Chao

 When we run the query on a 20B we only have a parallelism factor of 1.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8460) ORC SARG literal creation for double from float may lead to wrong evaluation of SARG

2014-10-16 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8460?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14174279#comment-14174279
 ] 

Hive QA commented on HIVE-8460:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12675323/HIVE-8460.2.patch

{color:red}ERROR:{color} -1 due to 5 failed/errored test(s), 6561 tests executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestHBaseCliDriver.testCliDriver_hbase_custom_key
org.apache.hadoop.hive.cli.TestHBaseCliDriver.testCliDriver_hbase_custom_key2
org.apache.hadoop.hive.cli.TestHBaseCliDriver.testCliDriver_hbase_custom_key3
org.apache.hive.beeline.TestSchemaTool.testSchemaInit
org.apache.hive.beeline.TestSchemaTool.testSchemaUpgrade
{noformat}

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/1302/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/1302/console
Test logs: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-1302/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 5 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12675323
 - PreCommit-HIVE-TRUNK-Build

 ORC SARG literal creation for double from float may lead to wrong evaluation 
 of SARG
 

 Key: HIVE-8460
 URL: https://issues.apache.org/jira/browse/HIVE-8460
 Project: Hive
  Issue Type: Bug
  Components: Query Processor
Affects Versions: 0.13.0, 0.13.1
Reporter: Ashutosh Chauhan
Assignee: Prasanth J
 Attachments: HIVE-8460.1.patch, HIVE-8460.2.patch


 e.g., expression like d = 0.22 where 0.22 is of float type, gets converted to 
 double d = 0.21986 in SARG creation. This will cause erroneous evaluation 
 of SARG expression.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8486) TPC-DS Query 96 parallelism is not set correcly

2014-10-16 Thread Chao (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8486?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14174292#comment-14174292
 ] 

Chao commented on HIVE-8486:


1) I discussed with [~szehon] about this, and seems we cannot change the 
calculation now, since it will affect the way how the buckets are calculated. 
2) In the Spark branch, we are not doing anything to estimate the number of 
reducers. In MR, in case this number is not set, it will estimate and set it at 
runtime. This is done in {{MapRedTask}}. Tez also uses Auto Reducer 
Parallelism (see HIVE-7158) to control this. As result, I think we should do 
something to resolve this. 

 TPC-DS Query 96 parallelism is not set correcly
 ---

 Key: HIVE-8486
 URL: https://issues.apache.org/jira/browse/HIVE-8486
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Reporter: Brock Noland
Assignee: Chao

 When we run the query on a 20B we only have a parallelism factor of 1.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8486) TPC-DS Query 96 parallelism is not set correcly

2014-10-16 Thread Brock Noland (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8486?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14174308#comment-14174308
 ] 

Brock Noland commented on HIVE-8486:


bq. In the Spark branch, we are not doing anything to estimate the number of 
reducers.

What is {{SetSparkReducerParallelism}} doing then?

 TPC-DS Query 96 parallelism is not set correcly
 ---

 Key: HIVE-8486
 URL: https://issues.apache.org/jira/browse/HIVE-8486
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Reporter: Brock Noland
Assignee: Chao

 When we run the query on a 20B we only have a parallelism factor of 1.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-8462) CBO duplicates columns

2014-10-16 Thread Sergey Shelukhin (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-8462?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sergey Shelukhin updated HIVE-8462:
---
   Resolution: Fixed
Fix Version/s: 0.15.0
   Status: Resolved  (was: Patch Available)

committed to trunk. [~vikram.dixit] ok for 14?

 CBO duplicates columns
 --

 Key: HIVE-8462
 URL: https://issues.apache.org/jira/browse/HIVE-8462
 Project: Hive
  Issue Type: Bug
  Components: CBO
Reporter: Sergey Shelukhin
Assignee: Sergey Shelukhin
Priority: Critical
 Fix For: 0.15.0

 Attachments: HIVE-8462.01.patch, HIVE-8462.patch


 {noformat}
 select *, rank() over(partition by key order by value) as rr from src1
 {noformat}
 Original plan appears to be incorrect:
 {noformat}
 HiveProjectRel(key=[$0], value=[$1], (tok_function rank (tok_windowspec 
 (tok_partitioningspec (tok_distributeby (tok_table_or_col key)) (tok_orderby 
 (tok_tabsortcolnameasc (tok_table_or_col value))=[$5], rr=[$5])
   HiveProjectRel(key=[$0], value=[$1], block__offset__inside__file=[$2], 
 input__file__name=[$3], row__id=[$4], (tok_function rank (tok_windowspec 
 (tok_partitioningspec (tok_distributeby (tok_table_or_col key)) (tok_orderby 
 (tok_tabsortcolnameasc (tok_table_or_col value))=[rank() OVER (PARTITION 
 BY $0 ORDER BY $1 ROWS BETWEEN 2147483647 FOLLOWING AND 2147483647 
 PRECEDING)])
 HiveTableScanRel(table=[[default.src1]])
 {noformat}
 and final AST has {noformat}
  TOK_SELEXPR
 .
TOK_TABLE_OR_COL
   $hdt$_0
(tok_function rank (tok_windowspec (tok_partitioningspec 
 (tok_distributeby (tok_table_or_col key)) (tok_orderby (tok_tabsortcolnameasc 
 (tok_table_or_col value))
 (tok_function rank (tok_windowspec (tok_partitioningspec 
 (tok_distributeby (tok_table_or_col key)) (tok_orderby (tok_tabsortcolnameasc 
 (tok_table_or_col value))
  TOK_SELEXPR
 .
TOK_TABLE_OR_COL
   $hdt$_0
(tok_function rank (tok_windowspec (tok_partitioningspec 
 (tok_distributeby (tok_table_or_col key)) (tok_orderby (tok_tabsortcolnameasc 
 (tok_table_or_col value))
 rr
 {noformat}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8486) TPC-DS Query 96 parallelism is not set correcly

2014-10-16 Thread Chao (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8486?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14174313#comment-14174313
 ] 

Chao commented on HIVE-8486:


Oops. I didn't notice this class. Let me take a look and come back.

 TPC-DS Query 96 parallelism is not set correcly
 ---

 Key: HIVE-8486
 URL: https://issues.apache.org/jira/browse/HIVE-8486
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Reporter: Brock Noland
Assignee: Chao

 When we run the query on a 20B we only have a parallelism factor of 1.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8486) TPC-DS Query 96 parallelism is not set correcly

2014-10-16 Thread Chao (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8486?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14174341#comment-14174341
 ] 

Chao commented on HIVE-8486:


OK, I debugged this query. In {{SetSparkReducerParallelism}}, in order to 
estimate, it needs to obtain statistics from the siblings of the current reduce 
sink, and adds up the total number of bytes. However, somehow the 
{{statistics}} field of all the siblings are null, and hence the number of 
bytes is 0 at end. As result, it will only use one reducer.

I'm wondering if this is something we haven't implemented yet, or is it a bug?

 TPC-DS Query 96 parallelism is not set correcly
 ---

 Key: HIVE-8486
 URL: https://issues.apache.org/jira/browse/HIVE-8486
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Reporter: Brock Noland
Assignee: Chao

 When we run the query on a 20B we only have a parallelism factor of 1.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HIVE-8489) Add sanity check to dynamic partition pruning

2014-10-16 Thread Gunther Hagleitner (JIRA)
Gunther Hagleitner created HIVE-8489:


 Summary: Add sanity check to dynamic partition pruning
 Key: HIVE-8489
 URL: https://issues.apache.org/jira/browse/HIVE-8489
 Project: Hive
  Issue Type: Bug
Reporter: Gunther Hagleitner
Assignee: Gunther Hagleitner
 Fix For: 0.14.0


For debugging/safety it would help to make sure we received the right number of 
events from tez tasks. This is always supposed to be the case, but would be 
really hard to debug if something goes wrong.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HIVE-8490) Constant folding should happen before partition pruning

2014-10-16 Thread Ashutosh Chauhan (JIRA)
Ashutosh Chauhan created HIVE-8490:
--

 Summary: Constant folding should happen before partition pruning
 Key: HIVE-8490
 URL: https://issues.apache.org/jira/browse/HIVE-8490
 Project: Hive
  Issue Type: Improvement
  Components: Logical Optimizer
Affects Versions: 0.14.0
Reporter: Ashutosh Chauhan
Assignee: Ashutosh Chauhan


so that partition pruning see simplified expressions and present simpler 
expressions to metastore



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-8490) Constant folding should happen before partition pruning

2014-10-16 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-8490?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-8490:
---
Attachment: HIVE-8490.patch

 Constant folding should happen before partition pruning
 ---

 Key: HIVE-8490
 URL: https://issues.apache.org/jira/browse/HIVE-8490
 Project: Hive
  Issue Type: Improvement
  Components: Logical Optimizer
Affects Versions: 0.14.0
Reporter: Ashutosh Chauhan
Assignee: Ashutosh Chauhan
 Attachments: HIVE-8490.patch


 so that partition pruning see simplified expressions and present simpler 
 expressions to metastore



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-8490) Constant folding should happen before partition pruning

2014-10-16 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-8490?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-8490:
---
Status: Patch Available  (was: Open)

 Constant folding should happen before partition pruning
 ---

 Key: HIVE-8490
 URL: https://issues.apache.org/jira/browse/HIVE-8490
 Project: Hive
  Issue Type: Improvement
  Components: Logical Optimizer
Affects Versions: 0.14.0
Reporter: Ashutosh Chauhan
Assignee: Ashutosh Chauhan
 Attachments: HIVE-8490.patch


 so that partition pruning see simplified expressions and present simpler 
 expressions to metastore



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8490) Constant folding should happen before partition pruning

2014-10-16 Thread Ashutosh Chauhan (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8490?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14174348#comment-14174348
 ] 

Ashutosh Chauhan commented on HIVE-8490:


Needs HIVE-6934

 Constant folding should happen before partition pruning
 ---

 Key: HIVE-8490
 URL: https://issues.apache.org/jira/browse/HIVE-8490
 Project: Hive
  Issue Type: Improvement
  Components: Logical Optimizer
Affects Versions: 0.14.0
Reporter: Ashutosh Chauhan
Assignee: Ashutosh Chauhan
 Attachments: HIVE-8490.patch


 so that partition pruning see simplified expressions and present simpler 
 expressions to metastore



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


  1   2   >