[jira] [Commented] (HIVE-10495) Hive index creation code throws NPE if index table is null
[ https://issues.apache.org/jira/browse/HIVE-10495?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14742156#comment-14742156 ] Ashutosh Chauhan commented on HIVE-10495: - I wonder why indexTable is null in MS. If that function was called, caller should have passed valid indexTable. I think bug lies with caller. Also, do you have a repro test case for this? > Hive index creation code throws NPE if index table is null > -- > > Key: HIVE-10495 > URL: https://issues.apache.org/jira/browse/HIVE-10495 > Project: Hive > Issue Type: Bug >Affects Versions: 1.0.0, 1.2.0, 1.2.1 >Reporter: Bing Li >Assignee: Bing Li > Attachments: HIVE-10495.1.patch, HIVE-10495.2.patch > > > The stack trace would be: > Caused by: java.lang.NullPointerException > at > org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.add_index(HiveMetaStore.java:2870) > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:60) > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:37) > at java.lang.reflect.Method.invoke(Method.java:611) > at > org.apache.hadoop.hive.metastore.RetryingHMSHandler.invoke(RetryingHMSHandler.java:102) > at $Proxy9.add_index(Unknown Source) > at > org.apache.hadoop.hive.metastore.HiveMetaStoreClient.createIndex(HiveMetaStoreClient.java:962) -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HIVE-11755) Incorrect method called with Kerberos enabled in AccumuloStorageHandler
[ https://issues.apache.org/jira/browse/HIVE-11755?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14742236#comment-14742236 ] Josh Elser commented on HIVE-11755: --- Anyone have some cycles that they can help me out with a review and (hopefully) commit, please? > Incorrect method called with Kerberos enabled in AccumuloStorageHandler > --- > > Key: HIVE-11755 > URL: https://issues.apache.org/jira/browse/HIVE-11755 > Project: Hive > Issue Type: Bug >Affects Versions: 1.2.1 >Reporter: Josh Elser >Assignee: Josh Elser > Fix For: 1.2.2 > > Attachments: HIVE-11755.001.patch, HIVE-11755.002.patch, > HIVE-11755.003.patch > > > The following exception was noticed in testing out the > AccumuloStorageHandler's OutputFormat: > {noformat} > java.lang.IllegalStateException: Connector info for AccumuloOutputFormat can > only be set once per job > at > org.apache.accumulo.core.client.mapreduce.lib.impl.ConfiguratorBase.setConnectorInfo(ConfiguratorBase.java:146) > at > org.apache.accumulo.core.client.mapred.AccumuloOutputFormat.setConnectorInfo(AccumuloOutputFormat.java:125) > at > org.apache.hadoop.hive.accumulo.mr.HiveAccumuloTableOutputFormat.configureAccumuloOutputFormat(HiveAccumuloTableOutputFormat.java:95) > at > org.apache.hadoop.hive.accumulo.mr.HiveAccumuloTableOutputFormat.checkOutputSpecs(HiveAccumuloTableOutputFormat.java:51) > at > org.apache.hadoop.hive.ql.io.HivePassThroughOutputFormat.checkOutputSpecs(HivePassThroughOutputFormat.java:46) > at > org.apache.hadoop.hive.ql.exec.FileSinkOperator.checkOutputSpecs(FileSinkOperator.java:1124) > at > org.apache.hadoop.hive.ql.io.HiveOutputFormatImpl.checkOutputSpecs(HiveOutputFormatImpl.java:67) > at > org.apache.hadoop.mapreduce.JobSubmitter.checkSpecs(JobSubmitter.java:268) > at > org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(JobSubmitter.java:139) > at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1290) > at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1287) > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Subject.java:415) > at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1657) > at org.apache.hadoop.mapreduce.Job.submit(Job.java:1287) > at org.apache.hadoop.mapred.JobClient$1.run(JobClient.java:575) > at org.apache.hadoop.mapred.JobClient$1.run(JobClient.java:570) > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Subject.java:415) > at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1657) > at org.apache.hadoop.mapred.JobClient.submitJobInternal(JobClient.java:570) > at org.apache.hadoop.mapred.JobClient.submitJob(JobClient.java:561) > at org.apache.hadoop.hive.ql.exec.mr.ExecDriver.execute(ExecDriver.java:431) > at org.apache.hadoop.hive.ql.exec.mr.MapRedTask.execute(MapRedTask.java:137) > at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:160) > at > org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:88) > at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:1653) > at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:1412) > at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1195) > at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1059) > at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1049) > at org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:213) > at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:165) > at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:376) > at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:311) > at org.apache.hadoop.hive.cli.CliDriver.processReader(CliDriver.java:409) > at org.apache.hadoop.hive.cli.CliDriver.processFile(CliDriver.java:425) > at org.apache.hadoop.hive.cli.CliDriver.executeDriver(CliDriver.java:714) > at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:681) > at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:621) > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > at java.lang.reflect.Method.invoke(Method.java:606) > at org.apache.hadoop.util.RunJar.run(RunJar.java:221) > at org.apache.hadoop.util.RunJar.main(RunJar.java:136) > Job Submission failed with exception > 'java.lang.IllegalStateException(Connector info for AccumuloOutputFormat can > only be set once per job)' > {noformat} > The OutputFormat implementation already had a method in place to
[jira] [Commented] (HIVE-11745) Alter table Exchange partition with multiple partition_spec is not working
[ https://issues.apache.org/jira/browse/HIVE-11745?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14742257#comment-14742257 ] Hive QA commented on HIVE-11745: {color:red}Overall{color}: -1 at least one tests failed Here are the results of testing the latest attachment: https://issues.apache.org/jira/secure/attachment/12755532/HIVE-11745.2.patch {color:red}ERROR:{color} -1 due to 8 failed/errored test(s), 9425 tests executed *Failed tests:* {noformat} org.apache.hadoop.hive.thrift.TestHadoop20SAuthBridge.testSaslWithHiveMetaStore org.apache.hive.hcatalog.api.TestHCatClient.testTableSchemaPropagation org.apache.hive.hcatalog.streaming.TestStreaming.testAddPartition org.apache.hive.hcatalog.streaming.TestStreaming.testTransactionBatchAbort org.apache.hive.hcatalog.streaming.TestStreaming.testTransactionBatchAbortAndCommit org.apache.hive.spark.client.TestSparkClient.testJobSubmission org.apache.hive.spark.client.TestSparkClient.testMetricsCollection org.apache.hive.spark.client.TestSparkClient.testSyncRpc {noformat} Test results: http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/5260/testReport Console output: http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/5260/console Test logs: http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-5260/ Messages: {noformat} Executing org.apache.hive.ptest.execution.PrepPhase Executing org.apache.hive.ptest.execution.ExecutionPhase Executing org.apache.hive.ptest.execution.ReportingPhase Tests exited with: TestsFailedException: 8 tests failed {noformat} This message is automatically generated. ATTACHMENT ID: 12755532 - PreCommit-HIVE-TRUNK-Build > Alter table Exchange partition with multiple partition_spec is not working > -- > > Key: HIVE-11745 > URL: https://issues.apache.org/jira/browse/HIVE-11745 > Project: Hive > Issue Type: Bug > Components: Metastore >Affects Versions: 1.2.0, 1.1.0, 2.0.0 >Reporter: Yongzhi Chen >Assignee: Yongzhi Chen > Attachments: HIVE-11745.1.patch, HIVE-11745.2.patch > > > Single partition works, but multiple partitions will not work. > Reproduce steps: > {noformat} > DROP TABLE IF EXISTS t1; > DROP TABLE IF EXISTS t2; > DROP TABLE IF EXISTS t3; > DROP TABLE IF EXISTS t4; > CREATE TABLE t1 (a int) PARTITIONED BY (d1 int); > CREATE TABLE t2 (a int) PARTITIONED BY (d1 int); > CREATE TABLE t3 (a int) PARTITIONED BY (d1 int, d2 int); > CREATE TABLE t4 (a int) PARTITIONED BY (d1 int, d2 int); > INSERT OVERWRITE TABLE t1 PARTITION (d1 = 1) SELECT salary FROM jsmall LIMIT > 10; > INSERT OVERWRITE TABLE t3 PARTITION (d1 = 1, d2 = 1) SELECT salary FROM > jsmall LIMIT 10; > SELECT * FROM t1; > SELECT * FROM t3; > ALTER TABLE t2 EXCHANGE PARTITION (d1 = 1) WITH TABLE t1; > SELECT * FROM t1; > SELECT * FROM t2; > ALTER TABLE t4 EXCHANGE PARTITION (d1 = 1, d2 = 1) WITH TABLE t3; > SELECT * FROM t3; > SELECT * FROM t4; > {noformat} > The output: > {noformat} > 0: jdbc:hive2://10.17.74.148:1/default> SELECT * FROM t3; > +---+++--+ > | t3.a | t3.d1 | t3.d2 | > +---+++--+ > +---+++--+ > No rows selected (0.227 seconds) > 0: jdbc:hive2://10.17.74.148:1/default> SELECT * FROM t4; > +---+++--+ > | t4.a | t4.d1 | t4.d2 | > +---+++--+ > +---+++--+ > No rows selected (0.266 seconds) > {noformat} -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HIVE-11794) GBY vectorization appears to process COMPLETE reduce-side GBY incorrectly
[ https://issues.apache.org/jira/browse/HIVE-11794?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14742226#comment-14742226 ] Hive QA commented on HIVE-11794: {color:red}Overall{color}: -1 at least one tests failed Here are the results of testing the latest attachment: https://issues.apache.org/jira/secure/attachment/12755514/HIVE-11794.patch {color:red}ERROR:{color} -1 due to 4 failed/errored test(s), 9416 tests executed *Failed tests:* {noformat} TestSparkClient - did not produce a TEST-*.xml file org.apache.hadoop.hive.cli.TestMiniTezCliDriver.testCliDriver_vectorization_limit org.apache.hadoop.hive.ql.optimizer.physical.TestVectorizer.testValidateNestedExpressions org.apache.hive.hcatalog.api.TestHCatClient.testTableSchemaPropagation {noformat} Test results: http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/5259/testReport Console output: http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/5259/console Test logs: http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-5259/ Messages: {noformat} Executing org.apache.hive.ptest.execution.PrepPhase Executing org.apache.hive.ptest.execution.ExecutionPhase Executing org.apache.hive.ptest.execution.ReportingPhase Tests exited with: TestsFailedException: 4 tests failed {noformat} This message is automatically generated. ATTACHMENT ID: 12755514 - PreCommit-HIVE-TRUNK-Build > GBY vectorization appears to process COMPLETE reduce-side GBY incorrectly > - > > Key: HIVE-11794 > URL: https://issues.apache.org/jira/browse/HIVE-11794 > Project: Hive > Issue Type: Bug >Reporter: Sergey Shelukhin >Assignee: Sergey Shelukhin > Attachments: HIVE-11794.patch > > > The code in Vectorizer is as such: > {noformat} > boolean isMergePartial = (desc.getMode() != GroupByDesc.Mode.HASH); > {noformat} > then, if it's reduce side: > {noformat} > if (isMergePartial) { > // Reduce Merge-Partial GROUP BY. > // A merge-partial GROUP BY is fed by grouping by keys from > reduce-shuffle. It is the > // first (or root) operator for its reduce task. > > } else { > // Reduce Hash GROUP BY or global aggregation. > ... > {noformat} > In fact, this logic is missing the COMPLETE mode. Both from the comment: > {noformat} > COMPLETE: complete 1-phase aggregation: iterate, terminate > ... > HASH: For non-distinct the same as PARTIAL1 but use hash-table-based > aggregation > ... > PARTIAL1: partial aggregation - first phase: iterate, terminatePartial > {noformat} > and from the explain plan like this (the query has multiple stages of > aggregations over a union; the mapper does a partial hash aggregation for > each side of the union, which is then followed by mergepartial, and 2nd stage > as complete): > {noformat} > Map Operator Tree: > ... > Group By Operator > keys: _col0 (type: int), _col1 (type: int), _col2 (type: int), > _col3 (type: int), _col4 (type: int), _col5 (type: bigint), _col6 (type: > bigint), _col7 (type: bigint), _col8 (type: bigint), _col9 (type: bigint), > _col10 (type: bigint), _col11 (type: bigint), _col12 (type: bigint) > mode: hash > outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, > _col7, _col8, _col9, _col10, _col11, _col12 > Reduce Output Operator > ... > feeding into > Reduce Operator Tree: > Group By Operator > keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: > int), KEY._col3 (type: int), KEY._col4 (type: int), KEY._col5 (type: bigint), > KEY._col6 (type: bigint), KEY._col7 (type: bigint), KEY._col8 (type: bigint), > KEY._col9 (type: bigint), KEY._col10 (type: bigint), KEY._col11 (type: > bigint), KEY._col12 (type: bigint) > mode: mergepartial > outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, > _col7, _col8, _col9, _col10, _col11, _col12 > Group By Operator > aggregations: sum(_col5), sum(_col6), sum(_col7), sum(_col8), > sum(_col9), sum(_col10), sum(_col11), sum(_col12) > keys: _col0 (type: int), _col1 (type: int), _col2 (type: int), _col3 > (type: int), _col4 (type: int) > mode: complete > outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, > _col7, _col8, _col9, _col10, _col11, _col12 > {noformat} > it seems like COMPLETE is actually the global aggregation, and HASH isn't (or > may not be). > So, it seems like reduce-side COMPLETE should be handled on the else-path of > the above if. For map-side, it doesn't check mode at all as far as I can see. > Not sure if additional code changes are necessary after that, it may just > work. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HIVE-11799) The output of explain query for multiple lateral views is huge
[ https://issues.apache.org/jira/browse/HIVE-11799?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14742191#comment-14742191 ] Hive QA commented on HIVE-11799: {color:red}Overall{color}: -1 at least one tests failed Here are the results of testing the latest attachment: https://issues.apache.org/jira/secure/attachment/12755473/HIVE-11799.patch {color:red}ERROR:{color} -1 due to 258 failed/errored test(s), 9422 tests executed *Failed tests:* {noformat} org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_annotate_stats_union org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_auto_join27 org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_auto_smb_mapjoin_14 org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_auto_sortmerge_join_10 org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_binary_output_format org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_bucketmapjoin1 org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_bucketmapjoin2 org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_bucketmapjoin3 org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_bucketmapjoin4 org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_bucketmapjoin5 org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_bucketmapjoin_negative org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_bucketmapjoin_negative2 org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_case_sensitivity org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_cast1 org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_cluster org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_column_access_stats org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_complex_alias org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_constprog_dp org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_constprog_type org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_correlationoptimizer10 org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_correlationoptimizer14 org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_correlationoptimizer15 org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_correlationoptimizer2 org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_correlationoptimizer3 org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_correlationoptimizer5 org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_correlationoptimizer6 org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_correlationoptimizer7 org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_correlationoptimizer8 org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_correlationoptimizer9 org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_cp_sel org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_ctas_colname org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_ctas_uses_database_location org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_dynamic_rdd_cache org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_groupby_ppd org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_groupby_sort_1_23 org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_groupby_sort_3 org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_groupby_sort_5 org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_groupby_sort_7 org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_groupby_sort_skew_1_23 org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_index_auto org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_index_auto_file_format org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_index_auto_mult_tables_compact org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_index_auto_multiple org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_index_auto_partitioned org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_index_auto_update org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_index_compression org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_index_serde org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_index_skewtable org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_input11 org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_input12 org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_input13 org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_input25 org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_input26 org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_input34 org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_input35 org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_input36 org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_input38 org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_input6 org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_input7 org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_input8 org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_input9 org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_input_dynamicserde org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_input_part1
[jira] [Commented] (HIVE-11802) Float-point numbers are displayed with different precision in Beeline/JDBC
[ https://issues.apache.org/jira/browse/HIVE-11802?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14742126#comment-14742126 ] Hive QA commented on HIVE-11802: {color:red}Overall{color}: -1 at least one tests failed Here are the results of testing the latest attachment: https://issues.apache.org/jira/secure/attachment/12755466/HIVE-11802.1.patch {color:red}ERROR:{color} -1 due to 1 failed/errored test(s), 9424 tests executed *Failed tests:* {noformat} org.apache.hive.hcatalog.api.TestHCatClient.testTableSchemaPropagation {noformat} Test results: http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/5256/testReport Console output: http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/5256/console Test logs: http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-5256/ Messages: {noformat} Executing org.apache.hive.ptest.execution.PrepPhase Executing org.apache.hive.ptest.execution.ExecutionPhase Executing org.apache.hive.ptest.execution.ReportingPhase Tests exited with: TestsFailedException: 1 tests failed {noformat} This message is automatically generated. ATTACHMENT ID: 12755466 - PreCommit-HIVE-TRUNK-Build > Float-point numbers are displayed with different precision in Beeline/JDBC > -- > > Key: HIVE-11802 > URL: https://issues.apache.org/jira/browse/HIVE-11802 > Project: Hive > Issue Type: Bug >Affects Versions: 1.2.1 >Reporter: Sergio Peña >Assignee: Sergio Peña > Attachments: HIVE-11802.1.patch > > > When inserting float-point numbers to a table, the values displayed on > beeline or jdbc are with different precision. > How to reproduce: > {noformat} > 0: jdbc:hive2://localhost:1> create table decimals (f float, af > array, d double, ad array) stored as parquet; > No rows affected (0.294 seconds) > 0: jdbc:hive2://localhost:1> insert into table decimals select 1.10058, > array(cast(1.10058 as float)), 2.0133, array(2.0133) from dummy limit 1; > ... > No rows affected (20.089 seconds) > 0: jdbc:hive2://localhost:1> select f, af, af[0], d, ad[0] from decimals; > +-++-+-+-+--+ > | f | af | _c2 |d| _c4 | > +-++-+-+-+--+ > | 1.1005799770355225 | [1.10058] | 1.1005799770355225 | 2.0133 | 2.0133 | > +-++-+-+-+--+ > {noformat} > When displaying arrays, the values are displayed correctly, but if I print a > specific element, it is then displayed with more decimal positions. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HIVE-10980) Merge of dynamic partitions loads all data to default partition
[ https://issues.apache.org/jira/browse/HIVE-10980?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14742145#comment-14742145 ] Ashutosh Chauhan commented on HIVE-10980: - +1 > Merge of dynamic partitions loads all data to default partition > --- > > Key: HIVE-10980 > URL: https://issues.apache.org/jira/browse/HIVE-10980 > Project: Hive > Issue Type: Bug > Components: Hive >Affects Versions: 0.14.0 > Environment: HDP 2.2.4 (also reproduced on apache hive built from > trunk) >Reporter: Illya Yalovyy >Assignee: Illya Yalovyy > Attachments: HIVE-10980.patch > > > Conditions that lead to the issue: > 1. Execution engine set to MapReduce > 2. Partition columns have different types > 3. Both static and dynamic partitions are used in the query > 4. Dynamically generated partitions require merge > Result: Final data is loaded to "__HIVE_DEFAULT_PARTITION__". > Steps to reproduce: > set hive.exec.dynamic.partition=true; > set hive.exec.dynamic.partition.mode=strict; > set hive.optimize.sort.dynamic.partition=false; > set hive.merge.mapfiles=true; > set hive.merge.mapredfiles=true; > set hive.input.format=org.apache.hadoop.hive.ql.io.HiveInputFormat; > set hive.execution.engine=mr; > create external table sdp ( > dataint bigint, > hour int, > req string, > cid string, > caid string > ) > row format delimited > fields terminated by ','; > load data local inpath '../../data/files/dynpartdata1.txt' into table sdp; > load data local inpath '../../data/files/dynpartdata2.txt' into table sdp; > ... > load data local inpath '../../data/files/dynpartdataN.txt' into table sdp; > create table tdp (cid string, caid string) > partitioned by (dataint bigint, hour int, req string); > insert overwrite table tdp partition (dataint=20150316, hour=16, req) > select cid, caid, req from sdp where dataint=20150316 and hour=16; > select * from tdp order by caid; > show partitions tdp; > Example of the input file: > 20150316,16,reqA,clusterIdA,cacheId1 > 20150316,16,reqB,clusterIdB,cacheId2 > 20150316,16,reqA,clusterIdC,cacheId3 > 20150316,16,reqD,clusterIdD,cacheId4 > 20150316,16,reqA,clusterIdA,cacheId5 > Actual result: > clusterIdA cacheId12015031616 > __HIVE_DEFAULT_PARTITION__ > clusterIdA cacheId12015031616 > __HIVE_DEFAULT_PARTITION__ > clusterIdB cacheId22015031616 > __HIVE_DEFAULT_PARTITION__ > clusterIdC cacheId32015031616 > __HIVE_DEFAULT_PARTITION__ > clusterIdD cacheId42015031616 > __HIVE_DEFAULT_PARTITION__ > clusterIdA cacheId52015031616 > __HIVE_DEFAULT_PARTITION__ > clusterIdD cacheId82015031616 > __HIVE_DEFAULT_PARTITION__ > clusterIdB cacheId92015031616 > __HIVE_DEFAULT_PARTITION__ > > dataint=20150316/hour=16/req=__HIVE_DEFAULT_PARTITION__ -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Updated] (HIVE-10328) Enable new return path for cbo
[ https://issues.apache.org/jira/browse/HIVE-10328?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Pengcheng Xiong updated HIVE-10328: --- Attachment: HIVE-10328.10.patch > Enable new return path for cbo > -- > > Key: HIVE-10328 > URL: https://issues.apache.org/jira/browse/HIVE-10328 > Project: Hive > Issue Type: Task > Components: CBO >Reporter: Ashutosh Chauhan >Assignee: Ashutosh Chauhan > Attachments: HIVE-10328.1.patch, HIVE-10328.10.patch, > HIVE-10328.2.patch, HIVE-10328.3.patch, HIVE-10328.4.patch, > HIVE-10328.4.patch, HIVE-10328.5.patch, HIVE-10328.6.patch, > HIVE-10328.7.patch, HIVE-10328.8.patch, HIVE-10328.9.patch, HIVE-10328.patch > > -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Updated] (HIVE-11802) Float-point numbers are displayed with different precision in Beeline/JDBC
[ https://issues.apache.org/jira/browse/HIVE-11802?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Sergio Peña updated HIVE-11802: --- Attachment: (was: HIVE-11802.1.patch) > Float-point numbers are displayed with different precision in Beeline/JDBC > -- > > Key: HIVE-11802 > URL: https://issues.apache.org/jira/browse/HIVE-11802 > Project: Hive > Issue Type: Bug >Affects Versions: 1.2.1 >Reporter: Sergio Peña >Assignee: Sergio Peña > Attachments: HIVE-11802.2.patch > > > When inserting float-point numbers to a table, the values displayed on > beeline or jdbc are with different precision. > How to reproduce: > {noformat} > 0: jdbc:hive2://localhost:1> create table decimals (f float, af > array, d double, ad array) stored as parquet; > No rows affected (0.294 seconds) > 0: jdbc:hive2://localhost:1> insert into table decimals select 1.10058, > array(cast(1.10058 as float)), 2.0133, array(2.0133) from dummy limit 1; > ... > No rows affected (20.089 seconds) > 0: jdbc:hive2://localhost:1> select f, af, af[0], d, ad[0] from decimals; > +-++-+-+-+--+ > | f | af | _c2 |d| _c4 | > +-++-+-+-+--+ > | 1.1005799770355225 | [1.10058] | 1.1005799770355225 | 2.0133 | 2.0133 | > +-++-+-+-+--+ > {noformat} > When displaying arrays, the values are displayed correctly, but if I print a > specific element, it is then displayed with more decimal positions. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Updated] (HIVE-11802) Float-point numbers are displayed with different precision in Beeline/JDBC
[ https://issues.apache.org/jira/browse/HIVE-11802?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Sergio Peña updated HIVE-11802: --- Attachment: HIVE-11802.2.patch > Float-point numbers are displayed with different precision in Beeline/JDBC > -- > > Key: HIVE-11802 > URL: https://issues.apache.org/jira/browse/HIVE-11802 > Project: Hive > Issue Type: Bug >Affects Versions: 1.2.1 >Reporter: Sergio Peña >Assignee: Sergio Peña > Attachments: HIVE-11802.2.patch > > > When inserting float-point numbers to a table, the values displayed on > beeline or jdbc are with different precision. > How to reproduce: > {noformat} > 0: jdbc:hive2://localhost:1> create table decimals (f float, af > array, d double, ad array) stored as parquet; > No rows affected (0.294 seconds) > 0: jdbc:hive2://localhost:1> insert into table decimals select 1.10058, > array(cast(1.10058 as float)), 2.0133, array(2.0133) from dummy limit 1; > ... > No rows affected (20.089 seconds) > 0: jdbc:hive2://localhost:1> select f, af, af[0], d, ad[0] from decimals; > +-++-+-+-+--+ > | f | af | _c2 |d| _c4 | > +-++-+-+-+--+ > | 1.1005799770355225 | [1.10058] | 1.1005799770355225 | 2.0133 | 2.0133 | > +-++-+-+-+--+ > {noformat} > When displaying arrays, the values are displayed correctly, but if I print a > specific element, it is then displayed with more decimal positions. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HIVE-11802) Float-point numbers are displayed with different precision in Beeline/JDBC
[ https://issues.apache.org/jira/browse/HIVE-11802?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14742346#comment-14742346 ] Sergio Peña commented on HIVE-11802: [~cwsteinbach] I submitted another patch with the unit tests. This issue happens only on beeline and JDBC connections. The values are correctly displayed on HiveCli. I believe the correct precision is corrected when converting the values to JSON string. But on JDBC, the values are sent with a different precision to the client. > Float-point numbers are displayed with different precision in Beeline/JDBC > -- > > Key: HIVE-11802 > URL: https://issues.apache.org/jira/browse/HIVE-11802 > Project: Hive > Issue Type: Bug >Affects Versions: 1.2.1 >Reporter: Sergio Peña >Assignee: Sergio Peña > Attachments: HIVE-11802.2.patch > > > When inserting float-point numbers to a table, the values displayed on > beeline or jdbc are with different precision. > How to reproduce: > {noformat} > 0: jdbc:hive2://localhost:1> create table decimals (f float, af > array, d double, ad array) stored as parquet; > No rows affected (0.294 seconds) > 0: jdbc:hive2://localhost:1> insert into table decimals select 1.10058, > array(cast(1.10058 as float)), 2.0133, array(2.0133) from dummy limit 1; > ... > No rows affected (20.089 seconds) > 0: jdbc:hive2://localhost:1> select f, af, af[0], d, ad[0] from decimals; > +-++-+-+-+--+ > | f | af | _c2 |d| _c4 | > +-++-+-+-+--+ > | 1.1005799770355225 | [1.10058] | 1.1005799770355225 | 2.0133 | 2.0133 | > +-++-+-+-+--+ > {noformat} > When displaying arrays, the values are displayed correctly, but if I print a > specific element, it is then displayed with more decimal positions. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Updated] (HIVE-3752) Add a non-sql API in hive to access data.
[ https://issues.apache.org/jira/browse/HIVE-3752?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Nitay Joffe updated HIVE-3752: -- Assignee: (was: Nitay Joffe) > Add a non-sql API in hive to access data. > - > > Key: HIVE-3752 > URL: https://issues.apache.org/jira/browse/HIVE-3752 > Project: Hive > Issue Type: Improvement >Reporter: Nitay Joffe > > We would like to add an input/output format for accessing Hive data in Hadoop > directly without having to use e.g. a transform. Using a transform > means having to do a whole map-reduce step with its own disk accesses and its > imposed structure. It also means needing to have Hive be the base > infrastructure for the entire system being developed which is not the right > fit as we only need a small part of it (access to the data). > So we propose adding an API level InputFormat and OutputFormat to Hive that > will make it trivially easy to select a table with partition spec and read > from / write to it. We chose this design to make it compatible with Hadoop so > that existing systems that work with Hadoop's IO API will just work out of > the box. > We need this system for the Giraph graph processing system > (http://giraph.apache.org/) as running graph jobs which read/write from Hive > is a common use case. > [~namitjain] [~aching] [~kevinwilfong] [~apresta] > Input-side (HiveApiInputFormat) review: https://reviews.facebook.net/D7401 -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HIVE-3752) Add a non-sql API in hive to access data.
[ https://issues.apache.org/jira/browse/HIVE-3752?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14742342#comment-14742342 ] Lefty Leverenz commented on HIVE-3752: -- [~nitay], I took the liberty of revising the last line of your proposal because it looked like a cut-&-paste error from the reading section. I changed "HiveApiInputFormat" to "HiveApiOutputFormat" and "Hadoop-compatible reading system" to "Hadoop-compatible writing system" -- okay? > Add a non-sql API in hive to access data. > - > > Key: HIVE-3752 > URL: https://issues.apache.org/jira/browse/HIVE-3752 > Project: Hive > Issue Type: Improvement >Reporter: Nitay Joffe > > We would like to add an input/output format for accessing Hive data in Hadoop > directly without having to use e.g. a transform. Using a transform > means having to do a whole map-reduce step with its own disk accesses and its > imposed structure. It also means needing to have Hive be the base > infrastructure for the entire system being developed which is not the right > fit as we only need a small part of it (access to the data). > So we propose adding an API level InputFormat and OutputFormat to Hive that > will make it trivially easy to select a table with partition spec and read > from / write to it. We chose this design to make it compatible with Hadoop so > that existing systems that work with Hadoop's IO API will just work out of > the box. > We need this system for the Giraph graph processing system > (http://giraph.apache.org/) as running graph jobs which read/write from Hive > is a common use case. > [~namitjain] [~aching] [~kevinwilfong] [~apresta] > Input-side (HiveApiInputFormat) review: https://reviews.facebook.net/D7401 -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HIVE-11774) Show macro definition for desc function
[ https://issues.apache.org/jira/browse/HIVE-11774?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14741955#comment-14741955 ] Hive QA commented on HIVE-11774: {color:red}Overall{color}: -1 at least one tests failed Here are the results of testing the latest attachment: https://issues.apache.org/jira/secure/attachment/12755330/HIVE-11774.2.patch.txt {color:red}ERROR:{color} -1 due to 2 failed/errored test(s), 9424 tests executed *Failed tests:* {noformat} org.apache.hadoop.hive.cli.TestNegativeMinimrCliDriver.testNegativeCliDriver_minimr_broken_pipe org.apache.hive.hcatalog.api.TestHCatClient.testTableSchemaPropagation {noformat} Test results: http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/5251/testReport Console output: http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/5251/console Test logs: http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-5251/ Messages: {noformat} Executing org.apache.hive.ptest.execution.PrepPhase Executing org.apache.hive.ptest.execution.ExecutionPhase Executing org.apache.hive.ptest.execution.ReportingPhase Tests exited with: TestsFailedException: 2 tests failed {noformat} This message is automatically generated. ATTACHMENT ID: 12755330 - PreCommit-HIVE-TRUNK-Build > Show macro definition for desc function > > > Key: HIVE-11774 > URL: https://issues.apache.org/jira/browse/HIVE-11774 > Project: Hive > Issue Type: Improvement > Components: Query Processor >Reporter: Navis >Assignee: Navis >Priority: Trivial > Attachments: HIVE-11774.1.patch.txt, HIVE-11774.2.patch.txt > > > Currently, desc function shows nothing for macro. It would be helpful if it > shows the definition of it. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HIVE-11762) TestHCatLoaderEncryption failures when using Hadoop 2.7
[ https://issues.apache.org/jira/browse/HIVE-11762?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14741918#comment-14741918 ] Hive QA commented on HIVE-11762: {color:red}Overall{color}: -1 at least one tests failed Here are the results of testing the latest attachment: https://issues.apache.org/jira/secure/attachment/12755278/HIVE-11762.3.patch {color:red}ERROR:{color} -1 due to 3 failed/errored test(s), 9424 tests executed *Failed tests:* {noformat} org.apache.hive.hcatalog.api.TestHCatClient.testTableSchemaPropagation org.apache.hive.hcatalog.streaming.TestStreaming.testEndpointConnection org.apache.hive.hcatalog.streaming.TestStreaming.testTransactionBatchEmptyCommit {noformat} Test results: http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/5249/testReport Console output: http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/5249/console Test logs: http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-5249/ Messages: {noformat} Executing org.apache.hive.ptest.execution.PrepPhase Executing org.apache.hive.ptest.execution.ExecutionPhase Executing org.apache.hive.ptest.execution.ReportingPhase Tests exited with: TestsFailedException: 3 tests failed {noformat} This message is automatically generated. ATTACHMENT ID: 12755278 - PreCommit-HIVE-TRUNK-Build > TestHCatLoaderEncryption failures when using Hadoop 2.7 > --- > > Key: HIVE-11762 > URL: https://issues.apache.org/jira/browse/HIVE-11762 > Project: Hive > Issue Type: Bug > Components: Shims, Tests >Reporter: Jason Dere >Assignee: Jason Dere > Attachments: HIVE-11762.1.patch, HIVE-11762.2.patch, > HIVE-11762.3.patch > > > When running TestHCatLoaderEncryption with -Dhadoop23.version=2.7.0, we get > the following error during setup(): > {noformat} > testReadDataFromEncryptedHiveTableByPig[5](org.apache.hive.hcatalog.pig.TestHCatLoaderEncryption) > Time elapsed: 3.648 sec <<< ERROR! > java.lang.NoSuchMethodError: > org.apache.hadoop.hdfs.DFSClient.setKeyProvider(Lorg/apache/hadoop/crypto/key/KeyProviderCryptoExtension;)V > at > org.apache.hadoop.hive.shims.Hadoop23Shims.getMiniDfs(Hadoop23Shims.java:534) > at > org.apache.hive.hcatalog.pig.TestHCatLoaderEncryption.initEncryptionShim(TestHCatLoaderEncryption.java:252) > at > org.apache.hive.hcatalog.pig.TestHCatLoaderEncryption.setup(TestHCatLoaderEncryption.java:200) > {noformat} > It looks like between Hadoop 2.6 and Hadoop 2.7, the argument to > DFSClient.setKeyProvider() changed: > {noformat} >@VisibleForTesting > - public void setKeyProvider(KeyProviderCryptoExtension provider) { > -this.provider = provider; > + public void setKeyProvider(KeyProvider provider) { > {noformat} -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HIVE-9756) LLAP: use log4j 2 for llap
[ https://issues.apache.org/jira/browse/HIVE-9756?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14741919#comment-14741919 ] Hive QA commented on HIVE-9756: --- {color:red}Overall{color}: -1 no tests executed Here are the results of testing the latest attachment: https://issues.apache.org/jira/secure/attachment/12754319/HIVE-9756.2.patch Test results: http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/5250/testReport Console output: http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/5250/console Test logs: http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-5250/ Messages: {noformat} Executing org.apache.hive.ptest.execution.PrepPhase Tests exited with: NonZeroExitCodeException Command 'bash /data/hive-ptest/working/scratch/source-prep.sh' failed with exit status 1 and output '+ [[ -n /usr/java/jdk1.7.0_45-cloudera ]] + export JAVA_HOME=/usr/java/jdk1.7.0_45-cloudera + JAVA_HOME=/usr/java/jdk1.7.0_45-cloudera + export PATH=/usr/java/jdk1.7.0_45-cloudera/bin/:/usr/local/apache-maven-3.0.5/bin:/usr/java/jdk1.7.0_45-cloudera/bin:/usr/local/apache-ant-1.9.1/bin:/usr/local/bin:/bin:/usr/bin:/usr/local/sbin:/usr/sbin:/sbin:/home/hiveptest/bin + PATH=/usr/java/jdk1.7.0_45-cloudera/bin/:/usr/local/apache-maven-3.0.5/bin:/usr/java/jdk1.7.0_45-cloudera/bin:/usr/local/apache-ant-1.9.1/bin:/usr/local/bin:/bin:/usr/bin:/usr/local/sbin:/usr/sbin:/sbin:/home/hiveptest/bin + export 'ANT_OPTS=-Xmx1g -XX:MaxPermSize=256m ' + ANT_OPTS='-Xmx1g -XX:MaxPermSize=256m ' + export 'M2_OPTS=-Xmx1g -XX:MaxPermSize=256m -Dhttp.proxyHost=localhost -Dhttp.proxyPort=3128' + M2_OPTS='-Xmx1g -XX:MaxPermSize=256m -Dhttp.proxyHost=localhost -Dhttp.proxyPort=3128' + cd /data/hive-ptest/working/ + tee /data/hive-ptest/logs/PreCommit-HIVE-TRUNK-Build-5250/source-prep.txt + [[ false == \t\r\u\e ]] + mkdir -p maven ivy + [[ git = \s\v\n ]] + [[ git = \g\i\t ]] + [[ -z master ]] + [[ -d apache-github-source-source ]] + [[ ! -d apache-github-source-source/.git ]] + [[ ! -d apache-github-source-source ]] + cd apache-github-source-source + git fetch origin + git reset --hard HEAD HEAD is now at b1fffd5 HIVE-11781 : Remove HiveLimit operator (Jesus Camacho Rodriguez via Ashutosh Chauhan) + git clean -f -d + git checkout master Already on 'master' + git reset --hard origin/master HEAD is now at b1fffd5 HIVE-11781 : Remove HiveLimit operator (Jesus Camacho Rodriguez via Ashutosh Chauhan) + git merge --ff-only origin/master Already up-to-date. + git gc + patchCommandPath=/data/hive-ptest/working/scratch/smart-apply-patch.sh + patchFilePath=/data/hive-ptest/working/scratch/build.patch + [[ -f /data/hive-ptest/working/scratch/build.patch ]] + chmod +x /data/hive-ptest/working/scratch/smart-apply-patch.sh + /data/hive-ptest/working/scratch/smart-apply-patch.sh /data/hive-ptest/working/scratch/build.patch The patch does not appear to apply with p0, p1, or p2 + exit 1 ' {noformat} This message is automatically generated. ATTACHMENT ID: 12754319 - PreCommit-HIVE-TRUNK-Build > LLAP: use log4j 2 for llap > -- > > Key: HIVE-9756 > URL: https://issues.apache.org/jira/browse/HIVE-9756 > Project: Hive > Issue Type: Sub-task >Reporter: Gunther Hagleitner >Assignee: Prasanth Jayachandran > Attachments: HIVE-9756.1.patch, HIVE-9756.2.patch > > > For the INFO logging, we'll need to use the log4j-jcl 2.x upgrade-path to get > throughput friendly logging. > http://logging.apache.org/log4j/2.0/manual/async.html#Performance -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Updated] (HIVE-11699) Support special characters in quoted table names
[ https://issues.apache.org/jira/browse/HIVE-11699?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Pengcheng Xiong updated HIVE-11699: --- Attachment: HIVE-11699.02.patch > Support special characters in quoted table names > > > Key: HIVE-11699 > URL: https://issues.apache.org/jira/browse/HIVE-11699 > Project: Hive > Issue Type: New Feature >Reporter: Pengcheng Xiong >Assignee: Pengcheng Xiong > Attachments: HIVE-11699.01.patch, HIVE-11699.02.patch > > > Right now table names can only be "[a-zA-z_0-9]+". This patch tries to > investigate how much change there should be if we would like to support > special characters, e.g., "/" in table names. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HIVE-11768) java.io.DeleteOnExitHook leaks memory on long running Hive Server2 Instances
[ https://issues.apache.org/jira/browse/HIVE-11768?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14741991#comment-14741991 ] Hive QA commented on HIVE-11768: {color:red}Overall{color}: -1 at least one tests failed Here are the results of testing the latest attachment: https://issues.apache.org/jira/secure/attachment/12755331/HIVE-11768.2.patch.txt {color:red}ERROR:{color} -1 due to 3 failed/errored test(s), 9424 tests executed *Failed tests:* {noformat} org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_index_auto_mult_tables_compact org.apache.hive.common.util.TestShutdownHookManager.shutdownHookManager org.apache.hive.hcatalog.api.TestHCatClient.testTableSchemaPropagation {noformat} Test results: http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/5252/testReport Console output: http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/5252/console Test logs: http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-5252/ Messages: {noformat} Executing org.apache.hive.ptest.execution.PrepPhase Executing org.apache.hive.ptest.execution.ExecutionPhase Executing org.apache.hive.ptest.execution.ReportingPhase Tests exited with: TestsFailedException: 3 tests failed {noformat} This message is automatically generated. ATTACHMENT ID: 12755331 - PreCommit-HIVE-TRUNK-Build > java.io.DeleteOnExitHook leaks memory on long running Hive Server2 Instances > > > Key: HIVE-11768 > URL: https://issues.apache.org/jira/browse/HIVE-11768 > Project: Hive > Issue Type: Bug > Components: HiveServer2 >Affects Versions: 1.2.1 >Reporter: Nemon Lou >Assignee: Navis > Attachments: HIVE-11768.1.patch.txt, HIVE-11768.2.patch.txt > > > More than 490,000 paths was added to java.io.DeleteOnExitHook on one of our > long running HiveServer2 instances,taken up more than 100MB on heap. > Most of the paths contains a suffix of ".pipeout". -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HIVE-11756) Avoid redundant key serialization in RS for distinct query
[ https://issues.apache.org/jira/browse/HIVE-11756?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14742014#comment-14742014 ] Hive QA commented on HIVE-11756: {color:red}Overall{color}: -1 at least one tests failed Here are the results of testing the latest attachment: https://issues.apache.org/jira/secure/attachment/12755347/HIVE-11756.3.patch.txt {color:red}ERROR:{color} -1 due to 108 failed/errored test(s), 9424 tests executed *Failed tests:* {noformat} org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_auto_join18 org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_auto_join18_multi_distinct org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_cbo_rp_udf_udaf org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_cbo_udf_udaf org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_count org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_groupby10 org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_groupby11 org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_groupby12 org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_groupby2_map org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_groupby2_map_multi_distinct org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_groupby2_map_skew org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_groupby2_noskew org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_groupby2_noskew_multi_distinct org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_groupby3 org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_groupby3_map org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_groupby3_map_multi_distinct org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_groupby3_map_skew org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_groupby3_noskew org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_groupby3_noskew_multi_distinct org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_groupby8 org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_groupby8_map org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_groupby8_map_skew org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_groupby8_noskew org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_groupby9 org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_groupby_bigdata org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_groupby_cube1 org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_groupby_distinct_samekey org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_groupby_map_ppr org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_groupby_map_ppr_multi_distinct org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_groupby_multi_insert_common_distinct org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_groupby_multi_single_reducer org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_groupby_multi_single_reducer2 org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_groupby_position org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_groupby_ppr org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_groupby_ppr_multi_distinct org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_groupby_rollup1 org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_groupby_sort_11 org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_groupby_sort_8 org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_join18 org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_join18_multi_distinct org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_limit_pushdown org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_lineage2 org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_metadataonly1 org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_multi_insert_gby3 org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_multi_insert_lateral_view org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_temp_table_gb1 org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_truncate_column_merge org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_udf_count org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_union17 org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_unionDistinct_1 org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_vector_count_distinct org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_vectorization_limit org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_vectorized_distinct_gby org.apache.hadoop.hive.cli.TestMiniTezCliDriver.testCliDriver_cbo_udf_udaf org.apache.hadoop.hive.cli.TestMiniTezCliDriver.testCliDriver_count org.apache.hadoop.hive.cli.TestMiniTezCliDriver.testCliDriver_groupby2 org.apache.hadoop.hive.cli.TestMiniTezCliDriver.testCliDriver_groupby3 org.apache.hadoop.hive.cli.TestMiniTezCliDriver.testCliDriver_limit_pushdown org.apache.hadoop.hive.cli.TestMiniTezCliDriver.testCliDriver_metadataonly1 org.apache.hadoop.hive.cli.TestMiniTezCliDriver.testCliDriver_tez_union_multiinsert org.apache.hadoop.hive.cli.TestMiniTezCliDriver.testCliDriver_unionDistinct_1
[jira] [Commented] (HIVE-7224) Set incremental printing to true by default in Beeline
[ https://issues.apache.org/jira/browse/HIVE-7224?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14742077#comment-14742077 ] Hive QA commented on HIVE-7224: --- {color:red}Overall{color}: -1 at least one tests failed Here are the results of testing the latest attachment: https://issues.apache.org/jira/secure/attachment/12755464/HIVE-7224.2.patch {color:red}ERROR:{color} -1 due to 14 failed/errored test(s), 9424 tests executed *Failed tests:* {noformat} org.apache.hadoop.hive.metastore.txn.TestCompactionTxnHandler.testRevokeTimedOutWorkers org.apache.hive.beeline.TestBeeLineWithArgs.testCSV2OutputWithDoubleQuotes org.apache.hive.beeline.TestBeeLineWithArgs.testCSVOutput org.apache.hive.beeline.TestBeeLineWithArgs.testCSVOutputWithDoubleQuotes org.apache.hive.beeline.TestBeeLineWithArgs.testDSVOutput org.apache.hive.beeline.TestBeeLineWithArgs.testDSVOutputWithDoubleQuotes org.apache.hive.beeline.TestBeeLineWithArgs.testNullEmpty org.apache.hive.beeline.TestBeeLineWithArgs.testNullEmptyCmdArg org.apache.hive.beeline.TestBeeLineWithArgs.testTSV2Output org.apache.hive.beeline.TestBeeLineWithArgs.testTSV2OutputWithDoubleQuotes org.apache.hive.beeline.TestBeeLineWithArgs.testTSVOutput org.apache.hive.beeline.TestBeeLineWithArgs.testTSVOutputWithDoubleQuotes org.apache.hive.hcatalog.api.TestHCatClient.testTableSchemaPropagation org.apache.hive.jdbc.TestSSL.testSSLFetchHttp {noformat} Test results: http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/5255/testReport Console output: http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/5255/console Test logs: http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-5255/ Messages: {noformat} Executing org.apache.hive.ptest.execution.PrepPhase Executing org.apache.hive.ptest.execution.ExecutionPhase Executing org.apache.hive.ptest.execution.ReportingPhase Tests exited with: TestsFailedException: 14 tests failed {noformat} This message is automatically generated. ATTACHMENT ID: 12755464 - PreCommit-HIVE-TRUNK-Build > Set incremental printing to true by default in Beeline > -- > > Key: HIVE-7224 > URL: https://issues.apache.org/jira/browse/HIVE-7224 > Project: Hive > Issue Type: Bug > Components: Clients, JDBC >Affects Versions: 0.13.0, 1.0.0, 1.2.0, 1.1.0 >Reporter: Vaibhav Gumashta >Assignee: Vaibhav Gumashta > Attachments: HIVE-7224.1.patch, HIVE-7224.2.patch, HIVE-7224.2.patch > > > See HIVE-7221. > By default beeline tries to buffer the entire output relation before printing > it on stdout. This can cause OOM when the output relation is large. However, > beeline has the option of incremental prints. We should keep that as the > default. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HIVE-11763) Use * instead of sum(hash(*)) on Parquet predicate (PPD) integration tests
[ https://issues.apache.org/jira/browse/HIVE-11763?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14742048#comment-14742048 ] Hive QA commented on HIVE-11763: {color:red}Overall{color}: -1 at least one tests failed Here are the results of testing the latest attachment: https://issues.apache.org/jira/secure/attachment/12755415/HIVE-11763.2.patch {color:red}ERROR:{color} -1 due to 1 failed/errored test(s), 9424 tests executed *Failed tests:* {noformat} org.apache.hive.hcatalog.api.TestHCatClient.testTableSchemaPropagation {noformat} Test results: http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/5254/testReport Console output: http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/5254/console Test logs: http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-5254/ Messages: {noformat} Executing org.apache.hive.ptest.execution.PrepPhase Executing org.apache.hive.ptest.execution.ExecutionPhase Executing org.apache.hive.ptest.execution.ReportingPhase Tests exited with: TestsFailedException: 1 tests failed {noformat} This message is automatically generated. ATTACHMENT ID: 12755415 - PreCommit-HIVE-TRUNK-Build > Use * instead of sum(hash(*)) on Parquet predicate (PPD) integration tests > -- > > Key: HIVE-11763 > URL: https://issues.apache.org/jira/browse/HIVE-11763 > Project: Hive > Issue Type: Sub-task >Reporter: Sergio Peña >Assignee: Sergio Peña > Attachments: HIVE-11763.2.patch > > > The integration tests for Parquet predicate push down (PPD) use the following > query to validate the values filtered: > {noformat} > select sum(hash(*)) from ... > {noformat} > It would be better if we use {{select * from ...}} instead to see that those > values are correct. It is difficult to see if a value was filtered by seeing > the hash. > Also, we can try to limit the number of rows of the INSERT ... SELECT > statmenet to avoid displaying many rows when validating the data. I think a > LIMIT 2 on each of the SELECT. > For example, the parquet_ppd_boolean.ppd has this: > {noformat} > insert overwrite table newtypestbl select * from (select cast("apple" as > char(10)), cast("bee" as varchar(10)), 0.22, true from src src1 union all > select cast("hello" as char(10)), cast("world" as varchar(10)), 11.22, false > from src src2) uniontbl; > {noformat} > If we use LIMIT 2, then we will reduce the # of rows: > {noformat} > insert overwrite table newtypestbl select * from (select cast("apple" as > char(10)), cast("bee" as varchar(10)), 0.22, true from src src1 LIMIT 2 union > all select cast("hello" as char(10)), cast("world" as varchar(10)), 11.22, > false from src src2 LIMIT 2) uniontbl; > {noformat} -- This message was sent by Atlassian JIRA (v6.3.4#6332)