Re: ORC separate project
Hi Owen, I'd like to get involved. Thanks, Xuefu On Thu, Mar 19, 2015 at 2:44 PM, Owen O'Malley omal...@apache.org wrote: All, Over the last year, there has been a fair number of projects that want to integrate with ORC, but don't want a dependence on Hive's exec jar. Additionally, we've been working on a C++ reader (and soon writer) and it would be great to host them both in the same project. Toward that end, I'd like to create a separate ORC project at Apache. There will be lots of technical details to work out, but I wanted to give the Hive community a chance to discuss it. Do any of the Hive committers want to be included on the proposal? Of the current Hive committers, my list looks like: * Alan * Gunther * Prasanth * Lefty * Owen * Sergey * Gopal * Kevin Did I miss anyone? Thanks! Owen
Re: Review Request 32254: HIVE-10007:Support qualified table name in analyze table compute statistics for columns
--- This is an automatically generated e-mail. To reply, visit: https://reviews.apache.org/r/32254/ --- (Updated March 20, 2015, 12:08 a.m.) Review request for hive, Ashutosh Chauhan, Szehon Ho, and Xuefu Zhang. Changes --- Fixed three failed tests and add a new case to compute column stats for a table not under default db. Bugs: HIVE-10007 https://issues.apache.org/jira/browse/HIVE-10007 Repository: hive-git Description --- Currently Currently analyze table compute statistics for columns command can not compute column stats for a table in a different database since it does not support qualified table name. You need switch to that table database in order to compute its column stats. So does it for ALTER TABLE .. UPDATE STATISTICS FOR COLUMN command. This JIRA will provide the support to qualified table name in analyze and update column stats commands. Diffs (updated) - ql/src/java/org/apache/hadoop/hive/ql/exec/ColumnStatsTask.java e2f696eea7b472d4e417629eedd33406b41caf98 ql/src/java/org/apache/hadoop/hive/ql/exec/ColumnStatsUpdateTask.java 28c17b75356eb4ffaac2193a80b27bdd67d98009 ql/src/java/org/apache/hadoop/hive/ql/parse/ColumnStatsSemanticAnalyzer.java c83523e507f1003eff1a821c216a7f6904f293ed ql/src/java/org/apache/hadoop/hive/ql/parse/DDLSemanticAnalyzer.java b9e15a1b2f60b291759444a4407acaf1e0d9c515 ql/src/test/queries/clientpositive/alter_partition_update_status.q 1eee9a506f3854eace7deaa3f4ec02c1f7bd3479 ql/src/test/queries/clientpositive/alter_table_update_status.q fd45cd426c17a7bfa1371081dd81ff6eb1b7353d ql/src/test/queries/clientpositive/columnstats_partlvl.q 82a9e0f0fef0c9bf909028f23ccd47ee07a61a2b ql/src/test/queries/clientpositive/columnstats_tbllvl.q 07cc959ec6bb9f5e1f1ba3184da5ce3fc8bf951b ql/src/test/results/clientpositive/alter_partition_update_status.q.out 7e33a7e8ecc07b2b9d8ab8b0f12b383e5a442cf3 ql/src/test/results/clientpositive/alter_table_update_status.q.out 361359807b9c8c1a2e25a629fd53f026e02957e7 ql/src/test/results/clientpositive/columnstats_partlvl.q.out a86c5fbd718a490c5973382810cc02c10184925e ql/src/test/results/clientpositive/columnstats_partlvl_dp.q.out 073b387c7b1d5f174928a110381faa851133c7ea ql/src/test/results/clientpositive/columnstats_tbllvl.q.out 0ef04f5f4b70d84cd6118c2407d29fc555bac116 ql/src/test/results/clientpositive/display_colstats_tbllvl.q.out a176d9a2a9db7eb848d1db022c4d99680db1b62c ql/src/test/results/clientpositive/temp_table_display_colstats_tbllvl.q.out 933d24b6cde934faa013f1ea102b9b753a7d44a4 Diff: https://reviews.apache.org/r/32254/diff/ Testing --- 1. Performed some manullay tests 2. Added new qtests for those cases 3. Precommit build Thanks, Chaoyu Tang
[jira] [Created] (HIVE-10026) LLAP: AM should get notifications on daemons going down or restarting
Siddharth Seth created HIVE-10026: - Summary: LLAP: AM should get notifications on daemons going down or restarting Key: HIVE-10026 URL: https://issues.apache.org/jira/browse/HIVE-10026 Project: Hive Issue Type: Sub-task Reporter: Siddharth Seth There's lost state otherwise, which can cause queries to hang. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
Review Request 32268: HIVE-9998 Vectorization support for interval types
--- This is an automatically generated e-mail. To reply, visit: https://reviews.apache.org/r/32268/ --- Review request for hive, Ashutosh Chauhan and Matt McCline. Bugs: HIVE-9998 https://issues.apache.org/jira/browse/HIVE-9998 Repository: hive-git Description --- Enables support for vectorized interval types This also fixes some vectorized comparisons for Date, when one of the left or right side is a constant expression. Diffs - ant/src/org/apache/hadoop/hive/ant/GenVectorCode.java 375c173 common/src/java/org/apache/hive/common/util/DateTimeMath.java 28030e6 common/src/java/org/apache/hive/common/util/DateUtils.java b4159d3 common/src/test/org/apache/hive/common/util/TestDateTimeMath.java 4886576 ql/src/gen/vectorization/ExpressionTemplates/ColumnArithmeticColumnWithConvert.txt PRE-CREATION ql/src/gen/vectorization/ExpressionTemplates/ColumnArithmeticScalarWithConvert.txt PRE-CREATION ql/src/gen/vectorization/ExpressionTemplates/ColumnUnaryMinus.txt 6bf6def ql/src/gen/vectorization/ExpressionTemplates/DTIColumnArithmeticDTIColumnNoConvert.txt PRE-CREATION ql/src/gen/vectorization/ExpressionTemplates/DTIColumnArithmeticDTIScalarNoConvert.txt PRE-CREATION ql/src/gen/vectorization/ExpressionTemplates/DTIColumnCompareScalar.txt PRE-CREATION ql/src/gen/vectorization/ExpressionTemplates/DTIScalarArithmeticDTIColumnNoConvert.txt PRE-CREATION ql/src/gen/vectorization/ExpressionTemplates/DTIScalarCompareColumn.txt PRE-CREATION ql/src/gen/vectorization/ExpressionTemplates/DateTimeColumnArithmeticIntervalColumnWithConvert.txt PRE-CREATION ql/src/gen/vectorization/ExpressionTemplates/DateTimeColumnArithmeticIntervalScalarWithConvert.txt PRE-CREATION ql/src/gen/vectorization/ExpressionTemplates/DateTimeScalarArithmeticIntervalColumnWithConvert.txt PRE-CREATION ql/src/gen/vectorization/ExpressionTemplates/FilterDTIColumnCompareScalar.txt PRE-CREATION ql/src/gen/vectorization/ExpressionTemplates/FilterDTIScalarCompareColumn.txt PRE-CREATION ql/src/gen/vectorization/ExpressionTemplates/IntervalColumnArithmeticDateTimeColumnWithConvert.txt PRE-CREATION ql/src/gen/vectorization/ExpressionTemplates/IntervalColumnArithmeticDateTimeScalarWithConvert.txt PRE-CREATION ql/src/gen/vectorization/ExpressionTemplates/IntervalScalarArithmeticDateTimeColumnWithConvert.txt PRE-CREATION ql/src/gen/vectorization/ExpressionTemplates/ScalarArithmeticColumnWithConvert.txt PRE-CREATION ql/src/java/org/apache/hadoop/hive/ql/exec/vector/TimestampUtils.java 352e43e ql/src/java/org/apache/hadoop/hive/ql/exec/vector/VectorColumnAssignFactory.java c915f72 ql/src/java/org/apache/hadoop/hive/ql/exec/vector/VectorExpressionDescriptor.java bb18b32 ql/src/java/org/apache/hadoop/hive/ql/exec/vector/VectorizationContext.java 5201c57 ql/src/java/org/apache/hadoop/hive/ql/exec/vector/VectorizedBatchUtil.java e304cf8 ql/src/java/org/apache/hadoop/hive/ql/exec/vector/VectorizedRowBatchCtx.java 88ec2b2 ql/src/java/org/apache/hadoop/hive/ql/exec/vector/expressions/CastStringToIntervalDayTime.java PRE-CREATION ql/src/java/org/apache/hadoop/hive/ql/exec/vector/expressions/CastStringToIntervalYearMonth.java PRE-CREATION ql/src/java/org/apache/hadoop/hive/ql/exec/vector/expressions/VectorExpression.java d7ace6d ql/src/java/org/apache/hadoop/hive/ql/exec/vector/expressions/VectorExpressionWriterFactory.java 94a47e0 ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/translator/ASTBuilder.java 10bf2bd ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/Vectorizer.java aca4273 ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFOPDTIMinus.java a32c133 ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFOPDTIPlus.java 9a5c3a9 ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFOPEqual.java 3870b51 ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFOPEqualOrGreaterThan.java 65e1835 ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFOPEqualOrLessThan.java 3e4a1d2 ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFOPGreaterThan.java df7a857 ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFOPLessThan.java fafd99b ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFOPMinus.java 18fbb5a ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFOPNotEqual.java 0436488 ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFOPPlus.java bfac5a8 ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFToIntervalDayTime.java 89c3988 ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFToIntervalYearMonth.java 5c05655 ql/src/java/org/apache/hadoop/hive/ql/util/DateTimeMath.java PRE-CREATION
[jira] [Created] (HIVE-10025) LLAP: Queued work times out
Siddharth Seth created HIVE-10025: - Summary: LLAP: Queued work times out Key: HIVE-10025 URL: https://issues.apache.org/jira/browse/HIVE-10025 Project: Hive Issue Type: Improvement Reporter: Siddharth Seth If a daemon holds a task in queue for a long time, it'll eventually time out - but isn't removed from the queue. Ideally, it shouldn't be allowed to time out. Otherwise, handle the timeout so that the task doesn't run - or starts and fails - likely a change in the TaskCommunicator. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Created] (HIVE-10016) Remove duplicated Hive table schema parsing in DataWritableReadSupport
Dong Chen created HIVE-10016: Summary: Remove duplicated Hive table schema parsing in DataWritableReadSupport Key: HIVE-10016 URL: https://issues.apache.org/jira/browse/HIVE-10016 Project: Hive Issue Type: Sub-task Reporter: Dong Chen Assignee: Dong Chen In {{DataWritableReadSupport.init()}}, the table schema is created and its string format is set in conf. When construct the {{ParquetRecordReaderWrapper}} , the schema is fetched from conf and parsed several times. We could remove these schema parsing, and improve the speed of getRecordReader a bit. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Created] (HIVE-10017) SparkTask log improvement [Spark Branch]
Chinna Rao Lalam created HIVE-10017: --- Summary: SparkTask log improvement [Spark Branch] Key: HIVE-10017 URL: https://issues.apache.org/jira/browse/HIVE-10017 Project: Hive Issue Type: Bug Components: Spark Reporter: Chinna Rao Lalam Priority: Minor Fix For: spark-branch Initialize log object in the own class for better log message. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
Re: Review Request 31800: HIVE-9658 Reduce parquet memory use by bypassing java primitive objects on ETypeConverter
--- This is an automatically generated e-mail. To reply, visit: https://reviews.apache.org/r/31800/#review77168 --- Hi Sergio, thank you for your update. Just few more minor suggestions. ql/src/java/org/apache/hadoop/hive/ql/exec/vector/VectorColumnAssignFactory.java https://reviews.apache.org/r/31800/#comment125044 Is that possible to pass down the switch block into buildObjectAssignMethod? In current implement, we already do this kind of thing in that method. ql/src/java/org/apache/hadoop/hive/ql/io/parquet/convert/HiveCollectionConverter.java https://reviews.apache.org/r/31800/#comment125045 How about moving this code block to the top of the class definition? Object objs[] = new Object[] { null, null }; ql/src/java/org/apache/hadoop/hive/ql/io/parquet/convert/HiveStructConverter.java https://reviews.apache.org/r/31800/#comment125046 Arrrays.fill(elements,null) ql/src/java/org/apache/hadoop/hive/ql/io/parquet/read/ParquetRecordReaderWrapper.java.orig https://reviews.apache.org/r/31800/#comment125047 Please remove this file. It shouldn't be committed into trunk. ql/src/test/org/apache/hadoop/hive/ql/io/parquet/TestDataWritableWriter.java https://reviews.apache.org/r/31800/#comment125048 Please only import needed packages. - cheng xu On March 10, 2015, 6:02 p.m., Sergio Pena wrote: --- This is an automatically generated e-mail. To reply, visit: https://reviews.apache.org/r/31800/ --- (Updated March 10, 2015, 6:02 p.m.) Review request for hive, Ryan Blue and cheng xu. Bugs: HIVE-9658 https://issues.apache.org/jira/browse/HIVE-9658 Repository: hive-git Description --- This patch bypasses primitive java objects to hive object inspectors without using primitive Writable objects. It helps to reduce memory usage. I did not bypass other complex objects, such as binaries, decimal and date/timestamp, because their Writable objects are needed in other parts of the code, and creating them later takes more ops/s to do it. Better save time at the beginning. Diffs - itests/hive-jmh/src/main/java/org/apache/hive/benchmark/storage/ColumnarStorageBench.java 4f6985cd13017ce37f4f0c100b16a27aa5b02f8b ql/src/java/org/apache/hadoop/hive/ql/exec/vector/VectorColumnAssignFactory.java c915f728fc9b27da0fabefab5d8f5faa53640b78 ql/src/java/org/apache/hadoop/hive/ql/io/parquet/MapredParquetInputFormat.java 0391229723cc3ecef551fa44b8456b0d2ac93fb5 ql/src/java/org/apache/hadoop/hive/ql/io/parquet/VectorizedParquetInputFormat.java d7edd52614771857d1b21971a66894841c248ef9 ql/src/java/org/apache/hadoop/hive/ql/io/parquet/convert/ConverterParent.java 6ff6b473c9f1867bc14bb597094ddb92487cc954 ql/src/java/org/apache/hadoop/hive/ql/io/parquet/convert/DataWritableRecordConverter.java a43661eb54ba29692c07c264584b5aecf648ef99 ql/src/java/org/apache/hadoop/hive/ql/io/parquet/convert/ETypeConverter.java 3fc012970e23bbc188ce2a2e2ba0b04bc6f22317 ql/src/java/org/apache/hadoop/hive/ql/io/parquet/convert/HiveCollectionConverter.java f1c8b6f13718b37f590263e5b35ed6c327f5cf4f ql/src/java/org/apache/hadoop/hive/ql/io/parquet/convert/HiveGroupConverter.java c6d03a19029d5bcc86b998dd7a8609973648c103 ql/src/java/org/apache/hadoop/hive/ql/io/parquet/convert/HiveStructConverter.java f95d15eddc21bc432fa53572de5756751a13341a ql/src/java/org/apache/hadoop/hive/ql/io/parquet/convert/Repeated.java ee57b31dac53d99af0c5a520f51102796ca32fd3 ql/src/java/org/apache/hadoop/hive/ql/io/parquet/read/DataWritableReadSupport.java 57ae7a9740d55b407cadfc8bc030593b29f90700 ql/src/java/org/apache/hadoop/hive/ql/io/parquet/read/ParquetRecordReaderWrapper.java a26199612cf338e336f210f29acb0398c536e1f9 ql/src/java/org/apache/hadoop/hive/ql/io/parquet/read/ParquetRecordReaderWrapper.java.orig PRE-CREATION ql/src/java/org/apache/hadoop/hive/ql/io/parquet/serde/AbstractParquetMapInspector.java 49bf1c5325833993f4c09efdf1546af560783c28 ql/src/java/org/apache/hadoop/hive/ql/io/parquet/serde/ArrayWritableObjectInspector.java 609188206f88e296d893b84bcaaab53f974e6b7d ql/src/java/org/apache/hadoop/hive/ql/io/parquet/serde/DeepParquetHiveMapInspector.java 143d72e76502d4877e8208181d9743259051dcea ql/src/java/org/apache/hadoop/hive/ql/io/parquet/serde/ObjectArrayWritableObjectInspector.java PRE-CREATION ql/src/java/org/apache/hadoop/hive/ql/io/parquet/serde/ParquetHiveArrayInspector.java bde0dcbb3978ba47b15ae2c9bbe2f87ed3984ab1 ql/src/java/org/apache/hadoop/hive/ql/io/parquet/serde/ParquetHiveSerDe.java 7fd5e9612d4e3c9bf3b816bc48dbdbe59fb8a5a8
[jira] [Created] (HIVE-10030) The unit test of testNewInputFormat got failure
Yi Zhou created HIVE-10030: -- Summary: The unit test of testNewInputFormat got failure Key: HIVE-10030 URL: https://issues.apache.org/jira/browse/HIVE-10030 Project: Hive Issue Type: Bug Components: Tests Affects Versions: 0.14.0 Reporter: Yi Zhou Running org.apache.hadoop.hive.ql.io.orc.TestNewInputOutputFormat Tests run: 5, Failures: 1, Errors: 0, Skipped: 0, Time elapsed: 10.126 sec FAILURE! - in org.apache.hadoop.hive.ql.io.orc.TestNewInputOutputFormat testNewInputFormat(org.apache.hadoop.hive.ql.io.orc.TestNewInputOutputFormat) Time elapsed: 4.938 sec FAILURE! junit.framework.ComparisonFailure: expected:...in}, {1234, hat}], {[mauddib={1, mauddib}, chani={5, chani]}}, 2000-03-12 15:00... but was:...in}, {1234, hat}], {[chani={5, chani}, mauddib={1, mauddib]}}, 2000-03-12 15:00... at junit.framework.Assert.assertEquals(Assert.java:100) at junit.framework.Assert.assertEquals(Assert.java:107) at org.apache.hadoop.hive.ql.io.orc.TestNewInputOutputFormat.testNewInputFormat(TestNewInputOutputFormat.java:125) For reproduce this test case, run mvn test -Dtest=TestNewInputOutputFormat#testNewInputFormat -Phadoop-2 -- This message was sent by Atlassian JIRA (v6.3.4#6332)
Re: ORC separate project
This is a great plan, +1! On Thursday, March 19, 2015, Owen O'Malley omal...@apache.org wrote: All, Over the last year, there has been a fair number of projects that want to integrate with ORC, but don't want a dependence on Hive's exec jar. Additionally, we've been working on a C++ reader (and soon writer) and it would be great to host them both in the same project. Toward that end, I'd like to create a separate ORC project at Apache. There will be lots of technical details to work out, but I wanted to give the Hive community a chance to discuss it. Do any of the Hive committers want to be included on the proposal? Of the current Hive committers, my list looks like: * Alan * Gunther * Prasanth * Lefty * Owen * Sergey * Gopal * Kevin Did I miss anyone? Thanks! Owen
[jira] [Created] (HIVE-10029) LLAP: Scheduling of work from different queries within the daemon
Siddharth Seth created HIVE-10029: - Summary: LLAP: Scheduling of work from different queries within the daemon Key: HIVE-10029 URL: https://issues.apache.org/jira/browse/HIVE-10029 Project: Hive Issue Type: Sub-task Reporter: Siddharth Seth Fix For: llap The current implementation is a simple queue - whichever query wins the race to submit work to a daemon will execute first. A policy around this may be useful - potentially a fair share, or a first query in gets all slots approach. Also, prioritiy associated with work within a query should be considered. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Created] (HIVE-10027) Use descriptions from Avro schema files in column comments
Jeremy Beard created HIVE-10027: --- Summary: Use descriptions from Avro schema files in column comments Key: HIVE-10027 URL: https://issues.apache.org/jira/browse/HIVE-10027 Project: Hive Issue Type: Improvement Components: Metastore Affects Versions: 0.13.1 Reporter: Jeremy Beard Priority: Minor Avro schema files can include field descriptions using the doc tag. It would be helpful if the Hive metastore would use these descriptions as the comments for a field when the table is backed by such a schema file, instead of the default from deserializer. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Created] (HIVE-10028) LLAP: Create a fixed size execution queue for daemons
Siddharth Seth created HIVE-10028: - Summary: LLAP: Create a fixed size execution queue for daemons Key: HIVE-10028 URL: https://issues.apache.org/jira/browse/HIVE-10028 Project: Hive Issue Type: Sub-task Reporter: Siddharth Seth Fix For: llap Currently, this is unbounded. This should be a configurable size. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
Re: Review Request 32189: HIVE-9859 Create bitwise left/right shift UDFs
--- This is an automatically generated e-mail. To reply, visit: https://reviews.apache.org/r/32189/ --- (Updated March 20, 2015, 5:24 a.m.) Review request for hive and Jason Dere. Changes --- I changed bitwise shift operators to -, -, - because conflicts with nested complex type declaration, e.g. ARRAYMAPSTRING,STRING I found that HiveParser.g has the following declaration for array type. : KW_ARRAY LESSTHAN type GREATERTHAN - ^(TOK_LIST type) I think it uses LESSTHAN and GREATERTHAN in array type declaration to avoid conflicts with operators I did not find a way how to do this trick with Bugs: HIVE-9859 https://issues.apache.org/jira/browse/HIVE-9859 Repository: hive-git Description --- HIVE-9859 Create bitwise left/right shift UDFs Diffs (updated) - ql/src/java/org/apache/hadoop/hive/ql/exec/FunctionRegistry.java fdbfab9e1c4f098766f58e2d07653a44f45d3350 ql/src/java/org/apache/hadoop/hive/ql/parse/HiveLexer.g e7de6c86a3c7a674b54f3678b00f34f2dd903dc8 ql/src/java/org/apache/hadoop/hive/ql/parse/HiveParser.g d2d998972b64a19bde28cf176b3f948c00ba492a ql/src/java/org/apache/hadoop/hive/ql/parse/IdentifiersParser.g 0a05cebf1f71bd32c8023cdb10c8393a0d871cc2 ql/src/java/org/apache/hadoop/hive/ql/udf/UDFOPBitLeftShift.java PRE-CREATION ql/src/java/org/apache/hadoop/hive/ql/udf/UDFOPBitRightShift.java PRE-CREATION ql/src/java/org/apache/hadoop/hive/ql/udf/UDFOPBitUnsignedRightShift.java PRE-CREATION ql/src/test/queries/clientpositive/udf_bitwise_left_shift.q PRE-CREATION ql/src/test/queries/clientpositive/udf_bitwise_right_shift.q PRE-CREATION ql/src/test/queries/clientpositive/udf_bitwise_unsigned_right_shift.q PRE-CREATION ql/src/test/results/clientpositive/show_functions.q.out 81abeb9be4fd47724be544c7bc8da8b25fcd6e75 ql/src/test/results/clientpositive/udf_bitwise_left_shift.q.out PRE-CREATION ql/src/test/results/clientpositive/udf_bitwise_right_shift.q.out PRE-CREATION ql/src/test/results/clientpositive/udf_bitwise_unsigned_right_shift.q.out PRE-CREATION Diff: https://reviews.apache.org/r/32189/diff/ Testing --- Thanks, Alexander Pivovarov
[jira] [Created] (HIVE-10031) Modify the using of jobConf variable in ParquetRecordReaderWrapper constructor
Dong Chen created HIVE-10031: Summary: Modify the using of jobConf variable in ParquetRecordReaderWrapper constructor Key: HIVE-10031 URL: https://issues.apache.org/jira/browse/HIVE-10031 Project: Hive Issue Type: Sub-task Reporter: Dong Chen Assignee: Dong Chen In {{ParquetRecordReaderWrapper}} constructor, it create splits, set projections and filters in conf, create task context, and then create Parquet record reader. In this procedure, we could improve the logic of conf usage: 1. the clone of jobConf is not necessary. This could speed up getRecordReader a little. 2. the updated jobConf is not passed to Parquet in one case. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Created] (HIVE-10021) Alter index rebuild statements submitted through HiveServer2 fail when Sentry is enabled
Richard Williams created HIVE-10021: --- Summary: Alter index rebuild statements submitted through HiveServer2 fail when Sentry is enabled Key: HIVE-10021 URL: https://issues.apache.org/jira/browse/HIVE-10021 Project: Hive Issue Type: Bug Components: HiveServer2, Indexing Affects Versions: 0.13.1 Environment: CDH 5.3.2 Reporter: Richard Williams When HiveServer2 is configured to authorize submitted queries and statements through Sentry, any attempt to issue an alter index rebuild statement fails with a SemanticException caused by a NullPointerException. This occurs regardless of whether the index is a compact or bitmap index. The root cause of the problem appears to be the fact that the static createRootTask function in org.apache.hadoop.hive.ql.optimizer.IndexUtils creates a new org.apache.hadoop.hive.ql.Driver object to compile the index builder query, and this new Driver object, unlike the one used by HiveServer2 to compile the submitted statement, is used without having its userName field initialized with the submitting user's username. Adding null checks to the Sentry code is insufficient to solve this problem, because Sentry needs the userName to determine whether or not the submitting user should be able to execute the index rebuild statement. Example stack trace from the HiveServer2 logs: FAILED: NullPointerException null java.lang.NullPointerException at java.util.concurrent.ConcurrentHashMap.hash(ConcurrentHashMap.java:333) at java.util.concurrent.ConcurrentHashMap.get(ConcurrentHashMap.java:988) at org.apache.hadoop.security.Groups.getGroups(Groups.java:161) at org.apache.sentry.provider.common.HadoopGroupMappingService.getGroups(HadoopGroupMappingService.java:46) at org.apache.sentry.binding.hive.authz.HiveAuthzBinding.getGroups(HiveAuthzBinding.java:370) at org.apache.sentry.binding.hive.HiveAuthzBindingHook.postAnalyze(HiveAuthzBindingHook.java:314) at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:440) at org.apache.hadoop.hive.ql.optimizer.IndexUtils.createRootTask(IndexUtils.java:258) at org.apache.hadoop.hive.ql.index.compact.CompactIndexHandler.getIndexBuilderMapRedTask(CompactIndexHandler.java:149) at org.apache.hadoop.hive.ql.index.TableBasedIndexHandler.generateIndexBuildTaskList(TableBasedIndexHandler.java:67) at org.apache.hadoop.hive.ql.parse.DDLSemanticAnalyzer.getIndexBuilderMapRed(DDLSemanticAnalyzer.java:1171) at org.apache.hadoop.hive.ql.parse.DDLSemanticAnalyzer.analyzeAlterIndexRebuild(DDLSemanticAnalyzer.java:1117) at org.apache.hadoop.hive.ql.parse.DDLSemanticAnalyzer.analyzeInternal(DDLSemanticAnalyzer.java:410) at org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer.analyze(BaseSemanticAnalyzer.java:204) at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:437) at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:335) at org.apache.hadoop.hive.ql.Driver.compileInternal(Driver.java:1026) at org.apache.hadoop.hive.ql.Driver.compileAndRespond(Driver.java:1019) at org.apache.hive.service.cli.operation.SQLOperation.prepare(SQLOperation.java:100) at org.apache.hive.service.cli.operation.SQLOperation.run(SQLOperation.java:173) at org.apache.hive.service.cli.session.HiveSessionImpl.runOperationWithLogCapture(HiveSessionImpl.java:715) at org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementInternal(HiveSessionImpl.java:370) at org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementAsync(HiveSessionImpl.java:357) at org.apache.hive.service.cli.CLIService.executeStatementAsync(CLIService.java:238) at org.apache.hive.service.cli.thrift.ThriftCLIService.ExecuteStatement(ThriftCLIService.java:393) at org.apache.hive.service.cli.thrift.TCLIService$Processor$ExecuteStatement.getResult(TCLIService.java:1373) at org.apache.hive.service.cli.thrift.TCLIService$Processor$ExecuteStatement.getResult(TCLIService.java:1358) at org.apache.thrift.ProcessFunction.process(ProcessFunction.java:39) at org.apache.thrift.TBaseProcessor.process(TBaseProcessor.java:39) at org.apache.thrift.server.TServlet.doPost(TServlet.java:83) at org.apache.hive.service.cli.thrift.ThriftHttpServlet.doPost(ThriftHttpServlet.java:99) at javax.servlet.http.HttpServlet.service(HttpServlet.java:727) at javax.servlet.http.HttpServlet.service(HttpServlet.java:820) at org.eclipse.jetty.servlet.ServletHolder.handle(ServletHolder.java:565) at org.eclipse.jetty.servlet.ServletHandler.doHandle(ServletHandler.java:479) at org.eclipse.jetty.server.session.SessionHandler.doHandle(SessionHandler.java:225) at
[jira] [Created] (HIVE-10020) Add Month() and Second() buildin functions
Alicia Ying Shu created HIVE-10020: -- Summary: Add Month() and Second() buildin functions Key: HIVE-10020 URL: https://issues.apache.org/jira/browse/HIVE-10020 Project: Hive Issue Type: Bug Reporter: Alicia Ying Shu From Oracle doc: Month(date) and Second(date). Very similar to Year(date) buildin. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
Review Request 32254: HIVE-10007:Support qualified table name in analyze table compute statistics for columns
--- This is an automatically generated e-mail. To reply, visit: https://reviews.apache.org/r/32254/ --- Review request for hive, Ashutosh Chauhan, Szehon Ho, and Xuefu Zhang. Bugs: HIVE-10007 https://issues.apache.org/jira/browse/HIVE-10007 Repository: hive-git Description --- Currently Currently analyze table compute statistics for columns command can not compute column stats for a table in a different database since it does not support qualified table name. You need switch to that table database in order to compute its column stats. So does it for ALTER TABLE .. UPDATE STATISTICS FOR COLUMN command. This JIRA will provide the support to qualified table name in analyze and update column stats commands. Diffs - ql/src/java/org/apache/hadoop/hive/ql/exec/ColumnStatsTask.java e2f696eea7b472d4e417629eedd33406b41caf98 ql/src/java/org/apache/hadoop/hive/ql/exec/ColumnStatsUpdateTask.java 28c17b75356eb4ffaac2193a80b27bdd67d98009 ql/src/java/org/apache/hadoop/hive/ql/parse/ColumnStatsSemanticAnalyzer.java c83523e507f1003eff1a821c216a7f6904f293ed ql/src/java/org/apache/hadoop/hive/ql/parse/DDLSemanticAnalyzer.java b9e15a1b2f60b291759444a4407acaf1e0d9c515 ql/src/test/queries/clientpositive/alter_partition_update_status.q 1eee9a506f3854eace7deaa3f4ec02c1f7bd3479 ql/src/test/queries/clientpositive/alter_table_update_status.q fd45cd426c17a7bfa1371081dd81ff6eb1b7353d ql/src/test/queries/clientpositive/columnstats_partlvl.q 82a9e0f0fef0c9bf909028f23ccd47ee07a61a2b ql/src/test/queries/clientpositive/columnstats_tbllvl.q 07cc959ec6bb9f5e1f1ba3184da5ce3fc8bf951b ql/src/test/results/clientpositive/alter_partition_update_status.q.out 7e33a7e8ecc07b2b9d8ab8b0f12b383e5a442cf3 ql/src/test/results/clientpositive/alter_table_update_status.q.out 361359807b9c8c1a2e25a629fd53f026e02957e7 ql/src/test/results/clientpositive/columnstats_partlvl.q.out a86c5fbd718a490c5973382810cc02c10184925e ql/src/test/results/clientpositive/columnstats_tbllvl.q.out 0ef04f5f4b70d84cd6118c2407d29fc555bac116 Diff: https://reviews.apache.org/r/32254/diff/ Testing --- 1. Performed some manullay tests 2. Added new qtests for those cases 3. Precommit build Thanks, Chaoyu Tang
Re: Review Request 31386: HIVE-9555 assorted ORC refactorings for LLAP on trunk
On March 18, 2015, 8:14 a.m., Gopal V wrote: common/src/java/org/apache/hadoop/hive/common/DiskRange.java, line 78 https://reviews.apache.org/r/31386/diff/3/?file=890915#file890915line78 Bad behaviour - the original DiskRange was written with final variables for easier debugging. actually only BC had finals, not DR. I will hide the fields and make modifications restricted in scope - Sergey --- This is an automatically generated e-mail. To reply, visit: https://reviews.apache.org/r/31386/#review76879 --- On March 11, 2015, 12:50 a.m., Sergey Shelukhin wrote: --- This is an automatically generated e-mail. To reply, visit: https://reviews.apache.org/r/31386/ --- (Updated March 11, 2015, 12:50 a.m.) Review request for hive and Prasanth_J. Repository: hive-git Description --- see jira Diffs - common/src/java/org/apache/hadoop/hive/common/DiskRange.java PRE-CREATION common/src/java/org/apache/hadoop/hive/common/DiskRangeList.java PRE-CREATION ql/src/java/org/apache/hadoop/hive/ql/io/orc/CompressionCodec.java 5e2d880 ql/src/java/org/apache/hadoop/hive/ql/io/orc/FileDump.java 9788c16 ql/src/java/org/apache/hadoop/hive/ql/io/orc/InStream.java 62c6f8d ql/src/java/org/apache/hadoop/hive/ql/io/orc/MetadataReader.java PRE-CREATION ql/src/java/org/apache/hadoop/hive/ql/io/orc/OrcFile.java 25bb15a ql/src/java/org/apache/hadoop/hive/ql/io/orc/OrcInputFormat.java 498ee14 ql/src/java/org/apache/hadoop/hive/ql/io/orc/PositionProvider.java 3daa9ba ql/src/java/org/apache/hadoop/hive/ql/io/orc/Reader.java f85c21b ql/src/java/org/apache/hadoop/hive/ql/io/orc/ReaderImpl.java 03f8085 ql/src/java/org/apache/hadoop/hive/ql/io/orc/RecordReaderImpl.java 458ad21 ql/src/java/org/apache/hadoop/hive/ql/io/orc/RecordReaderUtils.java PRE-CREATION ql/src/java/org/apache/hadoop/hive/ql/io/orc/RunLengthIntegerReaderV2.java 4057036 ql/src/java/org/apache/hadoop/hive/ql/io/orc/VectorizedOrcInputFormat.java 23e5f27 ql/src/java/org/apache/hadoop/hive/ql/io/orc/WriterImpl.java 79dc5a1 ql/src/java/org/apache/hadoop/hive/ql/io/sarg/SearchArgumentFactory.java f4a2e65 ql/src/test/org/apache/hadoop/hive/ql/io/orc/TestInStream.java 0ea4a7b ql/src/test/org/apache/hadoop/hive/ql/io/orc/TestInputOutputFormat.java 2cc3d7a ql/src/test/org/apache/hadoop/hive/ql/io/orc/TestIntegerCompressionReader.java 591ec3f ql/src/test/org/apache/hadoop/hive/ql/io/orc/TestOrcFile.java cd1d645 ql/src/test/org/apache/hadoop/hive/ql/io/orc/TestRecordReaderImpl.java 326dde4 Diff: https://reviews.apache.org/r/31386/diff/ Testing --- Thanks, Sergey Shelukhin
Re: Review Request 31386: HIVE-9555 assorted ORC refactorings for LLAP on trunk
--- This is an automatically generated e-mail. To reply, visit: https://reviews.apache.org/r/31386/ --- (Updated March 19, 2015, 9:12 p.m.) Review request for hive and Prasanth_J. Repository: hive-git Description --- see jira Diffs (updated) - b/common/src/java/org/apache/hadoop/hive/common/DiskRangeList.java PRE-CREATION common/src/java/org/apache/hadoop/hive/common/DiskRange.java PRE-CREATION common/src/java/org/apache/hadoop/hive/common/DiskRangeList.java PRE-CREATION ql/src/java/org/apache/hadoop/hive/ql/io/orc/CompressionCodec.java 5e2d880 ql/src/java/org/apache/hadoop/hive/ql/io/orc/FileDump.java 9788c16 ql/src/java/org/apache/hadoop/hive/ql/io/orc/InStream.java 62c6f8d ql/src/java/org/apache/hadoop/hive/ql/io/orc/MetadataReader.java PRE-CREATION ql/src/java/org/apache/hadoop/hive/ql/io/orc/OrcFile.java 25bb15a ql/src/java/org/apache/hadoop/hive/ql/io/orc/OrcInputFormat.java 498ee14 ql/src/java/org/apache/hadoop/hive/ql/io/orc/PositionProvider.java 3daa9ba ql/src/java/org/apache/hadoop/hive/ql/io/orc/Reader.java f85c21b ql/src/java/org/apache/hadoop/hive/ql/io/orc/ReaderImpl.java 03f8085 ql/src/java/org/apache/hadoop/hive/ql/io/orc/RecordReaderImpl.java 458ad21 ql/src/java/org/apache/hadoop/hive/ql/io/orc/RecordReaderUtils.java PRE-CREATION ql/src/java/org/apache/hadoop/hive/ql/io/orc/RunLengthIntegerReaderV2.java 4057036 ql/src/java/org/apache/hadoop/hive/ql/io/orc/VectorizedOrcInputFormat.java 23e5f27 ql/src/java/org/apache/hadoop/hive/ql/io/orc/WriterImpl.java 79dc5a1 ql/src/java/org/apache/hadoop/hive/ql/io/sarg/SearchArgumentFactory.java f4a2e65 ql/src/test/org/apache/hadoop/hive/ql/io/orc/TestInStream.java 0ea4a7b ql/src/test/org/apache/hadoop/hive/ql/io/orc/TestInputOutputFormat.java 2cc3d7a ql/src/test/org/apache/hadoop/hive/ql/io/orc/TestIntegerCompressionReader.java 591ec3f ql/src/test/org/apache/hadoop/hive/ql/io/orc/TestOrcFile.java cd1d645 ql/src/test/org/apache/hadoop/hive/ql/io/orc/TestRecordReaderImpl.java 326dde4 Diff: https://reviews.apache.org/r/31386/diff/ Testing --- Thanks, Sergey Shelukhin
[jira] [Created] (HIVE-10023) Fix more cache related concurrency issue
Jimmy Xiang created HIVE-10023: -- Summary: Fix more cache related concurrency issue Key: HIVE-10023 URL: https://issues.apache.org/jira/browse/HIVE-10023 Project: Hive Issue Type: Bug Reporter: Jimmy Xiang Assignee: Jimmy Xiang Searched the code and found couple more issues with cache, such as LazyBinaryObjectInspectorFactory, PrimitiveObjectInspectorFactory, and TypeInfoFactory. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Created] (HIVE-10022) DFS in authorization might take too long
Pankit Thapar created HIVE-10022: Summary: DFS in authorization might take too long Key: HIVE-10022 URL: https://issues.apache.org/jira/browse/HIVE-10022 Project: Hive Issue Type: Bug Components: Authorization Affects Versions: 0.14.0 Reporter: Pankit Thapar I am testing a query like : set hive.test.authz.sstd.hs2.mode=true; set hive.security.authorization.manager=org.apache.hadoop.hive.ql.security.authorization.plugin.sqlstd.SQLStdHiveAuthorizerFactoryForTest; set hive.security.authenticator.manager=org.apache.hadoop.hive.ql.security.SessionStateConfigUserAuthenticator; set hive.security.authorization.enabled=true; set user.name=user1; create table auth_noupd(i int) clustered by (i) into 2 buckets stored as orc location '${OUTPUT}' TBLPROPERTIES ('transactional'='true'); Now, in the above query, since authorization is true, we would end up calling doAuthorizationV2() which ultimately ends up calling SQLAuthorizationUtils.getPrivilegesFromFS() which calls a recursive method : FileUtils.isActionPermittedForFileHierarchy() with the object or the ancestor of the object we are trying to authorize if the object does not exist. The logic in FileUtils.isActionPermittedForFileHierarchy() is DFS. Now assume, we have a path as a/b/c/d that we are trying to authorize. In case, a/b/c/d does not exist, we would call FileUtils.isActionPermittedForFileHierarchy() with say a/b/ assuming a/b/c also does not exist. If under the subtree at a/b, we have millions of files, then FileUtils.isActionPermittedForFileHierarchy() is going to check file permission on each of those objects. I do not completely understand why do we have to check for file permissions in all the objects in branch of the tree that we are not trying to read from /write to. We could have checked file permission on the ancestor that exists and if it matches what we expect, the return true. Please confirm if this is a bug so that I can submit a patch else let me know what I am missing ? -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Created] (HIVE-10024) LLAP: q file test is broken again
Sergey Shelukhin created HIVE-10024: --- Summary: LLAP: q file test is broken again Key: HIVE-10024 URL: https://issues.apache.org/jira/browse/HIVE-10024 Project: Hive Issue Type: Sub-task Reporter: Sergey Shelukhin Assignee: Sergey Shelukhin -- This message was sent by Atlassian JIRA (v6.3.4#6332)
ORC separate project
All, Over the last year, there has been a fair number of projects that want to integrate with ORC, but don't want a dependence on Hive's exec jar. Additionally, we've been working on a C++ reader (and soon writer) and it would be great to host them both in the same project. Toward that end, I'd like to create a separate ORC project at Apache. There will be lots of technical details to work out, but I wanted to give the Hive community a chance to discuss it. Do any of the Hive committers want to be included on the proposal? Of the current Hive committers, my list looks like: * Alan * Gunther * Prasanth * Lefty * Owen * Sergey * Gopal * Kevin Did I miss anyone? Thanks! Owen
Re: Reading 2 table data in MapReduce for Performing Join
Hi All, I was successfully able to integrate HCatMultipleInputs with the patch for the tables created with TEXTFILE. But I get error when I read table created with ORC file. The error is below : 15/03/19 10:51:32 INFO mapreduce.Job: Task Id : attempt_1425012118520_9756_m_00_0, Status : FAILED Error: java.lang.ClassCastException: org.apache.hadoop.io.NullWritable cannot be cast to org.apache.hadoop.io.LongWritable at com.abccompany.mapreduce.MyMapper.map(MyMapper.java:15) at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:145) at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:764) at org.apache.hadoop.mapred.MapTask.run(MapTask.java:340) at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:167) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:415) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1557) at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:162) Can anyone help? Thanks in advance! On Wed, Mar 18, 2015 at 11:00 PM, Suraj Nayak snay...@gmail.com wrote: Hi All, https://issues.apache.org/jira/browse/HIVE-4997 patch helped! On Tue, Mar 17, 2015 at 1:05 AM, Suraj Nayak snay...@gmail.com wrote: Hi, I tried reading data via HCatalog for 1 Hive table in MapReduce using something similar to https://cwiki.apache.org/confluence/display/Hive/HCatalog+InputOutput#HCatalogInputOutput-RunningMapReducewithHCatalog. I was able to read successfully. Now am trying to read 2 tables, as the requirement is to join 2 tables. I did not find API similar to *FileInputFormat.addInputPaths* in *HCatInputFormat*. What is the equivalent of the same in HCat ? I had performed join using FilesInputFormat in HDFS(by getting split information in mapper). This article( http://www.codingjunkie.com/mapreduce-reduce-joins) helped me code join. http://www.codingjunkie.com/mapreduce-reduce-joins/ Can someone suggest how I can perform join operation using HCatalog ? Briefly, the aim is to - Read 2 tables (almost similar schema) - If key exists in both the table send it to same reducer. - Do some processing on the records in reducer. - Save the output into file/Hive table. *P.S : The reason for using MapReduce to perform join is because of complex requirement which can't be solved via Hive/Pig directly. * Any help will be greatly appreciated :) -- Thanks Suraj Nayak M -- Thanks Suraj Nayak M -- Thanks Suraj Nayak M
Re: Reading 2 table data in MapReduce for Performing Join
Is this related to https://issues.apache.org/jira/browse/HIVE-4329 ? Is there a workaround? On Thu, Mar 19, 2015 at 9:47 PM, Suraj Nayak snay...@gmail.com wrote: Hi All, I was successfully able to integrate HCatMultipleInputs with the patch for the tables created with TEXTFILE. But I get error when I read table created with ORC file. The error is below : 15/03/19 10:51:32 INFO mapreduce.Job: Task Id : attempt_1425012118520_9756_m_00_0, Status : FAILED Error: java.lang.ClassCastException: org.apache.hadoop.io.NullWritable cannot be cast to org.apache.hadoop.io.LongWritable at com.abccompany.mapreduce.MyMapper.map(MyMapper.java:15) at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:145) at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:764) at org.apache.hadoop.mapred.MapTask.run(MapTask.java:340) at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:167) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:415) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1557) at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:162) Can anyone help? Thanks in advance! On Wed, Mar 18, 2015 at 11:00 PM, Suraj Nayak snay...@gmail.com wrote: Hi All, https://issues.apache.org/jira/browse/HIVE-4997 patch helped! On Tue, Mar 17, 2015 at 1:05 AM, Suraj Nayak snay...@gmail.com wrote: Hi, I tried reading data via HCatalog for 1 Hive table in MapReduce using something similar to https://cwiki.apache.org/confluence/display/Hive/HCatalog+InputOutput#HCatalogInputOutput-RunningMapReducewithHCatalog. I was able to read successfully. Now am trying to read 2 tables, as the requirement is to join 2 tables. I did not find API similar to *FileInputFormat.addInputPaths* in *HCatInputFormat*. What is the equivalent of the same in HCat ? I had performed join using FilesInputFormat in HDFS(by getting split information in mapper). This article( http://www.codingjunkie.com/mapreduce-reduce-joins) helped me code join. http://www.codingjunkie.com/mapreduce-reduce-joins/ Can someone suggest how I can perform join operation using HCatalog ? Briefly, the aim is to - Read 2 tables (almost similar schema) - If key exists in both the table send it to same reducer. - Do some processing on the records in reducer. - Save the output into file/Hive table. *P.S : The reason for using MapReduce to perform join is because of complex requirement which can't be solved via Hive/Pig directly. * Any help will be greatly appreciated :) -- Thanks Suraj Nayak M -- Thanks Suraj Nayak M -- Thanks Suraj Nayak M -- Thanks Suraj Nayak M
[jira] [Created] (HIVE-10018) Activating SQLStandardAuth results in NPE [hbase-metastore branch]
Alan Gates created HIVE-10018: - Summary: Activating SQLStandardAuth results in NPE [hbase-metastore branch] Key: HIVE-10018 URL: https://issues.apache.org/jira/browse/HIVE-10018 Project: Hive Issue Type: Bug Components: Metastore Affects Versions: hbase-metastore-branch Reporter: Alan Gates Assignee: Alan Gates Setting the config to run SQLStandardAuth and then doing even simple SQL statements results in an NPE. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Created] (HIVE-10019) Configure jenkins precommit tests to run HMS upgrade tests
Sergio Peña created HIVE-10019: -- Summary: Configure jenkins precommit tests to run HMS upgrade tests Key: HIVE-10019 URL: https://issues.apache.org/jira/browse/HIVE-10019 Project: Hive Issue Type: Task Reporter: Sergio Peña Assignee: Sergio Peña This task is created to configure all jenkins precommit jobs with different branches to run HMS upgrade tests if there are changes on the metastore upgrade script. These tests are already created, so this task is only for final jenkins configuration. -- This message was sent by Atlassian JIRA (v6.3.4#6332)