[jira] [Updated] (HIVE-20476) CopyUtils used by REPL LOAD and EXPORT/IMPORT operations ignore distcp error.
[ https://issues.apache.org/jira/browse/HIVE-20476?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Sankar Hariappan updated HIVE-20476: Status: Patch Available (was: Open) > CopyUtils used by REPL LOAD and EXPORT/IMPORT operations ignore distcp error. > - > > Key: HIVE-20476 > URL: https://issues.apache.org/jira/browse/HIVE-20476 > Project: Hive > Issue Type: Bug > Components: HiveServer2, repl >Affects Versions: 3.1.0 >Reporter: Sankar Hariappan >Assignee: Sankar Hariappan >Priority: Major > Labels: DR, Export, Import, Replication, pull-request-available > Attachments: HIVE-20476.01.patch > > > CopyUtils uses FileUtils.distCp to copy files but doesn't check the return > value. It returns false if the copy fails. > Now, REPL LOAD and EXPORT/IMPORT commands internally uses CopyUtils to copy > data files across clusters and here it may return success even if file copy > fails and may cause data loss. > Need to throw error and retry. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (HIVE-20476) CopyUtils used by REPL LOAD and EXPORT/IMPORT operations ignore distcp error.
[ https://issues.apache.org/jira/browse/HIVE-20476?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Sankar Hariappan updated HIVE-20476: Attachment: HIVE-20476.01.patch > CopyUtils used by REPL LOAD and EXPORT/IMPORT operations ignore distcp error. > - > > Key: HIVE-20476 > URL: https://issues.apache.org/jira/browse/HIVE-20476 > Project: Hive > Issue Type: Bug > Components: HiveServer2, repl >Affects Versions: 3.1.0 >Reporter: Sankar Hariappan >Assignee: Sankar Hariappan >Priority: Major > Labels: DR, Export, Import, Replication, pull-request-available > Attachments: HIVE-20476.01.patch > > > CopyUtils uses FileUtils.distCp to copy files but doesn't check the return > value. It returns false if the copy fails. > Now, REPL LOAD and EXPORT/IMPORT commands internally uses CopyUtils to copy > data files across clusters and here it may return success even if file copy > fails and may cause data loss. > Need to throw error and retry. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (HIVE-20476) CopyUtils used by REPL LOAD and EXPORT/IMPORT operations ignore distcp error.
[ https://issues.apache.org/jira/browse/HIVE-20476?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Sankar Hariappan updated HIVE-20476: Attachment: (was: HIVE-20476.01.patch) > CopyUtils used by REPL LOAD and EXPORT/IMPORT operations ignore distcp error. > - > > Key: HIVE-20476 > URL: https://issues.apache.org/jira/browse/HIVE-20476 > Project: Hive > Issue Type: Bug > Components: HiveServer2, repl >Affects Versions: 3.1.0 >Reporter: Sankar Hariappan >Assignee: Sankar Hariappan >Priority: Major > Labels: DR, Export, Import, Replication, pull-request-available > > CopyUtils uses FileUtils.distCp to copy files but doesn't check the return > value. It returns false if the copy fails. > Now, REPL LOAD and EXPORT/IMPORT commands internally uses CopyUtils to copy > data files across clusters and here it may return success even if file copy > fails and may cause data loss. > Need to throw error and retry. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (HIVE-20476) CopyUtils used by REPL LOAD and EXPORT/IMPORT operations ignore distcp error.
[ https://issues.apache.org/jira/browse/HIVE-20476?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Sankar Hariappan updated HIVE-20476: Status: Open (was: Patch Available) > CopyUtils used by REPL LOAD and EXPORT/IMPORT operations ignore distcp error. > - > > Key: HIVE-20476 > URL: https://issues.apache.org/jira/browse/HIVE-20476 > Project: Hive > Issue Type: Bug > Components: HiveServer2, repl >Affects Versions: 3.1.0 >Reporter: Sankar Hariappan >Assignee: Sankar Hariappan >Priority: Major > Labels: DR, Export, Import, Replication, pull-request-available > > CopyUtils uses FileUtils.distCp to copy files but doesn't check the return > value. It returns false if the copy fails. > Now, REPL LOAD and EXPORT/IMPORT commands internally uses CopyUtils to copy > data files across clusters and here it may return success even if file copy > fails and may cause data loss. > Need to throw error and retry. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (HIVE-20476) CopyUtils used by REPL LOAD and EXPORT/IMPORT operations ignore distcp error.
[ https://issues.apache.org/jira/browse/HIVE-20476?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Sankar Hariappan updated HIVE-20476: Attachment: HIVE-20476.01.patch > CopyUtils used by REPL LOAD and EXPORT/IMPORT operations ignore distcp error. > - > > Key: HIVE-20476 > URL: https://issues.apache.org/jira/browse/HIVE-20476 > Project: Hive > Issue Type: Bug > Components: HiveServer2, repl >Affects Versions: 3.1.0 >Reporter: Sankar Hariappan >Assignee: Sankar Hariappan >Priority: Major > Labels: DR, Export, Import, Replication, pull-request-available > Attachments: HIVE-20476.01.patch > > > CopyUtils uses FileUtils.distCp to copy files but doesn't check the return > value. It returns false if the copy fails. > Now, REPL LOAD and EXPORT/IMPORT commands internally uses CopyUtils to copy > data files across clusters and here it may return success even if file copy > fails and may cause data loss. > Need to throw error and retry. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (HIVE-20476) CopyUtils used by REPL LOAD and EXPORT/IMPORT operations ignore distcp error.
[ https://issues.apache.org/jira/browse/HIVE-20476?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Sankar Hariappan updated HIVE-20476: Attachment: (was: HIVE-20476.01.patch) > CopyUtils used by REPL LOAD and EXPORT/IMPORT operations ignore distcp error. > - > > Key: HIVE-20476 > URL: https://issues.apache.org/jira/browse/HIVE-20476 > Project: Hive > Issue Type: Bug > Components: HiveServer2, repl >Affects Versions: 3.1.0 >Reporter: Sankar Hariappan >Assignee: Sankar Hariappan >Priority: Major > Labels: DR, Export, Import, Replication, pull-request-available > Attachments: HIVE-20476.01.patch > > > CopyUtils uses FileUtils.distCp to copy files but doesn't check the return > value. It returns false if the copy fails. > Now, REPL LOAD and EXPORT/IMPORT commands internally uses CopyUtils to copy > data files across clusters and here it may return success even if file copy > fails and may cause data loss. > Need to throw error and retry. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (HIVE-20476) CopyUtils used by REPL LOAD and EXPORT/IMPORT operations ignore distcp error.
[ https://issues.apache.org/jira/browse/HIVE-20476?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Sankar Hariappan updated HIVE-20476: Status: Patch Available (was: Open) Flaky test failure. Re-attaching same patch. > CopyUtils used by REPL LOAD and EXPORT/IMPORT operations ignore distcp error. > - > > Key: HIVE-20476 > URL: https://issues.apache.org/jira/browse/HIVE-20476 > Project: Hive > Issue Type: Bug > Components: HiveServer2, repl >Affects Versions: 3.1.0 >Reporter: Sankar Hariappan >Assignee: Sankar Hariappan >Priority: Major > Labels: DR, Export, Import, Replication, pull-request-available > Attachments: HIVE-20476.01.patch > > > CopyUtils uses FileUtils.distCp to copy files but doesn't check the return > value. It returns false if the copy fails. > Now, REPL LOAD and EXPORT/IMPORT commands internally uses CopyUtils to copy > data files across clusters and here it may return success even if file copy > fails and may cause data loss. > Need to throw error and retry. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (HIVE-20476) CopyUtils used by REPL LOAD and EXPORT/IMPORT operations ignore distcp error.
[ https://issues.apache.org/jira/browse/HIVE-20476?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Sankar Hariappan updated HIVE-20476: Status: Open (was: Patch Available) > CopyUtils used by REPL LOAD and EXPORT/IMPORT operations ignore distcp error. > - > > Key: HIVE-20476 > URL: https://issues.apache.org/jira/browse/HIVE-20476 > Project: Hive > Issue Type: Bug > Components: HiveServer2, repl >Affects Versions: 3.1.0 >Reporter: Sankar Hariappan >Assignee: Sankar Hariappan >Priority: Major > Labels: DR, Export, Import, Replication, pull-request-available > Attachments: HIVE-20476.01.patch > > > CopyUtils uses FileUtils.distCp to copy files but doesn't check the return > value. It returns false if the copy fails. > Now, REPL LOAD and EXPORT/IMPORT commands internally uses CopyUtils to copy > data files across clusters and here it may return success even if file copy > fails and may cause data loss. > Need to throw error and retry. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (HIVE-20405) Run all druid tests in one batch
[ https://issues.apache.org/jira/browse/HIVE-20405?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16595859#comment-16595859 ] slim bouguerra commented on HIVE-20405: --- I can see from the logs that the default batch size for miniDruidTests is 5 (grep for qFileTest.miniDruid.batchSize) at http://104.198.109.242/logs/PreCommit-HIVE-Build-13517/execution.txt This leads to running multiple tests setups for Druid tests, that causes issue if each batch is trying to startup a Druid Cluster. I think a short fix can be to increase the batch size to something like 50, this will ensure that only one Druid Cluster is used. {code} QFileTestBatch [batchId=192, size=5, driver=TestMiniDruidCliDriver, queryFilesProperty=qfile, name=192-TestMiniDruidCliDriver-druidmini_test_ts.q-druidmini_mv.q-druid_timestamptz2.q-and-2-more, tests=[druidmini_test_ts.q, druidmini_mv.q, druid_timestamptz2.q, druidmini_extractTime.q, druidmini_floorTime.q], isParallel=true, moduleName=itests/qtest] 2018-08-29 00:08:31,293 DEBUG [TestExecutor] ExecutionPhase.execute:98 PBatch: QFileTestBatch [batchId=193, size=5, driver=TestMiniDruidCliDriver, queryFilesProperty=qfile, name=193-TestMiniDruidCliDriver-druidmini_dynamic_partition.q-druidmini_expressions.q-druidmini_test_alter.q-and-2-more, tests=[druidmini_dynamic_partition.q, druidmini_expressions.q, druidmini_test_alter.q, druidmini_test1.q, druidmini_test_insert.q], isParallel=true, moduleName=itests/qtest] 2018-08-29 00:08:31,293 DEBUG [TestExecutor] ExecutionPhase.execute:98 PBatch: QFileTestBatch [batchId=194, size=3, driver=TestMiniDruidCliDriver, queryFilesProperty=qfile, name=194-TestMiniDruidCliDriver-druidmini_masking.q-druidmini_joins.q-druid_timestamptz.q, tests=[druidmini_masking.q, druidmini_joins.q, druid_timestamptz.q], isParallel=true, moduleName=itests/qtest] {code} > Run all druid tests in one batch > > > Key: HIVE-20405 > URL: https://issues.apache.org/jira/browse/HIVE-20405 > Project: Hive > Issue Type: Test > Components: Testing Infrastructure >Reporter: Vineet Garg >Assignee: Nishant Bangarwa >Priority: Major > > Running druid tests in parallel could cause issues so all of the tests should > be run in one batch. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (HIVE-20377) Hive Kafka Storage Handler
[ https://issues.apache.org/jira/browse/HIVE-20377?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] slim bouguerra updated HIVE-20377: -- Attachment: HIVE-20377.18.patch > Hive Kafka Storage Handler > -- > > Key: HIVE-20377 > URL: https://issues.apache.org/jira/browse/HIVE-20377 > Project: Hive > Issue Type: New Feature >Affects Versions: 4.0.0 >Reporter: slim bouguerra >Assignee: slim bouguerra >Priority: Major > Attachments: HIVE-20377.10.patch, HIVE-20377.11.patch, > HIVE-20377.12.patch, HIVE-20377.15.patch, HIVE-20377.18.patch, > HIVE-20377.18.patch, HIVE-20377.4.patch, HIVE-20377.5.patch, > HIVE-20377.6.patch, HIVE-20377.8.patch, HIVE-20377.8.patch, HIVE-20377.patch > > > h1. Goal > * Read streaming data form Kafka queue as an external table. > * Allow streaming navigation by pushing down filters on Kafka record > partition id, offset and timestamp. > * Insert streaming data form Kafka to an actual Hive internal table, using > CTAS statement. > h1. Example > h2. Create the external table > {code} > CREATE EXTERNAL TABLE kafka_table (`timestamp` timestamp, page string, `user` > string, language string, added int, deleted int, flags string,comment string, > namespace string) > STORED BY 'org.apache.hadoop.hive.kafka.KafkaStorageHandler' > TBLPROPERTIES > ("kafka.topic" = "wikipedia", > "kafka.bootstrap.servers"="brokeraddress:9092", > "kafka.serde.class"="org.apache.hadoop.hive.serde2.JsonSerDe"); > {code} > h2. Kafka Metadata > In order to keep track of Kafka records the storage handler will add > automatically the Kafka row metadata eg partition id, record offset and > record timestamp. > {code} > DESCRIBE EXTENDED kafka_table > timestamp timestamp from deserializer > page string from deserializer > user string from deserializer > language string from deserializer > country string from deserializer > continent string from deserializer > namespace string from deserializer > newpage boolean from deserializer > unpatrolled boolean from deserializer > anonymous boolean from deserializer > robot boolean from deserializer > added int from deserializer > deleted int from deserializer > delta bigint from deserializer > __partition int from deserializer > __offset bigint from deserializer > __timestamp bigint from deserializer > {code} > h2. Filter push down. > Newer Kafka consumers 0.11.0 and higher allow seeking on the stream based on > a given offset. The proposed storage handler will be able to leverage such > API by pushing down filters over metadata columns, namely __partition (int), > __offset(long) and __timestamp(long) > For instance Query like > {code} > select `__offset` from kafka_table where (`__offset` < 10 and `__offset`>3 > and `__partition` = 0) or (`__partition` = 0 and `__offset` < 105 and > `__offset` > 99) or (`__offset` = 109); > {code} > Will result on a scan of partition 0 only then read only records between > offset 4 and 109. > h2. With timestamp seeks > The seeking based on the internal timestamps allows the handler to run on > recently arrived data, by doing > {code} > select count(*) from kafka_table where `__timestamp` > 1000 * > to_unix_timestamp(CURRENT_TIMESTAMP - interval '20' hours) ; > {code} > This allows for implicit relationships between event timestamps and kafka > timestamps to be expressed in queries (i.e event_timestamp is always < than > kafka __timestamp and kafka __timestamp is never > 15 minutes from event etc). > h2. More examples with Avro > {code} > CREATE EXTERNAL TABLE wiki_kafka_avro_table > STORED BY 'org.apache.hadoop.hive.kafka.KafkaStorageHandler' > TBLPROPERTIES > ("kafka.topic" = "wiki_kafka_avro_table", > "kafka.bootstrap.servers"="localhost:9092", > "kafka.serde.class"="org.apache.hadoop.hive.serde2.avro.AvroSerDe", > 'avro.schema.literal'='{ > "type" : "record", > "name" : "Wikipedia", > "namespace" : "org.apache.hive.kafka", > "version": "1", > "fields" : [ { > "name" : "isrobot", > "type" : "boolean" > }, { > "name" : "channel", > "type" : "string" > }, { > "name" : "timestamp", > "type" : "string" > }, { > "name" : "flags", > "type" : "string" > }, { > "name" : "isunpatrolled", > "type" : "boolean" > },
[jira] [Commented] (HIVE-18767) Some alterPartitions invocations throw 'NumberFormatException: null'
[ https://issues.apache.org/jira/browse/HIVE-18767?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16595841#comment-16595841 ] Mass Dosage commented on HIVE-18767: No problem, OK, here we go one by one with the first one up being HIVE-18767.2-branch-2.3.patch. > Some alterPartitions invocations throw 'NumberFormatException: null' > > > Key: HIVE-18767 > URL: https://issues.apache.org/jira/browse/HIVE-18767 > Project: Hive > Issue Type: Bug > Components: Metastore >Affects Versions: 2.3.3, 3.1.0, 4.0.0, 3.2.0 >Reporter: Yuming Wang >Assignee: Mass Dosage >Priority: Major > Fix For: 2.3.3, 4.0.0 > > Attachments: HIVE-18767-branch-2.3.patch, HIVE-18767-branch-2.patch, > HIVE-18767-branch-3.1.patch, HIVE-18767-branch-3.patch, HIVE-18767.1.patch, > HIVE-18767.2-branch-2.3.patch, HIVE-18767.2.patch, HIVE-18767.3.patch, > HIVE-18767.4.patch, HIVE-18767.5.patch, HIVE-18767.6.patch > > > Error messages: > {noformat} > [info] Cause: java.lang.NumberFormatException: null > [info] at java.lang.Long.parseLong(Long.java:552) > [info] at java.lang.Long.parseLong(Long.java:631) > [info] at > org.apache.hadoop.hive.metastore.MetaStoreUtils.isFastStatsSame(MetaStoreUtils.java:315) > [info] at > org.apache.hadoop.hive.metastore.HiveAlterHandler.alterPartitions(HiveAlterHandler.java:605) > [info] at > org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.alter_partitions_with_environment_context(HiveMetaStore.java:3837) > [info] at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > [info] at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) > [info] at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > [info] at java.lang.reflect.Method.invoke(Method.java:498) > [info] at > org.apache.hadoop.hive.metastore.RetryingHMSHandler.invokeInternal(RetryingHMSHandler.java:148) > [info] at > org.apache.hadoop.hive.metastore.RetryingHMSHandler.invoke(RetryingHMSHandler.java:107) > [info] at > com.sun.proxy.$Proxy23.alter_partitions_with_environment_context(Unknown > Source) > [info] at > org.apache.hadoop.hive.metastore.HiveMetaStoreClient.alter_partitions(HiveMetaStoreClient.java:1527) > {noformat} -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (HIVE-18767) Some alterPartitions invocations throw 'NumberFormatException: null'
[ https://issues.apache.org/jira/browse/HIVE-18767?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Mass Dosage updated HIVE-18767: --- Attachment: HIVE-18767.2-branch-2.3.patch > Some alterPartitions invocations throw 'NumberFormatException: null' > > > Key: HIVE-18767 > URL: https://issues.apache.org/jira/browse/HIVE-18767 > Project: Hive > Issue Type: Bug > Components: Metastore >Affects Versions: 2.3.3, 3.1.0, 4.0.0, 3.2.0 >Reporter: Yuming Wang >Assignee: Mass Dosage >Priority: Major > Fix For: 2.3.3, 4.0.0 > > Attachments: HIVE-18767-branch-2.3.patch, HIVE-18767-branch-2.patch, > HIVE-18767-branch-3.1.patch, HIVE-18767-branch-3.patch, HIVE-18767.1.patch, > HIVE-18767.2-branch-2.3.patch, HIVE-18767.2.patch, HIVE-18767.3.patch, > HIVE-18767.4.patch, HIVE-18767.5.patch, HIVE-18767.6.patch > > > Error messages: > {noformat} > [info] Cause: java.lang.NumberFormatException: null > [info] at java.lang.Long.parseLong(Long.java:552) > [info] at java.lang.Long.parseLong(Long.java:631) > [info] at > org.apache.hadoop.hive.metastore.MetaStoreUtils.isFastStatsSame(MetaStoreUtils.java:315) > [info] at > org.apache.hadoop.hive.metastore.HiveAlterHandler.alterPartitions(HiveAlterHandler.java:605) > [info] at > org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.alter_partitions_with_environment_context(HiveMetaStore.java:3837) > [info] at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > [info] at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) > [info] at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > [info] at java.lang.reflect.Method.invoke(Method.java:498) > [info] at > org.apache.hadoop.hive.metastore.RetryingHMSHandler.invokeInternal(RetryingHMSHandler.java:148) > [info] at > org.apache.hadoop.hive.metastore.RetryingHMSHandler.invoke(RetryingHMSHandler.java:107) > [info] at > com.sun.proxy.$Proxy23.alter_partitions_with_environment_context(Unknown > Source) > [info] at > org.apache.hadoop.hive.metastore.HiveMetaStoreClient.alter_partitions(HiveMetaStoreClient.java:1527) > {noformat} -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (HIVE-18767) Some alterPartitions invocations throw 'NumberFormatException: null'
[ https://issues.apache.org/jira/browse/HIVE-18767?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Mass Dosage updated HIVE-18767: --- Target Version/s: 3.1.0, 2.3.3, 4.0.0, 3.2.0 (was: 2.3.3, 3.1.0, 4.0.0, 3.2.0) Status: In Progress (was: Patch Available) > Some alterPartitions invocations throw 'NumberFormatException: null' > > > Key: HIVE-18767 > URL: https://issues.apache.org/jira/browse/HIVE-18767 > Project: Hive > Issue Type: Bug > Components: Metastore >Affects Versions: 3.1.0, 2.3.3, 4.0.0, 3.2.0 >Reporter: Yuming Wang >Assignee: Mass Dosage >Priority: Major > Fix For: 2.3.3, 4.0.0 > > Attachments: HIVE-18767-branch-2.3.patch, HIVE-18767-branch-2.patch, > HIVE-18767-branch-3.1.patch, HIVE-18767-branch-3.patch, HIVE-18767.1.patch, > HIVE-18767.2-branch-2.3.patch, HIVE-18767.2.patch, HIVE-18767.3.patch, > HIVE-18767.4.patch, HIVE-18767.5.patch, HIVE-18767.6.patch > > > Error messages: > {noformat} > [info] Cause: java.lang.NumberFormatException: null > [info] at java.lang.Long.parseLong(Long.java:552) > [info] at java.lang.Long.parseLong(Long.java:631) > [info] at > org.apache.hadoop.hive.metastore.MetaStoreUtils.isFastStatsSame(MetaStoreUtils.java:315) > [info] at > org.apache.hadoop.hive.metastore.HiveAlterHandler.alterPartitions(HiveAlterHandler.java:605) > [info] at > org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.alter_partitions_with_environment_context(HiveMetaStore.java:3837) > [info] at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > [info] at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) > [info] at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > [info] at java.lang.reflect.Method.invoke(Method.java:498) > [info] at > org.apache.hadoop.hive.metastore.RetryingHMSHandler.invokeInternal(RetryingHMSHandler.java:148) > [info] at > org.apache.hadoop.hive.metastore.RetryingHMSHandler.invoke(RetryingHMSHandler.java:107) > [info] at > com.sun.proxy.$Proxy23.alter_partitions_with_environment_context(Unknown > Source) > [info] at > org.apache.hadoop.hive.metastore.HiveMetaStoreClient.alter_partitions(HiveMetaStoreClient.java:1527) > {noformat} -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (HIVE-20363) Use integer constants for frequently used serde classes.
[ https://issues.apache.org/jira/browse/HIVE-20363?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Anurag Mantripragada updated HIVE-20363: Attachment: (was: HIVE-20363.1.patch) > Use integer constants for frequently used serde classes. > > > Key: HIVE-20363 > URL: https://issues.apache.org/jira/browse/HIVE-20363 > Project: Hive > Issue Type: Improvement > Components: Metastore, Standalone Metastore >Reporter: Anurag Mantripragada >Assignee: Anurag Mantripragada >Priority: Major > Attachments: HIVE-20363.1.patch > > > Serde libraries are stored as fully qualified class names which are long > strings, we can get improvements in I/O and storage if we store integer > constants for frequently used serde classes in the backend DB. > For example: > {code:java} > org.apache.hadoop.hive.serde2.avro.AvroSerDe"---> 1 > parquet.hive.serde.ParquetHiveSerDe" ---> 2 > ...{code} > Review the patch at: > https://reviews.apache.org/r/68546/ -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (HIVE-20363) Use integer constants for frequently used serde classes.
[ https://issues.apache.org/jira/browse/HIVE-20363?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Anurag Mantripragada updated HIVE-20363: Attachment: HIVE-20363.1.patch Status: Patch Available (was: Open) > Use integer constants for frequently used serde classes. > > > Key: HIVE-20363 > URL: https://issues.apache.org/jira/browse/HIVE-20363 > Project: Hive > Issue Type: Improvement > Components: Metastore, Standalone Metastore >Reporter: Anurag Mantripragada >Assignee: Anurag Mantripragada >Priority: Major > Attachments: HIVE-20363.1.patch > > > Serde libraries are stored as fully qualified class names which are long > strings, we can get improvements in I/O and storage if we store integer > constants for frequently used serde classes in the backend DB. > For example: > {code:java} > org.apache.hadoop.hive.serde2.avro.AvroSerDe"---> 1 > parquet.hive.serde.ParquetHiveSerDe" ---> 2 > ...{code} > Review the patch at: > https://reviews.apache.org/r/68546/ -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (HIVE-20363) Use integer constants for frequently used serde classes.
[ https://issues.apache.org/jira/browse/HIVE-20363?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Anurag Mantripragada updated HIVE-20363: Attachment: HIVE-20363.1.patch > Use integer constants for frequently used serde classes. > > > Key: HIVE-20363 > URL: https://issues.apache.org/jira/browse/HIVE-20363 > Project: Hive > Issue Type: Improvement > Components: Metastore, Standalone Metastore >Reporter: Anurag Mantripragada >Assignee: Anurag Mantripragada >Priority: Major > Attachments: HIVE-20363.1.patch > > > Serde libraries are stored as fully qualified class names which are long > strings, we can get improvements in I/O and storage if we store integer > constants for frequently used serde classes in the backend DB. > For example: > {code:java} > org.apache.hadoop.hive.serde2.avro.AvroSerDe"---> 1 > parquet.hive.serde.ParquetHiveSerDe" ---> 2 > ...{code} > Review the patch at: > https://reviews.apache.org/r/68546/ -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (HIVE-20483) Really move metastore common classes into metastore-common
[ https://issues.apache.org/jira/browse/HIVE-20483?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16595822#comment-16595822 ] Alexander Kolbasov commented on HIVE-20483: --- [~pvary] Would you be able to review this patch? > Really move metastore common classes into metastore-common > -- > > Key: HIVE-20483 > URL: https://issues.apache.org/jira/browse/HIVE-20483 > Project: Hive > Issue Type: Sub-task > Components: Standalone Metastore >Affects Versions: 3.0.1, 4.0.0 >Reporter: Alexander Kolbasov >Assignee: Alexander Kolbasov >Priority: Major > Attachments: HIVE-20483.01.patch > > > HIVE-20482 patch was supposed to move a bunch of files from metastore-server > to metastore-common but for some reason it didn't happen, so now these files > should be moved. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (HIVE-20483) Really move metastore common classes into metastore-common
[ https://issues.apache.org/jira/browse/HIVE-20483?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16595821#comment-16595821 ] Alexander Kolbasov commented on HIVE-20483: --- This patch doesn't introduce any code changes - it just moves files around and changes pom files. It removes metastore-server dependency on hive-metastore and adds it to a couple of modules. > Really move metastore common classes into metastore-common > -- > > Key: HIVE-20483 > URL: https://issues.apache.org/jira/browse/HIVE-20483 > Project: Hive > Issue Type: Sub-task > Components: Standalone Metastore >Affects Versions: 3.0.1, 4.0.0 >Reporter: Alexander Kolbasov >Assignee: Alexander Kolbasov >Priority: Major > Attachments: HIVE-20483.01.patch > > > HIVE-20482 patch was supposed to move a bunch of files from metastore-server > to metastore-common but for some reason it didn't happen, so now these files > should be moved. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (HIVE-19253) HMS ignores tableType property for external tables
[ https://issues.apache.org/jira/browse/HIVE-19253?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Alexander Kolbasov updated HIVE-19253: -- Attachment: HIVE-19253.10.patch > HMS ignores tableType property for external tables > -- > > Key: HIVE-19253 > URL: https://issues.apache.org/jira/browse/HIVE-19253 > Project: Hive > Issue Type: Bug > Components: Metastore >Affects Versions: 3.0.0, 3.1.0, 4.0.0 >Reporter: Alexander Kolbasov >Assignee: Alexander Kolbasov >Priority: Major > Labels: newbie > Attachments: HIVE-19253.01.patch, HIVE-19253.02.patch, > HIVE-19253.03.patch, HIVE-19253.03.patch, HIVE-19253.04.patch, > HIVE-19253.05.patch, HIVE-19253.06.patch, HIVE-19253.07.patch, > HIVE-19253.08.patch, HIVE-19253.09.patch, HIVE-19253.10.patch > > > When someone creates a table using Thrift API they may think that setting > tableType to {{EXTERNAL_TABLE}} creates an external table. And boom - their > table is gone later because HMS will silently change it to managed table. > here is the offending code: > {code:java} > private MTable convertToMTable(Table tbl) throws InvalidObjectException, > MetaException { > ... > // If the table has property EXTERNAL set, update table type > // accordingly > String tableType = tbl.getTableType(); > boolean isExternal = > Boolean.parseBoolean(tbl.getParameters().get("EXTERNAL")); > if (TableType.MANAGED_TABLE.toString().equals(tableType)) { > if (isExternal) { > tableType = TableType.EXTERNAL_TABLE.toString(); > } > } > if (TableType.EXTERNAL_TABLE.toString().equals(tableType)) { > if (!isExternal) { // Here! > tableType = TableType.MANAGED_TABLE.toString(); > } > } > {code} > So if the EXTERNAL parameter is not set, table type is changed to managed > even if it was external in the first place - which is wrong. > More over, in other places code looks at the table property to decide table > type and some places look at parameter. HMS should really make its mind which > one to use. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (HIVE-20377) Hive Kafka Storage Handler
[ https://issues.apache.org/jira/browse/HIVE-20377?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16595819#comment-16595819 ] Hive QA commented on HIVE-20377: | (x) *{color:red}-1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || || || || || {color:brown} Prechecks {color} || | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 1s{color} | {color:green} The patch does not contain any @author tags. {color} | || || || || {color:brown} master Compile Tests {color} || | {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue} 0m 32s{color} | {color:blue} Maven dependency ordering for branch {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 8m 6s{color} | {color:green} master passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 9m 43s{color} | {color:green} master passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 2m 19s{color} | {color:green} master passed {color} | | {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue} 0m 42s{color} | {color:blue} serde in master has 195 extant Findbugs warnings. {color} | | {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue} 0m 19s{color} | {color:blue} itests/qtest-druid in master has 6 extant Findbugs warnings. {color} | | {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue} 0m 48s{color} | {color:blue} itests/util in master has 52 extant Findbugs warnings. {color} | | {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue} 0m 40s{color} | {color:blue} llap-server in master has 84 extant Findbugs warnings. {color} | | {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue} 4m 6s{color} | {color:blue} ql in master has 2310 extant Findbugs warnings. {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 9m 15s{color} | {color:green} master passed {color} | || || || || {color:brown} Patch Compile Tests {color} || | {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue} 0m 7s{color} | {color:blue} Maven dependency ordering for patch {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 11m 9s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 9m 53s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 9m 53s{color} | {color:green} the patch passed {color} | | {color:red}-1{color} | {color:red} checkstyle {color} | {color:red} 0m 10s{color} | {color:red} itests/qtest-druid: The patch generated 37 new + 3 unchanged - 0 fixed = 40 total (was 3) {color} | | {color:red}-1{color} | {color:red} checkstyle {color} | {color:red} 0m 9s{color} | {color:red} kafka-handler: The patch generated 33 new + 0 unchanged - 0 fixed = 33 total (was 0) {color} | | {color:red}-1{color} | {color:red} checkstyle {color} | {color:red} 0m 14s{color} | {color:red} llap-server: The patch generated 1 new + 26 unchanged - 4 fixed = 27 total (was 30) {color} | | {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 0s{color} | {color:green} The patch has no whitespace issues. {color} | | {color:green}+1{color} | {color:green} xml {color} | {color:green} 0m 5s{color} | {color:green} The patch has no ill-formed XML file. {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 7m 36s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 9m 35s{color} | {color:green} the patch passed {color} | || || || || {color:brown} Other Tests {color} || | {color:green}+1{color} | {color:green} asflicense {color} | {color:green} 0m 12s{color} | {color:green} The patch does not generate ASF License warnings. {color} | | {color:black}{color} | {color:black} {color} | {color:black} 79m 0s{color} | {color:black} {color} | \\ \\ || Subsystem || Report/Notes || | Optional Tests | asflicense javac javadoc xml compile findbugs checkstyle | | uname | Linux hiveptest-server-upstream 3.16.0-4-amd64 #1 SMP Debian 3.16.36-1+deb8u1 (2016-09-03) x86_64 GNU/Linux | | Build tool | maven | | Personality | /data/hiveptest/working/yetus_PreCommit-HIVE-Build-13517/dev-support/hive-personality.sh | | git revision | master / d2b81a9 | | Default Java | 1.8.0_111 | | findbugs | v3.0.0 | | checkstyle | http://104.198.109.242/logs//PreCommit-HIVE-Build-13517/yetus/diff-checkstyle-itests_qtest-druid.txt | | checkstyle | http://104.198.109.242/logs//PreCommit-HIVE-Build-13517/yetus/diff-checkstyle-kafka-handler.txt | | checkstyle | http://104.198.109.242/logs//PreCommit-HIVE-Build-13517/yetus/diff-checkstyle-llap-server.txt | | modules | C: serde .
[jira] [Commented] (HIVE-20377) Hive Kafka Storage Handler
[ https://issues.apache.org/jira/browse/HIVE-20377?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16595808#comment-16595808 ] Hive QA commented on HIVE-20377: Here are the results of testing the latest attachment: https://issues.apache.org/jira/secure/attachment/12937508/HIVE-20377.18.patch {color:green}SUCCESS:{color} +1 due to 8 test(s) being added or modified. {color:red}ERROR:{color} -1 due to 5 failed/errored test(s), 14900 tests executed *Failed tests:* {noformat} org.apache.hadoop.hive.cli.TestMiniDruidCliDriver.testCliDriver[druidmini_dynamic_partition] (batchId=193) org.apache.hadoop.hive.cli.TestMiniDruidCliDriver.testCliDriver[druidmini_expressions] (batchId=193) org.apache.hadoop.hive.cli.TestMiniDruidCliDriver.testCliDriver[druidmini_test1] (batchId=193) org.apache.hadoop.hive.cli.TestMiniDruidCliDriver.testCliDriver[druidmini_test_alter] (batchId=193) org.apache.hadoop.hive.cli.TestMiniDruidCliDriver.testCliDriver[druidmini_test_insert] (batchId=193) {noformat} Test results: https://builds.apache.org/job/PreCommit-HIVE-Build/13517/testReport Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/13517/console Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-13517/ Messages: {noformat} Executing org.apache.hive.ptest.execution.TestCheckPhase Executing org.apache.hive.ptest.execution.PrepPhase Executing org.apache.hive.ptest.execution.YetusPhase Executing org.apache.hive.ptest.execution.ExecutionPhase Executing org.apache.hive.ptest.execution.ReportingPhase Tests exited with: TestsFailedException: 5 tests failed {noformat} This message is automatically generated. ATTACHMENT ID: 12937508 - PreCommit-HIVE-Build > Hive Kafka Storage Handler > -- > > Key: HIVE-20377 > URL: https://issues.apache.org/jira/browse/HIVE-20377 > Project: Hive > Issue Type: New Feature >Affects Versions: 4.0.0 >Reporter: slim bouguerra >Assignee: slim bouguerra >Priority: Major > Attachments: HIVE-20377.10.patch, HIVE-20377.11.patch, > HIVE-20377.12.patch, HIVE-20377.15.patch, HIVE-20377.18.patch, > HIVE-20377.4.patch, HIVE-20377.5.patch, HIVE-20377.6.patch, > HIVE-20377.8.patch, HIVE-20377.8.patch, HIVE-20377.patch > > > h1. Goal > * Read streaming data form Kafka queue as an external table. > * Allow streaming navigation by pushing down filters on Kafka record > partition id, offset and timestamp. > * Insert streaming data form Kafka to an actual Hive internal table, using > CTAS statement. > h1. Example > h2. Create the external table > {code} > CREATE EXTERNAL TABLE kafka_table (`timestamp` timestamp, page string, `user` > string, language string, added int, deleted int, flags string,comment string, > namespace string) > STORED BY 'org.apache.hadoop.hive.kafka.KafkaStorageHandler' > TBLPROPERTIES > ("kafka.topic" = "wikipedia", > "kafka.bootstrap.servers"="brokeraddress:9092", > "kafka.serde.class"="org.apache.hadoop.hive.serde2.JsonSerDe"); > {code} > h2. Kafka Metadata > In order to keep track of Kafka records the storage handler will add > automatically the Kafka row metadata eg partition id, record offset and > record timestamp. > {code} > DESCRIBE EXTENDED kafka_table > timestamp timestamp from deserializer > page string from deserializer > user string from deserializer > language string from deserializer > country string from deserializer > continent string from deserializer > namespace string from deserializer > newpage boolean from deserializer > unpatrolled boolean from deserializer > anonymous boolean from deserializer > robot boolean from deserializer > added int from deserializer > deleted int from deserializer > delta bigint from deserializer > __partition int from deserializer > __offset bigint from deserializer > __timestamp bigint from deserializer > {code} > h2. Filter push down. > Newer Kafka consumers 0.11.0 and higher allow seeking on the stream based on > a given offset. The proposed storage handler will be able to leverage such > API by pushing down filters over metadata columns, namely __partition (int), > __offset(long) and __timestamp(long) > For instance Query like > {code} > select `__offset` from kafka_table where (`__offset` < 10 and `__offset`>3 > and `__partition` = 0) or
[jira] [Updated] (HIVE-20483) Really move metastore common classes into metastore-common
[ https://issues.apache.org/jira/browse/HIVE-20483?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Alexander Kolbasov updated HIVE-20483: -- Status: Patch Available (was: Open) > Really move metastore common classes into metastore-common > -- > > Key: HIVE-20483 > URL: https://issues.apache.org/jira/browse/HIVE-20483 > Project: Hive > Issue Type: Sub-task > Components: Standalone Metastore >Affects Versions: 3.0.1, 4.0.0 >Reporter: Alexander Kolbasov >Assignee: Alexander Kolbasov >Priority: Major > Attachments: HIVE-20483.01.patch > > > HIVE-20482 patch was supposed to move a bunch of files from metastore-server > to metastore-common but for some reason it didn't happen, so now these files > should be moved. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (HIVE-20483) Really move metastore common classes into metastore-common
[ https://issues.apache.org/jira/browse/HIVE-20483?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Alexander Kolbasov updated HIVE-20483: -- Attachment: HIVE-20483.01.patch > Really move metastore common classes into metastore-common > -- > > Key: HIVE-20483 > URL: https://issues.apache.org/jira/browse/HIVE-20483 > Project: Hive > Issue Type: Sub-task > Components: Standalone Metastore >Affects Versions: 3.0.1, 4.0.0 >Reporter: Alexander Kolbasov >Assignee: Alexander Kolbasov >Priority: Major > Attachments: HIVE-20483.01.patch > > > HIVE-20482 patch was supposed to move a bunch of files from metastore-server > to metastore-common but for some reason it didn't happen, so now these files > should be moved. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (HIVE-20451) Metastore client and server tarball issues
[ https://issues.apache.org/jira/browse/HIVE-20451?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16595806#comment-16595806 ] Alexander Kolbasov commented on HIVE-20451: --- [~alangates] Would you be able to review/commit the change? > Metastore client and server tarball issues > -- > > Key: HIVE-20451 > URL: https://issues.apache.org/jira/browse/HIVE-20451 > Project: Hive > Issue Type: Bug > Components: Standalone Metastore >Affects Versions: 4.0.0 >Reporter: Alan Gates >Assignee: Alexander Kolbasov >Priority: Major > Attachments: HIVE-20451.01.patch, HIVE-20451.02.patch, > HIVE-20451.03.patch > > > With the split of the metastore into common and server there are now two sets > of tarballs. There are a couple of issues here. > # It doesn't make sense to have separate source tarballs for each. The > source release should still be done from the standalone-metastore directory > and include all code for the metastore. > # The binary tarballs should have separate names. At the moment both are > named apache-hive-metastore. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Assigned] (HIVE-20483) Really move metastore common classes into metastore-common
[ https://issues.apache.org/jira/browse/HIVE-20483?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Alexander Kolbasov reassigned HIVE-20483: - > Really move metastore common classes into metastore-common > -- > > Key: HIVE-20483 > URL: https://issues.apache.org/jira/browse/HIVE-20483 > Project: Hive > Issue Type: Sub-task > Components: Standalone Metastore >Affects Versions: 3.0.1, 4.0.0 >Reporter: Alexander Kolbasov >Assignee: Alexander Kolbasov >Priority: Major > > HIVE-20482 patch was supposed to move a bunch of files from metastore-server > to metastore-common but for some reason it didn't happen, so now these files > should be moved. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Assigned] (HIVE-20482) Remove dependency on metastore-server
[ https://issues.apache.org/jira/browse/HIVE-20482?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Alexander Kolbasov reassigned HIVE-20482: - > Remove dependency on metastore-server > - > > Key: HIVE-20482 > URL: https://issues.apache.org/jira/browse/HIVE-20482 > Project: Hive > Issue Type: Sub-task > Components: Standalone Metastore >Affects Versions: 3.0.1, 4.0.0 >Reporter: Alexander Kolbasov >Assignee: Alexander Kolbasov >Priority: Major > > Now that we separated common and server classes we should remove dependency > on the server module from poms. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (HIVE-20472) mvn test failing for metastore-tool module
[ https://issues.apache.org/jira/browse/HIVE-20472?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16595764#comment-16595764 ] Alexander Kolbasov commented on HIVE-20472: --- [~pvary] Can you commit the patch? > mvn test failing for metastore-tool module > -- > > Key: HIVE-20472 > URL: https://issues.apache.org/jira/browse/HIVE-20472 > Project: Hive > Issue Type: Bug >Affects Versions: 4.0.0 >Reporter: Vihang Karajgaonkar >Assignee: Alexander Kolbasov >Priority: Trivial > Fix For: 4.0.0 > > Attachments: HIVE-20472.01.patch, HIVE-20472.02.patch > > > Fails because there are no applicable tests. > > {code} > [ERROR] Failed to execute goal > org.apache.maven.plugins:maven-surefire-plugin:2.20.1:test (default-test) on > project hive-metastore-benchmarks: No tests were executed! (Set > -DfailIfNoTests=false to ignore this error.) -> [Help 1] > [ERROR] > [ERROR] To see the full stack trace of the errors, re-run Maven with the -e > switch. > [ERROR] Re-run Maven using the -X switch to enable full debug logging. > [ERROR] > [ERROR] For more information about the errors and possible solutions, please > read the following articles: > [ERROR] [Help 1] > http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException > [ERROR] > [ERROR] After correcting the problems, you can resume the build with the > command > [ERROR] mvn -rf :hive-metastore-benchmarks > {code} -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (HIVE-20451) Metastore client and server tarball issues
[ https://issues.apache.org/jira/browse/HIVE-20451?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Alexander Kolbasov updated HIVE-20451: -- Attachment: HIVE-20451.03.patch > Metastore client and server tarball issues > -- > > Key: HIVE-20451 > URL: https://issues.apache.org/jira/browse/HIVE-20451 > Project: Hive > Issue Type: Bug > Components: Standalone Metastore >Affects Versions: 4.0.0 >Reporter: Alan Gates >Assignee: Alexander Kolbasov >Priority: Major > Attachments: HIVE-20451.01.patch, HIVE-20451.02.patch, > HIVE-20451.03.patch > > > With the split of the metastore into common and server there are now two sets > of tarballs. There are a couple of issues here. > # It doesn't make sense to have separate source tarballs for each. The > source release should still be done from the standalone-metastore directory > and include all code for the metastore. > # The binary tarballs should have separate names. At the moment both are > named apache-hive-metastore. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (HIVE-20451) Metastore client and server tarball issues
[ https://issues.apache.org/jira/browse/HIVE-20451?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16595762#comment-16595762 ] Alexander Kolbasov commented on HIVE-20451: --- Finally figured out the way to suppress generation of source tarballs in submodules, so with the latest patch a single source tarball is generated in standalone-metastore/target directory. > Metastore client and server tarball issues > -- > > Key: HIVE-20451 > URL: https://issues.apache.org/jira/browse/HIVE-20451 > Project: Hive > Issue Type: Bug > Components: Standalone Metastore >Affects Versions: 4.0.0 >Reporter: Alan Gates >Assignee: Alexander Kolbasov >Priority: Major > Attachments: HIVE-20451.01.patch, HIVE-20451.02.patch, > HIVE-20451.03.patch > > > With the split of the metastore into common and server there are now two sets > of tarballs. There are a couple of issues here. > # It doesn't make sense to have separate source tarballs for each. The > source release should still be done from the standalone-metastore directory > and include all code for the metastore. > # The binary tarballs should have separate names. At the moment both are > named apache-hive-metastore. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (HIVE-20472) mvn test failing for metastore-tool module
[ https://issues.apache.org/jira/browse/HIVE-20472?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16595756#comment-16595756 ] Hive QA commented on HIVE-20472: Here are the results of testing the latest attachment: https://issues.apache.org/jira/secure/attachment/12937473/HIVE-20472.02.patch {color:red}ERROR:{color} -1 due to no test(s) being added or modified. {color:green}SUCCESS:{color} +1 due to 14899 tests passed Test results: https://builds.apache.org/job/PreCommit-HIVE-Build/13516/testReport Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/13516/console Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-13516/ Messages: {noformat} Executing org.apache.hive.ptest.execution.TestCheckPhase Executing org.apache.hive.ptest.execution.PrepPhase Executing org.apache.hive.ptest.execution.YetusPhase Executing org.apache.hive.ptest.execution.ExecutionPhase Executing org.apache.hive.ptest.execution.ReportingPhase {noformat} This message is automatically generated. ATTACHMENT ID: 12937473 - PreCommit-HIVE-Build > mvn test failing for metastore-tool module > -- > > Key: HIVE-20472 > URL: https://issues.apache.org/jira/browse/HIVE-20472 > Project: Hive > Issue Type: Bug >Affects Versions: 4.0.0 >Reporter: Vihang Karajgaonkar >Assignee: Alexander Kolbasov >Priority: Trivial > Fix For: 4.0.0 > > Attachments: HIVE-20472.01.patch, HIVE-20472.02.patch > > > Fails because there are no applicable tests. > > {code} > [ERROR] Failed to execute goal > org.apache.maven.plugins:maven-surefire-plugin:2.20.1:test (default-test) on > project hive-metastore-benchmarks: No tests were executed! (Set > -DfailIfNoTests=false to ignore this error.) -> [Help 1] > [ERROR] > [ERROR] To see the full stack trace of the errors, re-run Maven with the -e > switch. > [ERROR] Re-run Maven using the -X switch to enable full debug logging. > [ERROR] > [ERROR] For more information about the errors and possible solutions, please > read the following articles: > [ERROR] [Help 1] > http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException > [ERROR] > [ERROR] After correcting the problems, you can resume the build with the > command > [ERROR] mvn -rf :hive-metastore-benchmarks > {code} -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (HIVE-20291) Allow HiveStreamingConnection to receive a WriteId
[ https://issues.apache.org/jira/browse/HIVE-20291?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Jaume M updated HIVE-20291: --- Attachment: HIVE-20291.3.patch Status: Patch Available (was: Open) > Allow HiveStreamingConnection to receive a WriteId > -- > > Key: HIVE-20291 > URL: https://issues.apache.org/jira/browse/HIVE-20291 > Project: Hive > Issue Type: Improvement >Reporter: Jaume M >Assignee: Jaume M >Priority: Major > Labels: pull-request-available > Attachments: HIVE-20291.1.patch, HIVE-20291.2.patch, > HIVE-20291.3.patch > > > If the writeId is received externally it won't need to open connections to > the metastore. It won't be able to the commit in this case as well so it must > be done by the entity passing the writeId. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (HIVE-20291) Allow HiveStreamingConnection to receive a WriteId
[ https://issues.apache.org/jira/browse/HIVE-20291?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Jaume M updated HIVE-20291: --- Attachment: TEZ-3985.3.patch Status: Patch Available (was: Open) > Allow HiveStreamingConnection to receive a WriteId > -- > > Key: HIVE-20291 > URL: https://issues.apache.org/jira/browse/HIVE-20291 > Project: Hive > Issue Type: Improvement >Reporter: Jaume M >Assignee: Jaume M >Priority: Major > Labels: pull-request-available > Attachments: HIVE-20291.1.patch, HIVE-20291.2.patch > > > If the writeId is received externally it won't need to open connections to > the metastore. It won't be able to the commit in this case as well so it must > be done by the entity passing the writeId. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (HIVE-20291) Allow HiveStreamingConnection to receive a WriteId
[ https://issues.apache.org/jira/browse/HIVE-20291?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Jaume M updated HIVE-20291: --- Attachment: (was: TEZ-3985.3.patch) > Allow HiveStreamingConnection to receive a WriteId > -- > > Key: HIVE-20291 > URL: https://issues.apache.org/jira/browse/HIVE-20291 > Project: Hive > Issue Type: Improvement >Reporter: Jaume M >Assignee: Jaume M >Priority: Major > Labels: pull-request-available > Attachments: HIVE-20291.1.patch, HIVE-20291.2.patch > > > If the writeId is received externally it won't need to open connections to > the metastore. It won't be able to the commit in this case as well so it must > be done by the entity passing the writeId. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (HIVE-20291) Allow HiveStreamingConnection to receive a WriteId
[ https://issues.apache.org/jira/browse/HIVE-20291?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Jaume M updated HIVE-20291: --- Status: Open (was: Patch Available) > Allow HiveStreamingConnection to receive a WriteId > -- > > Key: HIVE-20291 > URL: https://issues.apache.org/jira/browse/HIVE-20291 > Project: Hive > Issue Type: Improvement >Reporter: Jaume M >Assignee: Jaume M >Priority: Major > Labels: pull-request-available > Attachments: HIVE-20291.1.patch, HIVE-20291.2.patch > > > If the writeId is received externally it won't need to open connections to > the metastore. It won't be able to the commit in this case as well so it must > be done by the entity passing the writeId. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (HIVE-20291) Allow HiveStreamingConnection to receive a WriteId
[ https://issues.apache.org/jira/browse/HIVE-20291?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Jaume M updated HIVE-20291: --- Status: Open (was: Patch Available) > Allow HiveStreamingConnection to receive a WriteId > -- > > Key: HIVE-20291 > URL: https://issues.apache.org/jira/browse/HIVE-20291 > Project: Hive > Issue Type: Improvement >Reporter: Jaume M >Assignee: Jaume M >Priority: Major > Labels: pull-request-available > Attachments: HIVE-20291.1.patch, HIVE-20291.2.patch > > > If the writeId is received externally it won't need to open connections to > the metastore. It won't be able to the commit in this case as well so it must > be done by the entity passing the writeId. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (HIVE-18908) FULL OUTER JOIN to MapJoin
[ https://issues.apache.org/jira/browse/HIVE-18908?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Matt McCline updated HIVE-18908: Status: Patch Available (was: In Progress) > FULL OUTER JOIN to MapJoin > -- > > Key: HIVE-18908 > URL: https://issues.apache.org/jira/browse/HIVE-18908 > Project: Hive > Issue Type: Improvement > Components: Hive >Reporter: Matt McCline >Assignee: Matt McCline >Priority: Critical > Attachments: FULL OUTER MapJoin Code Changes.docx, > HIVE-18908.01.patch, HIVE-18908.02.patch, HIVE-18908.03.patch, > HIVE-18908.04.patch, HIVE-18908.05.patch, HIVE-18908.06.patch, > HIVE-18908.08.patch, HIVE-18908.09.patch, HIVE-18908.091.patch, > HIVE-18908.092.patch, HIVE-18908.093.patch, HIVE-18908.096.patch, > HIVE-18908.097.patch, HIVE-18908.098.patch, HIVE-18908.099.patch, JOIN to > MAPJOIN Transformation.pdf, SHARED-MEMORY FULL OUTER MapJoin.pdf > > > Currently, we do not support FULL OUTER JOIN in MapJoin. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (HIVE-18908) FULL OUTER JOIN to MapJoin
[ https://issues.apache.org/jira/browse/HIVE-18908?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Matt McCline updated HIVE-18908: Attachment: HIVE-18908.099.patch > FULL OUTER JOIN to MapJoin > -- > > Key: HIVE-18908 > URL: https://issues.apache.org/jira/browse/HIVE-18908 > Project: Hive > Issue Type: Improvement > Components: Hive >Reporter: Matt McCline >Assignee: Matt McCline >Priority: Critical > Attachments: FULL OUTER MapJoin Code Changes.docx, > HIVE-18908.01.patch, HIVE-18908.02.patch, HIVE-18908.03.patch, > HIVE-18908.04.patch, HIVE-18908.05.patch, HIVE-18908.06.patch, > HIVE-18908.08.patch, HIVE-18908.09.patch, HIVE-18908.091.patch, > HIVE-18908.092.patch, HIVE-18908.093.patch, HIVE-18908.096.patch, > HIVE-18908.097.patch, HIVE-18908.098.patch, HIVE-18908.099.patch, JOIN to > MAPJOIN Transformation.pdf, SHARED-MEMORY FULL OUTER MapJoin.pdf > > > Currently, we do not support FULL OUTER JOIN in MapJoin. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (HIVE-18908) FULL OUTER JOIN to MapJoin
[ https://issues.apache.org/jira/browse/HIVE-18908?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Matt McCline updated HIVE-18908: Status: In Progress (was: Patch Available) > FULL OUTER JOIN to MapJoin > -- > > Key: HIVE-18908 > URL: https://issues.apache.org/jira/browse/HIVE-18908 > Project: Hive > Issue Type: Improvement > Components: Hive >Reporter: Matt McCline >Assignee: Matt McCline >Priority: Critical > Attachments: FULL OUTER MapJoin Code Changes.docx, > HIVE-18908.01.patch, HIVE-18908.02.patch, HIVE-18908.03.patch, > HIVE-18908.04.patch, HIVE-18908.05.patch, HIVE-18908.06.patch, > HIVE-18908.08.patch, HIVE-18908.09.patch, HIVE-18908.091.patch, > HIVE-18908.092.patch, HIVE-18908.093.patch, HIVE-18908.096.patch, > HIVE-18908.097.patch, HIVE-18908.098.patch, JOIN to MAPJOIN > Transformation.pdf, SHARED-MEMORY FULL OUTER MapJoin.pdf > > > Currently, we do not support FULL OUTER JOIN in MapJoin. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (HIVE-20472) mvn test failing for metastore-tool module
[ https://issues.apache.org/jira/browse/HIVE-20472?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16595721#comment-16595721 ] Hive QA commented on HIVE-20472: | (/) *{color:green}+1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || || || || || {color:brown} Prechecks {color} || | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s{color} | {color:green} The patch does not contain any @author tags. {color} | || || || || {color:brown} master Compile Tests {color} || | {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue} 0m 31s{color} | {color:blue} Maven dependency ordering for branch {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 8m 10s{color} | {color:green} master passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 2m 23s{color} | {color:green} master passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 2m 40s{color} | {color:green} master passed {color} | || || || || {color:brown} Patch Compile Tests {color} || | {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue} 0m 7s{color} | {color:blue} Maven dependency ordering for patch {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 2m 45s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 2m 23s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 2m 23s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 0s{color} | {color:green} The patch has no whitespace issues. {color} | | {color:green}+1{color} | {color:green} xml {color} | {color:green} 0m 4s{color} | {color:green} The patch has no ill-formed XML file. {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 2m 37s{color} | {color:green} the patch passed {color} | || || || || {color:brown} Other Tests {color} || | {color:green}+1{color} | {color:green} asflicense {color} | {color:green} 0m 13s{color} | {color:green} The patch does not generate ASF License warnings. {color} | | {color:black}{color} | {color:black} {color} | {color:black} 22m 30s{color} | {color:black} {color} | \\ \\ || Subsystem || Report/Notes || | Optional Tests | asflicense javac javadoc xml compile | | uname | Linux hiveptest-server-upstream 3.16.0-4-amd64 #1 SMP Debian 3.16.36-1+deb8u1 (2016-09-03) x86_64 GNU/Linux | | Build tool | maven | | Personality | /data/hiveptest/working/yetus_PreCommit-HIVE-Build-13516/dev-support/hive-personality.sh | | git revision | master / d2b81a9 | | Default Java | 1.8.0_111 | | modules | C: standalone-metastore standalone-metastore/metastore-common standalone-metastore/metastore-server standalone-metastore/metastore-tools U: standalone-metastore | | Console output | http://104.198.109.242/logs//PreCommit-HIVE-Build-13516/yetus.txt | | Powered by | Apache Yetushttp://yetus.apache.org | This message was automatically generated. > mvn test failing for metastore-tool module > -- > > Key: HIVE-20472 > URL: https://issues.apache.org/jira/browse/HIVE-20472 > Project: Hive > Issue Type: Bug >Affects Versions: 4.0.0 >Reporter: Vihang Karajgaonkar >Assignee: Alexander Kolbasov >Priority: Trivial > Fix For: 4.0.0 > > Attachments: HIVE-20472.01.patch, HIVE-20472.02.patch > > > Fails because there are no applicable tests. > > {code} > [ERROR] Failed to execute goal > org.apache.maven.plugins:maven-surefire-plugin:2.20.1:test (default-test) on > project hive-metastore-benchmarks: No tests were executed! (Set > -DfailIfNoTests=false to ignore this error.) -> [Help 1] > [ERROR] > [ERROR] To see the full stack trace of the errors, re-run Maven with the -e > switch. > [ERROR] Re-run Maven using the -X switch to enable full debug logging. > [ERROR] > [ERROR] For more information about the errors and possible solutions, please > read the following articles: > [ERROR] [Help 1] > http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException > [ERROR] > [ERROR] After correcting the problems, you can resume the build with the > command > [ERROR] mvn -rf :hive-metastore-benchmarks > {code} -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (HIVE-18908) FULL OUTER JOIN to MapJoin
[ https://issues.apache.org/jira/browse/HIVE-18908?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16595688#comment-16595688 ] Hive QA commented on HIVE-18908: Here are the results of testing the latest attachment: https://issues.apache.org/jira/secure/attachment/12937469/HIVE-18908.098.patch {color:red}ERROR:{color} -1 due to build exiting with an error Test results: https://builds.apache.org/job/PreCommit-HIVE-Build/13515/testReport Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/13515/console Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-13515/ Messages: {noformat} Executing org.apache.hive.ptest.execution.TestCheckPhase Executing org.apache.hive.ptest.execution.PrepPhase Tests exited with: NonZeroExitCodeException Command 'bash /data/hiveptest/working/scratch/source-prep.sh' failed with exit status 1 and output '+ date '+%Y-%m-%d %T.%3N' 2018-08-28 22:54:14.251 + [[ -n /usr/lib/jvm/java-8-openjdk-amd64 ]] + export JAVA_HOME=/usr/lib/jvm/java-8-openjdk-amd64 + JAVA_HOME=/usr/lib/jvm/java-8-openjdk-amd64 + export PATH=/usr/lib/jvm/java-8-openjdk-amd64/bin/:/usr/local/bin:/usr/bin:/bin:/usr/local/games:/usr/games + PATH=/usr/lib/jvm/java-8-openjdk-amd64/bin/:/usr/local/bin:/usr/bin:/bin:/usr/local/games:/usr/games + export 'ANT_OPTS=-Xmx1g -XX:MaxPermSize=256m ' + ANT_OPTS='-Xmx1g -XX:MaxPermSize=256m ' + export 'MAVEN_OPTS=-Xmx1g ' + MAVEN_OPTS='-Xmx1g ' + cd /data/hiveptest/working/ + tee /data/hiveptest/logs/PreCommit-HIVE-Build-13515/source-prep.txt + [[ false == \t\r\u\e ]] + mkdir -p maven ivy + [[ git = \s\v\n ]] + [[ git = \g\i\t ]] + [[ -z master ]] + [[ -d apache-github-source-source ]] + [[ ! -d apache-github-source-source/.git ]] + [[ ! -d apache-github-source-source ]] + date '+%Y-%m-%d %T.%3N' 2018-08-28 22:54:14.255 + cd apache-github-source-source + git fetch origin + git reset --hard HEAD HEAD is now at d2b81a9 HIVE-19674: Group by Decimal Constants push down to Druid tables(Vineet Garg, reviewed by Ashutosh Chauhan) + git clean -f -d + git checkout master Already on 'master' Your branch is up-to-date with 'origin/master'. + git reset --hard origin/master HEAD is now at d2b81a9 HIVE-19674: Group by Decimal Constants push down to Druid tables(Vineet Garg, reviewed by Ashutosh Chauhan) + git merge --ff-only origin/master Already up-to-date. + date '+%Y-%m-%d %T.%3N' 2018-08-28 22:54:14.923 + rm -rf ../yetus_PreCommit-HIVE-Build-13515 + mkdir ../yetus_PreCommit-HIVE-Build-13515 + git gc + cp -R . ../yetus_PreCommit-HIVE-Build-13515 + mkdir /data/hiveptest/logs/PreCommit-HIVE-Build-13515/yetus + patchCommandPath=/data/hiveptest/working/scratch/smart-apply-patch.sh + patchFilePath=/data/hiveptest/working/scratch/build.patch + [[ -f /data/hiveptest/working/scratch/build.patch ]] + chmod +x /data/hiveptest/working/scratch/smart-apply-patch.sh + /data/hiveptest/working/scratch/smart-apply-patch.sh /data/hiveptest/working/scratch/build.patch error: patch failed: ql/src/test/results/clientpositive/llap/vector_groupby_sort_11.q.out:693 Falling back to three-way merge... Applied patch to 'ql/src/test/results/clientpositive/llap/vector_groupby_sort_11.q.out' cleanly. error: patch failed: ql/src/test/results/clientpositive/spark/spark_vectorized_dynamic_partition_pruning.q.out:4818 Falling back to three-way merge... Applied patch to 'ql/src/test/results/clientpositive/spark/spark_vectorized_dynamic_partition_pruning.q.out' with conflicts. Going to apply patch with: git apply -p0 /data/hiveptest/working/scratch/build.patch:18802: trailing whitespace. -- FROM /data/hiveptest/working/scratch/build.patch:18804: trailing whitespace. -- JOIN /data/hiveptest/working/scratch/build.patch:18818: trailing whitespace. -- FROM /data/hiveptest/working/scratch/build.patch:18820: trailing whitespace. -- LEFT OUTER JOIN /data/hiveptest/working/scratch/build.patch:18835: trailing whitespace. -- FROM error: patch failed: ql/src/test/results/clientpositive/llap/vector_groupby_sort_11.q.out:693 Falling back to three-way merge... Applied patch to 'ql/src/test/results/clientpositive/llap/vector_groupby_sort_11.q.out' cleanly. error: patch failed: ql/src/test/results/clientpositive/spark/spark_vectorized_dynamic_partition_pruning.q.out:4818 Falling back to three-way merge... Applied patch to 'ql/src/test/results/clientpositive/spark/spark_vectorized_dynamic_partition_pruning.q.out' with conflicts. U ql/src/test/results/clientpositive/spark/spark_vectorized_dynamic_partition_pruning.q.out warning: squelched 641 whitespace errors warning: 646 lines add whitespace errors. + result=1 + '[' 1 -ne 0 ']' + rm -rf yetus_PreCommit-HIVE-Build-13515 + exit 1 ' {noformat} This message is automatically generated. ATTACHMENT ID: 12937469 - PreCommit-HIVE-Build > FULL OUTER JOIN to MapJoin > -- > > Key: HIVE-18908 > URL:
[jira] [Commented] (HIVE-20476) CopyUtils used by REPL LOAD and EXPORT/IMPORT operations ignore distcp error.
[ https://issues.apache.org/jira/browse/HIVE-20476?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16595682#comment-16595682 ] Hive QA commented on HIVE-20476: Here are the results of testing the latest attachment: https://issues.apache.org/jira/secure/attachment/12937461/HIVE-20476.01.patch {color:green}SUCCESS:{color} +1 due to 1 test(s) being added or modified. {color:red}ERROR:{color} -1 due to 1 failed/errored test(s), 14900 tests executed *Failed tests:* {noformat} org.apache.hive.jdbc.miniHS2.TestHs2ConnectionMetricsBinary.testOpenConnectionMetrics (batchId=255) {noformat} Test results: https://builds.apache.org/job/PreCommit-HIVE-Build/13513/testReport Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/13513/console Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-13513/ Messages: {noformat} Executing org.apache.hive.ptest.execution.TestCheckPhase Executing org.apache.hive.ptest.execution.PrepPhase Executing org.apache.hive.ptest.execution.YetusPhase Executing org.apache.hive.ptest.execution.ExecutionPhase Executing org.apache.hive.ptest.execution.ReportingPhase Tests exited with: TestsFailedException: 1 tests failed {noformat} This message is automatically generated. ATTACHMENT ID: 12937461 - PreCommit-HIVE-Build > CopyUtils used by REPL LOAD and EXPORT/IMPORT operations ignore distcp error. > - > > Key: HIVE-20476 > URL: https://issues.apache.org/jira/browse/HIVE-20476 > Project: Hive > Issue Type: Bug > Components: HiveServer2, repl >Affects Versions: 3.1.0 >Reporter: Sankar Hariappan >Assignee: Sankar Hariappan >Priority: Major > Labels: DR, Export, Import, Replication, pull-request-available > Attachments: HIVE-20476.01.patch > > > CopyUtils uses FileUtils.distCp to copy files but doesn't check the return > value. It returns false if the copy fails. > Now, REPL LOAD and EXPORT/IMPORT commands internally uses CopyUtils to copy > data files across clusters and here it may return success even if file copy > fails and may cause data loss. > Need to throw error and retry. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Comment Edited] (HIVE-20451) Metastore client and server tarball issues
[ https://issues.apache.org/jira/browse/HIVE-20451?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16595648#comment-16595648 ] Alexander Kolbasov edited comment on HIVE-20451 at 8/28/18 9:54 PM: Posted an updated patch that seems a bit cleaner. Here is what it does. # Only one binary tarball is generated in metastore-server target directory. It is called {{apache-hive-standalone-metastore-server-4.0.0-SNAPSHOT-bin.tar.gz}}. I don't think that we need binary tarbal for the metastore-common - there are no binaries to run from there. # {{standalone-metastore/target}} has the overall tarball called {{apache-hive-standalone-metastore-4.0.0-SNAPSHOT-src.tar.gz}} which includes full source and can be used to build standalone metastore. # submodules have their own source tarballs that can be used to build these submodules. They all have different names. I don't know how the release process works, but basically it may choose to exclude these from the release. Is this an acceptable state of affairs? was (Author: akolb): Posted an updated patch that seems a bit cleaner. Here is what it does. # Only one binary tarball is generated in metastore-server target directory. It is called {{apache-hive-standalone-metastore-server-4.0.0-SNAPSHOT-bin.tar.gz}}. I don't think that we need binary tarbal for the metastore-common - there are no binaries to run from there. # {{standalone-metastore/target}} has the overall tarball called {{apache-hive-standalone-metastore-4.0.0-SNAPSHOT-src.tar.gz }} which includes full source and can be used to build standalone metastore. # submodules have their own source tarballs that can be used to build these submodules. They all have different names. I don't know how the release process works, but basically it may choose to exclude these from the release. Is this an acceptable state of affairs? > Metastore client and server tarball issues > -- > > Key: HIVE-20451 > URL: https://issues.apache.org/jira/browse/HIVE-20451 > Project: Hive > Issue Type: Bug > Components: Standalone Metastore >Affects Versions: 4.0.0 >Reporter: Alan Gates >Assignee: Alexander Kolbasov >Priority: Major > Attachments: HIVE-20451.01.patch, HIVE-20451.02.patch > > > With the split of the metastore into common and server there are now two sets > of tarballs. There are a couple of issues here. > # It doesn't make sense to have separate source tarballs for each. The > source release should still be done from the standalone-metastore directory > and include all code for the metastore. > # The binary tarballs should have separate names. At the moment both are > named apache-hive-metastore. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (HIVE-20451) Metastore client and server tarball issues
[ https://issues.apache.org/jira/browse/HIVE-20451?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16595648#comment-16595648 ] Alexander Kolbasov commented on HIVE-20451: --- Posted an updated patch that seems a bit cleaner. Here is what it does. # Only one binary tarball is generated in metastore-server target directory. It is called {{apache-hive-standalone-metastore-server-4.0.0-SNAPSHOT-bin.tar.gz}}. I don't think that we need binary tarbal for the metastore-common - there are no binaries to run from there. # {{standalone-metastore/target}} has the overall tarball called {{apache-hive-standalone-metastore-4.0.0-SNAPSHOT-src.tar.gz }} which includes full source and can be used to build standalone metastore. # submodules have their own source tarballs that can be used to build these submodules. They all have different names. I don't know how the release process works, but basically it may choose to exclude these from the release. Is this an acceptable state of affairs? > Metastore client and server tarball issues > -- > > Key: HIVE-20451 > URL: https://issues.apache.org/jira/browse/HIVE-20451 > Project: Hive > Issue Type: Bug > Components: Standalone Metastore >Affects Versions: 4.0.0 >Reporter: Alan Gates >Assignee: Alexander Kolbasov >Priority: Major > Attachments: HIVE-20451.01.patch, HIVE-20451.02.patch > > > With the split of the metastore into common and server there are now two sets > of tarballs. There are a couple of issues here. > # It doesn't make sense to have separate source tarballs for each. The > source release should still be done from the standalone-metastore directory > and include all code for the metastore. > # The binary tarballs should have separate names. At the moment both are > named apache-hive-metastore. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (HIVE-20451) Metastore client and server tarball issues
[ https://issues.apache.org/jira/browse/HIVE-20451?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Alexander Kolbasov updated HIVE-20451: -- Attachment: HIVE-20451.02.patch > Metastore client and server tarball issues > -- > > Key: HIVE-20451 > URL: https://issues.apache.org/jira/browse/HIVE-20451 > Project: Hive > Issue Type: Bug > Components: Standalone Metastore >Affects Versions: 4.0.0 >Reporter: Alan Gates >Assignee: Alexander Kolbasov >Priority: Major > Attachments: HIVE-20451.01.patch, HIVE-20451.02.patch > > > With the split of the metastore into common and server there are now two sets > of tarballs. There are a couple of issues here. > # It doesn't make sense to have separate source tarballs for each. The > source release should still be done from the standalone-metastore directory > and include all code for the metastore. > # The binary tarballs should have separate names. At the moment both are > named apache-hive-metastore. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (HIVE-20377) Hive Kafka Storage Handler
[ https://issues.apache.org/jira/browse/HIVE-20377?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16595622#comment-16595622 ] Gopal V commented on HIVE-20377: +1 tests pending > Hive Kafka Storage Handler > -- > > Key: HIVE-20377 > URL: https://issues.apache.org/jira/browse/HIVE-20377 > Project: Hive > Issue Type: New Feature >Affects Versions: 4.0.0 >Reporter: slim bouguerra >Assignee: slim bouguerra >Priority: Major > Attachments: HIVE-20377.10.patch, HIVE-20377.11.patch, > HIVE-20377.12.patch, HIVE-20377.15.patch, HIVE-20377.18.patch, > HIVE-20377.4.patch, HIVE-20377.5.patch, HIVE-20377.6.patch, > HIVE-20377.8.patch, HIVE-20377.8.patch, HIVE-20377.patch > > > h1. Goal > * Read streaming data form Kafka queue as an external table. > * Allow streaming navigation by pushing down filters on Kafka record > partition id, offset and timestamp. > * Insert streaming data form Kafka to an actual Hive internal table, using > CTAS statement. > h1. Example > h2. Create the external table > {code} > CREATE EXTERNAL TABLE kafka_table (`timestamp` timestamp, page string, `user` > string, language string, added int, deleted int, flags string,comment string, > namespace string) > STORED BY 'org.apache.hadoop.hive.kafka.KafkaStorageHandler' > TBLPROPERTIES > ("kafka.topic" = "wikipedia", > "kafka.bootstrap.servers"="brokeraddress:9092", > "kafka.serde.class"="org.apache.hadoop.hive.serde2.JsonSerDe"); > {code} > h2. Kafka Metadata > In order to keep track of Kafka records the storage handler will add > automatically the Kafka row metadata eg partition id, record offset and > record timestamp. > {code} > DESCRIBE EXTENDED kafka_table > timestamp timestamp from deserializer > page string from deserializer > user string from deserializer > language string from deserializer > country string from deserializer > continent string from deserializer > namespace string from deserializer > newpage boolean from deserializer > unpatrolled boolean from deserializer > anonymous boolean from deserializer > robot boolean from deserializer > added int from deserializer > deleted int from deserializer > delta bigint from deserializer > __partition int from deserializer > __offset bigint from deserializer > __timestamp bigint from deserializer > {code} > h2. Filter push down. > Newer Kafka consumers 0.11.0 and higher allow seeking on the stream based on > a given offset. The proposed storage handler will be able to leverage such > API by pushing down filters over metadata columns, namely __partition (int), > __offset(long) and __timestamp(long) > For instance Query like > {code} > select `__offset` from kafka_table where (`__offset` < 10 and `__offset`>3 > and `__partition` = 0) or (`__partition` = 0 and `__offset` < 105 and > `__offset` > 99) or (`__offset` = 109); > {code} > Will result on a scan of partition 0 only then read only records between > offset 4 and 109. > h2. With timestamp seeks > The seeking based on the internal timestamps allows the handler to run on > recently arrived data, by doing > {code} > select count(*) from kafka_table where `__timestamp` > 1000 * > to_unix_timestamp(CURRENT_TIMESTAMP - interval '20' hours) ; > {code} > This allows for implicit relationships between event timestamps and kafka > timestamps to be expressed in queries (i.e event_timestamp is always < than > kafka __timestamp and kafka __timestamp is never > 15 minutes from event etc). > h2. More examples with Avro > {code} > CREATE EXTERNAL TABLE wiki_kafka_avro_table > STORED BY 'org.apache.hadoop.hive.kafka.KafkaStorageHandler' > TBLPROPERTIES > ("kafka.topic" = "wiki_kafka_avro_table", > "kafka.bootstrap.servers"="localhost:9092", > "kafka.serde.class"="org.apache.hadoop.hive.serde2.avro.AvroSerDe", > 'avro.schema.literal'='{ > "type" : "record", > "name" : "Wikipedia", > "namespace" : "org.apache.hive.kafka", > "version": "1", > "fields" : [ { > "name" : "isrobot", > "type" : "boolean" > }, { > "name" : "channel", > "type" : "string" > }, { > "name" : "timestamp", > "type" : "string" > }, { > "name" : "flags", > "type" : "string" > }, { > "name" : "isunpatrolled", > "type" : "boolean" > }, { >
[jira] [Commented] (HIVE-20476) CopyUtils used by REPL LOAD and EXPORT/IMPORT operations ignore distcp error.
[ https://issues.apache.org/jira/browse/HIVE-20476?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16595618#comment-16595618 ] Hive QA commented on HIVE-20476: | (x) *{color:red}-1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || || || || || {color:brown} Prechecks {color} || | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s{color} | {color:green} The patch does not contain any @author tags. {color} | || || || || {color:brown} master Compile Tests {color} || | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 8m 42s{color} | {color:green} master passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 1m 5s{color} | {color:green} master passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 42s{color} | {color:green} master passed {color} | | {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue} 4m 9s{color} | {color:blue} ql in master has 2310 extant Findbugs warnings. {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 59s{color} | {color:green} master passed {color} | || || || || {color:brown} Patch Compile Tests {color} || | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 1m 29s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 1m 9s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 1m 9s{color} | {color:green} the patch passed {color} | | {color:red}-1{color} | {color:red} checkstyle {color} | {color:red} 0m 42s{color} | {color:red} ql: The patch generated 3 new + 227 unchanged - 1 fixed = 230 total (was 228) {color} | | {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 0s{color} | {color:green} The patch has no whitespace issues. {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 4m 22s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 1m 0s{color} | {color:green} the patch passed {color} | || || || || {color:brown} Other Tests {color} || | {color:green}+1{color} | {color:green} asflicense {color} | {color:green} 0m 13s{color} | {color:green} The patch does not generate ASF License warnings. {color} | | {color:black}{color} | {color:black} {color} | {color:black} 25m 4s{color} | {color:black} {color} | \\ \\ || Subsystem || Report/Notes || | Optional Tests | asflicense javac javadoc findbugs checkstyle compile | | uname | Linux hiveptest-server-upstream 3.16.0-4-amd64 #1 SMP Debian 3.16.36-1+deb8u1 (2016-09-03) x86_64 GNU/Linux | | Build tool | maven | | Personality | /data/hiveptest/working/yetus_PreCommit-HIVE-Build-13513/dev-support/hive-personality.sh | | git revision | master / d2b81a9 | | Default Java | 1.8.0_111 | | findbugs | v3.0.0 | | checkstyle | http://104.198.109.242/logs//PreCommit-HIVE-Build-13513/yetus/diff-checkstyle-ql.txt | | modules | C: ql U: ql | | Console output | http://104.198.109.242/logs//PreCommit-HIVE-Build-13513/yetus.txt | | Powered by | Apache Yetushttp://yetus.apache.org | This message was automatically generated. > CopyUtils used by REPL LOAD and EXPORT/IMPORT operations ignore distcp error. > - > > Key: HIVE-20476 > URL: https://issues.apache.org/jira/browse/HIVE-20476 > Project: Hive > Issue Type: Bug > Components: HiveServer2, repl >Affects Versions: 3.1.0 >Reporter: Sankar Hariappan >Assignee: Sankar Hariappan >Priority: Major > Labels: DR, Export, Import, Replication, pull-request-available > Attachments: HIVE-20476.01.patch > > > CopyUtils uses FileUtils.distCp to copy files but doesn't check the return > value. It returns false if the copy fails. > Now, REPL LOAD and EXPORT/IMPORT commands internally uses CopyUtils to copy > data files across clusters and here it may return success even if file copy > fails and may cause data loss. > Need to throw error and retry. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Assigned] (HIVE-20481) Add the Kafka Key record as part of the row.
[ https://issues.apache.org/jira/browse/HIVE-20481?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] slim bouguerra reassigned HIVE-20481: - > Add the Kafka Key record as part of the row. > > > Key: HIVE-20481 > URL: https://issues.apache.org/jira/browse/HIVE-20481 > Project: Hive > Issue Type: Sub-task >Reporter: slim bouguerra >Assignee: slim bouguerra >Priority: Major > > Kafka records are keyed, most of the case this key is null or used to route > records to the same partition. This patch adds this column as a binary column > {code} __record_key{code}. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (HIVE-20480) Implement column stats annotation rules for the UDTFOperator: Follow up for HIVE-20262
[ https://issues.apache.org/jira/browse/HIVE-20480?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16595588#comment-16595588 ] George Pachitariu commented on HIVE-20480: -- Hi [~ashutoshc], This is the follow-up task with a patch for your comment here: HIVE-20262 Is this what you meant? Can you also please give me an idea on how I could test this? Sorry for taking this long to come back to you. George :) > Implement column stats annotation rules for the UDTFOperator: Follow up for > HIVE-20262 > -- > > Key: HIVE-20480 > URL: https://issues.apache.org/jira/browse/HIVE-20480 > Project: Hive > Issue Type: Improvement > Components: Physical Optimizer >Reporter: George Pachitariu >Assignee: George Pachitariu >Priority: Minor > Attachments: HIVE-20480.patch > > > Implementing the rule for column stats: Follow-up task for > [HIVE-20262|http://example.com/] -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (HIVE-20480) Implement column stats annotation rules for the UDTFOperator: Follow up for HIVE-20262
[ https://issues.apache.org/jira/browse/HIVE-20480?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] George Pachitariu updated HIVE-20480: - Description: Implementing the rule for column stats: Follow-up task for [HIVE-20262|http://example.com/] (was: Implementing the rule for column stats: Follow up task for [HIVE-20262|http://example.com]) > Implement column stats annotation rules for the UDTFOperator: Follow up for > HIVE-20262 > -- > > Key: HIVE-20480 > URL: https://issues.apache.org/jira/browse/HIVE-20480 > Project: Hive > Issue Type: Improvement > Components: Physical Optimizer >Reporter: George Pachitariu >Assignee: George Pachitariu >Priority: Minor > Attachments: HIVE-20480.patch > > > Implementing the rule for column stats: Follow-up task for > [HIVE-20262|http://example.com/] -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (HIVE-20377) Hive Kafka Storage Handler
[ https://issues.apache.org/jira/browse/HIVE-20377?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16595586#comment-16595586 ] slim bouguerra commented on HIVE-20377: --- We have some hard coded timestamp records fields to ensure that q files are same across runs. [^HIVE-20377.18.patch] added a small fix to the test setup to avoid deleting old records by Kafka broker. {code}properties.setProperty("log.retention.hours", String.valueOf(Integer.MAX_VALUE));{code} > Hive Kafka Storage Handler > -- > > Key: HIVE-20377 > URL: https://issues.apache.org/jira/browse/HIVE-20377 > Project: Hive > Issue Type: New Feature >Affects Versions: 4.0.0 >Reporter: slim bouguerra >Assignee: slim bouguerra >Priority: Major > Attachments: HIVE-20377.10.patch, HIVE-20377.11.patch, > HIVE-20377.12.patch, HIVE-20377.15.patch, HIVE-20377.18.patch, > HIVE-20377.4.patch, HIVE-20377.5.patch, HIVE-20377.6.patch, > HIVE-20377.8.patch, HIVE-20377.8.patch, HIVE-20377.patch > > > h1. Goal > * Read streaming data form Kafka queue as an external table. > * Allow streaming navigation by pushing down filters on Kafka record > partition id, offset and timestamp. > * Insert streaming data form Kafka to an actual Hive internal table, using > CTAS statement. > h1. Example > h2. Create the external table > {code} > CREATE EXTERNAL TABLE kafka_table (`timestamp` timestamp, page string, `user` > string, language string, added int, deleted int, flags string,comment string, > namespace string) > STORED BY 'org.apache.hadoop.hive.kafka.KafkaStorageHandler' > TBLPROPERTIES > ("kafka.topic" = "wikipedia", > "kafka.bootstrap.servers"="brokeraddress:9092", > "kafka.serde.class"="org.apache.hadoop.hive.serde2.JsonSerDe"); > {code} > h2. Kafka Metadata > In order to keep track of Kafka records the storage handler will add > automatically the Kafka row metadata eg partition id, record offset and > record timestamp. > {code} > DESCRIBE EXTENDED kafka_table > timestamp timestamp from deserializer > page string from deserializer > user string from deserializer > language string from deserializer > country string from deserializer > continent string from deserializer > namespace string from deserializer > newpage boolean from deserializer > unpatrolled boolean from deserializer > anonymous boolean from deserializer > robot boolean from deserializer > added int from deserializer > deleted int from deserializer > delta bigint from deserializer > __partition int from deserializer > __offset bigint from deserializer > __timestamp bigint from deserializer > {code} > h2. Filter push down. > Newer Kafka consumers 0.11.0 and higher allow seeking on the stream based on > a given offset. The proposed storage handler will be able to leverage such > API by pushing down filters over metadata columns, namely __partition (int), > __offset(long) and __timestamp(long) > For instance Query like > {code} > select `__offset` from kafka_table where (`__offset` < 10 and `__offset`>3 > and `__partition` = 0) or (`__partition` = 0 and `__offset` < 105 and > `__offset` > 99) or (`__offset` = 109); > {code} > Will result on a scan of partition 0 only then read only records between > offset 4 and 109. > h2. With timestamp seeks > The seeking based on the internal timestamps allows the handler to run on > recently arrived data, by doing > {code} > select count(*) from kafka_table where `__timestamp` > 1000 * > to_unix_timestamp(CURRENT_TIMESTAMP - interval '20' hours) ; > {code} > This allows for implicit relationships between event timestamps and kafka > timestamps to be expressed in queries (i.e event_timestamp is always < than > kafka __timestamp and kafka __timestamp is never > 15 minutes from event etc). > h2. More examples with Avro > {code} > CREATE EXTERNAL TABLE wiki_kafka_avro_table > STORED BY 'org.apache.hadoop.hive.kafka.KafkaStorageHandler' > TBLPROPERTIES > ("kafka.topic" = "wiki_kafka_avro_table", > "kafka.bootstrap.servers"="localhost:9092", > "kafka.serde.class"="org.apache.hadoop.hive.serde2.avro.AvroSerDe", > 'avro.schema.literal'='{ > "type" : "record", > "name" : "Wikipedia", > "namespace" : "org.apache.hive.kafka", > "version": "1", > "fields" : [ { > "name" :
[jira] [Commented] (HIVE-17084) Turn on hive.stats.fetch.column.stats configuration flag
[ https://issues.apache.org/jira/browse/HIVE-17084?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16595583#comment-16595583 ] Hive QA commented on HIVE-17084: Here are the results of testing the latest attachment: https://issues.apache.org/jira/secure/attachment/12937453/HIVE-17084.08.patch {color:red}ERROR:{color} -1 due to build exiting with an error Test results: https://builds.apache.org/job/PreCommit-HIVE-Build/13512/testReport Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/13512/console Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-13512/ Messages: {noformat} Executing org.apache.hive.ptest.execution.TestCheckPhase Executing org.apache.hive.ptest.execution.PrepPhase Tests exited with: NonZeroExitCodeException Command 'bash /data/hiveptest/working/scratch/source-prep.sh' failed with exit status 1 and output '+ date '+%Y-%m-%d %T.%3N' 2018-08-28 20:49:18.722 + [[ -n /usr/lib/jvm/java-8-openjdk-amd64 ]] + export JAVA_HOME=/usr/lib/jvm/java-8-openjdk-amd64 + JAVA_HOME=/usr/lib/jvm/java-8-openjdk-amd64 + export PATH=/usr/lib/jvm/java-8-openjdk-amd64/bin/:/usr/local/bin:/usr/bin:/bin:/usr/local/games:/usr/games + PATH=/usr/lib/jvm/java-8-openjdk-amd64/bin/:/usr/local/bin:/usr/bin:/bin:/usr/local/games:/usr/games + export 'ANT_OPTS=-Xmx1g -XX:MaxPermSize=256m ' + ANT_OPTS='-Xmx1g -XX:MaxPermSize=256m ' + export 'MAVEN_OPTS=-Xmx1g ' + MAVEN_OPTS='-Xmx1g ' + cd /data/hiveptest/working/ + tee /data/hiveptest/logs/PreCommit-HIVE-Build-13512/source-prep.txt + [[ false == \t\r\u\e ]] + mkdir -p maven ivy + [[ git = \s\v\n ]] + [[ git = \g\i\t ]] + [[ -z master ]] + [[ -d apache-github-source-source ]] + [[ ! -d apache-github-source-source/.git ]] + [[ ! -d apache-github-source-source ]] + date '+%Y-%m-%d %T.%3N' 2018-08-28 20:49:18.726 + cd apache-github-source-source + git fetch origin >From https://github.com/apache/hive 7cf1d00..d2b81a9 master -> origin/master + git reset --hard HEAD HEAD is now at 7cf1d00 HIVE-20396: Test HS2 open_connection metrics (Laszlo Pinter via Peter Vary) + git clean -f -d Removing standalone-metastore/metastore-server/src/gen/ + git checkout master Already on 'master' Your branch is behind 'origin/master' by 1 commit, and can be fast-forwarded. (use "git pull" to update your local branch) + git reset --hard origin/master HEAD is now at d2b81a9 HIVE-19674: Group by Decimal Constants push down to Druid tables(Vineet Garg, reviewed by Ashutosh Chauhan) + git merge --ff-only origin/master Already up-to-date. + date '+%Y-%m-%d %T.%3N' 2018-08-28 20:49:20.487 + rm -rf ../yetus_PreCommit-HIVE-Build-13512 + mkdir ../yetus_PreCommit-HIVE-Build-13512 + git gc + cp -R . ../yetus_PreCommit-HIVE-Build-13512 + mkdir /data/hiveptest/logs/PreCommit-HIVE-Build-13512/yetus + patchCommandPath=/data/hiveptest/working/scratch/smart-apply-patch.sh + patchFilePath=/data/hiveptest/working/scratch/build.patch + [[ -f /data/hiveptest/working/scratch/build.patch ]] + chmod +x /data/hiveptest/working/scratch/smart-apply-patch.sh + /data/hiveptest/working/scratch/smart-apply-patch.sh /data/hiveptest/working/scratch/build.patch error: patch failed: ql/src/test/results/clientpositive/list_bucket_query_oneskew_2.q.out:533 Falling back to three-way merge... Applied patch to 'ql/src/test/results/clientpositive/list_bucket_query_oneskew_2.q.out' with conflicts. error: patch failed: ql/src/test/results/clientpositive/perf/spark/query44.q.out:69 Falling back to three-way merge... Applied patch to 'ql/src/test/results/clientpositive/perf/spark/query44.q.out' with conflicts. error: patch failed: ql/src/test/results/clientpositive/perf/tez/query44.q.out:86 Falling back to three-way merge... Applied patch to 'ql/src/test/results/clientpositive/perf/tez/query44.q.out' with conflicts. error: patch failed: ql/src/test/results/clientpositive/semijoin4.q.out:235 Falling back to three-way merge... Applied patch to 'ql/src/test/results/clientpositive/semijoin4.q.out' with conflicts. error: patch failed: ql/src/test/results/clientpositive/spark/spark_dynamic_partition_pruning.q.out:2514 Falling back to three-way merge... Applied patch to 'ql/src/test/results/clientpositive/spark/spark_dynamic_partition_pruning.q.out' with conflicts. error: patch failed: ql/src/test/results/clientpositive/spark/spark_vectorized_dynamic_partition_pruning.q.out:4811 Falling back to three-way merge... Applied patch to 'ql/src/test/results/clientpositive/spark/spark_vectorized_dynamic_partition_pruning.q.out' cleanly. error: patch failed: ql/src/test/results/clientpositive/spark/subquery_in.q.out:2636 Falling back to three-way merge... Applied patch to 'ql/src/test/results/clientpositive/spark/subquery_in.q.out' with conflicts. error: patch failed: ql/src/test/results/clientpositive/spark/subquery_notin.q.out:3600 Falling back to three-way merge... Applied patch to
[jira] [Updated] (HIVE-20480) Implement column stats annotation rules for the UDTFOperator: Follow up for HIVE-20262
[ https://issues.apache.org/jira/browse/HIVE-20480?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] George Pachitariu updated HIVE-20480: - Attachment: HIVE-20480.patch Status: Patch Available (was: Open) > Implement column stats annotation rules for the UDTFOperator: Follow up for > HIVE-20262 > -- > > Key: HIVE-20480 > URL: https://issues.apache.org/jira/browse/HIVE-20480 > Project: Hive > Issue Type: Improvement > Components: Physical Optimizer >Reporter: George Pachitariu >Assignee: George Pachitariu >Priority: Minor > Attachments: HIVE-20480.patch > > > > Implementing the rule for column stats: Follow up task for > [HIVE-20262|http://example.com] -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (HIVE-20377) Hive Kafka Storage Handler
[ https://issues.apache.org/jira/browse/HIVE-20377?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] slim bouguerra updated HIVE-20377: -- Attachment: HIVE-20377.18.patch > Hive Kafka Storage Handler > -- > > Key: HIVE-20377 > URL: https://issues.apache.org/jira/browse/HIVE-20377 > Project: Hive > Issue Type: New Feature >Affects Versions: 4.0.0 >Reporter: slim bouguerra >Assignee: slim bouguerra >Priority: Major > Attachments: HIVE-20377.10.patch, HIVE-20377.11.patch, > HIVE-20377.12.patch, HIVE-20377.15.patch, HIVE-20377.18.patch, > HIVE-20377.4.patch, HIVE-20377.5.patch, HIVE-20377.6.patch, > HIVE-20377.8.patch, HIVE-20377.8.patch, HIVE-20377.patch > > > h1. Goal > * Read streaming data form Kafka queue as an external table. > * Allow streaming navigation by pushing down filters on Kafka record > partition id, offset and timestamp. > * Insert streaming data form Kafka to an actual Hive internal table, using > CTAS statement. > h1. Example > h2. Create the external table > {code} > CREATE EXTERNAL TABLE kafka_table (`timestamp` timestamp, page string, `user` > string, language string, added int, deleted int, flags string,comment string, > namespace string) > STORED BY 'org.apache.hadoop.hive.kafka.KafkaStorageHandler' > TBLPROPERTIES > ("kafka.topic" = "wikipedia", > "kafka.bootstrap.servers"="brokeraddress:9092", > "kafka.serde.class"="org.apache.hadoop.hive.serde2.JsonSerDe"); > {code} > h2. Kafka Metadata > In order to keep track of Kafka records the storage handler will add > automatically the Kafka row metadata eg partition id, record offset and > record timestamp. > {code} > DESCRIBE EXTENDED kafka_table > timestamp timestamp from deserializer > page string from deserializer > user string from deserializer > language string from deserializer > country string from deserializer > continent string from deserializer > namespace string from deserializer > newpage boolean from deserializer > unpatrolled boolean from deserializer > anonymous boolean from deserializer > robot boolean from deserializer > added int from deserializer > deleted int from deserializer > delta bigint from deserializer > __partition int from deserializer > __offset bigint from deserializer > __timestamp bigint from deserializer > {code} > h2. Filter push down. > Newer Kafka consumers 0.11.0 and higher allow seeking on the stream based on > a given offset. The proposed storage handler will be able to leverage such > API by pushing down filters over metadata columns, namely __partition (int), > __offset(long) and __timestamp(long) > For instance Query like > {code} > select `__offset` from kafka_table where (`__offset` < 10 and `__offset`>3 > and `__partition` = 0) or (`__partition` = 0 and `__offset` < 105 and > `__offset` > 99) or (`__offset` = 109); > {code} > Will result on a scan of partition 0 only then read only records between > offset 4 and 109. > h2. With timestamp seeks > The seeking based on the internal timestamps allows the handler to run on > recently arrived data, by doing > {code} > select count(*) from kafka_table where `__timestamp` > 1000 * > to_unix_timestamp(CURRENT_TIMESTAMP - interval '20' hours) ; > {code} > This allows for implicit relationships between event timestamps and kafka > timestamps to be expressed in queries (i.e event_timestamp is always < than > kafka __timestamp and kafka __timestamp is never > 15 minutes from event etc). > h2. More examples with Avro > {code} > CREATE EXTERNAL TABLE wiki_kafka_avro_table > STORED BY 'org.apache.hadoop.hive.kafka.KafkaStorageHandler' > TBLPROPERTIES > ("kafka.topic" = "wiki_kafka_avro_table", > "kafka.bootstrap.servers"="localhost:9092", > "kafka.serde.class"="org.apache.hadoop.hive.serde2.avro.AvroSerDe", > 'avro.schema.literal'='{ > "type" : "record", > "name" : "Wikipedia", > "namespace" : "org.apache.hive.kafka", > "version": "1", > "fields" : [ { > "name" : "isrobot", > "type" : "boolean" > }, { > "name" : "channel", > "type" : "string" > }, { > "name" : "timestamp", > "type" : "string" > }, { > "name" : "flags", > "type" : "string" > }, { > "name" : "isunpatrolled", > "type" : "boolean" > }, { > "name" :
[jira] [Commented] (HIVE-13457) Create HS2 REST API endpoints for monitoring information
[ https://issues.apache.org/jira/browse/HIVE-13457?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16595576#comment-16595576 ] Hive QA commented on HIVE-13457: Here are the results of testing the latest attachment: https://issues.apache.org/jira/secure/attachment/12937447/HIVE-13457.5.patch {color:green}SUCCESS:{color} +1 due to 1 test(s) being added or modified. {color:green}SUCCESS:{color} +1 due to 14903 tests passed Test results: https://builds.apache.org/job/PreCommit-HIVE-Build/13511/testReport Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/13511/console Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-13511/ Messages: {noformat} Executing org.apache.hive.ptest.execution.TestCheckPhase Executing org.apache.hive.ptest.execution.PrepPhase Executing org.apache.hive.ptest.execution.YetusPhase Executing org.apache.hive.ptest.execution.ExecutionPhase Executing org.apache.hive.ptest.execution.ReportingPhase {noformat} This message is automatically generated. ATTACHMENT ID: 12937447 - PreCommit-HIVE-Build > Create HS2 REST API endpoints for monitoring information > > > Key: HIVE-13457 > URL: https://issues.apache.org/jira/browse/HIVE-13457 > Project: Hive > Issue Type: Improvement >Reporter: Szehon Ho >Assignee: Pawel Szostek >Priority: Major > Attachments: HIVE-13457.3.patch, HIVE-13457.4.patch, > HIVE-13457.5.patch, HIVE-13457.patch, HIVE-13457.patch > > > Similar to what is exposed in HS2 webui in HIVE-12338, it would be nice if > other UI's like admin tools or Hue can access and display this information as > well. Hence, we will create some REST endpoints to expose this information. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Assigned] (HIVE-20480) Implement column stats annotation rules for the UDTFOperator: Follow up for HIVE-20262
[ https://issues.apache.org/jira/browse/HIVE-20480?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] George Pachitariu reassigned HIVE-20480: > Implement column stats annotation rules for the UDTFOperator: Follow up for > HIVE-20262 > -- > > Key: HIVE-20480 > URL: https://issues.apache.org/jira/browse/HIVE-20480 > Project: Hive > Issue Type: Improvement > Components: Physical Optimizer >Reporter: George Pachitariu >Assignee: George Pachitariu >Priority: Minor > > > Implementing the rule for column stats: Follow up task for > [HIVE-20262|http://example.com] -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Assigned] (HIVE-19318) Improve Hive logging
[ https://issues.apache.org/jira/browse/HIVE-19318?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Aihua Xu reassigned HIVE-19318: --- Assignee: Alice Fan (was: Aihua Xu) > Improve Hive logging > > > Key: HIVE-19318 > URL: https://issues.apache.org/jira/browse/HIVE-19318 > Project: Hive > Issue Type: Improvement > Components: Logging >Affects Versions: 3.0.0 >Reporter: Aihua Xu >Assignee: Alice Fan >Priority: Major > > Use this jira to track some potential improvements on hive logging. What I > have noticed that some log entries may have incorrect log level, or may not > show in the correct places, e.g., some logs are printing to the STDERR/STDOUT > rather than the HS2 log file. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (HIVE-18725) Improve error handling for subqueries if there is wrong column reference
[ https://issues.apache.org/jira/browse/HIVE-18725?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16595528#comment-16595528 ] Vineet Garg commented on HIVE-18725: [~ikryvenko] Sorry I couldn't take a look the patch before. I will review it now and will shortly update. > Improve error handling for subqueries if there is wrong column reference > > > Key: HIVE-18725 > URL: https://issues.apache.org/jira/browse/HIVE-18725 > Project: Hive > Issue Type: Improvement > Components: Query Planning >Reporter: Vineet Garg >Assignee: Igor Kryvenko >Priority: Major > Attachments: HIVE-18725.01.patch, HIVE-18725.02.patch, > HIVE-18725.03.patch, HIVE-18725.04.patch, HIVE-18725.05.patch, > HIVE-18725.06.patch, HIVE-18725.07.patch, HIVE-18725.08.patch, > HIVE-18725.09.patch, HIVE-18725.10.patch > > > If there is a column reference within subquery which doesn't exist Hive > throws misleading error message. > e.g. > {code:sql} > select * from table1 where table1.col1 IN (select col2 from table2 where > table2.col1=table1.non_existing_column) and table1.col1 IN (select 4); > {code} > The above query, assuming table1 doesn't have non_existing_column, will throw > following misleading error: > {noformat} > FAILED: SemanticException Line 0:-1 Unsupported SubQuery Expression 'col1': > Only 1 SubQuery expression is supported. > {noformat} -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (HIVE-18725) Improve error handling for subqueries if there is wrong column reference
[ https://issues.apache.org/jira/browse/HIVE-18725?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16595523#comment-16595523 ] Igor Kryvenko commented on HIVE-18725: -- [~vgarg] Hi Vineet. Any updates on this issue? > Improve error handling for subqueries if there is wrong column reference > > > Key: HIVE-18725 > URL: https://issues.apache.org/jira/browse/HIVE-18725 > Project: Hive > Issue Type: Improvement > Components: Query Planning >Reporter: Vineet Garg >Assignee: Igor Kryvenko >Priority: Major > Attachments: HIVE-18725.01.patch, HIVE-18725.02.patch, > HIVE-18725.03.patch, HIVE-18725.04.patch, HIVE-18725.05.patch, > HIVE-18725.06.patch, HIVE-18725.07.patch, HIVE-18725.08.patch, > HIVE-18725.09.patch, HIVE-18725.10.patch > > > If there is a column reference within subquery which doesn't exist Hive > throws misleading error message. > e.g. > {code:sql} > select * from table1 where table1.col1 IN (select col2 from table2 where > table2.col1=table1.non_existing_column) and table1.col1 IN (select 4); > {code} > The above query, assuming table1 doesn't have non_existing_column, will throw > following misleading error: > {noformat} > FAILED: SemanticException Line 0:-1 Unsupported SubQuery Expression 'col1': > Only 1 SubQuery expression is supported. > {noformat} -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (HIVE-13457) Create HS2 REST API endpoints for monitoring information
[ https://issues.apache.org/jira/browse/HIVE-13457?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16595519#comment-16595519 ] Hive QA commented on HIVE-13457: | (x) *{color:red}-1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || || || || || {color:brown} Prechecks {color} || | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s{color} | {color:green} The patch does not contain any @author tags. {color} | || || || || {color:brown} master Compile Tests {color} || | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 8m 58s{color} | {color:green} master passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 26s{color} | {color:green} master passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 14s{color} | {color:green} master passed {color} | | {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue} 0m 41s{color} | {color:blue} service in master has 48 extant Findbugs warnings. {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 19s{color} | {color:green} master passed {color} | || || || || {color:brown} Patch Compile Tests {color} || | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 0m 24s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 27s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 0m 27s{color} | {color:green} the patch passed {color} | | {color:red}-1{color} | {color:red} checkstyle {color} | {color:red} 0m 15s{color} | {color:red} service: The patch generated 130 new + 36 unchanged - 0 fixed = 166 total (was 36) {color} | | {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 0s{color} | {color:green} The patch has no whitespace issues. {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 0m 48s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 18s{color} | {color:green} the patch passed {color} | || || || || {color:brown} Other Tests {color} || | {color:red}-1{color} | {color:red} asflicense {color} | {color:red} 0m 14s{color} | {color:red} The patch generated 1 ASF License warnings. {color} | | {color:black}{color} | {color:black} {color} | {color:black} 13m 25s{color} | {color:black} {color} | \\ \\ || Subsystem || Report/Notes || | Optional Tests | asflicense javac javadoc findbugs checkstyle compile | | uname | Linux hiveptest-server-upstream 3.16.0-4-amd64 #1 SMP Debian 3.16.36-1+deb8u1 (2016-09-03) x86_64 GNU/Linux | | Build tool | maven | | Personality | /data/hiveptest/working/yetus_PreCommit-HIVE-Build-13511/dev-support/hive-personality.sh | | git revision | master / 7cf1d00 | | Default Java | 1.8.0_111 | | findbugs | v3.0.0 | | checkstyle | http://104.198.109.242/logs//PreCommit-HIVE-Build-13511/yetus/diff-checkstyle-service.txt | | asflicense | http://104.198.109.242/logs//PreCommit-HIVE-Build-13511/yetus/patch-asflicense-problems.txt | | modules | C: service U: service | | Console output | http://104.198.109.242/logs//PreCommit-HIVE-Build-13511/yetus.txt | | Powered by | Apache Yetushttp://yetus.apache.org | This message was automatically generated. > Create HS2 REST API endpoints for monitoring information > > > Key: HIVE-13457 > URL: https://issues.apache.org/jira/browse/HIVE-13457 > Project: Hive > Issue Type: Improvement >Reporter: Szehon Ho >Assignee: Pawel Szostek >Priority: Major > Attachments: HIVE-13457.3.patch, HIVE-13457.4.patch, > HIVE-13457.5.patch, HIVE-13457.patch, HIVE-13457.patch > > > Similar to what is exposed in HS2 webui in HIVE-12338, it would be nice if > other UI's like admin tools or Hue can access and display this information as > well. Hence, we will create some REST endpoints to expose this information. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (HIVE-19674) Group by Decimal Constants push down to Druid tables.
[ https://issues.apache.org/jira/browse/HIVE-19674?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Vineet Garg updated HIVE-19674: --- Resolution: Fixed Fix Version/s: 4.0.0 Status: Resolved (was: Patch Available) Pushed to master. > Group by Decimal Constants push down to Druid tables. > - > > Key: HIVE-19674 > URL: https://issues.apache.org/jira/browse/HIVE-19674 > Project: Hive > Issue Type: Improvement > Components: Druid integration >Reporter: slim bouguerra >Assignee: Vineet Garg >Priority: Major > Fix For: 4.0.0 > > Attachments: HIVE-19674.1.patch, HIVE-19674.2.patch, > HIVE-19674.3.patch, HIVE-19674.4.patch, HIVE-19674.5.patch, HIVE-19674.6.patch > > > Queries like following gets generated by Tableau. > {code} > SELECT SUM(`ssb_druid_100`.`lo_revenue`) AS `sum_lo_revenue_ok` > FROM `druid_ssb`.`ssb_druid_100` `ssb_druid_100` > GROUP BY 1.1001; > {code} -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (HIVE-20363) Use integer constants for frequently used serde classes.
[ https://issues.apache.org/jira/browse/HIVE-20363?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Anurag Mantripragada updated HIVE-20363: Description: Serde libraries are stored as fully qualified class names which are long strings, we can get improvements in I/O and storage if we store integer constants for frequently used serde classes in the backend DB. For example: {code:java} org.apache.hadoop.hive.serde2.avro.AvroSerDe"---> 1 parquet.hive.serde.ParquetHiveSerDe" ---> 2 ...{code} Review the patch at: https://reviews.apache.org/r/68546/ was: Serde libraries are stored as fully qualified class names which are long strings, we can get improvements in I/O and storage if we store integer constants for frequently used serde classes in the backend DB. For example: {code:java} org.apache.hadoop.hive.serde2.avro.AvroSerDe"---> 1 parquet.hive.serde.ParquetHiveSerDe" ---> 2 ...{code} Review the patch at: https://issues.apache.org/jira/browse/HIVE-20363 > Use integer constants for frequently used serde classes. > > > Key: HIVE-20363 > URL: https://issues.apache.org/jira/browse/HIVE-20363 > Project: Hive > Issue Type: Improvement > Components: Metastore, Standalone Metastore >Reporter: Anurag Mantripragada >Assignee: Anurag Mantripragada >Priority: Major > > Serde libraries are stored as fully qualified class names which are long > strings, we can get improvements in I/O and storage if we store integer > constants for frequently used serde classes in the backend DB. > For example: > {code:java} > org.apache.hadoop.hive.serde2.avro.AvroSerDe"---> 1 > parquet.hive.serde.ParquetHiveSerDe" ---> 2 > ...{code} > Review the patch at: > https://reviews.apache.org/r/68546/ -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (HIVE-20363) Use integer constants for frequently used serde classes.
[ https://issues.apache.org/jira/browse/HIVE-20363?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Anurag Mantripragada updated HIVE-20363: Description: Serde libraries are stored as fully qualified class names which are long strings, we can get improvements in I/O and storage if we store integer constants for frequently used serde classes in the backend DB. For example: {code:java} org.apache.hadoop.hive.serde2.avro.AvroSerDe"---> 1 parquet.hive.serde.ParquetHiveSerDe" ---> 2 ...{code} Review the patch at: https://issues.apache.org/jira/browse/HIVE-20363 was: Serde libraries are stored as fully qualified class names which are long strings, we can get improvements in I/O and storage if we store integer constants for frequently used serde classes in the backend DB. For example: {code:java} org.apache.hadoop.hive.serde2.avro.AvroSerDe"---> 1 parquet.hive.serde.ParquetHiveSerDe" ---> 2 ...{code} > Use integer constants for frequently used serde classes. > > > Key: HIVE-20363 > URL: https://issues.apache.org/jira/browse/HIVE-20363 > Project: Hive > Issue Type: Improvement > Components: Metastore, Standalone Metastore >Reporter: Anurag Mantripragada >Assignee: Anurag Mantripragada >Priority: Major > > Serde libraries are stored as fully qualified class names which are long > strings, we can get improvements in I/O and storage if we store integer > constants for frequently used serde classes in the backend DB. > For example: > {code:java} > org.apache.hadoop.hive.serde2.avro.AvroSerDe"---> 1 > parquet.hive.serde.ParquetHiveSerDe" ---> 2 > ...{code} > Review the patch at: > https://issues.apache.org/jira/browse/HIVE-20363 -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Resolved] (HIVE-19585) Add UNKNOWN to PrincipalType
[ https://issues.apache.org/jira/browse/HIVE-19585?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Arjun Mishra resolved HIVE-19585. - Resolution: Not A Problem > Add UNKNOWN to PrincipalType > > > Key: HIVE-19585 > URL: https://issues.apache.org/jira/browse/HIVE-19585 > Project: Hive > Issue Type: Bug > Components: Hive >Affects Versions: 2.3.2 >Reporter: Arjun Mishra >Assignee: Arjun Mishra >Priority: Major > Fix For: 2.3.2 > > Attachments: HIVE-198585.01.patch > > > We need to include type UNKNOWN to PrincipalType to match with > HivePrincipal.HivePrincipalType.UKNOWN -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (HIVE-20205) Upgrade HBase dependencies off alpha4 release
[ https://issues.apache.org/jira/browse/HIVE-20205?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16595438#comment-16595438 ] Hive QA commented on HIVE-20205: Here are the results of testing the latest attachment: https://issues.apache.org/jira/secure/attachment/12937440/HIVE-20205.2.patch {color:red}ERROR:{color} -1 due to no test(s) being added or modified. {color:red}ERROR:{color} -1 due to 8 failed/errored test(s), 14868 tests executed *Failed tests:* {noformat} org.apache.hadoop.hive.cli.TestHBaseCliDriver.org.apache.hadoop.hive.cli.TestHBaseCliDriver (batchId=101) org.apache.hadoop.hive.cli.TestHBaseCliDriver.org.apache.hadoop.hive.cli.TestHBaseCliDriver (batchId=102) org.apache.hadoop.hive.cli.TestHBaseCliDriver.org.apache.hadoop.hive.cli.TestHBaseCliDriver (batchId=103) org.apache.hadoop.hive.cli.TestHBaseCliDriver.org.apache.hadoop.hive.cli.TestHBaseCliDriver (batchId=104) org.apache.hadoop.hive.cli.TestHBaseCliDriver.org.apache.hadoop.hive.cli.TestHBaseCliDriver (batchId=105) org.apache.hadoop.hive.cli.TestHBaseCliDriver.org.apache.hadoop.hive.cli.TestHBaseCliDriver (batchId=106) org.apache.hadoop.hive.cli.TestHBaseNegativeCliDriver.org.apache.hadoop.hive.cli.TestHBaseNegativeCliDriver (batchId=262) org.apache.hive.hcatalog.hbase.TestPigHBaseStorageHandler.org.apache.hive.hcatalog.hbase.TestPigHBaseStorageHandler (batchId=258) {noformat} Test results: https://builds.apache.org/job/PreCommit-HIVE-Build/13510/testReport Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/13510/console Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-13510/ Messages: {noformat} Executing org.apache.hive.ptest.execution.TestCheckPhase Executing org.apache.hive.ptest.execution.PrepPhase Executing org.apache.hive.ptest.execution.YetusPhase Executing org.apache.hive.ptest.execution.ExecutionPhase Executing org.apache.hive.ptest.execution.ReportingPhase Tests exited with: TestsFailedException: 8 tests failed {noformat} This message is automatically generated. ATTACHMENT ID: 12937440 - PreCommit-HIVE-Build > Upgrade HBase dependencies off alpha4 release > - > > Key: HIVE-20205 > URL: https://issues.apache.org/jira/browse/HIVE-20205 > Project: Hive > Issue Type: Bug > Components: Hive >Affects Versions: 3.0.0 >Reporter: Naveen Gangam >Assignee: Naveen Gangam >Priority: Minor > Attachments: HIVE-20205.1.patch, HIVE-20205.1.patch, > HIVE-20205.2.patch, HIVE-20205.2.patch, HIVE-20205.patch, HIVE-20205.patch > > > Appears Hive has dependencies on hbase 2.0.0-alpha4 releases. HBase 2.0.0 and > 2.0.1 have been released. HBase team recommends 2.0.1 and says there shouldnt > be any API surprises. (but we never know) -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (HIVE-20205) Upgrade HBase dependencies off alpha4 release
[ https://issues.apache.org/jira/browse/HIVE-20205?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16595392#comment-16595392 ] Hive QA commented on HIVE-20205: | (/) *{color:green}+1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || || || || || {color:brown} Prechecks {color} || | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s{color} | {color:green} The patch does not contain any @author tags. {color} | || || || || {color:brown} master Compile Tests {color} || | {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue} 0m 40s{color} | {color:blue} Maven dependency ordering for branch {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 8m 5s{color} | {color:green} master passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 7m 0s{color} | {color:green} master passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 7m 20s{color} | {color:green} master passed {color} | || || || || {color:brown} Patch Compile Tests {color} || | {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue} 0m 7s{color} | {color:blue} Maven dependency ordering for patch {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 7m 21s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 6m 46s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 6m 46s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 0s{color} | {color:green} The patch has no whitespace issues. {color} | | {color:green}+1{color} | {color:green} xml {color} | {color:green} 0m 2s{color} | {color:green} The patch has no ill-formed XML file. {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 7m 20s{color} | {color:green} the patch passed {color} | || || || || {color:brown} Other Tests {color} || | {color:green}+1{color} | {color:green} asflicense {color} | {color:green} 0m 13s{color} | {color:green} The patch does not generate ASF License warnings. {color} | | {color:black}{color} | {color:black} {color} | {color:black} 45m 19s{color} | {color:black} {color} | \\ \\ || Subsystem || Report/Notes || | Optional Tests | asflicense javac javadoc xml compile | | uname | Linux hiveptest-server-upstream 3.16.0-4-amd64 #1 SMP Debian 3.16.36-1+deb8u1 (2016-09-03) x86_64 GNU/Linux | | Build tool | maven | | Personality | /data/hiveptest/working/yetus_PreCommit-HIVE-Build-13510/dev-support/hive-personality.sh | | git revision | master / 7cf1d00 | | Default Java | 1.8.0_111 | | modules | C: . itests/util U: . | | Console output | http://104.198.109.242/logs//PreCommit-HIVE-Build-13510/yetus.txt | | Powered by | Apache Yetushttp://yetus.apache.org | This message was automatically generated. > Upgrade HBase dependencies off alpha4 release > - > > Key: HIVE-20205 > URL: https://issues.apache.org/jira/browse/HIVE-20205 > Project: Hive > Issue Type: Bug > Components: Hive >Affects Versions: 3.0.0 >Reporter: Naveen Gangam >Assignee: Naveen Gangam >Priority: Minor > Attachments: HIVE-20205.1.patch, HIVE-20205.1.patch, > HIVE-20205.2.patch, HIVE-20205.2.patch, HIVE-20205.patch, HIVE-20205.patch > > > Appears Hive has dependencies on hbase 2.0.0-alpha4 releases. HBase 2.0.0 and > 2.0.1 have been released. HBase team recommends 2.0.1 and says there shouldnt > be any API surprises. (but we never know) -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Resolved] (HIVE-20479) Update content/people.mdtext in cms
[ https://issues.apache.org/jira/browse/HIVE-20479?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Andrew Sherman resolved HIVE-20479. --- Resolution: Fixed Done already > Update content/people.mdtext in cms > > > Key: HIVE-20479 > URL: https://issues.apache.org/jira/browse/HIVE-20479 > Project: Hive > Issue Type: Task >Reporter: Andrew Sherman >Assignee: Andrew Sherman >Priority: Major > > I added myself to the committers list. > > {code:java} > asherman > Andrew Sherman > href="http://cloudera.com/;>Cloudera > > > {code} -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (HIVE-20479) Update content/people.mdtext in cms
[ https://issues.apache.org/jira/browse/HIVE-20479?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16595364#comment-16595364 ] Sahil Takiar commented on HIVE-20479: - +1 > Update content/people.mdtext in cms > > > Key: HIVE-20479 > URL: https://issues.apache.org/jira/browse/HIVE-20479 > Project: Hive > Issue Type: Task >Reporter: Andrew Sherman >Assignee: Andrew Sherman >Priority: Major > > I added myself to the committers list. > > {code:java} > asherman > Andrew Sherman > href="http://cloudera.com/;>Cloudera > > > {code} -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Assigned] (HIVE-20479) Update content/people.mdtext in cms
[ https://issues.apache.org/jira/browse/HIVE-20479?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Andrew Sherman reassigned HIVE-20479: - > Update content/people.mdtext in cms > > > Key: HIVE-20479 > URL: https://issues.apache.org/jira/browse/HIVE-20479 > Project: Hive > Issue Type: Task >Reporter: Andrew Sherman >Assignee: Andrew Sherman >Priority: Major > > I added myself to the committers list. > > {code:java} > asherman > Andrew Sherman > href="http://cloudera.com/;>Cloudera > > > {code} -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (HIVE-20466) Improve org.apache.hadoop.hive.ql.exec.FunctionTask Experience
[ https://issues.apache.org/jira/browse/HIVE-20466?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16595330#comment-16595330 ] Hive QA commented on HIVE-20466: Here are the results of testing the latest attachment: https://issues.apache.org/jira/secure/attachment/12937444/HIVE-20466.4.patch {color:red}ERROR:{color} -1 due to no test(s) being added or modified. {color:green}SUCCESS:{color} +1 due to 14899 tests passed Test results: https://builds.apache.org/job/PreCommit-HIVE-Build/13509/testReport Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/13509/console Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-13509/ Messages: {noformat} Executing org.apache.hive.ptest.execution.TestCheckPhase Executing org.apache.hive.ptest.execution.PrepPhase Executing org.apache.hive.ptest.execution.YetusPhase Executing org.apache.hive.ptest.execution.ExecutionPhase Executing org.apache.hive.ptest.execution.ReportingPhase {noformat} This message is automatically generated. ATTACHMENT ID: 12937444 - PreCommit-HIVE-Build > Improve org.apache.hadoop.hive.ql.exec.FunctionTask Experience > -- > > Key: HIVE-20466 > URL: https://issues.apache.org/jira/browse/HIVE-20466 > Project: Hive > Issue Type: Improvement > Components: Hive >Reporter: denys kuzmenko >Assignee: denys kuzmenko >Priority: Major > Attachments: HIVE-20466.1.patch, HIVE-20466.2.patch, > HIVE-20466.3.patch, HIVE-20466.4.patch > > -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (HIVE-20472) mvn test failing for metastore-tool module
[ https://issues.apache.org/jira/browse/HIVE-20472?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Alexander Kolbasov updated HIVE-20472: -- Attachment: HIVE-20472.02.patch > mvn test failing for metastore-tool module > -- > > Key: HIVE-20472 > URL: https://issues.apache.org/jira/browse/HIVE-20472 > Project: Hive > Issue Type: Bug >Affects Versions: 4.0.0 >Reporter: Vihang Karajgaonkar >Assignee: Alexander Kolbasov >Priority: Trivial > Fix For: 4.0.0 > > Attachments: HIVE-20472.01.patch, HIVE-20472.02.patch > > > Fails because there are no applicable tests. > > {code} > [ERROR] Failed to execute goal > org.apache.maven.plugins:maven-surefire-plugin:2.20.1:test (default-test) on > project hive-metastore-benchmarks: No tests were executed! (Set > -DfailIfNoTests=false to ignore this error.) -> [Help 1] > [ERROR] > [ERROR] To see the full stack trace of the errors, re-run Maven with the -e > switch. > [ERROR] Re-run Maven using the -X switch to enable full debug logging. > [ERROR] > [ERROR] For more information about the errors and possible solutions, please > read the following articles: > [ERROR] [Help 1] > http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException > [ERROR] > [ERROR] After correcting the problems, you can resume the build with the > command > [ERROR] mvn -rf :hive-metastore-benchmarks > {code} -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Assigned] (HIVE-20472) mvn test failing for metastore-tool module
[ https://issues.apache.org/jira/browse/HIVE-20472?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Alexander Kolbasov reassigned HIVE-20472: - Assignee: Alexander Kolbasov (was: Vihang Karajgaonkar) > mvn test failing for metastore-tool module > -- > > Key: HIVE-20472 > URL: https://issues.apache.org/jira/browse/HIVE-20472 > Project: Hive > Issue Type: Bug >Affects Versions: 4.0.0 >Reporter: Vihang Karajgaonkar >Assignee: Alexander Kolbasov >Priority: Trivial > Fix For: 4.0.0 > > Attachments: HIVE-20472.01.patch > > > Fails because there are no applicable tests. > > {code} > [ERROR] Failed to execute goal > org.apache.maven.plugins:maven-surefire-plugin:2.20.1:test (default-test) on > project hive-metastore-benchmarks: No tests were executed! (Set > -DfailIfNoTests=false to ignore this error.) -> [Help 1] > [ERROR] > [ERROR] To see the full stack trace of the errors, re-run Maven with the -e > switch. > [ERROR] Re-run Maven using the -X switch to enable full debug logging. > [ERROR] > [ERROR] For more information about the errors and possible solutions, please > read the following articles: > [ERROR] [Help 1] > http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException > [ERROR] > [ERROR] After correcting the problems, you can resume the build with the > command > [ERROR] mvn -rf :hive-metastore-benchmarks > {code} -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (HIVE-20466) Improve org.apache.hadoop.hive.ql.exec.FunctionTask Experience
[ https://issues.apache.org/jira/browse/HIVE-20466?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16595279#comment-16595279 ] Hive QA commented on HIVE-20466: | (x) *{color:red}-1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || || || || || {color:brown} Prechecks {color} || | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s{color} | {color:green} The patch does not contain any @author tags. {color} | || || || || {color:brown} master Compile Tests {color} || | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 7m 48s{color} | {color:green} master passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 1m 11s{color} | {color:green} master passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 42s{color} | {color:green} master passed {color} | | {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue} 4m 3s{color} | {color:blue} ql in master has 2310 extant Findbugs warnings. {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 1m 0s{color} | {color:green} master passed {color} | || || || || {color:brown} Patch Compile Tests {color} || | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 1m 25s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 1m 5s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 1m 5s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 40s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 0s{color} | {color:green} The patch has no whitespace issues. {color} | | {color:red}-1{color} | {color:red} findbugs {color} | {color:red} 4m 15s{color} | {color:red} ql generated 1 new + 2310 unchanged - 0 fixed = 2311 total (was 2310) {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 1m 1s{color} | {color:green} the patch passed {color} | || || || || {color:brown} Other Tests {color} || | {color:green}+1{color} | {color:green} asflicense {color} | {color:green} 0m 14s{color} | {color:green} The patch does not generate ASF License warnings. {color} | | {color:black}{color} | {color:black} {color} | {color:black} 23m 57s{color} | {color:black} {color} | \\ \\ || Reason || Tests || | FindBugs | module:ql | | | Exception is caught when Exception is not thrown in org.apache.hadoop.hive.ql.metadata.Hive.reloadFunctions() At Hive.java:is not thrown in org.apache.hadoop.hive.ql.metadata.Hive.reloadFunctions() At Hive.java:[line 246] | \\ \\ || Subsystem || Report/Notes || | Optional Tests | asflicense javac javadoc findbugs checkstyle compile | | uname | Linux hiveptest-server-upstream 3.16.0-4-amd64 #1 SMP Debian 3.16.36-1+deb8u1 (2016-09-03) x86_64 GNU/Linux | | Build tool | maven | | Personality | /data/hiveptest/working/yetus_PreCommit-HIVE-Build-13509/dev-support/hive-personality.sh | | git revision | master / 7cf1d00 | | Default Java | 1.8.0_111 | | findbugs | v3.0.0 | | findbugs | http://104.198.109.242/logs//PreCommit-HIVE-Build-13509/yetus/new-findbugs-ql.html | | modules | C: ql U: ql | | Console output | http://104.198.109.242/logs//PreCommit-HIVE-Build-13509/yetus.txt | | Powered by | Apache Yetushttp://yetus.apache.org | This message was automatically generated. > Improve org.apache.hadoop.hive.ql.exec.FunctionTask Experience > -- > > Key: HIVE-20466 > URL: https://issues.apache.org/jira/browse/HIVE-20466 > Project: Hive > Issue Type: Improvement > Components: Hive >Reporter: denys kuzmenko >Assignee: denys kuzmenko >Priority: Major > Attachments: HIVE-20466.1.patch, HIVE-20466.2.patch, > HIVE-20466.3.patch, HIVE-20466.4.patch > > -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (HIVE-18908) FULL OUTER JOIN to MapJoin
[ https://issues.apache.org/jira/browse/HIVE-18908?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Matt McCline updated HIVE-18908: Status: Patch Available (was: In Progress) > FULL OUTER JOIN to MapJoin > -- > > Key: HIVE-18908 > URL: https://issues.apache.org/jira/browse/HIVE-18908 > Project: Hive > Issue Type: Improvement > Components: Hive >Reporter: Matt McCline >Assignee: Matt McCline >Priority: Critical > Attachments: FULL OUTER MapJoin Code Changes.docx, > HIVE-18908.01.patch, HIVE-18908.02.patch, HIVE-18908.03.patch, > HIVE-18908.04.patch, HIVE-18908.05.patch, HIVE-18908.06.patch, > HIVE-18908.08.patch, HIVE-18908.09.patch, HIVE-18908.091.patch, > HIVE-18908.092.patch, HIVE-18908.093.patch, HIVE-18908.096.patch, > HIVE-18908.097.patch, HIVE-18908.098.patch, JOIN to MAPJOIN > Transformation.pdf, SHARED-MEMORY FULL OUTER MapJoin.pdf > > > Currently, we do not support FULL OUTER JOIN in MapJoin. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (HIVE-18908) FULL OUTER JOIN to MapJoin
[ https://issues.apache.org/jira/browse/HIVE-18908?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Matt McCline updated HIVE-18908: Attachment: HIVE-18908.098.patch > FULL OUTER JOIN to MapJoin > -- > > Key: HIVE-18908 > URL: https://issues.apache.org/jira/browse/HIVE-18908 > Project: Hive > Issue Type: Improvement > Components: Hive >Reporter: Matt McCline >Assignee: Matt McCline >Priority: Critical > Attachments: FULL OUTER MapJoin Code Changes.docx, > HIVE-18908.01.patch, HIVE-18908.02.patch, HIVE-18908.03.patch, > HIVE-18908.04.patch, HIVE-18908.05.patch, HIVE-18908.06.patch, > HIVE-18908.08.patch, HIVE-18908.09.patch, HIVE-18908.091.patch, > HIVE-18908.092.patch, HIVE-18908.093.patch, HIVE-18908.096.patch, > HIVE-18908.097.patch, HIVE-18908.098.patch, JOIN to MAPJOIN > Transformation.pdf, SHARED-MEMORY FULL OUTER MapJoin.pdf > > > Currently, we do not support FULL OUTER JOIN in MapJoin. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (HIVE-18908) FULL OUTER JOIN to MapJoin
[ https://issues.apache.org/jira/browse/HIVE-18908?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Matt McCline updated HIVE-18908: Status: In Progress (was: Patch Available) > FULL OUTER JOIN to MapJoin > -- > > Key: HIVE-18908 > URL: https://issues.apache.org/jira/browse/HIVE-18908 > Project: Hive > Issue Type: Improvement > Components: Hive >Reporter: Matt McCline >Assignee: Matt McCline >Priority: Critical > Attachments: FULL OUTER MapJoin Code Changes.docx, > HIVE-18908.01.patch, HIVE-18908.02.patch, HIVE-18908.03.patch, > HIVE-18908.04.patch, HIVE-18908.05.patch, HIVE-18908.06.patch, > HIVE-18908.08.patch, HIVE-18908.09.patch, HIVE-18908.091.patch, > HIVE-18908.092.patch, HIVE-18908.093.patch, HIVE-18908.096.patch, > HIVE-18908.097.patch, JOIN to MAPJOIN Transformation.pdf, SHARED-MEMORY FULL > OUTER MapJoin.pdf > > > Currently, we do not support FULL OUTER JOIN in MapJoin. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (HIVE-19748) Add appropriate null checks to DecimalColumnStatsAggregator
[ https://issues.apache.org/jira/browse/HIVE-19748?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Vaibhav Gumashta updated HIVE-19748: Status: Patch Available (was: Open) > Add appropriate null checks to DecimalColumnStatsAggregator > --- > > Key: HIVE-19748 > URL: https://issues.apache.org/jira/browse/HIVE-19748 > Project: Hive > Issue Type: Bug > Components: Metastore >Affects Versions: 3.0.0 >Reporter: Vaibhav Gumashta >Assignee: Vaibhav Gumashta >Priority: Major > Attachments: HIVE-19748.1.patch, HIVE-19748.1.patch > > > In some of our internal testing, we noticed that calls to > MetaStoreUtils.decimalToDoublee(Decimal decimal) from within > DecimalColumnStatsAggregator end up passing null Decimal values to the method. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (HIVE-19748) Add appropriate null checks to DecimalColumnStatsAggregator
[ https://issues.apache.org/jira/browse/HIVE-19748?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Vaibhav Gumashta updated HIVE-19748: Attachment: HIVE-19748.1.patch > Add appropriate null checks to DecimalColumnStatsAggregator > --- > > Key: HIVE-19748 > URL: https://issues.apache.org/jira/browse/HIVE-19748 > Project: Hive > Issue Type: Bug > Components: Metastore >Affects Versions: 3.0.0 >Reporter: Vaibhav Gumashta >Assignee: Vaibhav Gumashta >Priority: Major > Attachments: HIVE-19748.1.patch, HIVE-19748.1.patch > > > In some of our internal testing, we noticed that calls to > MetaStoreUtils.decimalToDoublee(Decimal decimal) from within > DecimalColumnStatsAggregator end up passing null Decimal values to the method. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (HIVE-19748) Add appropriate null checks to DecimalColumnStatsAggregator
[ https://issues.apache.org/jira/browse/HIVE-19748?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Vaibhav Gumashta updated HIVE-19748: Status: Open (was: Patch Available) > Add appropriate null checks to DecimalColumnStatsAggregator > --- > > Key: HIVE-19748 > URL: https://issues.apache.org/jira/browse/HIVE-19748 > Project: Hive > Issue Type: Bug > Components: Metastore >Affects Versions: 3.0.0 >Reporter: Vaibhav Gumashta >Assignee: Vaibhav Gumashta >Priority: Major > Attachments: HIVE-19748.1.patch, HIVE-19748.1.patch > > > In some of our internal testing, we noticed that calls to > MetaStoreUtils.decimalToDoublee(Decimal decimal) from within > DecimalColumnStatsAggregator end up passing null Decimal values to the method. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Resolved] (HIVE-20478) Metastore: Null checks needed in DecimalColumnStatsAggregator
[ https://issues.apache.org/jira/browse/HIVE-20478?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Vaibhav Gumashta resolved HIVE-20478. - Resolution: Duplicate > Metastore: Null checks needed in DecimalColumnStatsAggregator > - > > Key: HIVE-20478 > URL: https://issues.apache.org/jira/browse/HIVE-20478 > Project: Hive > Issue Type: Bug > Components: Standalone Metastore >Affects Versions: 3.1.0 >Reporter: Vaibhav Gumashta >Priority: Major > -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (HIVE-20433) Implicit String to Timestamp conversion is slow
[ https://issues.apache.org/jira/browse/HIVE-20433?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16595232#comment-16595232 ] Deepak Jaiswal commented on HIVE-20433: --- Ping [~gopalv] [~ashutoshc] > Implicit String to Timestamp conversion is slow > --- > > Key: HIVE-20433 > URL: https://issues.apache.org/jira/browse/HIVE-20433 > Project: Hive > Issue Type: Bug >Reporter: Deepak Jaiswal >Assignee: Deepak Jaiswal >Priority: Major > Attachments: HIVE-20433.1.patch, HIVE-20433.2.patch, > HIVE-20433.3.patch, HIVE-20433.4.patch, HIVE-20433.5.patch, > HIVE-20433.6.patch, HIVE-20433.7.patch, HIVE-20433.8.patch, HIVE-20433.9.patch > > > getTimestampFromString() is slow at casting dates. It throws twice before > date conversion can happen. > > cc [~gopalv] [~ashutoshc] -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (HIVE-20465) ProxyFileSystem.listStatusIterator function override required once migrated to Hadoop 3.2.0+
[ https://issues.apache.org/jira/browse/HIVE-20465?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16595220#comment-16595220 ] Hive QA commented on HIVE-20465: Here are the results of testing the latest attachment: https://issues.apache.org/jira/secure/attachment/12937439/HIVE-20465.7.patch {color:green}SUCCESS:{color} +1 due to 1 test(s) being added or modified. {color:green}SUCCESS:{color} +1 due to 14900 tests passed Test results: https://builds.apache.org/job/PreCommit-HIVE-Build/13508/testReport Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/13508/console Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-13508/ Messages: {noformat} Executing org.apache.hive.ptest.execution.TestCheckPhase Executing org.apache.hive.ptest.execution.PrepPhase Executing org.apache.hive.ptest.execution.YetusPhase Executing org.apache.hive.ptest.execution.ExecutionPhase Executing org.apache.hive.ptest.execution.ReportingPhase {noformat} This message is automatically generated. ATTACHMENT ID: 12937439 - PreCommit-HIVE-Build > ProxyFileSystem.listStatusIterator function override required once migrated > to Hadoop 3.2.0+ > > > Key: HIVE-20465 > URL: https://issues.apache.org/jira/browse/HIVE-20465 > Project: Hive > Issue Type: Test > Components: Tests >Reporter: denys kuzmenko >Assignee: denys kuzmenko >Priority: Major > Attachments: HIVE-20465.1.patch, HIVE-20465.2.patch, > HIVE-20465.3.patch, HIVE-20465.4.patch, HIVE-20465.5.patch, > HIVE-20465.6.patch, HIVE-20465.7.patch > > -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (HIVE-20377) Hive Kafka Storage Handler
[ https://issues.apache.org/jira/browse/HIVE-20377?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16595211#comment-16595211 ] Gopal V commented on HIVE-20377: LGTM - +1 > Hive Kafka Storage Handler > -- > > Key: HIVE-20377 > URL: https://issues.apache.org/jira/browse/HIVE-20377 > Project: Hive > Issue Type: New Feature >Affects Versions: 4.0.0 >Reporter: slim bouguerra >Assignee: slim bouguerra >Priority: Major > Attachments: HIVE-20377.10.patch, HIVE-20377.11.patch, > HIVE-20377.12.patch, HIVE-20377.15.patch, HIVE-20377.4.patch, > HIVE-20377.5.patch, HIVE-20377.6.patch, HIVE-20377.8.patch, > HIVE-20377.8.patch, HIVE-20377.patch > > > h1. Goal > * Read streaming data form Kafka queue as an external table. > * Allow streaming navigation by pushing down filters on Kafka record > partition id, offset and timestamp. > * Insert streaming data form Kafka to an actual Hive internal table, using > CTAS statement. > h1. Example > h2. Create the external table > {code} > CREATE EXTERNAL TABLE kafka_table (`timestamp` timestamp, page string, `user` > string, language string, added int, deleted int, flags string,comment string, > namespace string) > STORED BY 'org.apache.hadoop.hive.kafka.KafkaStorageHandler' > TBLPROPERTIES > ("kafka.topic" = "wikipedia", > "kafka.bootstrap.servers"="brokeraddress:9092", > "kafka.serde.class"="org.apache.hadoop.hive.serde2.JsonSerDe"); > {code} > h2. Kafka Metadata > In order to keep track of Kafka records the storage handler will add > automatically the Kafka row metadata eg partition id, record offset and > record timestamp. > {code} > DESCRIBE EXTENDED kafka_table > timestamp timestamp from deserializer > page string from deserializer > user string from deserializer > language string from deserializer > country string from deserializer > continent string from deserializer > namespace string from deserializer > newpage boolean from deserializer > unpatrolled boolean from deserializer > anonymous boolean from deserializer > robot boolean from deserializer > added int from deserializer > deleted int from deserializer > delta bigint from deserializer > __partition int from deserializer > __offset bigint from deserializer > __timestamp bigint from deserializer > {code} > h2. Filter push down. > Newer Kafka consumers 0.11.0 and higher allow seeking on the stream based on > a given offset. The proposed storage handler will be able to leverage such > API by pushing down filters over metadata columns, namely __partition (int), > __offset(long) and __timestamp(long) > For instance Query like > {code} > select `__offset` from kafka_table where (`__offset` < 10 and `__offset`>3 > and `__partition` = 0) or (`__partition` = 0 and `__offset` < 105 and > `__offset` > 99) or (`__offset` = 109); > {code} > Will result on a scan of partition 0 only then read only records between > offset 4 and 109. > h2. With timestamp seeks > The seeking based on the internal timestamps allows the handler to run on > recently arrived data, by doing > {code} > select count(*) from kafka_table where `__timestamp` > 1000 * > to_unix_timestamp(CURRENT_TIMESTAMP - interval '20' hours) ; > {code} > This allows for implicit relationships between event timestamps and kafka > timestamps to be expressed in queries (i.e event_timestamp is always < than > kafka __timestamp and kafka __timestamp is never > 15 minutes from event etc). > h2. More examples with Avro > {code} > CREATE EXTERNAL TABLE wiki_kafka_avro_table > STORED BY 'org.apache.hadoop.hive.kafka.KafkaStorageHandler' > TBLPROPERTIES > ("kafka.topic" = "wiki_kafka_avro_table", > "kafka.bootstrap.servers"="localhost:9092", > "kafka.serde.class"="org.apache.hadoop.hive.serde2.avro.AvroSerDe", > 'avro.schema.literal'='{ > "type" : "record", > "name" : "Wikipedia", > "namespace" : "org.apache.hive.kafka", > "version": "1", > "fields" : [ { > "name" : "isrobot", > "type" : "boolean" > }, { > "name" : "channel", > "type" : "string" > }, { > "name" : "timestamp", > "type" : "string" > }, { > "name" : "flags", > "type" : "string" > }, { > "name" : "isunpatrolled", > "type" : "boolean" > }, { > "name" : "page", > "type" :
[jira] [Commented] (HIVE-20476) CopyUtils used by REPL LOAD and EXPORT/IMPORT operations ignore distcp error.
[ https://issues.apache.org/jira/browse/HIVE-20476?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16595201#comment-16595201 ] Sankar Hariappan commented on HIVE-20476: - [~maheshk114], can you please take a look at the patch? > CopyUtils used by REPL LOAD and EXPORT/IMPORT operations ignore distcp error. > - > > Key: HIVE-20476 > URL: https://issues.apache.org/jira/browse/HIVE-20476 > Project: Hive > Issue Type: Bug > Components: HiveServer2, repl >Affects Versions: 3.1.0 >Reporter: Sankar Hariappan >Assignee: Sankar Hariappan >Priority: Major > Labels: DR, Export, Import, Replication, pull-request-available > Attachments: HIVE-20476.01.patch > > > CopyUtils uses FileUtils.distCp to copy files but doesn't check the return > value. It returns false if the copy fails. > Now, REPL LOAD and EXPORT/IMPORT commands internally uses CopyUtils to copy > data files across clusters and here it may return success even if file copy > fails and may cause data loss. > Need to throw error and retry. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (HIVE-20371) Fix definition for add_write_notification_log in hive_metastore.thrift.
[ https://issues.apache.org/jira/browse/HIVE-20371?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16595199#comment-16595199 ] ASF GitHub Bot commented on HIVE-20371: --- Github user sankarh closed the pull request at: https://github.com/apache/hive/pull/416 > Fix definition for add_write_notification_log in hive_metastore.thrift. > --- > > Key: HIVE-20371 > URL: https://issues.apache.org/jira/browse/HIVE-20371 > Project: Hive > Issue Type: Improvement > Components: repl >Affects Versions: 4.0.0, 3.2.0 >Reporter: Sankar Hariappan >Assignee: Sankar Hariappan >Priority: Minor > Labels: DR, pull-request-available, replication > Fix For: 4.0.0, 3.2.0 > > Attachments: HIVE-20371.01.patch > > > The definition of add_write_notification_log metastore api in > hive_metastore.thrift doesn't follow the standard format. > WriteNotificationLogResponse > add_write_notification_log(WriteNotificationLogRequest rqst) > to > WriteNotificationLogResponse > add_write_notification_log(1:WriteNotificationLogRequest rqst) -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (HIVE-20264) Bootstrap repl dump with concurrent write and drop of ACID table makes target inconsistent.
[ https://issues.apache.org/jira/browse/HIVE-20264?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16595198#comment-16595198 ] ASF GitHub Bot commented on HIVE-20264: --- Github user sankarh closed the pull request at: https://github.com/apache/hive/pull/410 > Bootstrap repl dump with concurrent write and drop of ACID table makes target > inconsistent. > --- > > Key: HIVE-20264 > URL: https://issues.apache.org/jira/browse/HIVE-20264 > Project: Hive > Issue Type: Sub-task > Components: HiveServer2, repl >Affects Versions: 4.0.0, 3.2.0 >Reporter: Sankar Hariappan >Assignee: Sankar Hariappan >Priority: Major > Labels: DR, pull-request-available, replication > Fix For: 4.0.0, 3.2.0 > > Attachments: HIVE-20264.01-branch-3.patch, HIVE-20264.01.patch, > HIVE-20264.02.patch > > > During bootstrap dump of ACID tables, let's consider the below sequence. > - Get lastReplId = last event ID logged. > - Current session (Thread-1), REPL DUMP -> Open txn (Txn1) - Event-10 > - Another session (Thread-2), Open txn (Txn2) - Event-11 > - Thread-2 -> Insert data (T1.D1) to ACID table. - Event-12 > - Thread-2 -> Commit Txn (Txn2) - Event-13 > - Thread-2 -> Drop table (T1) - Event-14 > - Thread-1 -> Dump ACID tables based on current list of tables. So, T1 will > be missing. > - Thread-1 -> Commit Txn (Txn1) > - REPL LOAD from bootstrap dump will skip T1. > - Incremental REPL DUMP will start from Event-10 and hence allocate write id > for table T1 and drop table(T1) is idempotent. So, at target, exist entries > in TXN_TO_WRITE_ID and NEXT_WRITE_ID metastore tables. > - Now, when we create another table at source with same name T1 and > replicate, then it may lead to incorrect data for readers at target on T1. > Couple of proposals: > 1. Make allocate write ID idempotent which is not possible as table doesn't > exist and MM table import may lead to allocate write id before creating > table. So, cannot differentiate these 2 cases. > 2. Make Drop table event to drop entries from TXN_TO_WRITE_ID and > NEXT_WRITE_ID tables irrespective of table exist or not at target. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (HIVE-20476) CopyUtils used by REPL LOAD and EXPORT/IMPORT operations ignore distcp error.
[ https://issues.apache.org/jira/browse/HIVE-20476?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16595177#comment-16595177 ] ASF GitHub Bot commented on HIVE-20476: --- GitHub user sankarh opened a pull request: https://github.com/apache/hive/pull/424 HIVE-20476: CopyUtils used by REPL LOAD and EXPORT/IMPORT operations ignore distcp error. You can merge this pull request into a Git repository by running: $ git pull https://github.com/sankarh/hive HIVE-20476 Alternatively you can review and apply these changes as the patch at: https://github.com/apache/hive/pull/424.patch To close this pull request, make a commit to your master/trunk branch with (at least) the following in the commit message: This closes #424 > CopyUtils used by REPL LOAD and EXPORT/IMPORT operations ignore distcp error. > - > > Key: HIVE-20476 > URL: https://issues.apache.org/jira/browse/HIVE-20476 > Project: Hive > Issue Type: Bug > Components: HiveServer2, repl >Affects Versions: 3.1.0 >Reporter: Sankar Hariappan >Assignee: Sankar Hariappan >Priority: Major > Labels: DR, Export, Import, Replication, pull-request-available > Attachments: HIVE-20476.01.patch > > > CopyUtils uses FileUtils.distCp to copy files but doesn't check the return > value. It returns false if the copy fails. > Now, REPL LOAD and EXPORT/IMPORT commands internally uses CopyUtils to copy > data files across clusters and here it may return success even if file copy > fails and may cause data loss. > Need to throw error and retry. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (HIVE-20476) CopyUtils used by REPL LOAD and EXPORT/IMPORT operations ignore distcp error.
[ https://issues.apache.org/jira/browse/HIVE-20476?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] ASF GitHub Bot updated HIVE-20476: -- Labels: DR Export Import Replication pull-request-available (was: DR Export Import Replication) > CopyUtils used by REPL LOAD and EXPORT/IMPORT operations ignore distcp error. > - > > Key: HIVE-20476 > URL: https://issues.apache.org/jira/browse/HIVE-20476 > Project: Hive > Issue Type: Bug > Components: HiveServer2, repl >Affects Versions: 3.1.0 >Reporter: Sankar Hariappan >Assignee: Sankar Hariappan >Priority: Major > Labels: DR, Export, Import, Replication, pull-request-available > Attachments: HIVE-20476.01.patch > > > CopyUtils uses FileUtils.distCp to copy files but doesn't check the return > value. It returns false if the copy fails. > Now, REPL LOAD and EXPORT/IMPORT commands internally uses CopyUtils to copy > data files across clusters and here it may return success even if file copy > fails and may cause data loss. > Need to throw error and retry. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (HIVE-20476) CopyUtils used by REPL LOAD and EXPORT/IMPORT operations ignore distcp error.
[ https://issues.apache.org/jira/browse/HIVE-20476?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Sankar Hariappan updated HIVE-20476: Status: Patch Available (was: Open) > CopyUtils used by REPL LOAD and EXPORT/IMPORT operations ignore distcp error. > - > > Key: HIVE-20476 > URL: https://issues.apache.org/jira/browse/HIVE-20476 > Project: Hive > Issue Type: Bug > Components: HiveServer2, repl >Affects Versions: 3.1.0 >Reporter: Sankar Hariappan >Assignee: Sankar Hariappan >Priority: Major > Labels: DR, Export, Import, Replication > Attachments: HIVE-20476.01.patch > > > CopyUtils uses FileUtils.distCp to copy files but doesn't check the return > value. It returns false if the copy fails. > Now, REPL LOAD and EXPORT/IMPORT commands internally uses CopyUtils to copy > data files across clusters and here it may return success even if file copy > fails and may cause data loss. > Need to throw error and retry. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (HIVE-20476) CopyUtils used by REPL LOAD and EXPORT/IMPORT operations ignore distcp error.
[ https://issues.apache.org/jira/browse/HIVE-20476?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Sankar Hariappan updated HIVE-20476: Attachment: (was: HIVE-20476.01.patch) > CopyUtils used by REPL LOAD and EXPORT/IMPORT operations ignore distcp error. > - > > Key: HIVE-20476 > URL: https://issues.apache.org/jira/browse/HIVE-20476 > Project: Hive > Issue Type: Bug > Components: HiveServer2, repl >Affects Versions: 3.1.0 >Reporter: Sankar Hariappan >Assignee: Sankar Hariappan >Priority: Major > Labels: DR, Export, Import, Replication > Attachments: HIVE-20476.01.patch > > > CopyUtils uses FileUtils.distCp to copy files but doesn't check the return > value. It returns false if the copy fails. > Now, REPL LOAD and EXPORT/IMPORT commands internally uses CopyUtils to copy > data files across clusters and here it may return success even if file copy > fails and may cause data loss. > Need to throw error and retry. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (HIVE-20476) CopyUtils used by REPL LOAD and EXPORT/IMPORT operations ignore distcp error.
[ https://issues.apache.org/jira/browse/HIVE-20476?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Sankar Hariappan updated HIVE-20476: Attachment: HIVE-20476.01.patch > CopyUtils used by REPL LOAD and EXPORT/IMPORT operations ignore distcp error. > - > > Key: HIVE-20476 > URL: https://issues.apache.org/jira/browse/HIVE-20476 > Project: Hive > Issue Type: Bug > Components: HiveServer2, repl >Affects Versions: 3.1.0 >Reporter: Sankar Hariappan >Assignee: Sankar Hariappan >Priority: Major > Labels: DR, Export, Import, Replication > Attachments: HIVE-20476.01.patch > > > CopyUtils uses FileUtils.distCp to copy files but doesn't check the return > value. It returns false if the copy fails. > Now, REPL LOAD and EXPORT/IMPORT commands internally uses CopyUtils to copy > data files across clusters and here it may return success even if file copy > fails and may cause data loss. > Need to throw error and retry. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (HIVE-20476) CopyUtils used by REPL LOAD and EXPORT/IMPORT operations ignore distcp error.
[ https://issues.apache.org/jira/browse/HIVE-20476?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Sankar Hariappan updated HIVE-20476: Attachment: HIVE-20476.01.patch > CopyUtils used by REPL LOAD and EXPORT/IMPORT operations ignore distcp error. > - > > Key: HIVE-20476 > URL: https://issues.apache.org/jira/browse/HIVE-20476 > Project: Hive > Issue Type: Bug > Components: HiveServer2, repl >Affects Versions: 3.1.0 >Reporter: Sankar Hariappan >Assignee: Sankar Hariappan >Priority: Major > Labels: DR, Export, Import, Replication > Attachments: HIVE-20476.01.patch > > > CopyUtils uses FileUtils.distCp to copy files but doesn't check the return > value. It returns false if the copy fails. > Now, REPL LOAD and EXPORT/IMPORT commands internally uses CopyUtils to copy > data files across clusters and here it may return success even if file copy > fails and may cause data loss. > Need to throw error and retry. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (HIVE-20476) CopyUtils used by REPL LOAD and EXPORT/IMPORT operations ignore distcp error.
[ https://issues.apache.org/jira/browse/HIVE-20476?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Sankar Hariappan updated HIVE-20476: Summary: CopyUtils used by REPL LOAD and EXPORT/IMPORT operations ignore distcp error. (was: REPL LOAD and EXPORT/IMPORT operations ignores distcp failures.) > CopyUtils used by REPL LOAD and EXPORT/IMPORT operations ignore distcp error. > - > > Key: HIVE-20476 > URL: https://issues.apache.org/jira/browse/HIVE-20476 > Project: Hive > Issue Type: Bug > Components: HiveServer2, repl >Affects Versions: 3.1.0 >Reporter: Sankar Hariappan >Assignee: Sankar Hariappan >Priority: Major > Labels: DR, Export, Import, Replication > > CopyUtils uses FileUtils.distCp to copy files but doesn't check the return > value. It returns false if the copy fails. > Now, REPL LOAD and EXPORT/IMPORT commands internally uses CopyUtils to copy > data files across clusters and here it may return success even if file copy > fails and may cause data loss. > Need to throw error and retry. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (HIVE-13457) Create HS2 REST API endpoints for monitoring information
[ https://issues.apache.org/jira/browse/HIVE-13457?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16595153#comment-16595153 ] Szehon Ho commented on HIVE-13457: -- Actually can you fix the checkstyle and findbugs? > Create HS2 REST API endpoints for monitoring information > > > Key: HIVE-13457 > URL: https://issues.apache.org/jira/browse/HIVE-13457 > Project: Hive > Issue Type: Improvement >Reporter: Szehon Ho >Assignee: Pawel Szostek >Priority: Major > Attachments: HIVE-13457.3.patch, HIVE-13457.4.patch, > HIVE-13457.5.patch, HIVE-13457.patch, HIVE-13457.patch > > > Similar to what is exposed in HS2 webui in HIVE-12338, it would be nice if > other UI's like admin tools or Hue can access and display this information as > well. Hence, we will create some REST endpoints to expose this information. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (HIVE-13457) Create HS2 REST API endpoints for monitoring information
[ https://issues.apache.org/jira/browse/HIVE-13457?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16595141#comment-16595141 ] Szehon Ho commented on HIVE-13457: -- Nice +1 > Create HS2 REST API endpoints for monitoring information > > > Key: HIVE-13457 > URL: https://issues.apache.org/jira/browse/HIVE-13457 > Project: Hive > Issue Type: Improvement >Reporter: Szehon Ho >Assignee: Pawel Szostek >Priority: Major > Attachments: HIVE-13457.3.patch, HIVE-13457.4.patch, > HIVE-13457.5.patch, HIVE-13457.patch, HIVE-13457.patch > > > Similar to what is exposed in HS2 webui in HIVE-12338, it would be nice if > other UI's like admin tools or Hue can access and display this information as > well. Hence, we will create some REST endpoints to expose this information. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (HIVE-20465) ProxyFileSystem.listStatusIterator function override required once migrated to Hadoop 3.2.0+
[ https://issues.apache.org/jira/browse/HIVE-20465?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16595139#comment-16595139 ] Hive QA commented on HIVE-20465: | (x) *{color:red}-1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || || || || || {color:brown} Prechecks {color} || | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s{color} | {color:green} The patch does not contain any @author tags. {color} | || || || || {color:brown} master Compile Tests {color} || | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 8m 38s{color} | {color:green} master passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 10s{color} | {color:green} master passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 10s{color} | {color:green} master passed {color} | | {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue} 0m 19s{color} | {color:blue} shims/common in master has 6 extant Findbugs warnings. {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 10s{color} | {color:green} master passed {color} | || || || || {color:brown} Patch Compile Tests {color} || | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 0m 12s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 11s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 0m 11s{color} | {color:green} the patch passed {color} | | {color:red}-1{color} | {color:red} checkstyle {color} | {color:red} 0m 9s{color} | {color:red} shims/common: The patch generated 2 new + 14 unchanged - 0 fixed = 16 total (was 14) {color} | | {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 0s{color} | {color:green} The patch has no whitespace issues. {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 0m 27s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 9s{color} | {color:green} the patch passed {color} | || || || || {color:brown} Other Tests {color} || | {color:green}+1{color} | {color:green} asflicense {color} | {color:green} 0m 14s{color} | {color:green} The patch does not generate ASF License warnings. {color} | | {color:black}{color} | {color:black} {color} | {color:black} 11m 17s{color} | {color:black} {color} | \\ \\ || Subsystem || Report/Notes || | Optional Tests | asflicense javac javadoc findbugs checkstyle compile | | uname | Linux hiveptest-server-upstream 3.16.0-4-amd64 #1 SMP Debian 3.16.36-1+deb8u1 (2016-09-03) x86_64 GNU/Linux | | Build tool | maven | | Personality | /data/hiveptest/working/yetus_PreCommit-HIVE-Build-13508/dev-support/hive-personality.sh | | git revision | master / 7cf1d00 | | Default Java | 1.8.0_111 | | findbugs | v3.0.0 | | checkstyle | http://104.198.109.242/logs//PreCommit-HIVE-Build-13508/yetus/diff-checkstyle-shims_common.txt | | modules | C: shims/common U: shims/common | | Console output | http://104.198.109.242/logs//PreCommit-HIVE-Build-13508/yetus.txt | | Powered by | Apache Yetushttp://yetus.apache.org | This message was automatically generated. > ProxyFileSystem.listStatusIterator function override required once migrated > to Hadoop 3.2.0+ > > > Key: HIVE-20465 > URL: https://issues.apache.org/jira/browse/HIVE-20465 > Project: Hive > Issue Type: Test > Components: Tests >Reporter: denys kuzmenko >Assignee: denys kuzmenko >Priority: Major > Attachments: HIVE-20465.1.patch, HIVE-20465.2.patch, > HIVE-20465.3.patch, HIVE-20465.4.patch, HIVE-20465.5.patch, > HIVE-20465.6.patch, HIVE-20465.7.patch > > -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (HIVE-13457) Create HS2 REST API endpoints for monitoring information
[ https://issues.apache.org/jira/browse/HIVE-13457?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16595116#comment-16595116 ] Hive QA commented on HIVE-13457: Here are the results of testing the latest attachment: https://issues.apache.org/jira/secure/attachment/12937431/HIVE-13457.4.patch {color:green}SUCCESS:{color} +1 due to 1 test(s) being added or modified. {color:green}SUCCESS:{color} +1 due to 14903 tests passed Test results: https://builds.apache.org/job/PreCommit-HIVE-Build/13507/testReport Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/13507/console Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-13507/ Messages: {noformat} Executing org.apache.hive.ptest.execution.TestCheckPhase Executing org.apache.hive.ptest.execution.PrepPhase Executing org.apache.hive.ptest.execution.YetusPhase Executing org.apache.hive.ptest.execution.ExecutionPhase Executing org.apache.hive.ptest.execution.ReportingPhase {noformat} This message is automatically generated. ATTACHMENT ID: 12937431 - PreCommit-HIVE-Build > Create HS2 REST API endpoints for monitoring information > > > Key: HIVE-13457 > URL: https://issues.apache.org/jira/browse/HIVE-13457 > Project: Hive > Issue Type: Improvement >Reporter: Szehon Ho >Assignee: Pawel Szostek >Priority: Major > Attachments: HIVE-13457.3.patch, HIVE-13457.4.patch, > HIVE-13457.5.patch, HIVE-13457.patch, HIVE-13457.patch > > > Similar to what is exposed in HS2 webui in HIVE-12338, it would be nice if > other UI's like admin tools or Hue can access and display this information as > well. Hence, we will create some REST endpoints to expose this information. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (HIVE-20296) Improve HivePointLookupOptimizerRule to be able to extract from more sophisticated contexts
[ https://issues.apache.org/jira/browse/HIVE-20296?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16595099#comment-16595099 ] Zoltan Haindrich commented on HIVE-20296: - [~ashutoshc] Could you please take a look? opened rb: https://reviews.apache.org/r/68525/ > Improve HivePointLookupOptimizerRule to be able to extract from more > sophisticated contexts > --- > > Key: HIVE-20296 > URL: https://issues.apache.org/jira/browse/HIVE-20296 > Project: Hive > Issue Type: Improvement > Components: Logical Optimizer >Reporter: Zoltan Haindrich >Assignee: Zoltan Haindrich >Priority: Major > Attachments: HIVE-20296.01.patch, HIVE-20296.01wip01.patch, > HIVE-20296.01wip02.patch, HIVE-20296.01wip03.patch, HIVE-20296.02.patch, > HIVE-20296.02.patch > > > Currently it could handle conditions in the form: > {code} > ( ( a=1 && b=1 && ... ) || (a=2 && b=2 && ...) || ... ) > {code} > If there is some noise somewhere in the or like: > {code} > ( x=18 || ( a=1 && b=1 && ... ) || (a=2 && b=2 && ...) || ... ) > {code} > it stops processing. > https://github.com/apache/hive/blob/2cabb8da150b8fb980223fbd6c2c93b842ca3ee5/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/rules/HivePointLookupOptimizerRule.java#L238 > HIVE-19097 right now unwinds some ors which can't be packed back; like query15 > https://github.com/apache/hive/blob/2cabb8da150b8fb980223fbd6c2c93b842ca3ee5/ql/src/test/queries/clientpositive/perf/query15.q#L14 -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (HIVE-17084) Turn on hive.stats.fetch.column.stats configuration flag
[ https://issues.apache.org/jira/browse/HIVE-17084?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Zoltan Haindrich updated HIVE-17084: Attachment: HIVE-17084.08.patch > Turn on hive.stats.fetch.column.stats configuration flag > > > Key: HIVE-17084 > URL: https://issues.apache.org/jira/browse/HIVE-17084 > Project: Hive > Issue Type: Task > Components: Statistics >Reporter: Vineet Garg >Assignee: Zoltan Haindrich >Priority: Major > Fix For: 3.2.0 > > Attachments: HIVE-17084.08.patch, HIVE-17084.1.patch, > HIVE-170884.4.patch, HIVE-170884.5.patch, HIVE-170884.7.patch > > > This flag is off by default and could result in bad plans due to missing > column statistics. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Assigned] (HIVE-17084) Turn on hive.stats.fetch.column.stats configuration flag
[ https://issues.apache.org/jira/browse/HIVE-17084?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Zoltan Haindrich reassigned HIVE-17084: --- Assignee: Vineet Garg (was: Zoltan Haindrich) > Turn on hive.stats.fetch.column.stats configuration flag > > > Key: HIVE-17084 > URL: https://issues.apache.org/jira/browse/HIVE-17084 > Project: Hive > Issue Type: Task > Components: Statistics >Reporter: Vineet Garg >Assignee: Vineet Garg >Priority: Major > Fix For: 3.2.0 > > Attachments: HIVE-17084.08.patch, HIVE-17084.1.patch, > HIVE-170884.4.patch, HIVE-170884.5.patch, HIVE-170884.7.patch > > > This flag is off by default and could result in bad plans due to missing > column statistics. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Assigned] (HIVE-17084) Turn on hive.stats.fetch.column.stats configuration flag
[ https://issues.apache.org/jira/browse/HIVE-17084?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Zoltan Haindrich reassigned HIVE-17084: --- Assignee: Zoltan Haindrich (was: Vineet Garg) > Turn on hive.stats.fetch.column.stats configuration flag > > > Key: HIVE-17084 > URL: https://issues.apache.org/jira/browse/HIVE-17084 > Project: Hive > Issue Type: Task > Components: Statistics >Reporter: Vineet Garg >Assignee: Zoltan Haindrich >Priority: Major > Fix For: 3.2.0 > > Attachments: HIVE-17084.1.patch, HIVE-170884.4.patch, > HIVE-170884.5.patch, HIVE-170884.7.patch > > > This flag is off by default and could result in bad plans due to missing > column statistics. -- This message was sent by Atlassian JIRA (v7.6.3#76005)