[jira] [Updated] (HIVE-20476) CopyUtils used by REPL LOAD and EXPORT/IMPORT operations ignore distcp error.

2018-08-28 Thread Sankar Hariappan (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-20476?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sankar Hariappan updated HIVE-20476:

Status: Patch Available  (was: Open)

> CopyUtils used by REPL LOAD and EXPORT/IMPORT operations ignore distcp error.
> -
>
> Key: HIVE-20476
> URL: https://issues.apache.org/jira/browse/HIVE-20476
> Project: Hive
>  Issue Type: Bug
>  Components: HiveServer2, repl
>Affects Versions: 3.1.0
>Reporter: Sankar Hariappan
>Assignee: Sankar Hariappan
>Priority: Major
>  Labels: DR, Export, Import, Replication, pull-request-available
> Attachments: HIVE-20476.01.patch
>
>
> CopyUtils uses FileUtils.distCp to copy files but doesn't check the return 
> value. It returns false if the copy fails.
> Now, REPL LOAD and EXPORT/IMPORT commands internally uses CopyUtils to copy 
> data files across clusters and here it may return success even if file copy 
> fails and may cause data loss.
> Need to throw error and retry.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-20476) CopyUtils used by REPL LOAD and EXPORT/IMPORT operations ignore distcp error.

2018-08-28 Thread Sankar Hariappan (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-20476?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sankar Hariappan updated HIVE-20476:

Attachment: HIVE-20476.01.patch

> CopyUtils used by REPL LOAD and EXPORT/IMPORT operations ignore distcp error.
> -
>
> Key: HIVE-20476
> URL: https://issues.apache.org/jira/browse/HIVE-20476
> Project: Hive
>  Issue Type: Bug
>  Components: HiveServer2, repl
>Affects Versions: 3.1.0
>Reporter: Sankar Hariappan
>Assignee: Sankar Hariappan
>Priority: Major
>  Labels: DR, Export, Import, Replication, pull-request-available
> Attachments: HIVE-20476.01.patch
>
>
> CopyUtils uses FileUtils.distCp to copy files but doesn't check the return 
> value. It returns false if the copy fails.
> Now, REPL LOAD and EXPORT/IMPORT commands internally uses CopyUtils to copy 
> data files across clusters and here it may return success even if file copy 
> fails and may cause data loss.
> Need to throw error and retry.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-20476) CopyUtils used by REPL LOAD and EXPORT/IMPORT operations ignore distcp error.

2018-08-28 Thread Sankar Hariappan (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-20476?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sankar Hariappan updated HIVE-20476:

Attachment: (was: HIVE-20476.01.patch)

> CopyUtils used by REPL LOAD and EXPORT/IMPORT operations ignore distcp error.
> -
>
> Key: HIVE-20476
> URL: https://issues.apache.org/jira/browse/HIVE-20476
> Project: Hive
>  Issue Type: Bug
>  Components: HiveServer2, repl
>Affects Versions: 3.1.0
>Reporter: Sankar Hariappan
>Assignee: Sankar Hariappan
>Priority: Major
>  Labels: DR, Export, Import, Replication, pull-request-available
>
> CopyUtils uses FileUtils.distCp to copy files but doesn't check the return 
> value. It returns false if the copy fails.
> Now, REPL LOAD and EXPORT/IMPORT commands internally uses CopyUtils to copy 
> data files across clusters and here it may return success even if file copy 
> fails and may cause data loss.
> Need to throw error and retry.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-20476) CopyUtils used by REPL LOAD and EXPORT/IMPORT operations ignore distcp error.

2018-08-28 Thread Sankar Hariappan (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-20476?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sankar Hariappan updated HIVE-20476:

Status: Open  (was: Patch Available)

> CopyUtils used by REPL LOAD and EXPORT/IMPORT operations ignore distcp error.
> -
>
> Key: HIVE-20476
> URL: https://issues.apache.org/jira/browse/HIVE-20476
> Project: Hive
>  Issue Type: Bug
>  Components: HiveServer2, repl
>Affects Versions: 3.1.0
>Reporter: Sankar Hariappan
>Assignee: Sankar Hariappan
>Priority: Major
>  Labels: DR, Export, Import, Replication, pull-request-available
>
> CopyUtils uses FileUtils.distCp to copy files but doesn't check the return 
> value. It returns false if the copy fails.
> Now, REPL LOAD and EXPORT/IMPORT commands internally uses CopyUtils to copy 
> data files across clusters and here it may return success even if file copy 
> fails and may cause data loss.
> Need to throw error and retry.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-20476) CopyUtils used by REPL LOAD and EXPORT/IMPORT operations ignore distcp error.

2018-08-28 Thread Sankar Hariappan (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-20476?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sankar Hariappan updated HIVE-20476:

Attachment: HIVE-20476.01.patch

> CopyUtils used by REPL LOAD and EXPORT/IMPORT operations ignore distcp error.
> -
>
> Key: HIVE-20476
> URL: https://issues.apache.org/jira/browse/HIVE-20476
> Project: Hive
>  Issue Type: Bug
>  Components: HiveServer2, repl
>Affects Versions: 3.1.0
>Reporter: Sankar Hariappan
>Assignee: Sankar Hariappan
>Priority: Major
>  Labels: DR, Export, Import, Replication, pull-request-available
> Attachments: HIVE-20476.01.patch
>
>
> CopyUtils uses FileUtils.distCp to copy files but doesn't check the return 
> value. It returns false if the copy fails.
> Now, REPL LOAD and EXPORT/IMPORT commands internally uses CopyUtils to copy 
> data files across clusters and here it may return success even if file copy 
> fails and may cause data loss.
> Need to throw error and retry.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-20476) CopyUtils used by REPL LOAD and EXPORT/IMPORT operations ignore distcp error.

2018-08-28 Thread Sankar Hariappan (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-20476?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sankar Hariappan updated HIVE-20476:

Attachment: (was: HIVE-20476.01.patch)

> CopyUtils used by REPL LOAD and EXPORT/IMPORT operations ignore distcp error.
> -
>
> Key: HIVE-20476
> URL: https://issues.apache.org/jira/browse/HIVE-20476
> Project: Hive
>  Issue Type: Bug
>  Components: HiveServer2, repl
>Affects Versions: 3.1.0
>Reporter: Sankar Hariappan
>Assignee: Sankar Hariappan
>Priority: Major
>  Labels: DR, Export, Import, Replication, pull-request-available
> Attachments: HIVE-20476.01.patch
>
>
> CopyUtils uses FileUtils.distCp to copy files but doesn't check the return 
> value. It returns false if the copy fails.
> Now, REPL LOAD and EXPORT/IMPORT commands internally uses CopyUtils to copy 
> data files across clusters and here it may return success even if file copy 
> fails and may cause data loss.
> Need to throw error and retry.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-20476) CopyUtils used by REPL LOAD and EXPORT/IMPORT operations ignore distcp error.

2018-08-28 Thread Sankar Hariappan (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-20476?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sankar Hariappan updated HIVE-20476:

Status: Patch Available  (was: Open)

Flaky test failure. Re-attaching same patch.

> CopyUtils used by REPL LOAD and EXPORT/IMPORT operations ignore distcp error.
> -
>
> Key: HIVE-20476
> URL: https://issues.apache.org/jira/browse/HIVE-20476
> Project: Hive
>  Issue Type: Bug
>  Components: HiveServer2, repl
>Affects Versions: 3.1.0
>Reporter: Sankar Hariappan
>Assignee: Sankar Hariappan
>Priority: Major
>  Labels: DR, Export, Import, Replication, pull-request-available
> Attachments: HIVE-20476.01.patch
>
>
> CopyUtils uses FileUtils.distCp to copy files but doesn't check the return 
> value. It returns false if the copy fails.
> Now, REPL LOAD and EXPORT/IMPORT commands internally uses CopyUtils to copy 
> data files across clusters and here it may return success even if file copy 
> fails and may cause data loss.
> Need to throw error and retry.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-20476) CopyUtils used by REPL LOAD and EXPORT/IMPORT operations ignore distcp error.

2018-08-28 Thread Sankar Hariappan (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-20476?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sankar Hariappan updated HIVE-20476:

Status: Open  (was: Patch Available)

> CopyUtils used by REPL LOAD and EXPORT/IMPORT operations ignore distcp error.
> -
>
> Key: HIVE-20476
> URL: https://issues.apache.org/jira/browse/HIVE-20476
> Project: Hive
>  Issue Type: Bug
>  Components: HiveServer2, repl
>Affects Versions: 3.1.0
>Reporter: Sankar Hariappan
>Assignee: Sankar Hariappan
>Priority: Major
>  Labels: DR, Export, Import, Replication, pull-request-available
> Attachments: HIVE-20476.01.patch
>
>
> CopyUtils uses FileUtils.distCp to copy files but doesn't check the return 
> value. It returns false if the copy fails.
> Now, REPL LOAD and EXPORT/IMPORT commands internally uses CopyUtils to copy 
> data files across clusters and here it may return success even if file copy 
> fails and may cause data loss.
> Need to throw error and retry.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-20405) Run all druid tests in one batch

2018-08-28 Thread slim bouguerra (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-20405?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16595859#comment-16595859
 ] 

slim bouguerra commented on HIVE-20405:
---

I can see from the logs that the default batch size for miniDruidTests is 5 
(grep for  qFileTest.miniDruid.batchSize) at 
http://104.198.109.242/logs/PreCommit-HIVE-Build-13517/execution.txt
This leads to running multiple tests setups for Druid tests, that causes issue 
if each batch is trying to startup a Druid Cluster. I think a short fix can be 
to increase the batch size to something like 50, this will ensure that only one 
Druid Cluster is used.
{code}
QFileTestBatch [batchId=192, size=5, driver=TestMiniDruidCliDriver, 
queryFilesProperty=qfile, 
name=192-TestMiniDruidCliDriver-druidmini_test_ts.q-druidmini_mv.q-druid_timestamptz2.q-and-2-more,
 tests=[druidmini_test_ts.q, druidmini_mv.q, druid_timestamptz2.q, 
druidmini_extractTime.q, druidmini_floorTime.q], isParallel=true, 
moduleName=itests/qtest]

2018-08-29 00:08:31,293 DEBUG [TestExecutor] ExecutionPhase.execute:98 PBatch: 
QFileTestBatch [batchId=193, size=5, driver=TestMiniDruidCliDriver, 
queryFilesProperty=qfile, 
name=193-TestMiniDruidCliDriver-druidmini_dynamic_partition.q-druidmini_expressions.q-druidmini_test_alter.q-and-2-more,
 tests=[druidmini_dynamic_partition.q, druidmini_expressions.q, 
druidmini_test_alter.q, druidmini_test1.q, druidmini_test_insert.q], 
isParallel=true, moduleName=itests/qtest]
2018-08-29 00:08:31,293 DEBUG [TestExecutor] ExecutionPhase.execute:98 PBatch: 
QFileTestBatch [batchId=194, size=3, driver=TestMiniDruidCliDriver, 
queryFilesProperty=qfile, 
name=194-TestMiniDruidCliDriver-druidmini_masking.q-druidmini_joins.q-druid_timestamptz.q,
 tests=[druidmini_masking.q, druidmini_joins.q, druid_timestamptz.q], 
isParallel=true, moduleName=itests/qtest]
{code}

> Run all druid tests in one batch
> 
>
> Key: HIVE-20405
> URL: https://issues.apache.org/jira/browse/HIVE-20405
> Project: Hive
>  Issue Type: Test
>  Components: Testing Infrastructure
>Reporter: Vineet Garg
>Assignee: Nishant Bangarwa
>Priority: Major
>
> Running druid tests in parallel could cause issues so all of the tests should 
> be run in one batch.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-20377) Hive Kafka Storage Handler

2018-08-28 Thread slim bouguerra (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-20377?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

slim bouguerra updated HIVE-20377:
--
Attachment: HIVE-20377.18.patch

> Hive Kafka Storage Handler
> --
>
> Key: HIVE-20377
> URL: https://issues.apache.org/jira/browse/HIVE-20377
> Project: Hive
>  Issue Type: New Feature
>Affects Versions: 4.0.0
>Reporter: slim bouguerra
>Assignee: slim bouguerra
>Priority: Major
> Attachments: HIVE-20377.10.patch, HIVE-20377.11.patch, 
> HIVE-20377.12.patch, HIVE-20377.15.patch, HIVE-20377.18.patch, 
> HIVE-20377.18.patch, HIVE-20377.4.patch, HIVE-20377.5.patch, 
> HIVE-20377.6.patch, HIVE-20377.8.patch, HIVE-20377.8.patch, HIVE-20377.patch
>
>
> h1. Goal
> * Read streaming data form Kafka queue as an external table.
> * Allow streaming navigation by pushing down filters on Kafka record 
> partition id, offset and timestamp. 
> * Insert streaming data form Kafka to an actual Hive internal table, using 
> CTAS statement.
> h1. Example
> h2. Create the external table
> {code} 
> CREATE EXTERNAL TABLE kafka_table (`timestamp` timestamp, page string, `user` 
> string, language string, added int, deleted int, flags string,comment string, 
> namespace string)
> STORED BY 'org.apache.hadoop.hive.kafka.KafkaStorageHandler'
> TBLPROPERTIES 
> ("kafka.topic" = "wikipedia", 
> "kafka.bootstrap.servers"="brokeraddress:9092",
> "kafka.serde.class"="org.apache.hadoop.hive.serde2.JsonSerDe");
> {code}
> h2. Kafka Metadata
> In order to keep track of Kafka records the storage handler will add 
> automatically the Kafka row metadata eg partition id, record offset and 
> record timestamp. 
> {code}
> DESCRIBE EXTENDED kafka_table
> timestamp timestamp   from deserializer   
> page  string  from deserializer   
> user  string  from deserializer   
> language  string  from deserializer   
> country   string  from deserializer   
> continent string  from deserializer   
> namespace string  from deserializer   
> newpage   boolean from deserializer   
> unpatrolled   boolean from deserializer   
> anonymous boolean from deserializer   
> robot boolean from deserializer   
> added int from deserializer   
> deleted   int from deserializer   
> delta bigint  from deserializer   
> __partition   int from deserializer   
> __offset  bigint  from deserializer   
> __timestamp   bigint  from deserializer   
> {code}
> h2. Filter push down.
> Newer Kafka consumers 0.11.0 and higher allow seeking on the stream based on 
> a given offset. The proposed storage handler will be able to leverage such 
> API by pushing down filters over metadata columns, namely __partition (int), 
> __offset(long) and __timestamp(long)
> For instance Query like
> {code} 
> select `__offset` from kafka_table where (`__offset` < 10 and `__offset`>3 
> and `__partition` = 0) or (`__partition` = 0 and `__offset` < 105 and 
> `__offset` > 99) or (`__offset` = 109);
> {code}
> Will result on a scan of partition 0 only then read only records between 
> offset 4 and 109. 
> h2. With timestamp seeks 
> The seeking based on the internal timestamps allows the handler to run on 
> recently arrived data, by doing
> {code}
> select count(*) from kafka_table where `__timestamp` >  1000 * 
> to_unix_timestamp(CURRENT_TIMESTAMP - interval '20' hours) ;
> {code}
> This allows for implicit relationships between event timestamps and kafka 
> timestamps to be expressed in queries (i.e event_timestamp is always < than 
> kafka __timestamp and kafka __timestamp is never > 15 minutes from event etc).
> h2. More examples with Avro 
> {code}
> CREATE EXTERNAL TABLE wiki_kafka_avro_table
> STORED BY 'org.apache.hadoop.hive.kafka.KafkaStorageHandler'
> TBLPROPERTIES
> ("kafka.topic" = "wiki_kafka_avro_table",
> "kafka.bootstrap.servers"="localhost:9092",
> "kafka.serde.class"="org.apache.hadoop.hive.serde2.avro.AvroSerDe",
> 'avro.schema.literal'='{
>   "type" : "record",
>   "name" : "Wikipedia",
>   "namespace" : "org.apache.hive.kafka",
>   "version": "1",
>   "fields" : [ {
> "name" : "isrobot",
> "type" : "boolean"
>   }, {
> "name" : "channel",
> "type" : "string"
>   }, {
> "name" : "timestamp",
> "type" : "string"
>   }, {
> "name" : "flags",
> "type" : "string"
>   }, {
> "name" : "isunpatrolled",
> "type" : "boolean"
>   }, 

[jira] [Commented] (HIVE-18767) Some alterPartitions invocations throw 'NumberFormatException: null'

2018-08-28 Thread Mass Dosage (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-18767?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16595841#comment-16595841
 ] 

Mass Dosage commented on HIVE-18767:


No problem, OK, here we go one by one with the first one up being 
HIVE-18767.2-branch-2.3.patch.

> Some alterPartitions invocations throw 'NumberFormatException: null'
> 
>
> Key: HIVE-18767
> URL: https://issues.apache.org/jira/browse/HIVE-18767
> Project: Hive
>  Issue Type: Bug
>  Components: Metastore
>Affects Versions: 2.3.3, 3.1.0, 4.0.0, 3.2.0
>Reporter: Yuming Wang
>Assignee: Mass Dosage
>Priority: Major
> Fix For: 2.3.3, 4.0.0
>
> Attachments: HIVE-18767-branch-2.3.patch, HIVE-18767-branch-2.patch, 
> HIVE-18767-branch-3.1.patch, HIVE-18767-branch-3.patch, HIVE-18767.1.patch, 
> HIVE-18767.2-branch-2.3.patch, HIVE-18767.2.patch, HIVE-18767.3.patch, 
> HIVE-18767.4.patch, HIVE-18767.5.patch, HIVE-18767.6.patch
>
>
> Error messages:
> {noformat}
> [info] Cause: java.lang.NumberFormatException: null
> [info] at java.lang.Long.parseLong(Long.java:552)
> [info] at java.lang.Long.parseLong(Long.java:631)
> [info] at 
> org.apache.hadoop.hive.metastore.MetaStoreUtils.isFastStatsSame(MetaStoreUtils.java:315)
> [info] at 
> org.apache.hadoop.hive.metastore.HiveAlterHandler.alterPartitions(HiveAlterHandler.java:605)
> [info] at 
> org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.alter_partitions_with_environment_context(HiveMetaStore.java:3837)
> [info] at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> [info] at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
> [info] at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> [info] at java.lang.reflect.Method.invoke(Method.java:498)
> [info] at 
> org.apache.hadoop.hive.metastore.RetryingHMSHandler.invokeInternal(RetryingHMSHandler.java:148)
> [info] at 
> org.apache.hadoop.hive.metastore.RetryingHMSHandler.invoke(RetryingHMSHandler.java:107)
> [info] at 
> com.sun.proxy.$Proxy23.alter_partitions_with_environment_context(Unknown 
> Source)
> [info] at 
> org.apache.hadoop.hive.metastore.HiveMetaStoreClient.alter_partitions(HiveMetaStoreClient.java:1527)
> {noformat}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-18767) Some alterPartitions invocations throw 'NumberFormatException: null'

2018-08-28 Thread Mass Dosage (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-18767?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mass Dosage updated HIVE-18767:
---
Attachment: HIVE-18767.2-branch-2.3.patch

> Some alterPartitions invocations throw 'NumberFormatException: null'
> 
>
> Key: HIVE-18767
> URL: https://issues.apache.org/jira/browse/HIVE-18767
> Project: Hive
>  Issue Type: Bug
>  Components: Metastore
>Affects Versions: 2.3.3, 3.1.0, 4.0.0, 3.2.0
>Reporter: Yuming Wang
>Assignee: Mass Dosage
>Priority: Major
> Fix For: 2.3.3, 4.0.0
>
> Attachments: HIVE-18767-branch-2.3.patch, HIVE-18767-branch-2.patch, 
> HIVE-18767-branch-3.1.patch, HIVE-18767-branch-3.patch, HIVE-18767.1.patch, 
> HIVE-18767.2-branch-2.3.patch, HIVE-18767.2.patch, HIVE-18767.3.patch, 
> HIVE-18767.4.patch, HIVE-18767.5.patch, HIVE-18767.6.patch
>
>
> Error messages:
> {noformat}
> [info] Cause: java.lang.NumberFormatException: null
> [info] at java.lang.Long.parseLong(Long.java:552)
> [info] at java.lang.Long.parseLong(Long.java:631)
> [info] at 
> org.apache.hadoop.hive.metastore.MetaStoreUtils.isFastStatsSame(MetaStoreUtils.java:315)
> [info] at 
> org.apache.hadoop.hive.metastore.HiveAlterHandler.alterPartitions(HiveAlterHandler.java:605)
> [info] at 
> org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.alter_partitions_with_environment_context(HiveMetaStore.java:3837)
> [info] at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> [info] at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
> [info] at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> [info] at java.lang.reflect.Method.invoke(Method.java:498)
> [info] at 
> org.apache.hadoop.hive.metastore.RetryingHMSHandler.invokeInternal(RetryingHMSHandler.java:148)
> [info] at 
> org.apache.hadoop.hive.metastore.RetryingHMSHandler.invoke(RetryingHMSHandler.java:107)
> [info] at 
> com.sun.proxy.$Proxy23.alter_partitions_with_environment_context(Unknown 
> Source)
> [info] at 
> org.apache.hadoop.hive.metastore.HiveMetaStoreClient.alter_partitions(HiveMetaStoreClient.java:1527)
> {noformat}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-18767) Some alterPartitions invocations throw 'NumberFormatException: null'

2018-08-28 Thread Mass Dosage (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-18767?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mass Dosage updated HIVE-18767:
---
Target Version/s: 3.1.0, 2.3.3, 4.0.0, 3.2.0  (was: 2.3.3, 3.1.0, 4.0.0, 
3.2.0)
  Status: In Progress  (was: Patch Available)

> Some alterPartitions invocations throw 'NumberFormatException: null'
> 
>
> Key: HIVE-18767
> URL: https://issues.apache.org/jira/browse/HIVE-18767
> Project: Hive
>  Issue Type: Bug
>  Components: Metastore
>Affects Versions: 3.1.0, 2.3.3, 4.0.0, 3.2.0
>Reporter: Yuming Wang
>Assignee: Mass Dosage
>Priority: Major
> Fix For: 2.3.3, 4.0.0
>
> Attachments: HIVE-18767-branch-2.3.patch, HIVE-18767-branch-2.patch, 
> HIVE-18767-branch-3.1.patch, HIVE-18767-branch-3.patch, HIVE-18767.1.patch, 
> HIVE-18767.2-branch-2.3.patch, HIVE-18767.2.patch, HIVE-18767.3.patch, 
> HIVE-18767.4.patch, HIVE-18767.5.patch, HIVE-18767.6.patch
>
>
> Error messages:
> {noformat}
> [info] Cause: java.lang.NumberFormatException: null
> [info] at java.lang.Long.parseLong(Long.java:552)
> [info] at java.lang.Long.parseLong(Long.java:631)
> [info] at 
> org.apache.hadoop.hive.metastore.MetaStoreUtils.isFastStatsSame(MetaStoreUtils.java:315)
> [info] at 
> org.apache.hadoop.hive.metastore.HiveAlterHandler.alterPartitions(HiveAlterHandler.java:605)
> [info] at 
> org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.alter_partitions_with_environment_context(HiveMetaStore.java:3837)
> [info] at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> [info] at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
> [info] at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> [info] at java.lang.reflect.Method.invoke(Method.java:498)
> [info] at 
> org.apache.hadoop.hive.metastore.RetryingHMSHandler.invokeInternal(RetryingHMSHandler.java:148)
> [info] at 
> org.apache.hadoop.hive.metastore.RetryingHMSHandler.invoke(RetryingHMSHandler.java:107)
> [info] at 
> com.sun.proxy.$Proxy23.alter_partitions_with_environment_context(Unknown 
> Source)
> [info] at 
> org.apache.hadoop.hive.metastore.HiveMetaStoreClient.alter_partitions(HiveMetaStoreClient.java:1527)
> {noformat}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-20363) Use integer constants for frequently used serde classes.

2018-08-28 Thread Anurag Mantripragada (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-20363?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Anurag Mantripragada updated HIVE-20363:

Attachment: (was: HIVE-20363.1.patch)

> Use integer constants for frequently used serde classes.
> 
>
> Key: HIVE-20363
> URL: https://issues.apache.org/jira/browse/HIVE-20363
> Project: Hive
>  Issue Type: Improvement
>  Components: Metastore, Standalone Metastore
>Reporter: Anurag Mantripragada
>Assignee: Anurag Mantripragada
>Priority: Major
> Attachments: HIVE-20363.1.patch
>
>
> Serde libraries are stored as fully qualified class names which are long 
> strings, we can get improvements in I/O and storage if we store integer 
> constants for frequently used serde classes in the backend DB.
> For example:
> {code:java}
> org.apache.hadoop.hive.serde2.avro.AvroSerDe"---> 1
> parquet.hive.serde.ParquetHiveSerDe" ---> 2
> ...{code}
>  Review the patch at: 
> https://reviews.apache.org/r/68546/



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-20363) Use integer constants for frequently used serde classes.

2018-08-28 Thread Anurag Mantripragada (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-20363?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Anurag Mantripragada updated HIVE-20363:

Attachment: HIVE-20363.1.patch
Status: Patch Available  (was: Open)

> Use integer constants for frequently used serde classes.
> 
>
> Key: HIVE-20363
> URL: https://issues.apache.org/jira/browse/HIVE-20363
> Project: Hive
>  Issue Type: Improvement
>  Components: Metastore, Standalone Metastore
>Reporter: Anurag Mantripragada
>Assignee: Anurag Mantripragada
>Priority: Major
> Attachments: HIVE-20363.1.patch
>
>
> Serde libraries are stored as fully qualified class names which are long 
> strings, we can get improvements in I/O and storage if we store integer 
> constants for frequently used serde classes in the backend DB.
> For example:
> {code:java}
> org.apache.hadoop.hive.serde2.avro.AvroSerDe"---> 1
> parquet.hive.serde.ParquetHiveSerDe" ---> 2
> ...{code}
>  Review the patch at: 
> https://reviews.apache.org/r/68546/



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-20363) Use integer constants for frequently used serde classes.

2018-08-28 Thread Anurag Mantripragada (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-20363?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Anurag Mantripragada updated HIVE-20363:

Attachment: HIVE-20363.1.patch

> Use integer constants for frequently used serde classes.
> 
>
> Key: HIVE-20363
> URL: https://issues.apache.org/jira/browse/HIVE-20363
> Project: Hive
>  Issue Type: Improvement
>  Components: Metastore, Standalone Metastore
>Reporter: Anurag Mantripragada
>Assignee: Anurag Mantripragada
>Priority: Major
> Attachments: HIVE-20363.1.patch
>
>
> Serde libraries are stored as fully qualified class names which are long 
> strings, we can get improvements in I/O and storage if we store integer 
> constants for frequently used serde classes in the backend DB.
> For example:
> {code:java}
> org.apache.hadoop.hive.serde2.avro.AvroSerDe"---> 1
> parquet.hive.serde.ParquetHiveSerDe" ---> 2
> ...{code}
>  Review the patch at: 
> https://reviews.apache.org/r/68546/



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-20483) Really move metastore common classes into metastore-common

2018-08-28 Thread Alexander Kolbasov (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-20483?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16595822#comment-16595822
 ] 

Alexander Kolbasov commented on HIVE-20483:
---

[~pvary] Would you be able to review this patch?

> Really move metastore common classes into metastore-common
> --
>
> Key: HIVE-20483
> URL: https://issues.apache.org/jira/browse/HIVE-20483
> Project: Hive
>  Issue Type: Sub-task
>  Components: Standalone Metastore
>Affects Versions: 3.0.1, 4.0.0
>Reporter: Alexander Kolbasov
>Assignee: Alexander Kolbasov
>Priority: Major
> Attachments: HIVE-20483.01.patch
>
>
> HIVE-20482 patch was supposed to move a bunch of files from metastore-server 
> to metastore-common but for some reason it didn't happen, so now these files 
> should be moved.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-20483) Really move metastore common classes into metastore-common

2018-08-28 Thread Alexander Kolbasov (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-20483?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16595821#comment-16595821
 ] 

Alexander Kolbasov commented on HIVE-20483:
---

This patch doesn't introduce any code changes - it just moves files around and 
changes pom files. It removes metastore-server dependency on hive-metastore and 
adds it to a couple of modules.

> Really move metastore common classes into metastore-common
> --
>
> Key: HIVE-20483
> URL: https://issues.apache.org/jira/browse/HIVE-20483
> Project: Hive
>  Issue Type: Sub-task
>  Components: Standalone Metastore
>Affects Versions: 3.0.1, 4.0.0
>Reporter: Alexander Kolbasov
>Assignee: Alexander Kolbasov
>Priority: Major
> Attachments: HIVE-20483.01.patch
>
>
> HIVE-20482 patch was supposed to move a bunch of files from metastore-server 
> to metastore-common but for some reason it didn't happen, so now these files 
> should be moved.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-19253) HMS ignores tableType property for external tables

2018-08-28 Thread Alexander Kolbasov (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-19253?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alexander Kolbasov updated HIVE-19253:
--
Attachment: HIVE-19253.10.patch

> HMS ignores tableType property for external tables
> --
>
> Key: HIVE-19253
> URL: https://issues.apache.org/jira/browse/HIVE-19253
> Project: Hive
>  Issue Type: Bug
>  Components: Metastore
>Affects Versions: 3.0.0, 3.1.0, 4.0.0
>Reporter: Alexander Kolbasov
>Assignee: Alexander Kolbasov
>Priority: Major
>  Labels: newbie
> Attachments: HIVE-19253.01.patch, HIVE-19253.02.patch, 
> HIVE-19253.03.patch, HIVE-19253.03.patch, HIVE-19253.04.patch, 
> HIVE-19253.05.patch, HIVE-19253.06.patch, HIVE-19253.07.patch, 
> HIVE-19253.08.patch, HIVE-19253.09.patch, HIVE-19253.10.patch
>
>
> When someone creates a table using Thrift API they may think that setting 
> tableType to {{EXTERNAL_TABLE}} creates an external table. And boom - their 
> table is gone later because HMS will silently change it to managed table.
> here is the offending code:
> {code:java}
>   private MTable convertToMTable(Table tbl) throws InvalidObjectException,
>   MetaException {
> ...
> // If the table has property EXTERNAL set, update table type
> // accordingly
> String tableType = tbl.getTableType();
> boolean isExternal = 
> Boolean.parseBoolean(tbl.getParameters().get("EXTERNAL"));
> if (TableType.MANAGED_TABLE.toString().equals(tableType)) {
>   if (isExternal) {
> tableType = TableType.EXTERNAL_TABLE.toString();
>   }
> }
> if (TableType.EXTERNAL_TABLE.toString().equals(tableType)) {
>   if (!isExternal) { // Here!
> tableType = TableType.MANAGED_TABLE.toString();
>   }
> }
> {code}
> So if the EXTERNAL parameter is not set, table type is changed to managed 
> even if it was external in the first place - which is wrong.
> More over, in other places code looks at the table property to decide table 
> type and some places look at parameter. HMS should really make its mind which 
> one to use.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-20377) Hive Kafka Storage Handler

2018-08-28 Thread Hive QA (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-20377?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16595819#comment-16595819
 ] 

Hive QA commented on HIVE-20377:


| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
|| || || || {color:brown} Prechecks {color} ||
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
1s{color} | {color:green} The patch does not contain any @author tags. {color} |
|| || || || {color:brown} master Compile Tests {color} ||
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  0m 
32s{color} | {color:blue} Maven dependency ordering for branch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  8m 
 6s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  9m 
43s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  2m 
19s{color} | {color:green} master passed {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  0m 
42s{color} | {color:blue} serde in master has 195 extant Findbugs warnings. 
{color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  0m 
19s{color} | {color:blue} itests/qtest-druid in master has 6 extant Findbugs 
warnings. {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  0m 
48s{color} | {color:blue} itests/util in master has 52 extant Findbugs 
warnings. {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  0m 
40s{color} | {color:blue} llap-server in master has 84 extant Findbugs 
warnings. {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  4m  
6s{color} | {color:blue} ql in master has 2310 extant Findbugs warnings. 
{color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  9m 
15s{color} | {color:green} master passed {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  0m  
7s{color} | {color:blue} Maven dependency ordering for patch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 11m 
 9s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  9m 
53s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  9m 
53s{color} | {color:green} the patch passed {color} |
| {color:red}-1{color} | {color:red} checkstyle {color} | {color:red}  0m 
10s{color} | {color:red} itests/qtest-druid: The patch generated 37 new + 3 
unchanged - 0 fixed = 40 total (was 3) {color} |
| {color:red}-1{color} | {color:red} checkstyle {color} | {color:red}  0m  
9s{color} | {color:red} kafka-handler: The patch generated 33 new + 0 unchanged 
- 0 fixed = 33 total (was 0) {color} |
| {color:red}-1{color} | {color:red} checkstyle {color} | {color:red}  0m 
14s{color} | {color:red} llap-server: The patch generated 1 new + 26 unchanged 
- 4 fixed = 27 total (was 30) {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} xml {color} | {color:green}  0m  
5s{color} | {color:green} The patch has no ill-formed XML file. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  7m 
36s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  9m 
35s{color} | {color:green} the patch passed {color} |
|| || || || {color:brown} Other Tests {color} ||
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
12s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black} 79m  0s{color} | 
{color:black} {color} |
\\
\\
|| Subsystem || Report/Notes ||
| Optional Tests |  asflicense  javac  javadoc  xml  compile  findbugs  
checkstyle  |
| uname | Linux hiveptest-server-upstream 3.16.0-4-amd64 #1 SMP Debian 
3.16.36-1+deb8u1 (2016-09-03) x86_64 GNU/Linux |
| Build tool | maven |
| Personality | 
/data/hiveptest/working/yetus_PreCommit-HIVE-Build-13517/dev-support/hive-personality.sh
 |
| git revision | master / d2b81a9 |
| Default Java | 1.8.0_111 |
| findbugs | v3.0.0 |
| checkstyle | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-13517/yetus/diff-checkstyle-itests_qtest-druid.txt
 |
| checkstyle | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-13517/yetus/diff-checkstyle-kafka-handler.txt
 |
| checkstyle | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-13517/yetus/diff-checkstyle-llap-server.txt
 |
| modules | C: serde . 

[jira] [Commented] (HIVE-20377) Hive Kafka Storage Handler

2018-08-28 Thread Hive QA (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-20377?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16595808#comment-16595808
 ] 

Hive QA commented on HIVE-20377:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12937508/HIVE-20377.18.patch

{color:green}SUCCESS:{color} +1 due to 8 test(s) being added or modified.

{color:red}ERROR:{color} -1 due to 5 failed/errored test(s), 14900 tests 
executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestMiniDruidCliDriver.testCliDriver[druidmini_dynamic_partition]
 (batchId=193)
org.apache.hadoop.hive.cli.TestMiniDruidCliDriver.testCliDriver[druidmini_expressions]
 (batchId=193)
org.apache.hadoop.hive.cli.TestMiniDruidCliDriver.testCliDriver[druidmini_test1]
 (batchId=193)
org.apache.hadoop.hive.cli.TestMiniDruidCliDriver.testCliDriver[druidmini_test_alter]
 (batchId=193)
org.apache.hadoop.hive.cli.TestMiniDruidCliDriver.testCliDriver[druidmini_test_insert]
 (batchId=193)
{noformat}

Test results: 
https://builds.apache.org/job/PreCommit-HIVE-Build/13517/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/13517/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-13517/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.YetusPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 5 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12937508 - PreCommit-HIVE-Build

> Hive Kafka Storage Handler
> --
>
> Key: HIVE-20377
> URL: https://issues.apache.org/jira/browse/HIVE-20377
> Project: Hive
>  Issue Type: New Feature
>Affects Versions: 4.0.0
>Reporter: slim bouguerra
>Assignee: slim bouguerra
>Priority: Major
> Attachments: HIVE-20377.10.patch, HIVE-20377.11.patch, 
> HIVE-20377.12.patch, HIVE-20377.15.patch, HIVE-20377.18.patch, 
> HIVE-20377.4.patch, HIVE-20377.5.patch, HIVE-20377.6.patch, 
> HIVE-20377.8.patch, HIVE-20377.8.patch, HIVE-20377.patch
>
>
> h1. Goal
> * Read streaming data form Kafka queue as an external table.
> * Allow streaming navigation by pushing down filters on Kafka record 
> partition id, offset and timestamp. 
> * Insert streaming data form Kafka to an actual Hive internal table, using 
> CTAS statement.
> h1. Example
> h2. Create the external table
> {code} 
> CREATE EXTERNAL TABLE kafka_table (`timestamp` timestamp, page string, `user` 
> string, language string, added int, deleted int, flags string,comment string, 
> namespace string)
> STORED BY 'org.apache.hadoop.hive.kafka.KafkaStorageHandler'
> TBLPROPERTIES 
> ("kafka.topic" = "wikipedia", 
> "kafka.bootstrap.servers"="brokeraddress:9092",
> "kafka.serde.class"="org.apache.hadoop.hive.serde2.JsonSerDe");
> {code}
> h2. Kafka Metadata
> In order to keep track of Kafka records the storage handler will add 
> automatically the Kafka row metadata eg partition id, record offset and 
> record timestamp. 
> {code}
> DESCRIBE EXTENDED kafka_table
> timestamp timestamp   from deserializer   
> page  string  from deserializer   
> user  string  from deserializer   
> language  string  from deserializer   
> country   string  from deserializer   
> continent string  from deserializer   
> namespace string  from deserializer   
> newpage   boolean from deserializer   
> unpatrolled   boolean from deserializer   
> anonymous boolean from deserializer   
> robot boolean from deserializer   
> added int from deserializer   
> deleted   int from deserializer   
> delta bigint  from deserializer   
> __partition   int from deserializer   
> __offset  bigint  from deserializer   
> __timestamp   bigint  from deserializer   
> {code}
> h2. Filter push down.
> Newer Kafka consumers 0.11.0 and higher allow seeking on the stream based on 
> a given offset. The proposed storage handler will be able to leverage such 
> API by pushing down filters over metadata columns, namely __partition (int), 
> __offset(long) and __timestamp(long)
> For instance Query like
> {code} 
> select `__offset` from kafka_table where (`__offset` < 10 and `__offset`>3 
> and `__partition` = 0) or 

[jira] [Updated] (HIVE-20483) Really move metastore common classes into metastore-common

2018-08-28 Thread Alexander Kolbasov (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-20483?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alexander Kolbasov updated HIVE-20483:
--
Status: Patch Available  (was: Open)

> Really move metastore common classes into metastore-common
> --
>
> Key: HIVE-20483
> URL: https://issues.apache.org/jira/browse/HIVE-20483
> Project: Hive
>  Issue Type: Sub-task
>  Components: Standalone Metastore
>Affects Versions: 3.0.1, 4.0.0
>Reporter: Alexander Kolbasov
>Assignee: Alexander Kolbasov
>Priority: Major
> Attachments: HIVE-20483.01.patch
>
>
> HIVE-20482 patch was supposed to move a bunch of files from metastore-server 
> to metastore-common but for some reason it didn't happen, so now these files 
> should be moved.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-20483) Really move metastore common classes into metastore-common

2018-08-28 Thread Alexander Kolbasov (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-20483?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alexander Kolbasov updated HIVE-20483:
--
Attachment: HIVE-20483.01.patch

> Really move metastore common classes into metastore-common
> --
>
> Key: HIVE-20483
> URL: https://issues.apache.org/jira/browse/HIVE-20483
> Project: Hive
>  Issue Type: Sub-task
>  Components: Standalone Metastore
>Affects Versions: 3.0.1, 4.0.0
>Reporter: Alexander Kolbasov
>Assignee: Alexander Kolbasov
>Priority: Major
> Attachments: HIVE-20483.01.patch
>
>
> HIVE-20482 patch was supposed to move a bunch of files from metastore-server 
> to metastore-common but for some reason it didn't happen, so now these files 
> should be moved.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-20451) Metastore client and server tarball issues

2018-08-28 Thread Alexander Kolbasov (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-20451?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16595806#comment-16595806
 ] 

Alexander Kolbasov commented on HIVE-20451:
---

[~alangates] Would you be able to review/commit the change? 

> Metastore client and server tarball issues
> --
>
> Key: HIVE-20451
> URL: https://issues.apache.org/jira/browse/HIVE-20451
> Project: Hive
>  Issue Type: Bug
>  Components: Standalone Metastore
>Affects Versions: 4.0.0
>Reporter: Alan Gates
>Assignee: Alexander Kolbasov
>Priority: Major
> Attachments: HIVE-20451.01.patch, HIVE-20451.02.patch, 
> HIVE-20451.03.patch
>
>
> With the split of the metastore into common and server there are now two sets 
> of tarballs.  There are a couple of issues here.
>  # It doesn't make sense to have separate source tarballs for each.  The 
> source release should still be done from the standalone-metastore directory 
> and include all code for the metastore.
>  # The binary tarballs should have separate names.  At the moment both are 
> named apache-hive-metastore.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Assigned] (HIVE-20483) Really move metastore common classes into metastore-common

2018-08-28 Thread Alexander Kolbasov (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-20483?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alexander Kolbasov reassigned HIVE-20483:
-


> Really move metastore common classes into metastore-common
> --
>
> Key: HIVE-20483
> URL: https://issues.apache.org/jira/browse/HIVE-20483
> Project: Hive
>  Issue Type: Sub-task
>  Components: Standalone Metastore
>Affects Versions: 3.0.1, 4.0.0
>Reporter: Alexander Kolbasov
>Assignee: Alexander Kolbasov
>Priority: Major
>
> HIVE-20482 patch was supposed to move a bunch of files from metastore-server 
> to metastore-common but for some reason it didn't happen, so now these files 
> should be moved.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Assigned] (HIVE-20482) Remove dependency on metastore-server

2018-08-28 Thread Alexander Kolbasov (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-20482?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alexander Kolbasov reassigned HIVE-20482:
-


> Remove dependency on metastore-server
> -
>
> Key: HIVE-20482
> URL: https://issues.apache.org/jira/browse/HIVE-20482
> Project: Hive
>  Issue Type: Sub-task
>  Components: Standalone Metastore
>Affects Versions: 3.0.1, 4.0.0
>Reporter: Alexander Kolbasov
>Assignee: Alexander Kolbasov
>Priority: Major
>
> Now that we separated common and server classes we should remove dependency 
> on the server module from poms.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-20472) mvn test failing for metastore-tool module

2018-08-28 Thread Alexander Kolbasov (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-20472?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16595764#comment-16595764
 ] 

Alexander Kolbasov commented on HIVE-20472:
---

[~pvary] Can you commit the patch?

> mvn test failing for metastore-tool module
> --
>
> Key: HIVE-20472
> URL: https://issues.apache.org/jira/browse/HIVE-20472
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 4.0.0
>Reporter: Vihang Karajgaonkar
>Assignee: Alexander Kolbasov
>Priority: Trivial
> Fix For: 4.0.0
>
> Attachments: HIVE-20472.01.patch, HIVE-20472.02.patch
>
>
> Fails because there are no applicable tests.
>  
> {code}
> [ERROR] Failed to execute goal 
> org.apache.maven.plugins:maven-surefire-plugin:2.20.1:test (default-test) on 
> project hive-metastore-benchmarks: No tests were executed! (Set 
> -DfailIfNoTests=false to ignore this error.) -> [Help 1]
> [ERROR]
> [ERROR] To see the full stack trace of the errors, re-run Maven with the -e 
> switch.
> [ERROR] Re-run Maven using the -X switch to enable full debug logging.
> [ERROR]
> [ERROR] For more information about the errors and possible solutions, please 
> read the following articles:
> [ERROR] [Help 1] 
> http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException
> [ERROR]
> [ERROR] After correcting the problems, you can resume the build with the 
> command
> [ERROR] mvn  -rf :hive-metastore-benchmarks
> {code}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-20451) Metastore client and server tarball issues

2018-08-28 Thread Alexander Kolbasov (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-20451?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alexander Kolbasov updated HIVE-20451:
--
Attachment: HIVE-20451.03.patch

> Metastore client and server tarball issues
> --
>
> Key: HIVE-20451
> URL: https://issues.apache.org/jira/browse/HIVE-20451
> Project: Hive
>  Issue Type: Bug
>  Components: Standalone Metastore
>Affects Versions: 4.0.0
>Reporter: Alan Gates
>Assignee: Alexander Kolbasov
>Priority: Major
> Attachments: HIVE-20451.01.patch, HIVE-20451.02.patch, 
> HIVE-20451.03.patch
>
>
> With the split of the metastore into common and server there are now two sets 
> of tarballs.  There are a couple of issues here.
>  # It doesn't make sense to have separate source tarballs for each.  The 
> source release should still be done from the standalone-metastore directory 
> and include all code for the metastore.
>  # The binary tarballs should have separate names.  At the moment both are 
> named apache-hive-metastore.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-20451) Metastore client and server tarball issues

2018-08-28 Thread Alexander Kolbasov (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-20451?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16595762#comment-16595762
 ] 

Alexander Kolbasov commented on HIVE-20451:
---

Finally figured out the way to suppress generation of source tarballs in 
submodules, so with the latest patch a single source tarball is generated in 
standalone-metastore/target directory.

> Metastore client and server tarball issues
> --
>
> Key: HIVE-20451
> URL: https://issues.apache.org/jira/browse/HIVE-20451
> Project: Hive
>  Issue Type: Bug
>  Components: Standalone Metastore
>Affects Versions: 4.0.0
>Reporter: Alan Gates
>Assignee: Alexander Kolbasov
>Priority: Major
> Attachments: HIVE-20451.01.patch, HIVE-20451.02.patch, 
> HIVE-20451.03.patch
>
>
> With the split of the metastore into common and server there are now two sets 
> of tarballs.  There are a couple of issues here.
>  # It doesn't make sense to have separate source tarballs for each.  The 
> source release should still be done from the standalone-metastore directory 
> and include all code for the metastore.
>  # The binary tarballs should have separate names.  At the moment both are 
> named apache-hive-metastore.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-20472) mvn test failing for metastore-tool module

2018-08-28 Thread Hive QA (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-20472?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16595756#comment-16595756
 ] 

Hive QA commented on HIVE-20472:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12937473/HIVE-20472.02.patch

{color:red}ERROR:{color} -1 due to no test(s) being added or modified.

{color:green}SUCCESS:{color} +1 due to 14899 tests passed

Test results: 
https://builds.apache.org/job/PreCommit-HIVE-Build/13516/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/13516/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-13516/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.YetusPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12937473 - PreCommit-HIVE-Build

> mvn test failing for metastore-tool module
> --
>
> Key: HIVE-20472
> URL: https://issues.apache.org/jira/browse/HIVE-20472
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 4.0.0
>Reporter: Vihang Karajgaonkar
>Assignee: Alexander Kolbasov
>Priority: Trivial
> Fix For: 4.0.0
>
> Attachments: HIVE-20472.01.patch, HIVE-20472.02.patch
>
>
> Fails because there are no applicable tests.
>  
> {code}
> [ERROR] Failed to execute goal 
> org.apache.maven.plugins:maven-surefire-plugin:2.20.1:test (default-test) on 
> project hive-metastore-benchmarks: No tests were executed! (Set 
> -DfailIfNoTests=false to ignore this error.) -> [Help 1]
> [ERROR]
> [ERROR] To see the full stack trace of the errors, re-run Maven with the -e 
> switch.
> [ERROR] Re-run Maven using the -X switch to enable full debug logging.
> [ERROR]
> [ERROR] For more information about the errors and possible solutions, please 
> read the following articles:
> [ERROR] [Help 1] 
> http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException
> [ERROR]
> [ERROR] After correcting the problems, you can resume the build with the 
> command
> [ERROR] mvn  -rf :hive-metastore-benchmarks
> {code}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-20291) Allow HiveStreamingConnection to receive a WriteId

2018-08-28 Thread Jaume M (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-20291?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jaume M updated HIVE-20291:
---
Attachment: HIVE-20291.3.patch
Status: Patch Available  (was: Open)

> Allow HiveStreamingConnection to receive a WriteId
> --
>
> Key: HIVE-20291
> URL: https://issues.apache.org/jira/browse/HIVE-20291
> Project: Hive
>  Issue Type: Improvement
>Reporter: Jaume M
>Assignee: Jaume M
>Priority: Major
>  Labels: pull-request-available
> Attachments: HIVE-20291.1.patch, HIVE-20291.2.patch, 
> HIVE-20291.3.patch
>
>
> If the writeId is received externally it won't need to open connections to 
> the metastore. It won't be able to the commit in this case as well so it must 
> be done by the entity passing the writeId.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-20291) Allow HiveStreamingConnection to receive a WriteId

2018-08-28 Thread Jaume M (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-20291?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jaume M updated HIVE-20291:
---
Attachment: TEZ-3985.3.patch
Status: Patch Available  (was: Open)

> Allow HiveStreamingConnection to receive a WriteId
> --
>
> Key: HIVE-20291
> URL: https://issues.apache.org/jira/browse/HIVE-20291
> Project: Hive
>  Issue Type: Improvement
>Reporter: Jaume M
>Assignee: Jaume M
>Priority: Major
>  Labels: pull-request-available
> Attachments: HIVE-20291.1.patch, HIVE-20291.2.patch
>
>
> If the writeId is received externally it won't need to open connections to 
> the metastore. It won't be able to the commit in this case as well so it must 
> be done by the entity passing the writeId.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-20291) Allow HiveStreamingConnection to receive a WriteId

2018-08-28 Thread Jaume M (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-20291?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jaume M updated HIVE-20291:
---
Attachment: (was: TEZ-3985.3.patch)

> Allow HiveStreamingConnection to receive a WriteId
> --
>
> Key: HIVE-20291
> URL: https://issues.apache.org/jira/browse/HIVE-20291
> Project: Hive
>  Issue Type: Improvement
>Reporter: Jaume M
>Assignee: Jaume M
>Priority: Major
>  Labels: pull-request-available
> Attachments: HIVE-20291.1.patch, HIVE-20291.2.patch
>
>
> If the writeId is received externally it won't need to open connections to 
> the metastore. It won't be able to the commit in this case as well so it must 
> be done by the entity passing the writeId.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-20291) Allow HiveStreamingConnection to receive a WriteId

2018-08-28 Thread Jaume M (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-20291?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jaume M updated HIVE-20291:
---
Status: Open  (was: Patch Available)

> Allow HiveStreamingConnection to receive a WriteId
> --
>
> Key: HIVE-20291
> URL: https://issues.apache.org/jira/browse/HIVE-20291
> Project: Hive
>  Issue Type: Improvement
>Reporter: Jaume M
>Assignee: Jaume M
>Priority: Major
>  Labels: pull-request-available
> Attachments: HIVE-20291.1.patch, HIVE-20291.2.patch
>
>
> If the writeId is received externally it won't need to open connections to 
> the metastore. It won't be able to the commit in this case as well so it must 
> be done by the entity passing the writeId.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-20291) Allow HiveStreamingConnection to receive a WriteId

2018-08-28 Thread Jaume M (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-20291?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jaume M updated HIVE-20291:
---
Status: Open  (was: Patch Available)

> Allow HiveStreamingConnection to receive a WriteId
> --
>
> Key: HIVE-20291
> URL: https://issues.apache.org/jira/browse/HIVE-20291
> Project: Hive
>  Issue Type: Improvement
>Reporter: Jaume M
>Assignee: Jaume M
>Priority: Major
>  Labels: pull-request-available
> Attachments: HIVE-20291.1.patch, HIVE-20291.2.patch
>
>
> If the writeId is received externally it won't need to open connections to 
> the metastore. It won't be able to the commit in this case as well so it must 
> be done by the entity passing the writeId.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-18908) FULL OUTER JOIN to MapJoin

2018-08-28 Thread Matt McCline (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-18908?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Matt McCline updated HIVE-18908:

Status: Patch Available  (was: In Progress)

> FULL OUTER JOIN to MapJoin
> --
>
> Key: HIVE-18908
> URL: https://issues.apache.org/jira/browse/HIVE-18908
> Project: Hive
>  Issue Type: Improvement
>  Components: Hive
>Reporter: Matt McCline
>Assignee: Matt McCline
>Priority: Critical
> Attachments: FULL OUTER MapJoin Code Changes.docx, 
> HIVE-18908.01.patch, HIVE-18908.02.patch, HIVE-18908.03.patch, 
> HIVE-18908.04.patch, HIVE-18908.05.patch, HIVE-18908.06.patch, 
> HIVE-18908.08.patch, HIVE-18908.09.patch, HIVE-18908.091.patch, 
> HIVE-18908.092.patch, HIVE-18908.093.patch, HIVE-18908.096.patch, 
> HIVE-18908.097.patch, HIVE-18908.098.patch, HIVE-18908.099.patch, JOIN to 
> MAPJOIN Transformation.pdf, SHARED-MEMORY FULL OUTER MapJoin.pdf
>
>
> Currently, we do not support FULL OUTER JOIN in MapJoin.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-18908) FULL OUTER JOIN to MapJoin

2018-08-28 Thread Matt McCline (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-18908?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Matt McCline updated HIVE-18908:

Attachment: HIVE-18908.099.patch

> FULL OUTER JOIN to MapJoin
> --
>
> Key: HIVE-18908
> URL: https://issues.apache.org/jira/browse/HIVE-18908
> Project: Hive
>  Issue Type: Improvement
>  Components: Hive
>Reporter: Matt McCline
>Assignee: Matt McCline
>Priority: Critical
> Attachments: FULL OUTER MapJoin Code Changes.docx, 
> HIVE-18908.01.patch, HIVE-18908.02.patch, HIVE-18908.03.patch, 
> HIVE-18908.04.patch, HIVE-18908.05.patch, HIVE-18908.06.patch, 
> HIVE-18908.08.patch, HIVE-18908.09.patch, HIVE-18908.091.patch, 
> HIVE-18908.092.patch, HIVE-18908.093.patch, HIVE-18908.096.patch, 
> HIVE-18908.097.patch, HIVE-18908.098.patch, HIVE-18908.099.patch, JOIN to 
> MAPJOIN Transformation.pdf, SHARED-MEMORY FULL OUTER MapJoin.pdf
>
>
> Currently, we do not support FULL OUTER JOIN in MapJoin.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-18908) FULL OUTER JOIN to MapJoin

2018-08-28 Thread Matt McCline (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-18908?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Matt McCline updated HIVE-18908:

Status: In Progress  (was: Patch Available)

> FULL OUTER JOIN to MapJoin
> --
>
> Key: HIVE-18908
> URL: https://issues.apache.org/jira/browse/HIVE-18908
> Project: Hive
>  Issue Type: Improvement
>  Components: Hive
>Reporter: Matt McCline
>Assignee: Matt McCline
>Priority: Critical
> Attachments: FULL OUTER MapJoin Code Changes.docx, 
> HIVE-18908.01.patch, HIVE-18908.02.patch, HIVE-18908.03.patch, 
> HIVE-18908.04.patch, HIVE-18908.05.patch, HIVE-18908.06.patch, 
> HIVE-18908.08.patch, HIVE-18908.09.patch, HIVE-18908.091.patch, 
> HIVE-18908.092.patch, HIVE-18908.093.patch, HIVE-18908.096.patch, 
> HIVE-18908.097.patch, HIVE-18908.098.patch, JOIN to MAPJOIN 
> Transformation.pdf, SHARED-MEMORY FULL OUTER MapJoin.pdf
>
>
> Currently, we do not support FULL OUTER JOIN in MapJoin.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-20472) mvn test failing for metastore-tool module

2018-08-28 Thread Hive QA (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-20472?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16595721#comment-16595721
 ] 

Hive QA commented on HIVE-20472:


| (/) *{color:green}+1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
|| || || || {color:brown} Prechecks {color} ||
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
|| || || || {color:brown} master Compile Tests {color} ||
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  0m 
31s{color} | {color:blue} Maven dependency ordering for branch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  8m 
10s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  2m 
23s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  2m 
40s{color} | {color:green} master passed {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  0m  
7s{color} | {color:blue} Maven dependency ordering for patch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  2m 
45s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  2m 
23s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  2m 
23s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} xml {color} | {color:green}  0m  
4s{color} | {color:green} The patch has no ill-formed XML file. {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  2m 
37s{color} | {color:green} the patch passed {color} |
|| || || || {color:brown} Other Tests {color} ||
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
13s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black} 22m 30s{color} | 
{color:black} {color} |
\\
\\
|| Subsystem || Report/Notes ||
| Optional Tests |  asflicense  javac  javadoc  xml  compile  |
| uname | Linux hiveptest-server-upstream 3.16.0-4-amd64 #1 SMP Debian 
3.16.36-1+deb8u1 (2016-09-03) x86_64 GNU/Linux |
| Build tool | maven |
| Personality | 
/data/hiveptest/working/yetus_PreCommit-HIVE-Build-13516/dev-support/hive-personality.sh
 |
| git revision | master / d2b81a9 |
| Default Java | 1.8.0_111 |
| modules | C: standalone-metastore standalone-metastore/metastore-common 
standalone-metastore/metastore-server standalone-metastore/metastore-tools U: 
standalone-metastore |
| Console output | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-13516/yetus.txt |
| Powered by | Apache Yetushttp://yetus.apache.org |


This message was automatically generated.



> mvn test failing for metastore-tool module
> --
>
> Key: HIVE-20472
> URL: https://issues.apache.org/jira/browse/HIVE-20472
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 4.0.0
>Reporter: Vihang Karajgaonkar
>Assignee: Alexander Kolbasov
>Priority: Trivial
> Fix For: 4.0.0
>
> Attachments: HIVE-20472.01.patch, HIVE-20472.02.patch
>
>
> Fails because there are no applicable tests.
>  
> {code}
> [ERROR] Failed to execute goal 
> org.apache.maven.plugins:maven-surefire-plugin:2.20.1:test (default-test) on 
> project hive-metastore-benchmarks: No tests were executed! (Set 
> -DfailIfNoTests=false to ignore this error.) -> [Help 1]
> [ERROR]
> [ERROR] To see the full stack trace of the errors, re-run Maven with the -e 
> switch.
> [ERROR] Re-run Maven using the -X switch to enable full debug logging.
> [ERROR]
> [ERROR] For more information about the errors and possible solutions, please 
> read the following articles:
> [ERROR] [Help 1] 
> http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException
> [ERROR]
> [ERROR] After correcting the problems, you can resume the build with the 
> command
> [ERROR] mvn  -rf :hive-metastore-benchmarks
> {code}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-18908) FULL OUTER JOIN to MapJoin

2018-08-28 Thread Hive QA (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-18908?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16595688#comment-16595688
 ] 

Hive QA commented on HIVE-18908:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12937469/HIVE-18908.098.patch

{color:red}ERROR:{color} -1 due to build exiting with an error

Test results: 
https://builds.apache.org/job/PreCommit-HIVE-Build/13515/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/13515/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-13515/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Tests exited with: NonZeroExitCodeException
Command 'bash /data/hiveptest/working/scratch/source-prep.sh' failed with exit 
status 1 and output '+ date '+%Y-%m-%d %T.%3N'
2018-08-28 22:54:14.251
+ [[ -n /usr/lib/jvm/java-8-openjdk-amd64 ]]
+ export JAVA_HOME=/usr/lib/jvm/java-8-openjdk-amd64
+ JAVA_HOME=/usr/lib/jvm/java-8-openjdk-amd64
+ export 
PATH=/usr/lib/jvm/java-8-openjdk-amd64/bin/:/usr/local/bin:/usr/bin:/bin:/usr/local/games:/usr/games
+ 
PATH=/usr/lib/jvm/java-8-openjdk-amd64/bin/:/usr/local/bin:/usr/bin:/bin:/usr/local/games:/usr/games
+ export 'ANT_OPTS=-Xmx1g -XX:MaxPermSize=256m '
+ ANT_OPTS='-Xmx1g -XX:MaxPermSize=256m '
+ export 'MAVEN_OPTS=-Xmx1g '
+ MAVEN_OPTS='-Xmx1g '
+ cd /data/hiveptest/working/
+ tee /data/hiveptest/logs/PreCommit-HIVE-Build-13515/source-prep.txt
+ [[ false == \t\r\u\e ]]
+ mkdir -p maven ivy
+ [[ git = \s\v\n ]]
+ [[ git = \g\i\t ]]
+ [[ -z master ]]
+ [[ -d apache-github-source-source ]]
+ [[ ! -d apache-github-source-source/.git ]]
+ [[ ! -d apache-github-source-source ]]
+ date '+%Y-%m-%d %T.%3N'
2018-08-28 22:54:14.255
+ cd apache-github-source-source
+ git fetch origin
+ git reset --hard HEAD
HEAD is now at d2b81a9 HIVE-19674: Group by Decimal Constants push down to 
Druid tables(Vineet Garg, reviewed by Ashutosh Chauhan)
+ git clean -f -d
+ git checkout master
Already on 'master'
Your branch is up-to-date with 'origin/master'.
+ git reset --hard origin/master
HEAD is now at d2b81a9 HIVE-19674: Group by Decimal Constants push down to 
Druid tables(Vineet Garg, reviewed by Ashutosh Chauhan)
+ git merge --ff-only origin/master
Already up-to-date.
+ date '+%Y-%m-%d %T.%3N'
2018-08-28 22:54:14.923
+ rm -rf ../yetus_PreCommit-HIVE-Build-13515
+ mkdir ../yetus_PreCommit-HIVE-Build-13515
+ git gc
+ cp -R . ../yetus_PreCommit-HIVE-Build-13515
+ mkdir /data/hiveptest/logs/PreCommit-HIVE-Build-13515/yetus
+ patchCommandPath=/data/hiveptest/working/scratch/smart-apply-patch.sh
+ patchFilePath=/data/hiveptest/working/scratch/build.patch
+ [[ -f /data/hiveptest/working/scratch/build.patch ]]
+ chmod +x /data/hiveptest/working/scratch/smart-apply-patch.sh
+ /data/hiveptest/working/scratch/smart-apply-patch.sh 
/data/hiveptest/working/scratch/build.patch
error: patch failed: 
ql/src/test/results/clientpositive/llap/vector_groupby_sort_11.q.out:693
Falling back to three-way merge...
Applied patch to 
'ql/src/test/results/clientpositive/llap/vector_groupby_sort_11.q.out' cleanly.
error: patch failed: 
ql/src/test/results/clientpositive/spark/spark_vectorized_dynamic_partition_pruning.q.out:4818
Falling back to three-way merge...
Applied patch to 
'ql/src/test/results/clientpositive/spark/spark_vectorized_dynamic_partition_pruning.q.out'
 with conflicts.
Going to apply patch with: git apply -p0
/data/hiveptest/working/scratch/build.patch:18802: trailing whitespace.
-- FROM 
/data/hiveptest/working/scratch/build.patch:18804: trailing whitespace.
-- JOIN 
/data/hiveptest/working/scratch/build.patch:18818: trailing whitespace.
-- FROM 
/data/hiveptest/working/scratch/build.patch:18820: trailing whitespace.
-- LEFT OUTER JOIN 
/data/hiveptest/working/scratch/build.patch:18835: trailing whitespace.
-- FROM 
error: patch failed: 
ql/src/test/results/clientpositive/llap/vector_groupby_sort_11.q.out:693
Falling back to three-way merge...
Applied patch to 
'ql/src/test/results/clientpositive/llap/vector_groupby_sort_11.q.out' cleanly.
error: patch failed: 
ql/src/test/results/clientpositive/spark/spark_vectorized_dynamic_partition_pruning.q.out:4818
Falling back to three-way merge...
Applied patch to 
'ql/src/test/results/clientpositive/spark/spark_vectorized_dynamic_partition_pruning.q.out'
 with conflicts.
U 
ql/src/test/results/clientpositive/spark/spark_vectorized_dynamic_partition_pruning.q.out
warning: squelched 641 whitespace errors
warning: 646 lines add whitespace errors.
+ result=1
+ '[' 1 -ne 0 ']'
+ rm -rf yetus_PreCommit-HIVE-Build-13515
+ exit 1
'
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12937469 - PreCommit-HIVE-Build

> FULL OUTER JOIN to MapJoin
> --
>
> Key: HIVE-18908
> URL: 

[jira] [Commented] (HIVE-20476) CopyUtils used by REPL LOAD and EXPORT/IMPORT operations ignore distcp error.

2018-08-28 Thread Hive QA (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-20476?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16595682#comment-16595682
 ] 

Hive QA commented on HIVE-20476:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12937461/HIVE-20476.01.patch

{color:green}SUCCESS:{color} +1 due to 1 test(s) being added or modified.

{color:red}ERROR:{color} -1 due to 1 failed/errored test(s), 14900 tests 
executed
*Failed tests:*
{noformat}
org.apache.hive.jdbc.miniHS2.TestHs2ConnectionMetricsBinary.testOpenConnectionMetrics
 (batchId=255)
{noformat}

Test results: 
https://builds.apache.org/job/PreCommit-HIVE-Build/13513/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/13513/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-13513/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.YetusPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 1 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12937461 - PreCommit-HIVE-Build

> CopyUtils used by REPL LOAD and EXPORT/IMPORT operations ignore distcp error.
> -
>
> Key: HIVE-20476
> URL: https://issues.apache.org/jira/browse/HIVE-20476
> Project: Hive
>  Issue Type: Bug
>  Components: HiveServer2, repl
>Affects Versions: 3.1.0
>Reporter: Sankar Hariappan
>Assignee: Sankar Hariappan
>Priority: Major
>  Labels: DR, Export, Import, Replication, pull-request-available
> Attachments: HIVE-20476.01.patch
>
>
> CopyUtils uses FileUtils.distCp to copy files but doesn't check the return 
> value. It returns false if the copy fails.
> Now, REPL LOAD and EXPORT/IMPORT commands internally uses CopyUtils to copy 
> data files across clusters and here it may return success even if file copy 
> fails and may cause data loss.
> Need to throw error and retry.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Comment Edited] (HIVE-20451) Metastore client and server tarball issues

2018-08-28 Thread Alexander Kolbasov (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-20451?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16595648#comment-16595648
 ] 

Alexander Kolbasov edited comment on HIVE-20451 at 8/28/18 9:54 PM:


Posted an updated patch that seems a bit cleaner. Here is what it does.

# Only one binary tarball is generated in metastore-server target directory. It 
is called 
{{apache-hive-standalone-metastore-server-4.0.0-SNAPSHOT-bin.tar.gz}}. I don't 
think that we need binary tarbal for the metastore-common - there are no 
binaries to run from there.
# {{standalone-metastore/target}} has the overall tarball called 
{{apache-hive-standalone-metastore-4.0.0-SNAPSHOT-src.tar.gz}} which includes 
full source and can be used to build standalone metastore.
# submodules have their own source tarballs that can be used to build these 
submodules. They all have different names. I don't know how the release process 
works, but basically it may choose to exclude these from the release. 

Is this an acceptable state of affairs?


was (Author: akolb):
Posted an updated patch that seems a bit cleaner. Here is what it does.

# Only one binary tarball is generated in metastore-server target directory. It 
is called 
{{apache-hive-standalone-metastore-server-4.0.0-SNAPSHOT-bin.tar.gz}}. I don't 
think that we need binary tarbal for the metastore-common - there are no 
binaries to run from there.
# {{standalone-metastore/target}} has the overall tarball called 
{{apache-hive-standalone-metastore-4.0.0-SNAPSHOT-src.tar.gz
}} which includes full source and can be used to build standalone metastore.
# submodules have their own source tarballs that can be used to build these 
submodules. They all have different names. I don't know how the release process 
works, but basically it may choose to exclude these from the release. 

Is this an acceptable state of affairs?

> Metastore client and server tarball issues
> --
>
> Key: HIVE-20451
> URL: https://issues.apache.org/jira/browse/HIVE-20451
> Project: Hive
>  Issue Type: Bug
>  Components: Standalone Metastore
>Affects Versions: 4.0.0
>Reporter: Alan Gates
>Assignee: Alexander Kolbasov
>Priority: Major
> Attachments: HIVE-20451.01.patch, HIVE-20451.02.patch
>
>
> With the split of the metastore into common and server there are now two sets 
> of tarballs.  There are a couple of issues here.
>  # It doesn't make sense to have separate source tarballs for each.  The 
> source release should still be done from the standalone-metastore directory 
> and include all code for the metastore.
>  # The binary tarballs should have separate names.  At the moment both are 
> named apache-hive-metastore.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-20451) Metastore client and server tarball issues

2018-08-28 Thread Alexander Kolbasov (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-20451?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16595648#comment-16595648
 ] 

Alexander Kolbasov commented on HIVE-20451:
---

Posted an updated patch that seems a bit cleaner. Here is what it does.

# Only one binary tarball is generated in metastore-server target directory. It 
is called 
{{apache-hive-standalone-metastore-server-4.0.0-SNAPSHOT-bin.tar.gz}}. I don't 
think that we need binary tarbal for the metastore-common - there are no 
binaries to run from there.
# {{standalone-metastore/target}} has the overall tarball called 
{{apache-hive-standalone-metastore-4.0.0-SNAPSHOT-src.tar.gz
}} which includes full source and can be used to build standalone metastore.
# submodules have their own source tarballs that can be used to build these 
submodules. They all have different names. I don't know how the release process 
works, but basically it may choose to exclude these from the release. 

Is this an acceptable state of affairs?

> Metastore client and server tarball issues
> --
>
> Key: HIVE-20451
> URL: https://issues.apache.org/jira/browse/HIVE-20451
> Project: Hive
>  Issue Type: Bug
>  Components: Standalone Metastore
>Affects Versions: 4.0.0
>Reporter: Alan Gates
>Assignee: Alexander Kolbasov
>Priority: Major
> Attachments: HIVE-20451.01.patch, HIVE-20451.02.patch
>
>
> With the split of the metastore into common and server there are now two sets 
> of tarballs.  There are a couple of issues here.
>  # It doesn't make sense to have separate source tarballs for each.  The 
> source release should still be done from the standalone-metastore directory 
> and include all code for the metastore.
>  # The binary tarballs should have separate names.  At the moment both are 
> named apache-hive-metastore.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-20451) Metastore client and server tarball issues

2018-08-28 Thread Alexander Kolbasov (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-20451?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alexander Kolbasov updated HIVE-20451:
--
Attachment: HIVE-20451.02.patch

> Metastore client and server tarball issues
> --
>
> Key: HIVE-20451
> URL: https://issues.apache.org/jira/browse/HIVE-20451
> Project: Hive
>  Issue Type: Bug
>  Components: Standalone Metastore
>Affects Versions: 4.0.0
>Reporter: Alan Gates
>Assignee: Alexander Kolbasov
>Priority: Major
> Attachments: HIVE-20451.01.patch, HIVE-20451.02.patch
>
>
> With the split of the metastore into common and server there are now two sets 
> of tarballs.  There are a couple of issues here.
>  # It doesn't make sense to have separate source tarballs for each.  The 
> source release should still be done from the standalone-metastore directory 
> and include all code for the metastore.
>  # The binary tarballs should have separate names.  At the moment both are 
> named apache-hive-metastore.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-20377) Hive Kafka Storage Handler

2018-08-28 Thread Gopal V (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-20377?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16595622#comment-16595622
 ] 

Gopal V commented on HIVE-20377:


+1 tests pending

> Hive Kafka Storage Handler
> --
>
> Key: HIVE-20377
> URL: https://issues.apache.org/jira/browse/HIVE-20377
> Project: Hive
>  Issue Type: New Feature
>Affects Versions: 4.0.0
>Reporter: slim bouguerra
>Assignee: slim bouguerra
>Priority: Major
> Attachments: HIVE-20377.10.patch, HIVE-20377.11.patch, 
> HIVE-20377.12.patch, HIVE-20377.15.patch, HIVE-20377.18.patch, 
> HIVE-20377.4.patch, HIVE-20377.5.patch, HIVE-20377.6.patch, 
> HIVE-20377.8.patch, HIVE-20377.8.patch, HIVE-20377.patch
>
>
> h1. Goal
> * Read streaming data form Kafka queue as an external table.
> * Allow streaming navigation by pushing down filters on Kafka record 
> partition id, offset and timestamp. 
> * Insert streaming data form Kafka to an actual Hive internal table, using 
> CTAS statement.
> h1. Example
> h2. Create the external table
> {code} 
> CREATE EXTERNAL TABLE kafka_table (`timestamp` timestamp, page string, `user` 
> string, language string, added int, deleted int, flags string,comment string, 
> namespace string)
> STORED BY 'org.apache.hadoop.hive.kafka.KafkaStorageHandler'
> TBLPROPERTIES 
> ("kafka.topic" = "wikipedia", 
> "kafka.bootstrap.servers"="brokeraddress:9092",
> "kafka.serde.class"="org.apache.hadoop.hive.serde2.JsonSerDe");
> {code}
> h2. Kafka Metadata
> In order to keep track of Kafka records the storage handler will add 
> automatically the Kafka row metadata eg partition id, record offset and 
> record timestamp. 
> {code}
> DESCRIBE EXTENDED kafka_table
> timestamp timestamp   from deserializer   
> page  string  from deserializer   
> user  string  from deserializer   
> language  string  from deserializer   
> country   string  from deserializer   
> continent string  from deserializer   
> namespace string  from deserializer   
> newpage   boolean from deserializer   
> unpatrolled   boolean from deserializer   
> anonymous boolean from deserializer   
> robot boolean from deserializer   
> added int from deserializer   
> deleted   int from deserializer   
> delta bigint  from deserializer   
> __partition   int from deserializer   
> __offset  bigint  from deserializer   
> __timestamp   bigint  from deserializer   
> {code}
> h2. Filter push down.
> Newer Kafka consumers 0.11.0 and higher allow seeking on the stream based on 
> a given offset. The proposed storage handler will be able to leverage such 
> API by pushing down filters over metadata columns, namely __partition (int), 
> __offset(long) and __timestamp(long)
> For instance Query like
> {code} 
> select `__offset` from kafka_table where (`__offset` < 10 and `__offset`>3 
> and `__partition` = 0) or (`__partition` = 0 and `__offset` < 105 and 
> `__offset` > 99) or (`__offset` = 109);
> {code}
> Will result on a scan of partition 0 only then read only records between 
> offset 4 and 109. 
> h2. With timestamp seeks 
> The seeking based on the internal timestamps allows the handler to run on 
> recently arrived data, by doing
> {code}
> select count(*) from kafka_table where `__timestamp` >  1000 * 
> to_unix_timestamp(CURRENT_TIMESTAMP - interval '20' hours) ;
> {code}
> This allows for implicit relationships between event timestamps and kafka 
> timestamps to be expressed in queries (i.e event_timestamp is always < than 
> kafka __timestamp and kafka __timestamp is never > 15 minutes from event etc).
> h2. More examples with Avro 
> {code}
> CREATE EXTERNAL TABLE wiki_kafka_avro_table
> STORED BY 'org.apache.hadoop.hive.kafka.KafkaStorageHandler'
> TBLPROPERTIES
> ("kafka.topic" = "wiki_kafka_avro_table",
> "kafka.bootstrap.servers"="localhost:9092",
> "kafka.serde.class"="org.apache.hadoop.hive.serde2.avro.AvroSerDe",
> 'avro.schema.literal'='{
>   "type" : "record",
>   "name" : "Wikipedia",
>   "namespace" : "org.apache.hive.kafka",
>   "version": "1",
>   "fields" : [ {
> "name" : "isrobot",
> "type" : "boolean"
>   }, {
> "name" : "channel",
> "type" : "string"
>   }, {
> "name" : "timestamp",
> "type" : "string"
>   }, {
> "name" : "flags",
> "type" : "string"
>   }, {
> "name" : "isunpatrolled",
> "type" : "boolean"
>   }, {
> 

[jira] [Commented] (HIVE-20476) CopyUtils used by REPL LOAD and EXPORT/IMPORT operations ignore distcp error.

2018-08-28 Thread Hive QA (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-20476?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16595618#comment-16595618
 ] 

Hive QA commented on HIVE-20476:


| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
|| || || || {color:brown} Prechecks {color} ||
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
|| || || || {color:brown} master Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  8m 
42s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  1m  
5s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
42s{color} | {color:green} master passed {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  4m  
9s{color} | {color:blue} ql in master has 2310 extant Findbugs warnings. 
{color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
59s{color} | {color:green} master passed {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  1m 
29s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  1m  
9s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  1m  
9s{color} | {color:green} the patch passed {color} |
| {color:red}-1{color} | {color:red} checkstyle {color} | {color:red}  0m 
42s{color} | {color:red} ql: The patch generated 3 new + 227 unchanged - 1 
fixed = 230 total (was 228) {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  4m 
22s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  1m  
0s{color} | {color:green} the patch passed {color} |
|| || || || {color:brown} Other Tests {color} ||
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
13s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black} 25m  4s{color} | 
{color:black} {color} |
\\
\\
|| Subsystem || Report/Notes ||
| Optional Tests |  asflicense  javac  javadoc  findbugs  checkstyle  compile  |
| uname | Linux hiveptest-server-upstream 3.16.0-4-amd64 #1 SMP Debian 
3.16.36-1+deb8u1 (2016-09-03) x86_64 GNU/Linux |
| Build tool | maven |
| Personality | 
/data/hiveptest/working/yetus_PreCommit-HIVE-Build-13513/dev-support/hive-personality.sh
 |
| git revision | master / d2b81a9 |
| Default Java | 1.8.0_111 |
| findbugs | v3.0.0 |
| checkstyle | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-13513/yetus/diff-checkstyle-ql.txt
 |
| modules | C: ql U: ql |
| Console output | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-13513/yetus.txt |
| Powered by | Apache Yetushttp://yetus.apache.org |


This message was automatically generated.



> CopyUtils used by REPL LOAD and EXPORT/IMPORT operations ignore distcp error.
> -
>
> Key: HIVE-20476
> URL: https://issues.apache.org/jira/browse/HIVE-20476
> Project: Hive
>  Issue Type: Bug
>  Components: HiveServer2, repl
>Affects Versions: 3.1.0
>Reporter: Sankar Hariappan
>Assignee: Sankar Hariappan
>Priority: Major
>  Labels: DR, Export, Import, Replication, pull-request-available
> Attachments: HIVE-20476.01.patch
>
>
> CopyUtils uses FileUtils.distCp to copy files but doesn't check the return 
> value. It returns false if the copy fails.
> Now, REPL LOAD and EXPORT/IMPORT commands internally uses CopyUtils to copy 
> data files across clusters and here it may return success even if file copy 
> fails and may cause data loss.
> Need to throw error and retry.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Assigned] (HIVE-20481) Add the Kafka Key record as part of the row.

2018-08-28 Thread slim bouguerra (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-20481?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

slim bouguerra reassigned HIVE-20481:
-


> Add the Kafka Key record as part of the row.
> 
>
> Key: HIVE-20481
> URL: https://issues.apache.org/jira/browse/HIVE-20481
> Project: Hive
>  Issue Type: Sub-task
>Reporter: slim bouguerra
>Assignee: slim bouguerra
>Priority: Major
>
> Kafka records are keyed, most of the case this key is null or used to route 
> records to the same partition. This patch adds this column as a binary column 
> {code} __record_key{code}.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-20480) Implement column stats annotation rules for the UDTFOperator: Follow up for HIVE-20262

2018-08-28 Thread George Pachitariu (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-20480?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16595588#comment-16595588
 ] 

George Pachitariu commented on HIVE-20480:
--

Hi [~ashutoshc],

This is the follow-up task with a patch for your comment here: HIVE-20262

Is this what you meant?

Can you also please give me an idea on how I could test this?

 

Sorry for taking this long to come back to you.

George :)

> Implement column stats annotation rules for the UDTFOperator: Follow up for 
> HIVE-20262
> --
>
> Key: HIVE-20480
> URL: https://issues.apache.org/jira/browse/HIVE-20480
> Project: Hive
>  Issue Type: Improvement
>  Components: Physical Optimizer
>Reporter: George Pachitariu
>Assignee: George Pachitariu
>Priority: Minor
> Attachments: HIVE-20480.patch
>
>
>  Implementing the rule for column stats: Follow-up task for 
> [HIVE-20262|http://example.com/]



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-20480) Implement column stats annotation rules for the UDTFOperator: Follow up for HIVE-20262

2018-08-28 Thread George Pachitariu (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-20480?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

George Pachitariu updated HIVE-20480:
-
Description:  Implementing the rule for column stats: Follow-up task for 
[HIVE-20262|http://example.com/]  (was:  

Implementing the rule for column stats: Follow up task for 
[HIVE-20262|http://example.com])

> Implement column stats annotation rules for the UDTFOperator: Follow up for 
> HIVE-20262
> --
>
> Key: HIVE-20480
> URL: https://issues.apache.org/jira/browse/HIVE-20480
> Project: Hive
>  Issue Type: Improvement
>  Components: Physical Optimizer
>Reporter: George Pachitariu
>Assignee: George Pachitariu
>Priority: Minor
> Attachments: HIVE-20480.patch
>
>
>  Implementing the rule for column stats: Follow-up task for 
> [HIVE-20262|http://example.com/]



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-20377) Hive Kafka Storage Handler

2018-08-28 Thread slim bouguerra (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-20377?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16595586#comment-16595586
 ] 

slim bouguerra commented on HIVE-20377:
---

We have some hard coded timestamp records fields to ensure that q files are 
same across runs.
 [^HIVE-20377.18.patch] added a small fix to the test setup to avoid deleting 
old records by Kafka broker.
{code}properties.setProperty("log.retention.hours", 
String.valueOf(Integer.MAX_VALUE));{code}



> Hive Kafka Storage Handler
> --
>
> Key: HIVE-20377
> URL: https://issues.apache.org/jira/browse/HIVE-20377
> Project: Hive
>  Issue Type: New Feature
>Affects Versions: 4.0.0
>Reporter: slim bouguerra
>Assignee: slim bouguerra
>Priority: Major
> Attachments: HIVE-20377.10.patch, HIVE-20377.11.patch, 
> HIVE-20377.12.patch, HIVE-20377.15.patch, HIVE-20377.18.patch, 
> HIVE-20377.4.patch, HIVE-20377.5.patch, HIVE-20377.6.patch, 
> HIVE-20377.8.patch, HIVE-20377.8.patch, HIVE-20377.patch
>
>
> h1. Goal
> * Read streaming data form Kafka queue as an external table.
> * Allow streaming navigation by pushing down filters on Kafka record 
> partition id, offset and timestamp. 
> * Insert streaming data form Kafka to an actual Hive internal table, using 
> CTAS statement.
> h1. Example
> h2. Create the external table
> {code} 
> CREATE EXTERNAL TABLE kafka_table (`timestamp` timestamp, page string, `user` 
> string, language string, added int, deleted int, flags string,comment string, 
> namespace string)
> STORED BY 'org.apache.hadoop.hive.kafka.KafkaStorageHandler'
> TBLPROPERTIES 
> ("kafka.topic" = "wikipedia", 
> "kafka.bootstrap.servers"="brokeraddress:9092",
> "kafka.serde.class"="org.apache.hadoop.hive.serde2.JsonSerDe");
> {code}
> h2. Kafka Metadata
> In order to keep track of Kafka records the storage handler will add 
> automatically the Kafka row metadata eg partition id, record offset and 
> record timestamp. 
> {code}
> DESCRIBE EXTENDED kafka_table
> timestamp timestamp   from deserializer   
> page  string  from deserializer   
> user  string  from deserializer   
> language  string  from deserializer   
> country   string  from deserializer   
> continent string  from deserializer   
> namespace string  from deserializer   
> newpage   boolean from deserializer   
> unpatrolled   boolean from deserializer   
> anonymous boolean from deserializer   
> robot boolean from deserializer   
> added int from deserializer   
> deleted   int from deserializer   
> delta bigint  from deserializer   
> __partition   int from deserializer   
> __offset  bigint  from deserializer   
> __timestamp   bigint  from deserializer   
> {code}
> h2. Filter push down.
> Newer Kafka consumers 0.11.0 and higher allow seeking on the stream based on 
> a given offset. The proposed storage handler will be able to leverage such 
> API by pushing down filters over metadata columns, namely __partition (int), 
> __offset(long) and __timestamp(long)
> For instance Query like
> {code} 
> select `__offset` from kafka_table where (`__offset` < 10 and `__offset`>3 
> and `__partition` = 0) or (`__partition` = 0 and `__offset` < 105 and 
> `__offset` > 99) or (`__offset` = 109);
> {code}
> Will result on a scan of partition 0 only then read only records between 
> offset 4 and 109. 
> h2. With timestamp seeks 
> The seeking based on the internal timestamps allows the handler to run on 
> recently arrived data, by doing
> {code}
> select count(*) from kafka_table where `__timestamp` >  1000 * 
> to_unix_timestamp(CURRENT_TIMESTAMP - interval '20' hours) ;
> {code}
> This allows for implicit relationships between event timestamps and kafka 
> timestamps to be expressed in queries (i.e event_timestamp is always < than 
> kafka __timestamp and kafka __timestamp is never > 15 minutes from event etc).
> h2. More examples with Avro 
> {code}
> CREATE EXTERNAL TABLE wiki_kafka_avro_table
> STORED BY 'org.apache.hadoop.hive.kafka.KafkaStorageHandler'
> TBLPROPERTIES
> ("kafka.topic" = "wiki_kafka_avro_table",
> "kafka.bootstrap.servers"="localhost:9092",
> "kafka.serde.class"="org.apache.hadoop.hive.serde2.avro.AvroSerDe",
> 'avro.schema.literal'='{
>   "type" : "record",
>   "name" : "Wikipedia",
>   "namespace" : "org.apache.hive.kafka",
>   "version": "1",
>   "fields" : [ {
> "name" : 

[jira] [Commented] (HIVE-17084) Turn on hive.stats.fetch.column.stats configuration flag

2018-08-28 Thread Hive QA (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-17084?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16595583#comment-16595583
 ] 

Hive QA commented on HIVE-17084:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12937453/HIVE-17084.08.patch

{color:red}ERROR:{color} -1 due to build exiting with an error

Test results: 
https://builds.apache.org/job/PreCommit-HIVE-Build/13512/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/13512/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-13512/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Tests exited with: NonZeroExitCodeException
Command 'bash /data/hiveptest/working/scratch/source-prep.sh' failed with exit 
status 1 and output '+ date '+%Y-%m-%d %T.%3N'
2018-08-28 20:49:18.722
+ [[ -n /usr/lib/jvm/java-8-openjdk-amd64 ]]
+ export JAVA_HOME=/usr/lib/jvm/java-8-openjdk-amd64
+ JAVA_HOME=/usr/lib/jvm/java-8-openjdk-amd64
+ export 
PATH=/usr/lib/jvm/java-8-openjdk-amd64/bin/:/usr/local/bin:/usr/bin:/bin:/usr/local/games:/usr/games
+ 
PATH=/usr/lib/jvm/java-8-openjdk-amd64/bin/:/usr/local/bin:/usr/bin:/bin:/usr/local/games:/usr/games
+ export 'ANT_OPTS=-Xmx1g -XX:MaxPermSize=256m '
+ ANT_OPTS='-Xmx1g -XX:MaxPermSize=256m '
+ export 'MAVEN_OPTS=-Xmx1g '
+ MAVEN_OPTS='-Xmx1g '
+ cd /data/hiveptest/working/
+ tee /data/hiveptest/logs/PreCommit-HIVE-Build-13512/source-prep.txt
+ [[ false == \t\r\u\e ]]
+ mkdir -p maven ivy
+ [[ git = \s\v\n ]]
+ [[ git = \g\i\t ]]
+ [[ -z master ]]
+ [[ -d apache-github-source-source ]]
+ [[ ! -d apache-github-source-source/.git ]]
+ [[ ! -d apache-github-source-source ]]
+ date '+%Y-%m-%d %T.%3N'
2018-08-28 20:49:18.726
+ cd apache-github-source-source
+ git fetch origin
>From https://github.com/apache/hive
   7cf1d00..d2b81a9  master -> origin/master
+ git reset --hard HEAD
HEAD is now at 7cf1d00 HIVE-20396: Test HS2 open_connection metrics (Laszlo 
Pinter via Peter Vary)
+ git clean -f -d
Removing standalone-metastore/metastore-server/src/gen/
+ git checkout master
Already on 'master'
Your branch is behind 'origin/master' by 1 commit, and can be fast-forwarded.
  (use "git pull" to update your local branch)
+ git reset --hard origin/master
HEAD is now at d2b81a9 HIVE-19674: Group by Decimal Constants push down to 
Druid tables(Vineet Garg, reviewed by Ashutosh Chauhan)
+ git merge --ff-only origin/master
Already up-to-date.
+ date '+%Y-%m-%d %T.%3N'
2018-08-28 20:49:20.487
+ rm -rf ../yetus_PreCommit-HIVE-Build-13512
+ mkdir ../yetus_PreCommit-HIVE-Build-13512
+ git gc
+ cp -R . ../yetus_PreCommit-HIVE-Build-13512
+ mkdir /data/hiveptest/logs/PreCommit-HIVE-Build-13512/yetus
+ patchCommandPath=/data/hiveptest/working/scratch/smart-apply-patch.sh
+ patchFilePath=/data/hiveptest/working/scratch/build.patch
+ [[ -f /data/hiveptest/working/scratch/build.patch ]]
+ chmod +x /data/hiveptest/working/scratch/smart-apply-patch.sh
+ /data/hiveptest/working/scratch/smart-apply-patch.sh 
/data/hiveptest/working/scratch/build.patch
error: patch failed: 
ql/src/test/results/clientpositive/list_bucket_query_oneskew_2.q.out:533
Falling back to three-way merge...
Applied patch to 
'ql/src/test/results/clientpositive/list_bucket_query_oneskew_2.q.out' with 
conflicts.
error: patch failed: 
ql/src/test/results/clientpositive/perf/spark/query44.q.out:69
Falling back to three-way merge...
Applied patch to 'ql/src/test/results/clientpositive/perf/spark/query44.q.out' 
with conflicts.
error: patch failed: 
ql/src/test/results/clientpositive/perf/tez/query44.q.out:86
Falling back to three-way merge...
Applied patch to 'ql/src/test/results/clientpositive/perf/tez/query44.q.out' 
with conflicts.
error: patch failed: ql/src/test/results/clientpositive/semijoin4.q.out:235
Falling back to three-way merge...
Applied patch to 'ql/src/test/results/clientpositive/semijoin4.q.out' with 
conflicts.
error: patch failed: 
ql/src/test/results/clientpositive/spark/spark_dynamic_partition_pruning.q.out:2514
Falling back to three-way merge...
Applied patch to 
'ql/src/test/results/clientpositive/spark/spark_dynamic_partition_pruning.q.out'
 with conflicts.
error: patch failed: 
ql/src/test/results/clientpositive/spark/spark_vectorized_dynamic_partition_pruning.q.out:4811
Falling back to three-way merge...
Applied patch to 
'ql/src/test/results/clientpositive/spark/spark_vectorized_dynamic_partition_pruning.q.out'
 cleanly.
error: patch failed: 
ql/src/test/results/clientpositive/spark/subquery_in.q.out:2636
Falling back to three-way merge...
Applied patch to 'ql/src/test/results/clientpositive/spark/subquery_in.q.out' 
with conflicts.
error: patch failed: 
ql/src/test/results/clientpositive/spark/subquery_notin.q.out:3600
Falling back to three-way merge...
Applied patch to 

[jira] [Updated] (HIVE-20480) Implement column stats annotation rules for the UDTFOperator: Follow up for HIVE-20262

2018-08-28 Thread George Pachitariu (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-20480?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

George Pachitariu updated HIVE-20480:
-
Attachment: HIVE-20480.patch
Status: Patch Available  (was: Open)

> Implement column stats annotation rules for the UDTFOperator: Follow up for 
> HIVE-20262
> --
>
> Key: HIVE-20480
> URL: https://issues.apache.org/jira/browse/HIVE-20480
> Project: Hive
>  Issue Type: Improvement
>  Components: Physical Optimizer
>Reporter: George Pachitariu
>Assignee: George Pachitariu
>Priority: Minor
> Attachments: HIVE-20480.patch
>
>
>  
> Implementing the rule for column stats: Follow up task for 
> [HIVE-20262|http://example.com]



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-20377) Hive Kafka Storage Handler

2018-08-28 Thread slim bouguerra (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-20377?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

slim bouguerra updated HIVE-20377:
--
Attachment: HIVE-20377.18.patch

> Hive Kafka Storage Handler
> --
>
> Key: HIVE-20377
> URL: https://issues.apache.org/jira/browse/HIVE-20377
> Project: Hive
>  Issue Type: New Feature
>Affects Versions: 4.0.0
>Reporter: slim bouguerra
>Assignee: slim bouguerra
>Priority: Major
> Attachments: HIVE-20377.10.patch, HIVE-20377.11.patch, 
> HIVE-20377.12.patch, HIVE-20377.15.patch, HIVE-20377.18.patch, 
> HIVE-20377.4.patch, HIVE-20377.5.patch, HIVE-20377.6.patch, 
> HIVE-20377.8.patch, HIVE-20377.8.patch, HIVE-20377.patch
>
>
> h1. Goal
> * Read streaming data form Kafka queue as an external table.
> * Allow streaming navigation by pushing down filters on Kafka record 
> partition id, offset and timestamp. 
> * Insert streaming data form Kafka to an actual Hive internal table, using 
> CTAS statement.
> h1. Example
> h2. Create the external table
> {code} 
> CREATE EXTERNAL TABLE kafka_table (`timestamp` timestamp, page string, `user` 
> string, language string, added int, deleted int, flags string,comment string, 
> namespace string)
> STORED BY 'org.apache.hadoop.hive.kafka.KafkaStorageHandler'
> TBLPROPERTIES 
> ("kafka.topic" = "wikipedia", 
> "kafka.bootstrap.servers"="brokeraddress:9092",
> "kafka.serde.class"="org.apache.hadoop.hive.serde2.JsonSerDe");
> {code}
> h2. Kafka Metadata
> In order to keep track of Kafka records the storage handler will add 
> automatically the Kafka row metadata eg partition id, record offset and 
> record timestamp. 
> {code}
> DESCRIBE EXTENDED kafka_table
> timestamp timestamp   from deserializer   
> page  string  from deserializer   
> user  string  from deserializer   
> language  string  from deserializer   
> country   string  from deserializer   
> continent string  from deserializer   
> namespace string  from deserializer   
> newpage   boolean from deserializer   
> unpatrolled   boolean from deserializer   
> anonymous boolean from deserializer   
> robot boolean from deserializer   
> added int from deserializer   
> deleted   int from deserializer   
> delta bigint  from deserializer   
> __partition   int from deserializer   
> __offset  bigint  from deserializer   
> __timestamp   bigint  from deserializer   
> {code}
> h2. Filter push down.
> Newer Kafka consumers 0.11.0 and higher allow seeking on the stream based on 
> a given offset. The proposed storage handler will be able to leverage such 
> API by pushing down filters over metadata columns, namely __partition (int), 
> __offset(long) and __timestamp(long)
> For instance Query like
> {code} 
> select `__offset` from kafka_table where (`__offset` < 10 and `__offset`>3 
> and `__partition` = 0) or (`__partition` = 0 and `__offset` < 105 and 
> `__offset` > 99) or (`__offset` = 109);
> {code}
> Will result on a scan of partition 0 only then read only records between 
> offset 4 and 109. 
> h2. With timestamp seeks 
> The seeking based on the internal timestamps allows the handler to run on 
> recently arrived data, by doing
> {code}
> select count(*) from kafka_table where `__timestamp` >  1000 * 
> to_unix_timestamp(CURRENT_TIMESTAMP - interval '20' hours) ;
> {code}
> This allows for implicit relationships between event timestamps and kafka 
> timestamps to be expressed in queries (i.e event_timestamp is always < than 
> kafka __timestamp and kafka __timestamp is never > 15 minutes from event etc).
> h2. More examples with Avro 
> {code}
> CREATE EXTERNAL TABLE wiki_kafka_avro_table
> STORED BY 'org.apache.hadoop.hive.kafka.KafkaStorageHandler'
> TBLPROPERTIES
> ("kafka.topic" = "wiki_kafka_avro_table",
> "kafka.bootstrap.servers"="localhost:9092",
> "kafka.serde.class"="org.apache.hadoop.hive.serde2.avro.AvroSerDe",
> 'avro.schema.literal'='{
>   "type" : "record",
>   "name" : "Wikipedia",
>   "namespace" : "org.apache.hive.kafka",
>   "version": "1",
>   "fields" : [ {
> "name" : "isrobot",
> "type" : "boolean"
>   }, {
> "name" : "channel",
> "type" : "string"
>   }, {
> "name" : "timestamp",
> "type" : "string"
>   }, {
> "name" : "flags",
> "type" : "string"
>   }, {
> "name" : "isunpatrolled",
> "type" : "boolean"
>   }, {
> "name" : 

[jira] [Commented] (HIVE-13457) Create HS2 REST API endpoints for monitoring information

2018-08-28 Thread Hive QA (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-13457?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16595576#comment-16595576
 ] 

Hive QA commented on HIVE-13457:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12937447/HIVE-13457.5.patch

{color:green}SUCCESS:{color} +1 due to 1 test(s) being added or modified.

{color:green}SUCCESS:{color} +1 due to 14903 tests passed

Test results: 
https://builds.apache.org/job/PreCommit-HIVE-Build/13511/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/13511/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-13511/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.YetusPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12937447 - PreCommit-HIVE-Build

> Create HS2 REST API endpoints for monitoring information
> 
>
> Key: HIVE-13457
> URL: https://issues.apache.org/jira/browse/HIVE-13457
> Project: Hive
>  Issue Type: Improvement
>Reporter: Szehon Ho
>Assignee: Pawel Szostek
>Priority: Major
> Attachments: HIVE-13457.3.patch, HIVE-13457.4.patch, 
> HIVE-13457.5.patch, HIVE-13457.patch, HIVE-13457.patch
>
>
> Similar to what is exposed in HS2 webui in HIVE-12338, it would be nice if 
> other UI's like admin tools or Hue can access and display this information as 
> well.  Hence, we will create some REST endpoints to expose this information.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Assigned] (HIVE-20480) Implement column stats annotation rules for the UDTFOperator: Follow up for HIVE-20262

2018-08-28 Thread George Pachitariu (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-20480?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

George Pachitariu reassigned HIVE-20480:



> Implement column stats annotation rules for the UDTFOperator: Follow up for 
> HIVE-20262
> --
>
> Key: HIVE-20480
> URL: https://issues.apache.org/jira/browse/HIVE-20480
> Project: Hive
>  Issue Type: Improvement
>  Components: Physical Optimizer
>Reporter: George Pachitariu
>Assignee: George Pachitariu
>Priority: Minor
>
>  
> Implementing the rule for column stats: Follow up task for 
> [HIVE-20262|http://example.com]



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Assigned] (HIVE-19318) Improve Hive logging

2018-08-28 Thread Aihua Xu (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-19318?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Aihua Xu reassigned HIVE-19318:
---

Assignee: Alice Fan  (was: Aihua Xu)

> Improve Hive logging
> 
>
> Key: HIVE-19318
> URL: https://issues.apache.org/jira/browse/HIVE-19318
> Project: Hive
>  Issue Type: Improvement
>  Components: Logging
>Affects Versions: 3.0.0
>Reporter: Aihua Xu
>Assignee: Alice Fan
>Priority: Major
>
> Use this jira to track some potential improvements on hive logging. What I 
> have noticed that some log entries may have incorrect log level, or may not 
> show in the correct places, e.g., some logs are printing to the STDERR/STDOUT 
> rather than the HS2 log file. 



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-18725) Improve error handling for subqueries if there is wrong column reference

2018-08-28 Thread Vineet Garg (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-18725?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16595528#comment-16595528
 ] 

Vineet Garg commented on HIVE-18725:


[~ikryvenko] Sorry I couldn't take a look the patch before. I will review it 
now and will shortly update.

> Improve error handling for subqueries if there is wrong column reference
> 
>
> Key: HIVE-18725
> URL: https://issues.apache.org/jira/browse/HIVE-18725
> Project: Hive
>  Issue Type: Improvement
>  Components: Query Planning
>Reporter: Vineet Garg
>Assignee: Igor Kryvenko
>Priority: Major
> Attachments: HIVE-18725.01.patch, HIVE-18725.02.patch, 
> HIVE-18725.03.patch, HIVE-18725.04.patch, HIVE-18725.05.patch, 
> HIVE-18725.06.patch, HIVE-18725.07.patch, HIVE-18725.08.patch, 
> HIVE-18725.09.patch, HIVE-18725.10.patch
>
>
> If there is a column reference within subquery which doesn't exist Hive 
> throws misleading error message.
> e.g. 
> {code:sql}
> select * from table1 where table1.col1 IN (select col2 from table2 where 
> table2.col1=table1.non_existing_column) and table1.col1 IN (select 4);
> {code}
> The above query, assuming table1 doesn't have non_existing_column, will throw 
> following misleading error:
> {noformat}
> FAILED: SemanticException Line 0:-1 Unsupported SubQuery Expression 'col1': 
> Only 1 SubQuery expression is supported.
> {noformat}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-18725) Improve error handling for subqueries if there is wrong column reference

2018-08-28 Thread Igor Kryvenko (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-18725?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16595523#comment-16595523
 ] 

Igor Kryvenko commented on HIVE-18725:
--

[~vgarg] Hi Vineet. Any updates on this issue?

> Improve error handling for subqueries if there is wrong column reference
> 
>
> Key: HIVE-18725
> URL: https://issues.apache.org/jira/browse/HIVE-18725
> Project: Hive
>  Issue Type: Improvement
>  Components: Query Planning
>Reporter: Vineet Garg
>Assignee: Igor Kryvenko
>Priority: Major
> Attachments: HIVE-18725.01.patch, HIVE-18725.02.patch, 
> HIVE-18725.03.patch, HIVE-18725.04.patch, HIVE-18725.05.patch, 
> HIVE-18725.06.patch, HIVE-18725.07.patch, HIVE-18725.08.patch, 
> HIVE-18725.09.patch, HIVE-18725.10.patch
>
>
> If there is a column reference within subquery which doesn't exist Hive 
> throws misleading error message.
> e.g. 
> {code:sql}
> select * from table1 where table1.col1 IN (select col2 from table2 where 
> table2.col1=table1.non_existing_column) and table1.col1 IN (select 4);
> {code}
> The above query, assuming table1 doesn't have non_existing_column, will throw 
> following misleading error:
> {noformat}
> FAILED: SemanticException Line 0:-1 Unsupported SubQuery Expression 'col1': 
> Only 1 SubQuery expression is supported.
> {noformat}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-13457) Create HS2 REST API endpoints for monitoring information

2018-08-28 Thread Hive QA (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-13457?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16595519#comment-16595519
 ] 

Hive QA commented on HIVE-13457:


| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
|| || || || {color:brown} Prechecks {color} ||
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
|| || || || {color:brown} master Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  8m 
58s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  0m 
26s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
14s{color} | {color:green} master passed {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  0m 
41s{color} | {color:blue} service in master has 48 extant Findbugs warnings. 
{color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
19s{color} | {color:green} master passed {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  0m 
24s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  0m 
27s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  0m 
27s{color} | {color:green} the patch passed {color} |
| {color:red}-1{color} | {color:red} checkstyle {color} | {color:red}  0m 
15s{color} | {color:red} service: The patch generated 130 new + 36 unchanged - 
0 fixed = 166 total (was 36) {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  0m 
48s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
18s{color} | {color:green} the patch passed {color} |
|| || || || {color:brown} Other Tests {color} ||
| {color:red}-1{color} | {color:red} asflicense {color} | {color:red}  0m 
14s{color} | {color:red} The patch generated 1 ASF License warnings. {color} |
| {color:black}{color} | {color:black} {color} | {color:black} 13m 25s{color} | 
{color:black} {color} |
\\
\\
|| Subsystem || Report/Notes ||
| Optional Tests |  asflicense  javac  javadoc  findbugs  checkstyle  compile  |
| uname | Linux hiveptest-server-upstream 3.16.0-4-amd64 #1 SMP Debian 
3.16.36-1+deb8u1 (2016-09-03) x86_64 GNU/Linux |
| Build tool | maven |
| Personality | 
/data/hiveptest/working/yetus_PreCommit-HIVE-Build-13511/dev-support/hive-personality.sh
 |
| git revision | master / 7cf1d00 |
| Default Java | 1.8.0_111 |
| findbugs | v3.0.0 |
| checkstyle | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-13511/yetus/diff-checkstyle-service.txt
 |
| asflicense | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-13511/yetus/patch-asflicense-problems.txt
 |
| modules | C: service U: service |
| Console output | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-13511/yetus.txt |
| Powered by | Apache Yetushttp://yetus.apache.org |


This message was automatically generated.



> Create HS2 REST API endpoints for monitoring information
> 
>
> Key: HIVE-13457
> URL: https://issues.apache.org/jira/browse/HIVE-13457
> Project: Hive
>  Issue Type: Improvement
>Reporter: Szehon Ho
>Assignee: Pawel Szostek
>Priority: Major
> Attachments: HIVE-13457.3.patch, HIVE-13457.4.patch, 
> HIVE-13457.5.patch, HIVE-13457.patch, HIVE-13457.patch
>
>
> Similar to what is exposed in HS2 webui in HIVE-12338, it would be nice if 
> other UI's like admin tools or Hue can access and display this information as 
> well.  Hence, we will create some REST endpoints to expose this information.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-19674) Group by Decimal Constants push down to Druid tables.

2018-08-28 Thread Vineet Garg (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-19674?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Vineet Garg updated HIVE-19674:
---
   Resolution: Fixed
Fix Version/s: 4.0.0
   Status: Resolved  (was: Patch Available)

Pushed to master.

> Group by Decimal Constants push down to Druid tables.
> -
>
> Key: HIVE-19674
> URL: https://issues.apache.org/jira/browse/HIVE-19674
> Project: Hive
>  Issue Type: Improvement
>  Components: Druid integration
>Reporter: slim bouguerra
>Assignee: Vineet Garg
>Priority: Major
> Fix For: 4.0.0
>
> Attachments: HIVE-19674.1.patch, HIVE-19674.2.patch, 
> HIVE-19674.3.patch, HIVE-19674.4.patch, HIVE-19674.5.patch, HIVE-19674.6.patch
>
>
> Queries like following gets generated by Tableau.
> {code}
> SELECT SUM(`ssb_druid_100`.`lo_revenue`) AS `sum_lo_revenue_ok`
>  FROM `druid_ssb`.`ssb_druid_100` `ssb_druid_100`
> GROUP BY 1.1001;
> {code}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-20363) Use integer constants for frequently used serde classes.

2018-08-28 Thread Anurag Mantripragada (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-20363?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Anurag Mantripragada updated HIVE-20363:

Description: 
Serde libraries are stored as fully qualified class names which are long 
strings, we can get improvements in I/O and storage if we store integer 
constants for frequently used serde classes in the backend DB.

For example:
{code:java}
org.apache.hadoop.hive.serde2.avro.AvroSerDe"---> 1
parquet.hive.serde.ParquetHiveSerDe" ---> 2
...{code}
 Review the patch at: 

https://reviews.apache.org/r/68546/

  was:
Serde libraries are stored as fully qualified class names which are long 
strings, we can get improvements in I/O and storage if we store integer 
constants for frequently used serde classes in the backend DB.

For example:
{code:java}
org.apache.hadoop.hive.serde2.avro.AvroSerDe"---> 1
parquet.hive.serde.ParquetHiveSerDe" ---> 2
...{code}
 Review the patch at: 

https://issues.apache.org/jira/browse/HIVE-20363


> Use integer constants for frequently used serde classes.
> 
>
> Key: HIVE-20363
> URL: https://issues.apache.org/jira/browse/HIVE-20363
> Project: Hive
>  Issue Type: Improvement
>  Components: Metastore, Standalone Metastore
>Reporter: Anurag Mantripragada
>Assignee: Anurag Mantripragada
>Priority: Major
>
> Serde libraries are stored as fully qualified class names which are long 
> strings, we can get improvements in I/O and storage if we store integer 
> constants for frequently used serde classes in the backend DB.
> For example:
> {code:java}
> org.apache.hadoop.hive.serde2.avro.AvroSerDe"---> 1
> parquet.hive.serde.ParquetHiveSerDe" ---> 2
> ...{code}
>  Review the patch at: 
> https://reviews.apache.org/r/68546/



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-20363) Use integer constants for frequently used serde classes.

2018-08-28 Thread Anurag Mantripragada (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-20363?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Anurag Mantripragada updated HIVE-20363:

Description: 
Serde libraries are stored as fully qualified class names which are long 
strings, we can get improvements in I/O and storage if we store integer 
constants for frequently used serde classes in the backend DB.

For example:
{code:java}
org.apache.hadoop.hive.serde2.avro.AvroSerDe"---> 1
parquet.hive.serde.ParquetHiveSerDe" ---> 2
...{code}
 Review the patch at: 

https://issues.apache.org/jira/browse/HIVE-20363

  was:
Serde libraries are stored as fully qualified class names which are long 
strings, we can get improvements in I/O and storage if we store integer 
constants for frequently used serde classes in the backend DB.

For example:
{code:java}
org.apache.hadoop.hive.serde2.avro.AvroSerDe"---> 1
parquet.hive.serde.ParquetHiveSerDe" ---> 2
...{code}
 


> Use integer constants for frequently used serde classes.
> 
>
> Key: HIVE-20363
> URL: https://issues.apache.org/jira/browse/HIVE-20363
> Project: Hive
>  Issue Type: Improvement
>  Components: Metastore, Standalone Metastore
>Reporter: Anurag Mantripragada
>Assignee: Anurag Mantripragada
>Priority: Major
>
> Serde libraries are stored as fully qualified class names which are long 
> strings, we can get improvements in I/O and storage if we store integer 
> constants for frequently used serde classes in the backend DB.
> For example:
> {code:java}
> org.apache.hadoop.hive.serde2.avro.AvroSerDe"---> 1
> parquet.hive.serde.ParquetHiveSerDe" ---> 2
> ...{code}
>  Review the patch at: 
> https://issues.apache.org/jira/browse/HIVE-20363



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Resolved] (HIVE-19585) Add UNKNOWN to PrincipalType

2018-08-28 Thread Arjun Mishra (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-19585?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arjun Mishra resolved HIVE-19585.
-
Resolution: Not A Problem

> Add UNKNOWN to PrincipalType
> 
>
> Key: HIVE-19585
> URL: https://issues.apache.org/jira/browse/HIVE-19585
> Project: Hive
>  Issue Type: Bug
>  Components: Hive
>Affects Versions: 2.3.2
>Reporter: Arjun Mishra
>Assignee: Arjun Mishra
>Priority: Major
> Fix For: 2.3.2
>
> Attachments: HIVE-198585.01.patch
>
>
> We need to include type UNKNOWN to PrincipalType to match with 
> HivePrincipal.HivePrincipalType.UKNOWN



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-20205) Upgrade HBase dependencies off alpha4 release

2018-08-28 Thread Hive QA (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-20205?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16595438#comment-16595438
 ] 

Hive QA commented on HIVE-20205:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12937440/HIVE-20205.2.patch

{color:red}ERROR:{color} -1 due to no test(s) being added or modified.

{color:red}ERROR:{color} -1 due to 8 failed/errored test(s), 14868 tests 
executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestHBaseCliDriver.org.apache.hadoop.hive.cli.TestHBaseCliDriver
 (batchId=101)
org.apache.hadoop.hive.cli.TestHBaseCliDriver.org.apache.hadoop.hive.cli.TestHBaseCliDriver
 (batchId=102)
org.apache.hadoop.hive.cli.TestHBaseCliDriver.org.apache.hadoop.hive.cli.TestHBaseCliDriver
 (batchId=103)
org.apache.hadoop.hive.cli.TestHBaseCliDriver.org.apache.hadoop.hive.cli.TestHBaseCliDriver
 (batchId=104)
org.apache.hadoop.hive.cli.TestHBaseCliDriver.org.apache.hadoop.hive.cli.TestHBaseCliDriver
 (batchId=105)
org.apache.hadoop.hive.cli.TestHBaseCliDriver.org.apache.hadoop.hive.cli.TestHBaseCliDriver
 (batchId=106)
org.apache.hadoop.hive.cli.TestHBaseNegativeCliDriver.org.apache.hadoop.hive.cli.TestHBaseNegativeCliDriver
 (batchId=262)
org.apache.hive.hcatalog.hbase.TestPigHBaseStorageHandler.org.apache.hive.hcatalog.hbase.TestPigHBaseStorageHandler
 (batchId=258)
{noformat}

Test results: 
https://builds.apache.org/job/PreCommit-HIVE-Build/13510/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/13510/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-13510/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.YetusPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 8 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12937440 - PreCommit-HIVE-Build

> Upgrade HBase dependencies off alpha4 release
> -
>
> Key: HIVE-20205
> URL: https://issues.apache.org/jira/browse/HIVE-20205
> Project: Hive
>  Issue Type: Bug
>  Components: Hive
>Affects Versions: 3.0.0
>Reporter: Naveen Gangam
>Assignee: Naveen Gangam
>Priority: Minor
> Attachments: HIVE-20205.1.patch, HIVE-20205.1.patch, 
> HIVE-20205.2.patch, HIVE-20205.2.patch, HIVE-20205.patch, HIVE-20205.patch
>
>
> Appears Hive has dependencies on hbase 2.0.0-alpha4 releases. HBase 2.0.0 and 
> 2.0.1 have been released. HBase team recommends 2.0.1 and says there shouldnt 
> be any API surprises. (but we never know)



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-20205) Upgrade HBase dependencies off alpha4 release

2018-08-28 Thread Hive QA (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-20205?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16595392#comment-16595392
 ] 

Hive QA commented on HIVE-20205:


| (/) *{color:green}+1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
|| || || || {color:brown} Prechecks {color} ||
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
|| || || || {color:brown} master Compile Tests {color} ||
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  0m 
40s{color} | {color:blue} Maven dependency ordering for branch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  8m 
 5s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  7m  
0s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  7m 
20s{color} | {color:green} master passed {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  0m  
7s{color} | {color:blue} Maven dependency ordering for patch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  7m 
21s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  6m 
46s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  6m 
46s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} xml {color} | {color:green}  0m  
2s{color} | {color:green} The patch has no ill-formed XML file. {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  7m 
20s{color} | {color:green} the patch passed {color} |
|| || || || {color:brown} Other Tests {color} ||
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
13s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black} 45m 19s{color} | 
{color:black} {color} |
\\
\\
|| Subsystem || Report/Notes ||
| Optional Tests |  asflicense  javac  javadoc  xml  compile  |
| uname | Linux hiveptest-server-upstream 3.16.0-4-amd64 #1 SMP Debian 
3.16.36-1+deb8u1 (2016-09-03) x86_64 GNU/Linux |
| Build tool | maven |
| Personality | 
/data/hiveptest/working/yetus_PreCommit-HIVE-Build-13510/dev-support/hive-personality.sh
 |
| git revision | master / 7cf1d00 |
| Default Java | 1.8.0_111 |
| modules | C: . itests/util U: . |
| Console output | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-13510/yetus.txt |
| Powered by | Apache Yetushttp://yetus.apache.org |


This message was automatically generated.



> Upgrade HBase dependencies off alpha4 release
> -
>
> Key: HIVE-20205
> URL: https://issues.apache.org/jira/browse/HIVE-20205
> Project: Hive
>  Issue Type: Bug
>  Components: Hive
>Affects Versions: 3.0.0
>Reporter: Naveen Gangam
>Assignee: Naveen Gangam
>Priority: Minor
> Attachments: HIVE-20205.1.patch, HIVE-20205.1.patch, 
> HIVE-20205.2.patch, HIVE-20205.2.patch, HIVE-20205.patch, HIVE-20205.patch
>
>
> Appears Hive has dependencies on hbase 2.0.0-alpha4 releases. HBase 2.0.0 and 
> 2.0.1 have been released. HBase team recommends 2.0.1 and says there shouldnt 
> be any API surprises. (but we never know)



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Resolved] (HIVE-20479) Update content/people.mdtext in cms

2018-08-28 Thread Andrew Sherman (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-20479?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Andrew Sherman resolved HIVE-20479.
---
Resolution: Fixed

Done already

> Update content/people.mdtext in cms 
> 
>
> Key: HIVE-20479
> URL: https://issues.apache.org/jira/browse/HIVE-20479
> Project: Hive
>  Issue Type: Task
>Reporter: Andrew Sherman
>Assignee: Andrew Sherman
>Priority: Major
>
> I added myself to the committers list. 
>  
> {code:java}
> asherman 
> Andrew Sherman 
>  href="http://cloudera.com/;>Cloudera 
>  
> 
> {code}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-20479) Update content/people.mdtext in cms

2018-08-28 Thread Sahil Takiar (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-20479?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16595364#comment-16595364
 ] 

Sahil Takiar commented on HIVE-20479:
-

+1

> Update content/people.mdtext in cms 
> 
>
> Key: HIVE-20479
> URL: https://issues.apache.org/jira/browse/HIVE-20479
> Project: Hive
>  Issue Type: Task
>Reporter: Andrew Sherman
>Assignee: Andrew Sherman
>Priority: Major
>
> I added myself to the committers list. 
>  
> {code:java}
> asherman 
> Andrew Sherman 
>  href="http://cloudera.com/;>Cloudera 
>  
> 
> {code}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Assigned] (HIVE-20479) Update content/people.mdtext in cms

2018-08-28 Thread Andrew Sherman (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-20479?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Andrew Sherman reassigned HIVE-20479:
-


> Update content/people.mdtext in cms 
> 
>
> Key: HIVE-20479
> URL: https://issues.apache.org/jira/browse/HIVE-20479
> Project: Hive
>  Issue Type: Task
>Reporter: Andrew Sherman
>Assignee: Andrew Sherman
>Priority: Major
>
> I added myself to the committers list. 
>  
> {code:java}
> asherman 
> Andrew Sherman 
>  href="http://cloudera.com/;>Cloudera 
>  
> 
> {code}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-20466) Improve org.apache.hadoop.hive.ql.exec.FunctionTask Experience

2018-08-28 Thread Hive QA (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-20466?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16595330#comment-16595330
 ] 

Hive QA commented on HIVE-20466:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12937444/HIVE-20466.4.patch

{color:red}ERROR:{color} -1 due to no test(s) being added or modified.

{color:green}SUCCESS:{color} +1 due to 14899 tests passed

Test results: 
https://builds.apache.org/job/PreCommit-HIVE-Build/13509/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/13509/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-13509/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.YetusPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12937444 - PreCommit-HIVE-Build

> Improve org.apache.hadoop.hive.ql.exec.FunctionTask Experience
> --
>
> Key: HIVE-20466
> URL: https://issues.apache.org/jira/browse/HIVE-20466
> Project: Hive
>  Issue Type: Improvement
>  Components: Hive
>Reporter: denys kuzmenko
>Assignee: denys kuzmenko
>Priority: Major
> Attachments: HIVE-20466.1.patch, HIVE-20466.2.patch, 
> HIVE-20466.3.patch, HIVE-20466.4.patch
>
>




--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-20472) mvn test failing for metastore-tool module

2018-08-28 Thread Alexander Kolbasov (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-20472?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alexander Kolbasov updated HIVE-20472:
--
Attachment: HIVE-20472.02.patch

> mvn test failing for metastore-tool module
> --
>
> Key: HIVE-20472
> URL: https://issues.apache.org/jira/browse/HIVE-20472
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 4.0.0
>Reporter: Vihang Karajgaonkar
>Assignee: Alexander Kolbasov
>Priority: Trivial
> Fix For: 4.0.0
>
> Attachments: HIVE-20472.01.patch, HIVE-20472.02.patch
>
>
> Fails because there are no applicable tests.
>  
> {code}
> [ERROR] Failed to execute goal 
> org.apache.maven.plugins:maven-surefire-plugin:2.20.1:test (default-test) on 
> project hive-metastore-benchmarks: No tests were executed! (Set 
> -DfailIfNoTests=false to ignore this error.) -> [Help 1]
> [ERROR]
> [ERROR] To see the full stack trace of the errors, re-run Maven with the -e 
> switch.
> [ERROR] Re-run Maven using the -X switch to enable full debug logging.
> [ERROR]
> [ERROR] For more information about the errors and possible solutions, please 
> read the following articles:
> [ERROR] [Help 1] 
> http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException
> [ERROR]
> [ERROR] After correcting the problems, you can resume the build with the 
> command
> [ERROR] mvn  -rf :hive-metastore-benchmarks
> {code}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Assigned] (HIVE-20472) mvn test failing for metastore-tool module

2018-08-28 Thread Alexander Kolbasov (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-20472?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alexander Kolbasov reassigned HIVE-20472:
-

Assignee: Alexander Kolbasov  (was: Vihang Karajgaonkar)

> mvn test failing for metastore-tool module
> --
>
> Key: HIVE-20472
> URL: https://issues.apache.org/jira/browse/HIVE-20472
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 4.0.0
>Reporter: Vihang Karajgaonkar
>Assignee: Alexander Kolbasov
>Priority: Trivial
> Fix For: 4.0.0
>
> Attachments: HIVE-20472.01.patch
>
>
> Fails because there are no applicable tests.
>  
> {code}
> [ERROR] Failed to execute goal 
> org.apache.maven.plugins:maven-surefire-plugin:2.20.1:test (default-test) on 
> project hive-metastore-benchmarks: No tests were executed! (Set 
> -DfailIfNoTests=false to ignore this error.) -> [Help 1]
> [ERROR]
> [ERROR] To see the full stack trace of the errors, re-run Maven with the -e 
> switch.
> [ERROR] Re-run Maven using the -X switch to enable full debug logging.
> [ERROR]
> [ERROR] For more information about the errors and possible solutions, please 
> read the following articles:
> [ERROR] [Help 1] 
> http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException
> [ERROR]
> [ERROR] After correcting the problems, you can resume the build with the 
> command
> [ERROR] mvn  -rf :hive-metastore-benchmarks
> {code}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-20466) Improve org.apache.hadoop.hive.ql.exec.FunctionTask Experience

2018-08-28 Thread Hive QA (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-20466?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16595279#comment-16595279
 ] 

Hive QA commented on HIVE-20466:


| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
|| || || || {color:brown} Prechecks {color} ||
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
|| || || || {color:brown} master Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  7m 
48s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  1m 
11s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
42s{color} | {color:green} master passed {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  4m  
3s{color} | {color:blue} ql in master has 2310 extant Findbugs warnings. 
{color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  1m  
0s{color} | {color:green} master passed {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  1m 
25s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  1m  
5s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  1m  
5s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
40s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:red}-1{color} | {color:red} findbugs {color} | {color:red}  4m 
15s{color} | {color:red} ql generated 1 new + 2310 unchanged - 0 fixed = 2311 
total (was 2310) {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  1m  
1s{color} | {color:green} the patch passed {color} |
|| || || || {color:brown} Other Tests {color} ||
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
14s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black} 23m 57s{color} | 
{color:black} {color} |
\\
\\
|| Reason || Tests ||
| FindBugs | module:ql |
|  |  Exception is caught when Exception is not thrown in 
org.apache.hadoop.hive.ql.metadata.Hive.reloadFunctions()  At Hive.java:is not 
thrown in org.apache.hadoop.hive.ql.metadata.Hive.reloadFunctions()  At 
Hive.java:[line 246] |
\\
\\
|| Subsystem || Report/Notes ||
| Optional Tests |  asflicense  javac  javadoc  findbugs  checkstyle  compile  |
| uname | Linux hiveptest-server-upstream 3.16.0-4-amd64 #1 SMP Debian 
3.16.36-1+deb8u1 (2016-09-03) x86_64 GNU/Linux |
| Build tool | maven |
| Personality | 
/data/hiveptest/working/yetus_PreCommit-HIVE-Build-13509/dev-support/hive-personality.sh
 |
| git revision | master / 7cf1d00 |
| Default Java | 1.8.0_111 |
| findbugs | v3.0.0 |
| findbugs | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-13509/yetus/new-findbugs-ql.html
 |
| modules | C: ql U: ql |
| Console output | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-13509/yetus.txt |
| Powered by | Apache Yetushttp://yetus.apache.org |


This message was automatically generated.



> Improve org.apache.hadoop.hive.ql.exec.FunctionTask Experience
> --
>
> Key: HIVE-20466
> URL: https://issues.apache.org/jira/browse/HIVE-20466
> Project: Hive
>  Issue Type: Improvement
>  Components: Hive
>Reporter: denys kuzmenko
>Assignee: denys kuzmenko
>Priority: Major
> Attachments: HIVE-20466.1.patch, HIVE-20466.2.patch, 
> HIVE-20466.3.patch, HIVE-20466.4.patch
>
>




--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-18908) FULL OUTER JOIN to MapJoin

2018-08-28 Thread Matt McCline (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-18908?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Matt McCline updated HIVE-18908:

Status: Patch Available  (was: In Progress)

> FULL OUTER JOIN to MapJoin
> --
>
> Key: HIVE-18908
> URL: https://issues.apache.org/jira/browse/HIVE-18908
> Project: Hive
>  Issue Type: Improvement
>  Components: Hive
>Reporter: Matt McCline
>Assignee: Matt McCline
>Priority: Critical
> Attachments: FULL OUTER MapJoin Code Changes.docx, 
> HIVE-18908.01.patch, HIVE-18908.02.patch, HIVE-18908.03.patch, 
> HIVE-18908.04.patch, HIVE-18908.05.patch, HIVE-18908.06.patch, 
> HIVE-18908.08.patch, HIVE-18908.09.patch, HIVE-18908.091.patch, 
> HIVE-18908.092.patch, HIVE-18908.093.patch, HIVE-18908.096.patch, 
> HIVE-18908.097.patch, HIVE-18908.098.patch, JOIN to MAPJOIN 
> Transformation.pdf, SHARED-MEMORY FULL OUTER MapJoin.pdf
>
>
> Currently, we do not support FULL OUTER JOIN in MapJoin.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-18908) FULL OUTER JOIN to MapJoin

2018-08-28 Thread Matt McCline (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-18908?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Matt McCline updated HIVE-18908:

Attachment: HIVE-18908.098.patch

> FULL OUTER JOIN to MapJoin
> --
>
> Key: HIVE-18908
> URL: https://issues.apache.org/jira/browse/HIVE-18908
> Project: Hive
>  Issue Type: Improvement
>  Components: Hive
>Reporter: Matt McCline
>Assignee: Matt McCline
>Priority: Critical
> Attachments: FULL OUTER MapJoin Code Changes.docx, 
> HIVE-18908.01.patch, HIVE-18908.02.patch, HIVE-18908.03.patch, 
> HIVE-18908.04.patch, HIVE-18908.05.patch, HIVE-18908.06.patch, 
> HIVE-18908.08.patch, HIVE-18908.09.patch, HIVE-18908.091.patch, 
> HIVE-18908.092.patch, HIVE-18908.093.patch, HIVE-18908.096.patch, 
> HIVE-18908.097.patch, HIVE-18908.098.patch, JOIN to MAPJOIN 
> Transformation.pdf, SHARED-MEMORY FULL OUTER MapJoin.pdf
>
>
> Currently, we do not support FULL OUTER JOIN in MapJoin.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-18908) FULL OUTER JOIN to MapJoin

2018-08-28 Thread Matt McCline (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-18908?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Matt McCline updated HIVE-18908:

Status: In Progress  (was: Patch Available)

> FULL OUTER JOIN to MapJoin
> --
>
> Key: HIVE-18908
> URL: https://issues.apache.org/jira/browse/HIVE-18908
> Project: Hive
>  Issue Type: Improvement
>  Components: Hive
>Reporter: Matt McCline
>Assignee: Matt McCline
>Priority: Critical
> Attachments: FULL OUTER MapJoin Code Changes.docx, 
> HIVE-18908.01.patch, HIVE-18908.02.patch, HIVE-18908.03.patch, 
> HIVE-18908.04.patch, HIVE-18908.05.patch, HIVE-18908.06.patch, 
> HIVE-18908.08.patch, HIVE-18908.09.patch, HIVE-18908.091.patch, 
> HIVE-18908.092.patch, HIVE-18908.093.patch, HIVE-18908.096.patch, 
> HIVE-18908.097.patch, JOIN to MAPJOIN Transformation.pdf, SHARED-MEMORY FULL 
> OUTER MapJoin.pdf
>
>
> Currently, we do not support FULL OUTER JOIN in MapJoin.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-19748) Add appropriate null checks to DecimalColumnStatsAggregator

2018-08-28 Thread Vaibhav Gumashta (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-19748?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Vaibhav Gumashta updated HIVE-19748:

Status: Patch Available  (was: Open)

> Add appropriate null checks to DecimalColumnStatsAggregator
> ---
>
> Key: HIVE-19748
> URL: https://issues.apache.org/jira/browse/HIVE-19748
> Project: Hive
>  Issue Type: Bug
>  Components: Metastore
>Affects Versions: 3.0.0
>Reporter: Vaibhav Gumashta
>Assignee: Vaibhav Gumashta
>Priority: Major
> Attachments: HIVE-19748.1.patch, HIVE-19748.1.patch
>
>
> In some of our internal testing, we noticed that calls to 
> MetaStoreUtils.decimalToDoublee(Decimal decimal)  from within 
> DecimalColumnStatsAggregator end up passing null Decimal values to the method.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-19748) Add appropriate null checks to DecimalColumnStatsAggregator

2018-08-28 Thread Vaibhav Gumashta (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-19748?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Vaibhav Gumashta updated HIVE-19748:

Attachment: HIVE-19748.1.patch

> Add appropriate null checks to DecimalColumnStatsAggregator
> ---
>
> Key: HIVE-19748
> URL: https://issues.apache.org/jira/browse/HIVE-19748
> Project: Hive
>  Issue Type: Bug
>  Components: Metastore
>Affects Versions: 3.0.0
>Reporter: Vaibhav Gumashta
>Assignee: Vaibhav Gumashta
>Priority: Major
> Attachments: HIVE-19748.1.patch, HIVE-19748.1.patch
>
>
> In some of our internal testing, we noticed that calls to 
> MetaStoreUtils.decimalToDoublee(Decimal decimal)  from within 
> DecimalColumnStatsAggregator end up passing null Decimal values to the method.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-19748) Add appropriate null checks to DecimalColumnStatsAggregator

2018-08-28 Thread Vaibhav Gumashta (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-19748?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Vaibhav Gumashta updated HIVE-19748:

Status: Open  (was: Patch Available)

> Add appropriate null checks to DecimalColumnStatsAggregator
> ---
>
> Key: HIVE-19748
> URL: https://issues.apache.org/jira/browse/HIVE-19748
> Project: Hive
>  Issue Type: Bug
>  Components: Metastore
>Affects Versions: 3.0.0
>Reporter: Vaibhav Gumashta
>Assignee: Vaibhav Gumashta
>Priority: Major
> Attachments: HIVE-19748.1.patch, HIVE-19748.1.patch
>
>
> In some of our internal testing, we noticed that calls to 
> MetaStoreUtils.decimalToDoublee(Decimal decimal)  from within 
> DecimalColumnStatsAggregator end up passing null Decimal values to the method.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Resolved] (HIVE-20478) Metastore: Null checks needed in DecimalColumnStatsAggregator

2018-08-28 Thread Vaibhav Gumashta (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-20478?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Vaibhav Gumashta resolved HIVE-20478.
-
Resolution: Duplicate

> Metastore: Null checks needed in DecimalColumnStatsAggregator
> -
>
> Key: HIVE-20478
> URL: https://issues.apache.org/jira/browse/HIVE-20478
> Project: Hive
>  Issue Type: Bug
>  Components: Standalone Metastore
>Affects Versions: 3.1.0
>Reporter: Vaibhav Gumashta
>Priority: Major
>




--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-20433) Implicit String to Timestamp conversion is slow

2018-08-28 Thread Deepak Jaiswal (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-20433?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16595232#comment-16595232
 ] 

Deepak Jaiswal commented on HIVE-20433:
---

Ping [~gopalv] [~ashutoshc]

> Implicit String to Timestamp conversion is slow
> ---
>
> Key: HIVE-20433
> URL: https://issues.apache.org/jira/browse/HIVE-20433
> Project: Hive
>  Issue Type: Bug
>Reporter: Deepak Jaiswal
>Assignee: Deepak Jaiswal
>Priority: Major
> Attachments: HIVE-20433.1.patch, HIVE-20433.2.patch, 
> HIVE-20433.3.patch, HIVE-20433.4.patch, HIVE-20433.5.patch, 
> HIVE-20433.6.patch, HIVE-20433.7.patch, HIVE-20433.8.patch, HIVE-20433.9.patch
>
>
> getTimestampFromString() is slow at casting dates. It throws twice before 
> date conversion can happen.
>  
> cc [~gopalv] [~ashutoshc]



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-20465) ProxyFileSystem.listStatusIterator function override required once migrated to Hadoop 3.2.0+

2018-08-28 Thread Hive QA (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-20465?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16595220#comment-16595220
 ] 

Hive QA commented on HIVE-20465:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12937439/HIVE-20465.7.patch

{color:green}SUCCESS:{color} +1 due to 1 test(s) being added or modified.

{color:green}SUCCESS:{color} +1 due to 14900 tests passed

Test results: 
https://builds.apache.org/job/PreCommit-HIVE-Build/13508/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/13508/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-13508/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.YetusPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12937439 - PreCommit-HIVE-Build

> ProxyFileSystem.listStatusIterator function override required once migrated 
> to Hadoop 3.2.0+
> 
>
> Key: HIVE-20465
> URL: https://issues.apache.org/jira/browse/HIVE-20465
> Project: Hive
>  Issue Type: Test
>  Components: Tests
>Reporter: denys kuzmenko
>Assignee: denys kuzmenko
>Priority: Major
> Attachments: HIVE-20465.1.patch, HIVE-20465.2.patch, 
> HIVE-20465.3.patch, HIVE-20465.4.patch, HIVE-20465.5.patch, 
> HIVE-20465.6.patch, HIVE-20465.7.patch
>
>




--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-20377) Hive Kafka Storage Handler

2018-08-28 Thread Gopal V (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-20377?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16595211#comment-16595211
 ] 

Gopal V commented on HIVE-20377:


LGTM - +1

> Hive Kafka Storage Handler
> --
>
> Key: HIVE-20377
> URL: https://issues.apache.org/jira/browse/HIVE-20377
> Project: Hive
>  Issue Type: New Feature
>Affects Versions: 4.0.0
>Reporter: slim bouguerra
>Assignee: slim bouguerra
>Priority: Major
> Attachments: HIVE-20377.10.patch, HIVE-20377.11.patch, 
> HIVE-20377.12.patch, HIVE-20377.15.patch, HIVE-20377.4.patch, 
> HIVE-20377.5.patch, HIVE-20377.6.patch, HIVE-20377.8.patch, 
> HIVE-20377.8.patch, HIVE-20377.patch
>
>
> h1. Goal
> * Read streaming data form Kafka queue as an external table.
> * Allow streaming navigation by pushing down filters on Kafka record 
> partition id, offset and timestamp. 
> * Insert streaming data form Kafka to an actual Hive internal table, using 
> CTAS statement.
> h1. Example
> h2. Create the external table
> {code} 
> CREATE EXTERNAL TABLE kafka_table (`timestamp` timestamp, page string, `user` 
> string, language string, added int, deleted int, flags string,comment string, 
> namespace string)
> STORED BY 'org.apache.hadoop.hive.kafka.KafkaStorageHandler'
> TBLPROPERTIES 
> ("kafka.topic" = "wikipedia", 
> "kafka.bootstrap.servers"="brokeraddress:9092",
> "kafka.serde.class"="org.apache.hadoop.hive.serde2.JsonSerDe");
> {code}
> h2. Kafka Metadata
> In order to keep track of Kafka records the storage handler will add 
> automatically the Kafka row metadata eg partition id, record offset and 
> record timestamp. 
> {code}
> DESCRIBE EXTENDED kafka_table
> timestamp timestamp   from deserializer   
> page  string  from deserializer   
> user  string  from deserializer   
> language  string  from deserializer   
> country   string  from deserializer   
> continent string  from deserializer   
> namespace string  from deserializer   
> newpage   boolean from deserializer   
> unpatrolled   boolean from deserializer   
> anonymous boolean from deserializer   
> robot boolean from deserializer   
> added int from deserializer   
> deleted   int from deserializer   
> delta bigint  from deserializer   
> __partition   int from deserializer   
> __offset  bigint  from deserializer   
> __timestamp   bigint  from deserializer   
> {code}
> h2. Filter push down.
> Newer Kafka consumers 0.11.0 and higher allow seeking on the stream based on 
> a given offset. The proposed storage handler will be able to leverage such 
> API by pushing down filters over metadata columns, namely __partition (int), 
> __offset(long) and __timestamp(long)
> For instance Query like
> {code} 
> select `__offset` from kafka_table where (`__offset` < 10 and `__offset`>3 
> and `__partition` = 0) or (`__partition` = 0 and `__offset` < 105 and 
> `__offset` > 99) or (`__offset` = 109);
> {code}
> Will result on a scan of partition 0 only then read only records between 
> offset 4 and 109. 
> h2. With timestamp seeks 
> The seeking based on the internal timestamps allows the handler to run on 
> recently arrived data, by doing
> {code}
> select count(*) from kafka_table where `__timestamp` >  1000 * 
> to_unix_timestamp(CURRENT_TIMESTAMP - interval '20' hours) ;
> {code}
> This allows for implicit relationships between event timestamps and kafka 
> timestamps to be expressed in queries (i.e event_timestamp is always < than 
> kafka __timestamp and kafka __timestamp is never > 15 minutes from event etc).
> h2. More examples with Avro 
> {code}
> CREATE EXTERNAL TABLE wiki_kafka_avro_table
> STORED BY 'org.apache.hadoop.hive.kafka.KafkaStorageHandler'
> TBLPROPERTIES
> ("kafka.topic" = "wiki_kafka_avro_table",
> "kafka.bootstrap.servers"="localhost:9092",
> "kafka.serde.class"="org.apache.hadoop.hive.serde2.avro.AvroSerDe",
> 'avro.schema.literal'='{
>   "type" : "record",
>   "name" : "Wikipedia",
>   "namespace" : "org.apache.hive.kafka",
>   "version": "1",
>   "fields" : [ {
> "name" : "isrobot",
> "type" : "boolean"
>   }, {
> "name" : "channel",
> "type" : "string"
>   }, {
> "name" : "timestamp",
> "type" : "string"
>   }, {
> "name" : "flags",
> "type" : "string"
>   }, {
> "name" : "isunpatrolled",
> "type" : "boolean"
>   }, {
> "name" : "page",
> "type" : 

[jira] [Commented] (HIVE-20476) CopyUtils used by REPL LOAD and EXPORT/IMPORT operations ignore distcp error.

2018-08-28 Thread Sankar Hariappan (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-20476?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16595201#comment-16595201
 ] 

Sankar Hariappan commented on HIVE-20476:
-

[~maheshk114], can you please take a look at the patch?

> CopyUtils used by REPL LOAD and EXPORT/IMPORT operations ignore distcp error.
> -
>
> Key: HIVE-20476
> URL: https://issues.apache.org/jira/browse/HIVE-20476
> Project: Hive
>  Issue Type: Bug
>  Components: HiveServer2, repl
>Affects Versions: 3.1.0
>Reporter: Sankar Hariappan
>Assignee: Sankar Hariappan
>Priority: Major
>  Labels: DR, Export, Import, Replication, pull-request-available
> Attachments: HIVE-20476.01.patch
>
>
> CopyUtils uses FileUtils.distCp to copy files but doesn't check the return 
> value. It returns false if the copy fails.
> Now, REPL LOAD and EXPORT/IMPORT commands internally uses CopyUtils to copy 
> data files across clusters and here it may return success even if file copy 
> fails and may cause data loss.
> Need to throw error and retry.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-20371) Fix definition for add_write_notification_log in hive_metastore.thrift.

2018-08-28 Thread ASF GitHub Bot (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-20371?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16595199#comment-16595199
 ] 

ASF GitHub Bot commented on HIVE-20371:
---

Github user sankarh closed the pull request at:

https://github.com/apache/hive/pull/416


> Fix definition for add_write_notification_log in hive_metastore.thrift.
> ---
>
> Key: HIVE-20371
> URL: https://issues.apache.org/jira/browse/HIVE-20371
> Project: Hive
>  Issue Type: Improvement
>  Components: repl
>Affects Versions: 4.0.0, 3.2.0
>Reporter: Sankar Hariappan
>Assignee: Sankar Hariappan
>Priority: Minor
>  Labels: DR, pull-request-available, replication
> Fix For: 4.0.0, 3.2.0
>
> Attachments: HIVE-20371.01.patch
>
>
> The definition of add_write_notification_log metastore api in 
> hive_metastore.thrift doesn't follow the standard format.
> WriteNotificationLogResponse 
> add_write_notification_log(WriteNotificationLogRequest rqst)
> to
> WriteNotificationLogResponse 
> add_write_notification_log(1:WriteNotificationLogRequest rqst)



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-20264) Bootstrap repl dump with concurrent write and drop of ACID table makes target inconsistent.

2018-08-28 Thread ASF GitHub Bot (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-20264?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16595198#comment-16595198
 ] 

ASF GitHub Bot commented on HIVE-20264:
---

Github user sankarh closed the pull request at:

https://github.com/apache/hive/pull/410


> Bootstrap repl dump with concurrent write and drop of ACID table makes target 
> inconsistent.
> ---
>
> Key: HIVE-20264
> URL: https://issues.apache.org/jira/browse/HIVE-20264
> Project: Hive
>  Issue Type: Sub-task
>  Components: HiveServer2, repl
>Affects Versions: 4.0.0, 3.2.0
>Reporter: Sankar Hariappan
>Assignee: Sankar Hariappan
>Priority: Major
>  Labels: DR, pull-request-available, replication
> Fix For: 4.0.0, 3.2.0
>
> Attachments: HIVE-20264.01-branch-3.patch, HIVE-20264.01.patch, 
> HIVE-20264.02.patch
>
>
> During bootstrap dump of ACID tables, let's consider the below sequence.
>  - Get lastReplId = last event ID logged.
>  - Current session (Thread-1), REPL DUMP -> Open txn (Txn1) - Event-10
>  - Another session (Thread-2), Open txn (Txn2) - Event-11
>  - Thread-2 -> Insert data (T1.D1) to ACID table. - Event-12
>  - Thread-2 -> Commit Txn (Txn2) - Event-13
>  - Thread-2 -> Drop table (T1) - Event-14
>  - Thread-1 -> Dump ACID tables based on current list of tables. So, T1 will 
> be missing.
>  - Thread-1 -> Commit Txn (Txn1)
>  - REPL LOAD from bootstrap dump will skip T1.
>  - Incremental REPL DUMP will start from Event-10 and hence allocate write id 
> for table T1 and drop table(T1) is idempotent. So, at target, exist entries 
> in TXN_TO_WRITE_ID and NEXT_WRITE_ID metastore tables.
>  - Now, when we create another table at source with same name T1 and 
> replicate, then it may lead to incorrect data for readers at target on T1.
> Couple of proposals:
>  1. Make allocate write ID idempotent which is not possible as table doesn't 
> exist and MM table import may lead to allocate write id before creating 
> table. So, cannot differentiate these 2 cases.
>  2. Make Drop table event to drop entries from TXN_TO_WRITE_ID and 
> NEXT_WRITE_ID tables irrespective of table exist or not at target.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-20476) CopyUtils used by REPL LOAD and EXPORT/IMPORT operations ignore distcp error.

2018-08-28 Thread ASF GitHub Bot (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-20476?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16595177#comment-16595177
 ] 

ASF GitHub Bot commented on HIVE-20476:
---

GitHub user sankarh opened a pull request:

https://github.com/apache/hive/pull/424

HIVE-20476: CopyUtils used by REPL LOAD and EXPORT/IMPORT operations ignore 
distcp error.



You can merge this pull request into a Git repository by running:

$ git pull https://github.com/sankarh/hive HIVE-20476

Alternatively you can review and apply these changes as the patch at:

https://github.com/apache/hive/pull/424.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

This closes #424






> CopyUtils used by REPL LOAD and EXPORT/IMPORT operations ignore distcp error.
> -
>
> Key: HIVE-20476
> URL: https://issues.apache.org/jira/browse/HIVE-20476
> Project: Hive
>  Issue Type: Bug
>  Components: HiveServer2, repl
>Affects Versions: 3.1.0
>Reporter: Sankar Hariappan
>Assignee: Sankar Hariappan
>Priority: Major
>  Labels: DR, Export, Import, Replication, pull-request-available
> Attachments: HIVE-20476.01.patch
>
>
> CopyUtils uses FileUtils.distCp to copy files but doesn't check the return 
> value. It returns false if the copy fails.
> Now, REPL LOAD and EXPORT/IMPORT commands internally uses CopyUtils to copy 
> data files across clusters and here it may return success even if file copy 
> fails and may cause data loss.
> Need to throw error and retry.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-20476) CopyUtils used by REPL LOAD and EXPORT/IMPORT operations ignore distcp error.

2018-08-28 Thread ASF GitHub Bot (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-20476?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ASF GitHub Bot updated HIVE-20476:
--
Labels: DR Export Import Replication pull-request-available  (was: DR 
Export Import Replication)

> CopyUtils used by REPL LOAD and EXPORT/IMPORT operations ignore distcp error.
> -
>
> Key: HIVE-20476
> URL: https://issues.apache.org/jira/browse/HIVE-20476
> Project: Hive
>  Issue Type: Bug
>  Components: HiveServer2, repl
>Affects Versions: 3.1.0
>Reporter: Sankar Hariappan
>Assignee: Sankar Hariappan
>Priority: Major
>  Labels: DR, Export, Import, Replication, pull-request-available
> Attachments: HIVE-20476.01.patch
>
>
> CopyUtils uses FileUtils.distCp to copy files but doesn't check the return 
> value. It returns false if the copy fails.
> Now, REPL LOAD and EXPORT/IMPORT commands internally uses CopyUtils to copy 
> data files across clusters and here it may return success even if file copy 
> fails and may cause data loss.
> Need to throw error and retry.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-20476) CopyUtils used by REPL LOAD and EXPORT/IMPORT operations ignore distcp error.

2018-08-28 Thread Sankar Hariappan (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-20476?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sankar Hariappan updated HIVE-20476:

Status: Patch Available  (was: Open)

> CopyUtils used by REPL LOAD and EXPORT/IMPORT operations ignore distcp error.
> -
>
> Key: HIVE-20476
> URL: https://issues.apache.org/jira/browse/HIVE-20476
> Project: Hive
>  Issue Type: Bug
>  Components: HiveServer2, repl
>Affects Versions: 3.1.0
>Reporter: Sankar Hariappan
>Assignee: Sankar Hariappan
>Priority: Major
>  Labels: DR, Export, Import, Replication
> Attachments: HIVE-20476.01.patch
>
>
> CopyUtils uses FileUtils.distCp to copy files but doesn't check the return 
> value. It returns false if the copy fails.
> Now, REPL LOAD and EXPORT/IMPORT commands internally uses CopyUtils to copy 
> data files across clusters and here it may return success even if file copy 
> fails and may cause data loss.
> Need to throw error and retry.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-20476) CopyUtils used by REPL LOAD and EXPORT/IMPORT operations ignore distcp error.

2018-08-28 Thread Sankar Hariappan (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-20476?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sankar Hariappan updated HIVE-20476:

Attachment: (was: HIVE-20476.01.patch)

> CopyUtils used by REPL LOAD and EXPORT/IMPORT operations ignore distcp error.
> -
>
> Key: HIVE-20476
> URL: https://issues.apache.org/jira/browse/HIVE-20476
> Project: Hive
>  Issue Type: Bug
>  Components: HiveServer2, repl
>Affects Versions: 3.1.0
>Reporter: Sankar Hariappan
>Assignee: Sankar Hariappan
>Priority: Major
>  Labels: DR, Export, Import, Replication
> Attachments: HIVE-20476.01.patch
>
>
> CopyUtils uses FileUtils.distCp to copy files but doesn't check the return 
> value. It returns false if the copy fails.
> Now, REPL LOAD and EXPORT/IMPORT commands internally uses CopyUtils to copy 
> data files across clusters and here it may return success even if file copy 
> fails and may cause data loss.
> Need to throw error and retry.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-20476) CopyUtils used by REPL LOAD and EXPORT/IMPORT operations ignore distcp error.

2018-08-28 Thread Sankar Hariappan (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-20476?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sankar Hariappan updated HIVE-20476:

Attachment: HIVE-20476.01.patch

> CopyUtils used by REPL LOAD and EXPORT/IMPORT operations ignore distcp error.
> -
>
> Key: HIVE-20476
> URL: https://issues.apache.org/jira/browse/HIVE-20476
> Project: Hive
>  Issue Type: Bug
>  Components: HiveServer2, repl
>Affects Versions: 3.1.0
>Reporter: Sankar Hariappan
>Assignee: Sankar Hariappan
>Priority: Major
>  Labels: DR, Export, Import, Replication
> Attachments: HIVE-20476.01.patch
>
>
> CopyUtils uses FileUtils.distCp to copy files but doesn't check the return 
> value. It returns false if the copy fails.
> Now, REPL LOAD and EXPORT/IMPORT commands internally uses CopyUtils to copy 
> data files across clusters and here it may return success even if file copy 
> fails and may cause data loss.
> Need to throw error and retry.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-20476) CopyUtils used by REPL LOAD and EXPORT/IMPORT operations ignore distcp error.

2018-08-28 Thread Sankar Hariappan (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-20476?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sankar Hariappan updated HIVE-20476:

Attachment: HIVE-20476.01.patch

> CopyUtils used by REPL LOAD and EXPORT/IMPORT operations ignore distcp error.
> -
>
> Key: HIVE-20476
> URL: https://issues.apache.org/jira/browse/HIVE-20476
> Project: Hive
>  Issue Type: Bug
>  Components: HiveServer2, repl
>Affects Versions: 3.1.0
>Reporter: Sankar Hariappan
>Assignee: Sankar Hariappan
>Priority: Major
>  Labels: DR, Export, Import, Replication
> Attachments: HIVE-20476.01.patch
>
>
> CopyUtils uses FileUtils.distCp to copy files but doesn't check the return 
> value. It returns false if the copy fails.
> Now, REPL LOAD and EXPORT/IMPORT commands internally uses CopyUtils to copy 
> data files across clusters and here it may return success even if file copy 
> fails and may cause data loss.
> Need to throw error and retry.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-20476) CopyUtils used by REPL LOAD and EXPORT/IMPORT operations ignore distcp error.

2018-08-28 Thread Sankar Hariappan (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-20476?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sankar Hariappan updated HIVE-20476:

Summary: CopyUtils used by REPL LOAD and EXPORT/IMPORT operations ignore 
distcp error.  (was: REPL LOAD and EXPORT/IMPORT operations ignores distcp 
failures.)

> CopyUtils used by REPL LOAD and EXPORT/IMPORT operations ignore distcp error.
> -
>
> Key: HIVE-20476
> URL: https://issues.apache.org/jira/browse/HIVE-20476
> Project: Hive
>  Issue Type: Bug
>  Components: HiveServer2, repl
>Affects Versions: 3.1.0
>Reporter: Sankar Hariappan
>Assignee: Sankar Hariappan
>Priority: Major
>  Labels: DR, Export, Import, Replication
>
> CopyUtils uses FileUtils.distCp to copy files but doesn't check the return 
> value. It returns false if the copy fails.
> Now, REPL LOAD and EXPORT/IMPORT commands internally uses CopyUtils to copy 
> data files across clusters and here it may return success even if file copy 
> fails and may cause data loss.
> Need to throw error and retry.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-13457) Create HS2 REST API endpoints for monitoring information

2018-08-28 Thread Szehon Ho (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-13457?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16595153#comment-16595153
 ] 

Szehon Ho commented on HIVE-13457:
--

Actually can you fix the checkstyle and findbugs?

> Create HS2 REST API endpoints for monitoring information
> 
>
> Key: HIVE-13457
> URL: https://issues.apache.org/jira/browse/HIVE-13457
> Project: Hive
>  Issue Type: Improvement
>Reporter: Szehon Ho
>Assignee: Pawel Szostek
>Priority: Major
> Attachments: HIVE-13457.3.patch, HIVE-13457.4.patch, 
> HIVE-13457.5.patch, HIVE-13457.patch, HIVE-13457.patch
>
>
> Similar to what is exposed in HS2 webui in HIVE-12338, it would be nice if 
> other UI's like admin tools or Hue can access and display this information as 
> well.  Hence, we will create some REST endpoints to expose this information.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-13457) Create HS2 REST API endpoints for monitoring information

2018-08-28 Thread Szehon Ho (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-13457?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16595141#comment-16595141
 ] 

Szehon Ho commented on HIVE-13457:
--

Nice +1

> Create HS2 REST API endpoints for monitoring information
> 
>
> Key: HIVE-13457
> URL: https://issues.apache.org/jira/browse/HIVE-13457
> Project: Hive
>  Issue Type: Improvement
>Reporter: Szehon Ho
>Assignee: Pawel Szostek
>Priority: Major
> Attachments: HIVE-13457.3.patch, HIVE-13457.4.patch, 
> HIVE-13457.5.patch, HIVE-13457.patch, HIVE-13457.patch
>
>
> Similar to what is exposed in HS2 webui in HIVE-12338, it would be nice if 
> other UI's like admin tools or Hue can access and display this information as 
> well.  Hence, we will create some REST endpoints to expose this information.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-20465) ProxyFileSystem.listStatusIterator function override required once migrated to Hadoop 3.2.0+

2018-08-28 Thread Hive QA (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-20465?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16595139#comment-16595139
 ] 

Hive QA commented on HIVE-20465:


| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
|| || || || {color:brown} Prechecks {color} ||
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
|| || || || {color:brown} master Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  8m 
38s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  0m 
10s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
10s{color} | {color:green} master passed {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  0m 
19s{color} | {color:blue} shims/common in master has 6 extant Findbugs 
warnings. {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
10s{color} | {color:green} master passed {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  0m 
12s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  0m 
11s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  0m 
11s{color} | {color:green} the patch passed {color} |
| {color:red}-1{color} | {color:red} checkstyle {color} | {color:red}  0m  
9s{color} | {color:red} shims/common: The patch generated 2 new + 14 unchanged 
- 0 fixed = 16 total (was 14) {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  0m 
27s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m  
9s{color} | {color:green} the patch passed {color} |
|| || || || {color:brown} Other Tests {color} ||
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
14s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black} 11m 17s{color} | 
{color:black} {color} |
\\
\\
|| Subsystem || Report/Notes ||
| Optional Tests |  asflicense  javac  javadoc  findbugs  checkstyle  compile  |
| uname | Linux hiveptest-server-upstream 3.16.0-4-amd64 #1 SMP Debian 
3.16.36-1+deb8u1 (2016-09-03) x86_64 GNU/Linux |
| Build tool | maven |
| Personality | 
/data/hiveptest/working/yetus_PreCommit-HIVE-Build-13508/dev-support/hive-personality.sh
 |
| git revision | master / 7cf1d00 |
| Default Java | 1.8.0_111 |
| findbugs | v3.0.0 |
| checkstyle | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-13508/yetus/diff-checkstyle-shims_common.txt
 |
| modules | C: shims/common U: shims/common |
| Console output | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-13508/yetus.txt |
| Powered by | Apache Yetushttp://yetus.apache.org |


This message was automatically generated.



> ProxyFileSystem.listStatusIterator function override required once migrated 
> to Hadoop 3.2.0+
> 
>
> Key: HIVE-20465
> URL: https://issues.apache.org/jira/browse/HIVE-20465
> Project: Hive
>  Issue Type: Test
>  Components: Tests
>Reporter: denys kuzmenko
>Assignee: denys kuzmenko
>Priority: Major
> Attachments: HIVE-20465.1.patch, HIVE-20465.2.patch, 
> HIVE-20465.3.patch, HIVE-20465.4.patch, HIVE-20465.5.patch, 
> HIVE-20465.6.patch, HIVE-20465.7.patch
>
>




--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-13457) Create HS2 REST API endpoints for monitoring information

2018-08-28 Thread Hive QA (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-13457?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16595116#comment-16595116
 ] 

Hive QA commented on HIVE-13457:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12937431/HIVE-13457.4.patch

{color:green}SUCCESS:{color} +1 due to 1 test(s) being added or modified.

{color:green}SUCCESS:{color} +1 due to 14903 tests passed

Test results: 
https://builds.apache.org/job/PreCommit-HIVE-Build/13507/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/13507/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-13507/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.YetusPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12937431 - PreCommit-HIVE-Build

> Create HS2 REST API endpoints for monitoring information
> 
>
> Key: HIVE-13457
> URL: https://issues.apache.org/jira/browse/HIVE-13457
> Project: Hive
>  Issue Type: Improvement
>Reporter: Szehon Ho
>Assignee: Pawel Szostek
>Priority: Major
> Attachments: HIVE-13457.3.patch, HIVE-13457.4.patch, 
> HIVE-13457.5.patch, HIVE-13457.patch, HIVE-13457.patch
>
>
> Similar to what is exposed in HS2 webui in HIVE-12338, it would be nice if 
> other UI's like admin tools or Hue can access and display this information as 
> well.  Hence, we will create some REST endpoints to expose this information.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-20296) Improve HivePointLookupOptimizerRule to be able to extract from more sophisticated contexts

2018-08-28 Thread Zoltan Haindrich (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-20296?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16595099#comment-16595099
 ] 

Zoltan Haindrich commented on HIVE-20296:
-

[~ashutoshc] Could you please take a look?
opened rb: https://reviews.apache.org/r/68525/

> Improve HivePointLookupOptimizerRule to be able to extract from more 
> sophisticated contexts
> ---
>
> Key: HIVE-20296
> URL: https://issues.apache.org/jira/browse/HIVE-20296
> Project: Hive
>  Issue Type: Improvement
>  Components: Logical Optimizer
>Reporter: Zoltan Haindrich
>Assignee: Zoltan Haindrich
>Priority: Major
> Attachments: HIVE-20296.01.patch, HIVE-20296.01wip01.patch, 
> HIVE-20296.01wip02.patch, HIVE-20296.01wip03.patch, HIVE-20296.02.patch, 
> HIVE-20296.02.patch
>
>
> Currently it could handle conditions in the form:
> {code}
> ( ( a=1 && b=1 && ... ) || (a=2 && b=2 && ...) || ... )
> {code}
> If there is some noise somewhere in the or like: 
> {code}
> ( x=18 || ( a=1 && b=1 && ... ) || (a=2 && b=2 && ...) || ... )
> {code}
> it stops processing.
> https://github.com/apache/hive/blob/2cabb8da150b8fb980223fbd6c2c93b842ca3ee5/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/rules/HivePointLookupOptimizerRule.java#L238
> HIVE-19097 right now unwinds some ors which can't be packed back; like query15
> https://github.com/apache/hive/blob/2cabb8da150b8fb980223fbd6c2c93b842ca3ee5/ql/src/test/queries/clientpositive/perf/query15.q#L14



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-17084) Turn on hive.stats.fetch.column.stats configuration flag

2018-08-28 Thread Zoltan Haindrich (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-17084?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zoltan Haindrich updated HIVE-17084:

Attachment: HIVE-17084.08.patch

> Turn on hive.stats.fetch.column.stats configuration flag
> 
>
> Key: HIVE-17084
> URL: https://issues.apache.org/jira/browse/HIVE-17084
> Project: Hive
>  Issue Type: Task
>  Components: Statistics
>Reporter: Vineet Garg
>Assignee: Zoltan Haindrich
>Priority: Major
> Fix For: 3.2.0
>
> Attachments: HIVE-17084.08.patch, HIVE-17084.1.patch, 
> HIVE-170884.4.patch, HIVE-170884.5.patch, HIVE-170884.7.patch
>
>
> This flag is off by default and could result in bad plans due to missing 
> column statistics.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Assigned] (HIVE-17084) Turn on hive.stats.fetch.column.stats configuration flag

2018-08-28 Thread Zoltan Haindrich (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-17084?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zoltan Haindrich reassigned HIVE-17084:
---

Assignee: Vineet Garg  (was: Zoltan Haindrich)

> Turn on hive.stats.fetch.column.stats configuration flag
> 
>
> Key: HIVE-17084
> URL: https://issues.apache.org/jira/browse/HIVE-17084
> Project: Hive
>  Issue Type: Task
>  Components: Statistics
>Reporter: Vineet Garg
>Assignee: Vineet Garg
>Priority: Major
> Fix For: 3.2.0
>
> Attachments: HIVE-17084.08.patch, HIVE-17084.1.patch, 
> HIVE-170884.4.patch, HIVE-170884.5.patch, HIVE-170884.7.patch
>
>
> This flag is off by default and could result in bad plans due to missing 
> column statistics.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Assigned] (HIVE-17084) Turn on hive.stats.fetch.column.stats configuration flag

2018-08-28 Thread Zoltan Haindrich (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-17084?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zoltan Haindrich reassigned HIVE-17084:
---

Assignee: Zoltan Haindrich  (was: Vineet Garg)

> Turn on hive.stats.fetch.column.stats configuration flag
> 
>
> Key: HIVE-17084
> URL: https://issues.apache.org/jira/browse/HIVE-17084
> Project: Hive
>  Issue Type: Task
>  Components: Statistics
>Reporter: Vineet Garg
>Assignee: Zoltan Haindrich
>Priority: Major
> Fix For: 3.2.0
>
> Attachments: HIVE-17084.1.patch, HIVE-170884.4.patch, 
> HIVE-170884.5.patch, HIVE-170884.7.patch
>
>
> This flag is off by default and could result in bad plans due to missing 
> column statistics.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


  1   2   >