Re: Need edit permission to Hive confluence

2018-06-18 Thread Lefty Leverenz
Done.  Welcome to the Hive wiki team, Janaki!

-- Lefty


On Mon, Jun 18, 2018 at 1:46 PM Janaki Latha Lahorani
 wrote:

> Hi,
>
> I would like to update documentation related to some minor enhancements.  I
> would be very grateful if I have the permission to edit.
>
> Username: janulatha
>
> Thanks,
> Janaki.
>


[jira] [Created] (HIVE-19942) Hive Notification: All events for indexes should have table name

2018-06-18 Thread Bharathkrishna Guruvayoor Murali (JIRA)
Bharathkrishna Guruvayoor Murali created HIVE-19942:
---

 Summary: Hive Notification: All events for indexes should have 
table name
 Key: HIVE-19942
 URL: https://issues.apache.org/jira/browse/HIVE-19942
 Project: Hive
  Issue Type: Improvement
  Components: Hive
Affects Versions: 2.3.2
Reporter: Bharathkrishna Guruvayoor Murali
Assignee: Bharathkrishna Guruvayoor Murali


All the events for indexes: Create Index, Alter Index, Drop Index have the 
TBL_NAME as null. The TBL_NAME should be populated with the table on which the 
index is created.

This makes it easier to decide whether to process the event or not without 
needing to parse the json message (which is a slower process).



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (HIVE-19941) Row based Filters added via Hive Ranger policies are not pushed to druid

2018-06-18 Thread Nishant Bangarwa (JIRA)
Nishant Bangarwa created HIVE-19941:
---

 Summary: Row based Filters added via Hive Ranger policies are not 
pushed to druid
 Key: HIVE-19941
 URL: https://issues.apache.org/jira/browse/HIVE-19941
 Project: Hive
  Issue Type: Bug
Reporter: Nishant Bangarwa
Assignee: Nishant Bangarwa


Issue is that when applying table mask we add virtual columns, however 
non-native tables do not have virtual columns, we need to skip adding virtual 
columns when generating masking query. 

Stack Trace - 
{code} 
org.apache.hadoop.hive.ql.parse.SemanticException: Line 1:79 Invalid table 
alias or column reference 'BLOCK__OFFSET__INSIDE__FILE'
: (possible column names are: __time, yearmonth, year, month, dayofmonth, 
dayofweek, weekofyear, hour, minute, second, payment_typ
e, fare_amount, surcharge, mta_tax, tip_amount, tolls_amount, total_amount, 
trip_time)
at 
org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genAllExprNodeDesc(SemanticAnalyzer.java:11830)
 ~[hive-exec-2.1.0.2.6.
4.0-91.jar:2.1.0.2.6.4.0-91]
at 
org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genExprNodeDesc(SemanticAnalyzer.java:11778)
 ~[hive-exec-2.1.0.2.6.4.0
-91.jar:2.1.0.2.6.4.0-91]
at 
org.apache.hadoop.hive.ql.parse.CalcitePlanner$CalcitePlannerAction.genSelectLogicalPlan(CalcitePlanner.java:3780)
 ~[hi
ve-exec-2.1.0.2.6.4.0-91.jar:2.1.0.2.6.4.0-91]
at 
org.apache.hadoop.hive.ql.parse.CalcitePlanner$CalcitePlannerAction.genLogicalPlan(CalcitePlanner.java:4117)
 ~[hive-exe
c-2.1.0.2.6.4.0-91.jar:2.1.0.2.6.4.0-91]
at 
org.apache.hadoop.hive.ql.parse.CalcitePlanner$CalcitePlannerAction.genLogicalPlan(CalcitePlanner.java:4016)
 ~[hive-exe
c-2.1.0.2.6.4.0-91.jar:2.1.0.2.6.4.0-91]
at 
org.apache.hadoop.hive.ql.parse.CalcitePlanner$CalcitePlannerAction.genLogicalPlan(CalcitePlanner.java:4060)
 ~[hive-exe
c-2.1.0.2.6.4.0-91.jar:2.1.0.2.6.4.0-91]
at 
org.apache.hadoop.hive.ql.parse.CalcitePlanner$CalcitePlannerAction.apply(CalcitePlanner.java:1340)
 ~[hive-exec-2.1.0.2
.6.4.0-91.jar:2.1.0.2.6.4.0-91]
at 
org.apache.hadoop.hive.ql.parse.CalcitePlanner$CalcitePlannerAction.apply(CalcitePlanner.java:1277)
 ~[hive-exec-2.1.0.2
.6.4.0-91.jar:2.1.0.2.6.4.0-91]
at org.apache.calcite.tools.Frameworks$1.apply(Frameworks.java:113) 
~[calcite-core-1.10.0.2.6.4.0-91.jar:1.10.0.2.6.4.0-91
]
at 
org.apache.calcite.prepare.CalcitePrepareImpl.perform(CalcitePrepareImpl.java:997)
 ~[calcite-core-1.10.0.2.6.4.0-91.jar
:1.10.0.2.6.4.0-91]
at org.apache.calcite.tools.Frameworks.withPrepare(Frameworks.java:149) 
~[calcite-core-1.10.0.2.6.4.0-91.jar:1.10.0.2.6.4.
0-91]
at org.apache.calcite.tools.Frameworks.withPlanner(Frameworks.java:106) 
~[calcite-core-1.10.0.2.6.4.0-91.jar:1.10.0.2.6.4.
0-91]
at 
org.apache.hadoop.hive.ql.parse.CalcitePlanner.logicalPlan(CalcitePlanner.java:1082)
 ~[hive-exec-2.1.0.2.6.4.0-91.jar:2
.1.0.2.6.4.0-91]
{code} 



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (HIVE-19940) Push predicates with deterministic UDFs with RBO

2018-06-18 Thread Janaki Lahorani (JIRA)
Janaki Lahorani created HIVE-19940:
--

 Summary: Push predicates with deterministic UDFs with RBO
 Key: HIVE-19940
 URL: https://issues.apache.org/jira/browse/HIVE-19940
 Project: Hive
  Issue Type: Bug
Reporter: Janaki Lahorani
Assignee: Janaki Lahorani


With RBO, predicates with any UDF doesn't get pushed down.  It makes sense to 
not pushdown the predicates with non-deterministic function as the meaning of 
the query changes after the predicate is resolved to use the function.  But 
pushing a deterministic function is beneficial.

Test Case:
{code}
set hive.cbo.enable=false;

CREATE TABLE `testb`(
   `cola` string COMMENT '',
   `colb` string COMMENT '',
   `colc` string COMMENT '')
PARTITIONED BY (
   `part1` string,
   `part2` string,
   `part3` string)

STORED AS AVRO;

CREATE TABLE `testa`(
   `col1` string COMMENT '',
   `col2` string COMMENT '',
   `col3` string COMMENT '',
   `col4` string COMMENT '',
   `col5` string COMMENT '')
PARTITIONED BY (
   `part1` string,
   `part2` string,
   `part3` string)
STORED AS AVRO;

insert into testA partition (part1='US', part2='ABC', part3='123')
values ('12.34', '100', '200', '300', 'abc'),
('12.341', '1001', '2001', '3001', 'abcd');

insert into testA partition (part1='UK', part2='DEF', part3='123')
values ('12.34', '100', '200', '300', 'abc'),
('12.341', '1001', '2001', '3001', 'abcd');

insert into testA partition (part1='US', part2='DEF', part3='200')
values ('12.34', '100', '200', '300', 'abc'),
('12.341', '1001', '2001', '3001', 'abcd');

insert into testA partition (part1='CA', part2='ABC', part3='300')
values ('12.34', '100', '200', '300', 'abc'),
('12.341', '1001', '2001', '3001', 'abcd');

insert into testB partition (part1='CA', part2='ABC', part3='300')
values ('600', '700', 'abc'), ('601', '701', 'abcd');

insert into testB partition (part1='CA', part2='ABC', part3='400')
values ( '600', '700', 'abc'), ( '601', '701', 'abcd');

insert into testB partition (part1='UK', part2='PQR', part3='500')
values ('600', '700', 'abc'), ('601', '701', 'abcd');

insert into testB partition (part1='US', part2='DEF', part3='200')
values ( '600', '700', 'abc'), ('601', '701', 'abcd');

insert into testB partition (part1='US', part2='PQR', part3='123')
values ( '600', '700', 'abc'), ('601', '701', 'abcd');

-- views with deterministic functions
create view viewDeterministicUDFA partitioned on (vpart1, vpart2, vpart3) as 
select
 cast(col1 as decimal(38,18)) as vcol1,
 cast(col2 as decimal(38,18)) as vcol2,
 cast(col3 as decimal(38,18)) as vcol3,
 cast(col4 as decimal(38,18)) as vcol4,
 cast(col5 as char(10)) as vcol5,
 cast(part1 as char(2)) as vpart1,
 cast(part2 as char(3)) as vpart2,
 cast(part3 as char(3)) as vpart3
 from testa
where part1 in ('US', 'CA');

create view viewDeterministicUDFB partitioned on (vpart1, vpart2, vpart3) as 
select
 cast(cola as decimal(38,18)) as vcolA,
 cast(colb as decimal(38,18)) as vcolB,
 cast(colc as char(10)) as vcolC,
 cast(part1 as char(2)) as vpart1,
 cast(part2 as char(3)) as vpart2,
 cast(part3 as char(3)) as vpart3
 from testb
where part1 in ('US', 'CA');

explain
select vcol1, vcol2, vcol3, vcola, vcolb
from viewDeterministicUDFA a inner join viewDeterministicUDFB b
on a.vpart1 = b.vpart1
and a.vpart2 = b.vpart2
and a.vpart3 = b.vpart3
and a.vpart1 = 'US'
and a.vpart2 = 'DEF'
and a.vpart3 = '200';
{code}

Plan where the CAST is not pushed down.
{code}
STAGE PLANS:
  Stage: Stage-1
Map Reduce
  Map Operator Tree:
  TableScan
alias: testa
filterExpr: (part1) IN ('US', 'CA') (type: boolean)
Statistics: Num rows: 6 Data size: 13740 Basic stats: COMPLETE 
Column stats: NONE
Select Operator
  expressions: CAST( col1 AS decimal(38,18)) (type: 
decimal(38,18)), CAST( col2 AS decimal(38,18)) (type: decimal(38,18)), CAST( 
col3 AS decimal(38,18)) (type: decimal(38,18)), CAST( part1 AS CHAR(2)) (type: 
char(2)), CAST( part2 AS CHAR(3)) (type: char(3)), CAST( part3 AS CHAR(3)) 
(type: char(3))
  outputColumnNames: _col0, _col1, _col2, _col5, _col6, _col7
  Statistics: Num rows: 6 Data size: 13740 Basic stats: COMPLETE 
Column stats: NONE
  Filter Operator
predicate: ((_col5 = 'US') and (_col6 = 'DEF') and (_col7 = 
'200')) (type: boolean)
Statistics: Num rows: 1 Data size: 2290 Basic stats: COMPLETE 
Column stats: NONE
Reduce Output Operator
  key expressions: 'US' (type: char(2)), 'DEF' (type: char(3)), 
'200' (type: char(3))
  sort order: +++
  Map-reduce partition columns: 'US' (type: char(2)), 'DEF' 
(type: char(3)), '200' (type: char(3))
  Statistics: Num rows: 1 Data size: 2290 Basic stats: COMPLETE 
Column stats: NONE
  value expressions: _col0 (type: 

[jira] [Created] (HIVE-19939) Verify any other aggregation functions other than COUNT

2018-06-18 Thread Steve Yeom (JIRA)
Steve Yeom created HIVE-19939:
-

 Summary: Verify any other aggregation functions other than COUNT
 Key: HIVE-19939
 URL: https://issues.apache.org/jira/browse/HIVE-19939
 Project: Hive
  Issue Type: Sub-task
  Components: Transactions
Affects Versions: 4.0.0
Reporter: Steve Yeom
 Fix For: 4.0.0


1. More on MAX
I have added MAX query into stats_part.q and stats_nonpart.q but showed a 
slightly different
explain.out which might be a bug.
2. Other functions than MAX and COUNT.
Also I think we need to check other possible aggregation functions than MAX and 
COUNT.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (HIVE-19938) Upgrade scripts for information schema

2018-06-18 Thread Daniel Dai (JIRA)
Daniel Dai created HIVE-19938:
-

 Summary: Upgrade scripts for information schema
 Key: HIVE-19938
 URL: https://issues.apache.org/jira/browse/HIVE-19938
 Project: Hive
  Issue Type: Bug
Reporter: Daniel Dai
Assignee: Daniel Dai


To make schematool -upgradeSchema work for information schema.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


Re: Review Request 67636: HIVE-19176 : Add HoS support to progress bar on Beeline client.

2018-06-18 Thread Bharathkrishna Guruvayoor Murali via Review Board

---
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/67636/#review204951
---




jdbc/src/java/org/apache/hive/jdbc/HiveStatement.java
Lines 379 (patched)


Added this condition because in the do-while loop, when this executes after 
the results are printed, the results get overwritten by progress bar. So if 
operation is complete, no need to update.


- Bharathkrishna Guruvayoor Murali


On June 18, 2018, 8:41 p.m., Bharathkrishna Guruvayoor Murali wrote:
> 
> ---
> This is an automatically generated e-mail. To reply, visit:
> https://reviews.apache.org/r/67636/
> ---
> 
> (Updated June 18, 2018, 8:41 p.m.)
> 
> 
> Review request for hive, Peter Vary, Sahil Takiar, and Vihang Karajgaonkar.
> 
> 
> Repository: hive-git
> 
> 
> Description
> ---
> 
> This logic is similar to the RenderStrategy used in Tez to print the progress 
> bar.
> 
> 
> Diffs
> -
> 
>   common/src/java/org/apache/hadoop/hive/conf/HiveConf.java 
> 933bda4ad01a6f7878019a7b4c971a0c39068ae2 
>   jdbc/src/java/org/apache/hive/jdbc/HiveStatement.java 
> ad8d1a7f1cca3a763bb7c07335998ab7d39d7598 
>   
> ql/src/java/org/apache/hadoop/hive/ql/exec/spark/status/LocalSparkJobMonitor.java
>  2a6c33bfd4824c96e7004cd1ecce48c62c97d685 
>   
> ql/src/java/org/apache/hadoop/hive/ql/exec/spark/status/RemoteSparkJobMonitor.java
>  004b50ba95934280cf302055a46a5d984b421e07 
>   ql/src/java/org/apache/hadoop/hive/ql/exec/spark/status/RenderStrategy.java 
> PRE-CREATION 
>   
> ql/src/java/org/apache/hadoop/hive/ql/exec/spark/status/SparkJobMonitor.java 
> 3531ac25a9959aacd5766a9a42316890c68a1cd5 
>   
> ql/src/test/org/apache/hadoop/hive/ql/exec/spark/status/TestSparkJobMonitor.java
>  e66354f0869738bd3cf0eb831c13fa6af1eda256 
>   service/src/java/org/apache/hive/service/ServiceUtils.java 
> 226e43244df10c22143b91f92ef312e56739d036 
>   
> service/src/java/org/apache/hive/service/cli/SparkProgressMonitorStatusMapper.java
>  PRE-CREATION 
>   service/src/java/org/apache/hive/service/cli/thrift/ThriftCLIService.java 
> 68fe8d8aa143fafbfc611253ce3a12065016a537 
> 
> 
> Diff: https://reviews.apache.org/r/67636/diff/1/
> 
> 
> Testing
> ---
> 
> 
> Thanks,
> 
> Bharathkrishna Guruvayoor Murali
> 
>



Review Request 67636: HIVE-19176 : Add HoS support to progress bar on Beeline client.

2018-06-18 Thread Bharathkrishna Guruvayoor Murali via Review Board

---
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/67636/
---

Review request for hive, Peter Vary, Sahil Takiar, and Vihang Karajgaonkar.


Repository: hive-git


Description
---

This logic is similar to the RenderStrategy used in Tez to print the progress 
bar.


Diffs
-

  common/src/java/org/apache/hadoop/hive/conf/HiveConf.java 
933bda4ad01a6f7878019a7b4c971a0c39068ae2 
  jdbc/src/java/org/apache/hive/jdbc/HiveStatement.java 
ad8d1a7f1cca3a763bb7c07335998ab7d39d7598 
  
ql/src/java/org/apache/hadoop/hive/ql/exec/spark/status/LocalSparkJobMonitor.java
 2a6c33bfd4824c96e7004cd1ecce48c62c97d685 
  
ql/src/java/org/apache/hadoop/hive/ql/exec/spark/status/RemoteSparkJobMonitor.java
 004b50ba95934280cf302055a46a5d984b421e07 
  ql/src/java/org/apache/hadoop/hive/ql/exec/spark/status/RenderStrategy.java 
PRE-CREATION 
  ql/src/java/org/apache/hadoop/hive/ql/exec/spark/status/SparkJobMonitor.java 
3531ac25a9959aacd5766a9a42316890c68a1cd5 
  
ql/src/test/org/apache/hadoop/hive/ql/exec/spark/status/TestSparkJobMonitor.java
 e66354f0869738bd3cf0eb831c13fa6af1eda256 
  service/src/java/org/apache/hive/service/ServiceUtils.java 
226e43244df10c22143b91f92ef312e56739d036 
  
service/src/java/org/apache/hive/service/cli/SparkProgressMonitorStatusMapper.java
 PRE-CREATION 
  service/src/java/org/apache/hive/service/cli/thrift/ThriftCLIService.java 
68fe8d8aa143fafbfc611253ce3a12065016a537 


Diff: https://reviews.apache.org/r/67636/diff/1/


Testing
---


Thanks,

Bharathkrishna Guruvayoor Murali



[jira] [Created] (HIVE-19937) Intern JobConf objects in Spark tasks

2018-06-18 Thread Sahil Takiar (JIRA)
Sahil Takiar created HIVE-19937:
---

 Summary: Intern JobConf objects in Spark tasks
 Key: HIVE-19937
 URL: https://issues.apache.org/jira/browse/HIVE-19937
 Project: Hive
  Issue Type: Improvement
  Components: Spark
Reporter: Sahil Takiar
Assignee: Sahil Takiar


When fixing HIVE-16395, we decided that each new Spark task should clone the 
{{JobConf}} object to prevent any {{ConcurrentModificationException}} from 
being thrown. However, setting this variable comes at a cost of storing a 
duplicate {{JobConf}} object for each Spark task. These objects can take up a 
significant amount of memory, we should intern them so that Spark tasks running 
in the same JVM don't store duplicate copies.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (HIVE-19936) explain on a query failing in secure cluster whereas query itself works

2018-06-18 Thread Aihua Xu (JIRA)
Aihua Xu created HIVE-19936:
---

 Summary: explain on a query failing in secure cluster whereas 
query itself works
 Key: HIVE-19936
 URL: https://issues.apache.org/jira/browse/HIVE-19936
 Project: Hive
  Issue Type: Bug
  Components: Hooks
Reporter: Aihua Xu


On a secured cluster with Sentry integrated run the following queries

{noformat}
create table foobar (id int) partitioned by (val int);
explain alter table foobar add partition (val=50);
{noformat}

The explain query will fail with the following exception while the query itself 
works with no issue.

Error while compiling statement: FAILED: SemanticException No valid 
privileges{color}
 Required privilege( Table) not available in output privileges
 The required privileges: (state=42000,code=4)



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (HIVE-19935) Hive WM session killed: Failed to update LLAP tasks count

2018-06-18 Thread Thai Bui (JIRA)
Thai Bui created HIVE-19935:
---

 Summary: Hive WM session killed: Failed to update LLAP tasks count
 Key: HIVE-19935
 URL: https://issues.apache.org/jira/browse/HIVE-19935
 Project: Hive
  Issue Type: Bug
  Components: HiveServer2
Affects Versions: 3.1.0
Reporter: Thai Bui
Assignee: Thai Bui


I'm getting this error with WM feature quite frequently. It causes AM 
containers to shut down and a new one created to replace it.
{noformat}
018-06-18T19:06:49,969 INFO [Thread-250] 
monitoring.RenderStrategy$LogToFileFunction: Map 1: 313(+270)/641
2018-06-18T19:06:49,988 INFO [NotificationEventPoll 0] metastore.HiveMetaStore: 
4: get_config_value: name=metastore.batch.retrieve.max defaultValue=50
2018-06-18T19:06:49,988 INFO [NotificationEventPoll 0] HiveMetaStore.audit: 
ugi=hive ip=unknown-ip-addr cmd=get_config_value: 
name=metastore.batch.retrieve.max defaultValue=50
2018-06-18T19:06:50,204 INFO [pool-29-thread-1] tez.TriggerValidatorRunnable: 
Query: hive_20180618190637_e65869b8-10be-4880-a8d3-84989bd055b4. Trigger { 
name: alluxio_medium, expression: ALLUXIO_BYTES_READ >
6442450944, action: MOVE TO medium } violated. Current value: 7184667126. 
Applying action.
2018-06-18T19:06:50,205 INFO [pool-29-thread-1] tez.WorkloadManager: Queued 
move session: 49be39e5-875c-4cfe-8601-7fe84dd57e0c moving from default to medium
2018-06-18T19:06:50,205 INFO [Workload management master] tez.WorkloadManager: 
Processing current events
2018-06-18T19:06:50,205 INFO [Workload management master] tez.WorkloadManager: 
Handling move session event: 49be39e5-875c-4cfe-8601-7fe84dd57e0c moving from 
default to medium
2018-06-18T19:06:50,205 INFO [Workload management master] tez.WorkloadManager: 
Subscribed to counters: [S3A_BYTES_READ, BYTES_READ, ALLUXIO_BYTES_READ]
2018-06-18T19:06:50,205 INFO [pool-29-thread-1] 
tez.KillMoveTriggerActionHandler: Moved session 
49be39e5-875c-4cfe-8601-7fe84dd57e0c to pool medium
2018-06-18T19:06:50,205 INFO [Workload management master] 
tez.GuaranteedTasksAllocator: Updating 49be39e5-875c-4cfe-8601-7fe84dd57e0c 
with 144 guaranteed tasks
2018-06-18T19:06:50,205 INFO [Workload management master] tez.WmEvent: Added 
WMEvent: EventType: MOVE EventStartTimestamp: 1529348810205 elapsedTime: 0 
wmTezSessionInfo:SessionId: 49be39e5-875c-4cfe-8601-7fe
84dd57e0c Pool: medium Cluster %: 30.0
2018-06-18T19:06:50,234 INFO [StateChangeNotificationHandler] 
impl.ZkRegistryBase$InstanceStateChangeListener: CHILD_UPDATED for zknode 
/user-hive/llap/workers/worker-001571
2018-06-18T19:06:50,235 INFO [StateChangeNotificationHandler] 
tez.TezSessionPool: AM for 49be39e5-875c-4cfe-8601-7fe84dd57e0c, v.1571 has 
updated; updating [sessionId=49be39e5-875c-4cfe-8601-7fe84dd57e0c, qu
eueName=llap, user=hive, doAs=false, isOpen=true, isDefault=true, expires in 
586277120ms, WM state poolName=medium, clusterFraction=0.3, 
queryId=hive_20180618190637_e65869b8-10be-4880-a8d3-84989bd055b4, killR
eason=null] with an endpoint at 32769
2018-06-18T19:06:50,235 INFO [StateChangeNotificationHandler] 
tez.TezSessionState: Ignoring an outdated info update 1571: TezAmInstance 
[49be39e5-875c-4cfe-8601-7fe84dd57e0c, host=ip-10-8-121-231.data.bazaar
voice.com, rpcPort=33365, pluginPort=32769, token=null]
2018-06-18T19:06:50,323 ERROR [TaskCommunicator # 4] 
tez.GuaranteedTasksAllocator: Failed to update guaranteed tasks count for the 
session sessionId=49be39e5-875c-4cfe-8601-7fe84dd57e0c, queueName=llap, user=
hive, doAs=false, isOpen=true, isDefault=true, expires in 586277032ms, WM state 
poolName=medium, clusterFraction=0.3, 
queryId=hive_20180618190637_e65869b8-10be-4880-a8d3-84989bd055b4, 
killReason=null
com.google.protobuf.ServiceException: 
org.apache.hadoop.ipc.RemoteException(java.lang.NullPointerException): 
java.lang.NullPointerException
at 
org.apache.hadoop.hive.llap.tezplugins.LlapTaskSchedulerService.checkAndSendGuaranteedStateUpdate(LlapTaskSchedulerService.java:596)
at 
org.apache.hadoop.hive.llap.tezplugins.LlapTaskSchedulerService.updateGuaranteedCount(LlapTaskSchedulerService.java:581)
at 
org.apache.hadoop.hive.llap.tezplugins.LlapTaskSchedulerService.updateQuery(LlapTaskSchedulerService.java:3041)
at 
org.apache.hadoop.hive.llap.tezplugins.endpoint.LlapPluginServerImpl.updateQuery(LlapPluginServerImpl.java:57)
at 
org.apache.hadoop.hive.llap.plugin.rpc.LlapPluginProtocolProtos$LlapPluginProtocol$2.callBlockingMethod(LlapPluginProtocolProtos.java:835)
at 
org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:523)
at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:991)
at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:872)
at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:818)
at java.security.AccessController.doPrivileged(Native Method)
at javax.security.auth.Subject.doAs(Subject.java:422)
at 

[jira] [Created] (HIVE-19934) Verify ACID table UPDATE/DELETE commands on transactional stats.

2018-06-18 Thread Steve Yeom (JIRA)
Steve Yeom created HIVE-19934:
-

 Summary: Verify ACID table UPDATE/DELETE commands on transactional 
stats.
 Key: HIVE-19934
 URL: https://issues.apache.org/jira/browse/HIVE-19934
 Project: Hive
  Issue Type: Sub-task
  Components: Transactions
Affects Versions: 4.0.0
Reporter: Steve Yeom
 Fix For: 4.0.0


ACID table UPDATE/DELETE 
1. should accurately compute table stats (if started from
accurate stats status and with hive.stats.autogather or 
hive.starts.column.autogather
being true) and keep "true" for "BASIC_STATS" of COLUMN_STATS_ACCURATE.
2. should turn off COLUMN_STATS (or turn to "false").



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (HIVE-19933) ALTER TABLE DROP PARTITION - Partition Not Found

2018-06-18 Thread BELUGA BEHR (JIRA)
BELUGA BEHR created HIVE-19933:
--

 Summary: ALTER TABLE DROP PARTITION - Partition Not Found
 Key: HIVE-19933
 URL: https://issues.apache.org/jira/browse/HIVE-19933
 Project: Hive
  Issue Type: Improvement
  Components: HiveServer2
Affects Versions: 1.2.2
Reporter: BELUGA BEHR


{code:sql}
ALTER TABLE web_logsz DROP PARTITION (`date`='xyz')
-- SemanticException [Error 10001]: Table not found web_logsz

ALTER TABLE web_logs DROP PARTITION (`date`='xyz')
-- Success.
{code}

There is no 'xyz' partition for the 'date' column.  To make this more 
consistent, the query should fail if the user tries to drop a partition that 
does not exist



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (HIVE-19932) INSERT INTO SELECT case test may be needed.

2018-06-18 Thread Steve Yeom (JIRA)
Steve Yeom created HIVE-19932:
-

 Summary: INSERT INTO SELECT case test may be needed.
 Key: HIVE-19932
 URL: https://issues.apache.org/jira/browse/HIVE-19932
 Project: Hive
  Issue Type: Sub-task
  Components: Transactions
Affects Versions: 4.0.0
Reporter: Steve Yeom
 Fix For: 4.0.0


INSERT INTO SELECT case test may be needed to make sure 
this command is executed correctly in that next SELECT COUNT/MAX/MIN works
only with stats.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


Need edit permission to Hive confluence

2018-06-18 Thread Janaki Latha Lahorani
Hi,

I would like to update documentation related to some minor enhancements.  I
would be very grateful if I have the permission to edit.

Username: janulatha

Thanks,
Janaki.


[jira] [Created] (HIVE-19931) stats_nonpart.q test run shows possibly wrong results.

2018-06-18 Thread Steve Yeom (JIRA)
Steve Yeom created HIVE-19931:
-

 Summary: stats_nonpart.q test run shows possibly wrong results.
 Key: HIVE-19931
 URL: https://issues.apache.org/jira/browse/HIVE-19931
 Project: Hive
  Issue Type: Sub-task
  Components: Transactions
Affects Versions: 4.0.0
Reporter: Steve Yeom
 Fix For: 4.0.0






--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (HIVE-19930) Resolve explain.out differences introduced by this fearture

2018-06-18 Thread Steve Yeom (JIRA)
Steve Yeom created HIVE-19930:
-

 Summary: Resolve explain.out differences introduced by this 
fearture
 Key: HIVE-19930
 URL: https://issues.apache.org/jira/browse/HIVE-19930
 Project: Hive
  Issue Type: Sub-task
  Components: Transactions
Affects Versions: 4.0.0
Reporter: Steve Yeom
 Fix For: 4.0.0






--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[GitHub] hive pull request #375: HIVE-19886: Logs may be directed to 2 files if --hiv...

2018-06-18 Thread beltran
GitHub user beltran opened a pull request:

https://github.com/apache/hive/pull/375

HIVE-19886: Logs may be directed to 2 files if --hiveconf hive.log.fi…

…le is used.

This is still dumping some logs for the metastore before the 
reconfiguration is done:
```
2018-06-18T16:39:17,109  INFO [main] conf.MetastoreConf: Found 
configuration file file:/hadoop-3.1.0/etc/hadoop/hive-site.xml
2018-06-18T16:39:17,364  INFO [main] conf.MetastoreConf: Unable to find 
config file hivemetastore-site.xml
2018-06-18T16:39:17,364  INFO [main] conf.MetastoreConf: Found 
configuration file null
2018-06-18T16:39:17,365  INFO [main] conf.MetastoreConf: Unable to find 
config file metastore-site.xml
2018-06-18T16:39:17,365  INFO [main] conf.MetastoreConf: Found 
configuration file null
```
(and that's it). Before, the loggers created before the logging 
configuration happened sent the logs to the wrong file forever.

It's hard to get rid of those because `MetastoreConf` is used to create the 
configure logging(and therefore spits logs before the configuration happens) so 
some refactoring would have to be done.

You can merge this pull request into a Git repository by running:

$ git pull https://github.com/beltran/hive HIVE-19886

Alternatively you can review and apply these changes as the patch at:

https://github.com/apache/hive/pull/375.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

This closes #375


commit a4b0fbff76ff912823e912b7f2104fbd7db88116
Author: Jaume Marhuenda 
Date:   2018-06-18T16:44:13Z

HIVE-19886: Logs may be directed to 2 files if --hiveconf hive.log.file is 
used




---


[jira] [Created] (HIVE-19929) Vectorization: Recheck for vectorization wrong results/execution failures

2018-06-18 Thread Matt McCline (JIRA)
Matt McCline created HIVE-19929:
---

 Summary: Vectorization: Recheck for vectorization wrong 
results/execution failures
 Key: HIVE-19929
 URL: https://issues.apache.org/jira/browse/HIVE-19929
 Project: Hive
  Issue Type: Bug
  Components: Hive
Reporter: Matt McCline
Assignee: Matt McCline


Use test variables hive.test.vectorized.execution.enabled.override=enable and 
hive.test.vectorization.suppress.explain.execution.mode=true to look for wrong 
results/execution failures when vectorization is forced ON and "Execution mode: 
vectorized" is suppressed.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)