[jira] [Updated] (HIVE-9896) \N un-recognized in AVRO format Hive tables

2015-03-09 Thread Madhan Sundararajan Devaki (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9896?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Madhan Sundararajan Devaki updated HIVE-9896:
-
Description: 
We Sqooped (1.4.5) data from many RDBMS into HDFS in text format with options 
--null-non-string '\\N' --null-string '\\N'.
When we load these into Hive tables in text format the \N is properly 
recognized as NULL and we are able to use SQL clauses such as IS NULL and IS 
NOT NULL against columns.
However, when we convert the text files into AVRO (1.7.6) with SNAPPY 
compression and try to query using the above SQL clauses, the query does not 
return results as expected.
Further, we have to use column_name = '\\N' or column_name  '\\N' as a 
workaround.

  was:
We Sqooped data from many RDBMS into HDFS in text format with options 
--null-non-string '\\N' --null-string '\\N'.
When we load these into Hive tables in text format the \N is properly 
recognized as NULL and we are able to use SQL clauses such as IS NULL and IS 
NOT NULL against columns.
However, when we convert the text files into AVRO (1.7.6) with SNAPPY 
compression and try to query using the above SQL clauses, the query does not 
return results as expected.
Further, we have to use column_name = '\\N' or column_name  '\\N' as a 
workaround.


 \N un-recognized in AVRO format Hive tables
 ---

 Key: HIVE-9896
 URL: https://issues.apache.org/jira/browse/HIVE-9896
 Project: Hive
  Issue Type: Bug
  Components: Database/Schema, File Formats, Hive
Affects Versions: 0.13.0
 Environment: CDH5.2.1, RHEL6.5, Java 7
Reporter: Madhan Sundararajan Devaki

 We Sqooped (1.4.5) data from many RDBMS into HDFS in text format with options 
 --null-non-string '\\N' --null-string '\\N'.
 When we load these into Hive tables in text format the \N is properly 
 recognized as NULL and we are able to use SQL clauses such as IS NULL and IS 
 NOT NULL against columns.
 However, when we convert the text files into AVRO (1.7.6) with SNAPPY 
 compression and try to query using the above SQL clauses, the query does not 
 return results as expected.
 Further, we have to use column_name = '\\N' or column_name  '\\N' as a 
 workaround.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9870) Add JvmPauseMonitor threads to HMS and HS2 daemons

2015-03-09 Thread Harsh J (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9870?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Harsh J updated HIVE-9870:
--
Attachment: HIVE-9870.patch

Sorry, missed the new try-catch block change in the previous attachment. Should 
work this time.

 Add JvmPauseMonitor threads to HMS and HS2 daemons
 --

 Key: HIVE-9870
 URL: https://issues.apache.org/jira/browse/HIVE-9870
 Project: Hive
  Issue Type: Improvement
  Components: HiveServer2, Metastore
Affects Versions: 1.1
Reporter: Harsh J
Assignee: Harsh J
Priority: Minor
 Attachments: HIVE-9870.patch, HIVE-9870.patch, HIVE-9870.patch


 The hadoop-common carries in it a nifty thread that prints GC or non-GC 
 pauses within the JVM if it exceeds a specific threshold.
 This has been immeasurably useful in supporting several clusters, in 
 identifying GC or other form of process pauses to be the root cause of some 
 event being investigated.
 The HMS and HS2 daemons are good targets for running similar threads within 
 it. It can be loaded in an if-available style.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9855) Runtime skew join doesn't work when skewed data only exists in big table

2015-03-09 Thread Rui Li (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9855?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14352955#comment-14352955
 ] 

Rui Li commented on HIVE-9855:
--

Merged into spark.

 Runtime skew join doesn't work when skewed data only exists in big table
 

 Key: HIVE-9855
 URL: https://issues.apache.org/jira/browse/HIVE-9855
 Project: Hive
  Issue Type: Bug
Reporter: Rui Li
Assignee: Rui Li
 Fix For: 1.2.0

 Attachments: HIVE-9855.1.patch


 To reproduce, enable runtime skew join and then join two tables that skewed 
 data only exists in one of them. The task will fail with the following 
 exception:
 {noformat}
 Error: java.lang.RuntimeException: Hive Runtime Error while closing 
 operators: java.io.IOException: Unable to rename output to: hdfs://..
 {noformat}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9889) Merge trunk to Spark branch 3/6/2015 [Spark Branch]

2015-03-09 Thread JIRA

[ 
https://issues.apache.org/jira/browse/HIVE-9889?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14353056#comment-14353056
 ] 

Sergio Peña commented on HIVE-9889:
---

I'll take a look.

 Merge trunk to Spark branch 3/6/2015 [Spark Branch]
 ---

 Key: HIVE-9889
 URL: https://issues.apache.org/jira/browse/HIVE-9889
 Project: Hive
  Issue Type: Task
  Components: Spark
Reporter: Xuefu Zhang
Assignee: Xuefu Zhang
 Attachments: HIVE-9889.1-spark.patch






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9894) Use new parquet Types API builder to construct DATE data type

2015-03-09 Thread JIRA

[ 
https://issues.apache.org/jira/browse/HIVE-9894?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14353055#comment-14353055
 ] 

Sergio Peña commented on HIVE-9894:
---

+1

 Use new parquet Types API builder to construct DATE data type
 -

 Key: HIVE-9894
 URL: https://issues.apache.org/jira/browse/HIVE-9894
 Project: Hive
  Issue Type: Sub-task
Reporter: Dong Chen
Assignee: Dong Chen
 Attachments: HIVE-9894.patch


 The DATE type was implemented in HIVE-8119. And new parquet Types API builder 
 was used in HIVE-9657 for all data types. But DATE is missed.
 We should also use new Types API for DATE type.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9659) 'Error while trying to create table container' occurs during hive query case execution when hive.optimize.skewjoin set to 'true' [Spark Branch]

2015-03-09 Thread Rui Li (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9659?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Rui Li updated HIVE-9659:
-
Attachment: HIVE-9659.3-spark.patch

Address RB comments

 'Error while trying to create table container' occurs during hive query case 
 execution when hive.optimize.skewjoin set to 'true' [Spark Branch]
 ---

 Key: HIVE-9659
 URL: https://issues.apache.org/jira/browse/HIVE-9659
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Reporter: Xin Hao
Assignee: Rui Li
 Attachments: HIVE-9659.1-spark.patch, HIVE-9659.2-spark.patch, 
 HIVE-9659.3-spark.patch


 We found that 'Error while trying to create table container'  occurs during 
 Big-Bench Q12 case execution when hive.optimize.skewjoin set to 'true'.
 If hive.optimize.skewjoin set to 'false', the case could pass.
 How to reproduce:
 1. set hive.optimize.skewjoin=true;
 2. Run BigBench case Q12 and it will fail. 
 Check the executor log (e.g. /usr/lib/spark/work/app-/2/stderr) and you 
 will found error 'Error while trying to create table container' in the log 
 and also a NullPointerException near the end of the log.
 (a) Detail error message for 'Error while trying to create table container':
 {noformat}
 15/02/12 01:29:49 ERROR SparkMapRecordHandler: Error processing row: 
 org.apache.hadoop.hive.ql.metadata.HiveException: 
 org.apache.hadoop.hive.ql.metadata.HiveException: Error while trying to 
 create table container
 org.apache.hadoop.hive.ql.metadata.HiveException: 
 org.apache.hadoop.hive.ql.metadata.HiveException: Error while trying to 
 create table container
   at 
 org.apache.hadoop.hive.ql.exec.spark.HashTableLoader.load(HashTableLoader.java:118)
   at 
 org.apache.hadoop.hive.ql.exec.MapJoinOperator.loadHashTable(MapJoinOperator.java:193)
   at 
 org.apache.hadoop.hive.ql.exec.MapJoinOperator.cleanUpInputFileChangedOp(MapJoinOperator.java:219)
   at 
 org.apache.hadoop.hive.ql.exec.Operator.cleanUpInputFileChanged(Operator.java:1051)
   at 
 org.apache.hadoop.hive.ql.exec.Operator.cleanUpInputFileChanged(Operator.java:1055)
   at 
 org.apache.hadoop.hive.ql.exec.Operator.cleanUpInputFileChanged(Operator.java:1055)
   at 
 org.apache.hadoop.hive.ql.exec.Operator.cleanUpInputFileChanged(Operator.java:1055)
   at 
 org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:486)
   at 
 org.apache.hadoop.hive.ql.exec.spark.SparkMapRecordHandler.processRow(SparkMapRecordHandler.java:141)
   at 
 org.apache.hadoop.hive.ql.exec.spark.HiveMapFunctionResultList.processNextRecord(HiveMapFunctionResultList.java:47)
   at 
 org.apache.hadoop.hive.ql.exec.spark.HiveMapFunctionResultList.processNextRecord(HiveMapFunctionResultList.java:27)
   at 
 org.apache.hadoop.hive.ql.exec.spark.HiveBaseFunctionResultList$ResultIterator.hasNext(HiveBaseFunctionResultList.java:98)
   at 
 scala.collection.convert.Wrappers$JIteratorWrapper.hasNext(Wrappers.scala:41)
   at 
 org.apache.spark.util.collection.ExternalSorter.insertAll(ExternalSorter.scala:217)
   at 
 org.apache.spark.shuffle.sort.SortShuffleWriter.write(SortShuffleWriter.scala:65)
   at 
 org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:68)
   at 
 org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:41)
   at org.apache.spark.scheduler.Task.run(Task.scala:56)
   at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:196)
   at 
 java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
   at 
 java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
   at java.lang.Thread.run(Thread.java:745)
 Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: Error while 
 trying to create table container
   at 
 org.apache.hadoop.hive.ql.exec.persistence.MapJoinTableContainerSerDe.load(MapJoinTableContainerSerDe.java:158)
   at 
 org.apache.hadoop.hive.ql.exec.spark.HashTableLoader.load(HashTableLoader.java:115)
   ... 21 more
 Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: Error, not a 
 directory: 
 hdfs://bhx1:8020/tmp/hive/root/d22ef465-bff5-4edb-a822-0a9f1c25b66c/hive_2015-02-12_01-28-10_008_6897031694580088767-1/-mr-10009/HashTable-Stage-6/MapJoin-mapfile01--.hashtable
   at 
 org.apache.hadoop.hive.ql.exec.persistence.MapJoinTableContainerSerDe.load(MapJoinTableContainerSerDe.java:106)
   ... 22 more
 15/02/12 01:29:49 INFO SparkRecordHandler: maximum memory = 40939028480
 15/02/12 01:29:49 INFO PerfLogger: PERFLOG method=SparkInitializeOperators 
 from=org.apache.hadoop.hive.ql.exec.spark.SparkRecordHandler
 {noformat}
 (b) Detail error message for 

[jira] [Updated] (HIVE-9601) New Beeline queries will hang If Beeline terminates in-properly [Spark Branch]

2015-03-09 Thread Jimmy Xiang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9601?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jimmy Xiang updated HIVE-9601:
--
Attachment: HIVE-9601.1-spark.patch

Re-attach the patch to trigger QA test.

 New Beeline queries will hang If Beeline terminates in-properly [Spark Branch]
 --

 Key: HIVE-9601
 URL: https://issues.apache.org/jira/browse/HIVE-9601
 Project: Hive
  Issue Type: Bug
  Components: Spark
Reporter: Szehon Ho
Assignee: Jimmy Xiang
 Fix For: spark-branch

 Attachments: HIVE-9601.1-spark.patch, HIVE-9601.1-spark.patch


 User session's Spark application seems to stay around if beeline is not quit 
 properly (!quit) because the user is not disconnected.
 If Beeline is started, it will create a new Spark application which will hang 
 waiting for the first one.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9897) Issue a warning when using an existing table/view name as an alias in a with statement.

2015-03-09 Thread Mario Konschake (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9897?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mario Konschake updated HIVE-9897:
--
Description: 
Consider the following query:

{code:sql}
WITH
table_a AS (
SELECT
'johndoe' AS name
FROM
my_table
)

SELECT
DISTINCT name
FROM
table_a;
{code}

Observation: 

If a table or a view with name `table_a` exists it is used instead of the one 
defined in the WITH statement.

Expectation:

As the expectation is ambiguous (using the alias in the WITH statement vs. 
using the existing table) issuing a warning when using a existing name in a 
WITH statement is recommended.

  was:
Consider the following query:

{code:sql}
WITH
table_a AS (
SELECT
'johndoe' AS name
FROM
my_table
)

SELECT
DISTINCT name
FROM
table_a;
{code:sql}

Observation: 

If a table or a view with name `table_a` exists it is used instead of the one 
defined in the WITH statement.

Expectation:

As the expectation is ambiguous (using the alias in the WITH statement vs. 
using the existing table) issuing a warning when using a existing name in a 
WITH statement is recommended.


 Issue a warning when using an existing table/view name as an alias in a with 
 statement. 
 

 Key: HIVE-9897
 URL: https://issues.apache.org/jira/browse/HIVE-9897
 Project: Hive
  Issue Type: Improvement
  Components: Hive
Affects Versions: 0.13.1
 Environment: cdh5.3.0
Reporter: Mario Konschake
Priority: Minor

 Consider the following query:
 {code:sql}
 WITH
 table_a AS (
 SELECT
 'johndoe' AS name
 FROM
 my_table
 )
 SELECT
 DISTINCT name
 FROM
 table_a;
 {code}
 Observation: 
 If a table or a view with name `table_a` exists it is used instead of the one 
 defined in the WITH statement.
 Expectation:
 As the expectation is ambiguous (using the alias in the WITH statement vs. 
 using the existing table) issuing a warning when using a existing name in a 
 WITH statement is recommended.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9897) Issue a warning when using an existing table/view name as an alias in a with statement.

2015-03-09 Thread Mario Konschake (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9897?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mario Konschake updated HIVE-9897:
--
Description: 
Consider the following query:

{code:sql}
WITH
table_a AS (
SELECT
'johndoe' AS name
FROM
my_table
)

SELECT
DISTINCT name
FROM
table_a;
{code:sql}

Observation: 

If a table or a view with name `table_a` exists it is used instead of the one 
defined in the WITH statement.

Expectation:

As the expectation is ambiguous (using the alias in the WITH statement vs. 
using the existing table) issuing a warning when using a existing name in a 
WITH statement is recommended.

  was:
Consider the following query:

WITH
table_a AS (
SELECT
'johndoe' AS name
FROM
my_table
)

SELECT
DISTINCT name
FROM
table_a;

Observation: 

If a table or a view with name `table_a` exists it is used instead of the one 
defined in the WITH statement.

Expectation:

As the expectation is ambiguous (using the alias in the WITH statement vs. 
using the existing table) issuing a warning when using a existing name in a 
WITH statement is recommended.


 Issue a warning when using an existing table/view name as an alias in a with 
 statement. 
 

 Key: HIVE-9897
 URL: https://issues.apache.org/jira/browse/HIVE-9897
 Project: Hive
  Issue Type: Improvement
  Components: Hive
Affects Versions: 0.13.1
 Environment: cdh5.3.0
Reporter: Mario Konschake
Priority: Minor

 Consider the following query:
 {code:sql}
 WITH
 table_a AS (
 SELECT
 'johndoe' AS name
 FROM
 my_table
 )
 SELECT
 DISTINCT name
 FROM
 table_a;
 {code:sql}
 Observation: 
 If a table or a view with name `table_a` exists it is used instead of the one 
 defined in the WITH statement.
 Expectation:
 As the expectation is ambiguous (using the alias in the WITH statement vs. 
 using the existing table) issuing a warning when using a existing name in a 
 WITH statement is recommended.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9808) LLAP: Push work into daemons instead of the current pull

2015-03-09 Thread Siddharth Seth (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9808?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Siddharth Seth updated HIVE-9808:
-
Attachment: HIVE-9808.2.txt

Rebased patch. Will commit shortly; this one was painful to rebase.
There's some UGI / closeAllForFileSystem changes which will need to be worked 
on in a follow up.

 LLAP: Push work into daemons instead of the current pull
 

 Key: HIVE-9808
 URL: https://issues.apache.org/jira/browse/HIVE-9808
 Project: Hive
  Issue Type: Sub-task
Reporter: Siddharth Seth
Assignee: Siddharth Seth
 Fix For: llap

 Attachments: HIVE-9808.1.txt, HIVE-9808.2.txt






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Resolved] (HIVE-9808) LLAP: Push work into daemons instead of the current pull

2015-03-09 Thread Siddharth Seth (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9808?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Siddharth Seth resolved HIVE-9808.
--
Resolution: Fixed

 LLAP: Push work into daemons instead of the current pull
 

 Key: HIVE-9808
 URL: https://issues.apache.org/jira/browse/HIVE-9808
 Project: Hive
  Issue Type: Sub-task
Reporter: Siddharth Seth
Assignee: Siddharth Seth
 Fix For: llap

 Attachments: HIVE-9808.1.txt, HIVE-9808.2.txt






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9897) Issue a warning when using an existing table/view name as an alias in a with statement.

2015-03-09 Thread RAUNAK JHAWAR (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9897?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14353208#comment-14353208
 ] 

RAUNAK JHAWAR commented on HIVE-9897:
-

I did a quick check on Hive on HDP:

Assumption users is an existing table in Hive (database: default)

{code:sql}
with users as (select * from products limit 10)select * from users limit 10;
{code}

The effective output is same as that of doing a {code:sql} select * from users 
limit 10{code}

The products data is not shown as the output neither any warning is issued that 
there already is an object by name users



 Issue a warning when using an existing table/view name as an alias in a with 
 statement. 
 

 Key: HIVE-9897
 URL: https://issues.apache.org/jira/browse/HIVE-9897
 Project: Hive
  Issue Type: Improvement
  Components: Hive
Affects Versions: 0.13.1
 Environment: cdh5.3.0
Reporter: Mario Konschake
Priority: Minor

 Consider the following query:
 WITH
 table_a AS (
 SELECT
 'johndoe' AS name
 FROM
 my_table
 )
 SELECT
 DISTINCT name
 FROM
 table_a;
 Observation: 
 If a table or a view with name `table_a` exists it is used instead of the one 
 defined in the WITH statement.
 Expectation:
 As the expectation is ambiguous (using the alias in the WITH statement vs. 
 using the existing table) issuing a warning when using a existing name in a 
 WITH statement is recommended.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9537) string expressions on a fixed length character do not preserve trailing spaces

2015-03-09 Thread Aihua Xu (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9537?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14353274#comment-14353274
 ] 

Aihua Xu commented on HIVE-9537:


HIVE-9899 doc bug is opened to keep track of the doc change. I'm closing this 
out. Let me know if you have other concerns. 

 string expressions on a fixed length character do not preserve trailing spaces
 --

 Key: HIVE-9537
 URL: https://issues.apache.org/jira/browse/HIVE-9537
 Project: Hive
  Issue Type: Bug
  Components: SQL
Reporter: N Campbell
Assignee: Aihua Xu

 When a string expression such as upper or lower is applied to a fixed length 
 column the trailing spaces of the fixed length character are not preserved.
 {code:sql}
 CREATE TABLE  if not exists TCHAR ( 
 RNUM int, 
 CCHAR char(32)
 )
 ROW FORMAT DELIMITED 
 FIELDS TERMINATED BY '|' 
 LINES TERMINATED BY '\n' 
 STORED AS TEXTFILE;
 {code}
 {{cchar}} as a {{char(32)}}.
 {code:sql}
 select cchar, concat(cchar, cchar), concat(lower(cchar), cchar), 
 concat(upper(cchar), cchar) 
 from tchar;
 {code}
 0|\N
 1|
 2| 
 3|BB
 4|EE
 5|FF



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9775) LLAP: Add a MiniLLAPCluster for tests

2015-03-09 Thread Siddharth Seth (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9775?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Siddharth Seth updated HIVE-9775:
-
Attachment: HIVE-9775.2.patch

Re-based patch.

 LLAP: Add a MiniLLAPCluster for tests
 -

 Key: HIVE-9775
 URL: https://issues.apache.org/jira/browse/HIVE-9775
 Project: Hive
  Issue Type: Sub-task
Reporter: Siddharth Seth
Assignee: Siddharth Seth
 Fix For: llap

 Attachments: HIVE-9775.1.patch, HIVE-9775.2.patch






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9273) Add option to fire metastore event on insert

2015-03-09 Thread Sushanth Sowmyan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9273?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sushanth Sowmyan updated HIVE-9273:
---
Issue Type: Sub-task  (was: New Feature)
Parent: HIVE-7973

 Add option to fire metastore event on insert
 

 Key: HIVE-9273
 URL: https://issues.apache.org/jira/browse/HIVE-9273
 Project: Hive
  Issue Type: Sub-task
Reporter: Alan Gates
Assignee: Alan Gates
  Labels: TODOC1.2
 Fix For: 1.2.0

 Attachments: HIVE-9273.2.patch, HIVE-9273.patch


 HIVE-9271 adds the ability for the client to request firing metastore events. 
  This can be used in the MoveTask to fire events when an insert is done that 
 does not add partitions to a table.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9550) ObjectStore.getNextNotification() can return events inside NotificationEventResponse as null which conflicts with its thrift required tag

2015-03-09 Thread Sushanth Sowmyan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9550?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sushanth Sowmyan updated HIVE-9550:
---
Issue Type: Sub-task  (was: Bug)
Parent: HIVE-7973

 ObjectStore.getNextNotification() can return events inside 
 NotificationEventResponse as null which conflicts with its thrift required 
 tag
 ---

 Key: HIVE-9550
 URL: https://issues.apache.org/jira/browse/HIVE-9550
 Project: Hive
  Issue Type: Sub-task
Reporter: Sushanth Sowmyan
Assignee: Sushanth Sowmyan
 Fix For: 1.2.0

 Attachments: HIVE-9550.patch


 Per hive_metastore.thrift, the events list inside NotificationEventResponse 
 is a required field that cannot be null.
 {code}
 struct NotificationEventResponse {
 1: required listNotificationEvent events,
 }
 {code}
 However, per ObjectStore.java, this events field can be uninitialized if the 
 events retrieved from the metastore is empty instead of null:
 {code}
   NotificationEventResponse result = new NotificationEventResponse();
   int maxEvents = rqst.getMaxEvents()  0 ? rqst.getMaxEvents() : 
 Integer.MAX_VALUE;
   int numEvents = 0; 
   while (i.hasNext()  numEvents++  maxEvents) {
 result.addToEvents(translateDbToThrift(i.next()));
   }
   return result;
 {code}
 The fix is simple enough - we need to call result.setEvents(new 
 ArrayListNotificationEvent()) before we begin the iteration to do 
 result.addToEvents(...).



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9658) Reduce parquet memory use by bypassing java primitive objects on ETypeConverter

2015-03-09 Thread JIRA

 [ 
https://issues.apache.org/jira/browse/HIVE-9658?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sergio Peña updated HIVE-9658:
--
Attachment: HIVE-9658.2.patch

 Reduce parquet memory use by bypassing java primitive objects on 
 ETypeConverter
 ---

 Key: HIVE-9658
 URL: https://issues.apache.org/jira/browse/HIVE-9658
 Project: Hive
  Issue Type: Sub-task
Reporter: Sergio Peña
Assignee: Sergio Peña
 Attachments: HIVE-9658.1.patch, HIVE-9658.2.patch


 The ETypeConverter class passes Writable objects to the collection converters 
 in order to be read later by the map/reduce functions. These objects are all 
 wrapped in a unique ArrayWritable object.
 We can save some memory by returning the java primitive objects instead in 
 order to prevent memory allocation. The only writable object needed by 
 map/reduce is ArrayWritable. If we create another writable class where to 
 store primitive objects (Object), then we can stop using all primitive 
 wirtables.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9577) Fire insert event on HCatalog appends

2015-03-09 Thread Sushanth Sowmyan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9577?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sushanth Sowmyan updated HIVE-9577:
---
Issue Type: Sub-task  (was: Bug)
Parent: HIVE-7973

 Fire insert event on HCatalog appends
 -

 Key: HIVE-9577
 URL: https://issues.apache.org/jira/browse/HIVE-9577
 Project: Hive
  Issue Type: Sub-task
  Components: HCatalog
Affects Versions: 1.2.0
Reporter: Sushanth Sowmyan
Assignee: Alan Gates

 HIVE-9271 added the ability to fire a listener event on the metastore to 
 notify on inserts, and HIVE-9273 added this ability to register data appends 
 in hive to trigger this event. HCatalog appends should also trigger this 
 event.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9601) New Beeline queries will hang If Beeline terminates in-properly [Spark Branch]

2015-03-09 Thread Jimmy Xiang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9601?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jimmy Xiang updated HIVE-9601:
--
Attachment: HIVE-9601.2-spark.patch

Attached v2 that fixed unit test failures. The patch works only for beeline 
client using binary Thrift transport.

 New Beeline queries will hang If Beeline terminates in-properly [Spark Branch]
 --

 Key: HIVE-9601
 URL: https://issues.apache.org/jira/browse/HIVE-9601
 Project: Hive
  Issue Type: Bug
  Components: Spark
Reporter: Szehon Ho
Assignee: Jimmy Xiang
 Fix For: spark-branch

 Attachments: HIVE-9601.1-spark.patch, HIVE-9601.1-spark.patch, 
 HIVE-9601.2-spark.patch


 User session's Spark application seems to stay around if beeline is not quit 
 properly (!quit) because the user is not disconnected.
 If Beeline is started, it will create a new Spark application which will hang 
 waiting for the first one.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9882) Add jar/file doesn't work with yarn-cluster mode [Spark Branch]

2015-03-09 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9882?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14353633#comment-14353633
 ] 

Hive QA commented on HIVE-9882:
---



{color:red}Overall{color}: -1 no tests executed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12703340/HIVE-9882.1-spark.patch

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-SPARK-Build/771/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-SPARK-Build/771/console
Test logs: 
http://ec2-50-18-27-0.us-west-1.compute.amazonaws.com/logs/PreCommit-HIVE-SPARK-Build-771/

Messages:
{noformat}
 This message was trimmed, see log for full details 
Reverted 
'hbase-handler/src/java/org/apache/hadoop/hive/hbase/HBaseStorageHandler.java'
Reverted 
'hbase-handler/src/java/org/apache/hadoop/hive/hbase/HBaseSerDeParameters.java'
Reverted 
'metastore/src/java/org/apache/hadoop/hive/metastore/RetryingHMSHandler.java'
Reverted 'metastore/src/java/org/apache/hadoop/hive/metastore/Warehouse.java'
Reverted 
'metastore/src/java/org/apache/hadoop/hive/metastore/IMetaStoreClient.java'
Reverted 
'metastore/src/java/org/apache/hadoop/hive/metastore/HiveMetaStoreClient.java'
Reverted 
'metastore/src/java/org/apache/hadoop/hive/metastore/MetaStoreDirectSql.java'
Reverted 
'metastore/src/java/org/apache/hadoop/hive/metastore/HiveMetaStore.java'
Reverted 'bin/hive'
Reverted 'data/files/parquet_types.txt'
Reverted 
'itests/hcatalog-unit/src/test/java/org/apache/hive/hcatalog/listener/TestDbNotificationListener.java'
Reverted 
'itests/hive-unit/src/test/java/org/apache/hive/beeline/TestBeeLineWithArgs.java'
Reverted 'LICENSE'
Reverted 
'hcatalog/core/src/main/java/org/apache/hive/hcatalog/mapreduce/FosterStorageHandler.java'
Reverted 
'hcatalog/core/src/main/java/org/apache/hive/hcatalog/mapreduce/HCatSplit.java'
Reverted 
'hcatalog/streaming/src/java/org/apache/hive/hcatalog/streaming/AbstractRecordWriter.java'
Reverted 
'hcatalog/server-extensions/src/main/java/org/apache/hive/hcatalog/messaging/MessageFactory.java'
Reverted 
'hcatalog/webhcat/svr/src/main/java/org/apache/hive/hcatalog/templeton/tool/JobState.java'
Reverted 
'hcatalog/webhcat/java-client/src/main/java/org/apache/hive/hcatalog/api/HCatClientHMSImpl.java'
Reverted 'hcatalog/webhcat/java-client/pom.xml'
Reverted 
'accumulo-handler/src/java/org/apache/hadoop/hive/accumulo/predicate/PrimitiveComparisonFilter.java'
Reverted 'accumulo-handler/src/java/org/apache/hadoop/hive/accumulo/Utils.java'
Reverted 
'accumulo-handler/src/java/org/apache/hadoop/hive/accumulo/serde/AccumuloSerDeParameters.java'
Reverted 'common/src/java/org/apache/hadoop/hive/conf/HiveConf.java'
Reverted 'common/src/java/org/apache/hadoop/hive/common/JavaUtils.java'
Reverted 
'service/src/java/org/apache/hive/service/server/ThreadFactoryWithGarbageCleanup.java'
Reverted 
'serde/src/test/org/apache/hadoop/hive/serde2/avro/TestAvroSerializer.java'
Reverted 'serde/src/java/org/apache/hadoop/hive/serde2/avro/AvroSerializer.java'
Reverted 'beeline/src/test/org/apache/hive/beeline/TestBeelineArgParsing.java'
Reverted 'beeline/src/java/org/apache/hive/beeline/BeeLine.java'
Reverted 
'beeline/src/java/org/apache/hive/beeline/DeprecatedSeparatedValuesOutputFormat.java'
Reverted 'beeline/src/java/org/apache/hive/beeline/DatabaseConnection.java'
Reverted 'beeline/src/java/org/apache/hive/beeline/Commands.java'
Reverted 'beeline/src/java/org/apache/hive/beeline/ClassNameCompleter.java'
Reverted 
'beeline/src/java/org/apache/hive/beeline/SeparatedValuesOutputFormat.java'
Reverted 'beeline/src/main/resources/BeeLine.properties'
Reverted 'ql/src/test/results/clientnegative/udf_add_months_error_2.q.out'
Reverted 'ql/src/test/results/clientnegative/udf_next_day_error_1.q.out'
Reverted 'ql/src/test/results/clientnegative/parquet_date.q.out'
Reverted 'ql/src/test/results/clientnegative/udf_next_day_error_2.q.out'
Reverted 'ql/src/test/results/clientnegative/udf_last_day_error_1.q.out'
Reverted 'ql/src/test/results/clientnegative/udf_last_day_error_2.q.out'
Reverted 'ql/src/test/results/clientnegative/invalid_arithmetic_type.q.out'
Reverted 'ql/src/test/results/clientnegative/udf_add_months_error_1.q.out'
Reverted 'ql/src/test/results/clientpositive/auto_join27.q.out'
Reverted 'ql/src/test/results/clientpositive/auto_join13.q.out'
Reverted 'ql/src/test/results/clientpositive/auto_join4.q.out'
Reverted 'ql/src/test/results/clientpositive/ql_rewrite_gbtoidx_cbo_1.q.out'
Reverted 'ql/src/test/results/clientpositive/ppd_join_filter.q.out'
Reverted 'ql/src/test/results/clientpositive/groupby_sort_skew_1_23.q.out'
Reverted 'ql/src/test/results/clientpositive/subquery_in_having.q.out'
Reverted 
'ql/src/test/results/clientpositive/tez/dynamic_partition_pruning.q.out'
Reverted 

[jira] [Updated] (HIVE-4256) JDBC2 HiveConnection does not use the specified database

2015-03-09 Thread Nick White (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-4256?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Nick White updated HIVE-4256:
-
Affects Version/s: 0.12.0

 JDBC2 HiveConnection does not use the specified database
 

 Key: HIVE-4256
 URL: https://issues.apache.org/jira/browse/HIVE-4256
 Project: Hive
  Issue Type: Bug
  Components: HiveServer2, JDBC
Affects Versions: 0.11.0, 0.12.0
Reporter: Chris Drome
Assignee: Anandha L Ranganathan
 Fix For: 0.13.0

 Attachments: HIVE-4256.1.patch, HIVE-4256.2.patch, HIVE-4256.3.patch, 
 HIVE-4256.4.patch, HIVE-4256.5.patch, HIVE-4256.6.patch, HIVE-4256.patch


 HiveConnection ignores the database specified in the connection string when 
 configuring the connection.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-6617) Reduce ambiguity in grammar

2015-03-09 Thread Pengcheng Xiong (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6617?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Pengcheng Xiong updated HIVE-6617:
--
Attachment: HIVE-6617.24.patch

 Reduce ambiguity in grammar
 ---

 Key: HIVE-6617
 URL: https://issues.apache.org/jira/browse/HIVE-6617
 Project: Hive
  Issue Type: Task
Reporter: Ashutosh Chauhan
Assignee: Pengcheng Xiong
 Attachments: HIVE-6617.01.patch, HIVE-6617.02.patch, 
 HIVE-6617.03.patch, HIVE-6617.04.patch, HIVE-6617.05.patch, 
 HIVE-6617.06.patch, HIVE-6617.07.patch, HIVE-6617.08.patch, 
 HIVE-6617.09.patch, HIVE-6617.10.patch, HIVE-6617.11.patch, 
 HIVE-6617.12.patch, HIVE-6617.13.patch, HIVE-6617.14.patch, 
 HIVE-6617.15.patch, HIVE-6617.16.patch, HIVE-6617.17.patch, 
 HIVE-6617.18.patch, HIVE-6617.19.patch, HIVE-6617.20.patch, 
 HIVE-6617.21.patch, HIVE-6617.22.patch, HIVE-6617.23.patch, 
 HIVE-6617.24.patch, parser.png


 CLEAR LIBRARY CACHE
 As of today, antlr reports 214 warnings. Need to bring down this number, 
 ideally to 0.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9271) Add ability for client to request metastore to fire an event

2015-03-09 Thread Sushanth Sowmyan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9271?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sushanth Sowmyan updated HIVE-9271:
---
Issue Type: Sub-task  (was: New Feature)
Parent: HIVE-7973

 Add ability for client to request metastore to fire an event
 

 Key: HIVE-9271
 URL: https://issues.apache.org/jira/browse/HIVE-9271
 Project: Hive
  Issue Type: Sub-task
  Components: Metastore
Reporter: Alan Gates
Assignee: Alan Gates
 Fix For: 1.2.0

 Attachments: HIVE-9271.patch


 Currently all events in Hive are fired by the metastore.  However, there are 
 events that only the client fully understands, such as DML operations.  There 
 should be a way for the client to request the metastore to fire a particular 
 event.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9175) Add alters to list of events handled by NotificationListener

2015-03-09 Thread Sushanth Sowmyan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9175?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sushanth Sowmyan updated HIVE-9175:
---
Issue Type: Sub-task  (was: New Feature)
Parent: HIVE-7973

 Add alters to list of events handled by NotificationListener
 

 Key: HIVE-9175
 URL: https://issues.apache.org/jira/browse/HIVE-9175
 Project: Hive
  Issue Type: Sub-task
  Components: HCatalog
Reporter: Alan Gates
Assignee: Alan Gates
  Labels: TODOC15
 Fix For: 1.1.0

 Attachments: HIVE-9175.patch


 HCatalog currently doesn't implement onAlterTable and onAlterPartition.  It 
 should.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9658) Reduce parquet memory use by bypassing java primitive objects on ETypeConverter

2015-03-09 Thread JIRA

[ 
https://issues.apache.org/jira/browse/HIVE-9658?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14353576#comment-14353576
 ] 

Sergio Peña commented on HIVE-9658:
---

Latest changes on 'trunk' need to be merged on 'parquet' branch so that this 
patch can be tested.

 Reduce parquet memory use by bypassing java primitive objects on 
 ETypeConverter
 ---

 Key: HIVE-9658
 URL: https://issues.apache.org/jira/browse/HIVE-9658
 Project: Hive
  Issue Type: Sub-task
Reporter: Sergio Peña
Assignee: Sergio Peña
 Attachments: HIVE-9658.1.patch, HIVE-9658.2.patch


 The ETypeConverter class passes Writable objects to the collection converters 
 in order to be read later by the map/reduce functions. These objects are all 
 wrapped in a unique ArrayWritable object.
 We can save some memory by returning the java primitive objects instead in 
 order to prevent memory allocation. The only writable object needed by 
 map/reduce is ArrayWritable. If we create another writable class where to 
 store primitive objects (Object), then we can stop using all primitive 
 wirtables.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9889) Merge trunk to Spark branch 3/6/2015 [Spark Branch]

2015-03-09 Thread Xuefu Zhang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9889?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Xuefu Zhang updated HIVE-9889:
--
Attachment: HIVE-9889.2-spark.patch

Regenerate the patch since some patches were merged individually.

 Merge trunk to Spark branch 3/6/2015 [Spark Branch]
 ---

 Key: HIVE-9889
 URL: https://issues.apache.org/jira/browse/HIVE-9889
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Reporter: Xuefu Zhang
Assignee: Xuefu Zhang
 Attachments: HIVE-9889.1-spark.patch, HIVE-9889.2-spark.patch






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-1161) Hive Replication

2015-03-09 Thread SHAILESH PILARE (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-1161?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14353377#comment-14353377
 ] 

SHAILESH PILARE commented on HIVE-1161:
---

Ahutosh,
Currently i am working on best feasible approach, out of Two available 
approaches
a) Use  Hive Import/Export along with Distcp with some background auditing in 
default DB .
b) Using Distcp for data and  Database replication for metastore  transfer .
After successful completion of POC  i would update the design doc .


 Hive Replication
 

 Key: HIVE-1161
 URL: https://issues.apache.org/jira/browse/HIVE-1161
 Project: Hive
  Issue Type: New Feature
  Components: Contrib
Reporter: Edward Capriolo
Assignee: SHAILESH PILARE
Priority: Minor

 Users may want to replicate data between two distinct hadoop clusters or two 
 hive warehouses on the same cluster.
 Users may want to replicate entire catalogs or possibly on a table by table 
 basis. Should this process be batch driven or a be a full time running 
 application? What are some practical requirements, what are the limitations?
 Comments?



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9902) Map join small table files need more replications [Spark Branch]

2015-03-09 Thread Jimmy Xiang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9902?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jimmy Xiang updated HIVE-9902:
--
Attachment: HIVE-9902.1-spark.patch

 Map join small table files need more replications [Spark Branch]
 

 Key: HIVE-9902
 URL: https://issues.apache.org/jira/browse/HIVE-9902
 Project: Hive
  Issue Type: Bug
Reporter: Jimmy Xiang
Assignee: Jimmy Xiang
 Fix For: spark-branch

 Attachments: HIVE-9902.1-spark.patch


 We have
 {noformat}
 replication = (short) Math.min(MIN_REPLICATION, numOfPartitions);
 {noformat}
 It should be
 {noformat}
 replication = (short) Math.max(MIN_REPLICATION, numOfPartitions);
 {noformat}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9664) Hive add jar command should be able to download and add jars from a repository

2015-03-09 Thread Carl Steinbach (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9664?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Carl Steinbach updated HIVE-9664:
-
Assignee: Anant Nag

 Hive add jar command should be able to download and add jars from a 
 repository
 

 Key: HIVE-9664
 URL: https://issues.apache.org/jira/browse/HIVE-9664
 Project: Hive
  Issue Type: Improvement
Affects Versions: 0.14.0
Reporter: Anant Nag
Assignee: Anant Nag
  Labels: hive, patch
 Attachments: HIVE-9664.patch


 Currently Hive's add jar command takes a local path to the dependency jar. 
 This clutters the local file-system as users may forget to remove this jar 
 later
 It would be nice if Hive supported a Gradle like notation to download the jar 
 from a repository.
 Example:  add jar org:module:version
 
 It should also be backward compatible and should take jar from the local 
 file-system as well. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9851) org.apache.hadoop.hive.serde2.avro.AvroSerializer should use org.apache.avro.generic.GenericData.Array when serializing a list

2015-03-09 Thread Carl Steinbach (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9851?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Carl Steinbach updated HIVE-9851:
-
Fix Version/s: 1.2.0

 org.apache.hadoop.hive.serde2.avro.AvroSerializer should use 
 org.apache.avro.generic.GenericData.Array when serializing a list
 --

 Key: HIVE-9851
 URL: https://issues.apache.org/jira/browse/HIVE-9851
 Project: Hive
  Issue Type: Bug
  Components: Hive, Serializers/Deserializers
Reporter: Ratandeep Ratti
 Fix For: 1.2.0

 Attachments: HIVE-9851.patch


 Currently AvroSerializer uses java.util.ArrayList for serializing a list in 
 Hive.
 This causes problems when we need to convert the avro object into some other 
 representation say a tuple in Pig.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9851) org.apache.hadoop.hive.serde2.avro.AvroSerializer should use org.apache.avro.generic.GenericData.Array when serializing a list

2015-03-09 Thread Carl Steinbach (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9851?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Carl Steinbach updated HIVE-9851:
-
Assignee: Ratandeep Ratti

 org.apache.hadoop.hive.serde2.avro.AvroSerializer should use 
 org.apache.avro.generic.GenericData.Array when serializing a list
 --

 Key: HIVE-9851
 URL: https://issues.apache.org/jira/browse/HIVE-9851
 Project: Hive
  Issue Type: Bug
  Components: Hive, Serializers/Deserializers
Reporter: Ratandeep Ratti
Assignee: Ratandeep Ratti
 Fix For: 1.2.0

 Attachments: HIVE-9851.patch


 Currently AvroSerializer uses java.util.ArrayList for serializing a list in 
 Hive.
 This causes problems when we need to convert the avro object into some other 
 representation say a tuple in Pig.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9871) Print spark job id in history file [spark branch]

2015-03-09 Thread Xuefu Zhang (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9871?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14353468#comment-14353468
 ] 

Xuefu Zhang commented on HIVE-9871:
---

[~chinnalalam], thanks for working on this. Patch looks good, but I'm wondering 
if you can come up with a better name for the private method added. Something 
like recordJobId() or addToHistory(), etc.

 Print spark job id in history file [spark branch]
 -

 Key: HIVE-9871
 URL: https://issues.apache.org/jira/browse/HIVE-9871
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Reporter: Chinna Rao Lalam
Assignee: Chinna Rao Lalam
 Attachments: HIVE-9871.1-spark.patch


 Maintain the spark job id in history file for the corresponding queries.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9658) Reduce parquet memory use by bypassing java primitive objects on ETypeConverter

2015-03-09 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9658?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14353565#comment-14353565
 ] 

Hive QA commented on HIVE-9658:
---



{color:red}Overall{color}: -1 no tests executed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12703476/HIVE-9658.2.patch

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/2979/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/2979/console
Test logs: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-2979/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Tests exited with: NonZeroExitCodeException
Command 'bash /data/hive-ptest/working/scratch/source-prep.sh' failed with exit 
status 1 and output '+ [[ -n /usr/java/jdk1.7.0_45-cloudera ]]
+ export JAVA_HOME=/usr/java/jdk1.7.0_45-cloudera
+ JAVA_HOME=/usr/java/jdk1.7.0_45-cloudera
+ export 
PATH=/usr/java/jdk1.7.0_45-cloudera/bin/:/usr/local/apache-maven-3.0.5/bin:/usr/java/jdk1.7.0_45-cloudera/bin:/usr/local/apache-ant-1.9.1/bin:/usr/local/bin:/bin:/usr/bin:/usr/local/sbin:/usr/sbin:/sbin:/home/hiveptest/bin
+ 
PATH=/usr/java/jdk1.7.0_45-cloudera/bin/:/usr/local/apache-maven-3.0.5/bin:/usr/java/jdk1.7.0_45-cloudera/bin:/usr/local/apache-ant-1.9.1/bin:/usr/local/bin:/bin:/usr/bin:/usr/local/sbin:/usr/sbin:/sbin:/home/hiveptest/bin
+ export 'ANT_OPTS=-Xmx1g -XX:MaxPermSize=256m '
+ ANT_OPTS='-Xmx1g -XX:MaxPermSize=256m '
+ export 'M2_OPTS=-Xmx1g -XX:MaxPermSize=256m -Dhttp.proxyHost=localhost 
-Dhttp.proxyPort=3128'
+ M2_OPTS='-Xmx1g -XX:MaxPermSize=256m -Dhttp.proxyHost=localhost 
-Dhttp.proxyPort=3128'
+ cd /data/hive-ptest/working/
+ tee /data/hive-ptest/logs/PreCommit-HIVE-TRUNK-Build-2979/source-prep.txt
+ [[ false == \t\r\u\e ]]
+ mkdir -p maven ivy
+ [[ svn = \s\v\n ]]
+ [[ -n '' ]]
+ [[ -d apache-svn-trunk-source ]]
+ [[ ! -d apache-svn-trunk-source/.svn ]]
+ [[ ! -d apache-svn-trunk-source ]]
+ cd apache-svn-trunk-source
+ svn revert -R .
Reverted 
'metastore/src/java/org/apache/hadoop/hive/metastore/HiveMetaStore.java'
Reverted 'service/src/java/org/apache/hive/service/server/HiveServer2.java'
++ awk '{print $2}'
++ egrep -v '^X|^Performing status on external'
++ svn status --no-ignore
+ rm -rf target datanucleus.log ant/target shims/target shims/0.20S/target 
shims/0.23/target shims/aggregator/target shims/common/target 
shims/scheduler/target packaging/target hbase-handler/target testutils/target 
jdbc/target metastore/target 
metastore/src/java/org/apache/hadoop/hive/metastore/HiveMetaStore.java.orig 
itests/target itests/thirdparty itests/hcatalog-unit/target 
itests/test-serde/target itests/qtest/target itests/hive-unit-hadoop2/target 
itests/hive-minikdc/target itests/hive-jmh/target itests/hive-unit/target 
itests/custom-serde/target itests/util/target itests/qtest-spark/target 
hcatalog/target hcatalog/core/target hcatalog/streaming/target 
hcatalog/server-extensions/target hcatalog/webhcat/svr/target 
hcatalog/webhcat/java-client/target hcatalog/hcatalog-pig-adapter/target 
accumulo-handler/target hwi/target common/target common/src/gen 
spark-client/target service/target contrib/target serde/target beeline/target 
odbc/target cli/target ql/dependency-reduced-pom.xml ql/target
+ svn update

Fetching external item into 'hcatalog/src/test/e2e/harness'
External at revision 1665350.

At revision 1665350.
+ patchCommandPath=/data/hive-ptest/working/scratch/smart-apply-patch.sh
+ patchFilePath=/data/hive-ptest/working/scratch/build.patch
+ [[ -f /data/hive-ptest/working/scratch/build.patch ]]
+ chmod +x /data/hive-ptest/working/scratch/smart-apply-patch.sh
+ /data/hive-ptest/working/scratch/smart-apply-patch.sh 
/data/hive-ptest/working/scratch/build.patch
The patch does not appear to apply with p0, p1, or p2
+ exit 1
'
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12703476 - PreCommit-HIVE-TRUNK-Build

 Reduce parquet memory use by bypassing java primitive objects on 
 ETypeConverter
 ---

 Key: HIVE-9658
 URL: https://issues.apache.org/jira/browse/HIVE-9658
 Project: Hive
  Issue Type: Sub-task
Reporter: Sergio Peña
Assignee: Sergio Peña
 Attachments: HIVE-9658.1.patch, HIVE-9658.2.patch


 The ETypeConverter class passes Writable objects to the collection converters 
 in order to be read later by the map/reduce functions. These objects are all 
 wrapped in a unique ArrayWritable object.
 We can save some memory by returning the java primitive objects instead in 
 order to prevent memory allocation. The only writable object needed by 
 map/reduce is ArrayWritable. If we create another writable 

[jira] [Updated] (HIVE-9892) various MSSQL upgrade scripts don't work

2015-03-09 Thread Sergey Shelukhin (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9892?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sergey Shelukhin updated HIVE-9892:
---
Attachment: HIVE-9892.ADDENDUM.patch

[~thejas] can you take a look at addendum? The pre-create script now conflicts 
with some other part of upgrade script

 various MSSQL upgrade scripts don't work
 

 Key: HIVE-9892
 URL: https://issues.apache.org/jira/browse/HIVE-9892
 Project: Hive
  Issue Type: Bug
  Components: Metastore
Affects Versions: 0.13.0, 0.14.0, 1.0.0
Reporter: Sergey Shelukhin
Assignee: Sergey Shelukhin
 Fix For: 1.1

 Attachments: HIVE-9892.ADDENDUM.patch, HIVE-9892.patch


  * Issue with GO statement when run through schematool - it results in syntax 
 error.
  * the create if not exists logic for PART_COL_STATS wasn't working
 NO PRECOMMIT TESTS



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9903) Update calcite version

2015-03-09 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9903?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-9903:
---
Attachment: HIVE-9903.patch

 Update calcite version
 --

 Key: HIVE-9903
 URL: https://issues.apache.org/jira/browse/HIVE-9903
 Project: Hive
  Issue Type: Task
  Components: CBO, Logical Optimizer
Reporter: Ashutosh Chauhan
Assignee: Ashutosh Chauhan
 Attachments: HIVE-9903.patch


 Calcite-1.1



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9852) virtual column “INPUT__FILE__NAME” does not exclude files before map

2015-03-09 Thread Nickolay Ponomarev (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9852?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14353912#comment-14353912
 ] 

Nickolay Ponomarev commented on HIVE-9852:
--

This appears to be a duplicate of HIVE-1662.

 virtual  column   “INPUT__FILE__NAME” does not exclude files before map 
 

 Key: HIVE-9852
 URL: https://issues.apache.org/jira/browse/HIVE-9852
 Project: Hive
  Issue Type: Bug
  Components: Build Infrastructure
Affects Versions: 0.13.1
Reporter: Zhang Jingpeng

 virtual  column   “INPUT__FILE__NAME” does not exclude files before map .
 If I have three files I only want it to deal one file of three files, Then 
 the map count will be 3 and will read all three files , so It's only tell 
 what's file the line in and then filter it but not filter before map input 
 path.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9569) Enable more unit tests for UNION ALL [Spark Branch]

2015-03-09 Thread Chao (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9569?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chao updated HIVE-9569:
---
Attachment: HIVE-9569.5.patch

Forgot about this one... Attaching a new patch.

 Enable more unit tests for UNION ALL [Spark Branch]
 ---

 Key: HIVE-9569
 URL: https://issues.apache.org/jira/browse/HIVE-9569
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Affects Versions: spark-branch
Reporter: Chao
Assignee: Chao
 Attachments: HIVE-9569.1-spark.patch, HIVE-9569.1.patch, 
 HIVE-9569.2.patch, HIVE-9569.3.patch, HIVE-9569.4.patch, HIVE-9569.5.patch


 Currently, we only enabled a subset of all the union tests. We should try to 
 enable the rest, and see if there's any issue.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-6617) Reduce ambiguity in grammar

2015-03-09 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6617?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14353868#comment-14353868
 ] 

Hive QA commented on HIVE-6617:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12703503/HIVE-6617.24.patch

{color:red}ERROR:{color} -1 due to 2 failed/errored test(s), 7761 tests executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_innerjoin
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_innerjoin
{noformat}

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/2980/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/2980/console
Test logs: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-2980/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 2 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12703503 - PreCommit-HIVE-TRUNK-Build

 Reduce ambiguity in grammar
 ---

 Key: HIVE-6617
 URL: https://issues.apache.org/jira/browse/HIVE-6617
 Project: Hive
  Issue Type: Task
Reporter: Ashutosh Chauhan
Assignee: Pengcheng Xiong
 Attachments: HIVE-6617.01.patch, HIVE-6617.02.patch, 
 HIVE-6617.03.patch, HIVE-6617.04.patch, HIVE-6617.05.patch, 
 HIVE-6617.06.patch, HIVE-6617.07.patch, HIVE-6617.08.patch, 
 HIVE-6617.09.patch, HIVE-6617.10.patch, HIVE-6617.11.patch, 
 HIVE-6617.12.patch, HIVE-6617.13.patch, HIVE-6617.14.patch, 
 HIVE-6617.15.patch, HIVE-6617.16.patch, HIVE-6617.17.patch, 
 HIVE-6617.18.patch, HIVE-6617.19.patch, HIVE-6617.20.patch, 
 HIVE-6617.21.patch, HIVE-6617.22.patch, HIVE-6617.23.patch, 
 HIVE-6617.24.patch, parser.png


 CLEAR LIBRARY CACHE
 As of today, antlr reports 214 warnings. Need to bring down this number, 
 ideally to 0.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9582) HCatalog should use IMetaStoreClient interface

2015-03-09 Thread Thiruvel Thirumoolan (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9582?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14353949#comment-14353949
 ] 

Thiruvel Thirumoolan commented on HIVE-9582:


Thanks Sushanth. Thejas looked at the patch and gave a first level review. I 
have addressed his comments and uploaded the patch.

 HCatalog should use IMetaStoreClient interface
 --

 Key: HIVE-9582
 URL: https://issues.apache.org/jira/browse/HIVE-9582
 Project: Hive
  Issue Type: Sub-task
  Components: HCatalog, Metastore
Affects Versions: 0.14.0, 0.13.1
Reporter: Thiruvel Thirumoolan
Assignee: Thiruvel Thirumoolan
  Labels: hcatalog, metastore, rolling_upgrade
 Fix For: 0.14.1

 Attachments: HIVE-9582.1.patch, HIVE-9582.2.patch, HIVE-9582.3.patch, 
 HIVE-9582.4.patch, HIVE-9583.1.patch


 Hive uses IMetaStoreClient and it makes using RetryingMetaStoreClient easy. 
 Hence during a failure, the client retries and possibly succeeds. But 
 HCatalog has long been using HiveMetaStoreClient directly and hence failures 
 are costly, especially if they are during the commit stage of a job. Its also 
 not possible to do rolling upgrade of MetaStore Server.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9678) create timediff UDF

2015-03-09 Thread Alexander Pivovarov (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9678?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alexander Pivovarov updated HIVE-9678:
--
Attachment: HIVE-9678.4.patch

patch #4 again

 create timediff UDF
 ---

 Key: HIVE-9678
 URL: https://issues.apache.org/jira/browse/HIVE-9678
 Project: Hive
  Issue Type: Improvement
  Components: UDF
Reporter: Alexander Pivovarov
Assignee: Alexander Pivovarov
 Attachments: HIVE-9678.1.patch, HIVE-9678.2.patch, HIVE-9678.3.patch, 
 HIVE-9678.4.patch, HIVE-9678.4.patch


 MySQL has very useful function timediff. We should have it in Hive
 {code}
 select timediff('2015-02-12 05:09:07.140', '2015-02-12 01:18:20');
 OK
 03:50:47.140
 {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-6617) Reduce ambiguity in grammar

2015-03-09 Thread Pengcheng Xiong (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6617?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Pengcheng Xiong updated HIVE-6617:
--
Attachment: HIVE-6617.25.patch

update golden file for keyword inner

 Reduce ambiguity in grammar
 ---

 Key: HIVE-6617
 URL: https://issues.apache.org/jira/browse/HIVE-6617
 Project: Hive
  Issue Type: Task
Reporter: Ashutosh Chauhan
Assignee: Pengcheng Xiong
 Attachments: HIVE-6617.01.patch, HIVE-6617.02.patch, 
 HIVE-6617.03.patch, HIVE-6617.04.patch, HIVE-6617.05.patch, 
 HIVE-6617.06.patch, HIVE-6617.07.patch, HIVE-6617.08.patch, 
 HIVE-6617.09.patch, HIVE-6617.10.patch, HIVE-6617.11.patch, 
 HIVE-6617.12.patch, HIVE-6617.13.patch, HIVE-6617.14.patch, 
 HIVE-6617.15.patch, HIVE-6617.16.patch, HIVE-6617.17.patch, 
 HIVE-6617.18.patch, HIVE-6617.19.patch, HIVE-6617.20.patch, 
 HIVE-6617.21.patch, HIVE-6617.22.patch, HIVE-6617.23.patch, 
 HIVE-6617.24.patch, HIVE-6617.25.patch, parser.png


 CLEAR LIBRARY CACHE
 As of today, antlr reports 214 warnings. Need to bring down this number, 
 ideally to 0.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9858) Create cbrt (cube root) UDF

2015-03-09 Thread Alexander Pivovarov (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9858?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alexander Pivovarov updated HIVE-9858:
--
Attachment: HIVE-9858.1.patch

patch #1 again

 Create cbrt (cube root) UDF
 ---

 Key: HIVE-9858
 URL: https://issues.apache.org/jira/browse/HIVE-9858
 Project: Hive
  Issue Type: Improvement
  Components: UDF
Reporter: Alexander Pivovarov
Assignee: Alexander Pivovarov
 Attachments: HIVE-9858.1.patch, HIVE-9858.1.patch


 returns the cube root of a double value
 cbrt(double a) : double
 For example:
 {code}
 select cbrt(87860583272930481.0);
 OK
 444561.0
 {code}
 I noticed that Math.pow(a, 1.0/3.0) and hive power UDF return 
 444560.965 for the example above.
 However Math.cbrt returns 444561.0
 This is why we should have hive cbrt function in hive



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9902) Map join small table files need more replications [Spark Branch]

2015-03-09 Thread Chao (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9902?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14353873#comment-14353873
 ] 

Chao commented on HIVE-9902:


+1

 Map join small table files need more replications [Spark Branch]
 

 Key: HIVE-9902
 URL: https://issues.apache.org/jira/browse/HIVE-9902
 Project: Hive
  Issue Type: Bug
Reporter: Jimmy Xiang
Assignee: Jimmy Xiang
 Fix For: spark-branch

 Attachments: HIVE-9902.1-spark.patch


 We have
 {noformat}
 replication = (short) Math.min(MIN_REPLICATION, numOfPartitions);
 {noformat}
 It should be
 {noformat}
 replication = (short) Math.max(MIN_REPLICATION, numOfPartitions);
 {noformat}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9892) various MSSQL upgrade scripts don't work

2015-03-09 Thread Thejas M Nair (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9892?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14353900#comment-14353900
 ] 

Thejas M Nair commented on HIVE-9892:
-

+1

 various MSSQL upgrade scripts don't work
 

 Key: HIVE-9892
 URL: https://issues.apache.org/jira/browse/HIVE-9892
 Project: Hive
  Issue Type: Bug
  Components: Metastore
Affects Versions: 0.13.0, 0.14.0, 1.0.0
Reporter: Sergey Shelukhin
Assignee: Sergey Shelukhin
 Fix For: 1.2.0

 Attachments: HIVE-9892.ADDENDUM.patch, HIVE-9892.patch


  * Issue with GO statement when run through schematool - it results in syntax 
 error.
  * the create if not exists logic for PART_COL_STATS wasn't working
 NO PRECOMMIT TESTS



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9659) 'Error while trying to create table container' occurs during hive query case execution when hive.optimize.skewjoin set to 'true' [Spark Branch]

2015-03-09 Thread Xuefu Zhang (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9659?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14354331#comment-14354331
 ] 

Xuefu Zhang commented on HIVE-9659:
---

[~ruili], let's create a JIRA for MR and move on. We enable the test only for 
Spark.

 'Error while trying to create table container' occurs during hive query case 
 execution when hive.optimize.skewjoin set to 'true' [Spark Branch]
 ---

 Key: HIVE-9659
 URL: https://issues.apache.org/jira/browse/HIVE-9659
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Reporter: Xin Hao
Assignee: Rui Li
 Attachments: HIVE-9659.1-spark.patch, HIVE-9659.2-spark.patch, 
 HIVE-9659.3-spark.patch


 We found that 'Error while trying to create table container'  occurs during 
 Big-Bench Q12 case execution when hive.optimize.skewjoin set to 'true'.
 If hive.optimize.skewjoin set to 'false', the case could pass.
 How to reproduce:
 1. set hive.optimize.skewjoin=true;
 2. Run BigBench case Q12 and it will fail. 
 Check the executor log (e.g. /usr/lib/spark/work/app-/2/stderr) and you 
 will found error 'Error while trying to create table container' in the log 
 and also a NullPointerException near the end of the log.
 (a) Detail error message for 'Error while trying to create table container':
 {noformat}
 15/02/12 01:29:49 ERROR SparkMapRecordHandler: Error processing row: 
 org.apache.hadoop.hive.ql.metadata.HiveException: 
 org.apache.hadoop.hive.ql.metadata.HiveException: Error while trying to 
 create table container
 org.apache.hadoop.hive.ql.metadata.HiveException: 
 org.apache.hadoop.hive.ql.metadata.HiveException: Error while trying to 
 create table container
   at 
 org.apache.hadoop.hive.ql.exec.spark.HashTableLoader.load(HashTableLoader.java:118)
   at 
 org.apache.hadoop.hive.ql.exec.MapJoinOperator.loadHashTable(MapJoinOperator.java:193)
   at 
 org.apache.hadoop.hive.ql.exec.MapJoinOperator.cleanUpInputFileChangedOp(MapJoinOperator.java:219)
   at 
 org.apache.hadoop.hive.ql.exec.Operator.cleanUpInputFileChanged(Operator.java:1051)
   at 
 org.apache.hadoop.hive.ql.exec.Operator.cleanUpInputFileChanged(Operator.java:1055)
   at 
 org.apache.hadoop.hive.ql.exec.Operator.cleanUpInputFileChanged(Operator.java:1055)
   at 
 org.apache.hadoop.hive.ql.exec.Operator.cleanUpInputFileChanged(Operator.java:1055)
   at 
 org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:486)
   at 
 org.apache.hadoop.hive.ql.exec.spark.SparkMapRecordHandler.processRow(SparkMapRecordHandler.java:141)
   at 
 org.apache.hadoop.hive.ql.exec.spark.HiveMapFunctionResultList.processNextRecord(HiveMapFunctionResultList.java:47)
   at 
 org.apache.hadoop.hive.ql.exec.spark.HiveMapFunctionResultList.processNextRecord(HiveMapFunctionResultList.java:27)
   at 
 org.apache.hadoop.hive.ql.exec.spark.HiveBaseFunctionResultList$ResultIterator.hasNext(HiveBaseFunctionResultList.java:98)
   at 
 scala.collection.convert.Wrappers$JIteratorWrapper.hasNext(Wrappers.scala:41)
   at 
 org.apache.spark.util.collection.ExternalSorter.insertAll(ExternalSorter.scala:217)
   at 
 org.apache.spark.shuffle.sort.SortShuffleWriter.write(SortShuffleWriter.scala:65)
   at 
 org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:68)
   at 
 org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:41)
   at org.apache.spark.scheduler.Task.run(Task.scala:56)
   at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:196)
   at 
 java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
   at 
 java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
   at java.lang.Thread.run(Thread.java:745)
 Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: Error while 
 trying to create table container
   at 
 org.apache.hadoop.hive.ql.exec.persistence.MapJoinTableContainerSerDe.load(MapJoinTableContainerSerDe.java:158)
   at 
 org.apache.hadoop.hive.ql.exec.spark.HashTableLoader.load(HashTableLoader.java:115)
   ... 21 more
 Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: Error, not a 
 directory: 
 hdfs://bhx1:8020/tmp/hive/root/d22ef465-bff5-4edb-a822-0a9f1c25b66c/hive_2015-02-12_01-28-10_008_6897031694580088767-1/-mr-10009/HashTable-Stage-6/MapJoin-mapfile01--.hashtable
   at 
 org.apache.hadoop.hive.ql.exec.persistence.MapJoinTableContainerSerDe.load(MapJoinTableContainerSerDe.java:106)
   ... 22 more
 15/02/12 01:29:49 INFO SparkRecordHandler: maximum memory = 40939028480
 15/02/12 01:29:49 INFO PerfLogger: PERFLOG method=SparkInitializeOperators 
 

[jira] [Updated] (HIVE-9904) Perform stats aggregation in HBase co-processor [hbase-metastore branch]

2015-03-09 Thread Vaibhav Gumashta (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9904?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Vaibhav Gumashta updated HIVE-9904:
---
Component/s: Metastore

 Perform stats aggregation in HBase co-processor [hbase-metastore branch] 
 -

 Key: HIVE-9904
 URL: https://issues.apache.org/jira/browse/HIVE-9904
 Project: Hive
  Issue Type: Sub-task
  Components: Metastore
Reporter: Vaibhav Gumashta
Assignee: Vaibhav Gumashta





--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-6617) Reduce ambiguity in grammar

2015-03-09 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6617?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14354387#comment-14354387
 ] 

Hive QA commented on HIVE-6617:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12703548/HIVE-6617.25.patch

{color:red}ERROR:{color} -1 due to 1 failed/errored test(s), 7762 tests executed
*Failed tests:*
{noformat}
org.apache.hive.hcatalog.streaming.TestStreaming.testTransactionBatchCommit_Json
{noformat}

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/2985/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/2985/console
Test logs: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-2985/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 1 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12703548 - PreCommit-HIVE-TRUNK-Build

 Reduce ambiguity in grammar
 ---

 Key: HIVE-6617
 URL: https://issues.apache.org/jira/browse/HIVE-6617
 Project: Hive
  Issue Type: Task
Reporter: Ashutosh Chauhan
Assignee: Pengcheng Xiong
 Attachments: HIVE-6617.01.patch, HIVE-6617.02.patch, 
 HIVE-6617.03.patch, HIVE-6617.04.patch, HIVE-6617.05.patch, 
 HIVE-6617.06.patch, HIVE-6617.07.patch, HIVE-6617.08.patch, 
 HIVE-6617.09.patch, HIVE-6617.10.patch, HIVE-6617.11.patch, 
 HIVE-6617.12.patch, HIVE-6617.13.patch, HIVE-6617.14.patch, 
 HIVE-6617.15.patch, HIVE-6617.16.patch, HIVE-6617.17.patch, 
 HIVE-6617.18.patch, HIVE-6617.19.patch, HIVE-6617.20.patch, 
 HIVE-6617.21.patch, HIVE-6617.22.patch, HIVE-6617.23.patch, 
 HIVE-6617.24.patch, HIVE-6617.25.patch, parser.png


 CLEAR LIBRARY CACHE
 As of today, antlr reports 214 warnings. Need to bring down this number, 
 ideally to 0.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9569) Enable more unit tests for UNION ALL [Spark Branch]

2015-03-09 Thread Xuefu Zhang (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9569?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14354339#comment-14354339
 ] 

Xuefu Zhang commented on HIVE-9569:
---

+1

 Enable more unit tests for UNION ALL [Spark Branch]
 ---

 Key: HIVE-9569
 URL: https://issues.apache.org/jira/browse/HIVE-9569
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Affects Versions: spark-branch
Reporter: Chao
Assignee: Chao
 Attachments: HIVE-9569.1-spark.patch, HIVE-9569.1.patch, 
 HIVE-9569.2.patch, HIVE-9569.3.patch, HIVE-9569.4.patch, HIVE-9569.5.patch


 Currently, we only enabled a subset of all the union tests. We should try to 
 enable the rest, and see if there's any issue.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9882) Add jar/file doesn't work with yarn-cluster mode [Spark Branch]

2015-03-09 Thread Xiaomin Zhang (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9882?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14354246#comment-14354246
 ] 

Xiaomin Zhang commented on HIVE-9882:
-

It now passed the test, good job!

 Add jar/file doesn't work with yarn-cluster mode [Spark Branch]
 ---

 Key: HIVE-9882
 URL: https://issues.apache.org/jira/browse/HIVE-9882
 Project: Hive
  Issue Type: Sub-task
  Components: Hive, spark-branch
Affects Versions: spark-branch
Reporter: Xiaomin Zhang
Assignee: Rui Li
 Fix For: spark-branch

 Attachments: HIVE-9882.1-spark.patch, HIVE-9882.1-spark.patch


 It seems current fix for HIVE-9425 only uploads the Jar/Files to HDFS, 
 however, they are not accessible by the Driver/Executor.
 I found below in the AM log:
 {noformat}
 15/02/26 15:10:36 INFO Configuration.deprecation: mapred.min.split.size is 
 deprecated. Instead, use mapreduce.input.fileinputformat.split.minsize
 15/02/26 15:10:36 INFO client.SparkClientUtilities: Added 
 jar[file:/data/hadoop-devel/data/nm/usercache/user/appcache/application_1424933948132_0002/container_1424933948132_0002_01_01/hdfs:/localhost:8020/tmp/hive/user/47040bca-1da4-49b6-b2c7-69be9bc92855/hive-exec-1.2.0-SNAPSHOT.jar]
  to classpath.
 15/02/26 15:10:36 INFO client.SparkClientUtilities: Added 
 jar[file:/data/hadoop-devel/data/nm/usercache/user/appcache/application_1424933948132_0002/container_1424933948132_0002_01_01/hdfs:/localhost:8020/tmp/hive/user/47040bca-1da4-49b6-b2c7-69be9bc92855/opennlp-maxent-3.0.3.jar]
  to classpath.
 15/02/26 15:10:36 INFO client.SparkClientUtilities: Added 
 jar[file:/data/hadoop-devel/data/nm/usercache/user/appcache/application_1424933948132_0002/container_1424933948132_0002_01_01/hdfs:/localhost:8020/tmp/hive/user/47040bca-1da4-49b6-b2c7-69be9bc92855/bigbenchqueriesmr.jar]
  to classpath.
 15/02/26 15:10:36 INFO client.SparkClientUtilities: Added 
 jar[file:/data/hadoop-devel/data/nm/usercache/user/appcache/application_1424933948132_0002/container_1424933948132_0002_01_01/hdfs:/localhost:8020/tmp/hive/user/47040bca-1da4-49b6-b2c7-69be9bc92855/opennlp-tools-1.5.3.jar]
  to classpath.
 15/02/26 15:10:36 INFO client.SparkClientUtilities: Added 
 jar[file:/data/hadoop-devel/data/nm/usercache/user/appcache/application_1424933948132_0002/container_1424933948132_0002_01_01/hdfs:/localhost:8020/tmp/hive/user/47040bca-1da4-49b6-b2c7-69be9bc92855/jcl-over-slf4j-1.7.5.jar]
  to classpath.
 15/02/26 15:10:36 INFO client.RemoteDriver: Failed to run job 
 6886df05-f430-456c-a0ff-c7621db712d6
 org.apache.hive.com.esotericsoftware.kryo.KryoException: Unable to find 
 class: de.bankmark.bigbench.queries.q10.SentimentUDF 
 {noformat}
 As above shows, the file path which was attempted to add to Classpath is 
 invalid, so actually all uploaded Jars/Files are still not available for 
 Driver/Executor.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9659) 'Error while trying to create table container' occurs during hive query case execution when hive.optimize.skewjoin set to 'true' [Spark Branch]

2015-03-09 Thread Rui Li (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9659?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14354321#comment-14354321
 ] 

Rui Li commented on HIVE-9659:
--

I tried to add golden file for MR for the added test. However MR doesn't 
produce a correct result for the new test (I used trunk to test MR). I think 
there's some issue with MR when map join and skew join are on. If I turn off 
skew join, MR produces the same result as spark does here.
[~xuefuz] - what's your opinion about this?

 'Error while trying to create table container' occurs during hive query case 
 execution when hive.optimize.skewjoin set to 'true' [Spark Branch]
 ---

 Key: HIVE-9659
 URL: https://issues.apache.org/jira/browse/HIVE-9659
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Reporter: Xin Hao
Assignee: Rui Li
 Attachments: HIVE-9659.1-spark.patch, HIVE-9659.2-spark.patch, 
 HIVE-9659.3-spark.patch


 We found that 'Error while trying to create table container'  occurs during 
 Big-Bench Q12 case execution when hive.optimize.skewjoin set to 'true'.
 If hive.optimize.skewjoin set to 'false', the case could pass.
 How to reproduce:
 1. set hive.optimize.skewjoin=true;
 2. Run BigBench case Q12 and it will fail. 
 Check the executor log (e.g. /usr/lib/spark/work/app-/2/stderr) and you 
 will found error 'Error while trying to create table container' in the log 
 and also a NullPointerException near the end of the log.
 (a) Detail error message for 'Error while trying to create table container':
 {noformat}
 15/02/12 01:29:49 ERROR SparkMapRecordHandler: Error processing row: 
 org.apache.hadoop.hive.ql.metadata.HiveException: 
 org.apache.hadoop.hive.ql.metadata.HiveException: Error while trying to 
 create table container
 org.apache.hadoop.hive.ql.metadata.HiveException: 
 org.apache.hadoop.hive.ql.metadata.HiveException: Error while trying to 
 create table container
   at 
 org.apache.hadoop.hive.ql.exec.spark.HashTableLoader.load(HashTableLoader.java:118)
   at 
 org.apache.hadoop.hive.ql.exec.MapJoinOperator.loadHashTable(MapJoinOperator.java:193)
   at 
 org.apache.hadoop.hive.ql.exec.MapJoinOperator.cleanUpInputFileChangedOp(MapJoinOperator.java:219)
   at 
 org.apache.hadoop.hive.ql.exec.Operator.cleanUpInputFileChanged(Operator.java:1051)
   at 
 org.apache.hadoop.hive.ql.exec.Operator.cleanUpInputFileChanged(Operator.java:1055)
   at 
 org.apache.hadoop.hive.ql.exec.Operator.cleanUpInputFileChanged(Operator.java:1055)
   at 
 org.apache.hadoop.hive.ql.exec.Operator.cleanUpInputFileChanged(Operator.java:1055)
   at 
 org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:486)
   at 
 org.apache.hadoop.hive.ql.exec.spark.SparkMapRecordHandler.processRow(SparkMapRecordHandler.java:141)
   at 
 org.apache.hadoop.hive.ql.exec.spark.HiveMapFunctionResultList.processNextRecord(HiveMapFunctionResultList.java:47)
   at 
 org.apache.hadoop.hive.ql.exec.spark.HiveMapFunctionResultList.processNextRecord(HiveMapFunctionResultList.java:27)
   at 
 org.apache.hadoop.hive.ql.exec.spark.HiveBaseFunctionResultList$ResultIterator.hasNext(HiveBaseFunctionResultList.java:98)
   at 
 scala.collection.convert.Wrappers$JIteratorWrapper.hasNext(Wrappers.scala:41)
   at 
 org.apache.spark.util.collection.ExternalSorter.insertAll(ExternalSorter.scala:217)
   at 
 org.apache.spark.shuffle.sort.SortShuffleWriter.write(SortShuffleWriter.scala:65)
   at 
 org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:68)
   at 
 org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:41)
   at org.apache.spark.scheduler.Task.run(Task.scala:56)
   at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:196)
   at 
 java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
   at 
 java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
   at java.lang.Thread.run(Thread.java:745)
 Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: Error while 
 trying to create table container
   at 
 org.apache.hadoop.hive.ql.exec.persistence.MapJoinTableContainerSerDe.load(MapJoinTableContainerSerDe.java:158)
   at 
 org.apache.hadoop.hive.ql.exec.spark.HashTableLoader.load(HashTableLoader.java:115)
   ... 21 more
 Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: Error, not a 
 directory: 
 hdfs://bhx1:8020/tmp/hive/root/d22ef465-bff5-4edb-a822-0a9f1c25b66c/hive_2015-02-12_01-28-10_008_6897031694580088767-1/-mr-10009/HashTable-Stage-6/MapJoin-mapfile01--.hashtable
   at 
 

[jira] [Commented] (HIVE-9659) 'Error while trying to create table container' occurs during hive query case execution when hive.optimize.skewjoin set to 'true' [Spark Branch]

2015-03-09 Thread Rui Li (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9659?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14354342#comment-14354342
 ] 

Rui Li commented on HIVE-9659:
--

Is there a way to enable the test only for spark? Seems I add the test and MR 
picks it up automatically.

 'Error while trying to create table container' occurs during hive query case 
 execution when hive.optimize.skewjoin set to 'true' [Spark Branch]
 ---

 Key: HIVE-9659
 URL: https://issues.apache.org/jira/browse/HIVE-9659
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Reporter: Xin Hao
Assignee: Rui Li
 Attachments: HIVE-9659.1-spark.patch, HIVE-9659.2-spark.patch, 
 HIVE-9659.3-spark.patch


 We found that 'Error while trying to create table container'  occurs during 
 Big-Bench Q12 case execution when hive.optimize.skewjoin set to 'true'.
 If hive.optimize.skewjoin set to 'false', the case could pass.
 How to reproduce:
 1. set hive.optimize.skewjoin=true;
 2. Run BigBench case Q12 and it will fail. 
 Check the executor log (e.g. /usr/lib/spark/work/app-/2/stderr) and you 
 will found error 'Error while trying to create table container' in the log 
 and also a NullPointerException near the end of the log.
 (a) Detail error message for 'Error while trying to create table container':
 {noformat}
 15/02/12 01:29:49 ERROR SparkMapRecordHandler: Error processing row: 
 org.apache.hadoop.hive.ql.metadata.HiveException: 
 org.apache.hadoop.hive.ql.metadata.HiveException: Error while trying to 
 create table container
 org.apache.hadoop.hive.ql.metadata.HiveException: 
 org.apache.hadoop.hive.ql.metadata.HiveException: Error while trying to 
 create table container
   at 
 org.apache.hadoop.hive.ql.exec.spark.HashTableLoader.load(HashTableLoader.java:118)
   at 
 org.apache.hadoop.hive.ql.exec.MapJoinOperator.loadHashTable(MapJoinOperator.java:193)
   at 
 org.apache.hadoop.hive.ql.exec.MapJoinOperator.cleanUpInputFileChangedOp(MapJoinOperator.java:219)
   at 
 org.apache.hadoop.hive.ql.exec.Operator.cleanUpInputFileChanged(Operator.java:1051)
   at 
 org.apache.hadoop.hive.ql.exec.Operator.cleanUpInputFileChanged(Operator.java:1055)
   at 
 org.apache.hadoop.hive.ql.exec.Operator.cleanUpInputFileChanged(Operator.java:1055)
   at 
 org.apache.hadoop.hive.ql.exec.Operator.cleanUpInputFileChanged(Operator.java:1055)
   at 
 org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:486)
   at 
 org.apache.hadoop.hive.ql.exec.spark.SparkMapRecordHandler.processRow(SparkMapRecordHandler.java:141)
   at 
 org.apache.hadoop.hive.ql.exec.spark.HiveMapFunctionResultList.processNextRecord(HiveMapFunctionResultList.java:47)
   at 
 org.apache.hadoop.hive.ql.exec.spark.HiveMapFunctionResultList.processNextRecord(HiveMapFunctionResultList.java:27)
   at 
 org.apache.hadoop.hive.ql.exec.spark.HiveBaseFunctionResultList$ResultIterator.hasNext(HiveBaseFunctionResultList.java:98)
   at 
 scala.collection.convert.Wrappers$JIteratorWrapper.hasNext(Wrappers.scala:41)
   at 
 org.apache.spark.util.collection.ExternalSorter.insertAll(ExternalSorter.scala:217)
   at 
 org.apache.spark.shuffle.sort.SortShuffleWriter.write(SortShuffleWriter.scala:65)
   at 
 org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:68)
   at 
 org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:41)
   at org.apache.spark.scheduler.Task.run(Task.scala:56)
   at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:196)
   at 
 java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
   at 
 java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
   at java.lang.Thread.run(Thread.java:745)
 Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: Error while 
 trying to create table container
   at 
 org.apache.hadoop.hive.ql.exec.persistence.MapJoinTableContainerSerDe.load(MapJoinTableContainerSerDe.java:158)
   at 
 org.apache.hadoop.hive.ql.exec.spark.HashTableLoader.load(HashTableLoader.java:115)
   ... 21 more
 Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: Error, not a 
 directory: 
 hdfs://bhx1:8020/tmp/hive/root/d22ef465-bff5-4edb-a822-0a9f1c25b66c/hive_2015-02-12_01-28-10_008_6897031694580088767-1/-mr-10009/HashTable-Stage-6/MapJoin-mapfile01--.hashtable
   at 
 org.apache.hadoop.hive.ql.exec.persistence.MapJoinTableContainerSerDe.load(MapJoinTableContainerSerDe.java:106)
   ... 22 more
 15/02/12 01:29:49 INFO SparkRecordHandler: maximum memory = 40939028480
 15/02/12 01:29:49 INFO PerfLogger: PERFLOG method=SparkInitializeOperators 
 

[jira] [Commented] (HIVE-9601) New Beeline queries will hang If Beeline terminates in-properly [Spark Branch]

2015-03-09 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9601?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14354020#comment-14354020
 ] 

Hive QA commented on HIVE-9601:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12703494/HIVE-9601.2-spark.patch

{color:red}ERROR:{color} -1 due to 2 failed/errored test(s), 7567 tests executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.ql.exec.spark.session.TestSparkSessionManagerImpl.testMultiSessionMultipleUse
org.apache.hadoop.hive.ql.exec.spark.session.TestSparkSessionManagerImpl.testSingleSessionMultipleUse
{noformat}

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-SPARK-Build/775/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-SPARK-Build/775/console
Test logs: 
http://ec2-50-18-27-0.us-west-1.compute.amazonaws.com/logs/PreCommit-HIVE-SPARK-Build-775/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 2 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12703494 - PreCommit-HIVE-SPARK-Build

 New Beeline queries will hang If Beeline terminates in-properly [Spark Branch]
 --

 Key: HIVE-9601
 URL: https://issues.apache.org/jira/browse/HIVE-9601
 Project: Hive
  Issue Type: Bug
  Components: Spark
Reporter: Szehon Ho
Assignee: Jimmy Xiang
 Fix For: spark-branch

 Attachments: HIVE-9601.1-spark.patch, HIVE-9601.1-spark.patch, 
 HIVE-9601.2-spark.patch


 User session's Spark application seems to stay around if beeline is not quit 
 properly (!quit) because the user is not disconnected.
 If Beeline is started, it will create a new Spark application which will hang 
 waiting for the first one.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9858) Create cbrt (cube root) UDF

2015-03-09 Thread Alexander Pivovarov (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9858?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alexander Pivovarov updated HIVE-9858:
--
Attachment: HIVE-9858.2.patch

patch #2 - fixed show_functions.q.out

 Create cbrt (cube root) UDF
 ---

 Key: HIVE-9858
 URL: https://issues.apache.org/jira/browse/HIVE-9858
 Project: Hive
  Issue Type: Improvement
  Components: UDF
Reporter: Alexander Pivovarov
Assignee: Alexander Pivovarov
 Attachments: HIVE-9858.1.patch, HIVE-9858.1.patch, HIVE-9858.2.patch


 returns the cube root of a double value
 cbrt(double a) : double
 For example:
 {code}
 select cbrt(87860583272930481.0);
 OK
 444561.0
 {code}
 I noticed that Math.pow(a, 1.0/3.0) and hive power UDF return 
 444560.965 for the example above.
 However Math.cbrt returns 444561.0
 This is why we should have hive cbrt function in hive



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9569) Enable more unit tests for UNION ALL [Spark Branch]

2015-03-09 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9569?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14354209#comment-14354209
 ] 

Hive QA commented on HIVE-9569:
---



{color:green}Overall{color}: +1 all checks pass

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12703545/HIVE-9569.5.patch

{color:green}SUCCESS:{color} +1 7635 tests passed

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/2983/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/2983/console
Test logs: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-2983/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12703545 - PreCommit-HIVE-TRUNK-Build

 Enable more unit tests for UNION ALL [Spark Branch]
 ---

 Key: HIVE-9569
 URL: https://issues.apache.org/jira/browse/HIVE-9569
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Affects Versions: spark-branch
Reporter: Chao
Assignee: Chao
 Attachments: HIVE-9569.1-spark.patch, HIVE-9569.1.patch, 
 HIVE-9569.2.patch, HIVE-9569.3.patch, HIVE-9569.4.patch, HIVE-9569.5.patch


 Currently, we only enabled a subset of all the union tests. We should try to 
 enable the rest, and see if there's any issue.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9788) Make double quote optional in tsv/csv/dsv output

2015-03-09 Thread Lefty Leverenz (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9788?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Lefty Leverenz updated HIVE-9788:
-
Labels: TODOC1.2  (was: )

 Make double quote optional in tsv/csv/dsv output
 

 Key: HIVE-9788
 URL: https://issues.apache.org/jira/browse/HIVE-9788
 Project: Hive
  Issue Type: Improvement
Reporter: Brock Noland
Assignee: Ferdinand Xu
  Labels: TODOC1.2
 Fix For: 1.2.0

 Attachments: HIVE-9788.1.patch, HIVE-9788.2.patch, HIVE-9788.patch


 Similar to HIVE-7390 some customers would like the double quotes to be 
 optional. So if the data is {{A}} then the output from beeline should be 
 {{A}} which is the same as the Hive CLI.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9788) Make double quote optional in tsv/csv/dsv output

2015-03-09 Thread Lefty Leverenz (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9788?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14352690#comment-14352690
 ] 

Lefty Leverenz commented on HIVE-9788:
--

Doc note:  The release note needs to be updated now that disableQuotingForSV is 
a system property.

(I don't know how users set the system properties, but the release note was 
written for patch 2 which had --disableQuotingForSV in BeeLine.properties and 
it's gone from patch 3, which has disable.quoting.for.sv in 
SeparatedValuesOutputFormat.java instead.)

 Make double quote optional in tsv/csv/dsv output
 

 Key: HIVE-9788
 URL: https://issues.apache.org/jira/browse/HIVE-9788
 Project: Hive
  Issue Type: Improvement
Reporter: Brock Noland
Assignee: Ferdinand Xu
  Labels: TODOC1.2
 Fix For: 1.2.0

 Attachments: HIVE-9788.1.patch, HIVE-9788.2.patch, HIVE-9788.patch


 Similar to HIVE-7390 some customers would like the double quotes to be 
 optional. So if the data is {{A}} then the output from beeline should be 
 {{A}} which is the same as the Hive CLI.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9588) Reimplement HCatClientHMSImpl.dropPartitions() with HMSC.dropPartitions()

2015-03-09 Thread Lefty Leverenz (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9588?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Lefty Leverenz updated HIVE-9588:
-
Labels: TODOC1.2  (was: )

 Reimplement HCatClientHMSImpl.dropPartitions() with HMSC.dropPartitions()
 -

 Key: HIVE-9588
 URL: https://issues.apache.org/jira/browse/HIVE-9588
 Project: Hive
  Issue Type: Bug
  Components: HCatalog, Metastore, Thrift API
Affects Versions: 0.14.0
Reporter: Mithun Radhakrishnan
Assignee: Mithun Radhakrishnan
  Labels: TODOC1.2
 Fix For: 1.2.0

 Attachments: HIVE-9588.1.patch, HIVE-9588.2.patch, HIVE-9588.3.patch, 
 HIVE-9588.4.patch


 {{HCatClientHMSImpl.dropPartitions()}} currently has an embarrassingly 
 inefficient implementation. The partial partition-spec is converted into a 
 filter-string. The partitions are fetched from the server, and then dropped 
 one by one.
 Here's a reimplementation that uses the {{ExprNode}}-based 
 {{HiveMetaStoreClient.dropPartitions()}}. It cuts out the excessive 
 back-and-forth between the HMS and the client-side. It also reduces the 
 memory footprint (from loading all the partitions that are to be dropped). 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-7100) Users of hive should be able to specify skipTrash when dropping tables.

2015-03-09 Thread Lefty Leverenz (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-7100?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14352730#comment-14352730
 ] 

Lefty Leverenz commented on HIVE-7100:
--

Doc note:  The documentation has been revised (please review).

* [LanguageManual DDL -- Drop Table | 
https://cwiki.apache.org/confluence/display/Hive/LanguageManual+DDL#LanguageManualDDL-DropTable]

 Users of hive should be able to specify skipTrash when dropping tables.
 ---

 Key: HIVE-7100
 URL: https://issues.apache.org/jira/browse/HIVE-7100
 Project: Hive
  Issue Type: Improvement
  Components: Query Processor
Affects Versions: 0.13.0
Reporter: Ravi Prakash
Assignee: david serafini
 Fix For: 0.14.0

 Attachments: HIVE-7100.1.patch, HIVE-7100.10.patch, 
 HIVE-7100.11.patch, HIVE-7100.2.patch, HIVE-7100.3.patch, HIVE-7100.4.patch, 
 HIVE-7100.5.patch, HIVE-7100.8.patch, HIVE-7100.9.patch, HIVE-7100.patch


 Users of our clusters are often running up against their quota limits because 
 of Hive tables. When they drop tables, they have to then manually delete the 
 files from HDFS using skipTrash. This is cumbersome and unnecessary. We 
 should enable users to skipTrash directly when dropping tables.
 We should also be able to provide this functionality without polluting SQL 
 syntax.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9118) Support auto-purge for tables, when dropping tables/partitions.

2015-03-09 Thread Lefty Leverenz (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9118?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14352734#comment-14352734
 ] 

Lefty Leverenz commented on HIVE-9118:
--

Doc note:  This is documented in the TBLPROPERTIES bullet item after the CREATE 
TABLE syntax in the DDL doc (please review).

* [LanguageManual DDL -- Create Table | 
https://cwiki.apache.org/confluence/display/Hive/LanguageManual+DDL#LanguageManualDDL-CreateTable]

 Support auto-purge for tables, when dropping tables/partitions.
 ---

 Key: HIVE-9118
 URL: https://issues.apache.org/jira/browse/HIVE-9118
 Project: Hive
  Issue Type: Bug
  Components: Metastore
Affects Versions: 1.0.0, 1.1
Reporter: Mithun Radhakrishnan
Assignee: Mithun Radhakrishnan
 Fix For: 1.2.0

 Attachments: HIVE-9118.1.patch, HIVE-9118.2.patch, HIVE-9118.3.patch


 HIVE-7100 introduced a way to skip the trash directory, when deleting 
 table-data, while dropping tables.
 In HIVE-9083/HIVE-9086, I extended this to work when partitions are dropped.
 Here, I propose a table-parameter ({{auto.purge}}) to set up tables to 
 skip-trash when table/partition data is deleted, without needing to say 
 PURGE on the Hive CLI. Apropos, on {{dropTable()}} and {{dropPartition()}}, 
 table data is deleted directly (and not moved to trash) if the following hold 
 true:
 # The table is MANAGED.
 # The {{deleteData}} parameter to the {{HMSC.drop*()}} methods is true.
 # Either PURGE is explicitly specified on the command-line (or rather, 
 {{ifPurge}} is set in the environment context, OR
 # TBLPROPERTIES contains {{auto.purge=true}}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9086) Add language support to PURGE data while dropping partitions.

2015-03-09 Thread Lefty Leverenz (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9086?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14352746#comment-14352746
 ] 

Lefty Leverenz commented on HIVE-9086:
--

Doc note:  The documentation has been revised (please review).

* [LanguageManual DDL -- Drop Partitions | 
https://cwiki.apache.org/confluence/display/Hive/LanguageManual+DDL#LanguageManualDDL-DropPartitions]

 Add language support to PURGE data while dropping partitions.
 -

 Key: HIVE-9086
 URL: https://issues.apache.org/jira/browse/HIVE-9086
 Project: Hive
  Issue Type: Bug
  Components: Metastore
Affects Versions: 0.15.0
Reporter: Mithun Radhakrishnan
Assignee: Mithun Radhakrishnan
 Fix For: 1.2.0

 Attachments: HIVE-9086.1.patch


 HIVE-9083 adds metastore-support to skip-trash while dropping partitions. 
 This patch includes language support to do the same.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)