[jira] [Resolved] (HIVE-10357) LLAP: Name IO elevator threads

2015-04-16 Thread Prasanth Jayachandran (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-10357?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Prasanth Jayachandran resolved HIVE-10357.
--
Resolution: Fixed

Committed to llap branch.

 LLAP: Name IO elevator threads
 --

 Key: HIVE-10357
 URL: https://issues.apache.org/jira/browse/HIVE-10357
 Project: Hive
  Issue Type: Sub-task
Affects Versions: llap
Reporter: Prasanth Jayachandran
Assignee: Prasanth Jayachandran
 Attachments: HIVE-10357.patch


 LLAP IO elevator thread pool thread names should be given a non-default name.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-10357) LLAP: Name IO elevator threads

2015-04-16 Thread Prasanth Jayachandran (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-10357?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Prasanth Jayachandran updated HIVE-10357:
-
Attachment: HIVE-10357.patch

 LLAP: Name IO elevator threads
 --

 Key: HIVE-10357
 URL: https://issues.apache.org/jira/browse/HIVE-10357
 Project: Hive
  Issue Type: Sub-task
Affects Versions: llap
Reporter: Prasanth Jayachandran
Assignee: Prasanth Jayachandran
 Attachments: HIVE-10357.patch


 LLAP IO elevator thread pool thread names should be given a non-default name.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-5672) Insert with custom separator not supported for non-local directory

2015-04-16 Thread Nemon Lou (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-5672?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Nemon Lou updated HIVE-5672:

Attachment: HIVE-5672.3.patch

 Insert with custom separator not supported for non-local directory
 --

 Key: HIVE-5672
 URL: https://issues.apache.org/jira/browse/HIVE-5672
 Project: Hive
  Issue Type: Bug
Affects Versions: 0.12.0, 1.0.0
Reporter: Romain Rigaux
Assignee: Nemon Lou
 Attachments: HIVE-5672.1.patch, HIVE-5672.2.patch, HIVE-5672.3.patch


 https://issues.apache.org/jira/browse/HIVE-3682 is great but non local 
 directory don't seem to be supported:
 {code}
 insert overwrite directory '/tmp/test-02'
 row format delimited
 FIELDS TERMINATED BY ':'
 select description FROM sample_07
 {code}
 {code}
 Error while compiling statement: FAILED: ParseException line 2:0 cannot 
 recognize input near 'row' 'format' 'delimited' in select clause
 {code}
 This works (with 'local'):
 {code}
 insert overwrite local directory '/tmp/test-02'
 row format delimited
 FIELDS TERMINATED BY ':'
 select code, description FROM sample_07
 {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-10358) LLAP: Tez in-place UI gets misaligned when state is TERMINATING

2015-04-16 Thread Prasanth Jayachandran (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-10358?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Prasanth Jayachandran updated HIVE-10358:
-
Attachment: HIVE-10358.patch

 LLAP: Tez in-place UI gets misaligned when state is TERMINATING
 ---

 Key: HIVE-10358
 URL: https://issues.apache.org/jira/browse/HIVE-10358
 Project: Hive
  Issue Type: Sub-task
Affects Versions: llap
Reporter: Prasanth Jayachandran
Assignee: Prasanth Jayachandran
 Attachments: HIVE-10358.patch


 In-place UI goes misaligned for TERMINATING state as length of TERMINATING 
 string is greater than header.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-7797) upgrade hive schema from 0.9.0 to 0.13.1 failed when back end db is postgres

2015-04-16 Thread Nemon Lou (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-7797?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Nemon Lou updated HIVE-7797:

Description: 
Using  the following command to upgrade hive schema :
schematool -dbType postgres -upgradeSchemaFrom 0.9.0

then got this:
ERROR: null value in column SCHEMA_VERSION violates not-null constraint

Log shows that the upgrade sql file 014-HIVE-3764.postgres.sql failed.
The sql in it is :
INSERT INTO VERSION (VER_ID, SCHEMA_VERSION, VERSION_COMMENT) VALUES 
(1, '', 'Initial value');
And the result is:
ERROR: null value in column SCHEMA_VERSION violates not-null constraint
DETAIL: Failing row contains (1, null, Initial value).

  was:
Using  the following command to upgrade hive schema :
schematool -dbType postgres -upgradeSchemaFrom 0.9.0

ERROR: null value in column SCHEMA_VERSION violates not-null constraint

Log shows that the upgrade sql file 014-HIVE-3764.postgres.sql failed.
The sql in it is :
INSERT INTO VERSION (VER_ID, SCHEMA_VERSION, VERSION_COMMENT) VALUES 
(1, '', 'Initial value');
And the result is:
ERROR: null value in column SCHEMA_VERSION violates not-null constraint
DETAIL: Failing row contains (1, null, Initial value).


  upgrade hive schema from 0.9.0 to 0.13.1 failed when back end db is postgres 
 --

 Key: HIVE-7797
 URL: https://issues.apache.org/jira/browse/HIVE-7797
 Project: Hive
  Issue Type: Bug
  Components: Metastore
Affects Versions: 0.14.0, 0.13.1
Reporter: Nemon Lou
Assignee: Nemon Lou
 Attachments: HIVE-7797.1.patch


 Using  the following command to upgrade hive schema :
 schematool -dbType postgres -upgradeSchemaFrom 0.9.0
 then got this:
 ERROR: null value in column SCHEMA_VERSION violates not-null constraint
 Log shows that the upgrade sql file 014-HIVE-3764.postgres.sql failed.
 The sql in it is :
 INSERT INTO VERSION (VER_ID, SCHEMA_VERSION, VERSION_COMMENT) VALUES 
 (1, '', 'Initial value');
 And the result is:
 ERROR: null value in column SCHEMA_VERSION violates not-null constraint
 DETAIL: Failing row contains (1, null, Initial value).



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-7797) upgrade hive schema from 0.9.0 to 0.13.1 failed when back end db is postgres

2015-04-16 Thread Nemon Lou (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-7797?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Nemon Lou updated HIVE-7797:

Description: 
Using  the following command to upgrade hive schema :
schematool -dbType postgres -upgradeSchemaFrom 0.9.0

ERROR: null value in column SCHEMA_VERSION violates not-null constraint

Log shows that the upgrade sql file 014-HIVE-3764.postgres.sql failed.
The sql in it is :
INSERT INTO VERSION (VER_ID, SCHEMA_VERSION, VERSION_COMMENT) VALUES 
(1, '', 'Initial value');
And the result is:
ERROR: null value in column SCHEMA_VERSION violates not-null constraint
DETAIL: Failing row contains (1, null, Initial value).

  was:
Using hive schema tool with the following command to upgrade hive schema failed:
schematool -dbType postgres -upgradeSchemaFrom 0.9.0

ERROR: null value in column SCHEMA_VERSION violates not-null constraint

Log shows that the upgrade sql file 014-HIVE-3764.postgres.sql failed.
The sql in it is :
INSERT INTO VERSION (VER_ID, SCHEMA_VERSION, VERSION_COMMENT) VALUES 
(1, '', 'Initial value');
And the result is:
ERROR: null value in column SCHEMA_VERSION violates not-null constraint
DETAIL: Failing row contains (1, null, Initial value).


  upgrade hive schema from 0.9.0 to 0.13.1 failed when back end db is postgres 
 --

 Key: HIVE-7797
 URL: https://issues.apache.org/jira/browse/HIVE-7797
 Project: Hive
  Issue Type: Bug
  Components: Metastore
Affects Versions: 0.14.0, 0.13.1
Reporter: Nemon Lou
Assignee: Nemon Lou
 Attachments: HIVE-7797.1.patch


 Using  the following command to upgrade hive schema :
 schematool -dbType postgres -upgradeSchemaFrom 0.9.0
 ERROR: null value in column SCHEMA_VERSION violates not-null constraint
 Log shows that the upgrade sql file 014-HIVE-3764.postgres.sql failed.
 The sql in it is :
 INSERT INTO VERSION (VER_ID, SCHEMA_VERSION, VERSION_COMMENT) VALUES 
 (1, '', 'Initial value');
 And the result is:
 ERROR: null value in column SCHEMA_VERSION violates not-null constraint
 DETAIL: Failing row contains (1, null, Initial value).



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Resolved] (HIVE-10358) LLAP: Tez in-place UI gets misaligned when state is TERMINATING

2015-04-16 Thread Prasanth Jayachandran (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-10358?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Prasanth Jayachandran resolved HIVE-10358.
--
Resolution: Fixed

Committed to llap branch.

 LLAP: Tez in-place UI gets misaligned when state is TERMINATING
 ---

 Key: HIVE-10358
 URL: https://issues.apache.org/jira/browse/HIVE-10358
 Project: Hive
  Issue Type: Sub-task
Affects Versions: llap
Reporter: Prasanth Jayachandran
Assignee: Prasanth Jayachandran
 Attachments: HIVE-10358.patch


 In-place UI goes misaligned for TERMINATING state as length of TERMINATING 
 string is greater than header.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-10361) CBO (Calcite Return Path): Overriding toString method in join algorithms

2015-04-16 Thread Jesus Camacho Rodriguez (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-10361?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jesus Camacho Rodriguez updated HIVE-10361:
---
Attachment: HIVE-10361.cbo.patch

 CBO (Calcite Return Path): Overriding toString method in join algorithms
 

 Key: HIVE-10361
 URL: https://issues.apache.org/jira/browse/HIVE-10361
 Project: Hive
  Issue Type: Sub-task
  Components: CBO
Affects Versions: cbo-branch
Reporter: Jesus Camacho Rodriguez
Assignee: Jesus Camacho Rodriguez
 Fix For: cbo-branch

 Attachments: HIVE-10361.cbo.patch






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Assigned] (HIVE-7797) using hive schema tool to upgrade hive schema in postgres from 0.9.0 to 0.13.1 failed

2015-04-16 Thread Nemon Lou (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-7797?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Nemon Lou reassigned HIVE-7797:
---

Assignee: Nemon Lou

  using hive schema tool to upgrade hive schema in postgres from 0.9.0 to 
 0.13.1 failed 
 ---

 Key: HIVE-7797
 URL: https://issues.apache.org/jira/browse/HIVE-7797
 Project: Hive
  Issue Type: Bug
  Components: Metastore
Affects Versions: 0.14.0, 0.13.1
Reporter: Nemon Lou
Assignee: Nemon Lou
 Attachments: HIVE-7797.1.patch


 Using hive schema tool with the following command to upgrade hive schema 
 failed:
 schematool -dbType postgres -upgradeSchemaFrom 0.9.0
 ERROR: null value in column SCHEMA_VERSION violates not-null constraint
 Log shows that the upgrade sql file 014-HIVE-3764.postgres.sql failed.
 The sql in it is :
 INSERT INTO VERSION (VER_ID, SCHEMA_VERSION, VERSION_COMMENT) VALUES 
 (1, '', 'Initial value');
 And the result is:
 ERROR: null value in column SCHEMA_VERSION violates not-null constraint
 DETAIL: Failing row contains (1, null, Initial value).



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-10331) ORC : Is null SARG filters out all row groups written in old ORC format

2015-04-16 Thread Mostafa Mokhtar (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-10331?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mostafa Mokhtar updated HIVE-10331:
---
Attachment: HIVE-10331.03.patch
HIVE-10331.03.patch

 ORC : Is null SARG filters out all row groups written in old ORC format
 ---

 Key: HIVE-10331
 URL: https://issues.apache.org/jira/browse/HIVE-10331
 Project: Hive
  Issue Type: Bug
  Components: Hive
Affects Versions: 1.1.0
Reporter: Mostafa Mokhtar
Assignee: Mostafa Mokhtar
 Fix For: 1.2.0

 Attachments: HIVE-10331.01.patch, HIVE-10331.02.patch, 
 HIVE-10331.03.patch, HIVE-10331.03.patch


 Queries are returning wrong results as all row groups gets filtered out and 
 no rows get scanned.
 {code}
 SELECT 
   count(*)
 FROM
 store_sales
 WHERE
 ss_addr_sk IS NULL
 {code}
 With hive.optimize.index.filter disabled we get the correct results
 In pickRowGroups stats show that hasNull_ is fales, while the rowgroup 
 actually has null.
 Same query runs fine for newly loaded ORC tables.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-10331) ORC : Is null SARG filters out all row groups written in old ORC format

2015-04-16 Thread Mostafa Mokhtar (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-10331?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mostafa Mokhtar updated HIVE-10331:
---
Attachment: HIVE-10331.04.patch

[~prasanth_j]

Makes sense. 

 ORC : Is null SARG filters out all row groups written in old ORC format
 ---

 Key: HIVE-10331
 URL: https://issues.apache.org/jira/browse/HIVE-10331
 Project: Hive
  Issue Type: Bug
  Components: Hive
Affects Versions: 1.1.0
Reporter: Mostafa Mokhtar
Assignee: Mostafa Mokhtar
 Fix For: 1.2.0

 Attachments: HIVE-10331.01.patch, HIVE-10331.02.patch, 
 HIVE-10331.03.patch, HIVE-10331.03.patch, HIVE-10331.04.patch


 Queries are returning wrong results as all row groups gets filtered out and 
 no rows get scanned.
 {code}
 SELECT 
   count(*)
 FROM
 store_sales
 WHERE
 ss_addr_sk IS NULL
 {code}
 With hive.optimize.index.filter disabled we get the correct results
 In pickRowGroups stats show that hasNull_ is fales, while the rowgroup 
 actually has null.
 Same query runs fine for newly loaded ORC tables.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-10361) CBO (Calcite Return Path): Overriding toString method in join algorithms

2015-04-16 Thread Jesus Camacho Rodriguez (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-10361?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jesus Camacho Rodriguez updated HIVE-10361:
---
Attachment: HIVE-10361.cbo.patch

 CBO (Calcite Return Path): Overriding toString method in join algorithms
 

 Key: HIVE-10361
 URL: https://issues.apache.org/jira/browse/HIVE-10361
 Project: Hive
  Issue Type: Sub-task
  Components: CBO
Affects Versions: cbo-branch
Reporter: Jesus Camacho Rodriguez
Assignee: Jesus Camacho Rodriguez
 Fix For: cbo-branch

 Attachments: HIVE-10361.cbo.patch






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-10222) Upgrade Calcite dependency to newest version

2015-04-16 Thread Jesus Camacho Rodriguez (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-10222?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jesus Camacho Rodriguez updated HIVE-10222:
---
Attachment: HIVE-10222.03.patch

Running QA on new release.

 Upgrade Calcite dependency to newest version
 

 Key: HIVE-10222
 URL: https://issues.apache.org/jira/browse/HIVE-10222
 Project: Hive
  Issue Type: Bug
Reporter: Jesus Camacho Rodriguez
Assignee: Jesus Camacho Rodriguez
 Attachments: HIVE-10222.01.patch, HIVE-10222.02.patch, 
 HIVE-10222.03.patch, HIVE-10222.patch


  Upgrade Calcite version to 1.2.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-10288) Cannot call permanent UDFs

2015-04-16 Thread Chinna Rao Lalam (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-10288?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14497858#comment-14497858
 ] 

Chinna Rao Lalam commented on HIVE-10288:
-

Failed tests seems not related to this patch.

 Cannot call permanent UDFs
 --

 Key: HIVE-10288
 URL: https://issues.apache.org/jira/browse/HIVE-10288
 Project: Hive
  Issue Type: Bug
Affects Versions: 1.2.0
Reporter: Nezih Yigitbasi
Assignee: Chinna Rao Lalam
 Attachments: HIVE-10288.1.patch, HIVE-10288.patch


 Just pulled the trunk and built the hive binary. If I create a permanent udf 
 and exit the cli, and then open the cli and try calling the udf it fails with 
 the exception below. However, the call succeeds if I call the udf right after 
 registering the permanent udf (without exiting the cli). The call also 
 succeeds with the apache-hive-1.0.0 release.
 {code}
 15-04-13 17:04:54,004 INFO  org.apache.hadoop.hive.ql.log.PerfLogger 
 (PerfLogger.java:PerfLogEnd(148)) - /PERFLOG method=parse 
 start=1428969893115 end=1428969894004 duration=889 
 from=org.apache.hadoop.hive.ql.Driver
 2015-04-13 17:04:54,007 DEBUG org.apache.hadoop.hive.ql.Driver 
 (Driver.java:recordValidTxns(939)) - Encoding valid txns info 
 9223372036854775807:
 2015-04-13 17:04:54,007 INFO  org.apache.hadoop.hive.ql.log.PerfLogger 
 (PerfLogger.java:PerfLogBegin(121)) - PERFLOG method=semanticAnalyze 
 from=org.apache.hadoop.hive.ql.Driver
 2015-04-13 17:04:54,052 INFO  org.apache.hadoop.hive.ql.parse.CalcitePlanner 
 (SemanticAnalyzer.java:analyzeInternal(9997)) - Starting Semantic Analysis
 2015-04-13 17:04:54,053 DEBUG org.apache.hadoop.hive.ql.exec.FunctionRegistry 
 (FunctionRegistry.java:getGenericUDAFResolver(942)) - Looking up GenericUDAF: 
 hour_now
 2015-04-13 17:04:54,053 INFO  org.apache.hadoop.hive.ql.parse.CalcitePlanner 
 (SemanticAnalyzer.java:genResolvedParseTree(9980)) - Completed phase 1 of 
 Semantic Analysis
 2015-04-13 17:04:54,053 INFO  org.apache.hadoop.hive.ql.parse.CalcitePlanner 
 (SemanticAnalyzer.java:getMetaData(1530)) - Get metadata for source tables
 2015-04-13 17:04:54,054 INFO  org.apache.hadoop.hive.metastore.HiveMetaStore 
 (HiveMetaStore.java:logInfo(744)) - 0: get_table : db=default tbl=test_table
 2015-04-13 17:04:54,054 INFO  
 org.apache.hadoop.hive.metastore.HiveMetaStore.audit 
 (HiveMetaStore.java:logAuditEvent(369)) - ugi=nyigitbasi   ip=unknown-ip-addr 
  cmd=get_table : db=default tbl=test_table
 2015-04-13 17:04:54,054 DEBUG org.apache.hadoop.hive.metastore.ObjectStore 
 (ObjectStore.java:debugLog(6776)) - Open transaction: count = 1, isActive = 
 true at:
   
 org.apache.hadoop.hive.metastore.ObjectStore.getTable(ObjectStore.java:927)
 2015-04-13 17:04:54,054 DEBUG org.apache.hadoop.hive.metastore.ObjectStore 
 (ObjectStore.java:debugLog(6776)) - Open transaction: count = 2, isActive = 
 true at:
   
 org.apache.hadoop.hive.metastore.ObjectStore.getMTable(ObjectStore.java:990)
 2015-04-13 17:04:54,104 DEBUG org.apache.hadoop.hive.metastore.ObjectStore 
 (ObjectStore.java:debugLog(6776)) - Commit transaction: count = 1, isactive 
 true at:
   
 org.apache.hadoop.hive.metastore.ObjectStore.getMTable(ObjectStore.java:998)
 2015-04-13 17:04:54,232 DEBUG org.apache.hadoop.hive.metastore.ObjectStore 
 (ObjectStore.java:debugLog(6776)) - Commit transaction: count = 0, isactive 
 true at:
   
 org.apache.hadoop.hive.metastore.ObjectStore.getTable(ObjectStore.java:929)
 2015-04-13 17:04:54,242 INFO  org.apache.hadoop.hive.ql.parse.CalcitePlanner 
 (SemanticAnalyzer.java:getMetaData(1682)) - Get metadata for subqueries
 2015-04-13 17:04:54,247 INFO  org.apache.hadoop.hive.ql.parse.CalcitePlanner 
 (SemanticAnalyzer.java:getMetaData(1706)) - Get metadata for destination 
 tables
 2015-04-13 17:04:54,256 INFO  org.apache.hadoop.hive.ql.parse.CalcitePlanner 
 (SemanticAnalyzer.java:genResolvedParseTree(9984)) - Completed getting 
 MetaData in Semantic Analysis
 2015-04-13 17:04:54,259 INFO  
 org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer 
 (CalcitePlanner.java:canHandleAstForCbo(369)) - Not invoking CBO because the 
 statement has too few joins
 2015-04-13 17:04:54,344 DEBUG 
 org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe 
 (LazySimpleSerDe.java:initialize(135)) - 
 org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe initialized with: 
 columnNames=[_c0, _c1] columnTypes=[int, int] separator=[[B@6e6d4780] 
 nullstring=\N lastColumnTakesRest=false timestampFormats=null
 2015-04-13 17:04:54,406 DEBUG org.apache.hadoop.hive.ql.parse.CalcitePlanner 
 (SemanticAnalyzer.java:genTablePlan(9458)) - Created Table Plan for 
 test_table TS[0]
 2015-04-13 17:04:54,410 DEBUG org.apache.hadoop.hive.ql.parse.CalcitePlanner 
 (SemanticAnalyzer.java:genBodyPlan(8815)) - RR before GB test_table{(_c0,_c0: 
 

[jira] [Commented] (HIVE-7594) Hive JDBC client: out of sequence response on large long running query

2015-04-16 Thread Shrey Mehrotra (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-7594?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14497954#comment-14497954
 ] 

Shrey Mehrotra commented on HIVE-7594:
--

I am also facing same issue with Hive 1.0.0. Any patch or workaround available?

 Hive JDBC client: out of sequence response on large long running query
 

 Key: HIVE-7594
 URL: https://issues.apache.org/jira/browse/HIVE-7594
 Project: Hive
  Issue Type: Bug
  Components: Clients, HiveServer2
Affects Versions: 0.13.0
 Environment: HDP2.1
Reporter: Hari Sekhon

 When executing a long running query in a JDBC client (Squirrel) to 
 HiveServer2 after several minutes I get this error in the client:
 {code}
 Error: org.apache.thrift.TApplicationException: ExecuteStatement failed: out 
 of sequence response
 SQLState:  08S01
 ErrorCode: 0
 {code}
 I've seen this before in, iirc when running 2 queries in 1 session but I've 
 closed the client and run only this single query in a new session each time. 
 I did a search and saw HIVE-6893 referring to a Metastore exception which I 
 have in some older logs but not corresponding / recent in these recent 
 instances, the error seems different in this case but may be related.
 The query to reproduce is select count(*) from myTable where myTable is a 
 1TB table of 620 million rows. This happens in both MR and Tez execution 
 engines running on Yarn.
 Here are all the jars I've added to the classpath (taken from Hortonworks doc 
 http://docs.hortonworks.com/HDPDocuments/HDP2/HDP-2.1-latest/bk_dataintegration/content/ch_using-hive-2.html,
  plus added hadoop-common, hive-exec and slf4j-api to solve class not found 
 issues on top of that):
 commons-codec-1.4.jar
 commons-logging-1.1.3.jar
 hadoop-common-2.4.0.2.1.3.0-563.jar
 hive-exec-0.13.0.2.1.3.0-563.jar
 hive-jdbc-0.13.0.2.1.3.0-563.jar
 hive-service-0.13.0.2.1.3.0-563.jar
 httpclient-4.2.5.jar
 httpcore-4.2.5.jar
 libthrift-0.9.0.jar
 slf4j-api-1.7.5.jar
 I am seeing errors like this in the hiveserver2.log:
 {code}
 2014-08-01 15:04:31,358 ERROR [pool-5-thread-3]: server.TThreadPoolServer 
 (TThreadPoolServer.java:run(215)) - Error occurred during processing of 
 message.
 java.lang.RuntimeException: org.apache.thrift.transport.TTransportException
 at 
 org.apache.thrift.transport.TSaslServerTransport$Factory.getTransport(TSaslServerTransport.java:219)
 at 
 org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPoolServer.java:189)
 at 
 java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
 at 
 java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
 at java.lang.Thread.run(Thread.java:745)
 Caused by: org.apache.thrift.transport.TTransportException
 at 
 org.apache.thrift.transport.TIOStreamTransport.read(TIOStreamTransport.java:132)
 at org.apache.thrift.transport.TTransport.readAll(TTransport.java:84)
 at 
 org.apache.thrift.transport.TSaslTransport.receiveSaslMessage(TSaslTransport.java:182)
 at 
 org.apache.thrift.transport.TSaslServerTransport.handleSaslStartMessage(TSaslServerTransport.java:125)
 at 
 org.apache.thrift.transport.TSaslTransport.open(TSaslTransport.java:253)
 at 
 org.apache.thrift.transport.TSaslServerTransport.open(TSaslServerTransport.java:41)
 at 
 org.apache.thrift.transport.TSaslServerTransport$Factory.getTransport(TSaslServerTransport.java:216)
 ... 4 more
 ...
 2014-08-01 15:06:31,520 ERROR [pool-5-thread-3]: server.TThreadPoolServer 
 (TThreadPoolServer.java:run(215)) - Error occurred during processing of 
 message.
 java.lang.RuntimeException: org.apache.thrift.transport.TTransportException
 at 
 org.apache.thrift.transport.TSaslServerTransport$Factory.getTransport(TSaslServerTransport.java:219)
 at 
 org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPoolServer.java:189)
 at 
 java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
 at 
 java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
 at java.lang.Thread.run(Thread.java:745)
 Caused by: org.apache.thrift.transport.TTransportException
 at 
 org.apache.thrift.transport.TIOStreamTransport.read(TIOStreamTransport.java:132)
 at org.apache.thrift.transport.TTransport.readAll(TTransport.java:84)
 at 
 org.apache.thrift.transport.TSaslTransport.receiveSaslMessage(TSaslTransport.java:182)
 at 
 org.apache.thrift.transport.TSaslServerTransport.handleSaslStartMessage(TSaslServerTransport.java:125)
 at 
 org.apache.thrift.transport.TSaslTransport.open(TSaslTransport.java:253)
 at 
 

[jira] [Commented] (HIVE-10228) Changes to Hive Export/Import/DropTable/DropPartition to support replication semantics

2015-04-16 Thread Alan Gates (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-10228?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14497895#comment-14497895
 ] 

Alan Gates commented on HIVE-10228:
---

Wow, when I saw it was a 150K patch I was hoping it was mostly generated code.  
No such luck.

Code level comments on review board, higher level below:

This stuff needs some major doc work as you're introducing a new concept of a 
table being replicated or generated from replication.  Is there a doc JIRA for 
the replication work yet?  If so we should link it to this JIRA.

Parser changes:
I don't understand why DROP TABLE needs the replication clause.  As far as I 
can tell from the changes in DDLSemanticAnalyzer this is semantically 
equivalent to IF EXISTS.  Why not use that?

Adding METADATA and REPLICATION as keywords is not backwards compatible.  We 
either need to explicitly note that in this JIRA or add them to the list of 
reserved keywords allowed as identifiers in IdentifiersParser.g.  I suspect the 
latter is a better choice.







 Changes to Hive Export/Import/DropTable/DropPartition to support replication 
 semantics
 --

 Key: HIVE-10228
 URL: https://issues.apache.org/jira/browse/HIVE-10228
 Project: Hive
  Issue Type: Sub-task
  Components: Import/Export
Affects Versions: 1.2.0
Reporter: Sushanth Sowmyan
Assignee: Sushanth Sowmyan
 Attachments: HIVE-10228.2.patch, HIVE-10228.3.patch, HIVE-10228.patch


 We need to update a couple of hive commands to support replication semantics. 
 To wit, we need the following:
 EXPORT ... [FOR [METADATA] REPLICATION(“comment”)]
 Export will now support an extra optional clause to tell it that this export 
 is being prepared for the purpose of replication. There is also an additional 
 optional clause here, that allows for the export to be a metadata-only 
 export, to handle cases of capturing the diff for alter statements, for 
 example.
 Also, if done for replication, the non-presence of a table, or a table being 
 a view/offline table/non-native table is not considered an error, and 
 instead, will result in a successful no-op.
 IMPORT ... (as normal) – but handles new semantics 
 No syntax changes for import, but import will have to change to be able to 
 handle all the permutations of export dumps possible. Also, import will have 
 to ensure that it should update the object only if the update being imported 
 is not older than the state of the object. Also, import currently does not 
 work with dbname.tablename kind of specification, this should be fixed to 
 work.
 DROP TABLE ... FOR REPLICATION('eventid')
 Drop Table now has an additional clause, to specify that this drop table is 
 being done for replication purposes, and that the dop should not actually 
 drop the table if the table is newer than that event id specified.
 ALTER TABLE ... DROP PARTITION (...) FOR REPLICATION('eventid')
 Similarly, Drop Partition also has an equivalent change to Drop Table.
 =
 In addition, we introduce a new property repl.last.id, which when tagged on 
 to table properties or partition properties on a replication-destination, 
 holds the effective state identifier of the object.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-10360) CBO (Calcite Return Path): Remove HiveCostUtil

2015-04-16 Thread Jesus Camacho Rodriguez (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-10360?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jesus Camacho Rodriguez updated HIVE-10360:
---
Summary: CBO (Calcite Return Path): Remove HiveCostUtil  (was: CBO (Calcite 
Return Path): Prevent cost overflow)

 CBO (Calcite Return Path): Remove HiveCostUtil
 --

 Key: HIVE-10360
 URL: https://issues.apache.org/jira/browse/HIVE-10360
 Project: Hive
  Issue Type: Sub-task
  Components: CBO
Affects Versions: cbo-branch
Reporter: Jesus Camacho Rodriguez
Assignee: Jesus Camacho Rodriguez
 Fix For: cbo-branch






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-10360) CBO (Calcite Return Path): Remove HiveCostUtil

2015-04-16 Thread Jesus Camacho Rodriguez (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-10360?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jesus Camacho Rodriguez updated HIVE-10360:
---
Description: The methods were already moved to {{HiveAlgorithmsUtil}}.

 CBO (Calcite Return Path): Remove HiveCostUtil
 --

 Key: HIVE-10360
 URL: https://issues.apache.org/jira/browse/HIVE-10360
 Project: Hive
  Issue Type: Sub-task
  Components: CBO
Affects Versions: cbo-branch
Reporter: Jesus Camacho Rodriguez
Assignee: Jesus Camacho Rodriguez
 Fix For: cbo-branch


 The methods were already moved to {{HiveAlgorithmsUtil}}.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-10121) Implement a hive --service udflint command to check UDF jars for common shading mistakes

2015-04-16 Thread Abdelrahman Shettia (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-10121?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Abdelrahman Shettia updated HIVE-10121:
---
Attachment: HIVE-10121.1.patch

 Implement a hive --service udflint command to check UDF jars for common 
 shading mistakes
 

 Key: HIVE-10121
 URL: https://issues.apache.org/jira/browse/HIVE-10121
 Project: Hive
  Issue Type: New Feature
  Components: UDF
Reporter: Gopal V
Assignee: Abdelrahman Shettia
 Fix For: 1.2.0

 Attachments: HIVE-10121.1.patch


 Several SerDe and UDF jars tend to shade in various parts of the dependencies 
 including hadoop-common or guava without relocation.
 Implement a simple udflint tool which automates some part of the class path 
 and shaded resources audit process required when upgrading a hive install 
 from an old version to a new one.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-10121) Implement a hive --service udflint command to check UDF jars for common shading mistakes

2015-04-16 Thread Abdelrahman Shettia (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-10121?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14498076#comment-14498076
 ] 

Abdelrahman Shettia commented on HIVE-10121:


Hi Gopal, 

I am attaching the following use cases output files: 

bad_udfs.out
bad_udfs_verbose.out
good_udfs.out
good_udfs_verbose.out

Usage: 

Normal mode: 
$ hive --service UDFLint -file /tmp/hive_udf-1.0.0.jar 
Verbose mode: 
hive --service UDFLint -jar yi/hive-json-serde-0.3.jar -v 
Without any options:
[root@sandbox test]#  hive --service UDFLint 
usage: udflint
 -h,--helpprint help message
--hiveconf property=value   Use value for given property
--jar arg   Comma separated list of jars to validate
 -v,--verbose Verbose mode (Run the tool in debug
  mode)

Please let me know if you have questions. 

Thanks
-Rahman

 Implement a hive --service udflint command to check UDF jars for common 
 shading mistakes
 

 Key: HIVE-10121
 URL: https://issues.apache.org/jira/browse/HIVE-10121
 Project: Hive
  Issue Type: New Feature
  Components: UDF
Reporter: Gopal V
Assignee: Abdelrahman Shettia
 Fix For: 1.2.0

 Attachments: HIVE-10121.1.patch, bad_udfs.out, bad_udfs_verbose.out, 
 good_udfs.out, good_udfs_verbose.out


 Several SerDe and UDF jars tend to shade in various parts of the dependencies 
 including hadoop-common or guava without relocation.
 Implement a simple udflint tool which automates some part of the class path 
 and shaded resources audit process required when upgrading a hive install 
 from an old version to a new one.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-10306) We need to print tez summary when hive.server2.logging.level = PERFORMANCE.

2015-04-16 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-10306?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14498118#comment-14498118
 ] 

Hive QA commented on HIVE-10306:




{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12725753/HIVE-10306.4.patch

{color:red}ERROR:{color} -1 due to 15 failed/errored test(s), 8707 tests 
executed
*Failed tests:*
{noformat}
TestMinimrCliDriver-bucketmapjoin6.q-constprog_partitioner.q-infer_bucket_sort_dyn_part.q-and-1-more
 - did not produce a TEST-*.xml file
TestMinimrCliDriver-external_table_with_space_in_location_path.q-infer_bucket_sort_merge.q-auto_sortmerge_join_16.q-and-1-more
 - did not produce a TEST-*.xml file
TestMinimrCliDriver-groupby2.q-import_exported_table.q-bucketizedhiveinputformat.q-and-1-more
 - did not produce a TEST-*.xml file
TestMinimrCliDriver-index_bitmap3.q-stats_counter_partitioned.q-temp_table_external.q-and-1-more
 - did not produce a TEST-*.xml file
TestMinimrCliDriver-infer_bucket_sort_map_operators.q-join1.q-bucketmapjoin7.q-and-1-more
 - did not produce a TEST-*.xml file
TestMinimrCliDriver-infer_bucket_sort_num_buckets.q-disable_merge_for_bucketing.q-uber_reduce.q-and-1-more
 - did not produce a TEST-*.xml file
TestMinimrCliDriver-infer_bucket_sort_reducers_power_two.q-scriptfile1.q-scriptfile1_win.q-and-1-more
 - did not produce a TEST-*.xml file
TestMinimrCliDriver-leftsemijoin_mr.q-load_hdfs_file_with_space_in_the_name.q-root_dir_external_table.q-and-1-more
 - did not produce a TEST-*.xml file
TestMinimrCliDriver-list_bucket_dml_10.q-bucket_num_reducers.q-bucket6.q-and-1-more
 - did not produce a TEST-*.xml file
TestMinimrCliDriver-load_fs2.q-file_with_header_footer.q-ql_rewrite_gbtoidx_cbo_1.q-and-1-more
 - did not produce a TEST-*.xml file
TestMinimrCliDriver-parallel_orderby.q-reduce_deduplicate.q-ql_rewrite_gbtoidx_cbo_2.q-and-1-more
 - did not produce a TEST-*.xml file
TestMinimrCliDriver-ql_rewrite_gbtoidx.q-smb_mapjoin_8.q - did not produce a 
TEST-*.xml file
TestMinimrCliDriver-schemeAuthority2.q-bucket4.q-input16_cc.q-and-1-more - did 
not produce a TEST-*.xml file
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_udaf_percentile_approx_23
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_union_view
{noformat}

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/3457/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/3457/console
Test logs: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-3457/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 15 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12725753 - PreCommit-HIVE-TRUNK-Build

 We need to print tez summary when hive.server2.logging.level = PERFORMANCE. 
 -

 Key: HIVE-10306
 URL: https://issues.apache.org/jira/browse/HIVE-10306
 Project: Hive
  Issue Type: Bug
Reporter: Hari Sankar Sivarama Subramaniyan
Assignee: Hari Sankar Sivarama Subramaniyan
 Attachments: HIVE-10306.1.patch, HIVE-10306.2.patch, 
 HIVE-10306.3.patch, HIVE-10306.4.patch


 We need to print tez summary when hive.server2.logging.level = PERFORMANCE. 
 We introduced this parameter via HIVE-10119.
 The logging param for levels is only relevant to HS2, so for hive-cli users 
 the hive.tez.exec.print.summary still makes sense. We can check for log-level 
 param as well, in places we are checking value of 
 hive.tez.exec.print.summary. Ie, consider hive.tez.exec.print.summary=true if 
 log.level = PERFORMANCE.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-10364) The HMS upgrade script test does not publish results when prepare.sh fails.

2015-04-16 Thread JIRA

 [ 
https://issues.apache.org/jira/browse/HIVE-10364?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sergio Peña updated HIVE-10364:
---
Attachment: HIVE-10364.1.patch

 The HMS upgrade script test does not publish results when prepare.sh fails.
 ---

 Key: HIVE-10364
 URL: https://issues.apache.org/jira/browse/HIVE-10364
 Project: Hive
  Issue Type: Bug
  Components: Testing Infrastructure
Reporter: Sergio Peña
Assignee: Sergio Peña
 Attachments: HIVE-10364.1.patch


 The HMS upgrade script must publish succeed or failure results to JIRA. This 
 bug is not publishing any results on JIRA is the prepare.sh script fails.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-10190) CBO: AST mode checks for TABLESAMPLE with AST.toString().contains(TOK_TABLESPLITSAMPLE)

2015-04-16 Thread JIRA

 [ 
https://issues.apache.org/jira/browse/HIVE-10190?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sergio Peña updated HIVE-10190:
---
Attachment: HIVE-10190.05.patch

Re-upload patch to allow jenkins run the tests again.

 CBO: AST mode checks for TABLESAMPLE with 
 AST.toString().contains(TOK_TABLESPLITSAMPLE)
 -

 Key: HIVE-10190
 URL: https://issues.apache.org/jira/browse/HIVE-10190
 Project: Hive
  Issue Type: Bug
  Components: CBO
Affects Versions: 1.2.0
Reporter: Gopal V
Assignee: Reuben Kuhnert
Priority: Trivial
  Labels: perfomance
 Attachments: HIVE-10190-querygen.py, HIVE-10190.01.patch, 
 HIVE-10190.02.patch, HIVE-10190.03.patch, HIVE-10190.04.patch, 
 HIVE-10190.05.patch, HIVE-10190.05.patch


 {code}
 public static boolean validateASTForUnsupportedTokens(ASTNode ast) {
 String astTree = ast.toStringTree();
 // if any of following tokens are present in AST, bail out
 String[] tokens = { TOK_CHARSETLITERAL, TOK_TABLESPLITSAMPLE };
 for (String token : tokens) {
   if (astTree.contains(token)) {
 return false;
   }
 }
 return true;
   }
 {code}
 This is an issue for a SQL query which is bigger in AST form than in text 
 (~700kb).



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9917) After HIVE-3454 is done, make int to timestamp conversion configurable

2015-04-16 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9917?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14498235#comment-14498235
 ] 

Hive QA commented on HIVE-9917:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12725748/HIVE-9917.patch

{color:red}ERROR:{color} -1 due to 16 failed/errored test(s), 8704 tests 
executed
*Failed tests:*
{noformat}
TestCustomAuthentication - did not produce a TEST-*.xml file
TestMinimrCliDriver-bucketmapjoin6.q-constprog_partitioner.q-infer_bucket_sort_dyn_part.q-and-1-more
 - did not produce a TEST-*.xml file
TestMinimrCliDriver-external_table_with_space_in_location_path.q-infer_bucket_sort_merge.q-auto_sortmerge_join_16.q-and-1-more
 - did not produce a TEST-*.xml file
TestMinimrCliDriver-groupby2.q-import_exported_table.q-bucketizedhiveinputformat.q-and-1-more
 - did not produce a TEST-*.xml file
TestMinimrCliDriver-index_bitmap3.q-stats_counter_partitioned.q-temp_table_external.q-and-1-more
 - did not produce a TEST-*.xml file
TestMinimrCliDriver-infer_bucket_sort_map_operators.q-join1.q-bucketmapjoin7.q-and-1-more
 - did not produce a TEST-*.xml file
TestMinimrCliDriver-infer_bucket_sort_num_buckets.q-disable_merge_for_bucketing.q-uber_reduce.q-and-1-more
 - did not produce a TEST-*.xml file
TestMinimrCliDriver-infer_bucket_sort_reducers_power_two.q-scriptfile1.q-scriptfile1_win.q-and-1-more
 - did not produce a TEST-*.xml file
TestMinimrCliDriver-leftsemijoin_mr.q-load_hdfs_file_with_space_in_the_name.q-root_dir_external_table.q-and-1-more
 - did not produce a TEST-*.xml file
TestMinimrCliDriver-list_bucket_dml_10.q-bucket_num_reducers.q-bucket6.q-and-1-more
 - did not produce a TEST-*.xml file
TestMinimrCliDriver-load_fs2.q-file_with_header_footer.q-ql_rewrite_gbtoidx_cbo_1.q-and-1-more
 - did not produce a TEST-*.xml file
TestMinimrCliDriver-parallel_orderby.q-reduce_deduplicate.q-ql_rewrite_gbtoidx_cbo_2.q-and-1-more
 - did not produce a TEST-*.xml file
TestMinimrCliDriver-ql_rewrite_gbtoidx.q-smb_mapjoin_8.q - did not produce a 
TEST-*.xml file
TestMinimrCliDriver-schemeAuthority2.q-bucket4.q-input16_cc.q-and-1-more - did 
not produce a TEST-*.xml file
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_udaf_percentile_approx_23
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_union_view
{noformat}

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/3458/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/3458/console
Test logs: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-3458/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 16 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12725748 - PreCommit-HIVE-TRUNK-Build

 After HIVE-3454 is done, make int to timestamp conversion configurable
 --

 Key: HIVE-9917
 URL: https://issues.apache.org/jira/browse/HIVE-9917
 Project: Hive
  Issue Type: Improvement
Reporter: Aihua Xu
Assignee: Aihua Xu
 Attachments: HIVE-9917.patch


 After HIVE-3454 is fixed, we will have correct behavior of converting int to 
 timestamp. While the customers are using such incorrect behavior for so long, 
 better to make it configurable so that in one release, it will default to 
 old/inconsistent way and the next release will default to new/consistent way. 
 And then we will deprecate it.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-10365) First job fails with StackOverflowError [Spark Branch]

2015-04-16 Thread Jimmy Xiang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-10365?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jimmy Xiang updated HIVE-10365:
---
Summary: First job fails with StackOverflowError [Spark Branch]  (was: 
First job fails with StackOverflowError)

 First job fails with StackOverflowError [Spark Branch]
 --

 Key: HIVE-10365
 URL: https://issues.apache.org/jira/browse/HIVE-10365
 Project: Hive
  Issue Type: Bug
Affects Versions: spark-branch
Reporter: Jimmy Xiang
Assignee: Jimmy Xiang

 When running some queries on Yarn with standalone Hadoop, the first query 
 fails with StackOverflowError:
 {noformat}
 java.lang.StackOverflowError
   at 
 java.util.concurrent.ConcurrentHashMap.hash(ConcurrentHashMap.java:333)
   at 
 java.util.concurrent.ConcurrentHashMap.putIfAbsent(ConcurrentHashMap.java:1145)
   at java.lang.ClassLoader.getClassLoadingLock(ClassLoader.java:464)
   at java.lang.ClassLoader.loadClass(ClassLoader.java:405)
   at java.lang.ClassLoader.loadClass(ClassLoader.java:412)
   at java.lang.ClassLoader.loadClass(ClassLoader.java:412)
   at java.lang.ClassLoader.loadClass(ClassLoader.java:412)
   at java.lang.ClassLoader.loadClass(ClassLoader.java:412)
   at java.lang.ClassLoader.loadClass(ClassLoader.java:412)
   at java.lang.ClassLoader.loadClass(ClassLoader.java:412)
   at java.lang.ClassLoader.loadClass(ClassLoader.java:412)
   at java.lang.ClassLoader.loadClass(ClassLoader.java:412)
   at java.lang.ClassLoader.loadClass(ClassLoader.java:412)
   at java.lang.ClassLoader.loadClass(ClassLoader.java:412)
 {noformat}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-10313) Literal Decimal ExprNodeConstantDesc should contain value of HiveDecimal instead of String

2015-04-16 Thread Jimmy Xiang (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-10313?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14498280#comment-14498280
 ] 

Jimmy Xiang commented on HIVE-10313:


+1

 Literal Decimal ExprNodeConstantDesc should contain value of HiveDecimal 
 instead of String
 --

 Key: HIVE-10313
 URL: https://issues.apache.org/jira/browse/HIVE-10313
 Project: Hive
  Issue Type: Bug
  Components: Query Processor
Affects Versions: 1.0.0
Reporter: Chaoyu Tang
Assignee: Chaoyu Tang
 Attachments: HIVE-10313.1.patch, HIVE-10313.patch


 In TyepCheckProcFactory.NumExprProcessor, the ExprNodeConstantDesc is created 
 from strVal:
 {code}
 else if (expr.getText().endsWith(BD)) {
   // Literal decimal
   String strVal = expr.getText().substring(0, expr.getText().length() 
 - 2);
   HiveDecimal hd = HiveDecimal.create(strVal);
   int prec = 1;
   int scale = 0;
   if (hd != null) {
 prec = hd.precision();
 scale = hd.scale();
   }
   DecimalTypeInfo typeInfo = TypeInfoFactory.getDecimalTypeInfo(prec, 
 scale);
   return new ExprNodeConstantDesc(typeInfo, strVal);
 } 
 {code}
 It should use HiveDecmal:
 return new ExprNodeConstantDesc(typeInfo, hd);



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8950) Add support in ParquetHiveSerde to create table schema from a parquet file

2015-04-16 Thread Ashish K Singh (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8950?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14498249#comment-14498249
 ] 

Ashish K Singh commented on HIVE-8950:
--

[~kamrul] for now you will have to create hive table with proper schema 
yourself. The patch is almost ready, it got side tracked for a bit. There are a 
few things I have to address. Will soon submit an updated patch.

 Add support in ParquetHiveSerde to create table schema from a parquet file
 --

 Key: HIVE-8950
 URL: https://issues.apache.org/jira/browse/HIVE-8950
 Project: Hive
  Issue Type: Improvement
Reporter: Ashish K Singh
Assignee: Ashish K Singh
 Attachments: HIVE-8950.1.patch, HIVE-8950.2.patch, HIVE-8950.3.patch, 
 HIVE-8950.4.patch, HIVE-8950.5.patch, HIVE-8950.6.patch, HIVE-8950.patch


 PARQUET-76 and PARQUET-47 ask for creating parquet backed tables without 
 having to specify the column names and types. As, parquet files store schema 
 in their footer, it is possible to generate hive schema from parquet file's 
 metadata. This will improve usability of parquet backed tables.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9923) No clear message when from is missing

2015-04-16 Thread Yongzhi Chen (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9923?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14498288#comment-14498288
 ] 

Yongzhi Chen commented on HIVE-9923:


Hive support simple select without from statement (when there is a UDF). When 
select  in an insert statement, hive need the from clause. The new patch only 
check from clause null or not in the insert statement. 

 No clear message when from is missing
 ---

 Key: HIVE-9923
 URL: https://issues.apache.org/jira/browse/HIVE-9923
 Project: Hive
  Issue Type: Bug
Affects Versions: 1.0.0
Reporter: Jeff Zhang
Assignee: Yongzhi Chen
 Attachments: HIVE-9923.1.patch


 For the following sql, from is missing but it throw NPE which is not clear 
 for user.
 {code}
 hive insert overwrite directory '/tmp/hive-3' select sb1.name, sb2.age 
 student_bucketed sb1 join student_bucketed sb2 on sb1.name=sb2.name;
 FAILED: NullPointerException null
 {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9917) After HIVE-3454 is done, make int to timestamp conversion configurable

2015-04-16 Thread Aihua Xu (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9917?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14498355#comment-14498355
 ] 

Aihua Xu commented on HIVE-9917:


The test failures are unrelated to the patch.

 After HIVE-3454 is done, make int to timestamp conversion configurable
 --

 Key: HIVE-9917
 URL: https://issues.apache.org/jira/browse/HIVE-9917
 Project: Hive
  Issue Type: Improvement
Reporter: Aihua Xu
Assignee: Aihua Xu
 Attachments: HIVE-9917.patch


 After HIVE-3454 is fixed, we will have correct behavior of converting int to 
 timestamp. While the customers are using such incorrect behavior for so long, 
 better to make it configurable so that in one release, it will default to 
 old/inconsistent way and the next release will default to new/consistent way. 
 And then we will deprecate it.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-10366) Multiple Date Functions inside Named Struct return the wrong values

2015-04-16 Thread Matt Traynham (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-10366?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Matt Traynham updated HIVE-10366:
-
Description: 
Using multiple date functions inside a named struct results in incorrect 
values.  The first date function is correct.  All subsequent date functions 
receive the value of the first date function, unless the function has a 
different valued argument.

As shown below, the 3rd query returns an incorrect value for `month(datetime)`. 
 A not so pleasant workaround is to substr datetime resulting in a different 
value (4th query).

 DESCRIBE my_datamart;
OK
datetimestring

 SELECT datetime FROM my_datamart LIMIT 1;
OK
2015-04-14T00:00:00

 SELECT named_struct('dayofmonth', day(datetime), 'monthofyear', 
 month(datetime)), day(datetime), month(datetime) FROM my_datamart LIMIT 1;
OK
{dayofmonth:14,monthofyear:14} 14   4

 SELECT named_struct('dayofmonth', day(datetime), 'monthofyear', 
 month(substr(datetime, 0, 10))), day(datetime), month(datetime) FROM 
 my_datamart LIMIT 1;
OK
{dayofmonth:14,monthofyear:4} 144

I havn't tried all the other date functions, but this is happens with  
weekofyear as well.  Looks like a caching issue...

  was:
Using multiple date functions inside a named struct results in incorrect 
values.  The first date function is correct.  All subsequent date functions 
receive the value of the first date function, unless the function has a 
different valued argument.

As shown below, the 3rd query returns an incorrect value for `month(datetime)`. 
 A not so pleasant workaround is to substr datetime resulting in a different 
value (4th query).

 DESCRIBE my_datamart;
OK
datetimestring

 SELECT datetime FROM my_datamart LIMIT 1;
OK
2015-04-14T00:00:00

 SELECT named_struct('dayofmonth', day(datetime), 'monthofyear', 
 month(datetime)), day(datetime), month(datetime) FROM my_datamart LIMIT 1;
OK
{dayofmonth:14,monthofyear:14}  14  4

 SELECT named_struct('dayofmonth', day(datetime), 'monthofyear', 
 month(substr(datetime, 0, 10))), day(datetime), month(datetime) FROM 
 my_datamart LIMIT 1;
OK
{dayofmonth:14,monthofyear:4}   14  4

I havn't tried all the other date functions, but this is happens with  
weekofyear as well.  Looks like a caching issue...


 Multiple Date Functions inside Named Struct return the wrong values
 ---

 Key: HIVE-10366
 URL: https://issues.apache.org/jira/browse/HIVE-10366
 Project: Hive
  Issue Type: Bug
  Components: Hive
Affects Versions: 0.14.0
Reporter: Matt Traynham

 Using multiple date functions inside a named struct results in incorrect 
 values.  The first date function is correct.  All subsequent date functions 
 receive the value of the first date function, unless the function has a 
 different valued argument.
 As shown below, the 3rd query returns an incorrect value for 
 `month(datetime)`.  A not so pleasant workaround is to substr datetime 
 resulting in a different value (4th query).
  DESCRIBE my_datamart;
 OK
 datetime  string
  SELECT datetime FROM my_datamart LIMIT 1;
 OK
 2015-04-14T00:00:00
  SELECT named_struct('dayofmonth', day(datetime), 'monthofyear', 
  month(datetime)), day(datetime), month(datetime) FROM my_datamart LIMIT 1;
 OK
 {dayofmonth:14,monthofyear:14} 14 4
  SELECT named_struct('dayofmonth', day(datetime), 'monthofyear', 
  month(substr(datetime, 0, 10))), day(datetime), month(datetime) FROM 
  my_datamart LIMIT 1;
 OK
 {dayofmonth:14,monthofyear:4} 14  4
 I havn't tried all the other date functions, but this is happens with  
 weekofyear as well.  Looks like a caching issue...



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-10059) Make udaf_percentile_approx_23.q test more stable

2015-04-16 Thread Aihua Xu (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-10059?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14498359#comment-14498359
 ] 

Aihua Xu commented on HIVE-10059:
-

There is a failure to have a different value.
 342.0
---
 341.5

 Make udaf_percentile_approx_23.q test more stable
 -

 Key: HIVE-10059
 URL: https://issues.apache.org/jira/browse/HIVE-10059
 Project: Hive
  Issue Type: Bug
  Components: Tests
Reporter: Alexander Pivovarov
Assignee: Alexander Pivovarov
 Fix For: 1.2.0

 Attachments: HIVE-10059.1.patch


 I think the result of the test is not deterministic
 Sometimes line 626 in the output shows 256.0 sometimes 255.5
 Both results are correct because this UDAF returns approximate result.
 Probably results in the range of 255-257 should be considered as correct
  



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-10346) Tez on HBase has problems with settings again

2015-04-16 Thread Gunther Hagleitner (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-10346?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14498375#comment-14498375
 ] 

Gunther Hagleitner commented on HIVE-10346:
---

LGTM +1

Things you might want to consider: Do you want to do the logging in the second 
code patch (checkOutputSpec) on failure too? Since you're creating the format 
at init time now, do we still need the checkOutputSpecs at all?

 Tez on HBase has problems with settings again
 -

 Key: HIVE-10346
 URL: https://issues.apache.org/jira/browse/HIVE-10346
 Project: Hive
  Issue Type: Bug
Reporter: Sergey Shelukhin
Assignee: Sergey Shelukhin
 Attachments: HIVE-10346.patch






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-10367) org.apache.hive.hcatalog.pig.TestHCatLoaderEncryption.* fails in Windows

2015-04-16 Thread Sushanth Sowmyan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-10367?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sushanth Sowmyan updated HIVE-10367:

Affects Version/s: 1.2.0

 org.apache.hive.hcatalog.pig.TestHCatLoaderEncryption.* fails in Windows
 

 Key: HIVE-10367
 URL: https://issues.apache.org/jira/browse/HIVE-10367
 Project: Hive
  Issue Type: Bug
Affects Versions: 1.2.0
Reporter: Hari Sankar Sivarama Subramaniyan
Assignee: Hari Sankar Sivarama Subramaniyan
 Attachments: HIVE-10367.1.patch


 {code}
 java.io.IOException: Failed to execute create table encrypted_table(a int, b 
 string) location 
 'D:/w/hv/hcatalog/hcatalog-pig-adapter/target/tmp/org.apache.hive.hcatalog.pig.TestHCatLoader-1429172673960/warehouse/encryptedTable'.
  Driver returned 1 Error: FAILED: Execution Error, return code 1 from 
 org.apache.hadoop.hive.ql.exec.DDLTask. 
 MetaException(message:java.lang.IllegalArgumentException: Pathname 
 /D:/w/hv/hcatalog/hcatalog-pig-adapter/target/tmp/org.apache.hive.hcatalog.pig.TestHCatLoader-1429172673960/warehouse/encryptedTable
  from 
 hdfs://127.0.0.1:64735/D:/w/hv/hcatalog/hcatalog-pig-adapter/target/tmp/org.apache.hive.hcatalog.pig.TestHCatLoader-1429172673960/warehouse/encryptedTable
  is not a valid DFS filename.)
   at 
 org.apache.hive.hcatalog.pig.TestHCatLoaderEncryption.executeStatementOnDriver(TestHCatLoaderEncryption.java:163)
   at 
 org.apache.hive.hcatalog.pig.TestHCatLoaderEncryption.createTableInSpecifiedPath(TestHCatLoaderEncryption.java:423)
   at 
 org.apache.hive.hcatalog.pig.TestHCatLoaderEncryption.setup(TestHCatLoaderEncryption.java:197)
 {code}
 It looks like the error is happening because of the below query:
 create table encrypted_table(a int, b string) location 
 'D:/w/hv/hcatalog/hcatalog-pig-adapter/target/tmp/org.apache.hive.hcatalog.pig.TestHCatLoader-1429056939184/warehouse/encryptedTable'
 This gets called from createTableInSpecifiedPath() in 
 TestHCatLoaderEncryption. I think  an invalid HDFS path is passed here when 
 we are running the tests on Windows OS, so in line 197, changing
 {code}
 createTableInSpecifiedPath(ENCRYPTED_TABLE, a int, b string, 
 encryptedTablePath, driver);
 {code}
 to the below line might resolve the issue.
 {code}
 createTableInSpecifiedPath(ENCRYPTED_TABLE, a int, b string, 
 WindowsPathUtil.getHdfsUriString(encryptedTablePath), driver);
 {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-10368) VectorExpressionWriter doesn't match vectorColumn during row spilling in HybridGraceHashJoin

2015-04-16 Thread Wei Zheng (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-10368?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Wei Zheng updated HIVE-10368:
-
Description: 
This problem was exposed by HIVE-10284, when testing vectorized_context.q

Below is the query and backtrace:
{code}
select store.s_city, ss_net_profit
from store_sales
JOIN store ON store_sales.ss_store_sk = store.s_store_sk
JOIN household_demographics ON store_sales.ss_hdemo_sk = 
household_demographics.hd_demo_sk
limit 100
{code}
Caused by: java.lang.ClassCastException: 
org.apache.hadoop.hive.ql.exec.vector.BytesColumnVector cannot be cast to 
org.apache.hadoop.hive.ql.exec.vector.LongColumnVector
at 
org.apache.hadoop.hive.ql.exec.vector.expressions.VectorExpressionWriterFactory$VectorExpressionWriterLong.writeValue(VectorExpressionWriterFactory.java:175)
at 
org.apache.hadoop.hive.ql.exec.vector.VectorMapJoinOperator.getRowObject(VectorMapJoinOperator.java:347)
at 
org.apache.hadoop.hive.ql.exec.vector.VectorMapJoinOperator.spillBigTableRow(VectorMapJoinOperator.java:306)
at 
org.apache.hadoop.hive.ql.exec.MapJoinOperator.process(MapJoinOperator.java:390)
... 24 more

  was:
This problem was exposed by HIVE-10284, when testing vectorized_context

Below is the query and backtrace:
{code}
select store.s_city, ss_net_profit
from store_sales
JOIN store ON store_sales.ss_store_sk = store.s_store_sk
JOIN household_demographics ON store_sales.ss_hdemo_sk = 
household_demographics.hd_demo_sk
limit 100
{code}
Caused by: java.lang.ClassCastException: 
org.apache.hadoop.hive.ql.exec.vector.BytesColumnVector cannot be cast to 
org.apache.hadoop.hive.ql.exec.vector.LongColumnVector
at 
org.apache.hadoop.hive.ql.exec.vector.expressions.VectorExpressionWriterFactory$VectorExpressionWriterLong.writeValue(VectorExpressionWriterFactory.java:175)
at 
org.apache.hadoop.hive.ql.exec.vector.VectorMapJoinOperator.getRowObject(VectorMapJoinOperator.java:347)
at 
org.apache.hadoop.hive.ql.exec.vector.VectorMapJoinOperator.spillBigTableRow(VectorMapJoinOperator.java:306)
at 
org.apache.hadoop.hive.ql.exec.MapJoinOperator.process(MapJoinOperator.java:390)
... 24 more


 VectorExpressionWriter doesn't match vectorColumn during row spilling in 
 HybridGraceHashJoin
 

 Key: HIVE-10368
 URL: https://issues.apache.org/jira/browse/HIVE-10368
 Project: Hive
  Issue Type: Bug
Affects Versions: 1.2.0
Reporter: Wei Zheng
Assignee: Wei Zheng

 This problem was exposed by HIVE-10284, when testing vectorized_context.q
 Below is the query and backtrace:
 {code}
 select store.s_city, ss_net_profit
 from store_sales
 JOIN store ON store_sales.ss_store_sk = store.s_store_sk
 JOIN household_demographics ON store_sales.ss_hdemo_sk = 
 household_demographics.hd_demo_sk
 limit 100
 {code}
 Caused by: java.lang.ClassCastException: 
 org.apache.hadoop.hive.ql.exec.vector.BytesColumnVector cannot be cast to 
 org.apache.hadoop.hive.ql.exec.vector.LongColumnVector
   at 
 org.apache.hadoop.hive.ql.exec.vector.expressions.VectorExpressionWriterFactory$VectorExpressionWriterLong.writeValue(VectorExpressionWriterFactory.java:175)
   at 
 org.apache.hadoop.hive.ql.exec.vector.VectorMapJoinOperator.getRowObject(VectorMapJoinOperator.java:347)
   at 
 org.apache.hadoop.hive.ql.exec.vector.VectorMapJoinOperator.spillBigTableRow(VectorMapJoinOperator.java:306)
   at 
 org.apache.hadoop.hive.ql.exec.MapJoinOperator.process(MapJoinOperator.java:390)
   ... 24 more



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-10284) enable container reuse for grace hash join

2015-04-16 Thread Wei Zheng (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-10284?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14498665#comment-14498665
 ] 

Wei Zheng commented on HIVE-10284:
--

HIVE-10368 has been created for this specific issue.

 enable container reuse for grace hash join 
 ---

 Key: HIVE-10284
 URL: https://issues.apache.org/jira/browse/HIVE-10284
 Project: Hive
  Issue Type: Bug
Reporter: Gunther Hagleitner
Assignee: Wei Zheng
 Attachments: HIVE-10284.1.patch, HIVE-10284.2.patch, 
 HIVE-10284.3.patch, HIVE-10284.4.patch, HIVE-10284.5.patch, 
 HIVE-10284.6.patch, HIVE-10284.7.patch, HIVE-10284.8.patch






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-10367) org.apache.hive.hcatalog.pig.TestHCatLoaderEncryption.* fails in Windows

2015-04-16 Thread Hari Sankar Sivarama Subramaniyan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-10367?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Hari Sankar Sivarama Subramaniyan updated HIVE-10367:
-
Description: 
java.io.IOException: Failed to execute create table encrypted_table(a int, b 
string) location 
'D:/w/hv/hcatalog/hcatalog-pig-adapter/target/tmp/org.apache.hive.hcatalog.pig.TestHCatLoader-1429172673960/warehouse/encryptedTable'.
 Driver returned 1 Error: FAILED: Execution Error, return code 1 from 
org.apache.hadoop.hive.ql.exec.DDLTask. 
MetaException(message:java.lang.IllegalArgumentException: Pathname 
/D:/w/hv/hcatalog/hcatalog-pig-adapter/target/tmp/org.apache.hive.hcatalog.pig.TestHCatLoader-1429172673960/warehouse/encryptedTable
 from 
hdfs://127.0.0.1:64735/D:/w/hv/hcatalog/hcatalog-pig-adapter/target/tmp/org.apache.hive.hcatalog.pig.TestHCatLoader-1429172673960/warehouse/encryptedTable
 is not a valid DFS filename.)
at 
org.apache.hive.hcatalog.pig.TestHCatLoaderEncryption.executeStatementOnDriver(TestHCatLoaderEncryption.java:163)
at 
org.apache.hive.hcatalog.pig.TestHCatLoaderEncryption.createTableInSpecifiedPath(TestHCatLoaderEncryption.java:423)
at 
org.apache.hive.hcatalog.pig.TestHCatLoaderEncryption.setup(TestHCatLoaderEncryption.java:197)


It looks like the error is happening because of the below query:
create table encrypted_table(a int, b string) location 
'D:/w/hv/hcatalog/hcatalog-pig-adapter/target/tmp/org.apache.hive.hcatalog.pig.TestHCatLoader-1429056939184/warehouse/encryptedTable'
This gets called from createTableInSpecifiedPath() in TestHCatLoaderEncryption. 
I think  an invalid HDFS path is passed here when we are running the tests on 
Windows OS, so in line 197, changing
createTableInSpecifiedPath(ENCRYPTED_TABLE, a int, b string, 
encryptedTablePath, driver);
to the below line might resolve the issue.
createTableInSpecifiedPath(ENCRYPTED_TABLE, a int, b string, 
WindowsPathUtil.getHdfsUriString(encryptedTablePath), driver);
Thanks
Hari

  was:
java.io.IOException: Failed to execute create table encrypted_table(a int, b 
string) location 
'D:/w/hv/hcatalog/hcatalog-pig-adapter/target/tmp/org.apache.hive.hcatalog.pig.TestHCatLoader-1429172673960/warehouse/encryptedTable'.
 Driver returned 1 Error: FAILED: Execution Error, return code 1 from 
org.apache.hadoop.hive.ql.exec.DDLTask. 
MetaException(message:java.lang.IllegalArgumentException: Pathname 
/D:/w/hv/hcatalog/hcatalog-pig-adapter/target/tmp/org.apache.hive.hcatalog.pig.TestHCatLoader-1429172673960/warehouse/encryptedTable
 from 
hdfs://127.0.0.1:64735/D:/w/hv/hcatalog/hcatalog-pig-adapter/target/tmp/org.apache.hive.hcatalog.pig.TestHCatLoader-1429172673960/warehouse/encryptedTable
 is not a valid DFS filename.)
at 
org.apache.hive.hcatalog.pig.TestHCatLoaderEncryption.executeStatementOnDriver(TestHCatLoaderEncryption.java:163)
at 
org.apache.hive.hcatalog.pig.TestHCatLoaderEncryption.createTableInSpecifiedPath(TestHCatLoaderEncryption.java:423)
at 
org.apache.hive.hcatalog.pig.TestHCatLoaderEncryption.setup(TestHCatLoaderEncryption.java:197)


 org.apache.hive.hcatalog.pig.TestHCatLoaderEncryption.* fails in Windows
 

 Key: HIVE-10367
 URL: https://issues.apache.org/jira/browse/HIVE-10367
 Project: Hive
  Issue Type: Bug
Reporter: Hari Sankar Sivarama Subramaniyan
Assignee: Hari Sankar Sivarama Subramaniyan

 java.io.IOException: Failed to execute create table encrypted_table(a int, b 
 string) location 
 'D:/w/hv/hcatalog/hcatalog-pig-adapter/target/tmp/org.apache.hive.hcatalog.pig.TestHCatLoader-1429172673960/warehouse/encryptedTable'.
  Driver returned 1 Error: FAILED: Execution Error, return code 1 from 
 org.apache.hadoop.hive.ql.exec.DDLTask. 
 MetaException(message:java.lang.IllegalArgumentException: Pathname 
 /D:/w/hv/hcatalog/hcatalog-pig-adapter/target/tmp/org.apache.hive.hcatalog.pig.TestHCatLoader-1429172673960/warehouse/encryptedTable
  from 
 hdfs://127.0.0.1:64735/D:/w/hv/hcatalog/hcatalog-pig-adapter/target/tmp/org.apache.hive.hcatalog.pig.TestHCatLoader-1429172673960/warehouse/encryptedTable
  is not a valid DFS filename.)
   at 
 org.apache.hive.hcatalog.pig.TestHCatLoaderEncryption.executeStatementOnDriver(TestHCatLoaderEncryption.java:163)
   at 
 org.apache.hive.hcatalog.pig.TestHCatLoaderEncryption.createTableInSpecifiedPath(TestHCatLoaderEncryption.java:423)
   at 
 org.apache.hive.hcatalog.pig.TestHCatLoaderEncryption.setup(TestHCatLoaderEncryption.java:197)
 It looks like the error is happening because of the below query:
 create table encrypted_table(a int, b string) location 
 'D:/w/hv/hcatalog/hcatalog-pig-adapter/target/tmp/org.apache.hive.hcatalog.pig.TestHCatLoader-1429056939184/warehouse/encryptedTable'
 This gets called from 

[jira] [Updated] (HIVE-10367) org.apache.hive.hcatalog.pig.TestHCatLoaderEncryption.* fails in Windows

2015-04-16 Thread Hari Sankar Sivarama Subramaniyan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-10367?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Hari Sankar Sivarama Subramaniyan updated HIVE-10367:
-
Description: 
{code}
java.io.IOException: Failed to execute create table encrypted_table(a int, b 
string) location 
'D:/w/hv/hcatalog/hcatalog-pig-adapter/target/tmp/org.apache.hive.hcatalog.pig.TestHCatLoader-1429172673960/warehouse/encryptedTable'.
 Driver returned 1 Error: FAILED: Execution Error, return code 1 from 
org.apache.hadoop.hive.ql.exec.DDLTask. 
MetaException(message:java.lang.IllegalArgumentException: Pathname 
/D:/w/hv/hcatalog/hcatalog-pig-adapter/target/tmp/org.apache.hive.hcatalog.pig.TestHCatLoader-1429172673960/warehouse/encryptedTable
 from 
hdfs://127.0.0.1:64735/D:/w/hv/hcatalog/hcatalog-pig-adapter/target/tmp/org.apache.hive.hcatalog.pig.TestHCatLoader-1429172673960/warehouse/encryptedTable
 is not a valid DFS filename.)
at 
org.apache.hive.hcatalog.pig.TestHCatLoaderEncryption.executeStatementOnDriver(TestHCatLoaderEncryption.java:163)
at 
org.apache.hive.hcatalog.pig.TestHCatLoaderEncryption.createTableInSpecifiedPath(TestHCatLoaderEncryption.java:423)
at 
org.apache.hive.hcatalog.pig.TestHCatLoaderEncryption.setup(TestHCatLoaderEncryption.java:197)
{code}

It looks like the error is happening because of the below query:
create table encrypted_table(a int, b string) location 
'D:/w/hv/hcatalog/hcatalog-pig-adapter/target/tmp/org.apache.hive.hcatalog.pig.TestHCatLoader-1429056939184/warehouse/encryptedTable'
This gets called from createTableInSpecifiedPath() in TestHCatLoaderEncryption. 
I think  an invalid HDFS path is passed here when we are running the tests on 
Windows OS, so in line 197, changing
{code}
createTableInSpecifiedPath(ENCRYPTED_TABLE, a int, b string, 
encryptedTablePath, driver);
{code}
to the below line might resolve the issue.
{code}
createTableInSpecifiedPath(ENCRYPTED_TABLE, a int, b string, 
WindowsPathUtil.getHdfsUriString(encryptedTablePath), driver);
{code}


  was:
java.io.IOException: Failed to execute create table encrypted_table(a int, b 
string) location 
'D:/w/hv/hcatalog/hcatalog-pig-adapter/target/tmp/org.apache.hive.hcatalog.pig.TestHCatLoader-1429172673960/warehouse/encryptedTable'.
 Driver returned 1 Error: FAILED: Execution Error, return code 1 from 
org.apache.hadoop.hive.ql.exec.DDLTask. 
MetaException(message:java.lang.IllegalArgumentException: Pathname 
/D:/w/hv/hcatalog/hcatalog-pig-adapter/target/tmp/org.apache.hive.hcatalog.pig.TestHCatLoader-1429172673960/warehouse/encryptedTable
 from 
hdfs://127.0.0.1:64735/D:/w/hv/hcatalog/hcatalog-pig-adapter/target/tmp/org.apache.hive.hcatalog.pig.TestHCatLoader-1429172673960/warehouse/encryptedTable
 is not a valid DFS filename.)
at 
org.apache.hive.hcatalog.pig.TestHCatLoaderEncryption.executeStatementOnDriver(TestHCatLoaderEncryption.java:163)
at 
org.apache.hive.hcatalog.pig.TestHCatLoaderEncryption.createTableInSpecifiedPath(TestHCatLoaderEncryption.java:423)
at 
org.apache.hive.hcatalog.pig.TestHCatLoaderEncryption.setup(TestHCatLoaderEncryption.java:197)


It looks like the error is happening because of the below query:
create table encrypted_table(a int, b string) location 
'D:/w/hv/hcatalog/hcatalog-pig-adapter/target/tmp/org.apache.hive.hcatalog.pig.TestHCatLoader-1429056939184/warehouse/encryptedTable'
This gets called from createTableInSpecifiedPath() in TestHCatLoaderEncryption. 
I think  an invalid HDFS path is passed here when we are running the tests on 
Windows OS, so in line 197, changing
createTableInSpecifiedPath(ENCRYPTED_TABLE, a int, b string, 
encryptedTablePath, driver);
to the below line might resolve the issue.
createTableInSpecifiedPath(ENCRYPTED_TABLE, a int, b string, 
WindowsPathUtil.getHdfsUriString(encryptedTablePath), driver);
Thanks
Hari


 org.apache.hive.hcatalog.pig.TestHCatLoaderEncryption.* fails in Windows
 

 Key: HIVE-10367
 URL: https://issues.apache.org/jira/browse/HIVE-10367
 Project: Hive
  Issue Type: Bug
Reporter: Hari Sankar Sivarama Subramaniyan
Assignee: Hari Sankar Sivarama Subramaniyan

 {code}
 java.io.IOException: Failed to execute create table encrypted_table(a int, b 
 string) location 
 'D:/w/hv/hcatalog/hcatalog-pig-adapter/target/tmp/org.apache.hive.hcatalog.pig.TestHCatLoader-1429172673960/warehouse/encryptedTable'.
  Driver returned 1 Error: FAILED: Execution Error, return code 1 from 
 org.apache.hadoop.hive.ql.exec.DDLTask. 
 MetaException(message:java.lang.IllegalArgumentException: Pathname 
 /D:/w/hv/hcatalog/hcatalog-pig-adapter/target/tmp/org.apache.hive.hcatalog.pig.TestHCatLoader-1429172673960/warehouse/encryptedTable
  from 
 

[jira] [Commented] (HIVE-10029) LLAP: Scheduling of work from different queries within the daemon

2015-04-16 Thread Siddharth Seth (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-10029?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14498716#comment-14498716
 ] 

Siddharth Seth commented on HIVE-10029:
---

Yes, to the most part. We'll likely a need a follow up to provide data to the 
pre-emption queue.

 LLAP: Scheduling of work from different queries within the daemon
 -

 Key: HIVE-10029
 URL: https://issues.apache.org/jira/browse/HIVE-10029
 Project: Hive
  Issue Type: Sub-task
Reporter: Siddharth Seth
 Fix For: llap


 The current implementation is a simple queue - whichever query wins the race 
 to submit work to a daemon will execute first.
 A policy around this may be useful - potentially a fair share, or a first 
 query in gets all slots approach.
 Also, prioritiy associated with work within a query should be considered.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-10303) HIVE-9471 broke forward compatibility of ORC files

2015-04-16 Thread Prasanth Jayachandran (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-10303?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14498816#comment-14498816
 ] 

Prasanth Jayachandran commented on HIVE-10303:
--

[~owen.omalley]/[~gopalv].. ping!

 HIVE-9471 broke forward compatibility of ORC files
 --

 Key: HIVE-10303
 URL: https://issues.apache.org/jira/browse/HIVE-10303
 Project: Hive
  Issue Type: Bug
  Components: File Formats
Affects Versions: 1.2.0
Reporter: Owen O'Malley
Assignee: Prasanth Jayachandran
 Fix For: 1.2.0

 Attachments: HIVE-10303.1.patch


 The change suppresses the streams in ORC files for ORC dictionaries with 0 
 entries. This causes NPE on ORC readers for all versions of Hive 0.11 to 1.1 
 and needs to be reverted.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-10222) Upgrade Calcite dependency to newest version

2015-04-16 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-10222?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-10222:

Attachment: HIVE-10222.04.patch

 Upgrade Calcite dependency to newest version
 

 Key: HIVE-10222
 URL: https://issues.apache.org/jira/browse/HIVE-10222
 Project: Hive
  Issue Type: Bug
Reporter: Jesus Camacho Rodriguez
Assignee: Jesus Camacho Rodriguez
 Attachments: HIVE-10222.01.patch, HIVE-10222.02.patch, 
 HIVE-10222.03.patch, HIVE-10222.04.patch, HIVE-10222.patch


  Upgrade Calcite version to 1.2.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-10367) org.apache.hive.hcatalog.pig.TestHCatLoaderEncryption.* fails in Windows

2015-04-16 Thread Sushanth Sowmyan (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-10367?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14498588#comment-14498588
 ] 

Sushanth Sowmyan commented on HIVE-10367:
-

+1

 org.apache.hive.hcatalog.pig.TestHCatLoaderEncryption.* fails in Windows
 

 Key: HIVE-10367
 URL: https://issues.apache.org/jira/browse/HIVE-10367
 Project: Hive
  Issue Type: Bug
Reporter: Hari Sankar Sivarama Subramaniyan
Assignee: Hari Sankar Sivarama Subramaniyan
 Attachments: HIVE-10367.1.patch


 {code}
 java.io.IOException: Failed to execute create table encrypted_table(a int, b 
 string) location 
 'D:/w/hv/hcatalog/hcatalog-pig-adapter/target/tmp/org.apache.hive.hcatalog.pig.TestHCatLoader-1429172673960/warehouse/encryptedTable'.
  Driver returned 1 Error: FAILED: Execution Error, return code 1 from 
 org.apache.hadoop.hive.ql.exec.DDLTask. 
 MetaException(message:java.lang.IllegalArgumentException: Pathname 
 /D:/w/hv/hcatalog/hcatalog-pig-adapter/target/tmp/org.apache.hive.hcatalog.pig.TestHCatLoader-1429172673960/warehouse/encryptedTable
  from 
 hdfs://127.0.0.1:64735/D:/w/hv/hcatalog/hcatalog-pig-adapter/target/tmp/org.apache.hive.hcatalog.pig.TestHCatLoader-1429172673960/warehouse/encryptedTable
  is not a valid DFS filename.)
   at 
 org.apache.hive.hcatalog.pig.TestHCatLoaderEncryption.executeStatementOnDriver(TestHCatLoaderEncryption.java:163)
   at 
 org.apache.hive.hcatalog.pig.TestHCatLoaderEncryption.createTableInSpecifiedPath(TestHCatLoaderEncryption.java:423)
   at 
 org.apache.hive.hcatalog.pig.TestHCatLoaderEncryption.setup(TestHCatLoaderEncryption.java:197)
 {code}
 It looks like the error is happening because of the below query:
 create table encrypted_table(a int, b string) location 
 'D:/w/hv/hcatalog/hcatalog-pig-adapter/target/tmp/org.apache.hive.hcatalog.pig.TestHCatLoader-1429056939184/warehouse/encryptedTable'
 This gets called from createTableInSpecifiedPath() in 
 TestHCatLoaderEncryption. I think  an invalid HDFS path is passed here when 
 we are running the tests on Windows OS, so in line 197, changing
 {code}
 createTableInSpecifiedPath(ENCRYPTED_TABLE, a int, b string, 
 encryptedTablePath, driver);
 {code}
 to the below line might resolve the issue.
 {code}
 createTableInSpecifiedPath(ENCRYPTED_TABLE, a int, b string, 
 WindowsPathUtil.getHdfsUriString(encryptedTablePath), driver);
 {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-10366) Multiple Date Functions inside Named Struct return the wrong values

2015-04-16 Thread Matt Traynham (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-10366?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14498696#comment-14498696
 ] 

Matt Traynham commented on HIVE-10366:
--

Possible dupe of https://issues.apache.org/jira/browse/HIVE-9278.  Took a while 
skimming to find it.

 Multiple Date Functions inside Named Struct return the wrong values
 ---

 Key: HIVE-10366
 URL: https://issues.apache.org/jira/browse/HIVE-10366
 Project: Hive
  Issue Type: Bug
  Components: Hive
Affects Versions: 0.14.0
Reporter: Matt Traynham

 Using multiple date functions inside a named struct results in incorrect 
 values.  The first date function is correct.  All subsequent date functions 
 receive the value of the first date function, unless the function has a 
 different valued argument.
 As shown below, the 3rd query returns an incorrect value for 
 `month(datetime)`.  A not so pleasant workaround is to substr datetime 
 resulting in a different value (4th query).
  DESCRIBE my_datamart;
 OK
 datetime  string
  SELECT datetime FROM my_datamart LIMIT 1;
 OK
 2015-04-14T00:00:00
  SELECT named_struct('dayofmonth', day(datetime), 'monthofyear', 
  month(datetime)), day(datetime), month(datetime) FROM my_datamart LIMIT 1;
 OK
 {dayofmonth:14,monthofyear:14} 14 4
  SELECT named_struct('dayofmonth', day(datetime), 'monthofyear', 
  month(substr(datetime, 0, 10))), day(datetime), month(datetime) FROM 
  my_datamart LIMIT 1;
 OK
 {dayofmonth:14,monthofyear:4} 14  4
 I havn't tried all the other date functions, but this is happens with  
 weekofyear as well.  Looks like a caching issue...



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-10239) Create scripts to do metastore upgrade tests on jenkins for Derby, Oracle and PostgreSQL

2015-04-16 Thread Naveen Gangam (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-10239?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Naveen Gangam updated HIVE-10239:
-
Attachment: HIVE-10239.00.patch

This appears to also be an issue for postgreSQL scripts. Appears that this is 
an issue when wget gets redirected from http to https URL.

I have added a check to ensure that the apt-transport-https package gets 
installed before installing the corresponding DB packages. Thanks

 Create scripts to do metastore upgrade tests on jenkins for Derby, Oracle and 
 PostgreSQL
 

 Key: HIVE-10239
 URL: https://issues.apache.org/jira/browse/HIVE-10239
 Project: Hive
  Issue Type: Improvement
Affects Versions: 1.1.0
Reporter: Naveen Gangam
Assignee: Naveen Gangam
 Attachments: HIVE-10239-donotcommit.patch, HIVE-10239.0.patch, 
 HIVE-10239.0.patch, HIVE-10239.00.patch, HIVE-10239.patch


 Need to create DB-implementation specific scripts to use the framework 
 introduced in HIVE-9800 to have any metastore schema changes tested across 
 all supported databases.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-10331) ORC : Is null SARG filters out all row groups written in old ORC format

2015-04-16 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-10331?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14498648#comment-14498648
 ] 

Hive QA commented on HIVE-10331:




{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12725797/HIVE-10331.04.patch

{color:red}ERROR:{color} -1 due to 36 failed/errored test(s), 8711 tests 
executed
*Failed tests:*
{noformat}
TestMinimrCliDriver-bucketmapjoin6.q-constprog_partitioner.q-infer_bucket_sort_dyn_part.q-and-1-more
 - did not produce a TEST-*.xml file
TestMinimrCliDriver-external_table_with_space_in_location_path.q-infer_bucket_sort_merge.q-auto_sortmerge_join_16.q-and-1-more
 - did not produce a TEST-*.xml file
TestMinimrCliDriver-groupby2.q-import_exported_table.q-bucketizedhiveinputformat.q-and-1-more
 - did not produce a TEST-*.xml file
TestMinimrCliDriver-index_bitmap3.q-stats_counter_partitioned.q-temp_table_external.q-and-1-more
 - did not produce a TEST-*.xml file
TestMinimrCliDriver-infer_bucket_sort_map_operators.q-join1.q-bucketmapjoin7.q-and-1-more
 - did not produce a TEST-*.xml file
TestMinimrCliDriver-infer_bucket_sort_num_buckets.q-disable_merge_for_bucketing.q-uber_reduce.q-and-1-more
 - did not produce a TEST-*.xml file
TestMinimrCliDriver-infer_bucket_sort_reducers_power_two.q-scriptfile1.q-scriptfile1_win.q-and-1-more
 - did not produce a TEST-*.xml file
TestMinimrCliDriver-leftsemijoin_mr.q-load_hdfs_file_with_space_in_the_name.q-root_dir_external_table.q-and-1-more
 - did not produce a TEST-*.xml file
TestMinimrCliDriver-list_bucket_dml_10.q-bucket_num_reducers.q-bucket6.q-and-1-more
 - did not produce a TEST-*.xml file
TestMinimrCliDriver-load_fs2.q-file_with_header_footer.q-ql_rewrite_gbtoidx_cbo_1.q-and-1-more
 - did not produce a TEST-*.xml file
TestMinimrCliDriver-parallel_orderby.q-reduce_deduplicate.q-ql_rewrite_gbtoidx_cbo_2.q-and-1-more
 - did not produce a TEST-*.xml file
TestMinimrCliDriver-ql_rewrite_gbtoidx.q-smb_mapjoin_8.q - did not produce a 
TEST-*.xml file
TestMinimrCliDriver-schemeAuthority2.q-bucket4.q-input16_cc.q-and-1-more - did 
not produce a TEST-*.xml file
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_union_view
org.apache.hadoop.hive.ql.io.orc.TestOrcFile.testReadFormat_0_11[0]
org.apache.hadoop.hive.ql.io.orc.TestOrcFile.testReadFormat_0_11[1]
org.apache.hadoop.hive.ql.io.orc.TestOrcSerDeStats.testSerdeStatsOldFormat
org.apache.hadoop.hive.ql.io.orc.TestRecordReaderImpl.testBetween
org.apache.hadoop.hive.ql.io.orc.TestRecordReaderImpl.testDateWritableEqualsBloomFilter
org.apache.hadoop.hive.ql.io.orc.TestRecordReaderImpl.testDateWritableInBloomFilter
org.apache.hadoop.hive.ql.io.orc.TestRecordReaderImpl.testDecimalEqualsBloomFilter
org.apache.hadoop.hive.ql.io.orc.TestRecordReaderImpl.testDecimalInBloomFilter
org.apache.hadoop.hive.ql.io.orc.TestRecordReaderImpl.testDoubleEqualsBloomFilter
org.apache.hadoop.hive.ql.io.orc.TestRecordReaderImpl.testDoubleInBloomFilter
org.apache.hadoop.hive.ql.io.orc.TestRecordReaderImpl.testEquals
org.apache.hadoop.hive.ql.io.orc.TestRecordReaderImpl.testIn
org.apache.hadoop.hive.ql.io.orc.TestRecordReaderImpl.testIntEqualsBloomFilter
org.apache.hadoop.hive.ql.io.orc.TestRecordReaderImpl.testIntInBloomFilter
org.apache.hadoop.hive.ql.io.orc.TestRecordReaderImpl.testIsNull
org.apache.hadoop.hive.ql.io.orc.TestRecordReaderImpl.testLessThan
org.apache.hadoop.hive.ql.io.orc.TestRecordReaderImpl.testLessThanEquals
org.apache.hadoop.hive.ql.io.orc.TestRecordReaderImpl.testNullsInBloomFilter
org.apache.hadoop.hive.ql.io.orc.TestRecordReaderImpl.testStringEqualsBloomFilter
org.apache.hadoop.hive.ql.io.orc.TestRecordReaderImpl.testStringInBloomFilter
org.apache.hadoop.hive.ql.io.orc.TestRecordReaderImpl.testTimestampEqualsBloomFilter
org.apache.hadoop.hive.ql.io.orc.TestRecordReaderImpl.testTimestampInBloomFilter
{noformat}

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/3461/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/3461/console
Test logs: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-3461/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 36 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12725797 - PreCommit-HIVE-TRUNK-Build

 ORC : Is null SARG filters out all row groups written in old ORC format
 ---

 Key: HIVE-10331
 URL: https://issues.apache.org/jira/browse/HIVE-10331
 Project: Hive
  Issue Type: Bug
 

[jira] [Commented] (HIVE-10222) Upgrade Calcite dependency to newest version

2015-04-16 Thread Ashutosh Chauhan (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-10222?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14498503#comment-14498503
 ] 

Ashutosh Chauhan commented on HIVE-10222:
-

+1

 Upgrade Calcite dependency to newest version
 

 Key: HIVE-10222
 URL: https://issues.apache.org/jira/browse/HIVE-10222
 Project: Hive
  Issue Type: Bug
Reporter: Jesus Camacho Rodriguez
Assignee: Jesus Camacho Rodriguez
 Attachments: HIVE-10222.01.patch, HIVE-10222.02.patch, 
 HIVE-10222.03.patch, HIVE-10222.patch


  Upgrade Calcite version to 1.2.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-10324) Hive metatool should take table_param_key to allow for changes to avro serde's schema url key

2015-04-16 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-10324?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14498516#comment-14498516
 ] 

Hive QA commented on HIVE-10324:




{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12725773/HIVE-10324.1.patch

{color:red}ERROR:{color} -1 due to 15 failed/errored test(s), 8702 tests 
executed
*Failed tests:*
{noformat}
TestMinimrCliDriver-bucketmapjoin6.q-constprog_partitioner.q-infer_bucket_sort_dyn_part.q-and-1-more
 - did not produce a TEST-*.xml file
TestMinimrCliDriver-external_table_with_space_in_location_path.q-infer_bucket_sort_merge.q-auto_sortmerge_join_16.q-and-1-more
 - did not produce a TEST-*.xml file
TestMinimrCliDriver-groupby2.q-import_exported_table.q-bucketizedhiveinputformat.q-and-1-more
 - did not produce a TEST-*.xml file
TestMinimrCliDriver-index_bitmap3.q-stats_counter_partitioned.q-temp_table_external.q-and-1-more
 - did not produce a TEST-*.xml file
TestMinimrCliDriver-infer_bucket_sort_map_operators.q-join1.q-bucketmapjoin7.q-and-1-more
 - did not produce a TEST-*.xml file
TestMinimrCliDriver-infer_bucket_sort_num_buckets.q-disable_merge_for_bucketing.q-uber_reduce.q-and-1-more
 - did not produce a TEST-*.xml file
TestMinimrCliDriver-infer_bucket_sort_reducers_power_two.q-scriptfile1.q-scriptfile1_win.q-and-1-more
 - did not produce a TEST-*.xml file
TestMinimrCliDriver-leftsemijoin_mr.q-load_hdfs_file_with_space_in_the_name.q-root_dir_external_table.q-and-1-more
 - did not produce a TEST-*.xml file
TestMinimrCliDriver-list_bucket_dml_10.q-bucket_num_reducers.q-bucket6.q-and-1-more
 - did not produce a TEST-*.xml file
TestMinimrCliDriver-load_fs2.q-file_with_header_footer.q-ql_rewrite_gbtoidx_cbo_1.q-and-1-more
 - did not produce a TEST-*.xml file
TestMinimrCliDriver-parallel_orderby.q-reduce_deduplicate.q-ql_rewrite_gbtoidx_cbo_2.q-and-1-more
 - did not produce a TEST-*.xml file
TestMinimrCliDriver-ql_rewrite_gbtoidx.q-smb_mapjoin_8.q - did not produce a 
TEST-*.xml file
TestMinimrCliDriver-schemeAuthority2.q-bucket4.q-input16_cc.q-and-1-more - did 
not produce a TEST-*.xml file
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_union_view
org.apache.hive.jdbc.TestSSL.testSSLFetchHttp
{noformat}

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/3460/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/3460/console
Test logs: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-3460/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 15 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12725773 - PreCommit-HIVE-TRUNK-Build

 Hive metatool should take table_param_key to allow for changes to avro 
 serde's schema url key
 -

 Key: HIVE-10324
 URL: https://issues.apache.org/jira/browse/HIVE-10324
 Project: Hive
  Issue Type: Bug
  Components: Metastore
Affects Versions: 1.1.0
Reporter: Szehon Ho
Assignee: Ferdinand Xu
 Attachments: HIVE-10324.1.patch, HIVE-10324.patch, 
 HIVE-10324.patch.WIP


 HIVE-3443 added support to change the serdeParams from 'metatool 
 updateLocation' command.
 However, in avro it is possible to specify the schema via the tableParams:
 {noformat}
 CREATE  TABLE `testavro`(
   `test` string COMMENT 'from deserializer')
 ROW FORMAT SERDE 
   'org.apache.hadoop.hive.serde2.avro.AvroSerDe' 
 STORED AS INPUTFORMAT 
   'org.apache.hadoop.hive.ql.io.avro.AvroContainerInputFormat' 
 OUTPUTFORMAT 
   'org.apache.hadoop.hive.ql.io.avro.AvroContainerOutputFormat'
 TBLPROPERTIES (
   'avro.schema.url'='hdfs://namenode:8020/tmp/test.avsc', 
   'kite.compression.type'='snappy', 
   'transient_lastDdlTime'='1427996456')
 {noformat}
 Hence for those tables the 'metatool updateLocation' will not help.
 This is necessary in case like upgrade the namenode to HA where the absolute 
 paths have changed.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-10367) org.apache.hive.hcatalog.pig.TestHCatLoaderEncryption.* fails in Windows

2015-04-16 Thread Hari Sankar Sivarama Subramaniyan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-10367?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Hari Sankar Sivarama Subramaniyan updated HIVE-10367:
-
Attachment: HIVE-10367.1.patch

 org.apache.hive.hcatalog.pig.TestHCatLoaderEncryption.* fails in Windows
 

 Key: HIVE-10367
 URL: https://issues.apache.org/jira/browse/HIVE-10367
 Project: Hive
  Issue Type: Bug
Reporter: Hari Sankar Sivarama Subramaniyan
Assignee: Hari Sankar Sivarama Subramaniyan
 Attachments: HIVE-10367.1.patch


 {code}
 java.io.IOException: Failed to execute create table encrypted_table(a int, b 
 string) location 
 'D:/w/hv/hcatalog/hcatalog-pig-adapter/target/tmp/org.apache.hive.hcatalog.pig.TestHCatLoader-1429172673960/warehouse/encryptedTable'.
  Driver returned 1 Error: FAILED: Execution Error, return code 1 from 
 org.apache.hadoop.hive.ql.exec.DDLTask. 
 MetaException(message:java.lang.IllegalArgumentException: Pathname 
 /D:/w/hv/hcatalog/hcatalog-pig-adapter/target/tmp/org.apache.hive.hcatalog.pig.TestHCatLoader-1429172673960/warehouse/encryptedTable
  from 
 hdfs://127.0.0.1:64735/D:/w/hv/hcatalog/hcatalog-pig-adapter/target/tmp/org.apache.hive.hcatalog.pig.TestHCatLoader-1429172673960/warehouse/encryptedTable
  is not a valid DFS filename.)
   at 
 org.apache.hive.hcatalog.pig.TestHCatLoaderEncryption.executeStatementOnDriver(TestHCatLoaderEncryption.java:163)
   at 
 org.apache.hive.hcatalog.pig.TestHCatLoaderEncryption.createTableInSpecifiedPath(TestHCatLoaderEncryption.java:423)
   at 
 org.apache.hive.hcatalog.pig.TestHCatLoaderEncryption.setup(TestHCatLoaderEncryption.java:197)
 {code}
 It looks like the error is happening because of the below query:
 create table encrypted_table(a int, b string) location 
 'D:/w/hv/hcatalog/hcatalog-pig-adapter/target/tmp/org.apache.hive.hcatalog.pig.TestHCatLoader-1429056939184/warehouse/encryptedTable'
 This gets called from createTableInSpecifiedPath() in 
 TestHCatLoaderEncryption. I think  an invalid HDFS path is passed here when 
 we are running the tests on Windows OS, so in line 197, changing
 {code}
 createTableInSpecifiedPath(ENCRYPTED_TABLE, a int, b string, 
 encryptedTablePath, driver);
 {code}
 to the below line might resolve the issue.
 {code}
 createTableInSpecifiedPath(ENCRYPTED_TABLE, a int, b string, 
 WindowsPathUtil.getHdfsUriString(encryptedTablePath), driver);
 {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-7797) upgrade hive schema from 0.9.0 to 0.13.1 failed when back end db is postgres

2015-04-16 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-7797?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14498779#comment-14498779
 ] 

Hive QA commented on HIVE-7797:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12687447/HIVE-7797.1.patch

{color:red}ERROR:{color} -1 due to 16 failed/errored test(s), 8711 tests 
executed
*Failed tests:*
{noformat}
TestMinimrCliDriver-bucketmapjoin6.q-constprog_partitioner.q-infer_bucket_sort_dyn_part.q-and-1-more
 - did not produce a TEST-*.xml file
TestMinimrCliDriver-external_table_with_space_in_location_path.q-infer_bucket_sort_merge.q-auto_sortmerge_join_16.q-and-1-more
 - did not produce a TEST-*.xml file
TestMinimrCliDriver-groupby2.q-import_exported_table.q-bucketizedhiveinputformat.q-and-1-more
 - did not produce a TEST-*.xml file
TestMinimrCliDriver-index_bitmap3.q-stats_counter_partitioned.q-temp_table_external.q-and-1-more
 - did not produce a TEST-*.xml file
TestMinimrCliDriver-infer_bucket_sort_map_operators.q-join1.q-bucketmapjoin7.q-and-1-more
 - did not produce a TEST-*.xml file
TestMinimrCliDriver-infer_bucket_sort_num_buckets.q-disable_merge_for_bucketing.q-uber_reduce.q-and-1-more
 - did not produce a TEST-*.xml file
TestMinimrCliDriver-infer_bucket_sort_reducers_power_two.q-scriptfile1.q-scriptfile1_win.q-and-1-more
 - did not produce a TEST-*.xml file
TestMinimrCliDriver-leftsemijoin_mr.q-load_hdfs_file_with_space_in_the_name.q-root_dir_external_table.q-and-1-more
 - did not produce a TEST-*.xml file
TestMinimrCliDriver-list_bucket_dml_10.q-bucket_num_reducers.q-bucket6.q-and-1-more
 - did not produce a TEST-*.xml file
TestMinimrCliDriver-load_fs2.q-file_with_header_footer.q-ql_rewrite_gbtoidx_cbo_1.q-and-1-more
 - did not produce a TEST-*.xml file
TestMinimrCliDriver-parallel_orderby.q-reduce_deduplicate.q-ql_rewrite_gbtoidx_cbo_2.q-and-1-more
 - did not produce a TEST-*.xml file
TestMinimrCliDriver-ql_rewrite_gbtoidx.q-smb_mapjoin_8.q - did not produce a 
TEST-*.xml file
TestMinimrCliDriver-schemeAuthority2.q-bucket4.q-input16_cc.q-and-1-more - did 
not produce a TEST-*.xml file
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_union_view
org.apache.hadoop.hive.thrift.TestHadoop20SAuthBridge.testMetastoreProxyUser
org.apache.hadoop.hive.thrift.TestHadoop20SAuthBridge.testSaslWithHiveMetaStore
{noformat}

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/3462/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/3462/console
Test logs: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-3462/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 16 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12687447 - PreCommit-HIVE-TRUNK-Build

  upgrade hive schema from 0.9.0 to 0.13.1 failed when back end db is postgres 
 --

 Key: HIVE-7797
 URL: https://issues.apache.org/jira/browse/HIVE-7797
 Project: Hive
  Issue Type: Bug
  Components: Metastore
Affects Versions: 0.14.0, 0.13.1
Reporter: Nemon Lou
Assignee: Nemon Lou
 Attachments: HIVE-7797.1.patch


 Using  the following command to upgrade hive schema :
 schematool -dbType postgres -upgradeSchemaFrom 0.9.0
 then got this:
 ERROR: null value in column SCHEMA_VERSION violates not-null constraint
 Log shows that the upgrade sql file 014-HIVE-3764.postgres.sql failed.
 The sql in it is :
 INSERT INTO VERSION (VER_ID, SCHEMA_VERSION, VERSION_COMMENT) VALUES 
 (1, '', 'Initial value');
 And the result is:
 ERROR: null value in column SCHEMA_VERSION violates not-null constraint
 DETAIL: Failing row contains (1, null, Initial value).



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-10367) org.apache.hive.hcatalog.pig.TestHCatLoaderEncryption.* fails in Windows

2015-04-16 Thread Hari Sankar Sivarama Subramaniyan (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-10367?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14498545#comment-14498545
 ] 

Hari Sankar Sivarama Subramaniyan commented on HIVE-10367:
--

Either [~sushanth] or [~thejas] can review this minor change.

Thanks
Hari

 org.apache.hive.hcatalog.pig.TestHCatLoaderEncryption.* fails in Windows
 

 Key: HIVE-10367
 URL: https://issues.apache.org/jira/browse/HIVE-10367
 Project: Hive
  Issue Type: Bug
Reporter: Hari Sankar Sivarama Subramaniyan
Assignee: Hari Sankar Sivarama Subramaniyan
 Attachments: HIVE-10367.1.patch


 {code}
 java.io.IOException: Failed to execute create table encrypted_table(a int, b 
 string) location 
 'D:/w/hv/hcatalog/hcatalog-pig-adapter/target/tmp/org.apache.hive.hcatalog.pig.TestHCatLoader-1429172673960/warehouse/encryptedTable'.
  Driver returned 1 Error: FAILED: Execution Error, return code 1 from 
 org.apache.hadoop.hive.ql.exec.DDLTask. 
 MetaException(message:java.lang.IllegalArgumentException: Pathname 
 /D:/w/hv/hcatalog/hcatalog-pig-adapter/target/tmp/org.apache.hive.hcatalog.pig.TestHCatLoader-1429172673960/warehouse/encryptedTable
  from 
 hdfs://127.0.0.1:64735/D:/w/hv/hcatalog/hcatalog-pig-adapter/target/tmp/org.apache.hive.hcatalog.pig.TestHCatLoader-1429172673960/warehouse/encryptedTable
  is not a valid DFS filename.)
   at 
 org.apache.hive.hcatalog.pig.TestHCatLoaderEncryption.executeStatementOnDriver(TestHCatLoaderEncryption.java:163)
   at 
 org.apache.hive.hcatalog.pig.TestHCatLoaderEncryption.createTableInSpecifiedPath(TestHCatLoaderEncryption.java:423)
   at 
 org.apache.hive.hcatalog.pig.TestHCatLoaderEncryption.setup(TestHCatLoaderEncryption.java:197)
 {code}
 It looks like the error is happening because of the below query:
 create table encrypted_table(a int, b string) location 
 'D:/w/hv/hcatalog/hcatalog-pig-adapter/target/tmp/org.apache.hive.hcatalog.pig.TestHCatLoader-1429056939184/warehouse/encryptedTable'
 This gets called from createTableInSpecifiedPath() in 
 TestHCatLoaderEncryption. I think  an invalid HDFS path is passed here when 
 we are running the tests on Windows OS, so in line 197, changing
 {code}
 createTableInSpecifiedPath(ENCRYPTED_TABLE, a int, b string, 
 encryptedTablePath, driver);
 {code}
 to the below line might resolve the issue.
 {code}
 createTableInSpecifiedPath(ENCRYPTED_TABLE, a int, b string, 
 WindowsPathUtil.getHdfsUriString(encryptedTablePath), driver);
 {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-10239) Create scripts to do metastore upgrade tests on jenkins for Derby, Oracle and PostgreSQL

2015-04-16 Thread Naveen Gangam (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-10239?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14498596#comment-14498596
 ] 

Naveen Gangam commented on HIVE-10239:
--

I have been able to replicate the issue on a brand new VM by removing the 
apt-transport-https package before I did anything else.
{code}
100%[]
 1,743   --.-K/s   in 0s  

2015-04-16 12:56:00 (15.8 MB/s) - written to stdout [1743/1743]

OK
+ apt-get update
E: The method driver /usr/lib/apt/methods/https could not be found.
N: Is the package apt-transport-https installed?
+ /bin/true
+ apt-get install -y --force-yes oracle-xe
Reading package lists... Done
Building dependency tree   
Reading state information... Done
E: Unable to locate package oracle-xe
{code}

Will upload a fix shortly.

 Create scripts to do metastore upgrade tests on jenkins for Derby, Oracle and 
 PostgreSQL
 

 Key: HIVE-10239
 URL: https://issues.apache.org/jira/browse/HIVE-10239
 Project: Hive
  Issue Type: Improvement
Affects Versions: 1.1.0
Reporter: Naveen Gangam
Assignee: Naveen Gangam
 Attachments: HIVE-10239-donotcommit.patch, HIVE-10239.0.patch, 
 HIVE-10239.0.patch, HIVE-10239.patch


 Need to create DB-implementation specific scripts to use the framework 
 introduced in HIVE-9800 to have any metastore schema changes tested across 
 all supported databases.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-10284) enable container reuse for grace hash join

2015-04-16 Thread Wei Zheng (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-10284?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Wei Zheng updated HIVE-10284:
-
Attachment: HIVE-10284.9.patch

Updated test outputs for several q tests.
Upload path 9 for testing.

 enable container reuse for grace hash join 
 ---

 Key: HIVE-10284
 URL: https://issues.apache.org/jira/browse/HIVE-10284
 Project: Hive
  Issue Type: Bug
Reporter: Gunther Hagleitner
Assignee: Wei Zheng
 Attachments: HIVE-10284.1.patch, HIVE-10284.2.patch, 
 HIVE-10284.3.patch, HIVE-10284.4.patch, HIVE-10284.5.patch, 
 HIVE-10284.6.patch, HIVE-10284.7.patch, HIVE-10284.8.patch, HIVE-10284.9.patch






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-10222) Upgrade Calcite dependency to newest version

2015-04-16 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-10222?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14498921#comment-14498921
 ] 

Hive QA commented on HIVE-10222:




{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12725949/HIVE-10222.04.patch

{color:red}ERROR:{color} -1 due to 14 failed/errored test(s), 8711 tests 
executed
*Failed tests:*
{noformat}
TestMinimrCliDriver-bucketmapjoin6.q-constprog_partitioner.q-infer_bucket_sort_dyn_part.q-and-1-more
 - did not produce a TEST-*.xml file
TestMinimrCliDriver-external_table_with_space_in_location_path.q-infer_bucket_sort_merge.q-auto_sortmerge_join_16.q-and-1-more
 - did not produce a TEST-*.xml file
TestMinimrCliDriver-groupby2.q-import_exported_table.q-bucketizedhiveinputformat.q-and-1-more
 - did not produce a TEST-*.xml file
TestMinimrCliDriver-index_bitmap3.q-stats_counter_partitioned.q-temp_table_external.q-and-1-more
 - did not produce a TEST-*.xml file
TestMinimrCliDriver-infer_bucket_sort_map_operators.q-join1.q-bucketmapjoin7.q-and-1-more
 - did not produce a TEST-*.xml file
TestMinimrCliDriver-infer_bucket_sort_num_buckets.q-disable_merge_for_bucketing.q-uber_reduce.q-and-1-more
 - did not produce a TEST-*.xml file
TestMinimrCliDriver-infer_bucket_sort_reducers_power_two.q-scriptfile1.q-scriptfile1_win.q-and-1-more
 - did not produce a TEST-*.xml file
TestMinimrCliDriver-leftsemijoin_mr.q-load_hdfs_file_with_space_in_the_name.q-root_dir_external_table.q-and-1-more
 - did not produce a TEST-*.xml file
TestMinimrCliDriver-list_bucket_dml_10.q-bucket_num_reducers.q-bucket6.q-and-1-more
 - did not produce a TEST-*.xml file
TestMinimrCliDriver-load_fs2.q-file_with_header_footer.q-ql_rewrite_gbtoidx_cbo_1.q-and-1-more
 - did not produce a TEST-*.xml file
TestMinimrCliDriver-parallel_orderby.q-reduce_deduplicate.q-ql_rewrite_gbtoidx_cbo_2.q-and-1-more
 - did not produce a TEST-*.xml file
TestMinimrCliDriver-ql_rewrite_gbtoidx.q-smb_mapjoin_8.q - did not produce a 
TEST-*.xml file
TestMinimrCliDriver-schemeAuthority2.q-bucket4.q-input16_cc.q-and-1-more - did 
not produce a TEST-*.xml file
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_union_view
{noformat}

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/3463/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/3463/console
Test logs: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-3463/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 14 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12725949 - PreCommit-HIVE-TRUNK-Build

 Upgrade Calcite dependency to newest version
 

 Key: HIVE-10222
 URL: https://issues.apache.org/jira/browse/HIVE-10222
 Project: Hive
  Issue Type: Bug
Reporter: Jesus Camacho Rodriguez
Assignee: Jesus Camacho Rodriguez
 Attachments: HIVE-10222.01.patch, HIVE-10222.02.patch, 
 HIVE-10222.03.patch, HIVE-10222.04.patch, HIVE-10222.patch


  Upgrade Calcite version to 1.2.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-10284) enable container reuse for grace hash join

2015-04-16 Thread Gunther Hagleitner (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-10284?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14498961#comment-14498961
 ] 

Gunther Hagleitner commented on HIVE-10284:
---

+1 assuming tests will pass.

 enable container reuse for grace hash join 
 ---

 Key: HIVE-10284
 URL: https://issues.apache.org/jira/browse/HIVE-10284
 Project: Hive
  Issue Type: Bug
Reporter: Gunther Hagleitner
Assignee: Wei Zheng
 Attachments: HIVE-10284.1.patch, HIVE-10284.2.patch, 
 HIVE-10284.3.patch, HIVE-10284.4.patch, HIVE-10284.5.patch, 
 HIVE-10284.6.patch, HIVE-10284.7.patch, HIVE-10284.8.patch, HIVE-10284.9.patch






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-10374) Revert deprecation message to HiveCLI

2015-04-16 Thread Szehon Ho (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-10374?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Szehon Ho updated HIVE-10374:
-
Attachment: HIVE-10374.patch

 Revert deprecation message to HiveCLI
 -

 Key: HIVE-10374
 URL: https://issues.apache.org/jira/browse/HIVE-10374
 Project: Hive
  Issue Type: Bug
Reporter: Szehon Ho
Assignee: Szehon Ho
 Attachments: HIVE-10374.patch


 Reverting HIVE-10304 while there is a discussion about this.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-10277) Unable to process Comment line '--' in HIVE-1.1.0

2015-04-16 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-10277?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14499016#comment-14499016
 ] 

Hive QA commented on HIVE-10277:




{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12725857/HIVE-10277.2.patch

{color:red}ERROR:{color} -1 due to 632 failed/errored test(s), 8711 tests 
executed
*Failed tests:*
{noformat}
TestMinimrCliDriver-bucketmapjoin6.q-constprog_partitioner.q-infer_bucket_sort_dyn_part.q-and-1-more
 - did not produce a TEST-*.xml file
TestMinimrCliDriver-external_table_with_space_in_location_path.q-infer_bucket_sort_merge.q-auto_sortmerge_join_16.q-and-1-more
 - did not produce a TEST-*.xml file
TestMinimrCliDriver-groupby2.q-import_exported_table.q-bucketizedhiveinputformat.q-and-1-more
 - did not produce a TEST-*.xml file
TestMinimrCliDriver-index_bitmap3.q-stats_counter_partitioned.q-temp_table_external.q-and-1-more
 - did not produce a TEST-*.xml file
TestMinimrCliDriver-infer_bucket_sort_map_operators.q-join1.q-bucketmapjoin7.q-and-1-more
 - did not produce a TEST-*.xml file
TestMinimrCliDriver-infer_bucket_sort_num_buckets.q-disable_merge_for_bucketing.q-uber_reduce.q-and-1-more
 - did not produce a TEST-*.xml file
TestMinimrCliDriver-infer_bucket_sort_reducers_power_two.q-scriptfile1.q-scriptfile1_win.q-and-1-more
 - did not produce a TEST-*.xml file
TestMinimrCliDriver-leftsemijoin_mr.q-load_hdfs_file_with_space_in_the_name.q-root_dir_external_table.q-and-1-more
 - did not produce a TEST-*.xml file
TestMinimrCliDriver-list_bucket_dml_10.q-bucket_num_reducers.q-bucket6.q-and-1-more
 - did not produce a TEST-*.xml file
TestMinimrCliDriver-load_fs2.q-file_with_header_footer.q-ql_rewrite_gbtoidx_cbo_1.q-and-1-more
 - did not produce a TEST-*.xml file
TestMinimrCliDriver-parallel_orderby.q-reduce_deduplicate.q-ql_rewrite_gbtoidx_cbo_2.q-and-1-more
 - did not produce a TEST-*.xml file
TestMinimrCliDriver-ql_rewrite_gbtoidx.q-smb_mapjoin_8.q - did not produce a 
TEST-*.xml file
TestMinimrCliDriver-schemeAuthority2.q-bucket4.q-input16_cc.q-and-1-more - did 
not produce a TEST-*.xml file
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_add_part_multiple
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_alias_casted_column
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_alter5
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_alter_char1
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_alter_numbuckets_partitioned_table2_h23
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_alter_numbuckets_partitioned_table_h23
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_alter_partition_coltype
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_alter_partition_protect_mode
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_alter_rename_partition
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_alter_rename_partition_authorization
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_alter_table_serde
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_alter_table_serde2
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_alter_varchar1
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_authorization_1
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_authorization_2
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_authorization_3
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_authorization_4
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_authorization_5
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_authorization_6
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_authorization_9
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_auto_join26
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_auto_join_reordering_values
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_auto_sortmerge_join_1
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_auto_sortmerge_join_11
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_auto_sortmerge_join_12
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_auto_sortmerge_join_2
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_auto_sortmerge_join_3
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_auto_sortmerge_join_4
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_auto_sortmerge_join_5
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_auto_sortmerge_join_7
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_avro_add_column
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_avro_add_column2
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_avro_add_column3
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_avro_change_schema
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_avro_comments
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_avro_compression_enabled

[jira] [Resolved] (HIVE-10371) CBO (Calcite Return Path): Tag not set up correctly for Join operator in HiveOpConverter

2015-04-16 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-10371?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan resolved HIVE-10371.
-
Resolution: Fixed

Committed to branch. Thanks, Jesus!

 CBO (Calcite Return Path): Tag not set up correctly for Join operator in 
 HiveOpConverter
 

 Key: HIVE-10371
 URL: https://issues.apache.org/jira/browse/HIVE-10371
 Project: Hive
  Issue Type: Sub-task
  Components: CBO
Affects Versions: cbo-branch
Reporter: Jesus Camacho Rodriguez
Assignee: Jesus Camacho Rodriguez
 Fix For: cbo-branch

 Attachments: HIVE-10371.cbo.patch






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-10268) Merge cbo branch into trunk

2015-04-16 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-10268?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-10268:

Affects Version/s: (was: cbo-branch)

 Merge cbo branch into trunk
 ---

 Key: HIVE-10268
 URL: https://issues.apache.org/jira/browse/HIVE-10268
 Project: Hive
  Issue Type: Task
  Components: CBO
Reporter: Ashutosh Chauhan
Assignee: Ashutosh Chauhan
 Attachments: HIVE-10268.1.patch, HIVE-10268.2.patch, 
 HIVE-10268.3.patch, HIVE-10268.4.patch, HIVE-10268.patch


 Merge patch generated on basis of diffs of trunk with cbo-branch



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-10222) Upgrade Calcite dependency to newest version

2015-04-16 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-10222?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-10222:

Issue Type: Task  (was: Bug)

 Upgrade Calcite dependency to newest version
 

 Key: HIVE-10222
 URL: https://issues.apache.org/jira/browse/HIVE-10222
 Project: Hive
  Issue Type: Task
  Components: CBO
Reporter: Jesus Camacho Rodriguez
Assignee: Jesus Camacho Rodriguez
 Fix For: 1.2.0

 Attachments: HIVE-10222.01.patch, HIVE-10222.02.patch, 
 HIVE-10222.03.patch, HIVE-10222.04.patch, HIVE-10222.patch


  Upgrade Calcite version to 1.2.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-10264) Document Replication support on wiki

2015-04-16 Thread Sushanth Sowmyan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-10264?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sushanth Sowmyan updated HIVE-10264:

Attachment: EXIMReplicationReplayProtocol.pdf

EXIM Replay Protocol

 Document Replication support on wiki
 

 Key: HIVE-10264
 URL: https://issues.apache.org/jira/browse/HIVE-10264
 Project: Hive
  Issue Type: Sub-task
  Components: Import/Export
Reporter: Sushanth Sowmyan
Assignee: Sushanth Sowmyan
 Attachments: EXIMReplicationReplayProtocol.pdf, apache_hivedr.0.pdf






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-10228) Changes to Hive Export/Import/DropTable/DropPartition to support replication semantics

2015-04-16 Thread Sushanth Sowmyan (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-10228?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14498877#comment-14498877
 ] 

Sushanth Sowmyan commented on HIVE-10228:
-

Sorry, yeah, this is a big patch. :)

It's really a cumulative patch of a bunch of work, but a lot of that was 
overwriting itself so much that splitting them out into a bunch of patches 
would have been difficult. Forking hive to do dev of this on a separate branch 
and merging in one go might have been easier.

I'd created https://issues.apache.org/jira/browse/HIVE-10264 as a doc jira, and 
I've attached a presentation-like document there outlining various points of 
why we're doing a bunch of what we're doing, but that still needs some 
wiki-fication that I am working on. I've also attached the replay-protocol 
document on that jira after updating it slightly with your question on DROP 
TABLE here.

I'll reply to code-level comments on review board, and reply to your 
higher-level comments here.

DROP TABLE : This is not quite a DROP TABLE IF EXISTS, it's a DROP TABLE IF 
OLDER THAN(x). There are a couple of cases this can happen in:

a) To make it more resilient in cases of parallelization of events (in the 
cases of a worker that times out and does not respond back, for eg., but might 
still be running, albeit slowly in the background), one of the goals of all 
Commands generated by Replication is that they should be idempotent, and 
reprocessing of events older than the state of an object should not cause any 
error. So, if one drone that's processing events (41,42,43) might perform 41 
and then not respond back for a significant amount of time, causing Falcon to 
queue another HiveDR job that starts performing (41,42,43), and 43 might return 
successfully before the other job performs 42, and then failing. So, one of the 
early design goals was that all commands should be resilient to repeats. This 
is a way of achieving that goal.

b) In the case of a 
CREATE1-DROP1-CREATE2-REPL(CREATE1)-REPL(DROP1)-REPL(CREATE2), since the 
REPL(CREATE1) occurs after CREATE2, it picks up a newer state of the table, and 
the destination is at a newer state than the table which was dropped. Thus, by 
making the DROP ignore the destination table if it's already newer than the 
event that spawned the DROP, we can optimize away a bit of re-importing that 
REPL(CREATE2) would have needed to do. In the future, we'll add in 
event-nullification, and can do it at a higher level if we batch events, but 
this helps out even when processing at an individual level.

c) In addition to a DROP-IF-OLDER, it also acts like a recursive 
DROP-TABLE-IF-OLDER for cases where it doesn't result in the dropping of the 
table, it will still result in dropping older partitions in a newer table. For 
eg., if a T(state=50) has partitions P1(state=45) and P2(state=53), then 
DROP_TABLE_IF_OLDER_THAN(47) will drop P1 but not P2. This is because a 
Drop-table event does not result in a series of DropPtn events that are 
associated with the appropriate table. So, given that our replication works on 
an per-object basis, if DropTable should not drop the destination table because 
the destination table is newer than the origin table at the time of the drop, 
it might still contain older partitions which should be nuked. (This mode is 
tested in one of the tests in TestCommands in HIVE-10227 if you want to have a 
look at an example of what's expected)

--

Regarding the kewword addition, thanks for the feedback, it was not my intent 
to make them reserved keywords. I talked to [~pxiong] and [~ashutoshc] about 
it, and the latter is the way that makes sense. As long as I add them to the 
nonReserved entry in IdentifiersParser.g, it should be good. So, I'll add that 
in and have another update here.


 Changes to Hive Export/Import/DropTable/DropPartition to support replication 
 semantics
 --

 Key: HIVE-10228
 URL: https://issues.apache.org/jira/browse/HIVE-10228
 Project: Hive
  Issue Type: Sub-task
  Components: Import/Export
Affects Versions: 1.2.0
Reporter: Sushanth Sowmyan
Assignee: Sushanth Sowmyan
 Attachments: HIVE-10228.2.patch, HIVE-10228.3.patch, HIVE-10228.patch


 We need to update a couple of hive commands to support replication semantics. 
 To wit, we need the following:
 EXPORT ... [FOR [METADATA] REPLICATION(“comment”)]
 Export will now support an extra optional clause to tell it that this export 
 is being prepared for the purpose of replication. There is also an additional 
 optional clause here, that allows for the export to be a metadata-only 
 export, to handle cases of capturing the diff for alter statements, for 
 example.
 Also, if done for replication, the non-presence of a table, or a table 

[jira] [Updated] (HIVE-10371) CBO (Calcite Return Path): Tag not set up correctly for Join operator in HiveOpConverter

2015-04-16 Thread Jesus Camacho Rodriguez (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-10371?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jesus Camacho Rodriguez updated HIVE-10371:
---
Attachment: HIVE-10371.cbo.patch

[~ashutoshc], could you take a look? Thanks

 CBO (Calcite Return Path): Tag not set up correctly for Join operator in 
 HiveOpConverter
 

 Key: HIVE-10371
 URL: https://issues.apache.org/jira/browse/HIVE-10371
 Project: Hive
  Issue Type: Sub-task
  Components: CBO
Affects Versions: cbo-branch
Reporter: Jesus Camacho Rodriguez
Assignee: Jesus Camacho Rodriguez
 Fix For: cbo-branch

 Attachments: HIVE-10371.cbo.patch






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-10373) CBO (Calcite Return Path): Complete translation for all kinds of RexNode in Windowing

2015-04-16 Thread Jesus Camacho Rodriguez (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-10373?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jesus Camacho Rodriguez updated HIVE-10373:
---
Attachment: HIVE-10373.cbo.patch

[~jpullokkaran], could you review it? Thanks

 CBO (Calcite Return Path): Complete translation for all kinds of RexNode in 
 Windowing
 -

 Key: HIVE-10373
 URL: https://issues.apache.org/jira/browse/HIVE-10373
 Project: Hive
  Issue Type: Sub-task
  Components: CBO
Affects Versions: cbo-branch
Reporter: Jesus Camacho Rodriguez
Assignee: Jesus Camacho Rodriguez
 Fix For: cbo-branch

 Attachments: HIVE-10373.cbo.patch






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-10350) CBO: Use total size instead of bucket count to determine number of splits parallelism

2015-04-16 Thread Laljo John Pullokkaran (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-10350?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14499159#comment-14499159
 ] 

Laljo John Pullokkaran commented on HIVE-10350:
---

Fixed by HIVE-10369

 CBO: Use total size instead of bucket count to determine number of splits  
 parallelism 
 

 Key: HIVE-10350
 URL: https://issues.apache.org/jira/browse/HIVE-10350
 Project: Hive
  Issue Type: Sub-task
  Components: CBO
Affects Versions: 1.2.0
Reporter: Mostafa Mokhtar
Assignee: Mostafa Mokhtar
 Fix For: 1.2.0

 Attachments: HIVE-10331.01.patch, HIVE-10350.2.patch


 Not an overflow but parallelism ends up being -1 as it uses number of buckets
 {code}
  final int parallelism = RelMetadataQuery.splitCount(join) == null
   ? 1 : RelMetadataQuery.splitCount(join);
 {code}
 {code}
 2015-04-13 18:19:09,154 DEBUG [main]: cost.HiveCostModel 
 (HiveCostModel.java:getJoinCost(62)) - COMMON_JOIN cost: {1600892.857142857 
 rows, 2.4463782008994658E7 cpu, 8.54445445875E10 io}
 2015-04-13 18:19:09,155 DEBUG [main]: cost.HiveCostModel 
 (HiveCostModel.java:getJoinCost(62)) - MAP_JOIN cost: {1600892.857142857 
 rows, 1601785.714285714 cpu, -1698787.48 io}
 2015-04-13 18:19:09,155 DEBUG [main]: cost.HiveCostModel 
 (HiveCostModel.java:getJoinCost(72)) - MAP_JOIN selected
 2015-04-13 18:19:09,157 DEBUG [main]: parse.CalcitePlanner 
 (CalcitePlanner.java:apply(862)) - Plan After Join Reordering:
 HiveSort(fetch=[100]): rowcount = 6006.726049749041, cumulative cost = 
 {1.1468867492063493E8 rows, 1.166177684126984E8 cpu, -1.1757664816220238E9 
 io}, id = 3000
   HiveSort(sort0=[$0], dir0=[ASC]): rowcount = 6006.726049749041, cumulative 
 cost = {1.1468867492063493E8 rows, 1.166177684126984E8 cpu, 
 -1.1757664816220238E9 io}, id = 2998
 HiveProject(customer_id=[$4], customername=[concat($9, ', ', $8)]): 
 rowcount = 6006.726049749041, cumulative cost = {1.1468867492063493E8 rows, 
 1.166177684126984E8 cpu, -1.1757664816220238E9 io}, id = 3136
   HiveJoin(condition=[=($1, $5)], joinType=[inner], 
 joinAlgorithm=[map_join], cost=[{5.557820341269841E7 rows, 
 5.557840182539682E7 cpu, -4299694.122023809 io}]): rowcount = 
 6006.726049749041, cumulative cost = {1.1468867492063493E8 rows, 
 1.166177684126984E8 cpu, -1.1757664816220238E9 io}, id = 3132
 HiveJoin(condition=[=($0, $1)], joinType=[inner], 
 joinAlgorithm=[map_join], cost=[{5.7498805E7 rows, 5.9419605E7 cpu, 
 -1.15248E9 io}]): rowcount = 5.5578005E7, cumulative cost = {5.7498805E7 
 rows, 5.9419605E7 cpu, -1.15248E9 io}, id = 3100
   HiveProject(sr_cdemo_sk=[$4]): rowcount = 5.5578005E7, cumulative 
 cost = {0.0 rows, 0.0 cpu, 0.0 io}, id = 2992
 HiveTableScan(table=[[tpcds_bin_orc_200.store_returns]]): 
 rowcount = 5.5578005E7, cumulative cost = {0}, id = 2878
   HiveProject(cd_demo_sk=[$0]): rowcount = 1920800.0, cumulative cost 
 = {0.0 rows, 0.0 cpu, 0.0 io}, id = 2978
 HiveTableScan(table=[[tpcds_bin_orc_200.customer_demographics]]): 
 rowcount = 1920800.0, cumulative cost = {0}, id = 2868
 HiveJoin(condition=[=($10, $1)], joinType=[inner], 
 joinAlgorithm=[map_join], cost=[{1787.9365079365077 rows, 1790.15873015873 
 cpu, -8000.0 io}]): rowcount = 198.4126984126984, cumulative cost = 
 {1611666.507936508 rows, 1619761.5873015872 cpu, -1.89867875E7 io}, id = 3130
   HiveJoin(condition=[=($0, $4)], joinType=[inner], 
 joinAlgorithm=[map_join], cost=[{8985.714285714286 rows, 16185.714285714286 
 cpu, -1.728E7 io}]): rowcount = 1785.7142857142856, cumulative cost = 
 {1609878.5714285714 rows, 1617971.4285714284 cpu, -1.89787875E7 io}, id = 3128
 HiveProject(hd_demo_sk=[$0], hd_income_band_sk=[$1]): rowcount = 
 7200.0, cumulative cost = {0.0 rows, 0.0 cpu, 0.0 io}, id = 2982
   
 HiveTableScan(table=[[tpcds_bin_orc_200.household_demographics]]): rowcount = 
 7200.0, cumulative cost = {0}, id = 2871
 HiveJoin(condition=[=($3, $6)], joinType=[inner], 
 joinAlgorithm=[map_join], cost=[{1600892.857142857 rows, 1601785.714285714 
 cpu, -1698787.48 io}]): rowcount = 1785.7142857142856, cumulative 
 cost = {1600892.857142857 rows, 1601785.714285714 cpu, -1698787.48 
 io}, id = 3105
   HiveProject(c_customer_id=[$1], c_current_cdemo_sk=[$2], 
 c_current_hdemo_sk=[$3], c_current_addr_sk=[$4], c_first_name=[$8], 
 c_last_name=[$9]): rowcount = 160.0, cumulative cost = {0.0 rows, 0.0 
 cpu, 0.0 io}, id = 2970
 HiveTableScan(table=[[tpcds_bin_orc_200.customer]]): rowcount 
 = 160.0, cumulative cost = {0}, id = 2862
   HiveProject(ca_address_sk=[$0], ca_city=[$6]): rowcount = 
 892.8571428571428, cumulative 

[jira] [Updated] (HIVE-10306) We need to print tez summary when hive.server2.logging.level = PERFORMANCE.

2015-04-16 Thread Hari Sankar Sivarama Subramaniyan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-10306?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Hari Sankar Sivarama Subramaniyan updated HIVE-10306:
-
Attachment: HIVE-10306.5.patch

[~thejas] Made the change to move dependencies to hadoop-2 profile.

 We need to print tez summary when hive.server2.logging.level = PERFORMANCE. 
 -

 Key: HIVE-10306
 URL: https://issues.apache.org/jira/browse/HIVE-10306
 Project: Hive
  Issue Type: Bug
Reporter: Hari Sankar Sivarama Subramaniyan
Assignee: Hari Sankar Sivarama Subramaniyan
 Attachments: HIVE-10306.1.patch, HIVE-10306.2.patch, 
 HIVE-10306.3.patch, HIVE-10306.4.patch, HIVE-10306.5.patch


 We need to print tez summary when hive.server2.logging.level = PERFORMANCE. 
 We introduced this parameter via HIVE-10119.
 The logging param for levels is only relevant to HS2, so for hive-cli users 
 the hive.tez.exec.print.summary still makes sense. We can check for log-level 
 param as well, in places we are checking value of 
 hive.tez.exec.print.summary. Ie, consider hive.tez.exec.print.summary=true if 
 log.level = PERFORMANCE.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-8950) Add support in ParquetHiveSerde to create table schema from a parquet file

2015-04-16 Thread Ashish K Singh (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-8950?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashish K Singh updated HIVE-8950:
-
Attachment: HIVE-8950.7.patch

 Add support in ParquetHiveSerde to create table schema from a parquet file
 --

 Key: HIVE-8950
 URL: https://issues.apache.org/jira/browse/HIVE-8950
 Project: Hive
  Issue Type: Improvement
Reporter: Ashish K Singh
Assignee: Ashish K Singh
 Attachments: HIVE-8950.1.patch, HIVE-8950.2.patch, HIVE-8950.3.patch, 
 HIVE-8950.4.patch, HIVE-8950.5.patch, HIVE-8950.6.patch, HIVE-8950.7.patch, 
 HIVE-8950.patch


 PARQUET-76 and PARQUET-47 ask for creating parquet backed tables without 
 having to specify the column names and types. As, parquet files store schema 
 in their footer, it is possible to generate hive schema from parquet file's 
 metadata. This will improve usability of parquet backed tables.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Resolved] (HIVE-10369) CBO: Don't use HiveDefaultCostModel when With Tez and hive.cbo.costmodel.extended enabled

2015-04-16 Thread Laljo John Pullokkaran (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-10369?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Laljo John Pullokkaran resolved HIVE-10369.
---
Resolution: Fixed

 CBO: Don't use HiveDefaultCostModel when With Tez and 
 hive.cbo.costmodel.extended enabled 
 --

 Key: HIVE-10369
 URL: https://issues.apache.org/jira/browse/HIVE-10369
 Project: Hive
  Issue Type: Sub-task
  Components: CBO
Affects Versions: 1.2.0
Reporter: Mostafa Mokhtar
Assignee: Laljo John Pullokkaran
 Fix For: 1.2.0

 Attachments: HIVE-10369.patch


 When calculating parallelism, we end up using  HiveDefaultCostModel. 
 getSplitCount which returns null instead of  HiveOnTezCostModel.getSplitCount 
 which results in wrong parallelism.
 This happens for this join 
 {code}
 org.apache.calcite.plan.RelOptUtil.toString(join)
(java.lang.String) HiveJoin(condition=[=($1, $3)], joinType=[inner], 
 algorithm=[none], cost=[not available])
   HiveProject(cs_sold_date_sk=[$0], cs_bill_customer_sk=[$3], 
 cs_sales_price=[$21])
 HiveTableScan(table=[[tpcds_bin_orc_200.catalog_sales]])
   HiveJoin(condition=[=($1, $2)], joinType=[inner], algorithm=[MapJoin], 
 cost=[{240.0 rows, 6.48E11 cpu, 1294.6098 io}])
 HiveProject(c_customer_sk=[$0], c_current_addr_sk=[$4])
   HiveTableScan(table=[[tpcds_bin_orc_200.customer]])
 HiveProject(ca_address_sk=[$0], ca_state=[$8], ca_zip=[$9])
   HiveTableScan(table=[[tpcds_bin_orc_200.customer_address]])
 {code}
 The issue appears to be happening very early when calling 
 {code}
 if (pushDownTree != null) {
   costPushDown =
   RelMetadataQuery.getCumulativeCost(pushDownTree.getJoinTree());
 }
 {code}
 As pushDownTree.getJoinTree().joinAlgorithm = 
 HiveOnTezCostModel$TezMapJoinAlgorithm
 Call stack.
 {code}
 HiveDefaultCostModel$DefaultJoinAlgorithm.getSplitCount(HiveJoin) line: 114   
 HiveJoin.getSplitCount() line: 136
 HiveRelMdParallelism.splitCount(HiveJoin) line: 63
 NativeMethodAccessorImpl.invoke0(Method, Object, Object[]) line: not 
 available [native method]
 NativeMethodAccessorImpl.invoke(Object, Object[]) line: 57
 DelegatingMethodAccessorImpl.invoke(Object, Object[]) line: 43
 Method.invoke(Object, Object...) line: 606
 ReflectiveRelMetadataProvider$1$1.invoke(Object, Method, Object[]) line: 182  
 $Proxy46.splitCount() line: not available 
 GeneratedMethodAccessor26.invoke(Object, Object[]) line: not available
 DelegatingMethodAccessorImpl.invoke(Object, Object[]) line: 43
 Method.invoke(Object, Object...) line: 606
 ChainedRelMetadataProvider$ChainedInvocationHandler.invoke(Object, Method, 
 Object[]) line: 109
 $Proxy46.splitCount() line: not available 
 GeneratedMethodAccessor26.invoke(Object, Object[]) line: not available
 DelegatingMethodAccessorImpl.invoke(Object, Object[]) line: 43
 Method.invoke(Object, Object...) line: 606
 ChainedRelMetadataProvider$ChainedInvocationHandler.invoke(Object, Method, 
 Object[]) line: 109
 $Proxy46.splitCount() line: not available 
 GeneratedMethodAccessor26.invoke(Object, Object[]) line: not available
 DelegatingMethodAccessorImpl.invoke(Object, Object[]) line: 43
 Method.invoke(Object, Object...) line: 606
 CachingRelMetadataProvider$CachingInvocationHandler.invoke(Object, Method, 
 Object[]) line: 132
 $Proxy46.splitCount() line: not available 
 RelMetadataQuery.splitCount(RelNode) line: 401
 HiveOnTezCostModel$TezMapJoinAlgorithm.getCost(HiveJoin) line: 255
 HiveOnTezCostModel(HiveCostModel).getJoinCost(HiveJoin) line: 64  
 HiveRelMdCost.getNonCumulativeCost(HiveJoin) line: 56 
 NativeMethodAccessorImpl.invoke0(Method, Object, Object[]) line: not 
 available [native method]
 NativeMethodAccessorImpl.invoke(Object, Object[]) line: 57
 DelegatingMethodAccessorImpl.invoke(Object, Object[]) line: 43
 Method.invoke(Object, Object...) line: 606
 ReflectiveRelMetadataProvider$1$1.invoke(Object, Method, Object[]) line: 182  
 $Proxy41.getNonCumulativeCost() line: not available   
 GeneratedMethodAccessor22.invoke(Object, Object[]) line: not available
 DelegatingMethodAccessorImpl.invoke(Object, Object[]) line: 43
 Method.invoke(Object, Object...) line: 606
 ChainedRelMetadataProvider$ChainedInvocationHandler.invoke(Object, Method, 
 Object[]) line: 109
 $Proxy41.getNonCumulativeCost() line: not available   
 GeneratedMethodAccessor22.invoke(Object, Object[]) line: not available
 DelegatingMethodAccessorImpl.invoke(Object, Object[]) line: 43
 Method.invoke(Object, Object...) line: 606
 

[jira] [Commented] (HIVE-10304) Add deprecation message to HiveCLI

2015-04-16 Thread Mithun Radhakrishnan (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-10304?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14498900#comment-14498900
 ] 

Mithun Radhakrishnan commented on HIVE-10304:
-

Hello, [~szehon]. Pardon the delay; I wish I'd responded sooner.

There was a discussion on the dev user-list that concluded that we shouldn't be 
deprecating the Hive command-line until we have interface/error-code parity 
between beeline and the CLI. 
[Here|http://mail-archives.apache.org/mod_mbox/hive-dev/201412.mbox/%3ccabgngzfnjhnfv0p15+glmznf-gogw6dm9xotgoqh+dnyg3z...@mail.gmail.com%3E]
 is one thread. To quote:

bq. +1 to the idea of embedding beeline within hive cli, and retaining core 
behavior such as exit codes in hive-cli while doing that... users don't have to 
specify parameters like jdbc url, username etc.

The issue I see here is that there are still Hive installations that depend on 
the CLI, and don't depend entirely on HS2 deploys. (Where I work, for 
instance.) I'd be very keen to see the embedded-beeline option in working order.

Could we please discuss this check-in? I don't know if it's a good idea to push 
this into the impending release. I fear that the deprecation will be too 
disruptive, without proper recourse.

 Add deprecation message to HiveCLI
 --

 Key: HIVE-10304
 URL: https://issues.apache.org/jira/browse/HIVE-10304
 Project: Hive
  Issue Type: Improvement
  Components: CLI
Affects Versions: 1.1.0
Reporter: Szehon Ho
Assignee: Szehon Ho
  Labels: TODOC1.2
 Fix For: 1.2.0

 Attachments: HIVE-10304.2.patch, HIVE-10304.3.patch, HIVE-10304.patch


 As Beeline is now the recommended command line tool to Hive, we should add a 
 message to HiveCLI to indicate that it is deprecated and redirect them to 
 Beeline.  
 This is not suggesting to remove HiveCLI for now, but just a helpful 
 direction for user to know the direction to focus attention in Beeline.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-5672) Insert with custom separator not supported for non-local directory

2015-04-16 Thread Lefty Leverenz (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-5672?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14499023#comment-14499023
 ] 

Lefty Leverenz commented on HIVE-5672:
--

How's this?

{quote}
As of Hive 0.11.0 the separator used can be specified, in earlier versions it 
was always the ^A character (\001). However, custom separators are currently 
only supported for LOCAL writes; this is a bug that will be fixed by HIVE-5672.
{quote}

* [DML -- Writing data into the filesystem from queries -- Notes | 
https://cwiki.apache.org/confluence/display/Hive/LanguageManual+DML#LanguageManualDML-Notes.2]

 Insert with custom separator not supported for non-local directory
 --

 Key: HIVE-5672
 URL: https://issues.apache.org/jira/browse/HIVE-5672
 Project: Hive
  Issue Type: Bug
Affects Versions: 0.12.0, 1.0.0
Reporter: Romain Rigaux
Assignee: Nemon Lou
 Attachments: HIVE-5672.1.patch, HIVE-5672.2.patch, HIVE-5672.3.patch


 https://issues.apache.org/jira/browse/HIVE-3682 is great but non local 
 directory don't seem to be supported:
 {code}
 insert overwrite directory '/tmp/test-02'
 row format delimited
 FIELDS TERMINATED BY ':'
 select description FROM sample_07
 {code}
 {code}
 Error while compiling statement: FAILED: ParseException line 2:0 cannot 
 recognize input near 'row' 'format' 'delimited' in select clause
 {code}
 This works (with 'local'):
 {code}
 insert overwrite local directory '/tmp/test-02'
 row format delimited
 FIELDS TERMINATED BY ':'
 select code, description FROM sample_07
 {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-10364) The HMS upgrade script test does not publish results when prepare.sh fails.

2015-04-16 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-10364?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14499102#comment-14499102
 ] 

Hive QA commented on HIVE-10364:




{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12725894/HIVE-10364.1.patch

{color:red}ERROR:{color} -1 due to 17 failed/errored test(s), 8711 tests 
executed
*Failed tests:*
{noformat}
TestMinimrCliDriver-bucketmapjoin6.q-constprog_partitioner.q-infer_bucket_sort_dyn_part.q-and-1-more
 - did not produce a TEST-*.xml file
TestMinimrCliDriver-external_table_with_space_in_location_path.q-infer_bucket_sort_merge.q-auto_sortmerge_join_16.q-and-1-more
 - did not produce a TEST-*.xml file
TestMinimrCliDriver-groupby2.q-import_exported_table.q-bucketizedhiveinputformat.q-and-1-more
 - did not produce a TEST-*.xml file
TestMinimrCliDriver-index_bitmap3.q-stats_counter_partitioned.q-temp_table_external.q-and-1-more
 - did not produce a TEST-*.xml file
TestMinimrCliDriver-infer_bucket_sort_map_operators.q-join1.q-bucketmapjoin7.q-and-1-more
 - did not produce a TEST-*.xml file
TestMinimrCliDriver-infer_bucket_sort_num_buckets.q-disable_merge_for_bucketing.q-uber_reduce.q-and-1-more
 - did not produce a TEST-*.xml file
TestMinimrCliDriver-infer_bucket_sort_reducers_power_two.q-scriptfile1.q-scriptfile1_win.q-and-1-more
 - did not produce a TEST-*.xml file
TestMinimrCliDriver-leftsemijoin_mr.q-load_hdfs_file_with_space_in_the_name.q-root_dir_external_table.q-and-1-more
 - did not produce a TEST-*.xml file
TestMinimrCliDriver-list_bucket_dml_10.q-bucket_num_reducers.q-bucket6.q-and-1-more
 - did not produce a TEST-*.xml file
TestMinimrCliDriver-load_fs2.q-file_with_header_footer.q-ql_rewrite_gbtoidx_cbo_1.q-and-1-more
 - did not produce a TEST-*.xml file
TestMinimrCliDriver-parallel_orderby.q-reduce_deduplicate.q-ql_rewrite_gbtoidx_cbo_2.q-and-1-more
 - did not produce a TEST-*.xml file
TestMinimrCliDriver-ql_rewrite_gbtoidx.q-smb_mapjoin_8.q - did not produce a 
TEST-*.xml file
TestMinimrCliDriver-schemeAuthority2.q-bucket4.q-input16_cc.q-and-1-more - did 
not produce a TEST-*.xml file
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_union_view
org.apache.hadoop.hive.thrift.TestHadoop20SAuthBridge.testMetastoreProxyUser
org.apache.hadoop.hive.thrift.TestHadoop20SAuthBridge.testSaslWithHiveMetaStore
org.apache.hive.hcatalog.streaming.TestStreaming.testTransactionBatchCommit_Json
{noformat}

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/3466/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/3466/console
Test logs: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-3466/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 17 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12725894 - PreCommit-HIVE-TRUNK-Build

 The HMS upgrade script test does not publish results when prepare.sh fails.
 ---

 Key: HIVE-10364
 URL: https://issues.apache.org/jira/browse/HIVE-10364
 Project: Hive
  Issue Type: Bug
  Components: Testing Infrastructure
Reporter: Sergio Peña
Assignee: Sergio Peña
 Attachments: HIVE-10364.1.patch


 The HMS upgrade script must publish succeed or failure results to JIRA. This 
 bug is not publishing any results on JIRA is the prepare.sh script fails.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9710) HiveServer2 should support cookie based authentication, when using HTTP transport.

2015-04-16 Thread Lefty Leverenz (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9710?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14498989#comment-14498989
 ] 

Lefty Leverenz commented on HIVE-9710:
--

Doc note:  In addition to general documentation (release notes), these new 
configuration parameters need to be documented:

* hive.server2.thrift.http.cookie.auth.enabled
* hive.server2.thrift.http.cookie.max.age
* hive.server2.thrift.http.cookie.domain
* hive.server2.thrift.http.cookie.path
* hive.server2.thrift.http.cookie.is.secure
* hive.server2.thrift.http.cookie.is.httponly

Here are the wikidocs for HS2 and its configs:

* [Setting Up HiveServer2 | 
https://cwiki.apache.org/confluence/display/Hive/Setting+Up+HiveServer2]
* [HiveServer2 Clients | 
https://cwiki.apache.org/confluence/display/Hive/HiveServer2+Clients]
* [Configuration Properties -- HiveServer2 | 
https://cwiki.apache.org/confluence/display/Hive/Configuration+Properties#ConfigurationProperties-HiveServer2]

 HiveServer2 should support cookie based authentication, when using HTTP 
 transport.
 --

 Key: HIVE-9710
 URL: https://issues.apache.org/jira/browse/HIVE-9710
 Project: Hive
  Issue Type: Improvement
  Components: HiveServer2
Affects Versions: 1.2.0
Reporter: Hari Sankar Sivarama Subramaniyan
Assignee: Hari Sankar Sivarama Subramaniyan
  Labels: TODOC1.2
 Fix For: 1.2.0

 Attachments: HIVE-9710.1.patch, HIVE-9710.2.patch, HIVE-9710.3.patch, 
 HIVE-9710.4.patch, HIVE-9710.5.patch, HIVE-9710.6.patch, HIVE-9710.7.patch, 
 HIVE-9710.8.patch


 HiveServer2 should generate cookies and validate the client cookie send to it 
 so that it need not perform User/Password or a Kerberos based authentication 
 on each HTTP request. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-10369) CBO: Don't use HiveDefaultCostModel when With Tez and hive.cbo.costmodel.extended enabled

2015-04-16 Thread Mostafa Mokhtar (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-10369?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14498904#comment-14498904
 ] 

Mostafa Mokhtar commented on HIVE-10369:


[~jcamachorodriguez]

 CBO: Don't use HiveDefaultCostModel when With Tez and 
 hive.cbo.costmodel.extended enabled 
 --

 Key: HIVE-10369
 URL: https://issues.apache.org/jira/browse/HIVE-10369
 Project: Hive
  Issue Type: Sub-task
  Components: CBO
Affects Versions: 1.2.0
Reporter: Mostafa Mokhtar
Assignee: Laljo John Pullokkaran
 Fix For: 1.2.0


 When calculating parallelism, we end up using  HiveDefaultCostModel. 
 getSplitCount which returns null instead of  HiveOnTezCostModel.getSplitCount 
 which results in wrong parallelism.
 This happens for this join 
 {code}
 org.apache.calcite.plan.RelOptUtil.toString(join)
(java.lang.String) HiveJoin(condition=[=($1, $3)], joinType=[inner], 
 algorithm=[none], cost=[not available])
   HiveProject(cs_sold_date_sk=[$0], cs_bill_customer_sk=[$3], 
 cs_sales_price=[$21])
 HiveTableScan(table=[[tpcds_bin_orc_200.catalog_sales]])
   HiveJoin(condition=[=($1, $2)], joinType=[inner], algorithm=[MapJoin], 
 cost=[{240.0 rows, 6.48E11 cpu, 1294.6098 io}])
 HiveProject(c_customer_sk=[$0], c_current_addr_sk=[$4])
   HiveTableScan(table=[[tpcds_bin_orc_200.customer]])
 HiveProject(ca_address_sk=[$0], ca_state=[$8], ca_zip=[$9])
   HiveTableScan(table=[[tpcds_bin_orc_200.customer_address]])
 {code}
 The issue appears to be happening very early when calling 
 {code}
 if (pushDownTree != null) {
   costPushDown =
   RelMetadataQuery.getCumulativeCost(pushDownTree.getJoinTree());
 }
 {code}
 As pushDownTree.getJoinTree().joinAlgorithm = 
 HiveOnTezCostModel$TezMapJoinAlgorithm
 Call stack.
 {code}
 HiveDefaultCostModel$DefaultJoinAlgorithm.getSplitCount(HiveJoin) line: 114   
 HiveJoin.getSplitCount() line: 136
 HiveRelMdParallelism.splitCount(HiveJoin) line: 63
 NativeMethodAccessorImpl.invoke0(Method, Object, Object[]) line: not 
 available [native method]
 NativeMethodAccessorImpl.invoke(Object, Object[]) line: 57
 DelegatingMethodAccessorImpl.invoke(Object, Object[]) line: 43
 Method.invoke(Object, Object...) line: 606
 ReflectiveRelMetadataProvider$1$1.invoke(Object, Method, Object[]) line: 182  
 $Proxy46.splitCount() line: not available 
 GeneratedMethodAccessor26.invoke(Object, Object[]) line: not available
 DelegatingMethodAccessorImpl.invoke(Object, Object[]) line: 43
 Method.invoke(Object, Object...) line: 606
 ChainedRelMetadataProvider$ChainedInvocationHandler.invoke(Object, Method, 
 Object[]) line: 109
 $Proxy46.splitCount() line: not available 
 GeneratedMethodAccessor26.invoke(Object, Object[]) line: not available
 DelegatingMethodAccessorImpl.invoke(Object, Object[]) line: 43
 Method.invoke(Object, Object...) line: 606
 ChainedRelMetadataProvider$ChainedInvocationHandler.invoke(Object, Method, 
 Object[]) line: 109
 $Proxy46.splitCount() line: not available 
 GeneratedMethodAccessor26.invoke(Object, Object[]) line: not available
 DelegatingMethodAccessorImpl.invoke(Object, Object[]) line: 43
 Method.invoke(Object, Object...) line: 606
 CachingRelMetadataProvider$CachingInvocationHandler.invoke(Object, Method, 
 Object[]) line: 132
 $Proxy46.splitCount() line: not available 
 RelMetadataQuery.splitCount(RelNode) line: 401
 HiveOnTezCostModel$TezMapJoinAlgorithm.getCost(HiveJoin) line: 255
 HiveOnTezCostModel(HiveCostModel).getJoinCost(HiveJoin) line: 64  
 HiveRelMdCost.getNonCumulativeCost(HiveJoin) line: 56 
 NativeMethodAccessorImpl.invoke0(Method, Object, Object[]) line: not 
 available [native method]
 NativeMethodAccessorImpl.invoke(Object, Object[]) line: 57
 DelegatingMethodAccessorImpl.invoke(Object, Object[]) line: 43
 Method.invoke(Object, Object...) line: 606
 ReflectiveRelMetadataProvider$1$1.invoke(Object, Method, Object[]) line: 182  
 $Proxy41.getNonCumulativeCost() line: not available   
 GeneratedMethodAccessor22.invoke(Object, Object[]) line: not available
 DelegatingMethodAccessorImpl.invoke(Object, Object[]) line: 43
 Method.invoke(Object, Object...) line: 606
 ChainedRelMetadataProvider$ChainedInvocationHandler.invoke(Object, Method, 
 Object[]) line: 109
 $Proxy41.getNonCumulativeCost() line: not available   
 GeneratedMethodAccessor22.invoke(Object, Object[]) line: not available
 DelegatingMethodAccessorImpl.invoke(Object, Object[]) line: 43
 Method.invoke(Object, Object...) line: 606
 

[jira] [Updated] (HIVE-10268) Merge cbo branch into trunk

2015-04-16 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-10268?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-10268:

Attachment: HIVE-10268.4.patch

 Merge cbo branch into trunk
 ---

 Key: HIVE-10268
 URL: https://issues.apache.org/jira/browse/HIVE-10268
 Project: Hive
  Issue Type: Task
  Components: CBO
Affects Versions: cbo-branch
Reporter: Ashutosh Chauhan
Assignee: Ashutosh Chauhan
 Attachments: HIVE-10268.1.patch, HIVE-10268.2.patch, 
 HIVE-10268.3.patch, HIVE-10268.4.patch, HIVE-10268.patch


 Merge patch generated on basis of diffs of trunk with cbo-branch



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-10371) CBO (Calcite Return Path): Tag not set up correctly for Join operator in HiveOpConverter

2015-04-16 Thread Jesus Camacho Rodriguez (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-10371?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jesus Camacho Rodriguez updated HIVE-10371:
---
Attachment: HIVE-10371.cbo.patch

 CBO (Calcite Return Path): Tag not set up correctly for Join operator in 
 HiveOpConverter
 

 Key: HIVE-10371
 URL: https://issues.apache.org/jira/browse/HIVE-10371
 Project: Hive
  Issue Type: Sub-task
  Components: CBO
Affects Versions: cbo-branch
Reporter: Jesus Camacho Rodriguez
Assignee: Jesus Camacho Rodriguez
 Fix For: cbo-branch

 Attachments: HIVE-10371.cbo.patch






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-10222) Upgrade Calcite dependency to newest version

2015-04-16 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-10222?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-10222:

Component/s: CBO

 Upgrade Calcite dependency to newest version
 

 Key: HIVE-10222
 URL: https://issues.apache.org/jira/browse/HIVE-10222
 Project: Hive
  Issue Type: Bug
  Components: CBO
Reporter: Jesus Camacho Rodriguez
Assignee: Jesus Camacho Rodriguez
 Fix For: 1.2.0

 Attachments: HIVE-10222.01.patch, HIVE-10222.02.patch, 
 HIVE-10222.03.patch, HIVE-10222.04.patch, HIVE-10222.patch


  Upgrade Calcite version to 1.2.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-10304) Add deprecation message to HiveCLI

2015-04-16 Thread Szehon Ho (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-10304?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14499176#comment-14499176
 ] 

Szehon Ho commented on HIVE-10304:
--

OK, filed HIVE-10374 for reversion while discussion is ongoing.

 Add deprecation message to HiveCLI
 --

 Key: HIVE-10304
 URL: https://issues.apache.org/jira/browse/HIVE-10304
 Project: Hive
  Issue Type: Improvement
  Components: CLI
Affects Versions: 1.1.0
Reporter: Szehon Ho
Assignee: Szehon Ho
  Labels: TODOC1.2
 Fix For: 1.2.0

 Attachments: HIVE-10304.2.patch, HIVE-10304.3.patch, HIVE-10304.patch


 As Beeline is now the recommended command line tool to Hive, we should add a 
 message to HiveCLI to indicate that it is deprecated and redirect them to 
 Beeline.  
 This is not suggesting to remove HiveCLI for now, but just a helpful 
 direction for user to know the direction to focus attention in Beeline.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-10190) CBO: AST mode checks for TABLESAMPLE with AST.toString().contains(TOK_TABLESPLITSAMPLE)

2015-04-16 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-10190?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14499153#comment-14499153
 ] 

Hive QA commented on HIVE-10190:




{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12725898/HIVE-10190.05.patch

{color:red}ERROR:{color} -1 due to 25 failed/errored test(s), 8711 tests 
executed
*Failed tests:*
{noformat}
TestMinimrCliDriver-bucketmapjoin6.q-constprog_partitioner.q-infer_bucket_sort_dyn_part.q-and-1-more
 - did not produce a TEST-*.xml file
TestMinimrCliDriver-external_table_with_space_in_location_path.q-infer_bucket_sort_merge.q-auto_sortmerge_join_16.q-and-1-more
 - did not produce a TEST-*.xml file
TestMinimrCliDriver-groupby2.q-import_exported_table.q-bucketizedhiveinputformat.q-and-1-more
 - did not produce a TEST-*.xml file
TestMinimrCliDriver-index_bitmap3.q-stats_counter_partitioned.q-temp_table_external.q-and-1-more
 - did not produce a TEST-*.xml file
TestMinimrCliDriver-infer_bucket_sort_map_operators.q-join1.q-bucketmapjoin7.q-and-1-more
 - did not produce a TEST-*.xml file
TestMinimrCliDriver-infer_bucket_sort_num_buckets.q-disable_merge_for_bucketing.q-uber_reduce.q-and-1-more
 - did not produce a TEST-*.xml file
TestMinimrCliDriver-infer_bucket_sort_reducers_power_two.q-scriptfile1.q-scriptfile1_win.q-and-1-more
 - did not produce a TEST-*.xml file
TestMinimrCliDriver-leftsemijoin_mr.q-load_hdfs_file_with_space_in_the_name.q-root_dir_external_table.q-and-1-more
 - did not produce a TEST-*.xml file
TestMinimrCliDriver-list_bucket_dml_10.q-bucket_num_reducers.q-bucket6.q-and-1-more
 - did not produce a TEST-*.xml file
TestMinimrCliDriver-load_fs2.q-file_with_header_footer.q-ql_rewrite_gbtoidx_cbo_1.q-and-1-more
 - did not produce a TEST-*.xml file
TestMinimrCliDriver-parallel_orderby.q-reduce_deduplicate.q-ql_rewrite_gbtoidx_cbo_2.q-and-1-more
 - did not produce a TEST-*.xml file
TestMinimrCliDriver-ql_rewrite_gbtoidx.q-smb_mapjoin_8.q - did not produce a 
TEST-*.xml file
TestMinimrCliDriver-schemeAuthority2.q-bucket4.q-input16_cc.q-and-1-more - did 
not produce a TEST-*.xml file
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_inputddl5
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_partition_decode_name
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_partition_special_char
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_ppr_pushdown
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_udf_printf
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_udf_reverse
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_udf_substr
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_udtf_json_tuple
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_udtf_parse_url_tuple
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_union35
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_union_view
org.apache.hadoop.hive.cli.TestMiniTezCliDriver.testCliDriver_tez_union_decimal
{noformat}

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/3467/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/3467/console
Test logs: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-3467/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 25 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12725898 - PreCommit-HIVE-TRUNK-Build

 CBO: AST mode checks for TABLESAMPLE with 
 AST.toString().contains(TOK_TABLESPLITSAMPLE)
 -

 Key: HIVE-10190
 URL: https://issues.apache.org/jira/browse/HIVE-10190
 Project: Hive
  Issue Type: Bug
  Components: CBO
Affects Versions: 1.2.0
Reporter: Gopal V
Assignee: Reuben Kuhnert
Priority: Trivial
  Labels: perfomance
 Attachments: HIVE-10190-querygen.py, HIVE-10190.01.patch, 
 HIVE-10190.02.patch, HIVE-10190.03.patch, HIVE-10190.04.patch, 
 HIVE-10190.05.patch, HIVE-10190.05.patch


 {code}
 public static boolean validateASTForUnsupportedTokens(ASTNode ast) {
 String astTree = ast.toStringTree();
 // if any of following tokens are present in AST, bail out
 String[] tokens = { TOK_CHARSETLITERAL, TOK_TABLESPLITSAMPLE };
 for (String token : tokens) {
   if (astTree.contains(token)) {
 return false;
   }
 }
 return true;
   }
 {code}
 This is an issue for a SQL query which is bigger in AST form than in text 
 (~700kb).




[jira] [Updated] (HIVE-10372) Bump parquet version to 1.6.0

2015-04-16 Thread Ferdinand Xu (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-10372?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ferdinand Xu updated HIVE-10372:

Attachment: HIVE-10372.patch

Hi [~spena], the maven central repo is not updated yet. Will be pending on that.

 Bump parquet version to 1.6.0
 -

 Key: HIVE-10372
 URL: https://issues.apache.org/jira/browse/HIVE-10372
 Project: Hive
  Issue Type: Bug
Reporter: Ferdinand Xu
Assignee: Ferdinand Xu
 Attachments: HIVE-10372.patch






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-10324) Hive metatool should take table_param_key to allow for changes to avro serde's schema url key

2015-04-16 Thread Ferdinand Xu (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-10324?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14499055#comment-14499055
 ] 

Ferdinand Xu commented on HIVE-10324:
-

 Thanks [~szehon] for your review. And the failed cases are irrelevant.

 Hive metatool should take table_param_key to allow for changes to avro 
 serde's schema url key
 -

 Key: HIVE-10324
 URL: https://issues.apache.org/jira/browse/HIVE-10324
 Project: Hive
  Issue Type: Bug
  Components: Metastore
Affects Versions: 1.1.0
Reporter: Szehon Ho
Assignee: Ferdinand Xu
 Attachments: HIVE-10324.1.patch, HIVE-10324.patch, 
 HIVE-10324.patch.WIP


 HIVE-3443 added support to change the serdeParams from 'metatool 
 updateLocation' command.
 However, in avro it is possible to specify the schema via the tableParams:
 {noformat}
 CREATE  TABLE `testavro`(
   `test` string COMMENT 'from deserializer')
 ROW FORMAT SERDE 
   'org.apache.hadoop.hive.serde2.avro.AvroSerDe' 
 STORED AS INPUTFORMAT 
   'org.apache.hadoop.hive.ql.io.avro.AvroContainerInputFormat' 
 OUTPUTFORMAT 
   'org.apache.hadoop.hive.ql.io.avro.AvroContainerOutputFormat'
 TBLPROPERTIES (
   'avro.schema.url'='hdfs://namenode:8020/tmp/test.avsc', 
   'kite.compression.type'='snappy', 
   'transient_lastDdlTime'='1427996456')
 {noformat}
 Hence for those tables the 'metatool updateLocation' will not help.
 This is necessary in case like upgrade the namenode to HA where the absolute 
 paths have changed.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-10369) CBO: Don't use HiveDefaultCostModel when With Tez and hive.cbo.costmodel.extended enabled

2015-04-16 Thread Laljo John Pullokkaran (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-10369?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Laljo John Pullokkaran updated HIVE-10369:
--
Attachment: HIVE-10369.patch

 CBO: Don't use HiveDefaultCostModel when With Tez and 
 hive.cbo.costmodel.extended enabled 
 --

 Key: HIVE-10369
 URL: https://issues.apache.org/jira/browse/HIVE-10369
 Project: Hive
  Issue Type: Sub-task
  Components: CBO
Affects Versions: 1.2.0
Reporter: Mostafa Mokhtar
Assignee: Laljo John Pullokkaran
 Fix For: 1.2.0

 Attachments: HIVE-10369.patch


 When calculating parallelism, we end up using  HiveDefaultCostModel. 
 getSplitCount which returns null instead of  HiveOnTezCostModel.getSplitCount 
 which results in wrong parallelism.
 This happens for this join 
 {code}
 org.apache.calcite.plan.RelOptUtil.toString(join)
(java.lang.String) HiveJoin(condition=[=($1, $3)], joinType=[inner], 
 algorithm=[none], cost=[not available])
   HiveProject(cs_sold_date_sk=[$0], cs_bill_customer_sk=[$3], 
 cs_sales_price=[$21])
 HiveTableScan(table=[[tpcds_bin_orc_200.catalog_sales]])
   HiveJoin(condition=[=($1, $2)], joinType=[inner], algorithm=[MapJoin], 
 cost=[{240.0 rows, 6.48E11 cpu, 1294.6098 io}])
 HiveProject(c_customer_sk=[$0], c_current_addr_sk=[$4])
   HiveTableScan(table=[[tpcds_bin_orc_200.customer]])
 HiveProject(ca_address_sk=[$0], ca_state=[$8], ca_zip=[$9])
   HiveTableScan(table=[[tpcds_bin_orc_200.customer_address]])
 {code}
 The issue appears to be happening very early when calling 
 {code}
 if (pushDownTree != null) {
   costPushDown =
   RelMetadataQuery.getCumulativeCost(pushDownTree.getJoinTree());
 }
 {code}
 As pushDownTree.getJoinTree().joinAlgorithm = 
 HiveOnTezCostModel$TezMapJoinAlgorithm
 Call stack.
 {code}
 HiveDefaultCostModel$DefaultJoinAlgorithm.getSplitCount(HiveJoin) line: 114   
 HiveJoin.getSplitCount() line: 136
 HiveRelMdParallelism.splitCount(HiveJoin) line: 63
 NativeMethodAccessorImpl.invoke0(Method, Object, Object[]) line: not 
 available [native method]
 NativeMethodAccessorImpl.invoke(Object, Object[]) line: 57
 DelegatingMethodAccessorImpl.invoke(Object, Object[]) line: 43
 Method.invoke(Object, Object...) line: 606
 ReflectiveRelMetadataProvider$1$1.invoke(Object, Method, Object[]) line: 182  
 $Proxy46.splitCount() line: not available 
 GeneratedMethodAccessor26.invoke(Object, Object[]) line: not available
 DelegatingMethodAccessorImpl.invoke(Object, Object[]) line: 43
 Method.invoke(Object, Object...) line: 606
 ChainedRelMetadataProvider$ChainedInvocationHandler.invoke(Object, Method, 
 Object[]) line: 109
 $Proxy46.splitCount() line: not available 
 GeneratedMethodAccessor26.invoke(Object, Object[]) line: not available
 DelegatingMethodAccessorImpl.invoke(Object, Object[]) line: 43
 Method.invoke(Object, Object...) line: 606
 ChainedRelMetadataProvider$ChainedInvocationHandler.invoke(Object, Method, 
 Object[]) line: 109
 $Proxy46.splitCount() line: not available 
 GeneratedMethodAccessor26.invoke(Object, Object[]) line: not available
 DelegatingMethodAccessorImpl.invoke(Object, Object[]) line: 43
 Method.invoke(Object, Object...) line: 606
 CachingRelMetadataProvider$CachingInvocationHandler.invoke(Object, Method, 
 Object[]) line: 132
 $Proxy46.splitCount() line: not available 
 RelMetadataQuery.splitCount(RelNode) line: 401
 HiveOnTezCostModel$TezMapJoinAlgorithm.getCost(HiveJoin) line: 255
 HiveOnTezCostModel(HiveCostModel).getJoinCost(HiveJoin) line: 64  
 HiveRelMdCost.getNonCumulativeCost(HiveJoin) line: 56 
 NativeMethodAccessorImpl.invoke0(Method, Object, Object[]) line: not 
 available [native method]
 NativeMethodAccessorImpl.invoke(Object, Object[]) line: 57
 DelegatingMethodAccessorImpl.invoke(Object, Object[]) line: 43
 Method.invoke(Object, Object...) line: 606
 ReflectiveRelMetadataProvider$1$1.invoke(Object, Method, Object[]) line: 182  
 $Proxy41.getNonCumulativeCost() line: not available   
 GeneratedMethodAccessor22.invoke(Object, Object[]) line: not available
 DelegatingMethodAccessorImpl.invoke(Object, Object[]) line: 43
 Method.invoke(Object, Object...) line: 606
 ChainedRelMetadataProvider$ChainedInvocationHandler.invoke(Object, Method, 
 Object[]) line: 109
 $Proxy41.getNonCumulativeCost() line: not available   
 GeneratedMethodAccessor22.invoke(Object, Object[]) line: not available
 DelegatingMethodAccessorImpl.invoke(Object, Object[]) line: 43
 Method.invoke(Object, Object...) line: 606
 

[jira] [Updated] (HIVE-10307) Support to use number literals in partition column

2015-04-16 Thread Chaoyu Tang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-10307?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chaoyu Tang updated HIVE-10307:
---
Attachment: HIVE-10307.2.patch

Uploaded revised patch based on Jimmy's review.

 Support to use number literals in partition column
 --

 Key: HIVE-10307
 URL: https://issues.apache.org/jira/browse/HIVE-10307
 Project: Hive
  Issue Type: Improvement
  Components: Query Processor
Affects Versions: 1.0.0
Reporter: Chaoyu Tang
Assignee: Chaoyu Tang
 Attachments: HIVE-10307.1.patch, HIVE-10307.2.patch, HIVE-10307.patch


 Data types like TinyInt, SmallInt, BigInt or Decimal can be expressed as 
 literals with postfix like Y, S, L, or BD appended to the number. These 
 literals work in most Hive queries, but do not when they are used as 
 partition column value. For a partitioned table like:
 create table partcoltypenum (key int, value string) partitioned by (tint 
 tinyint, sint smallint, bint bigint);
 insert into partcoltypenum partition (tint=100Y, sint=1S, 
 bint=1000L) select key, value from src limit 30;
 Queries like select, describe and drop partition do not work. For an example
 select * from partcoltypenum where tint=100Y and sint=1S and 
 bint=1000L;
 does not return any rows.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-10242) ACID: insert overwrite prevents create table command

2015-04-16 Thread Eugene Koifman (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-10242?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14498411#comment-14498411
 ] 

Eugene Koifman commented on HIVE-10242:
---

The fix also logging improvements and a fix to the following situation where a 
lock (by client3) could be acquired when it should not be.
client1: update table T
client2: select * from T
client3: update table T

with all 3 clients executing concurrently and starting in above order, client 3 
should be blocked until client1 finishes.

 ACID: insert overwrite prevents create table command
 

 Key: HIVE-10242
 URL: https://issues.apache.org/jira/browse/HIVE-10242
 Project: Hive
  Issue Type: Bug
  Components: Transactions
Affects Versions: 1.0.0
Reporter: Eugene Koifman
Assignee: Eugene Koifman
 Fix For: 1.2.0

 Attachments: HIVE-10242.2.patch, HIVE-10242.3.patch, 
 HIVE-10242.4.patch, HIVE-10242.patch


 1. insert overwirte table DB.T1 select ... from T2: this takes X lock on 
 DB.T1 and S lock on T2.
 X lock makes sense because we don't want anyone reading T1 while it's 
 overwritten. S lock on T2 prevents if from being dropped while the query is 
 in progress.
 2. create table DB.T3: takes S lock on DB.
 This S lock gets blocked by X lock on T1. S lock prevents the DB from being 
 dropped while create table is executed.
 If the insert statement is long running, this blocks DDL ops on the same 
 database.  This is a usability issue.  
 There is no good reason why X lock on a table within a DB and S lock on DB 
 should be in conflict.  
 (this is different from a situation where X lock is on a partition and S lock 
 is on the table to which this partition belongs.  Here it makes sense.  
 Basically there is no SQL way to address all tables in a DB but you can 
 easily refer to all partitions of a table)



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-10270) Cannot use Decimal constants less than 0.1BD

2015-04-16 Thread Gunther Hagleitner (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-10270?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14498417#comment-14498417
 ] 

Gunther Hagleitner commented on HIVE-10270:
---

LGTM +1

 Cannot use Decimal constants less than 0.1BD
 

 Key: HIVE-10270
 URL: https://issues.apache.org/jira/browse/HIVE-10270
 Project: Hive
  Issue Type: Bug
  Components: Types
Reporter: Jason Dere
Assignee: Jason Dere
 Attachments: HIVE-10270.1.patch, HIVE-10270.2.patch, 
 HIVE-10270.3.patch, HIVE-10270.4.patch, HIVE-10270.5.patch


 {noformat}
 hive select 0.09765625BD;
 FAILED: IllegalArgumentException Decimal scale must be less than or equal to 
 precision
 {noformat}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9710) HiveServer2 should support cookie based authentication, when using HTTP transport.

2015-04-16 Thread Vaibhav Gumashta (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9710?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14498430#comment-14498430
 ] 

Vaibhav Gumashta commented on HIVE-9710:


[~hsubramaniyan] It will be good to add documentation in the release notes 
section on how to enable/disable and use this feature. Once [~leftylev] reviews 
it, we can add it to the wiki. Thanks.

 HiveServer2 should support cookie based authentication, when using HTTP 
 transport.
 --

 Key: HIVE-9710
 URL: https://issues.apache.org/jira/browse/HIVE-9710
 Project: Hive
  Issue Type: Improvement
  Components: HiveServer2
Affects Versions: 1.2.0
Reporter: Hari Sankar Sivarama Subramaniyan
Assignee: Hari Sankar Sivarama Subramaniyan
  Labels: TODOC1.2
 Fix For: 1.2.0

 Attachments: HIVE-9710.1.patch, HIVE-9710.2.patch, HIVE-9710.3.patch, 
 HIVE-9710.4.patch, HIVE-9710.5.patch, HIVE-9710.6.patch, HIVE-9710.7.patch, 
 HIVE-9710.8.patch


 HiveServer2 should generate cookies and validate the client cookie send to it 
 so that it need not perform User/Password or a Kerberos based authentication 
 on each HTTP request. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-10346) Tez on HBase has problems with settings again

2015-04-16 Thread Sergey Shelukhin (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-10346?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14498448#comment-14498448
 ] 

Sergey Shelukhin commented on HIVE-10346:
-

checkOutputSpec is actually an API on outputformat; the one of FileSink is just 
a translation call between HiveOutputFormatImpl and real output formats inside 
the filesink... so needs to be called.

 Tez on HBase has problems with settings again
 -

 Key: HIVE-10346
 URL: https://issues.apache.org/jira/browse/HIVE-10346
 Project: Hive
  Issue Type: Bug
Reporter: Sergey Shelukhin
Assignee: Sergey Shelukhin
 Attachments: HIVE-10346.patch






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-10356) LLAP: query80 fails with vectorization cast issue

2015-04-16 Thread Sergey Shelukhin (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-10356?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14498473#comment-14498473
 ] 

Sergey Shelukhin commented on HIVE-10356:
-

Is that one going to be fixed? :)

 LLAP: query80 fails with vectorization cast issue 
 --

 Key: HIVE-10356
 URL: https://issues.apache.org/jira/browse/HIVE-10356
 Project: Hive
  Issue Type: Sub-task
Reporter: Sergey Shelukhin
Assignee: Matt McCline

 Reducer 6 fails:
 {noformat}
 Error: Failure while running task:java.lang.RuntimeException: 
 java.lang.RuntimeException: org.apache.hadoop.hive.ql.metadata.HiveException: 
 Hive Runtime Error while processing vector batch (tag=0) 
 \N\N09.285817653506076E84.639990363237801E7-1.1814318134524737E8
 \N\N01.2847032699693155E96.41569738480791E7-5.956161019898126E8
 \N\N04.682909323885761E82.288924051203157E7-5.995957665973593E7
   at 
 org.apache.hadoop.hive.ql.exec.tez.TezProcessor.initializeAndRunProcessor(TezProcessor.java:171)
   at 
 org.apache.hadoop.hive.ql.exec.tez.TezProcessor.run(TezProcessor.java:137)
   at 
 org.apache.tez.runtime.LogicalIOProcessorRuntimeTask.run(LogicalIOProcessorRuntimeTask.java:332)
   at 
 org.apache.tez.runtime.task.TezTaskRunner$TaskRunnerCallable$1.run(TezTaskRunner.java:180)
   at 
 org.apache.tez.runtime.task.TezTaskRunner$TaskRunnerCallable$1.run(TezTaskRunner.java:172)
   at java.security.AccessController.doPrivileged(Native Method)
   at javax.security.auth.Subject.doAs(Subject.java:422)
   at 
 org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1657)
   at 
 org.apache.tez.runtime.task.TezTaskRunner$TaskRunnerCallable.callInternal(TezTaskRunner.java:172)
   at 
 org.apache.tez.runtime.task.TezTaskRunner$TaskRunnerCallable.callInternal(TezTaskRunner.java:168)
   at org.apache.tez.common.CallableWithNdc.call(CallableWithNdc.java:36)
   at java.util.concurrent.FutureTask.run(FutureTask.java:266)
   at 
 java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
   at 
 java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
   at java.lang.Thread.run(Thread.java:745)
 Caused by: java.lang.RuntimeException: 
 org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime Error while 
 processing vector batch (tag=0) 
 \N\N09.285817653506076E84.639990363237801E7-1.1814318134524737E8
 \N\N01.2847032699693155E96.41569738480791E7-5.956161019898126E8
 \N\N04.682909323885761E82.288924051203157E7-5.995957665973593E7
   at 
 org.apache.hadoop.hive.ql.exec.tez.ReduceRecordSource.pushRecord(ReduceRecordSource.java:267)
   at 
 org.apache.hadoop.hive.ql.exec.tez.ReduceRecordProcessor.run(ReduceRecordProcessor.java:254)
   at 
 org.apache.hadoop.hive.ql.exec.tez.TezProcessor.initializeAndRunProcessor(TezProcessor.java:148)
   ... 14 more
 Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime 
 Error while processing vector batch (tag=0) 
 \N\N09.285817653506076E84.639990363237801E7-1.1814318134524737E8
 \N\N01.2847032699693155E96.41569738480791E7-5.956161019898126E8
 \N\N04.682909323885761E82.288924051203157E7-5.995957665973593E7
   at 
 org.apache.hadoop.hive.ql.exec.tez.ReduceRecordSource.processVectors(ReduceRecordSource.java:394)
   at 
 org.apache.hadoop.hive.ql.exec.tez.ReduceRecordSource.pushRecord(ReduceRecordSource.java:252)
   ... 16 more
 Caused by: java.lang.ClassCastException: 
 org.apache.hadoop.hive.ql.exec.vector.DoubleColumnVector cannot be cast to 
 org.apache.hadoop.hive.ql.exec.vector.BytesColumnVector
   at 
 org.apache.hadoop.hive.ql.exec.vector.VectorGroupKeyHelper.copyGroupKey(VectorGroupKeyHelper.java:94)
   at 
 org.apache.hadoop.hive.ql.exec.vector.VectorGroupByOperator$ProcessingModeGroupBatches.processBatch(VectorGroupByOperator.java:729)
   at 
 org.apache.hadoop.hive.ql.exec.vector.VectorGroupByOperator.process(VectorGroupByOperator.java:878)
   at 
 org.apache.hadoop.hive.ql.exec.tez.ReduceRecordSource.processVectors(ReduceRecordSource.java:378)
   ... 17 more
 ]], Vertex failed as one or more tasks failed. failedTasks:1, Vertex 
 vertex_1428572510173_0231_1_24 [Reducer 5] killed/failed due to:null]Vertex 
 killed, vertexName=Reducer 6, vertexId=vertex_1428572510173_0231_1_25, 
 diagnostics=[Vertex received Kill while in RUNNING state., Vertex killed as 
 other vertex failed. failedTasks:0, Vertex vertex_1428572510173_0231_1_25 
 [Reducer 6] killed/failed due to:null]DAG failed due to vertex failure. 
 failedVertices:1 killedVertices:1
 {noformat}
 How to repro: run query80 on scale factor 200. I might look tomorrow to see 
 if this is specific to LLAP or not



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-4625) HS2 should not attempt to get delegation token from metastore if using embedded metastore

2015-04-16 Thread Hari Sankar Sivarama Subramaniyan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-4625?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Hari Sankar Sivarama Subramaniyan updated HIVE-4625:

Attachment: HIVE-4625.2.patch

 HS2 should not attempt to get delegation token from metastore if using 
 embedded metastore
 -

 Key: HIVE-4625
 URL: https://issues.apache.org/jira/browse/HIVE-4625
 Project: Hive
  Issue Type: Bug
  Components: HiveServer2
Affects Versions: 0.11.0
Reporter: Thejas M Nair
Assignee: Hari Sankar Sivarama Subramaniyan
 Attachments: HIVE-4625.1.patch, HIVE-4625.2.patch


 In kerberos secure mode, with doas enabled, Hive server2 tries to get 
 delegation token from metastore even if the metastore is being used in 
 embedded mode. 
 To avoid failure in that case, it uses catch block for 
 UnsupportedOperationException thrown that does nothing. But this leads to an 
 error being logged  by lower levels and can mislead users into thinking that 
 there is a problem.
 It should check if delegation token mode is supported with current 
 configuration before calling the function.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8136) Reduce table locking

2015-04-16 Thread Chaoyu Tang (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8136?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14498498#comment-14498498
 ] 

Chaoyu Tang commented on HIVE-8136:
---

[~Ferd] I have a question, when Hive processes a query, it needs fetch 
Table/Partition/Serde metadata from HMS, is this metadata cached at HMS client 
side (as Table/Partition) or to be retrieved from HMS each time it is needed, 
no matter the metadata is cached at HMS (as JDO entity object)? If the first 
one is the case, it is easer and I think the change should be fine. 
Otherwise, I think we should be more conservative and might even need add 
DDL_EXCLUSIVE on anything which may affect other query behaviors (e.g. ADDPROP 
etc). For example, querying data for an Avro or HBase backed tables, the Avro 
schema url/literal or HBase column mapping is specified as a property. Can we 
guarantee that Hive only need access these data once or all the metadata has 
been cached at its HMS client?

 Reduce table locking
 

 Key: HIVE-8136
 URL: https://issues.apache.org/jira/browse/HIVE-8136
 Project: Hive
  Issue Type: Sub-task
Reporter: Brock Noland
Assignee: Ferdinand Xu
 Attachments: HIVE-8136.1.patch, HIVE-8136.patch


 When using ZK for concurrency control, some statements require an exclusive 
 table lock when they are atomic. Such as setting a tables location.
 This JIRA is to analyze the scope of statements like ALTER TABLE and see if 
 we can reduce the locking required.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)