[jira] [Commented] (HIVE-9480) Build UDF TRUNC to implement FIRST_DAY as compared with LAST_DAY

2015-03-17 Thread Alexander Pivovarov (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9480?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14366640#comment-14366640
 ] 

Alexander Pivovarov commented on HIVE-9480:
---

Added the function description to the wiki 
https://cwiki.apache.org/confluence/display/Hive/LanguageManual+UDF#LanguageManualUDF-DateFunctions

 Build UDF TRUNC to implement FIRST_DAY as compared with LAST_DAY
 

 Key: HIVE-9480
 URL: https://issues.apache.org/jira/browse/HIVE-9480
 Project: Hive
  Issue Type: Improvement
  Components: UDF
Reporter: Xiaobing Zhou
Assignee: Xiaobing Zhou
  Labels: TODOC1.2
 Fix For: 1.2.0

 Attachments: HIVE-9480.1.patch, HIVE-9480.3.patch, HIVE-9480.4.patch, 
 HIVE-9480.5.patch, HIVE-9480.6.patch, HIVE-9480.7.patch, HIVE-9480.8.patch, 
 HIVE-9480.9.patch


 Hive already supports LAST_DAY UDF, in some cases, FIRST_DAY is necessary to 
 do date/timestamp related computation. This JIRA is to track such an 
 implementation. Choose to impl TRUNC, a more standard way to get the first 
 day of a a month, e.g., SELECT TRUNC('2009-12-12', 'MM'); will return 
 2009-12-01, SELECT TRUNC('2009-12-12', 'YEAR'); will return 2009-01-01.
 BTW, this TRUNC is not as feature complete as aligned with Oracle one. only 
 'MM' and 'YEAR' are supported as format, however, it's a base to add on other 
 formats.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9975) Renaming a nonexisting partition should not throw out NullPointerException

2015-03-17 Thread Szehon Ho (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9975?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14366583#comment-14366583
 ] 

Szehon Ho commented on HIVE-9975:
-

+1, thanks Chaoyu

 Renaming a nonexisting partition should not throw out NullPointerException
 --

 Key: HIVE-9975
 URL: https://issues.apache.org/jira/browse/HIVE-9975
 Project: Hive
  Issue Type: Bug
Reporter: Chaoyu Tang
Assignee: Chaoyu Tang
Priority: Minor
 Attachments: HIVE-9975.1.patch, HIVE-9975.2.patch, HIVE-9975.patch


 Renaming a nonexisting partition should not throw out NullPointerException. 
 create table testpart (col1 int, col2 string, col3 string) partitioned by 
 (part string);
 alter table testpart partition (part = 'nonexisting') rename to partition 
 (part = 'existing');
 we get NPE like following:
 {code}
 15/03/16 10:16:11 ERROR exec.DDLTask: java.lang.NullPointerException
   at 
 org.apache.hadoop.hive.ql.exec.DDLTask.renamePartition(DDLTask.java:944)
   at org.apache.hadoop.hive.ql.exec.DDLTask.execute(DDLTask.java:350)
   at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:160)
   at 
 org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:88)
   at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:1642)
   at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:1402)
   at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1187)
   at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1053)
   at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1043)
   at 
 org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:207)
   at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:159)
   at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:370)
   at 
 org.apache.hadoop.hive.cli.CliDriver.executeDriver(CliDriver.java:754)
   at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:675)
   at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:615)
 FAILED: Execution Error, return code 1 from 
 org.apache.hadoop.hive.ql.exec.DDLTask. null
 {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9984) JoinReorder's getOutputSize is exponential

2015-03-17 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9984?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14364731#comment-14364731
 ] 

Hive QA commented on HIVE-9984:
---



{color:green}Overall{color}: +1 all checks pass

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12704937/HIVE-9984.2.patch

{color:green}SUCCESS:{color} +1 7769 tests passed

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/3050/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/3050/console
Test logs: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-3050/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12704937 - PreCommit-HIVE-TRUNK-Build

 JoinReorder's getOutputSize is exponential
 --

 Key: HIVE-9984
 URL: https://issues.apache.org/jira/browse/HIVE-9984
 Project: Hive
  Issue Type: Bug
Reporter: Gunther Hagleitner
Assignee: Gunther Hagleitner
 Attachments: HIVE-9984.1.patch, HIVE-9984.2.patch


 Found by [~mmokhtar]. Causes major issues in large plans (50+ joins). Simple 
 fix would be to memoize the recursion. There should also be a flag to switch 
 this opt off.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9993) Retrying task could use cached bad operators [Spark Branch]

2015-03-17 Thread Szehon Ho (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9993?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14365946#comment-14365946
 ] 

Szehon Ho commented on HIVE-9993:
-

+1, good we caught it

 Retrying task could use cached bad operators [Spark Branch]
 ---

 Key: HIVE-9993
 URL: https://issues.apache.org/jira/browse/HIVE-9993
 Project: Hive
  Issue Type: Bug
Affects Versions: spark-branch
Reporter: Jimmy Xiang
Assignee: Jimmy Xiang
 Fix For: spark-branch

 Attachments: HIVE-9993.1-spark.patch


 For a Spark task, it could be retried on the same executor in case some 
 failures. In retrying, the cache task could be used. Since the operators in 
 the task are already initialized, they won't be initialized again. The 
 partial data in these operators could lead to wrong final results.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9986) LLAP: EOFException in reader

2015-03-17 Thread Sergey Shelukhin (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9986?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14366040#comment-14366040
 ] 

Sergey Shelukhin commented on HIVE-9986:


May be fixed in HIVE-9978, unless it's a different EOF. I'll take a look at the 
logs

 LLAP: EOFException in reader
 

 Key: HIVE-9986
 URL: https://issues.apache.org/jira/browse/HIVE-9986
 Project: Hive
  Issue Type: Sub-task
Reporter: Gopal V
Assignee: Sergey Shelukhin
 Fix For: llap


 From HIVE-9979
 {noformat}
 2015-03-16 10:20:51,439 
 [pool-2-thread-3(container_1_1141_01_000192_gopal_20150316102020_c8c92488-6a61-401e-8298-401dace286dc:1_Map
  1_191_0)] INFO org.apache.hadoop.hive.ql.io.orc.EncodedReaderImpl: Getting 
 data for column 9 RG 112 stream DATA at 62278935, 1057137 index position 0: 
 compressed [62614934, 63139228)
 2015-03-16 10:20:51,439 
 [pool-2-thread-6(container_1_1141_01_000211_gopal_20150316102020_c8c92488-6a61-401e-8298-401dace286dc:1_Map
  1_210_0)] INFO org.apache.hadoop.hive.ql.io.orc.EncodedReaderImpl: Getting 
 stripe-level stream [LENGTH, kind: DICTIONARY_V2
 dictionarySize: 3
 ] for column 9 RG 91 at 64139927, 5
 ...
 Caused by: java.io.EOFException
 at 
 org.apache.hadoop.hive.ql.io.orc.RecordReaderUtils.readDirect(RecordReaderUtils.java:286)
 at 
 org.apache.hadoop.hive.ql.io.orc.RecordReaderUtils.readDiskRanges(RecordReaderUtils.java:266)
 at 
 org.apache.hadoop.hive.ql.io.orc.EncodedReaderImpl.readEncodedColumns(EncodedReaderImpl.java:234)
 at 
 org.apache.hadoop.hive.llap.io.encoded.OrcEncodedDataReader.callInternal(OrcEncodedDataReader.java:280)
 at 
 org.apache.hadoop.hive.llap.io.encoded.OrcEncodedDataReader.callInternal(OrcEncodedDataReader.java:44)
 at 
 org.apache.hadoop.hive.common.CallableWithNdc.call(CallableWithNdc.java:37)
 ... 4 more
 {noformat}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Resolved] (HIVE-9729) LLAP: design and implement proper metadata cache

2015-03-17 Thread Sergey Shelukhin (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9729?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sergey Shelukhin resolved HIVE-9729.

Resolution: Fixed

 LLAP: design and implement proper metadata cache
 

 Key: HIVE-9729
 URL: https://issues.apache.org/jira/browse/HIVE-9729
 Project: Hive
  Issue Type: Sub-task
Reporter: Sergey Shelukhin
Assignee: Sergey Shelukhin
 Fix For: llap


 Simple approach: add external priorities to data cache, read metadata parts 
 of orc file into it. Advantage: simple; consistent management (no need to 
 coordinate sizes and eviction between data and metadata caches, etc); 
 disadvantage - have to decode every time.
 Maybe add decoded metadata cache on top - fixed size, small and 
 opportunistic? Or some other approach.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9996) [CBO] Generate appropriate join operator as per algorithm selected by CBO

2015-03-17 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9996?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-9996:
---
Attachment: HIVE-9996.cbo.patch

 [CBO] Generate appropriate join operator as per algorithm selected by CBO
 -

 Key: HIVE-9996
 URL: https://issues.apache.org/jira/browse/HIVE-9996
 Project: Hive
  Issue Type: Task
  Components: CBO, Query Planning
Affects Versions: cbo-branch
Reporter: Ashutosh Chauhan
Assignee: Ashutosh Chauhan
 Attachments: HIVE-9996.cbo.patch






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9989) hive on tez group by NPE

2015-03-17 Thread tangjunjie (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9989?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

tangjunjie updated HIVE-9989:
-
Attachment: test.sql
errorlog

 hive on tez group by NPE
 

 Key: HIVE-9989
 URL: https://issues.apache.org/jira/browse/HIVE-9989
 Project: Hive
  Issue Type: Bug
Affects Versions: 0.14.0
 Environment: hive 0.14 tez 0.53
Reporter: tangjunjie
 Attachments: errorlog, test.sql


 set hive.execution.engine=tez   run sql throw NPE   when I modify my sql (see 
 my test.sql file line 14) to  cast(NULL AS BIGINT).The 
 sql run ok.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9971) Clean up operator class

2015-03-17 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9971?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14364893#comment-14364893
 ] 

Hive QA commented on HIVE-9971:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12705002/HIVE-9971.4.patch

{color:red}ERROR:{color} -1 due to 2 failed/errored test(s), 7769 tests executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestMiniTezCliDriver.testCliDriver_tez_bmj_schema_evolution
org.apache.hadoop.hive.cli.TestMiniTezCliDriver.testCliDriver_tez_smb_main
{noformat}

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/3052/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/3052/console
Test logs: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-3052/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 2 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12705002 - PreCommit-HIVE-TRUNK-Build

 Clean up operator class
 ---

 Key: HIVE-9971
 URL: https://issues.apache.org/jira/browse/HIVE-9971
 Project: Hive
  Issue Type: Bug
Reporter: Gunther Hagleitner
Assignee: Gunther Hagleitner
 Attachments: HIVE-9971.1.patch, HIVE-9971.2.patch, HIVE-9971.3.patch, 
 HIVE-9971.4.patch


 This is mostly cleanup although it does enhance the pipeline in one respect. 
 It introduces asyn init for operators and uses it for hash table loading 
 where desired.
 There's a bunch of weird code associated with the operator class:
 - initialize isn't recursive, rather initializeOp is supposed to call 
 initializeChildren. That has led to bugs in the past.
 - setExecContext and passExecContext. Both are recursive, but passExecContext 
 calls setExecContext and then recurses again. Boo.
 - lots of (getChildren() != null) although that can't happen anymore
 - TezCacheAccess is a hack. We should just leave init of inputs up to the 
 operator that needs it.
 - Need some sanity checks that make sure that operators were all initialized.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9989) hive on tez group by NPE

2015-03-17 Thread tangjunjie (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9989?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

tangjunjie updated HIVE-9989:
-
Attachment: errorlog.TXT

 hive on tez group by NPE
 

 Key: HIVE-9989
 URL: https://issues.apache.org/jira/browse/HIVE-9989
 Project: Hive
  Issue Type: Bug
Affects Versions: 0.14.0
 Environment: hive 0.14 tez 0.53
Reporter: tangjunjie
 Attachments: errorlog.TXT, test.sql


 set hive.execution.engine=tez   run sql throw NPE   when I modify my sql (see 
 my test.sql file line 14) to  cast(NULL AS BIGINT).The 
 sql run ok.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9989) hive on tez group by NPE

2015-03-17 Thread tangjunjie (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9989?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

tangjunjie updated HIVE-9989:
-
Attachment: (was: errorlog)

 hive on tez group by NPE
 

 Key: HIVE-9989
 URL: https://issues.apache.org/jira/browse/HIVE-9989
 Project: Hive
  Issue Type: Bug
Affects Versions: 0.14.0
 Environment: hive 0.14 tez 0.53
Reporter: tangjunjie
 Attachments: errorlog.TXT, test.sql


 set hive.execution.engine=tez   run sql throw NPE   when I modify my sql (see 
 my test.sql file line 14) to  cast(NULL AS BIGINT).The 
 sql run ok.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9543) MetaException(message:Metastore contains multiple versions)

2015-03-17 Thread Pavan Srinivas (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9543?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14364857#comment-14364857
 ] 

Pavan Srinivas commented on HIVE-9543:
--

Hey [~brocknoland], as [~lijy83] pointed out, when ever there is a 
MysqlException, a duplicate entry will be created. The problem here is with 
https://db.apache.org/jdo/api20/apidocs/javax/jdo/Query.html#execute%28%29 
execute() function of Query.java.  execute() does not throw any exceptions. 

Some reproducible steps: 
1) Make sure metastore db is reachablerunning. 
2) Enable debugger and stop at a breakpoint inside checkSchema().
3) Stop the metastore db or make it unreachable. 
4) Continue on the debugger, you would see execute call of getMSchemaVersion() 
does not throw exceptions of DB reachability. It just returns an empty array as 
if there are no entries in the table. Thus making the getMSchemaVersion() throw 
NoSuchObjectException. 


This issue happens when we have a flacky network connection to DB. 



 MetaException(message:Metastore contains multiple versions)
 ---

 Key: HIVE-9543
 URL: https://issues.apache.org/jira/browse/HIVE-9543
 Project: Hive
  Issue Type: Bug
  Components: Metastore
Affects Versions: 0.13.1
Reporter: Junyong Li

 When i run bin/hive command, i got the following exception:
 {noformat}
 Logging initialized using configuration in 
 jar:file:/home/hadoop/apache-hive-0.13.1-bin/lib/hive-common-0.13.1.jar!/hive-log4j.properties
 Exception in thread main java.lang.RuntimeException: 
 java.lang.RuntimeException: Unable to instantiate 
 org.apache.hadoop.hive.metastore.HiveMetaStoreClient
 at 
 org.apache.hadoop.hive.ql.session.SessionState.start(SessionState.java:346)
 at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:681)
 at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:625)
 at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
 at 
 sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
 at 
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
 at java.lang.reflect.Method.invoke(Method.java:606)
 at org.apache.hadoop.util.RunJar.main(RunJar.java:212)
 Caused by: java.lang.RuntimeException: Unable to instantiate 
 org.apache.hadoop.hive.metastore.HiveMetaStoreClient
 at 
 org.apache.hadoop.hive.metastore.MetaStoreUtils.newInstance(MetaStoreUtils.java:1412)
 at 
 org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.init(RetryingMetaStoreClient.java:62)
 at 
 org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:72)
 at 
 org.apache.hadoop.hive.ql.metadata.Hive.createMetaStoreClient(Hive.java:2453)
 at org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:2465)
 at 
 org.apache.hadoop.hive.ql.session.SessionState.start(SessionState.java:340)
 ... 7 more
 Caused by: java.lang.reflect.InvocationTargetException
 at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
 at 
 sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:57)
 at 
 sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
 at java.lang.reflect.Constructor.newInstance(Constructor.java:526)
 at 
 org.apache.hadoop.hive.metastore.MetaStoreUtils.newInstance(MetaStoreUtils.java:1410)
 ... 12 more
 Caused by: MetaException(message:Metastore contains multiple versions)
 at 
 org.apache.hadoop.hive.metastore.ObjectStore.getMSchemaVersion(ObjectStore.java:6368)
 at 
 org.apache.hadoop.hive.metastore.ObjectStore.getMetaStoreSchemaVersion(ObjectStore.java:6330)
 at 
 org.apache.hadoop.hive.metastore.ObjectStore.checkSchema(ObjectStore.java:6289)
 at 
 org.apache.hadoop.hive.metastore.ObjectStore.verifySchema(ObjectStore.java:6277)
 at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
 at 
 sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
 at 
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
 at java.lang.reflect.Method.invoke(Method.java:606)
 at 
 org.apache.hadoop.hive.metastore.RawStoreProxy.invoke(RawStoreProxy.java:108)
 at com.sun.proxy.$Proxy9.verifySchema(Unknown Source)
 at 
 org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.getMS(HiveMetaStore.java:476)
 at 
 org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.createDefaultDB(HiveMetaStore.java:523)
 at 
 org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.init(HiveMetaStore.java:397)
 at 
 org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.init(HiveMetaStore.java:356)
 at 
 

[jira] [Commented] (HIVE-9985) Vectorization: NPE for added columns in ORC non-partitioned tables

2015-03-17 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9985?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14364800#comment-14364800
 ] 

Hive QA commented on HIVE-9985:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12704929/HIVE-9985.01.patch

{color:red}ERROR:{color} -1 due to 5 failed/errored test(s), 7770 tests executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_vector_partitioned_date_time
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_vectorization_part
org.apache.hadoop.hive.cli.TestMiniTezCliDriver.testCliDriver_vector_partitioned_date_time
org.apache.hadoop.hive.cli.TestMiniTezCliDriver.testCliDriver_vectorization_part
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_vectorization_part
{noformat}

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/3051/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/3051/console
Test logs: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-3051/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 5 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12704929 - PreCommit-HIVE-TRUNK-Build

 Vectorization: NPE for added columns in ORC non-partitioned tables
 --

 Key: HIVE-9985
 URL: https://issues.apache.org/jira/browse/HIVE-9985
 Project: Hive
  Issue Type: Bug
  Components: Vectorization
Affects Versions: 0.14.0
Reporter: Matt McCline
Assignee: Matt McCline
Priority: Critical
 Attachments: HIVE-9985.01.patch


 If you add STRING columns to a non-partitioned table (ORC format) and try to 
 read the added STRING column using vectorization, you will get a 
 NullPointerException.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9989) hive on tez group by NPE

2015-03-17 Thread tangjunjie (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9989?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

tangjunjie updated HIVE-9989:
-
Description: 
set hive.execution.engine=tez   run sql throw NPE   when I modify my sql (see 
my test.sql file line 14) to  cast(NULL AS BIGINT).The 
sql run ok.

  was:set hive.execution.engine=tez   run sql throw NPE 


 hive on tez group by NPE
 

 Key: HIVE-9989
 URL: https://issues.apache.org/jira/browse/HIVE-9989
 Project: Hive
  Issue Type: Bug
Affects Versions: 0.14.0
 Environment: hive 0.14 tez 0.53
Reporter: tangjunjie

 set hive.execution.engine=tez   run sql throw NPE   when I modify my sql (see 
 my test.sql file line 14) to  cast(NULL AS BIGINT).The 
 sql run ok.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9792) Support interval type in expressions/predicates

2015-03-17 Thread Ashutosh Chauhan (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9792?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14365600#comment-14365600
 ] 

Ashutosh Chauhan commented on HIVE-9792:


Patch mostly looks good.
Left few minor comments on RB.
Thanks for excellent system  unit tests!

 Support interval type in expressions/predicates 
 

 Key: HIVE-9792
 URL: https://issues.apache.org/jira/browse/HIVE-9792
 Project: Hive
  Issue Type: Sub-task
  Components: Types
Reporter: Jason Dere
Assignee: Jason Dere
 Attachments: HIVE-9792.1.patch, HIVE-9792.2.patch, HIVE-9792.3.patch, 
 HIVE-9792.4.patch, HIVE-9792.5.patch, HIVE-9792.6.patch


 Provide partial support for the interval year-month/interval day-time types 
 in Hive. Intervals will be usable in expressions/predicates/joins:
 {noformat}
   select birthdate + interval '30-0' year to month as thirtieth_birthday
   from table
   where (current_timestamp - ts1  interval '3 0:0:0' day to second)
 {noformat}
 This stops short of adding making the interval types usable as a storable 
 column type.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9991) Cannot do a SELECT on external tables that are on S3 due to Encryption error

2015-03-17 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9991?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14365779#comment-14365779
 ] 

Hive QA commented on HIVE-9991:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12705104/HIVE-9991.1.patch

{color:red}ERROR:{color} -1 due to 2 failed/errored test(s), 7770 tests executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_encryption_unencrypted_nonhdfs_external_tables
org.apache.hadoop.hive.cli.TestNegativeCliDriver.testNegativeCliDriver_fs_default_name2
{noformat}

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/3058/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/3058/console
Test logs: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-3058/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 2 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12705104 - PreCommit-HIVE-TRUNK-Build

 Cannot do a SELECT on external tables that are on S3 due to Encryption error
 

 Key: HIVE-9991
 URL: https://issues.apache.org/jira/browse/HIVE-9991
 Project: Hive
  Issue Type: Bug
Affects Versions: 1.0.0
Reporter: Sergio Peña
Assignee: Sergio Peña
 Attachments: HIVE-9991.1.patch, HIVE-9991.2.patch


 I cannot do any select query on external tables that are not part of HDFS. 
 For example S3.
 {code}
 Select * from my_table limit 10;
 FAILED: SemanticException Unable to determine if s3n://my-bucket/is 
 encrypted: java.lang.IllegalArgumentException: Wrong FS: s3n://my-bucket/, 
 expected: hdfs://0.0.0.0:8020
 {code}
 This error is due to a internal function that checks if a table is encrypted 
 or not. This is only supported on HDFS files, but the check is happening on 
 any external table as well causing the above error.
 To fix this, we should check for encrypted tables only for HDFS tables. And 
 skip the check for any other file schema.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9934) Vulnerability in LdapAuthenticationProviderImpl enables HiveServer2 client to degrade the authentication mechanism to none, allowing authentication without password

2015-03-17 Thread Xuefu Zhang (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9934?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14365832#comment-14365832
 ] 

Xuefu Zhang commented on HIVE-9934:
---

[~prasadm], I think lacking @Test seems fine in this case, as the class is 
extended from TestCase. I also saw the added test case was run in previous test 
result. Thus, patch #3 is good as far as I can see. Let me know if you see 
differently.
 

 Vulnerability in LdapAuthenticationProviderImpl enables HiveServer2 client to 
 degrade the authentication mechanism to none, allowing authentication 
 without password
 --

 Key: HIVE-9934
 URL: https://issues.apache.org/jira/browse/HIVE-9934
 Project: Hive
  Issue Type: Bug
  Components: Security
Affects Versions: 1.1.0
Reporter: Chao
Assignee: Chao
 Attachments: HIVE-9934.1.patch, HIVE-9934.2.patch, HIVE-9934.3.patch, 
 HIVE-9934.3.patch, HIVE-9934.4.patch


 Vulnerability in LdapAuthenticationProviderImpl enables HiveServer2 client to 
 degrade the authentication mechanism to none, allowing authentication 
 without password.
 See: http://docs.oracle.com/javase/jndi/tutorial/ldap/security/simple.html
 “If you supply an empty string, an empty byte/char array, or null to the 
 Context.SECURITY_CREDENTIALS environment property, then the authentication 
 mechanism will be none. This is because the LDAP requires the password to 
 be nonempty for simple authentication. The protocol automatically converts 
 the authentication to none if a password is not supplied.”
  
 Since the LdapAuthenticationProviderImpl.Authenticate method is relying on a 
 NamingException being thrown during creation of initial context, it does not 
 fail when the context result is an “unauthenticated” positive response from 
 the LDAP server. The end result is, one can authenticate with HiveServer2 
 using the LdapAuthenticationProviderImpl with only a user name and an empty 
 password.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9991) Cannot do a SELECT on external tables that are on S3 due to Encryption error

2015-03-17 Thread Xuefu Zhang (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9991?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14365606#comment-14365606
 ] 

Xuefu Zhang commented on HIVE-9991:
---

+1 pending on test

 Cannot do a SELECT on external tables that are on S3 due to Encryption error
 

 Key: HIVE-9991
 URL: https://issues.apache.org/jira/browse/HIVE-9991
 Project: Hive
  Issue Type: Bug
Affects Versions: 1.0.0
Reporter: Sergio Peña
Assignee: Sergio Peña
 Attachments: HIVE-9991.1.patch


 I cannot do any select query on external tables that are not part of HDFS. 
 For example S3.
 {code}
 Select * from my_table limit 10;
 FAILED: SemanticException Unable to determine if s3n://my-bucket/is 
 encrypted: java.lang.IllegalArgumentException: Wrong FS: s3n://my-bucket/, 
 expected: hdfs://0.0.0.0:8020
 {code}
 This error is due to a internal function that checks if a table is encrypted 
 or not. This is only supported on HDFS files, but the check is happening on 
 any external table as well causing the above error.
 To fix this, we should check for encrypted tables only for HDFS tables. And 
 skip the check for any other file schema.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9975) Renaming a nonexisting partition should not throw out NullPointerException

2015-03-17 Thread Chaoyu Tang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9975?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chaoyu Tang updated HIVE-9975:
--
Attachment: HIVE-9975.1.patch

Fixed alter_rename_partition_failure.q negative test failure. The error message 
should be:
FAILED: Execution Error, return code 1 from 
org.apache.hadoop.hive.ql.exec.DDLTask. Renamed partition does not exist in 
default.alter_rename_partition
instead of the previous 
FAILED: Execution Error, return code 1 from 
org.apache.hadoop.hive.ql.exec.DDLTask. null

 Renaming a nonexisting partition should not throw out NullPointerException
 --

 Key: HIVE-9975
 URL: https://issues.apache.org/jira/browse/HIVE-9975
 Project: Hive
  Issue Type: Bug
Reporter: Chaoyu Tang
Assignee: Chaoyu Tang
Priority: Minor
 Attachments: HIVE-9975.1.patch, HIVE-9975.patch


 Renaming a nonexisting partition should not throw out NullPointerException. 
 create table testpart (col1 int, col2 string, col3 string) partitioned by 
 (part string);
 alter table testpart partition (part = 'nonexisting') rename to partition 
 (part = 'existing');
 we get NPE like following:
 {code}
 15/03/16 10:16:11 ERROR exec.DDLTask: java.lang.NullPointerException
   at 
 org.apache.hadoop.hive.ql.exec.DDLTask.renamePartition(DDLTask.java:944)
   at org.apache.hadoop.hive.ql.exec.DDLTask.execute(DDLTask.java:350)
   at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:160)
   at 
 org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:88)
   at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:1642)
   at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:1402)
   at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1187)
   at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1053)
   at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1043)
   at 
 org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:207)
   at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:159)
   at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:370)
   at 
 org.apache.hadoop.hive.cli.CliDriver.executeDriver(CliDriver.java:754)
   at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:675)
   at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:615)
 FAILED: Execution Error, return code 1 from 
 org.apache.hadoop.hive.ql.exec.DDLTask. null
 {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9993) Retrying task could use cached bad operators [Spark Branch]

2015-03-17 Thread Jimmy Xiang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9993?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jimmy Xiang updated HIVE-9993:
--
Attachment: HIVE-9993.1-spark.patch

 Retrying task could use cached bad operators [Spark Branch]
 ---

 Key: HIVE-9993
 URL: https://issues.apache.org/jira/browse/HIVE-9993
 Project: Hive
  Issue Type: Bug
Affects Versions: spark-branch
Reporter: Jimmy Xiang
Assignee: Jimmy Xiang
 Fix For: spark-branch

 Attachments: HIVE-9993.1-spark.patch


 For a Spark task, it could be retried on the same executor in case some 
 failures. In retrying, the cache task could be used. Since the operators in 
 the task are already initialized, they won't be initialized again. The 
 partial data in these operators could lead to wrong final results.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9978) LLAP: OrcColumnVectorProducer should handle reading isPresent columns only

2015-03-17 Thread Sergey Shelukhin (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9978?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sergey Shelukhin updated HIVE-9978:
---
Attachment: HIVE-9978.patch

 LLAP: OrcColumnVectorProducer should handle reading isPresent columns only
 --

 Key: HIVE-9978
 URL: https://issues.apache.org/jira/browse/HIVE-9978
 Project: Hive
  Issue Type: Sub-task
  Components: File Formats
Affects Versions: llap
Reporter: Gopal V
Assignee: Sergey Shelukhin
 Attachments: HIVE-9978.patch


 LlapInputFormat does not understand the difference between empty columns list 
 and null columns list.
 The empty columns list indicates no columns read except the root struct 
 isPresent column, while the null columns list indicates that all columns are 
 being read.
 {code}
 select count(1) from store_sales join date_dim on ss_sold_date_sk = d_date_sk 
 where d_date = '1998-01-01';
 ...
 Caused by: java.lang.NullPointerException
 at 
 org.apache.hadoop.hive.llap.io.decode.OrcColumnVectorProducer.createReadPipeline(OrcColumnVectorProducer.java:72)
 at 
 org.apache.hadoop.hive.llap.io.api.impl.LlapInputFormat$LlapRecordReader.startRead(LlapInputFormat.java:181)
 at 
 org.apache.hadoop.hive.llap.io.api.impl.LlapInputFormat$LlapRecordReader.next(LlapInputFormat.java:140)
 at 
 org.apache.hadoop.hive.llap.io.api.impl.LlapInputFormat$LlapRecordReader.next(LlapInputFormat.java:99)
 at 
 org.apache.hadoop.hive.ql.io.HiveContextAwareRecordReader.doNext(HiveContextAwareRecordReader.java:350)
 ... 22 more
 {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9934) Vulnerability in LdapAuthenticationProviderImpl enables HiveServer2 client to degrade the authentication mechanism to none, allowing authentication without password

2015-03-17 Thread Xuefu Zhang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9934?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Xuefu Zhang updated HIVE-9934:
--
Attachment: HIVE-9934.4.patch

Update the patch, adding @Test annotation.

 Vulnerability in LdapAuthenticationProviderImpl enables HiveServer2 client to 
 degrade the authentication mechanism to none, allowing authentication 
 without password
 --

 Key: HIVE-9934
 URL: https://issues.apache.org/jira/browse/HIVE-9934
 Project: Hive
  Issue Type: Bug
  Components: Security
Affects Versions: 1.1.0
Reporter: Chao
Assignee: Chao
 Attachments: HIVE-9934.1.patch, HIVE-9934.2.patch, HIVE-9934.3.patch, 
 HIVE-9934.3.patch, HIVE-9934.4.patch


 Vulnerability in LdapAuthenticationProviderImpl enables HiveServer2 client to 
 degrade the authentication mechanism to none, allowing authentication 
 without password.
 See: http://docs.oracle.com/javase/jndi/tutorial/ldap/security/simple.html
 “If you supply an empty string, an empty byte/char array, or null to the 
 Context.SECURITY_CREDENTIALS environment property, then the authentication 
 mechanism will be none. This is because the LDAP requires the password to 
 be nonempty for simple authentication. The protocol automatically converts 
 the authentication to none if a password is not supplied.”
  
 Since the LdapAuthenticationProviderImpl.Authenticate method is relying on a 
 NamingException being thrown during creation of initial context, it does not 
 fail when the context result is an “unauthenticated” positive response from 
 the LDAP server. The end result is, one can authenticate with HiveServer2 
 using the LdapAuthenticationProviderImpl with only a user name and an empty 
 password.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9828) Semantic analyzer does not capture view parent entity for tables referred in view with union all

2015-03-17 Thread Prasad Mujumdar (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9828?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Prasad Mujumdar updated HIVE-9828:
--
Attachment: HIVE-9828.1-npf.patch

Reattaching for pre-commit run

 Semantic analyzer does not capture view parent entity for tables referred in 
 view with union all 
 -

 Key: HIVE-9828
 URL: https://issues.apache.org/jira/browse/HIVE-9828
 Project: Hive
  Issue Type: Bug
  Components: Parser
Affects Versions: 1.1.0
Reporter: Prasad Mujumdar
 Fix For: 1.2.0

 Attachments: HIVE-9828.1-npf.patch, HIVE-9828.1-npf.patch


 Hive compiler adds tables used in a view definition in the input entity list, 
 with the view as parent entity for the table.
 In case of a view with union all query, this is not being done property. For 
 example,
 {noformat}
 create view view1 as select t.id from (select tab1.id from db.tab1 union all 
 select tab2.id from db.tab2 ) t;
 {noformat}
 This query will capture tab1 and tab2 as read entity without view1 as parent.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9991) Cannot do a SELECT on external tables that are on S3 due to Encryption error

2015-03-17 Thread JIRA

 [ 
https://issues.apache.org/jira/browse/HIVE-9991?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sergio Peña updated HIVE-9991:
--
Attachment: HIVE-9991.2.patch

New patch with just a small changes:
- Change equals() to equalsIgnoreCase()

 Cannot do a SELECT on external tables that are on S3 due to Encryption error
 

 Key: HIVE-9991
 URL: https://issues.apache.org/jira/browse/HIVE-9991
 Project: Hive
  Issue Type: Bug
Affects Versions: 1.0.0
Reporter: Sergio Peña
Assignee: Sergio Peña
 Attachments: HIVE-9991.1.patch, HIVE-9991.2.patch


 I cannot do any select query on external tables that are not part of HDFS. 
 For example S3.
 {code}
 Select * from my_table limit 10;
 FAILED: SemanticException Unable to determine if s3n://my-bucket/is 
 encrypted: java.lang.IllegalArgumentException: Wrong FS: s3n://my-bucket/, 
 expected: hdfs://0.0.0.0:8020
 {code}
 This error is due to a internal function that checks if a table is encrypted 
 or not. This is only supported on HDFS files, but the check is happening on 
 any external table as well causing the above error.
 To fix this, we should check for encrypted tables only for HDFS tables. And 
 skip the check for any other file schema.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9964) CBO (Calcite Return Path): Traits propagation for Aggregate operator [CBO Branch]

2015-03-17 Thread Jesus Camacho Rodriguez (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9964?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14365824#comment-14365824
 ] 

Jesus Camacho Rodriguez commented on HIVE-9964:
---

[~ashutoshc], I think this patch could go in so Mostafa gets the correct trait 
set propagation.

 CBO (Calcite Return Path): Traits propagation for Aggregate operator [CBO 
 Branch]
 -

 Key: HIVE-9964
 URL: https://issues.apache.org/jira/browse/HIVE-9964
 Project: Hive
  Issue Type: Sub-task
  Components: CBO
Affects Versions: cbo-branch
Reporter: Jesus Camacho Rodriguez
Assignee: Jesus Camacho Rodriguez
 Fix For: cbo-branch

 Attachments: HIVE-9964.cbo.patch






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Resolved] (HIVE-9992) Operator Precedence is incorrect. bitwise XOR should be after minus/plus

2015-03-17 Thread Alexander Pivovarov (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9992?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alexander Pivovarov resolved HIVE-9992.
---
Resolution: Not a Problem

I checked MySQL doc again. Bitwise XOR ^ has higher precedence than -/+ in 
MySQL.
http://dev.mysql.com/doc/refman/5.0/en/operator-precedence.html
Hive implemented similar way.

 Operator Precedence is incorrect. bitwise XOR should be after minus/plus
 

 Key: HIVE-9992
 URL: https://issues.apache.org/jira/browse/HIVE-9992
 Project: Hive
  Issue Type: Bug
  Components: Query Processor
Affects Versions: 1.2.0
Reporter: Alexander Pivovarov
Assignee: Alexander Pivovarov

 bitwise XOR ^ has higher precedence than -/+ in hive
 But in java bitwise XOR has lower precedence
 http://docs.oracle.com/javase/tutorial/java/nutsandbolts/operators.html
 {code}
 // now
 select 4 - 2 ^ 8 - 7;
 -13
 // should be
 select 4 - 2 ^ 8 - 7;
 3
 {code}
 need to fix
 IdentifiersParser.g  BITWISEXOR



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9975) Renaming a nonexisting partition should not throw out NullPointerException

2015-03-17 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9975?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14365598#comment-14365598
 ] 

Hive QA commented on HIVE-9975:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12704962/HIVE-9975.patch

{color:red}ERROR:{color} -1 due to 1 failed/errored test(s), 7769 tests executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestNegativeCliDriver.testNegativeCliDriver_alter_rename_partition_failure
{noformat}

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/3057/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/3057/console
Test logs: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-3057/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 1 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12704962 - PreCommit-HIVE-TRUNK-Build

 Renaming a nonexisting partition should not throw out NullPointerException
 --

 Key: HIVE-9975
 URL: https://issues.apache.org/jira/browse/HIVE-9975
 Project: Hive
  Issue Type: Bug
Reporter: Chaoyu Tang
Assignee: Chaoyu Tang
Priority: Minor
 Attachments: HIVE-9975.patch


 Renaming a nonexisting partition should not throw out NullPointerException. 
 create table testpart (col1 int, col2 string, col3 string) partitioned by 
 (part string);
 alter table testpart partition (part = 'nonexisting') rename to partition 
 (part = 'existing');
 we get NPE like following:
 {code}
 15/03/16 10:16:11 ERROR exec.DDLTask: java.lang.NullPointerException
   at 
 org.apache.hadoop.hive.ql.exec.DDLTask.renamePartition(DDLTask.java:944)
   at org.apache.hadoop.hive.ql.exec.DDLTask.execute(DDLTask.java:350)
   at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:160)
   at 
 org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:88)
   at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:1642)
   at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:1402)
   at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1187)
   at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1053)
   at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1043)
   at 
 org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:207)
   at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:159)
   at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:370)
   at 
 org.apache.hadoop.hive.cli.CliDriver.executeDriver(CliDriver.java:754)
   at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:675)
   at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:615)
 FAILED: Execution Error, return code 1 from 
 org.apache.hadoop.hive.ql.exec.DDLTask. null
 {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9792) Support interval type in expressions/predicates

2015-03-17 Thread Ashutosh Chauhan (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9792?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14365770#comment-14365770
 ] 

Ashutosh Chauhan commented on HIVE-9792:


+1

 Support interval type in expressions/predicates 
 

 Key: HIVE-9792
 URL: https://issues.apache.org/jira/browse/HIVE-9792
 Project: Hive
  Issue Type: Sub-task
  Components: Types
Reporter: Jason Dere
Assignee: Jason Dere
 Attachments: HIVE-9792.1.patch, HIVE-9792.2.patch, HIVE-9792.3.patch, 
 HIVE-9792.4.patch, HIVE-9792.5.patch, HIVE-9792.6.patch, HIVE-9792.7.patch


 Provide partial support for the interval year-month/interval day-time types 
 in Hive. Intervals will be usable in expressions/predicates/joins:
 {noformat}
   select birthdate + interval '30-0' year to month as thirtieth_birthday
   from table
   where (current_timestamp - ts1  interval '3 0:0:0' day to second)
 {noformat}
 This stops short of adding making the interval types usable as a storable 
 column type.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Issue Comment Deleted] (HIVE-9934) Vulnerability in LdapAuthenticationProviderImpl enables HiveServer2 client to degrade the authentication mechanism to none, allowing authentication without

2015-03-17 Thread Xuefu Zhang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9934?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Xuefu Zhang updated HIVE-9934:
--
Comment: was deleted

(was: 

{color:red}Overall{color}: -1 no tests executed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12705130/HIVE-9934.4.patch

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/3059/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/3059/console
Test logs: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-3059/

Messages:
{noformat}
 This message was trimmed, see log for full details 
[INFO] Excluding org.scala-lang:scala-compiler:jar:2.10.0 from the shaded jar.
[INFO] Excluding org.scala-lang:scala-reflect:jar:2.10.0 from the shaded jar.
[INFO] Excluding com.fasterxml.jackson.core:jackson-databind:jar:2.3.1 from the 
shaded jar.
[INFO] Excluding com.fasterxml.jackson.core:jackson-annotations:jar:2.3.0 from 
the shaded jar.
[INFO] Excluding com.fasterxml.jackson.core:jackson-core:jar:2.3.1 from the 
shaded jar.
[INFO] Excluding org.apache.mesos:mesos:jar:shaded-protobuf:0.18.1 from the 
shaded jar.
[INFO] Excluding com.clearspring.analytics:stream:jar:2.7.0 from the shaded jar.
[INFO] Excluding com.codahale.metrics:metrics-core:jar:3.0.0 from the shaded 
jar.
[INFO] Excluding com.codahale.metrics:metrics-jvm:jar:3.0.0 from the shaded jar.
[INFO] Excluding com.codahale.metrics:metrics-json:jar:3.0.0 from the shaded 
jar.
[INFO] Excluding com.codahale.metrics:metrics-graphite:jar:3.0.0 from the 
shaded jar.
[INFO] Excluding org.tachyonproject:tachyon-client:jar:0.5.0 from the shaded 
jar.
[INFO] Excluding org.tachyonproject:tachyon:jar:0.5.0 from the shaded jar.
[INFO] Excluding org.spark-project:pyrolite:jar:2.0.1 from the shaded jar.
[INFO] Excluding net.sf.py4j:py4j:jar:0.8.2.1 from the shaded jar.
[INFO] Excluding org.spark-project.spark:unused:jar:1.0.0 from the shaded jar.
[INFO] Excluding org.apache.hadoop:hadoop-common:jar:2.6.0 from the shaded jar.
[INFO] Excluding xmlenc:xmlenc:jar:0.52 from the shaded jar.
[INFO] Excluding javax.servlet:servlet-api:jar:2.5 from the shaded jar.
[INFO] Excluding org.mortbay.jetty:jetty:jar:6.1.26 from the shaded jar.
[INFO] Excluding org.mortbay.jetty:jetty-util:jar:6.1.26 from the shaded jar.
[INFO] Excluding com.sun.jersey:jersey-core:jar:1.14 from the shaded jar.
[INFO] Excluding com.sun.jersey:jersey-json:jar:1.14 from the shaded jar.
[INFO] Excluding org.codehaus.jettison:jettison:jar:1.1 from the shaded jar.
[INFO] Excluding com.sun.xml.bind:jaxb-impl:jar:2.2.3-1 from the shaded jar.
[INFO] Excluding com.sun.jersey:jersey-server:jar:1.14 from the shaded jar.
[INFO] Excluding asm:asm:jar:3.1 from the shaded jar.
[INFO] Excluding tomcat:jasper-compiler:jar:5.5.23 from the shaded jar.
[INFO] Excluding tomcat:jasper-runtime:jar:5.5.23 from the shaded jar.
[INFO] Excluding javax.servlet.jsp:jsp-api:jar:2.1 from the shaded jar.
[INFO] Excluding commons-el:commons-el:jar:1.0 from the shaded jar.
[INFO] Excluding commons-configuration:commons-configuration:jar:1.6 from the 
shaded jar.
[INFO] Excluding commons-digester:commons-digester:jar:1.8 from the shaded jar.
[INFO] Excluding commons-beanutils:commons-beanutils:jar:1.7.0 from the shaded 
jar.
[INFO] Excluding commons-beanutils:commons-beanutils-core:jar:1.8.0 from the 
shaded jar.
[INFO] Excluding com.google.code.gson:gson:jar:2.2.4 from the shaded jar.
[INFO] Excluding org.apache.hadoop:hadoop-auth:jar:2.6.0 from the shaded jar.
[INFO] Excluding 
org.apache.directory.server:apacheds-kerberos-codec:jar:2.0.0-M15 from the 
shaded jar.
[INFO] Excluding org.apache.directory.server:apacheds-i18n:jar:2.0.0-M15 from 
the shaded jar.
[INFO] Excluding org.apache.directory.api:api-asn1-api:jar:1.0.0-M20 from the 
shaded jar.
[INFO] Excluding org.apache.directory.api:api-util:jar:1.0.0-M20 from the 
shaded jar.
[INFO] Excluding com.jcraft:jsch:jar:0.1.42 from the shaded jar.
[INFO] Excluding org.htrace:htrace-core:jar:3.0.4 from the shaded jar.
[INFO] Excluding org.apache.hadoop:hadoop-archives:jar:2.6.0 from the shaded 
jar.
[INFO] Excluding org.apache.hadoop:hadoop-mapreduce-client-core:jar:2.6.0 from 
the shaded jar.
[INFO] Excluding com.google.inject.extensions:guice-servlet:jar:3.0 from the 
shaded jar.
[INFO] Excluding org.apache.hadoop:hadoop-yarn-server-common:jar:2.6.0 from the 
shaded jar.
[INFO] Excluding org.fusesource.leveldbjni:leveldbjni-all:jar:1.8 from the 
shaded jar.
[INFO] Excluding org.apache.hadoop:hadoop-hdfs:jar:2.6.0 from the shaded jar.
[INFO] Excluding commons-daemon:commons-daemon:jar:1.0.13 from the shaded jar.
[INFO] Excluding xerces:xercesImpl:jar:2.9.1 from the shaded jar.
[INFO] Excluding xml-apis:xml-apis:jar:1.3.04 from the shaded jar.
[INFO] Excluding 

[jira] [Updated] (HIVE-9934) Vulnerability in LdapAuthenticationProviderImpl enables HiveServer2 client to degrade the authentication mechanism to none, allowing authentication without password

2015-03-17 Thread Xuefu Zhang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9934?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Xuefu Zhang updated HIVE-9934:
--
Attachment: (was: HIVE-9934.4.patch)

 Vulnerability in LdapAuthenticationProviderImpl enables HiveServer2 client to 
 degrade the authentication mechanism to none, allowing authentication 
 without password
 --

 Key: HIVE-9934
 URL: https://issues.apache.org/jira/browse/HIVE-9934
 Project: Hive
  Issue Type: Bug
  Components: Security
Affects Versions: 1.1.0
Reporter: Chao
Assignee: Chao
 Attachments: HIVE-9934.1.patch, HIVE-9934.2.patch, HIVE-9934.3.patch, 
 HIVE-9934.3.patch


 Vulnerability in LdapAuthenticationProviderImpl enables HiveServer2 client to 
 degrade the authentication mechanism to none, allowing authentication 
 without password.
 See: http://docs.oracle.com/javase/jndi/tutorial/ldap/security/simple.html
 “If you supply an empty string, an empty byte/char array, or null to the 
 Context.SECURITY_CREDENTIALS environment property, then the authentication 
 mechanism will be none. This is because the LDAP requires the password to 
 be nonempty for simple authentication. The protocol automatically converts 
 the authentication to none if a password is not supplied.”
  
 Since the LdapAuthenticationProviderImpl.Authenticate method is relying on a 
 NamingException being thrown during creation of initial context, it does not 
 fail when the context result is an “unauthenticated” positive response from 
 the LDAP server. The end result is, one can authenticate with HiveServer2 
 using the LdapAuthenticationProviderImpl with only a user name and an empty 
 password.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9859) Create bitwise left/right shift UDFs

2015-03-17 Thread Alexander Pivovarov (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9859?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alexander Pivovarov updated HIVE-9859:
--
Summary: Create bitwise left/right shift UDFs  (was: Create bitwise shift 
left/right UDFs)

 Create bitwise left/right shift UDFs
 

 Key: HIVE-9859
 URL: https://issues.apache.org/jira/browse/HIVE-9859
 Project: Hive
  Issue Type: Improvement
  Components: UDF
Reporter: Alexander Pivovarov
Assignee: Alexander Pivovarov
 Attachments: HIVE-9859.1.patch


 Signature:
 a  b
 a  b
 a  b
 For example:
 {code}
 select 1  4, 8  2, 8  2;
 OK
 16   2   2
 {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9991) Cannot do a SELECT on external tables that are on S3 due to Encryption error

2015-03-17 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9991?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14366556#comment-14366556
 ] 

Hive QA commented on HIVE-9991:
---



{color:green}Overall{color}: +1 all checks pass

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12705207/HIVE-9991.4.patch

{color:green}SUCCESS:{color} +1 7771 tests passed

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/3065/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/3065/console
Test logs: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-3065/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12705207 - PreCommit-HIVE-TRUNK-Build

 Cannot do a SELECT on external tables that are on S3 due to Encryption error
 

 Key: HIVE-9991
 URL: https://issues.apache.org/jira/browse/HIVE-9991
 Project: Hive
  Issue Type: Bug
Affects Versions: 1.0.0
Reporter: Sergio Peña
Assignee: Sergio Peña
 Attachments: HIVE-9991.1.patch, HIVE-9991.2.patch, HIVE-9991.3.patch, 
 HIVE-9991.4.patch


 I cannot do any select query on external tables that are not part of HDFS. 
 For example S3.
 {code}
 Select * from my_table limit 10;
 FAILED: SemanticException Unable to determine if s3n://my-bucket/is 
 encrypted: java.lang.IllegalArgumentException: Wrong FS: s3n://my-bucket/, 
 expected: hdfs://0.0.0.0:8020
 {code}
 This error is due to a internal function that checks if a table is encrypted 
 or not. This is only supported on HDFS files, but the check is happening on 
 any external table as well causing the above error.
 To fix this, we should check for encrypted tables only for HDFS tables. And 
 skip the check for any other file schema.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9999) LLAP: Handle task rejection from daemons in the AM

2015-03-17 Thread Siddharth Seth (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Siddharth Seth updated HIVE-:
-
Attachment: HIVE-.1.patch

 LLAP: Handle task rejection from daemons in the AM
 --

 Key: HIVE-
 URL: https://issues.apache.org/jira/browse/HIVE-
 Project: Hive
  Issue Type: Sub-task
Reporter: Siddharth Seth
Assignee: Siddharth Seth
 Fix For: llap

 Attachments: HIVE-.1.patch






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-10000) 10000 whoooohooo

2015-03-17 Thread Sergey Shelukhin (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-1?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14366335#comment-14366335
 ] 

Sergey Shelukhin commented on HIVE-1:
-

Here I was hoping to file some aspirational LLAP JIRA under than number

 1 whhooo
 

 Key: HIVE-1
 URL: https://issues.apache.org/jira/browse/HIVE-1
 Project: Hive
  Issue Type: Bug
Reporter: Gunther Hagleitner





--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Comment Edited] (HIVE-10000) 10000 whoooohooo

2015-03-17 Thread Sergey Shelukhin (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-1?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14366335#comment-14366335
 ] 

Sergey Shelukhin edited comment on HIVE-1 at 3/17/15 11:56 PM:
---

Here I was hoping to file some aspirational LLAP JIRA under this number


was (Author: sershe):
Here I was hoping to file some aspirational LLAP JIRA under than number

 1 whhooo
 

 Key: HIVE-1
 URL: https://issues.apache.org/jira/browse/HIVE-1
 Project: Hive
  Issue Type: Bug
Reporter: Gunther Hagleitner





--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-5016) Local mode FileNotFoundException: File does not exist

2015-03-17 Thread Jian Fang (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-5016?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14366243#comment-14366243
 ] 

Jian Fang commented on HIVE-5016:
-

Hi Abin, seems this issue still exits in the latest hadoop 2.6.0 code. Could 
you please create a Jira to hadoop? Thanks.


 Local mode FileNotFoundException: File does not exist
 -

 Key: HIVE-5016
 URL: https://issues.apache.org/jira/browse/HIVE-5016
 Project: Hive
  Issue Type: Bug
Affects Versions: 0.10.0
 Environment: Centos 6.3 (final)
 Hadoop 2.0.2-alpha
 Java(TM) SE Runtime Environment (build 1.6.0_31-b04)
 Hive libs:
 ls -1 lib/
 antlr-2.7.7.jar
 antlr-runtime-3.0.1.jar
 avro-1.7.1.jar
 avro-mapred-1.7.1.jar
 commons-cli-1.2.jar
 commons-codec-1.4.jar
 commons-collections-3.2.1.jar
 commons-compress-1.4.1.jar
 commons-configuration-1.6.jar
 commons-dbcp-1.4.jar
 commons-lang-2.4.jar
 commons-logging-1.0.4.jar
 commons-logging-api-1.0.4.jar
 commons-pool-1.5.4.jar
 datanucleus-connectionpool-2.0.3.jar
 datanucleus-core-2.0.3.jar
 datanucleus-enhancer-2.0.3.jar
 datanucleus-rdbms-2.0.3.jar
 derby-10.4.2.0.jar
 guava-r09.jar
 hbase-0.92.0.jar
 hbase-0.92.0-tests.jar
 hive-builtins-0.10.0.jar
 hive-cli-0.10.0.jar
 hive-common-0.10.0.jar
 hive-contrib-0.10.0.jar
 hive-exec-0.10.0.jar
 hive-hbase-handler-0.10.0.jar
 hive-hwi-0.10.0.jar
 hive-hwi-0.10.0.war
 hive-jdbc-0.10.0.jar
 hive-metastore-0.10.0.jar
 hive-pdk-0.10.0.jar
 hive-serde-0.10.0.jar
 hive-service-0.10.0.jar
 hive-shims-0.10.0.jar
 jackson-core-asl-1.8.8.jar
 jackson-jaxrs-1.8.8.jar
 jackson-mapper-asl-1.8.8.jar
 jackson-xc-1.8.8.jar
 JavaEWAH-0.3.2.jar
 javolution-5.5.1.jar
 jdo2-api-2.3-ec.jar
 jetty-6.1.26.jar
 jetty-util-6.1.26.jar
 jline-0.9.94.jar
 json-20090211.jar
 libfb303-0.9.0.jar
 libthrift-0.9.0.jar
 log4j-1.2.16.jar
 php
 py
 servlet-api-2.5-20081211.jar
 slf4j-api-1.6.1.jar
 slf4j-log4j12-1.6.1.jar
 sqlline-1_0_2.jar
 stringtemplate-3.1-b1.jar
 xz-1.0.jar
 zookeeper-3.4.3.jar
Reporter: Abin Shahab
Priority: Critical
 Attachments: HIVE-5016.patch, HIVE-5016.patch


 Hive jobs in local mode fail with the error posted below. The jar file that's 
 not being found exists and has the following access:
  ls -l hive-0.10.0/lib/hive-builtins-0.10.0.jar
 rw-rw-r-- 1 ashahab ashahab 3914 Dec 18  2012 
 hive-0.10.0/lib/hive-builtins-0.10.0.jar
 Steps to reproduce:
 hive set hive.exec.mode.local.auto=true;
 hive set hive.exec.mode.local.auto;
 hive.exec.mode.local.auto=true
 hive select count(*) from abin_test_table;
 Automatically selecting local only mode for query
 Total MapReduce jobs = 1
 Launching Job 1 out of 1
 Number of reduce tasks determined at compile time: 1
 In order to change the average load for a reducer (in bytes):
   set hive.exec.reducers.bytes.per.reducer=number
 In order to limit the maximum number of reducers:
   set hive.exec.reducers.max=number
 In order to set a constant number of reducers:
   set mapred.reduce.tasks=number
 13/08/06 21:37:11 WARN conf.Configuration: 
 file:/tmp/ashahab/hive_2013-08-06_21-37-09_046_3263640403676309186/-local-10002/jobconf.xml:an
  attempt to override final parameter: 
 mapreduce.job.end-notification.max.retry.interval;  Ignoring.
 13/08/06 21:37:11 WARN conf.Configuration: 
 file:/tmp/ashahab/hive_2013-08-06_21-37-09_046_3263640403676309186/-local-10002/jobconf.xml:an
  attempt to override final parameter: 
 mapreduce.job.end-notification.max.attempts;  Ignoring.
 WARNING: org.apache.hadoop.metrics.jvm.EventCounter is deprecated. Please use 
 org.apache.hadoop.log.metrics.EventCounter in all the log4j.properties files.
 Execution log at: 
 /tmp/ashahab/ashahab_20130806213737_7d26b796-5f55-44ca-a755-8898153d963b.log
 java.io.FileNotFoundException: File does not exist: 
 /home/ashahab/dev/hive-0.10.0/lib/hive-builtins-0.10.0.jar
   at 
 org.apache.hadoop.hdfs.DistributedFileSystem.getFileStatus(DistributedFileSystem.java:782)
   at 
 org.apache.hadoop.mapreduce.filecache.ClientDistributedCacheManager.getFileStatus(ClientDistributedCacheManager.java:208)
   at 
 org.apache.hadoop.mapreduce.filecache.ClientDistributedCacheManager.determineTimestamps(ClientDistributedCacheManager.java:71)
   at 
 org.apache.hadoop.mapreduce.JobSubmitter.copyAndConfigureFiles(JobSubmitter.java:252)
   at 
 org.apache.hadoop.mapreduce.JobSubmitter.copyAndConfigureFiles(JobSubmitter.java:290)
   at 
 org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(JobSubmitter.java:361)
   at org.apache.hadoop.mapreduce.Job$11.run(Job.java:1218)
   at org.apache.hadoop.mapreduce.Job$11.run(Job.java:1215)
   at java.security.AccessController.doPrivileged(Native Method)
   at javax.security.auth.Subject.doAs(Subject.java:396)
   at 
 

[jira] [Updated] (HIVE-9919) upgrade scripts don't work on some auto-created DBs due to absence of tables

2015-03-17 Thread Sergey Shelukhin (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9919?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sergey Shelukhin updated HIVE-9919:
---
Attachment: HIVE-9919.01.patch

Updated the patch to add all the indexes/etc. from all the respective schemas, 
in case future scripts decide to drop them (which they do)

 upgrade scripts don't work on some auto-created DBs due to absence of tables
 

 Key: HIVE-9919
 URL: https://issues.apache.org/jira/browse/HIVE-9919
 Project: Hive
  Issue Type: Bug
Reporter: Sergey Shelukhin
Assignee: Sergey Shelukhin
 Attachments: HIVE-9919.01.patch, HIVE-9919.patch


 DataNucleus in its infinite wisdom doesn't create all tables.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9991) Cannot do a SELECT on external tables that are on S3 due to Encryption error

2015-03-17 Thread Xuefu Zhang (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9991?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14366213#comment-14366213
 ] 

Xuefu Zhang commented on HIVE-9991:
---

[~spena], it seems the above failed test has a result diff. You might need to 
regenerate the test output.

 Cannot do a SELECT on external tables that are on S3 due to Encryption error
 

 Key: HIVE-9991
 URL: https://issues.apache.org/jira/browse/HIVE-9991
 Project: Hive
  Issue Type: Bug
Affects Versions: 1.0.0
Reporter: Sergio Peña
Assignee: Sergio Peña
 Attachments: HIVE-9991.1.patch, HIVE-9991.2.patch, HIVE-9991.3.patch


 I cannot do any select query on external tables that are not part of HDFS. 
 For example S3.
 {code}
 Select * from my_table limit 10;
 FAILED: SemanticException Unable to determine if s3n://my-bucket/is 
 encrypted: java.lang.IllegalArgumentException: Wrong FS: s3n://my-bucket/, 
 expected: hdfs://0.0.0.0:8020
 {code}
 This error is due to a internal function that checks if a table is encrypted 
 or not. This is only supported on HDFS files, but the check is happening on 
 any external table as well causing the above error.
 To fix this, we should check for encrypted tables only for HDFS tables. And 
 skip the check for any other file schema.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-10000) 10000 whoooohooo

2015-03-17 Thread Gunther Hagleitner (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-1?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14366337#comment-14366337
 ] 

Gunther Hagleitner commented on HIVE-1:
---

[~sershe] what do you have in mind. I can s till change the title :-P

 1 whhooo
 

 Key: HIVE-1
 URL: https://issues.apache.org/jira/browse/HIVE-1
 Project: Hive
  Issue Type: Bug
Reporter: Gunther Hagleitner





--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9997) minor tweaks for bytes mapjoin hash table

2015-03-17 Thread Sergey Shelukhin (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9997?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sergey Shelukhin updated HIVE-9997:
---
Attachment: HIVE-9997.patch

first attempt; need to see HiveQA

 minor tweaks for bytes mapjoin hash table
 -

 Key: HIVE-9997
 URL: https://issues.apache.org/jira/browse/HIVE-9997
 Project: Hive
  Issue Type: Bug
Reporter: Sergey Shelukhin
Assignee: Sergey Shelukhin
 Attachments: HIVE-9997.patch


 From HIVE-7617



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-7018) Table and Partition tables have column LINK_TARGET_ID in Mysql scripts but not others

2015-03-17 Thread Yongzhi Chen (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-7018?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14366209#comment-14366209
 ] 

Yongzhi Chen commented on HIVE-7018:


Here is my plan: I will find a way to check if the two tables have 
LINK_TARGE_ID column or not, if have it, I will drop the column. Is that a safe 
plan? Thanks

 Table and Partition tables have column LINK_TARGET_ID in Mysql scripts but 
 not others
 -

 Key: HIVE-7018
 URL: https://issues.apache.org/jira/browse/HIVE-7018
 Project: Hive
  Issue Type: Bug
Reporter: Brock Noland
Assignee: Yongzhi Chen
 Attachments: HIVE-7018.1.patch


 It appears that at least postgres and oracle do not have the LINK_TARGET_ID 
 column while mysql does.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9997) minor tweaks for bytes mapjoin hash table

2015-03-17 Thread Sergey Shelukhin (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9997?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14366272#comment-14366272
 ] 

Sergey Shelukhin commented on HIVE-9997:


https://reviews.apache.org/r/32178


 minor tweaks for bytes mapjoin hash table
 -

 Key: HIVE-9997
 URL: https://issues.apache.org/jira/browse/HIVE-9997
 Project: Hive
  Issue Type: Bug
Reporter: Sergey Shelukhin
Assignee: Sergey Shelukhin
 Attachments: HIVE-9997.patch


 From HIVE-7617



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-10000) 10000 whoooohooo

2015-03-17 Thread Sergey Shelukhin (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-1?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14366371#comment-14366371
 ] 

Sergey Shelukhin commented on HIVE-1:
-

Define APIs? :) [~sseth]

 1 whhooo
 

 Key: HIVE-1
 URL: https://issues.apache.org/jira/browse/HIVE-1
 Project: Hive
  Issue Type: Bug
Reporter: Gunther Hagleitner





--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Resolved] (HIVE-9999) LLAP: Handle task rejection from daemons in the AM

2015-03-17 Thread Siddharth Seth (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Siddharth Seth resolved HIVE-.
--
Resolution: Fixed

 LLAP: Handle task rejection from daemons in the AM
 --

 Key: HIVE-
 URL: https://issues.apache.org/jira/browse/HIVE-
 Project: Hive
  Issue Type: Sub-task
Reporter: Siddharth Seth
Assignee: Siddharth Seth
 Fix For: llap

 Attachments: HIVE-.1.patch






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9998) Vectorization support for interval types

2015-03-17 Thread Jason Dere (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9998?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jason Dere updated HIVE-9998:
-
Attachment: HIVE-9998.1.patch

Initial patch, dependent on HIVE-9792.
This also fixes some vectorized comparisons for Date, when one of the left or 
right side is a constant expression.

 Vectorization support for interval types
 

 Key: HIVE-9998
 URL: https://issues.apache.org/jira/browse/HIVE-9998
 Project: Hive
  Issue Type: Sub-task
  Components: Types, Vectorization
Reporter: Jason Dere
Assignee: Jason Dere
 Attachments: HIVE-9998.1.patch


 Allow intervals to work during vectorized query execution



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9997) minor tweaks for bytes mapjoin hash table

2015-03-17 Thread Mostafa Mokhtar (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9997?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14366340#comment-14366340
 ] 

Mostafa Mokhtar commented on HIVE-9997:
---

[~sershe]

Can these be addressed as well?
  private int getOffset(long offset) {
return (int)(offset % wbSize);
  }

  private int getBufferIndex(long offset) {
return (int)(offset / wbSize);
  }

Called from BytesBytesMultiHashMap.isSameKey


 minor tweaks for bytes mapjoin hash table
 -

 Key: HIVE-9997
 URL: https://issues.apache.org/jira/browse/HIVE-9997
 Project: Hive
  Issue Type: Bug
Reporter: Sergey Shelukhin
Assignee: Sergey Shelukhin
 Attachments: HIVE-9997.patch


 From HIVE-7617



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9991) Cannot do a SELECT on external tables that are on S3 due to Encryption error

2015-03-17 Thread JIRA

 [ 
https://issues.apache.org/jira/browse/HIVE-9991?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sergio Peña updated HIVE-9991:
--
Attachment: HIVE-9991.3.patch

Fixed patch to include the test into TestEncryptedHdfsCliDriver tests. 

It also uses pfile:/// as a location instead of file:/// in order to avoid 
change ownership errors.

 Cannot do a SELECT on external tables that are on S3 due to Encryption error
 

 Key: HIVE-9991
 URL: https://issues.apache.org/jira/browse/HIVE-9991
 Project: Hive
  Issue Type: Bug
Affects Versions: 1.0.0
Reporter: Sergio Peña
Assignee: Sergio Peña
 Attachments: HIVE-9991.1.patch, HIVE-9991.2.patch, HIVE-9991.3.patch


 I cannot do any select query on external tables that are not part of HDFS. 
 For example S3.
 {code}
 Select * from my_table limit 10;
 FAILED: SemanticException Unable to determine if s3n://my-bucket/is 
 encrypted: java.lang.IllegalArgumentException: Wrong FS: s3n://my-bucket/, 
 expected: hdfs://0.0.0.0:8020
 {code}
 This error is due to a internal function that checks if a table is encrypted 
 or not. This is only supported on HDFS files, but the check is happening on 
 any external table as well causing the above error.
 To fix this, we should check for encrypted tables only for HDFS tables. And 
 skip the check for any other file schema.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9992) Operator Precedence is incorrect. bitwise XOR should be after minus/plus

2015-03-17 Thread Alexander Pivovarov (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9992?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alexander Pivovarov updated HIVE-9992:
--
Summary: Operator Precedence is incorrect. bitwise XOR should be after 
minus/plus  (was: Operator Precedence is incorrect. XOR should be after 
minus/plus)

 Operator Precedence is incorrect. bitwise XOR should be after minus/plus
 

 Key: HIVE-9992
 URL: https://issues.apache.org/jira/browse/HIVE-9992
 Project: Hive
  Issue Type: Bug
  Components: Query Processor
Affects Versions: 1.2.0
Reporter: Alexander Pivovarov
Assignee: Alexander Pivovarov

 XOR ^ has higher precedence than -/+ in hive
 But in java and in MySQL XOR has lower precedence
 http://dev.mysql.com/doc/refman/5.0/en/operator-precedence.html
 http://docs.oracle.com/javase/tutorial/java/nutsandbolts/operators.html
 {code}
 // now
 select 4 - 2 ^ 8 - 7;
 -13
 // should be
 select 4 - 2 ^ 8 - 7;
 3
 {code}
 need to fix
 IdentifiersParser.g  BITWISEXOR



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9992) Operator Precedence is incorrect. bitwise XOR should be after minus/plus

2015-03-17 Thread Alexander Pivovarov (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9992?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alexander Pivovarov updated HIVE-9992:
--
Description: 
bitwise XOR ^ has higher precedence than -/+ in hive
But in java bitwise XOR has lower precedence
http://docs.oracle.com/javase/tutorial/java/nutsandbolts/operators.html

{code}
// now
select 4 - 2 ^ 8 - 7;
-13

// should be
select 4 - 2 ^ 8 - 7;
3
{code}

need to fix
IdentifiersParser.g  BITWISEXOR

  was:
XOR ^ has higher precedence than -/+ in hive
But in java and in MySQL XOR has lower precedence
http://dev.mysql.com/doc/refman/5.0/en/operator-precedence.html
http://docs.oracle.com/javase/tutorial/java/nutsandbolts/operators.html

{code}
// now
select 4 - 2 ^ 8 - 7;
-13

// should be
select 4 - 2 ^ 8 - 7;
3
{code}

need to fix
IdentifiersParser.g  BITWISEXOR


 Operator Precedence is incorrect. bitwise XOR should be after minus/plus
 

 Key: HIVE-9992
 URL: https://issues.apache.org/jira/browse/HIVE-9992
 Project: Hive
  Issue Type: Bug
  Components: Query Processor
Affects Versions: 1.2.0
Reporter: Alexander Pivovarov
Assignee: Alexander Pivovarov

 bitwise XOR ^ has higher precedence than -/+ in hive
 But in java bitwise XOR has lower precedence
 http://docs.oracle.com/javase/tutorial/java/nutsandbolts/operators.html
 {code}
 // now
 select 4 - 2 ^ 8 - 7;
 -13
 // should be
 select 4 - 2 ^ 8 - 7;
 3
 {code}
 need to fix
 IdentifiersParser.g  BITWISEXOR



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9792) Support interval type in expressions/predicates

2015-03-17 Thread Jason Dere (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9792?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jason Dere updated HIVE-9792:
-
Attachment: HIVE-9792.7.patch

Patch v7 - RB feedback from Ashutosh

 Support interval type in expressions/predicates 
 

 Key: HIVE-9792
 URL: https://issues.apache.org/jira/browse/HIVE-9792
 Project: Hive
  Issue Type: Sub-task
  Components: Types
Reporter: Jason Dere
Assignee: Jason Dere
 Attachments: HIVE-9792.1.patch, HIVE-9792.2.patch, HIVE-9792.3.patch, 
 HIVE-9792.4.patch, HIVE-9792.5.patch, HIVE-9792.6.patch, HIVE-9792.7.patch


 Provide partial support for the interval year-month/interval day-time types 
 in Hive. Intervals will be usable in expressions/predicates/joins:
 {noformat}
   select birthdate + interval '30-0' year to month as thirtieth_birthday
   from table
   where (current_timestamp - ts1  interval '3 0:0:0' day to second)
 {noformat}
 This stops short of adding making the interval types usable as a storable 
 column type.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9934) Vulnerability in LdapAuthenticationProviderImpl enables HiveServer2 client to degrade the authentication mechanism to none, allowing authentication without password

2015-03-17 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9934?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14365794#comment-14365794
 ] 

Hive QA commented on HIVE-9934:
---



{color:red}Overall{color}: -1 no tests executed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12705130/HIVE-9934.4.patch

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/3059/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/3059/console
Test logs: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-3059/

Messages:
{noformat}
 This message was trimmed, see log for full details 
[INFO] Excluding org.scala-lang:scala-compiler:jar:2.10.0 from the shaded jar.
[INFO] Excluding org.scala-lang:scala-reflect:jar:2.10.0 from the shaded jar.
[INFO] Excluding com.fasterxml.jackson.core:jackson-databind:jar:2.3.1 from the 
shaded jar.
[INFO] Excluding com.fasterxml.jackson.core:jackson-annotations:jar:2.3.0 from 
the shaded jar.
[INFO] Excluding com.fasterxml.jackson.core:jackson-core:jar:2.3.1 from the 
shaded jar.
[INFO] Excluding org.apache.mesos:mesos:jar:shaded-protobuf:0.18.1 from the 
shaded jar.
[INFO] Excluding com.clearspring.analytics:stream:jar:2.7.0 from the shaded jar.
[INFO] Excluding com.codahale.metrics:metrics-core:jar:3.0.0 from the shaded 
jar.
[INFO] Excluding com.codahale.metrics:metrics-jvm:jar:3.0.0 from the shaded jar.
[INFO] Excluding com.codahale.metrics:metrics-json:jar:3.0.0 from the shaded 
jar.
[INFO] Excluding com.codahale.metrics:metrics-graphite:jar:3.0.0 from the 
shaded jar.
[INFO] Excluding org.tachyonproject:tachyon-client:jar:0.5.0 from the shaded 
jar.
[INFO] Excluding org.tachyonproject:tachyon:jar:0.5.0 from the shaded jar.
[INFO] Excluding org.spark-project:pyrolite:jar:2.0.1 from the shaded jar.
[INFO] Excluding net.sf.py4j:py4j:jar:0.8.2.1 from the shaded jar.
[INFO] Excluding org.spark-project.spark:unused:jar:1.0.0 from the shaded jar.
[INFO] Excluding org.apache.hadoop:hadoop-common:jar:2.6.0 from the shaded jar.
[INFO] Excluding xmlenc:xmlenc:jar:0.52 from the shaded jar.
[INFO] Excluding javax.servlet:servlet-api:jar:2.5 from the shaded jar.
[INFO] Excluding org.mortbay.jetty:jetty:jar:6.1.26 from the shaded jar.
[INFO] Excluding org.mortbay.jetty:jetty-util:jar:6.1.26 from the shaded jar.
[INFO] Excluding com.sun.jersey:jersey-core:jar:1.14 from the shaded jar.
[INFO] Excluding com.sun.jersey:jersey-json:jar:1.14 from the shaded jar.
[INFO] Excluding org.codehaus.jettison:jettison:jar:1.1 from the shaded jar.
[INFO] Excluding com.sun.xml.bind:jaxb-impl:jar:2.2.3-1 from the shaded jar.
[INFO] Excluding com.sun.jersey:jersey-server:jar:1.14 from the shaded jar.
[INFO] Excluding asm:asm:jar:3.1 from the shaded jar.
[INFO] Excluding tomcat:jasper-compiler:jar:5.5.23 from the shaded jar.
[INFO] Excluding tomcat:jasper-runtime:jar:5.5.23 from the shaded jar.
[INFO] Excluding javax.servlet.jsp:jsp-api:jar:2.1 from the shaded jar.
[INFO] Excluding commons-el:commons-el:jar:1.0 from the shaded jar.
[INFO] Excluding commons-configuration:commons-configuration:jar:1.6 from the 
shaded jar.
[INFO] Excluding commons-digester:commons-digester:jar:1.8 from the shaded jar.
[INFO] Excluding commons-beanutils:commons-beanutils:jar:1.7.0 from the shaded 
jar.
[INFO] Excluding commons-beanutils:commons-beanutils-core:jar:1.8.0 from the 
shaded jar.
[INFO] Excluding com.google.code.gson:gson:jar:2.2.4 from the shaded jar.
[INFO] Excluding org.apache.hadoop:hadoop-auth:jar:2.6.0 from the shaded jar.
[INFO] Excluding 
org.apache.directory.server:apacheds-kerberos-codec:jar:2.0.0-M15 from the 
shaded jar.
[INFO] Excluding org.apache.directory.server:apacheds-i18n:jar:2.0.0-M15 from 
the shaded jar.
[INFO] Excluding org.apache.directory.api:api-asn1-api:jar:1.0.0-M20 from the 
shaded jar.
[INFO] Excluding org.apache.directory.api:api-util:jar:1.0.0-M20 from the 
shaded jar.
[INFO] Excluding com.jcraft:jsch:jar:0.1.42 from the shaded jar.
[INFO] Excluding org.htrace:htrace-core:jar:3.0.4 from the shaded jar.
[INFO] Excluding org.apache.hadoop:hadoop-archives:jar:2.6.0 from the shaded 
jar.
[INFO] Excluding org.apache.hadoop:hadoop-mapreduce-client-core:jar:2.6.0 from 
the shaded jar.
[INFO] Excluding com.google.inject.extensions:guice-servlet:jar:3.0 from the 
shaded jar.
[INFO] Excluding org.apache.hadoop:hadoop-yarn-server-common:jar:2.6.0 from the 
shaded jar.
[INFO] Excluding org.fusesource.leveldbjni:leveldbjni-all:jar:1.8 from the 
shaded jar.
[INFO] Excluding org.apache.hadoop:hadoop-hdfs:jar:2.6.0 from the shaded jar.
[INFO] Excluding commons-daemon:commons-daemon:jar:1.0.13 from the shaded jar.
[INFO] Excluding xerces:xercesImpl:jar:2.9.1 from the shaded jar.
[INFO] Excluding xml-apis:xml-apis:jar:1.3.04 from the shaded jar.

[jira] [Commented] (HIVE-9934) Vulnerability in LdapAuthenticationProviderImpl enables HiveServer2 client to degrade the authentication mechanism to none, allowing authentication without password

2015-03-17 Thread Prasad Mujumdar (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9934?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14365839#comment-14365839
 ] 

Prasad Mujumdar commented on HIVE-9934:
---

That's fine. The test did get run in the pre-commit run for patch #3. sorry 
about the noise.

+1


 Vulnerability in LdapAuthenticationProviderImpl enables HiveServer2 client to 
 degrade the authentication mechanism to none, allowing authentication 
 without password
 --

 Key: HIVE-9934
 URL: https://issues.apache.org/jira/browse/HIVE-9934
 Project: Hive
  Issue Type: Bug
  Components: Security
Affects Versions: 1.1.0
Reporter: Chao
Assignee: Chao
 Attachments: HIVE-9934.1.patch, HIVE-9934.2.patch, HIVE-9934.3.patch, 
 HIVE-9934.3.patch


 Vulnerability in LdapAuthenticationProviderImpl enables HiveServer2 client to 
 degrade the authentication mechanism to none, allowing authentication 
 without password.
 See: http://docs.oracle.com/javase/jndi/tutorial/ldap/security/simple.html
 “If you supply an empty string, an empty byte/char array, or null to the 
 Context.SECURITY_CREDENTIALS environment property, then the authentication 
 mechanism will be none. This is because the LDAP requires the password to 
 be nonempty for simple authentication. The protocol automatically converts 
 the authentication to none if a password is not supplied.”
  
 Since the LdapAuthenticationProviderImpl.Authenticate method is relying on a 
 NamingException being thrown during creation of initial context, it does not 
 fail when the context result is an “unauthenticated” positive response from 
 the LDAP server. The end result is, one can authenticate with HiveServer2 
 using the LdapAuthenticationProviderImpl with only a user name and an empty 
 password.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9981) Avoid throwing many exceptions when attempting to create new hdfs encryption shim

2015-03-17 Thread JIRA

[ 
https://issues.apache.org/jira/browse/HIVE-9981?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14365920#comment-14365920
 ] 

Sergio Peña commented on HIVE-9981:
---

This test is not related with this patch.

 Avoid throwing many exceptions when attempting to create new hdfs encryption 
 shim
 -

 Key: HIVE-9981
 URL: https://issues.apache.org/jira/browse/HIVE-9981
 Project: Hive
  Issue Type: Improvement
  Components: Encryption
Reporter: Sergio Peña
Assignee: Sergio Peña
 Attachments: HIVE-9981.1.patch, HIVE-9981.2.patch


 Hadoop23Shims.createHdfsEncryptionShim() is throwing an exception for verions 
 lower than 2.6.0 everytime a query is executed.
 Exceptions are expensive, so rather than throwing them every time, we can use 
 this design pattern followed for some other functions in Hadoop23Shims -
 {code}
   protected static final Method accessMethod;
   protected static final Method getPasswordMethod;
   static {
 Method m = null;
 try {
   m = FileSystem.class.getMethod(access, Path.class, FsAction.class);
 } catch (NoSuchMethodException err) {
   // This version of Hadoop does not support FileSystem.access().
 }
 accessMethod = m;
 try {
   m = Configuration.class.getMethod(getPassword, String.class);
 } catch (NoSuchMethodException err) {
   // This version of Hadoop does not support getPassword(), just retrieve 
 password from conf.
   m = null;
 }
 getPasswordMethod = m;
   }
 {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9971) Clean up operator class

2015-03-17 Thread Gunther Hagleitner (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9971?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gunther Hagleitner updated HIVE-9971:
-
Attachment: HIVE-9971.5.patch

 Clean up operator class
 ---

 Key: HIVE-9971
 URL: https://issues.apache.org/jira/browse/HIVE-9971
 Project: Hive
  Issue Type: Bug
Reporter: Gunther Hagleitner
Assignee: Gunther Hagleitner
 Attachments: HIVE-9971.1.patch, HIVE-9971.2.patch, HIVE-9971.3.patch, 
 HIVE-9971.4.patch, HIVE-9971.5.patch


 This is mostly cleanup although it does enhance the pipeline in one respect. 
 It introduces asyn init for operators and uses it for hash table loading 
 where desired.
 There's a bunch of weird code associated with the operator class:
 - initialize isn't recursive, rather initializeOp is supposed to call 
 initializeChildren. That has led to bugs in the past.
 - setExecContext and passExecContext. Both are recursive, but passExecContext 
 calls setExecContext and then recurses again. Boo.
 - lots of (getChildren() != null) although that can't happen anymore
 - TezCacheAccess is a hack. We should just leave init of inputs up to the 
 operator that needs it.
 - Need some sanity checks that make sure that operators were all initialized.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Resolved] (HIVE-9978) LLAP: OrcColumnVectorProducer should handle reading isPresent columns only

2015-03-17 Thread Sergey Shelukhin (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9978?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sergey Shelukhin resolved HIVE-9978.

   Resolution: Fixed
Fix Version/s: llap

committed to branch. Should work now

 LLAP: OrcColumnVectorProducer should handle reading isPresent columns only
 --

 Key: HIVE-9978
 URL: https://issues.apache.org/jira/browse/HIVE-9978
 Project: Hive
  Issue Type: Sub-task
  Components: File Formats
Affects Versions: llap
Reporter: Gopal V
Assignee: Sergey Shelukhin
 Fix For: llap

 Attachments: HIVE-9978.patch


 LlapInputFormat does not understand the difference between empty columns list 
 and null columns list.
 The empty columns list indicates no columns read except the root struct 
 isPresent column, while the null columns list indicates that all columns are 
 being read.
 {code}
 select count(1) from store_sales join date_dim on ss_sold_date_sk = d_date_sk 
 where d_date = '1998-01-01';
 ...
 Caused by: java.lang.NullPointerException
 at 
 org.apache.hadoop.hive.llap.io.decode.OrcColumnVectorProducer.createReadPipeline(OrcColumnVectorProducer.java:72)
 at 
 org.apache.hadoop.hive.llap.io.api.impl.LlapInputFormat$LlapRecordReader.startRead(LlapInputFormat.java:181)
 at 
 org.apache.hadoop.hive.llap.io.api.impl.LlapInputFormat$LlapRecordReader.next(LlapInputFormat.java:140)
 at 
 org.apache.hadoop.hive.llap.io.api.impl.LlapInputFormat$LlapRecordReader.next(LlapInputFormat.java:99)
 at 
 org.apache.hadoop.hive.ql.io.HiveContextAwareRecordReader.doNext(HiveContextAwareRecordReader.java:350)
 ... 22 more
 {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9934) Vulnerability in LdapAuthenticationProviderImpl enables HiveServer2 client to degrade the authentication mechanism to none, allowing authentication without password

2015-03-17 Thread Chao (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9934?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14365037#comment-14365037
 ] 

Chao commented on HIVE-9934:


Found this in log: 

{noformat}
2015-03-17 04:33:32,728 INFO  [stdout-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(537)) - 2015-03-17 04:33:32,725 INFO  
[pool-1-thread-1] client.RemoteDriver (RemoteDriver.java:call(371)) - Failed to 
run job 681ccfbe-bf9f-491c-a2e7-ad513f62d1dc
2015-03-17 04:33:32,728 INFO  [stdout-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(537)) - java.util.concurrent.ExecutionException: 
Exception thrown by job
2015-03-17 04:33:32,728 INFO  [stdout-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(537)) -at 
org.apache.spark.JavaFutureActionWrapper.getImpl(FutureAction.scala:311)
2015-03-17 04:33:32,728 INFO  [stdout-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(537)) -at 
org.apache.spark.JavaFutureActionWrapper.get(FutureAction.scala:316)
2015-03-17 04:33:32,728 INFO  [stdout-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(537)) -at 
org.apache.hive.spark.client.RemoteDriver$JobWrapper.call(RemoteDriver.java:364)
2015-03-17 04:33:32,728 INFO  [stdout-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(537)) -at 
org.apache.hive.spark.client.RemoteDriver$JobWrapper.call(RemoteDriver.java:317)
2015-03-17 04:33:32,729 INFO  [stdout-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(537)) -at 
java.util.concurrent.FutureTask.run(FutureTask.java:262)
2015-03-17 04:33:32,729 INFO  [stdout-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(537)) -at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
2015-03-17 04:33:32,729 INFO  [stdout-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(537)) -at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
2015-03-17 04:33:32,729 INFO  [stdout-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(537)) -at 
java.lang.Thread.run(Thread.java:744)
2015-03-17 04:33:32,729 INFO  [stdout-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(537)) - Caused by: org.apache.spark.SparkException: 
Job aborted due to stage failure: Task 0 in stage 0.0 failed 4 times, most 
recent failure: Lost task 0.3 in stage 0.0 (TID 3, 
ip-10-182-56-7.ec2.internal): java.io.FileNotFoundException: 
http://10.182.56.7:34690/jars/hive-exec-1.2.0-SNAPSHOT.jar
2015-03-17 04:33:32,729 INFO  [stdout-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(537)) -at 
sun.net.www.protocol.http.HttpURLConnection.getInputStream(HttpURLConnection.java:1624)
2015-03-17 04:33:32,729 INFO  [stdout-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(537)) -at 
org.apache.spark.util.Utils$.doFetchFile(Utils.scala:452)
2015-03-17 04:33:32,729 INFO  [stdout-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(537)) -at 
org.apache.spark.util.Utils$.fetchFile(Utils.scala:383)
2015-03-17 04:33:32,729 INFO  [stdout-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(537)) -at 
org.apache.spark.executor.Executor$$anonfun$org$apache$spark$executor$Executor$$updateDependencies$6.apply(Executor.scala:350)
2015-03-17 04:33:32,729 INFO  [stdout-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(537)) -at 
org.apache.spark.executor.Executor$$anonfun$org$apache$spark$executor$Executor$$updateDependencies$6.apply(Executor.scala:347)
2015-03-17 04:33:32,729 INFO  [stdout-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(537)) -at 
scala.collection.TraversableLike$WithFilter$$anonfun$foreach$1.apply(TraversableLike.scala:772)
2015-03-17 04:33:32,729 INFO  [stdout-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(537)) -at 
scala.collection.mutable.HashMap$$anonfun$foreach$1.apply(HashMap.scala:98)
2015-03-17 04:33:32,729 INFO  [stdout-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(537)) -at 
scala.collection.mutable.HashMap$$anonfun$foreach$1.apply(HashMap.scala:98)
2015-03-17 04:33:32,729 INFO  [stdout-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(537)) -at 
scala.collection.mutable.HashTable$class.foreachEntry(HashTable.scala:226)
2015-03-17 04:33:32,729 INFO  [stdout-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(537)) -at 
scala.collection.mutable.HashMap.foreachEntry(HashMap.scala:39)
2015-03-17 04:33:32,730 INFO  [stdout-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(537)) -at 
scala.collection.mutable.HashMap.foreach(HashMap.scala:98)
2015-03-17 04:33:32,730 INFO  [stdout-redir-1]: client.SparkClientImpl 
(SparkClientImpl.java:run(537)) -at 
scala.collection.TraversableLike$WithFilter.foreach(TraversableLike.scala:771)
2015-03-17 

[jira] [Commented] (HIVE-9697) Hive on Spark is not as aggressive as MR on map join [Spark Branch]

2015-03-17 Thread Chao (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9697?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14365072#comment-14365072
 ] 

Chao commented on HIVE-9697:


Can we use hive.stats.collect.rawdatasize to control this? I think in default 
Hive uses rawDataSize, right? And if that is not available (like the issue in 
HIVE-9560), it will switch to totalSize. Is that right?

 Hive on Spark is not as aggressive as MR on map join [Spark Branch]
 ---

 Key: HIVE-9697
 URL: https://issues.apache.org/jira/browse/HIVE-9697
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Reporter: Xin Hao

 We have a finding during running some Big-Bench cases:
 when the same small table size threshold is used, Map Join operator will not 
 be generated in Stage Plans for Hive on Spark, while will be generated for 
 Hive on MR.
 For example, When we run BigBench Q25, the meta info of one input ORC table 
 is as below:
 totalSize=1748955 (about 1.5M)
 rawDataSize=123050375 (about 120M)
 If we use the following parameter settings,
 set hive.auto.convert.join=true;
 set hive.mapjoin.smalltable.filesize=2500;
 set hive.auto.convert.join.noconditionaltask=true;
 set hive.auto.convert.join.noconditionaltask.size=1; (100M)
 Map Join will be enabled for Hive on MR mode, while will not be enabled for 
 Hive on Spark.
 We found that for Hive on MR, the HDFS file size for the table 
 (ContentSummary.getLength(), should approximate the value of ‘totalSize’) 
 will be used to compare with the threshold 100M (smaller than 100M), while 
 for Hive on Spark 'rawDataSize' will be used to compare with the threshold 
 100M (larger than 100M). That's why MapJoin is not enabled for Hive on Spark 
 for this case. And as a result Hive on Spark will get much lower performance 
 data than Hive on MR for this case.
 When we set  hive.auto.convert.join.noconditionaltask.size=15000; (150M), 
 MapJoin will be enabled for Hive on Spark mode, and Hive on Spark will have 
 similar performance data with Hive on MR by then.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9970) Hive on spark

2015-03-17 Thread Amithsha (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9970?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14364951#comment-14364951
 ] 

Amithsha commented on HIVE-9970:



hive insert into table test values(6,8797);
Query ID = hadoop2_20150317163636_4692aa68-56b6-4ea9-ad21-e0f46efe4bfc
Total jobs = 1
Launching Job 1 out of 1
In order to change the average load for a reducer (in bytes):
  set hive.exec.reducers.bytes.per.reducer=number
In order to limit the maximum number of reducers:
  set hive.exec.reducers.max=number
In order to set a constant number of reducers:
  set mapreduce.job.reduces=number
Failed to execute spark task, with exception 
'org.apache.hadoop.hive.ql.metadata.HiveException(Failed to create spark 
client.)'
FAILED: Execution Error, return code 1 from 
org.apache.hadoop.hive.ql.exec.spark.SparkTask


 Hive on spark
 -

 Key: HIVE-9970
 URL: https://issues.apache.org/jira/browse/HIVE-9970
 Project: Hive
  Issue Type: Bug
Reporter: Amithsha

 Hi all,
 Recently i have configured Spark 1.2.0 and my environment is hadoop
 2.6.0 hive 1.1.0 Here i have tried hive on Spark while executing
 insert into i am getting the following g error.
 Query ID = hadoop2_20150313162828_8764adad-a8e4-49da-9ef5-35e4ebd6bc63
 Total jobs = 1
 Launching Job 1 out of 1
 In order to change the average load for a reducer (in bytes):
 set hive.exec.reducers.bytes.per.reducer=number
 In order to limit the maximum number of reducers:
 set hive.exec.reducers.max=number
 In order to set a constant number of reducers:
 set mapreduce.job.reduces=number
 Failed to execute spark task, with exception
 'org.apache.hadoop.hive.ql.metadata.HiveException(Failed to create
 spark client.)'
 FAILED: Execution Error, return code 1 from
 org.apache.hadoop.hive.ql.exec.spark.SparkTask
 Have added the spark-assembly jar in hive lib
 And also in hive console using the command add jar followed by the steps
 set spark.home=/opt/spark-1.2.1/;
 add jar 
 /opt/spark-1.2.1/assembly/target/scala-2.10/spark-assembly-1.2.1-hadoop2.4.0.jar;
 set hive.execution.engine=spark;
 set spark.master=spark://xxx:7077;
 set spark.eventLog.enabled=true;
 set spark.executor.memory=512m;
 set spark.serializer=org.apache.spark.serializer.KryoSerializer;
 Can anyone suggest
 Thanks  Regards
 Amithsha



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-5341) Link doesn't work. Needs to be updated as mentioned in the Description

2015-03-17 Thread Fahd Khan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-5341?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Fahd Khan updated HIVE-5341:


The links isn't working. please provide an alternate link to download the files.

 Link doesn't work. Needs to be updated as mentioned in the Description
 --

 Key: HIVE-5341
 URL: https://issues.apache.org/jira/browse/HIVE-5341
 Project: Hive
  Issue Type: Bug
  Components: Documentation
Reporter: Rakesh Chouhan
Assignee: Lefty Leverenz
  Labels: documentation

 Go to.. Apache HIVE Getting Started Documentation
 https://cwiki.apache.org/confluence/display/Hive/GettingStarted
 Under Section ...
 Simple Example Use Cases
 MovieLens User Ratings
 wget http://www.grouplens.org/system/files/ml-data.tar+0.gz
 The link mentioned as per the document does not work. It needs to be updated 
 to the below URL.
 http://www.grouplens.org/sites/www.grouplens.org/external_files/data/ml-data.tar.gz
 I am setting this defect's priority as a Blocker because, user's will not be 
 able to continue their hands on exercises, unless they find the correct URL 
 to download the mentioned file.
 Referenced from:
 http://mail-archives.apache.org/mod_mbox/hive-user/201302.mbox/%3c8a0c145b-4db9-4d26-8613-8ca1bd741...@daum.net%3E.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9934) Vulnerability in LdapAuthenticationProviderImpl enables HiveServer2 client to degrade the authentication mechanism to none, allowing authentication without password

2015-03-17 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9934?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14365024#comment-14365024
 ] 

Hive QA commented on HIVE-9934:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12704981/HIVE-9934.3.patch

{color:red}ERROR:{color} -1 due to 1 failed/errored test(s), 7770 tests executed
*Failed tests:*
{noformat}
org.apache.hive.jdbc.TestMultiSessionsHS2WithLocalClusterSpark.testSparkQuery
{noformat}

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/3053/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/3053/console
Test logs: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-3053/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 1 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12704981 - PreCommit-HIVE-TRUNK-Build

 Vulnerability in LdapAuthenticationProviderImpl enables HiveServer2 client to 
 degrade the authentication mechanism to none, allowing authentication 
 without password
 --

 Key: HIVE-9934
 URL: https://issues.apache.org/jira/browse/HIVE-9934
 Project: Hive
  Issue Type: Bug
  Components: Security
Affects Versions: 1.1.0
Reporter: Chao
Assignee: Chao
 Attachments: HIVE-9934.1.patch, HIVE-9934.2.patch, HIVE-9934.3.patch


 Vulnerability in LdapAuthenticationProviderImpl enables HiveServer2 client to 
 degrade the authentication mechanism to none, allowing authentication 
 without password.
 See: http://docs.oracle.com/javase/jndi/tutorial/ldap/security/simple.html
 “If you supply an empty string, an empty byte/char array, or null to the 
 Context.SECURITY_CREDENTIALS environment property, then the authentication 
 mechanism will be none. This is because the LDAP requires the password to 
 be nonempty for simple authentication. The protocol automatically converts 
 the authentication to none if a password is not supplied.”
  
 Since the LdapAuthenticationProviderImpl.Authenticate method is relying on a 
 NamingException being thrown during creation of initial context, it does not 
 fail when the context result is an “unauthenticated” positive response from 
 the LDAP server. The end result is, one can authenticate with HiveServer2 
 using the LdapAuthenticationProviderImpl with only a user name and an empty 
 password.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9977) Compactor not running on partitions after dynamic partitioned insert

2015-03-17 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9977?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14365090#comment-14365090
 ] 

Hive QA commented on HIVE-9977:
---



{color:green}Overall{color}: +1 all checks pass

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12704984/HIVE-9977.patch

{color:green}SUCCESS:{color} +1 7774 tests passed

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/3054/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/3054/console
Test logs: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-3054/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12704984 - PreCommit-HIVE-TRUNK-Build

 Compactor not running on partitions after dynamic partitioned insert
 

 Key: HIVE-9977
 URL: https://issues.apache.org/jira/browse/HIVE-9977
 Project: Hive
  Issue Type: Bug
  Components: Transactions
Affects Versions: 0.14.0, 1.0.0, 1.1.0
Reporter: Alan Gates
Assignee: Alan Gates
 Attachments: HIVE-9977.patch


 When an insert, update, or delete is done using dynamic partitioning the lock 
 is obtained on the table instead of on the individual partitions, since the 
 partitions are not known at lock acquisition time.  The compactor is using 
 the locks to determine which partitions to check to see if they need 
 compacted.  Since the individual partitions aren't locked they aren't checked.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-7018) Table and Partition tables have column LINK_TARGET_ID in Mysql scripts but not others

2015-03-17 Thread Yongzhi Chen (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-7018?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14365099#comment-14365099
 ] 

Yongzhi Chen commented on HIVE-7018:


[~xuefuz] , [~csun] could you review the change and commit it if looks fine? 
Thanks

 Table and Partition tables have column LINK_TARGET_ID in Mysql scripts but 
 not others
 -

 Key: HIVE-7018
 URL: https://issues.apache.org/jira/browse/HIVE-7018
 Project: Hive
  Issue Type: Bug
Reporter: Brock Noland
Assignee: Yongzhi Chen
 Attachments: HIVE-7018.1.patch


 It appears that at least postgres and oracle do not have the LINK_TARGET_ID 
 column while mysql does.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9697) Hive on Spark is not as aggressive as MR on map join [Spark Branch]

2015-03-17 Thread Rui Li (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9697?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14365102#comment-14365102
 ] 

Rui Li commented on HIVE-9697:
--

[~csun] - I think MR doesn't use rawDataSize even when it's available. Seems it 
just uses ContentSummary.

 Hive on Spark is not as aggressive as MR on map join [Spark Branch]
 ---

 Key: HIVE-9697
 URL: https://issues.apache.org/jira/browse/HIVE-9697
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Reporter: Xin Hao

 We have a finding during running some Big-Bench cases:
 when the same small table size threshold is used, Map Join operator will not 
 be generated in Stage Plans for Hive on Spark, while will be generated for 
 Hive on MR.
 For example, When we run BigBench Q25, the meta info of one input ORC table 
 is as below:
 totalSize=1748955 (about 1.5M)
 rawDataSize=123050375 (about 120M)
 If we use the following parameter settings,
 set hive.auto.convert.join=true;
 set hive.mapjoin.smalltable.filesize=2500;
 set hive.auto.convert.join.noconditionaltask=true;
 set hive.auto.convert.join.noconditionaltask.size=1; (100M)
 Map Join will be enabled for Hive on MR mode, while will not be enabled for 
 Hive on Spark.
 We found that for Hive on MR, the HDFS file size for the table 
 (ContentSummary.getLength(), should approximate the value of ‘totalSize’) 
 will be used to compare with the threshold 100M (smaller than 100M), while 
 for Hive on Spark 'rawDataSize' will be used to compare with the threshold 
 100M (larger than 100M). That's why MapJoin is not enabled for Hive on Spark 
 for this case. And as a result Hive on Spark will get much lower performance 
 data than Hive on MR for this case.
 When we set  hive.auto.convert.join.noconditionaltask.size=15000; (150M), 
 MapJoin will be enabled for Hive on Spark mode, and Hive on Spark will have 
 similar performance data with Hive on MR by then.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-7018) Table and Partition tables have column LINK_TARGET_ID in Mysql scripts but not others

2015-03-17 Thread Xuefu Zhang (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-7018?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14365134#comment-14365134
 ] 

Xuefu Zhang commented on HIVE-7018:
---

Patch looks fine. However, I don't quite understand why we are also removing 
the following:
{code}
-  CONSTRAINT `PARTITIONS_FK2` FOREIGN KEY (`SD_ID`) REFERENCES `SDS` (`SD_ID`),
...
-  CONSTRAINT `TBLS_FK2` FOREIGN KEY (`DB_ID`) REFERENCES `DBS` (`DB_ID`),
{code}

This doesn't seem related to LINK_TARGET_ID.

 Table and Partition tables have column LINK_TARGET_ID in Mysql scripts but 
 not others
 -

 Key: HIVE-7018
 URL: https://issues.apache.org/jira/browse/HIVE-7018
 Project: Hive
  Issue Type: Bug
Reporter: Brock Noland
Assignee: Yongzhi Chen
 Attachments: HIVE-7018.1.patch


 It appears that at least postgres and oracle do not have the LINK_TARGET_ID 
 column while mysql does.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9664) Hive add jar command should be able to download and add jars from a repository

2015-03-17 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9664?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14365100#comment-14365100
 ] 

Hive QA commented on HIVE-9664:
---



{color:red}Overall{color}: -1 no tests executed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12705028/HIVE-9664.patch

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/3055/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/3055/console
Test logs: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-3055/

Messages:
{noformat}
 This message was trimmed, see log for full details 
[INFO] Installing 
/data/hive-ptest/working/apache-svn-trunk-source/itests/custom-serde/target/hive-it-custom-serde-1.2.0-SNAPSHOT.jar
 to 
/data/hive-ptest/working/maven/org/apache/hive/hive-it-custom-serde/1.2.0-SNAPSHOT/hive-it-custom-serde-1.2.0-SNAPSHOT.jar
[INFO] Installing 
/data/hive-ptest/working/apache-svn-trunk-source/itests/custom-serde/pom.xml to 
/data/hive-ptest/working/maven/org/apache/hive/hive-it-custom-serde/1.2.0-SNAPSHOT/hive-it-custom-serde-1.2.0-SNAPSHOT.pom
[INFO] 
[INFO] 
[INFO] Building Hive Integration - HCatalog Unit Tests 1.2.0-SNAPSHOT
[INFO] 
[INFO] 
[INFO] --- maven-clean-plugin:2.5:clean (default-clean) @ hive-hcatalog-it-unit 
---
[INFO] Deleting 
/data/hive-ptest/working/apache-svn-trunk-source/itests/hcatalog-unit (includes 
= [datanucleus.log, derby.log], excludes = [])
[INFO] 
[INFO] --- maven-enforcer-plugin:1.3.1:enforce (enforce-no-snapshots) @ 
hive-hcatalog-it-unit ---
[INFO] 
[INFO] --- maven-antrun-plugin:1.7:run (download-spark) @ hive-hcatalog-it-unit 
---
[INFO] Executing tasks

main:
[INFO] Executed tasks
[INFO] 
[INFO] --- maven-remote-resources-plugin:1.5:process (default) @ 
hive-hcatalog-it-unit ---
[INFO] 
[INFO] --- maven-resources-plugin:2.6:resources (default-resources) @ 
hive-hcatalog-it-unit ---
[INFO] Using 'UTF-8' encoding to copy filtered resources.
[INFO] skip non existing resourceDirectory 
/data/hive-ptest/working/apache-svn-trunk-source/itests/hcatalog-unit/src/main/resources
[INFO] Copying 3 resources
[INFO] 
[INFO] --- maven-antrun-plugin:1.7:run (define-classpath) @ 
hive-hcatalog-it-unit ---
[INFO] Executing tasks

main:
[INFO] Executed tasks
[INFO] 
[INFO] --- maven-compiler-plugin:3.1:compile (default-compile) @ 
hive-hcatalog-it-unit ---
[INFO] No sources to compile
[INFO] 
[INFO] --- maven-resources-plugin:2.6:testResources (default-testResources) @ 
hive-hcatalog-it-unit ---
[INFO] Using 'UTF-8' encoding to copy filtered resources.
[INFO] skip non existing resourceDirectory 
/data/hive-ptest/working/apache-svn-trunk-source/itests/hcatalog-unit/src/test/resources
[INFO] Copying 3 resources
[INFO] 
[INFO] --- maven-antrun-plugin:1.7:run (setup-test-dirs) @ 
hive-hcatalog-it-unit ---
[INFO] Executing tasks

main:
[mkdir] Created dir: 
/data/hive-ptest/working/apache-svn-trunk-source/itests/hcatalog-unit/target/tmp
[mkdir] Created dir: 
/data/hive-ptest/working/apache-svn-trunk-source/itests/hcatalog-unit/target/warehouse
[mkdir] Created dir: 
/data/hive-ptest/working/apache-svn-trunk-source/itests/hcatalog-unit/target/tmp/conf
 [copy] Copying 11 files to 
/data/hive-ptest/working/apache-svn-trunk-source/itests/hcatalog-unit/target/tmp/conf
[INFO] Executed tasks
[INFO] 
[INFO] --- maven-compiler-plugin:3.1:testCompile (default-testCompile) @ 
hive-hcatalog-it-unit ---
[INFO] Compiling 8 source files to 
/data/hive-ptest/working/apache-svn-trunk-source/itests/hcatalog-unit/target/test-classes
[WARNING] 
/data/hive-ptest/working/apache-svn-trunk-source/itests/hcatalog-unit/src/test/java/org/apache/hive/hcatalog/hbase/ManyMiniCluster.java:
 Some input files use or override a deprecated API.
[WARNING] 
/data/hive-ptest/working/apache-svn-trunk-source/itests/hcatalog-unit/src/test/java/org/apache/hive/hcatalog/hbase/ManyMiniCluster.java:
 Recompile with -Xlint:deprecation for details.
[INFO] 
[INFO] --- maven-surefire-plugin:2.16:test (default-test) @ 
hive-hcatalog-it-unit ---
[INFO] Tests are skipped.
[INFO] 
[INFO] --- maven-jar-plugin:2.2:jar (default-jar) @ hive-hcatalog-it-unit ---
[INFO] Building jar: 
/data/hive-ptest/working/apache-svn-trunk-source/itests/hcatalog-unit/target/hive-hcatalog-it-unit-1.2.0-SNAPSHOT.jar
[INFO] 
[INFO] --- maven-site-plugin:3.3:attach-descriptor (attach-descriptor) @ 
hive-hcatalog-it-unit ---
[INFO] 
[INFO] --- maven-jar-plugin:2.2:test-jar (default) @ hive-hcatalog-it-unit ---
[INFO] Building jar: 

[jira] [Commented] (HIVE-5341) Link doesn't work. Needs to be updated as mentioned in the Description

2015-03-17 Thread Fahd Khan (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-5341?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14365112#comment-14365112
 ] 

Fahd Khan commented on HIVE-5341:
-

you can find the file on the following link as well.


 Link doesn't work. Needs to be updated as mentioned in the Description
 --

 Key: HIVE-5341
 URL: https://issues.apache.org/jira/browse/HIVE-5341
 Project: Hive
  Issue Type: Bug
  Components: Documentation
Reporter: Rakesh Chouhan
Assignee: Lefty Leverenz
  Labels: documentation

 Go to.. Apache HIVE Getting Started Documentation
 https://cwiki.apache.org/confluence/display/Hive/GettingStarted
 Under Section ...
 Simple Example Use Cases
 MovieLens User Ratings
 wget http://www.grouplens.org/system/files/ml-data.tar+0.gz
 The link mentioned as per the document does not work. It needs to be updated 
 to the below URL.
 http://www.grouplens.org/sites/www.grouplens.org/external_files/data/ml-data.tar.gz
 I am setting this defect's priority as a Blocker because, user's will not be 
 able to continue their hands on exercises, unless they find the correct URL 
 to download the mentioned file.
 Referenced from:
 http://mail-archives.apache.org/mod_mbox/hive-user/201302.mbox/%3c8a0c145b-4db9-4d26-8613-8ca1bd741...@daum.net%3E.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-5341) Link doesn't work. Needs to be updated as mentioned in the Description

2015-03-17 Thread Fahd Khan (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-5341?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14365113#comment-14365113
 ] 

Fahd Khan commented on HIVE-5341:
-

https://code.google.com/p/hadoop-clusternet/source/browse/trunk/clusternet/thirdparty/data/ml-data.tar__0.gz?r=20

 Link doesn't work. Needs to be updated as mentioned in the Description
 --

 Key: HIVE-5341
 URL: https://issues.apache.org/jira/browse/HIVE-5341
 Project: Hive
  Issue Type: Bug
  Components: Documentation
Reporter: Rakesh Chouhan
Assignee: Lefty Leverenz
  Labels: documentation

 Go to.. Apache HIVE Getting Started Documentation
 https://cwiki.apache.org/confluence/display/Hive/GettingStarted
 Under Section ...
 Simple Example Use Cases
 MovieLens User Ratings
 wget http://www.grouplens.org/system/files/ml-data.tar+0.gz
 The link mentioned as per the document does not work. It needs to be updated 
 to the below URL.
 http://www.grouplens.org/sites/www.grouplens.org/external_files/data/ml-data.tar.gz
 I am setting this defect's priority as a Blocker because, user's will not be 
 able to continue their hands on exercises, unless they find the correct URL 
 to download the mentioned file.
 Referenced from:
 http://mail-archives.apache.org/mod_mbox/hive-user/201302.mbox/%3c8a0c145b-4db9-4d26-8613-8ca1bd741...@daum.net%3E.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9697) Hive on Spark is not as aggressive as MR on map join [Spark Branch]

2015-03-17 Thread Xuefu Zhang (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9697?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14365114#comment-14365114
 ] 

Xuefu Zhang commented on HIVE-9697:
---

It seems that we all agree that rawDataSize is more practical for Spark. Could 
anyone give a summary on if it's the default or how to make it as default? If 
code change is required, we can propose a patch here. Thanks.

 Hive on Spark is not as aggressive as MR on map join [Spark Branch]
 ---

 Key: HIVE-9697
 URL: https://issues.apache.org/jira/browse/HIVE-9697
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Reporter: Xin Hao

 We have a finding during running some Big-Bench cases:
 when the same small table size threshold is used, Map Join operator will not 
 be generated in Stage Plans for Hive on Spark, while will be generated for 
 Hive on MR.
 For example, When we run BigBench Q25, the meta info of one input ORC table 
 is as below:
 totalSize=1748955 (about 1.5M)
 rawDataSize=123050375 (about 120M)
 If we use the following parameter settings,
 set hive.auto.convert.join=true;
 set hive.mapjoin.smalltable.filesize=2500;
 set hive.auto.convert.join.noconditionaltask=true;
 set hive.auto.convert.join.noconditionaltask.size=1; (100M)
 Map Join will be enabled for Hive on MR mode, while will not be enabled for 
 Hive on Spark.
 We found that for Hive on MR, the HDFS file size for the table 
 (ContentSummary.getLength(), should approximate the value of ‘totalSize’) 
 will be used to compare with the threshold 100M (smaller than 100M), while 
 for Hive on Spark 'rawDataSize' will be used to compare with the threshold 
 100M (larger than 100M). That's why MapJoin is not enabled for Hive on Spark 
 for this case. And as a result Hive on Spark will get much lower performance 
 data than Hive on MR for this case.
 When we set  hive.auto.convert.join.noconditionaltask.size=15000; (150M), 
 MapJoin will be enabled for Hive on Spark mode, and Hive on Spark will have 
 similar performance data with Hive on MR by then.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-7018) Table and Partition tables have column LINK_TARGET_ID in Mysql scripts but not others

2015-03-17 Thread Aihua Xu (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-7018?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14365141#comment-14365141
 ] 

Aihua Xu commented on HIVE-7018:


I think it's just removing the ending ',' , not removing the complete line.

 Table and Partition tables have column LINK_TARGET_ID in Mysql scripts but 
 not others
 -

 Key: HIVE-7018
 URL: https://issues.apache.org/jira/browse/HIVE-7018
 Project: Hive
  Issue Type: Bug
Reporter: Brock Noland
Assignee: Yongzhi Chen
 Attachments: HIVE-7018.1.patch


 It appears that at least postgres and oracle do not have the LINK_TARGET_ID 
 column while mysql does.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9990) TestMultiSessionsHS2WithLocalClusterSpark is failing

2015-03-17 Thread Xuefu Zhang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9990?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Xuefu Zhang updated HIVE-9990:
--
Description: 
At least sometimes. I can reproduce it with mvn test 
-Dtest=TestMultiSessionsHS2WithLocalClusterSpark -Phadoop-2 consistently on my 
local box (both trunk and spark branch).
{code}
---
 T E S T S
---
Running org.apache.hive.jdbc.TestMultiSessionsHS2WithLocalClusterSpark
Tests run: 2, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 92.438 sec  
FAILURE! - in org.apache.hive.jdbc.TestMultiSessionsHS2WithLocalClusterSpark
testSparkQuery(org.apache.hive.jdbc.TestMultiSessionsHS2WithLocalClusterSpark)  
Time elapsed: 21.514 sec   ERROR!
java.util.concurrent.ExecutionException: java.sql.SQLException: Error while 
processing statement: FAILED: Execution Error, return code 3 from 
org.apache.hadoop.hive.ql.exec.spark.SparkTask
at org.apache.hive.jdbc.HiveStatement.execute(HiveStatement.java:296)
at 
org.apache.hive.jdbc.HiveStatement.executeQuery(HiveStatement.java:392)
at 
org.apache.hive.jdbc.TestMultiSessionsHS2WithLocalClusterSpark.verifyResult(TestMultiSessionsHS2WithLocalClusterSpark.java:244)
at 
org.apache.hive.jdbc.TestMultiSessionsHS2WithLocalClusterSpark.testKvQuery(TestMultiSessionsHS2WithLocalClusterSpark.java:220)
at 
org.apache.hive.jdbc.TestMultiSessionsHS2WithLocalClusterSpark.access$000(TestMultiSessionsHS2WithLocalClusterSpark.java:53)
{code}

The error was also seen in HIVE-9934 test run.

  was:
At least sometimes. I can reproduce it with mvn test 
-Dtest=TestMultiSessionsHS2WithLocalClusterSpark -Phadoop-2 consistently on my 
local box.
{code}
---
 T E S T S
---
Running org.apache.hive.jdbc.TestMultiSessionsHS2WithLocalClusterSpark
Tests run: 2, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 92.438 sec  
FAILURE! - in org.apache.hive.jdbc.TestMultiSessionsHS2WithLocalClusterSpark
testSparkQuery(org.apache.hive.jdbc.TestMultiSessionsHS2WithLocalClusterSpark)  
Time elapsed: 21.514 sec   ERROR!
java.util.concurrent.ExecutionException: java.sql.SQLException: Error while 
processing statement: FAILED: Execution Error, return code 3 from 
org.apache.hadoop.hive.ql.exec.spark.SparkTask
at org.apache.hive.jdbc.HiveStatement.execute(HiveStatement.java:296)
at 
org.apache.hive.jdbc.HiveStatement.executeQuery(HiveStatement.java:392)
at 
org.apache.hive.jdbc.TestMultiSessionsHS2WithLocalClusterSpark.verifyResult(TestMultiSessionsHS2WithLocalClusterSpark.java:244)
at 
org.apache.hive.jdbc.TestMultiSessionsHS2WithLocalClusterSpark.testKvQuery(TestMultiSessionsHS2WithLocalClusterSpark.java:220)
at 
org.apache.hive.jdbc.TestMultiSessionsHS2WithLocalClusterSpark.access$000(TestMultiSessionsHS2WithLocalClusterSpark.java:53)
{code}

The error was also seen in HIVE-9934 test run.


 TestMultiSessionsHS2WithLocalClusterSpark is failing
 

 Key: HIVE-9990
 URL: https://issues.apache.org/jira/browse/HIVE-9990
 Project: Hive
  Issue Type: Bug
  Components: Spark
Affects Versions: 1.2.0
Reporter: Xuefu Zhang

 At least sometimes. I can reproduce it with mvn test 
 -Dtest=TestMultiSessionsHS2WithLocalClusterSpark -Phadoop-2 consistently on 
 my local box (both trunk and spark branch).
 {code}
 ---
  T E S T S
 ---
 Running org.apache.hive.jdbc.TestMultiSessionsHS2WithLocalClusterSpark
 Tests run: 2, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 92.438 sec 
  FAILURE! - in 
 org.apache.hive.jdbc.TestMultiSessionsHS2WithLocalClusterSpark
 testSparkQuery(org.apache.hive.jdbc.TestMultiSessionsHS2WithLocalClusterSpark)
   Time elapsed: 21.514 sec   ERROR!
 java.util.concurrent.ExecutionException: java.sql.SQLException: Error while 
 processing statement: FAILED: Execution Error, return code 3 from 
 org.apache.hadoop.hive.ql.exec.spark.SparkTask
   at org.apache.hive.jdbc.HiveStatement.execute(HiveStatement.java:296)
   at 
 org.apache.hive.jdbc.HiveStatement.executeQuery(HiveStatement.java:392)
   at 
 org.apache.hive.jdbc.TestMultiSessionsHS2WithLocalClusterSpark.verifyResult(TestMultiSessionsHS2WithLocalClusterSpark.java:244)
   at 
 org.apache.hive.jdbc.TestMultiSessionsHS2WithLocalClusterSpark.testKvQuery(TestMultiSessionsHS2WithLocalClusterSpark.java:220)
   at 
 org.apache.hive.jdbc.TestMultiSessionsHS2WithLocalClusterSpark.access$000(TestMultiSessionsHS2WithLocalClusterSpark.java:53)
 {code}
 The error was also seen in HIVE-9934 test run.



--
This message was sent by 

[jira] [Updated] (HIVE-9934) Vulnerability in LdapAuthenticationProviderImpl enables HiveServer2 client to degrade the authentication mechanism to none, allowing authentication without password

2015-03-17 Thread Xuefu Zhang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9934?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Xuefu Zhang updated HIVE-9934:
--
Attachment: HIVE-9934.3.patch

Attached the same patch for another test run.

 Vulnerability in LdapAuthenticationProviderImpl enables HiveServer2 client to 
 degrade the authentication mechanism to none, allowing authentication 
 without password
 --

 Key: HIVE-9934
 URL: https://issues.apache.org/jira/browse/HIVE-9934
 Project: Hive
  Issue Type: Bug
  Components: Security
Affects Versions: 1.1.0
Reporter: Chao
Assignee: Chao
 Attachments: HIVE-9934.1.patch, HIVE-9934.2.patch, HIVE-9934.3.patch, 
 HIVE-9934.3.patch


 Vulnerability in LdapAuthenticationProviderImpl enables HiveServer2 client to 
 degrade the authentication mechanism to none, allowing authentication 
 without password.
 See: http://docs.oracle.com/javase/jndi/tutorial/ldap/security/simple.html
 “If you supply an empty string, an empty byte/char array, or null to the 
 Context.SECURITY_CREDENTIALS environment property, then the authentication 
 mechanism will be none. This is because the LDAP requires the password to 
 be nonempty for simple authentication. The protocol automatically converts 
 the authentication to none if a password is not supplied.”
  
 Since the LdapAuthenticationProviderImpl.Authenticate method is relying on a 
 NamingException being thrown during creation of initial context, it does not 
 fail when the context result is an “unauthenticated” positive response from 
 the LDAP server. The end result is, one can authenticate with HiveServer2 
 using the LdapAuthenticationProviderImpl with only a user name and an empty 
 password.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9957) Hive 1.1.0 not compatible with Hadoop 2.4.0

2015-03-17 Thread JIRA

[ 
https://issues.apache.org/jira/browse/HIVE-9957?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14365216#comment-14365216
 ] 

Sergio Peña commented on HIVE-9957:
---

Yes. Encryption was merged on 1.1.0.

 Hive 1.1.0 not compatible with Hadoop 2.4.0
 ---

 Key: HIVE-9957
 URL: https://issues.apache.org/jira/browse/HIVE-9957
 Project: Hive
  Issue Type: Bug
  Components: Encryption
Reporter: Vivek Shrivastava
Assignee: Sergio Peña
  Labels: TODOC1.2
 Fix For: 1.2.0

 Attachments: HIVE-9957.1.patch


 Getting this exception while accessing data through Hive. 
 Exception in thread main java.lang.NoSuchMethodError: 
 org.apache.hadoop.hdfs.DFSClient.getKeyProvider()Lorg/apache/hadoop/crypto/key/KeyProvider;
 at 
 org.apache.hadoop.hive.shims.Hadoop23Shims$HdfsEncryptionShim.init(Hadoop23Shims.java:1152)
 at 
 org.apache.hadoop.hive.shims.Hadoop23Shims.createHdfsEncryptionShim(Hadoop23Shims.java:1279)
 at 
 org.apache.hadoop.hive.ql.session.SessionState.getHdfsEncryptionShim(SessionState.java:392)
 at 
 org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.isPathEncrypted(SemanticAnalyzer.java:1756)
 at 
 org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.getStagingDirectoryPathname(SemanticAnalyzer.java:1875)
 at 
 org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.getMetaData(SemanticAnalyzer.java:1689)
 at 
 org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.getMetaData(SemanticAnalyzer.java:1427)
 at 
 org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genResolvedParseTree(SemanticAnalyzer.java:10132)
 at 
 org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.analyzeInternal(SemanticAnalyzer.java:10147)
 at 
 org.apache.hadoop.hive.ql.parse.CalcitePlanner.analyzeInternal(CalcitePlanner.java:192)
 at 
 org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer.analyze(BaseSemanticAnalyzer.java:222)
 at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:421)
 at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:307)
 at org.apache.hadoop.hive.ql.Driver.compileInternal(Driver.java:1112)
 at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1160)
 at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1049)
 at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1039)
 at 
 org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:207)
 at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:159)
 at 
 org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:370)
 at 
 org.apache.hadoop.hive.cli.CliDriver.executeDriver(CliDriver.java:754)
 at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:675)
 at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:615)
 at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
 at 
 sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
 at 
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
 at java.lang.reflect.Method.invoke(Method.java:606)
 at org.apache.hadoop.util.RunJar.main(RunJar.java:212)



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9945) FunctionTask.conf hides Task.conf field

2015-03-17 Thread Alan Gates (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9945?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14365376#comment-14365376
 ] 

Alan Gates commented on HIVE-9945:
--

+1

 FunctionTask.conf hides Task.conf field
 ---

 Key: HIVE-9945
 URL: https://issues.apache.org/jira/browse/HIVE-9945
 Project: Hive
  Issue Type: Bug
Reporter: Alexander Pivovarov
Assignee: Alexander Pivovarov
Priority: Minor
 Attachments: HIVE-9945.1.patch


 Task class has protected field conf.
 FunctionTask can use it instead of creating another conf field.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9991) Cannot do a SELECT on external tables that are on S3 due to Encryption error

2015-03-17 Thread JIRA

 [ 
https://issues.apache.org/jira/browse/HIVE-9991?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sergio Peña updated HIVE-9991:
--
Attachment: HIVE-9991.1.patch

 Cannot do a SELECT on external tables that are on S3 due to Encryption error
 

 Key: HIVE-9991
 URL: https://issues.apache.org/jira/browse/HIVE-9991
 Project: Hive
  Issue Type: Bug
Affects Versions: 1.0.0
Reporter: Sergio Peña
Assignee: Sergio Peña
 Attachments: HIVE-9991.1.patch


 I cannot do any select query on external tables that are not part of HDFS. 
 For example S3.
 {code}
 Select * from my_table limit 10;
 FAILED: SemanticException Unable to determine if s3n://my-bucket/is 
 encrypted: java.lang.IllegalArgumentException: Wrong FS: s3n://my-bucket/, 
 expected: hdfs://0.0.0.0:8020
 {code}
 This error is due to a internal function that checks if a table is encrypted 
 or not. This is only supported on HDFS files, but the check is happening on 
 any external table as well causing the above error.
 To fix this, we should check for encrypted tables only for HDFS tables. And 
 skip the check for any other file schema.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9934) Vulnerability in LdapAuthenticationProviderImpl enables HiveServer2 client to degrade the authentication mechanism to none, allowing authentication without password

2015-03-17 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9934?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14365402#comment-14365402
 ] 

Hive QA commented on HIVE-9934:
---



{color:green}Overall{color}: +1 all checks pass

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12705088/HIVE-9934.3.patch

{color:green}SUCCESS:{color} +1 7770 tests passed

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/3056/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/3056/console
Test logs: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-3056/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12705088 - PreCommit-HIVE-TRUNK-Build

 Vulnerability in LdapAuthenticationProviderImpl enables HiveServer2 client to 
 degrade the authentication mechanism to none, allowing authentication 
 without password
 --

 Key: HIVE-9934
 URL: https://issues.apache.org/jira/browse/HIVE-9934
 Project: Hive
  Issue Type: Bug
  Components: Security
Affects Versions: 1.1.0
Reporter: Chao
Assignee: Chao
 Attachments: HIVE-9934.1.patch, HIVE-9934.2.patch, HIVE-9934.3.patch, 
 HIVE-9934.3.patch


 Vulnerability in LdapAuthenticationProviderImpl enables HiveServer2 client to 
 degrade the authentication mechanism to none, allowing authentication 
 without password.
 See: http://docs.oracle.com/javase/jndi/tutorial/ldap/security/simple.html
 “If you supply an empty string, an empty byte/char array, or null to the 
 Context.SECURITY_CREDENTIALS environment property, then the authentication 
 mechanism will be none. This is because the LDAP requires the password to 
 be nonempty for simple authentication. The protocol automatically converts 
 the authentication to none if a password is not supplied.”
  
 Since the LdapAuthenticationProviderImpl.Authenticate method is relying on a 
 NamingException being thrown during creation of initial context, it does not 
 fail when the context result is an “unauthenticated” positive response from 
 the LDAP server. The end result is, one can authenticate with HiveServer2 
 using the LdapAuthenticationProviderImpl with only a user name and an empty 
 password.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9555) assorted ORC refactorings for LLAP on trunk

2015-03-17 Thread Prasanth Jayachandran (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9555?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14366676#comment-14366676
 ] 

Prasanth Jayachandran commented on HIVE-9555:
-

And for 
https://github.com/apache/hive/blob/llap/llap-server/src/java/org/apache/hadoop/hive/llap/io/decode/orc/stream/readers/StringStreamReader.java#L37

 assorted ORC refactorings for LLAP on trunk
 ---

 Key: HIVE-9555
 URL: https://issues.apache.org/jira/browse/HIVE-9555
 Project: Hive
  Issue Type: Bug
Reporter: Sergey Shelukhin
Assignee: Sergey Shelukhin
 Attachments: HIVE-9555.01.patch, HIVE-9555.02.patch, 
 HIVE-9555.03.patch, HIVE-9555.04.patch, HIVE-9555.05.patch, 
 HIVE-9555.06.patch, HIVE-9555.07.patch, HIVE-9555.patch


 To minimize conflicts and given that ORC is being developed rapidly on trunk, 
 I would like to refactor some parts of ORC in advance based on the changes 
 in LLAP branch. Mostly it concerns making parts of ORC code (esp. SARG, but 
 also some internal methods) more modular and easier to use from alternative 
 codepaths. There's also significant change to how data reading is handled - 
 BufferChunk inherits from DiskRange; the reader receives a list of 
 DiskRange-s (as before), but instead of making a list of buffer chunks it 
 replaces ranges with buffer chunks in the original (linked) list. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9975) Renaming a nonexisting partition should not throw out NullPointerException

2015-03-17 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9975?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14366687#comment-14366687
 ] 

Hive QA commented on HIVE-9975:
---



{color:green}Overall{color}: +1 all checks pass

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12705245/HIVE-9975.2.patch

{color:green}SUCCESS:{color} +1 7770 tests passed

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/3067/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/3067/console
Test logs: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-3067/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12705245 - PreCommit-HIVE-TRUNK-Build

 Renaming a nonexisting partition should not throw out NullPointerException
 --

 Key: HIVE-9975
 URL: https://issues.apache.org/jira/browse/HIVE-9975
 Project: Hive
  Issue Type: Bug
Reporter: Chaoyu Tang
Assignee: Chaoyu Tang
Priority: Minor
 Attachments: HIVE-9975.1.patch, HIVE-9975.2.patch, HIVE-9975.patch


 Renaming a nonexisting partition should not throw out NullPointerException. 
 create table testpart (col1 int, col2 string, col3 string) partitioned by 
 (part string);
 alter table testpart partition (part = 'nonexisting') rename to partition 
 (part = 'existing');
 we get NPE like following:
 {code}
 15/03/16 10:16:11 ERROR exec.DDLTask: java.lang.NullPointerException
   at 
 org.apache.hadoop.hive.ql.exec.DDLTask.renamePartition(DDLTask.java:944)
   at org.apache.hadoop.hive.ql.exec.DDLTask.execute(DDLTask.java:350)
   at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:160)
   at 
 org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:88)
   at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:1642)
   at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:1402)
   at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1187)
   at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1053)
   at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1043)
   at 
 org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:207)
   at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:159)
   at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:370)
   at 
 org.apache.hadoop.hive.cli.CliDriver.executeDriver(CliDriver.java:754)
   at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:675)
   at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:615)
 FAILED: Execution Error, return code 1 from 
 org.apache.hadoop.hive.ql.exec.DDLTask. null
 {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9555) assorted ORC refactorings for LLAP on trunk

2015-03-17 Thread Prasanth Jayachandran (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9555?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14366672#comment-14366672
 ] 

Prasanth Jayachandran commented on HIVE-9555:
-

Its needed for 
https://github.com/apache/hive/blob/llap/llap-server/src/java/org/apache/hadoop/hive/llap/io/decode/orc/stream/readers/CharacterStreamReader.java#L37
to handle char and varchar in the same class.

 assorted ORC refactorings for LLAP on trunk
 ---

 Key: HIVE-9555
 URL: https://issues.apache.org/jira/browse/HIVE-9555
 Project: Hive
  Issue Type: Bug
Reporter: Sergey Shelukhin
Assignee: Sergey Shelukhin
 Attachments: HIVE-9555.01.patch, HIVE-9555.02.patch, 
 HIVE-9555.03.patch, HIVE-9555.04.patch, HIVE-9555.05.patch, 
 HIVE-9555.06.patch, HIVE-9555.07.patch, HIVE-9555.patch


 To minimize conflicts and given that ORC is being developed rapidly on trunk, 
 I would like to refactor some parts of ORC in advance based on the changes 
 in LLAP branch. Mostly it concerns making parts of ORC code (esp. SARG, but 
 also some internal methods) more modular and easier to use from alternative 
 codepaths. There's also significant change to how data reading is handled - 
 BufferChunk inherits from DiskRange; the reader receives a list of 
 DiskRange-s (as before), but instead of making a list of buffer chunks it 
 replaces ranges with buffer chunks in the original (linked) list. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)