[jira] [Commented] (HIVE-2055) Hive should add HBase classpath dependencies when available
[ https://issues.apache.org/jira/browse/HIVE-2055?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13824558#comment-13824558 ] Hudson commented on HIVE-2055: -- FAILURE: Integrated in HBase-0.94 #1203 (See [https://builds.apache.org/job/HBase-0.94/1203/]) HBASE-9165 [mapreduce] Modularize building dependency jars Separate adding HBase and dependencies from adding other job dependencies, and expose it as a separate method that other projects can use (for PIG-3285, HIVE-2055). (ndimiduk: rev 1542414) * /hbase/branches/0.94/src/main/java/org/apache/hadoop/hbase/mapred/TableMapReduceUtil.java * /hbase/branches/0.94/src/main/java/org/apache/hadoop/hbase/mapreduce/TableMapReduceUtil.java * /hbase/branches/0.94/src/test/java/org/apache/hadoop/hbase/mapreduce/TestTableMapReduce.java Hive should add HBase classpath dependencies when available --- Key: HIVE-2055 URL: https://issues.apache.org/jira/browse/HIVE-2055 Project: Hive Issue Type: Bug Components: HBase Handler Affects Versions: 0.10.0 Reporter: sajith v Attachments: 0001-HIVE-2055-include-hbase-dependencies-in-launch-scrip.patch, HIVE-2055.patch Created an external table in hive , which points to the HBase table. When tried to query a column using the column name in select clause got the following exception : ( java.lang.ClassNotFoundException: org.apache.hadoop.hive.hbase.HiveHBaseTableInputFormat), errorCode:12, SQLState:42000) -- This message was sent by Atlassian JIRA (v6.1#6144)
[jira] [Commented] (HIVE-2055) Hive should add HBase classpath dependencies when available
[ https://issues.apache.org/jira/browse/HIVE-2055?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13824586#comment-13824586 ] Hudson commented on HIVE-2055: -- SUCCESS: Integrated in HBase-0.94-security #337 (See [https://builds.apache.org/job/HBase-0.94-security/337/]) HBASE-9165 [mapreduce] Modularize building dependency jars Separate adding HBase and dependencies from adding other job dependencies, and expose it as a separate method that other projects can use (for PIG-3285, HIVE-2055). (ndimiduk: rev 1542414) * /hbase/branches/0.94/src/main/java/org/apache/hadoop/hbase/mapred/TableMapReduceUtil.java * /hbase/branches/0.94/src/main/java/org/apache/hadoop/hbase/mapreduce/TableMapReduceUtil.java * /hbase/branches/0.94/src/test/java/org/apache/hadoop/hbase/mapreduce/TestTableMapReduce.java Hive should add HBase classpath dependencies when available --- Key: HIVE-2055 URL: https://issues.apache.org/jira/browse/HIVE-2055 Project: Hive Issue Type: Bug Components: HBase Handler Affects Versions: 0.10.0 Reporter: sajith v Attachments: 0001-HIVE-2055-include-hbase-dependencies-in-launch-scrip.patch, HIVE-2055.patch Created an external table in hive , which points to the HBase table. When tried to query a column using the column name in select clause got the following exception : ( java.lang.ClassNotFoundException: org.apache.hadoop.hive.hbase.HiveHBaseTableInputFormat), errorCode:12, SQLState:42000) -- This message was sent by Atlassian JIRA (v6.1#6144)
[jira] [Commented] (HIVE-5601) NPE in ORC's PPD when using select * from table with where predicate
[ https://issues.apache.org/jira/browse/HIVE-5601?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13823765#comment-13823765 ] Hudson commented on HIVE-5601: -- FAILURE: Integrated in Hive-branch-0.12-hadoop1 #32 (See [https://builds.apache.org/job/Hive-branch-0.12-hadoop1/32/]) HIVE-5601: NPE in ORC's PPD when using select * from table with where predicate pushdown (Prasanth J via Owen O'Malley and Gunther Hagleitner) (omalley: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1542024) * /hive/branches/branch-0.12/ql/src/java/org/apache/hadoop/hive/ql/io/orc/ReaderImpl.java * /hive/branches/branch-0.12/ql/src/test/queries/clientpositive/orc_predicate_pushdown.q * /hive/branches/branch-0.12/ql/src/test/results/clientpositive/orc_predicate_pushdown.q.out NPE in ORC's PPD when using select * from table with where predicate - Key: HIVE-5601 URL: https://issues.apache.org/jira/browse/HIVE-5601 Project: Hive Issue Type: Bug Affects Versions: 0.12.0 Reporter: Prasanth J Assignee: Prasanth J Priority: Critical Labels: ORC Fix For: 0.13.0, 0.12.1 Attachments: HIVE-5601.4-branch-0.12.patch.txt, HIVE-5601.5.patch.txt, HIVE-5601.branch-0.12.2.patch.txt, HIVE-5601.branch-0.12.3.patch.txt, HIVE-5601.branch-0.12.4.patch.txt, HIVE-5601.branch-12.1.patch.txt, HIVE-5601.trunk.1.patch.txt, HIVE-5601.trunk.2.patch.txt, HIVE-5601.trunk.3.patch.txt, HIVE-5601.trunk.4.patch.txt, HIVE-5601.trunk.5.patch.txt ORCInputFormat has a method findIncludedColumns() which returns boolean array of included columns. In case of the following query {code}select * from qlog_orc where id1000 limit 10;{code} where all columns are selected the findIncludedColumns() returns null. This will result in a NPE when PPD is enabled. Following is the stack trace {code}Caused by: java.lang.NullPointerException at org.apache.hadoop.hive.ql.io.orc.RecordReaderImpl.planReadPartialDataStreams(RecordReaderImpl.java:2387) at org.apache.hadoop.hive.ql.io.orc.RecordReaderImpl.readPartialDataStreams(RecordReaderImpl.java:2543) at org.apache.hadoop.hive.ql.io.orc.RecordReaderImpl.readStripe(RecordReaderImpl.java:2200) at org.apache.hadoop.hive.ql.io.orc.RecordReaderImpl.advanceStripe(RecordReaderImpl.java:2573) at org.apache.hadoop.hive.ql.io.orc.RecordReaderImpl.advanceToNextRow(RecordReaderImpl.java:2615) at org.apache.hadoop.hive.ql.io.orc.RecordReaderImpl.init(RecordReaderImpl.java:132) at org.apache.hadoop.hive.ql.io.orc.ReaderImpl.rows(ReaderImpl.java:348) at org.apache.hadoop.hive.ql.io.orc.OrcInputFormat$OrcRecordReader.init(OrcInputFormat.java:99) at org.apache.hadoop.hive.ql.io.orc.OrcInputFormat.getRecordReader(OrcInputFormat.java:241) at org.apache.hadoop.hive.ql.io.HiveInputFormat.getRecordReader(HiveInputFormat.java:237) ... 8 more{code} -- This message was sent by Atlassian JIRA (v6.1#6144)
[jira] [Commented] (HIVE-5601) NPE in ORC's PPD when using select * from table with where predicate
[ https://issues.apache.org/jira/browse/HIVE-5601?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13824409#comment-13824409 ] Hudson commented on HIVE-5601: -- FAILURE: Integrated in Hive-branch-0.12-hadoop2 #23 (See [https://builds.apache.org/job/Hive-branch-0.12-hadoop2/23/]) HIVE-5601: NPE in ORC's PPD when using select * from table with where predicate pushdown (Prasanth J via Owen O'Malley and Gunther Hagleitner) (omalley: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1542024) * /hive/branches/branch-0.12/ql/src/java/org/apache/hadoop/hive/ql/io/orc/ReaderImpl.java * /hive/branches/branch-0.12/ql/src/test/queries/clientpositive/orc_predicate_pushdown.q * /hive/branches/branch-0.12/ql/src/test/results/clientpositive/orc_predicate_pushdown.q.out NPE in ORC's PPD when using select * from table with where predicate - Key: HIVE-5601 URL: https://issues.apache.org/jira/browse/HIVE-5601 Project: Hive Issue Type: Bug Affects Versions: 0.12.0 Reporter: Prasanth J Assignee: Prasanth J Priority: Critical Labels: ORC Fix For: 0.13.0, 0.12.1 Attachments: HIVE-5601.4-branch-0.12.patch.txt, HIVE-5601.5.patch.txt, HIVE-5601.branch-0.12.2.patch.txt, HIVE-5601.branch-0.12.3.patch.txt, HIVE-5601.branch-0.12.4.patch.txt, HIVE-5601.branch-12.1.patch.txt, HIVE-5601.trunk.1.patch.txt, HIVE-5601.trunk.2.patch.txt, HIVE-5601.trunk.3.patch.txt, HIVE-5601.trunk.4.patch.txt, HIVE-5601.trunk.5.patch.txt ORCInputFormat has a method findIncludedColumns() which returns boolean array of included columns. In case of the following query {code}select * from qlog_orc where id1000 limit 10;{code} where all columns are selected the findIncludedColumns() returns null. This will result in a NPE when PPD is enabled. Following is the stack trace {code}Caused by: java.lang.NullPointerException at org.apache.hadoop.hive.ql.io.orc.RecordReaderImpl.planReadPartialDataStreams(RecordReaderImpl.java:2387) at org.apache.hadoop.hive.ql.io.orc.RecordReaderImpl.readPartialDataStreams(RecordReaderImpl.java:2543) at org.apache.hadoop.hive.ql.io.orc.RecordReaderImpl.readStripe(RecordReaderImpl.java:2200) at org.apache.hadoop.hive.ql.io.orc.RecordReaderImpl.advanceStripe(RecordReaderImpl.java:2573) at org.apache.hadoop.hive.ql.io.orc.RecordReaderImpl.advanceToNextRow(RecordReaderImpl.java:2615) at org.apache.hadoop.hive.ql.io.orc.RecordReaderImpl.init(RecordReaderImpl.java:132) at org.apache.hadoop.hive.ql.io.orc.ReaderImpl.rows(ReaderImpl.java:348) at org.apache.hadoop.hive.ql.io.orc.OrcInputFormat$OrcRecordReader.init(OrcInputFormat.java:99) at org.apache.hadoop.hive.ql.io.orc.OrcInputFormat.getRecordReader(OrcInputFormat.java:241) at org.apache.hadoop.hive.ql.io.HiveInputFormat.getRecordReader(HiveInputFormat.java:237) ... 8 more{code} -- This message was sent by Atlassian JIRA (v6.1#6144)
[jira] [Commented] (HIVE-5667) ThriftCLIService log messages jumbled up
[ https://issues.apache.org/jira/browse/HIVE-5667?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13807053#comment-13807053 ] Hudson commented on HIVE-5667: -- FAILURE: Integrated in Hive-trunk-hadoop2-ptest #154 (See [https://builds.apache.org/job/Hive-trunk-hadoop2-ptest/154/]) HIVE-5667 - ThriftCLIService log messages jumbled up (Vaibhav Gumashta via Brock Noland) (brock: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1536361) * /hive/trunk/service/src/java/org/apache/hive/service/cli/thrift/ThriftCLIService.java ThriftCLIService log messages jumbled up Key: HIVE-5667 URL: https://issues.apache.org/jira/browse/HIVE-5667 Project: Hive Issue Type: Bug Components: HiveServer2 Reporter: Vaibhav Gumashta Assignee: Vaibhav Gumashta Fix For: 0.13.0 Attachments: HIVE-5667.1.patch ThriftCLIService log messages are not aligned with the methods correctly -- This message was sent by Atlassian JIRA (v6.1#6144)
[jira] [Commented] (HIVE-5666) use Path instead of String for IOContext.inputPath
[ https://issues.apache.org/jira/browse/HIVE-5666?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13807267#comment-13807267 ] Hudson commented on HIVE-5666: -- ABORTED: Integrated in Hive-trunk-hadoop2 #526 (See [https://builds.apache.org/job/Hive-trunk-hadoop2/526/]) HIVE-5666 : use Path instead of String for IOContext.inputPath (Thejas Nair via Ashutosh Chauhan) (hashutosh: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1536478) * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/MapJoinOperator.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/MapOperator.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/SMBMapJoinOperator.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/TableScanOperator.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/mr/ExecMapperContext.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/io/HiveContextAwareRecordReader.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/io/IOContext.java use Path instead of String for IOContext.inputPath -- Key: HIVE-5666 URL: https://issues.apache.org/jira/browse/HIVE-5666 Project: Hive Issue Type: Improvement Reporter: Thejas M Nair Assignee: Thejas M Nair Fix For: 0.13.0 Attachments: HIVE-5666.1.patch Path is converted to string in HiveContextAwareRecordReader to be stored in IOContext.inputPath, then in MapOperator normalizePath gets called on it which converts it back to Path. Path creation is expensive, so it is better to use Path instead of string through the call stack. This is also a step towards HIVE-3616. -- This message was sent by Atlassian JIRA (v6.1#6144)
[jira] [Commented] (HIVE-5667) ThriftCLIService log messages jumbled up
[ https://issues.apache.org/jira/browse/HIVE-5667?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13807268#comment-13807268 ] Hudson commented on HIVE-5667: -- ABORTED: Integrated in Hive-trunk-hadoop2 #526 (See [https://builds.apache.org/job/Hive-trunk-hadoop2/526/]) HIVE-5667 - ThriftCLIService log messages jumbled up (Vaibhav Gumashta via Brock Noland) (brock: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1536361) * /hive/trunk/service/src/java/org/apache/hive/service/cli/thrift/ThriftCLIService.java ThriftCLIService log messages jumbled up Key: HIVE-5667 URL: https://issues.apache.org/jira/browse/HIVE-5667 Project: Hive Issue Type: Bug Components: HiveServer2 Reporter: Vaibhav Gumashta Assignee: Vaibhav Gumashta Fix For: 0.13.0 Attachments: HIVE-5667.1.patch ThriftCLIService log messages are not aligned with the methods correctly -- This message was sent by Atlassian JIRA (v6.1#6144)
[jira] [Commented] (HIVE-5450) pTest2 TestReportParser is failing due to .svn directory
[ https://issues.apache.org/jira/browse/HIVE-5450?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13807269#comment-13807269 ] Hudson commented on HIVE-5450: -- ABORTED: Integrated in Hive-trunk-hadoop2 #526 (See [https://builds.apache.org/job/Hive-trunk-hadoop2/526/]) HIVE-5450 - pTest2 TestReportParser is failing due to .svn directory (Brock Noland reviewed by Ashutosh Chauhan) (brock: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1536508) * /hive/trunk/testutils/ptest2/src/test/java/org/apache/hive/ptest/execution/TestReportParser.java pTest2 TestReportParser is failing due to .svn directory Key: HIVE-5450 URL: https://issues.apache.org/jira/browse/HIVE-5450 Project: Hive Issue Type: Task Components: Testing Infrastructure Affects Versions: 0.13.0 Reporter: Ashutosh Chauhan Assignee: Brock Noland Fix For: 0.13.0 Attachments: HIVE-5450.patch Following test failed when I ran mvn package: {code} Running org.apache.hive.ptest.execution.TestPhase 2013-10-04 22:57:20,150 ERROR HostExecutor$5.call:379 Aborting drone during exec echo org.apache.hive.ptest.execution.AbortDroneException: Drone Drone [user=someuser, host=somehost, instance=0] exited with 255: SSHResult [command=echo, getExitCode()=255, getException()=null, getUser()=someuser, getHost()=somehost, getInstance()=0] at org.apache.hive.ptest.execution.HostExecutor$5.call(HostExecutor.java:379) at org.apache.hive.ptest.execution.HostExecutor$5.call(HostExecutor.java:368) at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) at java.util.concurrent.FutureTask.run(FutureTask.java:138) at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) at java.lang.Thread.run(Thread.java:662) Tests in error: test(org.apache.hive.ptest.execution.TestReportParser): src/test/resources/test-outputs/.svn (Is a directory) Tests run: 44, Failures: 0, Errors: 1, Skipped: 0 {code} NO PRECOMMIT TESTS -- This message was sent by Atlassian JIRA (v6.1#6144)
[jira] [Commented] (HIVE-5554) add more comments to CombineHiveInputFormat.java, BucketizedHiveInputFormat.java
[ https://issues.apache.org/jira/browse/HIVE-5554?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13807274#comment-13807274 ] Hudson commented on HIVE-5554: -- ABORTED: Integrated in Hive-trunk-hadoop2 #526 (See [https://builds.apache.org/job/Hive-trunk-hadoop2/526/]) HIVE-5554 : add more comments to CombineHiveInputFormat.java, BucketizedHiveInputFormat.java (Thejas Nair via Ashutosh Chauhan) (hashutosh: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1536517) * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/io/BucketizedHiveInputFormat.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/io/CombineHiveInputFormat.java add more comments to CombineHiveInputFormat.java, BucketizedHiveInputFormat.java Key: HIVE-5554 URL: https://issues.apache.org/jira/browse/HIVE-5554 Project: Hive Issue Type: Improvement Affects Versions: 0.12.0 Reporter: Thejas M Nair Assignee: Thejas M Nair Fix For: 0.13.0 Attachments: HIVE-5554.1.patch more comments to CombineHiveInputFormat.java, BucketizedHiveInputFormat.java to make it easier to understand. NO PRECOMMIT TESTS (tested build locally) -- This message was sent by Atlassian JIRA (v6.1#6144)
[jira] [Commented] (HIVE-5653) Vectorized Shuffle Join produces incorrect results
[ https://issues.apache.org/jira/browse/HIVE-5653?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13807271#comment-13807271 ] Hudson commented on HIVE-5653: -- ABORTED: Integrated in Hive-trunk-hadoop2 #526 (See [https://builds.apache.org/job/Hive-trunk-hadoop2/526/]) HIVE-5653 : Vectorized Shuffle Join produces incorrect results (Remus Rusanu via Ashutosh Chauhan) (hashutosh: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1536474) * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/VectorReduceSinkOperator.java * /hive/trunk/ql/src/test/queries/clientpositive/vectorized_shufflejoin.q * /hive/trunk/ql/src/test/results/clientpositive/vectorized_shufflejoin.q.out Vectorized Shuffle Join produces incorrect results -- Key: HIVE-5653 URL: https://issues.apache.org/jira/browse/HIVE-5653 Project: Hive Issue Type: Sub-task Affects Versions: 0.13.0 Reporter: Remus Rusanu Assignee: Remus Rusanu Fix For: 0.13.0 Attachments: HIVE-5653.1.patch Vectorized shuffle join should work out-of-the-box, but it produces empty result set. Investigating. -- This message was sent by Atlassian JIRA (v6.1#6144)
[jira] [Commented] (HIVE-5648) error when casting partition column to varchar in where clause
[ https://issues.apache.org/jira/browse/HIVE-5648?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13807273#comment-13807273 ] Hudson commented on HIVE-5648: -- ABORTED: Integrated in Hive-trunk-hadoop2 #526 (See [https://builds.apache.org/job/Hive-trunk-hadoop2/526/]) HIVE-5648 : error when casting partition column to varchar in where clause (Jason Dere via Ashutosh Chauhan) (hashutosh: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1536471) * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/ExprNodeConstantDesc.java * /hive/trunk/ql/src/test/queries/clientpositive/partition_varchar2.q * /hive/trunk/ql/src/test/results/clientpositive/partition_varchar2.q.out * /hive/trunk/serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/primitive/PrimitiveObjectInspectorFactory.java * /hive/trunk/serde/src/java/org/apache/hadoop/hive/serde2/typeinfo/TypeInfoFactory.java * /hive/trunk/serde/src/test/org/apache/hadoop/hive/serde2/objectinspector/primitive/TestPrimitiveObjectInspectorFactory.java error when casting partition column to varchar in where clause --- Key: HIVE-5648 URL: https://issues.apache.org/jira/browse/HIVE-5648 Project: Hive Issue Type: Bug Reporter: Jason Dere Assignee: Jason Dere Fix For: 0.13.0 Attachments: HIVE-5648.1.patch, HIVE-5648.2.patch hive select * from partition_varchar_2 where cast(dt as varchar(10)) = '2000-01-01'; FAILED: RuntimeException org.apache.hadoop.hive.ql.metadata.HiveException: java.lang.RuntimeException: Internal error: Cannot find ObjectInspector for VARCHAR -- This message was sent by Atlassian JIRA (v6.1#6144)
[jira] [Commented] (HIVE-5576) Blank lines missing from .q.out files created on Windows for testcase=TestCliDriver
[ https://issues.apache.org/jira/browse/HIVE-5576?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13807272#comment-13807272 ] Hudson commented on HIVE-5576: -- ABORTED: Integrated in Hive-trunk-hadoop2 #526 (See [https://builds.apache.org/job/Hive-trunk-hadoop2/526/]) HIVE-5576 : Blank lines missing from .q.out files created on Windows for testcase=TestCliDriver (Remus Rusanu via Ashutosh Chauhan) (hashutosh: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1536426) * /hive/trunk/ql/src/test/org/apache/hadoop/hive/ql/QTestUtil.java Blank lines missing from .q.out files created on Windows for testcase=TestCliDriver --- Key: HIVE-5576 URL: https://issues.apache.org/jira/browse/HIVE-5576 Project: Hive Issue Type: Bug Components: Testing Infrastructure Affects Versions: 0.13.0 Environment: Windows 8 using Hive Monarch build environment Reporter: Eric Hanson Assignee: Remus Rusanu Priority: Minor Fix For: 0.13.0 Attachments: HIVE-5576.1.patch, vectorized_math_funcs.q, vectorized_math_funcs.q.out.unix, vectorized_math_funcs.q.out.windows If you create a .q.out file on Windows using a command like this: ant test -Dhadoop.security.version=1.1.0-SNAPSHOT -Dhadoop.root=c:\hw\project\hadoop-monarch -Dresolvers=internal -Dhadoop-0.20S.version=1.1.0-SNAPSHOT -Dhadoop.mr.rev=20S -Dhive.support.concurrency=false -Dshims.include=0.20S -Dtest.continue.on.failure=true -Dtest.halt.on.failure=no -Dtest.print.classpath=true -Dtestcase=TestCliDriver -Dqfile=vectorized_math_funcs.q,vectorized_string_funcs.q,vectorized_casts.q -Doverwrite=true -Dtest.silent=false Then the .q.out files generated in the hive directory under ql\src\test\results\clientpositive having missing blank lines. So, the .q tests will pass on your Windows machine. But when you upload them in a patch, they fail on the automated build server. See HIVE-5517 for an example. HIVE-5517.3.patch has .q.out files with missing blank lines. Hive-5517.4.patch has .q.out files created on a Linux or Mac system. Those have blank lines. -- This message was sent by Atlassian JIRA (v6.1#6144)
[jira] [Commented] (HIVE-5667) ThriftCLIService log messages jumbled up
[ https://issues.apache.org/jira/browse/HIVE-5667?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13807275#comment-13807275 ] Hudson commented on HIVE-5667: -- ABORTED: Integrated in Hive-trunk-h0.21 #2426 (See [https://builds.apache.org/job/Hive-trunk-h0.21/2426/]) HIVE-5667 - ThriftCLIService log messages jumbled up (Vaibhav Gumashta via Brock Noland) (brock: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1536361) * /hive/trunk/service/src/java/org/apache/hive/service/cli/thrift/ThriftCLIService.java ThriftCLIService log messages jumbled up Key: HIVE-5667 URL: https://issues.apache.org/jira/browse/HIVE-5667 Project: Hive Issue Type: Bug Components: HiveServer2 Reporter: Vaibhav Gumashta Assignee: Vaibhav Gumashta Fix For: 0.13.0 Attachments: HIVE-5667.1.patch ThriftCLIService log messages are not aligned with the methods correctly -- This message was sent by Atlassian JIRA (v6.1#6144)
[jira] [Commented] (HIVE-5656) Hive produces unclear, confusing SemanticException when dealing with mod or pmod by zero
[ https://issues.apache.org/jira/browse/HIVE-5656?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13807270#comment-13807270 ] Hudson commented on HIVE-5656: -- ABORTED: Integrated in Hive-trunk-hadoop2 #526 (See [https://builds.apache.org/job/Hive-trunk-hadoop2/526/]) HIVE-5656 : Hive produces unclear, confusing SemanticException when dealing with mod or pmod by zero (Xuefu Zhang via Ashutosh Chauhan) (hashutosh: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1536476) * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/UDFOPMod.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/UDFPosMod.java * /hive/trunk/ql/src/test/org/apache/hadoop/hive/ql/udf/TestUDFOPMod.java * /hive/trunk/ql/src/test/org/apache/hadoop/hive/ql/udf/TestUDFPosMod.java * /hive/trunk/ql/src/test/results/clientpositive/vectorization_14.q.out Hive produces unclear, confusing SemanticException when dealing with mod or pmod by zero Key: HIVE-5656 URL: https://issues.apache.org/jira/browse/HIVE-5656 Project: Hive Issue Type: Bug Components: Types Affects Versions: 0.12.0 Reporter: Xuefu Zhang Assignee: Xuefu Zhang Fix For: 0.13.0 Attachments: HIVE-5656.patch {code} hive select 5%0 from tmp2 limit 1; FAILED: SemanticException [Error 10014]: Line 1:7 Wrong arguments '0': org.apache.hadoop.hive.ql.metadata.HiveException: Unable to execute method public org.apache.hadoop.io.IntWritable org.apache.hadoop.hive.ql.udf.UDFOPMod.evaluate(org.apache.hadoop.io.IntWritable,org.apache.hadoop.io.IntWritable) on object org.apache.hadoop.hive.ql.udf.UDFOPMod@21b594a9 of class org.apache.hadoop.hive.ql.udf.UDFOPMod with arguments {5:org.apache.hadoop.io.IntWritable, 0:org.apache.hadoop.io.IntWritable} of size 2 hive select pmod(5,0) from tmp2 limit 1; FAILED: SemanticException [Error 10014]: Line 1:7 Wrong arguments '0': org.apache.hadoop.hive.ql.metadata.HiveException: Unable to execute method public org.apache.hadoop.io.IntWritable org.apache.hadoop.hive.ql.udf.UDFPosMod.evaluate(org.apache.hadoop.io.IntWritable,org.apache.hadoop.io.IntWritable) on object org.apache.hadoop.hive.ql.udf.UDFPosMod@174ed99a of class org.apache.hadoop.hive.ql.udf.UDFPosMod with arguments {5:org.apache.hadoop.io.IntWritable, 0:org.apache.hadoop.io.IntWritable} of size 2 {code} Exception stack: {code} at org.apache.hadoop.hive.ql.parse.TypeCheckProcFactory$DefaultExprProcessor.process(TypeCheckProcFactory.java:1112) at org.apache.hadoop.hive.ql.lib.DefaultRuleDispatcher.dispatch(DefaultRuleDispatcher.java:90) at org.apache.hadoop.hive.ql.lib.DefaultGraphWalker.dispatchAndReturn(DefaultGraphWalker.java:94) at org.apache.hadoop.hive.ql.lib.DefaultGraphWalker.dispatch(DefaultGraphWalker.java:78) at org.apache.hadoop.hive.ql.lib.DefaultGraphWalker.walk(DefaultGraphWalker.java:132) at org.apache.hadoop.hive.ql.lib.DefaultGraphWalker.startWalking(DefaultGraphWalker.java:109) at org.apache.hadoop.hive.ql.parse.TypeCheckProcFactory.genExprNode(TypeCheckProcFactory.java:181) at org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genAllExprNodeDesc(SemanticAnalyzer.java:8870) at org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genExprNodeDesc(SemanticAnalyzer.java:8826) at org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genSelectPlan(SemanticAnalyzer.java:2734) at org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genSelectPlan(SemanticAnalyzer.java:2531) at org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genPostGroupByBodyPlan(SemanticAnalyzer.java:7606) at org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genBodyPlan(SemanticAnalyzer.java:7562) at org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genPlan(SemanticAnalyzer.java:8365) at org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.analyzeInternal(SemanticAnalyzer.java:8591) at org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer.analyze(BaseSemanticAnalyzer.java:284) at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:451) at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:351) at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1004) at org.apache.hadoop.hive.ql.Driver.run(Driver.java:915) at org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:268) at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:220) at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:422) at org.apache.hadoop.hive.cli.CliDriver.executeDriver(CliDriver.java:790) at
[jira] [Commented] (HIVE-5653) Vectorized Shuffle Join produces incorrect results
[ https://issues.apache.org/jira/browse/HIVE-5653?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13807405#comment-13807405 ] Hudson commented on HIVE-5653: -- SUCCESS: Integrated in Hive-trunk-hadoop1-ptest #217 (See [https://builds.apache.org/job/Hive-trunk-hadoop1-ptest/217/]) HIVE-5653 : Vectorized Shuffle Join produces incorrect results (Remus Rusanu via Ashutosh Chauhan) (hashutosh: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1536474) * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/VectorReduceSinkOperator.java * /hive/trunk/ql/src/test/queries/clientpositive/vectorized_shufflejoin.q * /hive/trunk/ql/src/test/results/clientpositive/vectorized_shufflejoin.q.out Vectorized Shuffle Join produces incorrect results -- Key: HIVE-5653 URL: https://issues.apache.org/jira/browse/HIVE-5653 Project: Hive Issue Type: Sub-task Affects Versions: 0.13.0 Reporter: Remus Rusanu Assignee: Remus Rusanu Fix For: 0.13.0 Attachments: HIVE-5653.1.patch Vectorized shuffle join should work out-of-the-box, but it produces empty result set. Investigating. -- This message was sent by Atlassian JIRA (v6.1#6144)
[jira] [Commented] (HIVE-5656) Hive produces unclear, confusing SemanticException when dealing with mod or pmod by zero
[ https://issues.apache.org/jira/browse/HIVE-5656?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13807404#comment-13807404 ] Hudson commented on HIVE-5656: -- SUCCESS: Integrated in Hive-trunk-hadoop1-ptest #217 (See [https://builds.apache.org/job/Hive-trunk-hadoop1-ptest/217/]) HIVE-5656 : Hive produces unclear, confusing SemanticException when dealing with mod or pmod by zero (Xuefu Zhang via Ashutosh Chauhan) (hashutosh: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1536476) * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/UDFOPMod.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/UDFPosMod.java * /hive/trunk/ql/src/test/org/apache/hadoop/hive/ql/udf/TestUDFOPMod.java * /hive/trunk/ql/src/test/org/apache/hadoop/hive/ql/udf/TestUDFPosMod.java * /hive/trunk/ql/src/test/results/clientpositive/vectorization_14.q.out Hive produces unclear, confusing SemanticException when dealing with mod or pmod by zero Key: HIVE-5656 URL: https://issues.apache.org/jira/browse/HIVE-5656 Project: Hive Issue Type: Bug Components: Types Affects Versions: 0.12.0 Reporter: Xuefu Zhang Assignee: Xuefu Zhang Fix For: 0.13.0 Attachments: HIVE-5656.patch {code} hive select 5%0 from tmp2 limit 1; FAILED: SemanticException [Error 10014]: Line 1:7 Wrong arguments '0': org.apache.hadoop.hive.ql.metadata.HiveException: Unable to execute method public org.apache.hadoop.io.IntWritable org.apache.hadoop.hive.ql.udf.UDFOPMod.evaluate(org.apache.hadoop.io.IntWritable,org.apache.hadoop.io.IntWritable) on object org.apache.hadoop.hive.ql.udf.UDFOPMod@21b594a9 of class org.apache.hadoop.hive.ql.udf.UDFOPMod with arguments {5:org.apache.hadoop.io.IntWritable, 0:org.apache.hadoop.io.IntWritable} of size 2 hive select pmod(5,0) from tmp2 limit 1; FAILED: SemanticException [Error 10014]: Line 1:7 Wrong arguments '0': org.apache.hadoop.hive.ql.metadata.HiveException: Unable to execute method public org.apache.hadoop.io.IntWritable org.apache.hadoop.hive.ql.udf.UDFPosMod.evaluate(org.apache.hadoop.io.IntWritable,org.apache.hadoop.io.IntWritable) on object org.apache.hadoop.hive.ql.udf.UDFPosMod@174ed99a of class org.apache.hadoop.hive.ql.udf.UDFPosMod with arguments {5:org.apache.hadoop.io.IntWritable, 0:org.apache.hadoop.io.IntWritable} of size 2 {code} Exception stack: {code} at org.apache.hadoop.hive.ql.parse.TypeCheckProcFactory$DefaultExprProcessor.process(TypeCheckProcFactory.java:1112) at org.apache.hadoop.hive.ql.lib.DefaultRuleDispatcher.dispatch(DefaultRuleDispatcher.java:90) at org.apache.hadoop.hive.ql.lib.DefaultGraphWalker.dispatchAndReturn(DefaultGraphWalker.java:94) at org.apache.hadoop.hive.ql.lib.DefaultGraphWalker.dispatch(DefaultGraphWalker.java:78) at org.apache.hadoop.hive.ql.lib.DefaultGraphWalker.walk(DefaultGraphWalker.java:132) at org.apache.hadoop.hive.ql.lib.DefaultGraphWalker.startWalking(DefaultGraphWalker.java:109) at org.apache.hadoop.hive.ql.parse.TypeCheckProcFactory.genExprNode(TypeCheckProcFactory.java:181) at org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genAllExprNodeDesc(SemanticAnalyzer.java:8870) at org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genExprNodeDesc(SemanticAnalyzer.java:8826) at org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genSelectPlan(SemanticAnalyzer.java:2734) at org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genSelectPlan(SemanticAnalyzer.java:2531) at org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genPostGroupByBodyPlan(SemanticAnalyzer.java:7606) at org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genBodyPlan(SemanticAnalyzer.java:7562) at org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genPlan(SemanticAnalyzer.java:8365) at org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.analyzeInternal(SemanticAnalyzer.java:8591) at org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer.analyze(BaseSemanticAnalyzer.java:284) at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:451) at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:351) at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1004) at org.apache.hadoop.hive.ql.Driver.run(Driver.java:915) at org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:268) at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:220) at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:422) at org.apache.hadoop.hive.cli.CliDriver.executeDriver(CliDriver.java:790) at
[jira] [Commented] (HIVE-5648) error when casting partition column to varchar in where clause
[ https://issues.apache.org/jira/browse/HIVE-5648?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13807407#comment-13807407 ] Hudson commented on HIVE-5648: -- SUCCESS: Integrated in Hive-trunk-hadoop1-ptest #217 (See [https://builds.apache.org/job/Hive-trunk-hadoop1-ptest/217/]) HIVE-5648 : error when casting partition column to varchar in where clause (Jason Dere via Ashutosh Chauhan) (hashutosh: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1536471) * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/ExprNodeConstantDesc.java * /hive/trunk/ql/src/test/queries/clientpositive/partition_varchar2.q * /hive/trunk/ql/src/test/results/clientpositive/partition_varchar2.q.out * /hive/trunk/serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/primitive/PrimitiveObjectInspectorFactory.java * /hive/trunk/serde/src/java/org/apache/hadoop/hive/serde2/typeinfo/TypeInfoFactory.java * /hive/trunk/serde/src/test/org/apache/hadoop/hive/serde2/objectinspector/primitive/TestPrimitiveObjectInspectorFactory.java error when casting partition column to varchar in where clause --- Key: HIVE-5648 URL: https://issues.apache.org/jira/browse/HIVE-5648 Project: Hive Issue Type: Bug Reporter: Jason Dere Assignee: Jason Dere Fix For: 0.13.0 Attachments: HIVE-5648.1.patch, HIVE-5648.2.patch hive select * from partition_varchar_2 where cast(dt as varchar(10)) = '2000-01-01'; FAILED: RuntimeException org.apache.hadoop.hive.ql.metadata.HiveException: java.lang.RuntimeException: Internal error: Cannot find ObjectInspector for VARCHAR -- This message was sent by Atlassian JIRA (v6.1#6144)
[jira] [Commented] (HIVE-5666) use Path instead of String for IOContext.inputPath
[ https://issues.apache.org/jira/browse/HIVE-5666?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13807402#comment-13807402 ] Hudson commented on HIVE-5666: -- SUCCESS: Integrated in Hive-trunk-hadoop1-ptest #217 (See [https://builds.apache.org/job/Hive-trunk-hadoop1-ptest/217/]) HIVE-5666 : use Path instead of String for IOContext.inputPath (Thejas Nair via Ashutosh Chauhan) (hashutosh: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1536478) * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/MapJoinOperator.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/MapOperator.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/SMBMapJoinOperator.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/TableScanOperator.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/mr/ExecMapperContext.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/io/HiveContextAwareRecordReader.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/io/IOContext.java use Path instead of String for IOContext.inputPath -- Key: HIVE-5666 URL: https://issues.apache.org/jira/browse/HIVE-5666 Project: Hive Issue Type: Improvement Reporter: Thejas M Nair Assignee: Thejas M Nair Fix For: 0.13.0 Attachments: HIVE-5666.1.patch Path is converted to string in HiveContextAwareRecordReader to be stored in IOContext.inputPath, then in MapOperator normalizePath gets called on it which converts it back to Path. Path creation is expensive, so it is better to use Path instead of string through the call stack. This is also a step towards HIVE-3616. -- This message was sent by Atlassian JIRA (v6.1#6144)
[jira] [Commented] (HIVE-5576) Blank lines missing from .q.out files created on Windows for testcase=TestCliDriver
[ https://issues.apache.org/jira/browse/HIVE-5576?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13807406#comment-13807406 ] Hudson commented on HIVE-5576: -- SUCCESS: Integrated in Hive-trunk-hadoop1-ptest #217 (See [https://builds.apache.org/job/Hive-trunk-hadoop1-ptest/217/]) HIVE-5576 : Blank lines missing from .q.out files created on Windows for testcase=TestCliDriver (Remus Rusanu via Ashutosh Chauhan) (hashutosh: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1536426) * /hive/trunk/ql/src/test/org/apache/hadoop/hive/ql/QTestUtil.java Blank lines missing from .q.out files created on Windows for testcase=TestCliDriver --- Key: HIVE-5576 URL: https://issues.apache.org/jira/browse/HIVE-5576 Project: Hive Issue Type: Bug Components: Testing Infrastructure Affects Versions: 0.13.0 Environment: Windows 8 using Hive Monarch build environment Reporter: Eric Hanson Assignee: Remus Rusanu Priority: Minor Fix For: 0.13.0 Attachments: HIVE-5576.1.patch, vectorized_math_funcs.q, vectorized_math_funcs.q.out.unix, vectorized_math_funcs.q.out.windows If you create a .q.out file on Windows using a command like this: ant test -Dhadoop.security.version=1.1.0-SNAPSHOT -Dhadoop.root=c:\hw\project\hadoop-monarch -Dresolvers=internal -Dhadoop-0.20S.version=1.1.0-SNAPSHOT -Dhadoop.mr.rev=20S -Dhive.support.concurrency=false -Dshims.include=0.20S -Dtest.continue.on.failure=true -Dtest.halt.on.failure=no -Dtest.print.classpath=true -Dtestcase=TestCliDriver -Dqfile=vectorized_math_funcs.q,vectorized_string_funcs.q,vectorized_casts.q -Doverwrite=true -Dtest.silent=false Then the .q.out files generated in the hive directory under ql\src\test\results\clientpositive having missing blank lines. So, the .q tests will pass on your Windows machine. But when you upload them in a patch, they fail on the automated build server. See HIVE-5517 for an example. HIVE-5517.3.patch has .q.out files with missing blank lines. Hive-5517.4.patch has .q.out files created on a Linux or Mac system. Those have blank lines. -- This message was sent by Atlassian JIRA (v6.1#6144)
[jira] [Commented] (HIVE-5667) ThriftCLIService log messages jumbled up
[ https://issues.apache.org/jira/browse/HIVE-5667?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13807403#comment-13807403 ] Hudson commented on HIVE-5667: -- SUCCESS: Integrated in Hive-trunk-hadoop1-ptest #217 (See [https://builds.apache.org/job/Hive-trunk-hadoop1-ptest/217/]) HIVE-5667 - ThriftCLIService log messages jumbled up (Vaibhav Gumashta via Brock Noland) (brock: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1536361) * /hive/trunk/service/src/java/org/apache/hive/service/cli/thrift/ThriftCLIService.java ThriftCLIService log messages jumbled up Key: HIVE-5667 URL: https://issues.apache.org/jira/browse/HIVE-5667 Project: Hive Issue Type: Bug Components: HiveServer2 Reporter: Vaibhav Gumashta Assignee: Vaibhav Gumashta Fix For: 0.13.0 Attachments: HIVE-5667.1.patch ThriftCLIService log messages are not aligned with the methods correctly -- This message was sent by Atlassian JIRA (v6.1#6144)
[jira] [Commented] (HIVE-5653) Vectorized Shuffle Join produces incorrect results
[ https://issues.apache.org/jira/browse/HIVE-5653?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13807432#comment-13807432 ] Hudson commented on HIVE-5653: -- ABORTED: Integrated in Hive-trunk-h0.21 #2427 (See [https://builds.apache.org/job/Hive-trunk-h0.21/2427/]) HIVE-5653 : Vectorized Shuffle Join produces incorrect results (Remus Rusanu via Ashutosh Chauhan) (hashutosh: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1536474) * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/VectorReduceSinkOperator.java * /hive/trunk/ql/src/test/queries/clientpositive/vectorized_shufflejoin.q * /hive/trunk/ql/src/test/results/clientpositive/vectorized_shufflejoin.q.out Vectorized Shuffle Join produces incorrect results -- Key: HIVE-5653 URL: https://issues.apache.org/jira/browse/HIVE-5653 Project: Hive Issue Type: Sub-task Affects Versions: 0.13.0 Reporter: Remus Rusanu Assignee: Remus Rusanu Fix For: 0.13.0 Attachments: HIVE-5653.1.patch Vectorized shuffle join should work out-of-the-box, but it produces empty result set. Investigating. -- This message was sent by Atlassian JIRA (v6.1#6144)
[jira] [Commented] (HIVE-5656) Hive produces unclear, confusing SemanticException when dealing with mod or pmod by zero
[ https://issues.apache.org/jira/browse/HIVE-5656?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13807431#comment-13807431 ] Hudson commented on HIVE-5656: -- ABORTED: Integrated in Hive-trunk-h0.21 #2427 (See [https://builds.apache.org/job/Hive-trunk-h0.21/2427/]) HIVE-5656 : Hive produces unclear, confusing SemanticException when dealing with mod or pmod by zero (Xuefu Zhang via Ashutosh Chauhan) (hashutosh: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1536476) * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/UDFOPMod.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/UDFPosMod.java * /hive/trunk/ql/src/test/org/apache/hadoop/hive/ql/udf/TestUDFOPMod.java * /hive/trunk/ql/src/test/org/apache/hadoop/hive/ql/udf/TestUDFPosMod.java * /hive/trunk/ql/src/test/results/clientpositive/vectorization_14.q.out Hive produces unclear, confusing SemanticException when dealing with mod or pmod by zero Key: HIVE-5656 URL: https://issues.apache.org/jira/browse/HIVE-5656 Project: Hive Issue Type: Bug Components: Types Affects Versions: 0.12.0 Reporter: Xuefu Zhang Assignee: Xuefu Zhang Fix For: 0.13.0 Attachments: HIVE-5656.patch {code} hive select 5%0 from tmp2 limit 1; FAILED: SemanticException [Error 10014]: Line 1:7 Wrong arguments '0': org.apache.hadoop.hive.ql.metadata.HiveException: Unable to execute method public org.apache.hadoop.io.IntWritable org.apache.hadoop.hive.ql.udf.UDFOPMod.evaluate(org.apache.hadoop.io.IntWritable,org.apache.hadoop.io.IntWritable) on object org.apache.hadoop.hive.ql.udf.UDFOPMod@21b594a9 of class org.apache.hadoop.hive.ql.udf.UDFOPMod with arguments {5:org.apache.hadoop.io.IntWritable, 0:org.apache.hadoop.io.IntWritable} of size 2 hive select pmod(5,0) from tmp2 limit 1; FAILED: SemanticException [Error 10014]: Line 1:7 Wrong arguments '0': org.apache.hadoop.hive.ql.metadata.HiveException: Unable to execute method public org.apache.hadoop.io.IntWritable org.apache.hadoop.hive.ql.udf.UDFPosMod.evaluate(org.apache.hadoop.io.IntWritable,org.apache.hadoop.io.IntWritable) on object org.apache.hadoop.hive.ql.udf.UDFPosMod@174ed99a of class org.apache.hadoop.hive.ql.udf.UDFPosMod with arguments {5:org.apache.hadoop.io.IntWritable, 0:org.apache.hadoop.io.IntWritable} of size 2 {code} Exception stack: {code} at org.apache.hadoop.hive.ql.parse.TypeCheckProcFactory$DefaultExprProcessor.process(TypeCheckProcFactory.java:1112) at org.apache.hadoop.hive.ql.lib.DefaultRuleDispatcher.dispatch(DefaultRuleDispatcher.java:90) at org.apache.hadoop.hive.ql.lib.DefaultGraphWalker.dispatchAndReturn(DefaultGraphWalker.java:94) at org.apache.hadoop.hive.ql.lib.DefaultGraphWalker.dispatch(DefaultGraphWalker.java:78) at org.apache.hadoop.hive.ql.lib.DefaultGraphWalker.walk(DefaultGraphWalker.java:132) at org.apache.hadoop.hive.ql.lib.DefaultGraphWalker.startWalking(DefaultGraphWalker.java:109) at org.apache.hadoop.hive.ql.parse.TypeCheckProcFactory.genExprNode(TypeCheckProcFactory.java:181) at org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genAllExprNodeDesc(SemanticAnalyzer.java:8870) at org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genExprNodeDesc(SemanticAnalyzer.java:8826) at org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genSelectPlan(SemanticAnalyzer.java:2734) at org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genSelectPlan(SemanticAnalyzer.java:2531) at org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genPostGroupByBodyPlan(SemanticAnalyzer.java:7606) at org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genBodyPlan(SemanticAnalyzer.java:7562) at org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genPlan(SemanticAnalyzer.java:8365) at org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.analyzeInternal(SemanticAnalyzer.java:8591) at org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer.analyze(BaseSemanticAnalyzer.java:284) at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:451) at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:351) at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1004) at org.apache.hadoop.hive.ql.Driver.run(Driver.java:915) at org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:268) at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:220) at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:422) at org.apache.hadoop.hive.cli.CliDriver.executeDriver(CliDriver.java:790) at
[jira] [Commented] (HIVE-5648) error when casting partition column to varchar in where clause
[ https://issues.apache.org/jira/browse/HIVE-5648?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13807434#comment-13807434 ] Hudson commented on HIVE-5648: -- ABORTED: Integrated in Hive-trunk-h0.21 #2427 (See [https://builds.apache.org/job/Hive-trunk-h0.21/2427/]) HIVE-5648 : error when casting partition column to varchar in where clause (Jason Dere via Ashutosh Chauhan) (hashutosh: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1536471) * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/ExprNodeConstantDesc.java * /hive/trunk/ql/src/test/queries/clientpositive/partition_varchar2.q * /hive/trunk/ql/src/test/results/clientpositive/partition_varchar2.q.out * /hive/trunk/serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/primitive/PrimitiveObjectInspectorFactory.java * /hive/trunk/serde/src/java/org/apache/hadoop/hive/serde2/typeinfo/TypeInfoFactory.java * /hive/trunk/serde/src/test/org/apache/hadoop/hive/serde2/objectinspector/primitive/TestPrimitiveObjectInspectorFactory.java error when casting partition column to varchar in where clause --- Key: HIVE-5648 URL: https://issues.apache.org/jira/browse/HIVE-5648 Project: Hive Issue Type: Bug Reporter: Jason Dere Assignee: Jason Dere Fix For: 0.13.0 Attachments: HIVE-5648.1.patch, HIVE-5648.2.patch hive select * from partition_varchar_2 where cast(dt as varchar(10)) = '2000-01-01'; FAILED: RuntimeException org.apache.hadoop.hive.ql.metadata.HiveException: java.lang.RuntimeException: Internal error: Cannot find ObjectInspector for VARCHAR -- This message was sent by Atlassian JIRA (v6.1#6144)
[jira] [Commented] (HIVE-5295) HiveConnection#configureConnection tries to execute statement even after it is closed
[ https://issues.apache.org/jira/browse/HIVE-5295?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13807429#comment-13807429 ] Hudson commented on HIVE-5295: -- ABORTED: Integrated in Hive-trunk-h0.21 #2427 (See [https://builds.apache.org/job/Hive-trunk-h0.21/2427/]) HIVE-5295 : HiveConnection#configureConnection tries to execute statement even after it is closed (Vaibhav Gumashta via Thejas Nair) (thejas: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1536533) * /hive/trunk/jdbc/src/test/org/apache/hive/jdbc/TestJdbcDriver2.java HiveConnection#configureConnection tries to execute statement even after it is closed - Key: HIVE-5295 URL: https://issues.apache.org/jira/browse/HIVE-5295 Project: Hive Issue Type: Bug Components: HiveServer2 Reporter: Vaibhav Gumashta Assignee: Vaibhav Gumashta Fix For: 0.13.0 Attachments: D12957.1.patch, D12957.2.patch, D12957.3.patch, HIVE-5295.5.patch, HIVE-5295.D12957.3.patch, HIVE-5295.D12957.3.patch, HIVE-5295.D12957.4.patch HiveConnection#configureConnection tries to execute statement even after it is closed. For remote JDBC client, it tries to set the conf var using 'set foo=bar' by calling HiveStatement.execute for each conf var pair, but closes the statement after the 1st iteration through the conf var pairs. -- This message was sent by Atlassian JIRA (v6.1#6144)
[jira] [Commented] (HIVE-5576) Blank lines missing from .q.out files created on Windows for testcase=TestCliDriver
[ https://issues.apache.org/jira/browse/HIVE-5576?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13807433#comment-13807433 ] Hudson commented on HIVE-5576: -- ABORTED: Integrated in Hive-trunk-h0.21 #2427 (See [https://builds.apache.org/job/Hive-trunk-h0.21/2427/]) HIVE-5576 : Blank lines missing from .q.out files created on Windows for testcase=TestCliDriver (Remus Rusanu via Ashutosh Chauhan) (hashutosh: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1536426) * /hive/trunk/ql/src/test/org/apache/hadoop/hive/ql/QTestUtil.java Blank lines missing from .q.out files created on Windows for testcase=TestCliDriver --- Key: HIVE-5576 URL: https://issues.apache.org/jira/browse/HIVE-5576 Project: Hive Issue Type: Bug Components: Testing Infrastructure Affects Versions: 0.13.0 Environment: Windows 8 using Hive Monarch build environment Reporter: Eric Hanson Assignee: Remus Rusanu Priority: Minor Fix For: 0.13.0 Attachments: HIVE-5576.1.patch, vectorized_math_funcs.q, vectorized_math_funcs.q.out.unix, vectorized_math_funcs.q.out.windows If you create a .q.out file on Windows using a command like this: ant test -Dhadoop.security.version=1.1.0-SNAPSHOT -Dhadoop.root=c:\hw\project\hadoop-monarch -Dresolvers=internal -Dhadoop-0.20S.version=1.1.0-SNAPSHOT -Dhadoop.mr.rev=20S -Dhive.support.concurrency=false -Dshims.include=0.20S -Dtest.continue.on.failure=true -Dtest.halt.on.failure=no -Dtest.print.classpath=true -Dtestcase=TestCliDriver -Dqfile=vectorized_math_funcs.q,vectorized_string_funcs.q,vectorized_casts.q -Doverwrite=true -Dtest.silent=false Then the .q.out files generated in the hive directory under ql\src\test\results\clientpositive having missing blank lines. So, the .q tests will pass on your Windows machine. But when you upload them in a patch, they fail on the automated build server. See HIVE-5517 for an example. HIVE-5517.3.patch has .q.out files with missing blank lines. Hive-5517.4.patch has .q.out files created on a Linux or Mac system. Those have blank lines. -- This message was sent by Atlassian JIRA (v6.1#6144)
[jira] [Commented] (HIVE-5666) use Path instead of String for IOContext.inputPath
[ https://issues.apache.org/jira/browse/HIVE-5666?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13807428#comment-13807428 ] Hudson commented on HIVE-5666: -- ABORTED: Integrated in Hive-trunk-h0.21 #2427 (See [https://builds.apache.org/job/Hive-trunk-h0.21/2427/]) HIVE-5666 : use Path instead of String for IOContext.inputPath (Thejas Nair via Ashutosh Chauhan) (hashutosh: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1536478) * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/MapJoinOperator.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/MapOperator.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/SMBMapJoinOperator.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/TableScanOperator.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/mr/ExecMapperContext.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/io/HiveContextAwareRecordReader.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/io/IOContext.java use Path instead of String for IOContext.inputPath -- Key: HIVE-5666 URL: https://issues.apache.org/jira/browse/HIVE-5666 Project: Hive Issue Type: Improvement Reporter: Thejas M Nair Assignee: Thejas M Nair Fix For: 0.13.0 Attachments: HIVE-5666.1.patch Path is converted to string in HiveContextAwareRecordReader to be stored in IOContext.inputPath, then in MapOperator normalizePath gets called on it which converts it back to Path. Path creation is expensive, so it is better to use Path instead of string through the call stack. This is also a step towards HIVE-3616. -- This message was sent by Atlassian JIRA (v6.1#6144)
[jira] [Commented] (HIVE-5450) pTest2 TestReportParser is failing due to .svn directory
[ https://issues.apache.org/jira/browse/HIVE-5450?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13807430#comment-13807430 ] Hudson commented on HIVE-5450: -- ABORTED: Integrated in Hive-trunk-h0.21 #2427 (See [https://builds.apache.org/job/Hive-trunk-h0.21/2427/]) HIVE-5450 - pTest2 TestReportParser is failing due to .svn directory (Brock Noland reviewed by Ashutosh Chauhan) (brock: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1536508) * /hive/trunk/testutils/ptest2/src/test/java/org/apache/hive/ptest/execution/TestReportParser.java pTest2 TestReportParser is failing due to .svn directory Key: HIVE-5450 URL: https://issues.apache.org/jira/browse/HIVE-5450 Project: Hive Issue Type: Task Components: Testing Infrastructure Affects Versions: 0.13.0 Reporter: Ashutosh Chauhan Assignee: Brock Noland Fix For: 0.13.0 Attachments: HIVE-5450.patch Following test failed when I ran mvn package: {code} Running org.apache.hive.ptest.execution.TestPhase 2013-10-04 22:57:20,150 ERROR HostExecutor$5.call:379 Aborting drone during exec echo org.apache.hive.ptest.execution.AbortDroneException: Drone Drone [user=someuser, host=somehost, instance=0] exited with 255: SSHResult [command=echo, getExitCode()=255, getException()=null, getUser()=someuser, getHost()=somehost, getInstance()=0] at org.apache.hive.ptest.execution.HostExecutor$5.call(HostExecutor.java:379) at org.apache.hive.ptest.execution.HostExecutor$5.call(HostExecutor.java:368) at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) at java.util.concurrent.FutureTask.run(FutureTask.java:138) at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) at java.lang.Thread.run(Thread.java:662) Tests in error: test(org.apache.hive.ptest.execution.TestReportParser): src/test/resources/test-outputs/.svn (Is a directory) Tests run: 44, Failures: 0, Errors: 1, Skipped: 0 {code} NO PRECOMMIT TESTS -- This message was sent by Atlassian JIRA (v6.1#6144)
[jira] [Commented] (HIVE-5511) percentComplete returned by job status from WebHCat is null
[ https://issues.apache.org/jira/browse/HIVE-5511?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13806249#comment-13806249 ] Hudson commented on HIVE-5511: -- FAILURE: Integrated in Hive-trunk-hadoop2-ptest #152 (See [https://builds.apache.org/job/Hive-trunk-hadoop2-ptest/152/]) HIVE-5511 : percentComplete returned by job status from WebHCat is null (Eugene Koifman via Thejas Nair) (thejas: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1535796) * /hive/trunk/hcatalog/src/test/e2e/templeton/README.txt * /hive/trunk/hcatalog/src/test/e2e/templeton/drivers/TestDriverCurl.pm * /hive/trunk/hcatalog/src/test/e2e/templeton/tests/jobsubmission.conf * /hive/trunk/hcatalog/src/test/e2e/templeton/tests/jobsubmission_streaming.conf * /hive/trunk/hcatalog/webhcat/svr/src/main/bin/webhcat_config.sh * /hive/trunk/hcatalog/webhcat/svr/src/main/config/override-container-log4j.properties * /hive/trunk/hcatalog/webhcat/svr/src/main/java/org/apache/hive/hcatalog/templeton/AppConfig.java * /hive/trunk/hcatalog/webhcat/svr/src/main/java/org/apache/hive/hcatalog/templeton/CompleteDelegator.java * /hive/trunk/hcatalog/webhcat/svr/src/main/java/org/apache/hive/hcatalog/templeton/tool/HDFSStorage.java * /hive/trunk/hcatalog/webhcat/svr/src/main/java/org/apache/hive/hcatalog/templeton/tool/HiveJobIDParser.java * /hive/trunk/hcatalog/webhcat/svr/src/main/java/org/apache/hive/hcatalog/templeton/tool/JarJobIDParser.java * /hive/trunk/hcatalog/webhcat/svr/src/main/java/org/apache/hive/hcatalog/templeton/tool/JobSubmissionConstants.java * /hive/trunk/hcatalog/webhcat/svr/src/main/java/org/apache/hive/hcatalog/templeton/tool/LaunchMapper.java * /hive/trunk/hcatalog/webhcat/svr/src/main/java/org/apache/hive/hcatalog/templeton/tool/PigJobIDParser.java * /hive/trunk/hcatalog/webhcat/svr/src/main/java/org/apache/hive/hcatalog/templeton/tool/TempletonControllerJob.java * /hive/trunk/hcatalog/webhcat/svr/src/main/java/org/apache/hive/hcatalog/templeton/tool/TempletonUtils.java * /hive/trunk/hcatalog/webhcat/svr/src/main/java/org/apache/hive/hcatalog/templeton/tool/TrivialExecService.java * /hive/trunk/hcatalog/webhcat/svr/src/test/java/org/apache/hive/hcatalog/templeton/tool/TestTrivialExecService.java * /hive/trunk/shims/src/0.20S/java/org/apache/hadoop/mapred/WebHCatJTShim20S.java * /hive/trunk/shims/src/0.23/java/org/apache/hadoop/mapred/WebHCatJTShim23.java * /hive/trunk/shims/src/common/java/org/apache/hadoop/hive/shims/HadoopShims.java percentComplete returned by job status from WebHCat is null --- Key: HIVE-5511 URL: https://issues.apache.org/jira/browse/HIVE-5511 Project: Hive Issue Type: Bug Components: WebHCat Affects Versions: 0.12.0 Reporter: Eugene Koifman Assignee: Eugene Koifman Fix For: 0.13.0 Attachments: HIVE-5511.3.patch, HIVE-5511.5.patch In hadoop1 the logging from MR is sent to stderr. In H2, by default, to syslog. templeton.tool.LaunchMapper expects to see the output on stderr to produce 'percentComplete' in job status. -- This message was sent by Atlassian JIRA (v6.1#6144)
[jira] [Commented] (HIVE-5629) Fix two javadoc failures in HCatalog
[ https://issues.apache.org/jira/browse/HIVE-5629?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13806252#comment-13806252 ] Hudson commented on HIVE-5629: -- FAILURE: Integrated in Hive-trunk-hadoop2-ptest #152 (See [https://builds.apache.org/job/Hive-trunk-hadoop2-ptest/152/]) HIVE-5629 : Fix two javadoc failures in HCatalog (Brock Noland via Ashutosh Chauhan) (hashutosh: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1535513) * /hive/trunk/hcatalog/core/src/main/java/org/apache/hive/hcatalog/mapreduce/InputJobInfo.java * /hive/trunk/hcatalog/webhcat/svr/src/main/java/org/apache/hive/hcatalog/templeton/PigDelegator.java Fix two javadoc failures in HCatalog Key: HIVE-5629 URL: https://issues.apache.org/jira/browse/HIVE-5629 Project: Hive Issue Type: Bug Reporter: Brock Noland Assignee: Brock Noland Fix For: 0.13.0 Attachments: HIVE-5629.patch I am seeing two javadoc failures on HCatalog. These are not being seen by PTest and indeed I cannot reproduce on my Mac but can on Linux. Regardless they should be fixed. -- This message was sent by Atlassian JIRA (v6.1#6144)
[jira] [Commented] (HIVE-5403) Move loading of filesystem, ugi, metastore client to hive session
[ https://issues.apache.org/jira/browse/HIVE-5403?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13806263#comment-13806263 ] Hudson commented on HIVE-5403: -- FAILURE: Integrated in Hive-trunk-hadoop2-ptest #152 (See [https://builds.apache.org/job/Hive-trunk-hadoop2-ptest/152/]) HIVE-5403: Perflogger broken due to HIVE-5403 (Vikram Dixit K via Gunther Hagleitner) (gunther: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1535598) * /hive/trunk/metastore/src/test/org/apache/hadoop/hive/metastore/TestMetastoreVersion.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/session/SessionState.java HIVE-5403 : Move loading of filesystem, ugi, metastore client to hive session (Vikram Dixit via Ashutosh Chauhan) (hashutosh: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1535039) * /hive/trunk/metastore/src/test/org/apache/hadoop/hive/metastore/TestMetastoreVersion.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/metadata/Hive.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/session/SessionState.java Move loading of filesystem, ugi, metastore client to hive session - Key: HIVE-5403 URL: https://issues.apache.org/jira/browse/HIVE-5403 Project: Hive Issue Type: Bug Affects Versions: 0.13.0 Reporter: Vikram Dixit K Assignee: Vikram Dixit K Fix For: 0.13.0 Attachments: HIVE-5403.1.patch, HIVE-5403.2.patch, HIVE-5403.3.patch, HIVE-5403.4.patch As part of HIVE-5184, the metastore connection, loading filesystem were done as part of the tez session so as to speed up query times while paying a cost at startup. We can do this more generally in hive to apply to both the mapreduce and tez side of things. -- This message was sent by Atlassian JIRA (v6.1#6144)
[jira] [Commented] (HIVE-5482) JDBC should depend on httpclient.version and httpcore.version 4.1.3 to be consistent with other modules
[ https://issues.apache.org/jira/browse/HIVE-5482?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13806256#comment-13806256 ] Hudson commented on HIVE-5482: -- FAILURE: Integrated in Hive-trunk-hadoop2-ptest #152 (See [https://builds.apache.org/job/Hive-trunk-hadoop2-ptest/152/]) HIVE-5482 : JDBC should depend on httpclient.version and httpcore.version 4.1.3 to be consistent with other modules (Vaibhav Gumashta via Thejas Nair) (thejas: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1535620) * /hive/trunk/ivy/libraries.properties JDBC should depend on httpclient.version and httpcore.version 4.1.3 to be consistent with other modules --- Key: HIVE-5482 URL: https://issues.apache.org/jira/browse/HIVE-5482 Project: Hive Issue Type: Bug Components: HiveServer2, JDBC Reporter: Vaibhav Gumashta Assignee: Vaibhav Gumashta Fix For: 0.13.0 Attachments: HIVE-5482.1.patch Currently depends on 4.2.4 and 4.2.5 which conflicts with thrift-0.9 which depends on 4.1.3 -- This message was sent by Atlassian JIRA (v6.1#6144)
[jira] [Commented] (HIVE-5625) Fix issue with metastore version restriction test.
[ https://issues.apache.org/jira/browse/HIVE-5625?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13806254#comment-13806254 ] Hudson commented on HIVE-5625: -- FAILURE: Integrated in Hive-trunk-hadoop2-ptest #152 (See [https://builds.apache.org/job/Hive-trunk-hadoop2-ptest/152/]) HIVE-5625 - Fix issue with metastore version restriction test. (Vikram Dixit K via Brock Noland) (brock: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1535402) * /hive/trunk/metastore/src/test/org/apache/hadoop/hive/metastore/TestMetastoreVersion.java Fix issue with metastore version restriction test. -- Key: HIVE-5625 URL: https://issues.apache.org/jira/browse/HIVE-5625 Project: Hive Issue Type: Bug Components: Tests Affects Versions: 0.13.0 Reporter: Vikram Dixit K Assignee: Vikram Dixit K Fix For: 0.13.0 Attachments: HIVE-5625.1.patch Based on Brock's comments, the change made in HIVE-5403 change the nature of the test. -- This message was sent by Atlassian JIRA (v6.1#6144)
[jira] [Commented] (HIVE-5637) Sporadic minimr test failure
[ https://issues.apache.org/jira/browse/HIVE-5637?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13806266#comment-13806266 ] Hudson commented on HIVE-5637: -- FAILURE: Integrated in Hive-trunk-hadoop2-ptest #152 (See [https://builds.apache.org/job/Hive-trunk-hadoop2-ptest/152/]) HIVE-5637 : Sporadic minimr test failure (Navis via Ashutosh Chauhan) (hashutosh: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1535510) * /hive/trunk/ql/src/test/queries/clientpositive/import_exported_table.q Sporadic minimr test failure Key: HIVE-5637 URL: https://issues.apache.org/jira/browse/HIVE-5637 Project: Hive Issue Type: Test Components: Tests Reporter: Navis Assignee: Navis Priority: Trivial Fix For: 0.13.0 Attachments: HIVE-5637.1.patch.txt {noformat} ant test -Dtestcase=TestMinimrCliDriver -Dqfile=import_exported_table.q,load_hdfs_file_with_space_in_the_name.q -Dmodule=ql {noformat} Fails with message like this. {noformat} Begin query: load_hdfs_file_with_space_in_the_name.q mkdir: cannot create directory hdfs:///tmp/test/: File exists Exception: Client Execution failed with error code = -1 See build/ql/tmp/hive.log, or try ant test ... -Dtest.silent=false to get more logs. junit.framework.AssertionFailedError: Client Execution failed with error code = -1 See build/ql/tmp/hive.log, or try ant test ... -Dtest.silent=false to get more logs. at junit.framework.Assert.fail(Assert.java:47) at org.apache.hadoop.hive.cli.TestMinimrCliDriver.runTest(TestMinimrCliDriver.java:349) at org.apache.hadoop.hive.cli.TestMinimrCliDriver.testCliDriver_load_hdfs_file_with_space_in_the_name(TestMinimrCliDriver.java:291) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25) at java.lang.reflect.Method.invoke(Method.java:597) at junit.framework.TestCase.runTest(TestCase.java:154) at junit.framework.TestCase.runBare(TestCase.java:127) at junit.framework.TestResult$1.protect(TestResult.java:106) at junit.framework.TestResult.runProtected(TestResult.java:124) at junit.framework.TestResult.run(TestResult.java:109) at junit.framework.TestCase.run(TestCase.java:118) at junit.framework.TestSuite.runTest(TestSuite.java:208) at junit.framework.TestSuite.run(TestSuite.java:203) at org.apache.tools.ant.taskdefs.optional.junit.JUnitTestRunner.run(JUnitTestRunner.java:520) at org.apache.tools.ant.taskdefs.optional.junit.JUnitTestRunner.launch(JUnitTestRunner.java:1060) at org.apache.tools.ant.taskdefs.optional.junit.JUnitTestRunner.main(JUnitTestRunner.java:911) {noformat} -- This message was sent by Atlassian JIRA (v6.1#6144)
[jira] [Commented] (HIVE-5599) Change default logging level to INFO
[ https://issues.apache.org/jira/browse/HIVE-5599?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13806260#comment-13806260 ] Hudson commented on HIVE-5599: -- FAILURE: Integrated in Hive-trunk-hadoop2-ptest #152 (See [https://builds.apache.org/job/Hive-trunk-hadoop2-ptest/152/]) HIVE-5599 - Change default logging level to INFO (Brock Noland, Reviewed by Thejas M Nair) (brock: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1535047) * /hive/trunk/common/src/java/conf/hive-log4j.properties Change default logging level to INFO Key: HIVE-5599 URL: https://issues.apache.org/jira/browse/HIVE-5599 Project: Hive Issue Type: Bug Affects Versions: 0.12.0 Reporter: Brock Noland Assignee: Brock Noland Fix For: 0.13.0 Attachments: HIVE-5599.patch The default logging level is warn: https://github.com/apache/hive/blob/trunk/common/src/java/conf/hive-log4j.properties#L19 but hive logs lot's of good information at INFO level. Additionally most hadoop projects log at INFO by default. Let's change the logging level to INFO by default. -- This message was sent by Atlassian JIRA (v6.1#6144)
[jira] [Commented] (HIVE-5619) Allow concat() to accept mixed string/binary args
[ https://issues.apache.org/jira/browse/HIVE-5619?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13806251#comment-13806251 ] Hudson commented on HIVE-5619: -- FAILURE: Integrated in Hive-trunk-hadoop2-ptest #152 (See [https://builds.apache.org/job/Hive-trunk-hadoop2-ptest/152/]) HIVE-5619 : Allow concat() to accept mixed string/binary args (Jason Dere via Ashutosh Chauhan) (hashutosh: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1535621) * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFConcat.java * /hive/trunk/ql/src/test/queries/clientpositive/udf_concat.q * /hive/trunk/ql/src/test/results/clientpositive/udf_concat.q.out * /hive/trunk/serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/primitive/PrimitiveObjectInspectorUtils.java Allow concat() to accept mixed string/binary args - Key: HIVE-5619 URL: https://issues.apache.org/jira/browse/HIVE-5619 Project: Hive Issue Type: Improvement Reporter: Jason Dere Assignee: Jason Dere Fix For: 0.13.0 Attachments: HIVE-5619.1.patch concat() is currently strict about allowing either all binary or all non-binary arguments. Loosen this to permit mixed params. -- This message was sent by Atlassian JIRA (v6.1#6144)
[jira] [Commented] (HIVE-784) Support uncorrelated subqueries in the WHERE clause
[ https://issues.apache.org/jira/browse/HIVE-784?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13806257#comment-13806257 ] Hudson commented on HIVE-784: - FAILURE: Integrated in Hive-trunk-hadoop2-ptest #152 (See [https://builds.apache.org/job/Hive-trunk-hadoop2-ptest/152/]) HIVE-784 : Support uncorrelated subqueries in the WHERE clause (Harish Butani via Ashutosh Chauhan) (hashutosh: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1535040) * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/ErrorMsg.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/parse/HiveParser.g * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/parse/IdentifiersParser.g * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/parse/QBSubQuery.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/parse/SemanticAnalyzer.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/parse/SubQueryUtils.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/parse/TypeCheckProcFactory.java * /hive/trunk/ql/src/test/queries/clientnegative/subquery_exists_implicit_gby.q * /hive/trunk/ql/src/test/queries/clientnegative/subquery_in_groupby.q * /hive/trunk/ql/src/test/queries/clientnegative/subquery_in_select.q * /hive/trunk/ql/src/test/queries/clientnegative/subquery_multiple_cols_in_select.q * /hive/trunk/ql/src/test/queries/clientnegative/subquery_notexists_implicit_gby.q * /hive/trunk/ql/src/test/queries/clientnegative/subquery_subquery_chain.q * /hive/trunk/ql/src/test/queries/clientnegative/subquery_windowing_corr.q * /hive/trunk/ql/src/test/queries/clientnegative/subquery_with_or_cond.q * /hive/trunk/ql/src/test/queries/clientpositive/subquery_exists.q * /hive/trunk/ql/src/test/queries/clientpositive/subquery_in.q * /hive/trunk/ql/src/test/queries/clientpositive/subquery_multiinsert.q * /hive/trunk/ql/src/test/queries/clientpositive/subquery_notexists.q * /hive/trunk/ql/src/test/queries/clientpositive/subquery_notin.q * /hive/trunk/ql/src/test/results/clientnegative/subquery_exists_implicit_gby.q.out * /hive/trunk/ql/src/test/results/clientnegative/subquery_in_groupby.q.out * /hive/trunk/ql/src/test/results/clientnegative/subquery_in_select.q.out * /hive/trunk/ql/src/test/results/clientnegative/subquery_multiple_cols_in_select.q.out * /hive/trunk/ql/src/test/results/clientnegative/subquery_notexists_implicit_gby.q.out * /hive/trunk/ql/src/test/results/clientnegative/subquery_subquery_chain.q.out * /hive/trunk/ql/src/test/results/clientnegative/subquery_windowing_corr.q.out * /hive/trunk/ql/src/test/results/clientnegative/subquery_with_or_cond.q.out * /hive/trunk/ql/src/test/results/clientpositive/subquery_exists.q.out * /hive/trunk/ql/src/test/results/clientpositive/subquery_in.q.out * /hive/trunk/ql/src/test/results/clientpositive/subquery_multiinsert.q.out * /hive/trunk/ql/src/test/results/clientpositive/subquery_notexists.q.out * /hive/trunk/ql/src/test/results/clientpositive/subquery_notin.q.out Support uncorrelated subqueries in the WHERE clause --- Key: HIVE-784 URL: https://issues.apache.org/jira/browse/HIVE-784 Project: Hive Issue Type: New Feature Components: Query Processor Reporter: Ning Zhang Assignee: Harish Butani Fix For: 0.13.0 Attachments: D13443.1.patch, D13443.2.patch, HIVE-784.1.patch.txt, HIVE-784.2.patch, SubQuerySpec.pdf, tpchQueriesUsingSubQueryClauses.sql Hive currently only support views in the FROM-clause, some Facebook use cases suggest that Hive should support subqueries such as those connected by IN/EXISTS in the WHERE-clause. -- This message was sent by Atlassian JIRA (v6.1#6144)
[jira] [Commented] (HIVE-5454) HCatalog runs a partition listing with an empty filter
[ https://issues.apache.org/jira/browse/HIVE-5454?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13806264#comment-13806264 ] Hudson commented on HIVE-5454: -- FAILURE: Integrated in Hive-trunk-hadoop2-ptest #152 (See [https://builds.apache.org/job/Hive-trunk-hadoop2-ptest/152/]) HIVE-5454 - HCatalog runs a partition listing with an empty filter (Harsh J via Brock Noland) (brock: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1535051) * /hive/trunk/hcatalog/core/src/main/java/org/apache/hive/hcatalog/data/transfer/impl/HCatInputFormatReader.java * /hive/trunk/hcatalog/core/src/main/java/org/apache/hive/hcatalog/mapreduce/HCatInputFormat.java * /hive/trunk/hcatalog/core/src/test/java/org/apache/hive/hcatalog/mapreduce/HCatMapReduceTest.java * /hive/trunk/hcatalog/hcatalog-pig-adapter/src/main/java/org/apache/hive/hcatalog/pig/HCatLoader.java * /hive/trunk/hcatalog/src/docs/src/documentation/content/xdocs/inputoutput.xml * /hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hcatalog/utils/HBaseReadWrite.java * /hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/GroupByAge.java * /hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/ReadJson.java * /hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/ReadRC.java * /hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/ReadText.java * /hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/ReadWrite.java * /hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/SimpleRead.java * /hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/StoreComplex.java * /hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/StoreDemo.java * /hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/StoreNumbers.java * /hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/SumNumbers.java * /hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/TypeDataCheck.java * /hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/WriteJson.java * /hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/WriteRC.java * /hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/WriteText.java * /hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/WriteTextPartitioned.java * /hive/trunk/hcatalog/storage-handlers/hbase/src/test/org/apache/hive/hcatalog/hbase/TestHBaseInputFormat.java HCatalog runs a partition listing with an empty filter -- Key: HIVE-5454 URL: https://issues.apache.org/jira/browse/HIVE-5454 Project: Hive Issue Type: Bug Components: HCatalog Affects Versions: 0.12.0 Reporter: Harsh J Assignee: Harsh J Fix For: 0.13.0 Attachments: D13317.1.patch, D13317.2.patch, D13317.3.patch This is a HCATALOG-527 caused regression, wherein the HCatLoader's way of calling HCatInputFormat causes it to do 2x partition lookups - once without the filter, and then again with the filter. For tables with large number partitions (10, say), the non-filter lookup proves fatal both to the client (Read timed out errors from ThriftMetaStoreClient cause the server doesn't respond) and to the server (too much data loaded into the cache, OOME, or slowdown). The fix would be to use a single call that also passes a partition filter information, as was in the case of HCatalog 0.4 sources before HCATALOG-527. (HCatalog-release-wise, this affects all 0.5.x users) -- This message was sent by Atlassian JIRA (v6.1#6144)
[jira] [Commented] (HIVE-5430) Refactor VectorizationContext and handle NOT expression with nulls.
[ https://issues.apache.org/jira/browse/HIVE-5430?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13806262#comment-13806262 ] Hudson commented on HIVE-5430: -- FAILURE: Integrated in Hive-trunk-hadoop2-ptest #152 (See [https://builds.apache.org/job/Hive-trunk-hadoop2-ptest/152/]) HIVE-5430 : Refactor VectorizationContext and handle NOT expression with nulls. (Jitendra Nath Pandey via Ashutosh Chauhan) (hashutosh: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1535055) * /hive/trunk/ant/src/org/apache/hadoop/hive/ant/GenVectorTestCode.java * /hive/trunk/ql/src/gen/vectorization/ExpressionTemplates/ColumnArithmeticColumn.txt * /hive/trunk/ql/src/gen/vectorization/ExpressionTemplates/ColumnArithmeticScalar.txt * /hive/trunk/ql/src/gen/vectorization/ExpressionTemplates/ColumnCompareColumn.txt * /hive/trunk/ql/src/gen/vectorization/ExpressionTemplates/ColumnCompareScalar.txt * /hive/trunk/ql/src/gen/vectorization/ExpressionTemplates/ColumnUnaryFunc.txt * /hive/trunk/ql/src/gen/vectorization/ExpressionTemplates/ColumnUnaryMinus.txt * /hive/trunk/ql/src/gen/vectorization/ExpressionTemplates/FilterColumnCompareColumn.txt * /hive/trunk/ql/src/gen/vectorization/ExpressionTemplates/FilterColumnCompareScalar.txt * /hive/trunk/ql/src/gen/vectorization/ExpressionTemplates/FilterScalarCompareColumn.txt * /hive/trunk/ql/src/gen/vectorization/ExpressionTemplates/FilterStringColumnCompareColumn.txt * /hive/trunk/ql/src/gen/vectorization/ExpressionTemplates/FilterStringColumnCompareScalar.txt * /hive/trunk/ql/src/gen/vectorization/ExpressionTemplates/FilterStringScalarCompareColumn.txt * /hive/trunk/ql/src/gen/vectorization/ExpressionTemplates/ScalarArithmeticColumn.txt * /hive/trunk/ql/src/gen/vectorization/ExpressionTemplates/ScalarCompareColumn.txt * /hive/trunk/ql/src/gen/vectorization/ExpressionTemplates/StringColumnCompareColumn.txt * /hive/trunk/ql/src/gen/vectorization/ExpressionTemplates/StringColumnCompareScalar.txt * /hive/trunk/ql/src/gen/vectorization/ExpressionTemplates/StringScalarCompareColumn.txt * /hive/trunk/ql/src/gen/vectorization/TestTemplates/TestColumnScalarFilterVectorExpressionEvaluation.txt * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/OperatorFactory.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/VectorExpressionDescriptor.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/VectorFilterOperator.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/VectorGroupByOperator.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/VectorMapJoinOperator.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/VectorReduceSinkOperator.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/VectorSelectOperator.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/VectorizationContext.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/VectorizedExpressions.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/expressions/AbstractFilterStringColLikeStringScalar.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/expressions/ColAndCol.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/expressions/ColOrCol.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/expressions/ConstantVectorExpression.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/expressions/FilterColAndScalar.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/expressions/FilterColOrScalar.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/expressions/FilterExprAndExpr.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/expressions/FilterExprOrExpr.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/expressions/FilterNotExpr.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/expressions/FilterScalarAndColumn.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/expressions/FilterScalarOrColumn.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/expressions/FilterStringColLikeStringScalar.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/expressions/FilterStringColRegExpStringScalar.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/expressions/FuncLogWithBaseDoubleToDouble.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/expressions/FuncLogWithBaseLongToDouble.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/expressions/FuncLongToString.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/expressions/FuncPowerDoubleToDouble.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/expressions/FuncPowerLongToDouble.java *
[jira] [Commented] (HIVE-5220) Add option for removing intermediate directory for partition, which is empty
[ https://issues.apache.org/jira/browse/HIVE-5220?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13806258#comment-13806258 ] Hudson commented on HIVE-5220: -- FAILURE: Integrated in Hive-trunk-hadoop2-ptest #152 (See [https://builds.apache.org/job/Hive-trunk-hadoop2-ptest/152/]) HIVE-5220 : Add option for removing intermediate directory for partition, which is empty (Navis via Ashutosh Chauhan) (hashutosh: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1535072) * /hive/trunk/metastore/src/java/org/apache/hadoop/hive/metastore/HiveMetaStore.java * /hive/trunk/metastore/src/java/org/apache/hadoop/hive/metastore/Warehouse.java Add option for removing intermediate directory for partition, which is empty Key: HIVE-5220 URL: https://issues.apache.org/jira/browse/HIVE-5220 Project: Hive Issue Type: Improvement Components: Query Processor Reporter: Navis Assignee: Navis Priority: Minor Fix For: 0.13.0 Attachments: D12729.2.patch, HIVE-5220.D12729.1.patch For deeply nested partitioned table, intermediate directories are not removed even if there is no partitions in it by removing them. {noformat} /deep_part/c=09/d=01 /deep_part/c=09/d=01/e=01 /deep_part/c=09/d=01/e=02 /deep_part/c=09/d=02 /deep_part/c=09/d=02/e=01 /deep_part/c=09/d=02/e=02 {noformat} After removing partition (c='09'), directory remains like this, {noformat} /deep_part/c=09/d=01 /deep_part/c=09/d=02 {noformat} -- This message was sent by Atlassian JIRA (v6.1#6144)
[jira] [Commented] (HIVE-5216) Need to annotate public API in HCatalog
[ https://issues.apache.org/jira/browse/HIVE-5216?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13806259#comment-13806259 ] Hudson commented on HIVE-5216: -- FAILURE: Integrated in Hive-trunk-hadoop2-ptest #152 (See [https://builds.apache.org/job/Hive-trunk-hadoop2-ptest/152/]) HIVE-5216 : Need to annotate public API in HCatalog (Eugene Koifman via Thejas Nair) (thejas: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1535535) * /hive/trunk/hcatalog/core/src/main/java/org/apache/hive/hcatalog/data/HCatRecord.java * /hive/trunk/hcatalog/core/src/main/java/org/apache/hive/hcatalog/data/schema/HCatSchema.java * /hive/trunk/hcatalog/core/src/main/java/org/apache/hive/hcatalog/data/transfer/HCatReader.java * /hive/trunk/hcatalog/core/src/main/java/org/apache/hive/hcatalog/data/transfer/HCatWriter.java * /hive/trunk/hcatalog/core/src/main/java/org/apache/hive/hcatalog/mapreduce/HCatOutputFormat.java * /hive/trunk/hcatalog/hcatalog-pig-adapter/src/main/java/org/apache/hive/hcatalog/pig/HCatLoader.java * /hive/trunk/hcatalog/hcatalog-pig-adapter/src/main/java/org/apache/hive/hcatalog/pig/HCatStorer.java * /hive/trunk/hcatalog/webhcat/svr/src/main/java/org/apache/hive/hcatalog/templeton/Main.java Need to annotate public API in HCatalog --- Key: HIVE-5216 URL: https://issues.apache.org/jira/browse/HIVE-5216 Project: Hive Issue Type: Bug Components: HCatalog, WebHCat Affects Versions: 0.12.0 Reporter: Eugene Koifman Assignee: Eugene Koifman Fix For: 0.13.0 Attachments: HIVE-5216.2.patch, HIVE-5216.patch need to annotate which API is considered public using something like @InterfaceAudience.Public @InterfaceStability.Evolving Currently this is what is considered (at a minimum) public API HCatLoader HCatStorer HCatInputFormat HCatOutputFormat HCatReader HCatWriter HCatRecord HCatSchema This is needed so that clients/dependent projects know which API they can rely on and which can change w/o notice. -- This message was sent by Atlassian JIRA (v6.1#6144)
[jira] [Commented] (HIVE-5628) ListBucketingPrunnerTest and DynamicMultiDimeCollectionTest should start with Test not end with it
[ https://issues.apache.org/jira/browse/HIVE-5628?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13806250#comment-13806250 ] Hudson commented on HIVE-5628: -- FAILURE: Integrated in Hive-trunk-hadoop2-ptest #152 (See [https://builds.apache.org/job/Hive-trunk-hadoop2-ptest/152/]) HIVE-5628 : ListBucketingPrunnerTest and DynamicMultiDimeCollectionTest should start with Test not end with it (Brock Noland via Ashutosh Chauhan) (hashutosh: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1535761) * /hive/trunk/ql/src/test/org/apache/hadoop/hive/ql/optimizer/listbucketingpruner/DynamicMultiDimeCollectionTest.java * /hive/trunk/ql/src/test/org/apache/hadoop/hive/ql/optimizer/listbucketingpruner/ListBucketingPrunnerTest.java * /hive/trunk/ql/src/test/org/apache/hadoop/hive/ql/optimizer/listbucketingpruner/TestDynamicMultiDimeCollection.java * /hive/trunk/ql/src/test/org/apache/hadoop/hive/ql/optimizer/listbucketingpruner/TestListBucketingPrunner.java ListBucketingPrunnerTest and DynamicMultiDimeCollectionTest should start with Test not end with it -- Key: HIVE-5628 URL: https://issues.apache.org/jira/browse/HIVE-5628 Project: Hive Issue Type: Bug Reporter: Brock Noland Assignee: Brock Noland Fix For: 0.13.0 Attachments: HIVE-5628.patch ListBucketingPrunnerTest and DynamicMultiDimeCollectionTest will not be run by PTest because they end with Test and PTest requires tests start with Test. -- This message was sent by Atlassian JIRA (v6.1#6144)
[jira] [Commented] (HIVE-5605) AddResourceOperation, DeleteResourceOperation, DfsOperation, SetOperation should be removed from org.apache.hive.service.cli.operation
[ https://issues.apache.org/jira/browse/HIVE-5605?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13806253#comment-13806253 ] Hudson commented on HIVE-5605: -- FAILURE: Integrated in Hive-trunk-hadoop2-ptest #152 (See [https://builds.apache.org/job/Hive-trunk-hadoop2-ptest/152/]) HIVE-5605 - AddResourceOperation, DeleteResourceOperation, DfsOperation, SetOperation should be removed from org.apache.hive.service.cli.operation (Vaibhav Gumashta via Brock Noland) (brock: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1535043) * /hive/trunk/service/src/java/org/apache/hive/service/cli/operation/AddResourceOperation.java * /hive/trunk/service/src/java/org/apache/hive/service/cli/operation/DeleteResourceOperation.java * /hive/trunk/service/src/java/org/apache/hive/service/cli/operation/DfsOperation.java * /hive/trunk/service/src/java/org/apache/hive/service/cli/operation/SetOperation.java AddResourceOperation, DeleteResourceOperation, DfsOperation, SetOperation should be removed from org.apache.hive.service.cli.operation --- Key: HIVE-5605 URL: https://issues.apache.org/jira/browse/HIVE-5605 Project: Hive Issue Type: Improvement Components: HiveServer2 Affects Versions: 0.13.0 Reporter: Vaibhav Gumashta Assignee: Vaibhav Gumashta Priority: Minor Fix For: 0.13.0 Attachments: HIVE-5605.1.patch These classes are not used as the processing for Add, Delete, DFS and Set commands is done by HiveCommandOperation -- This message was sent by Atlassian JIRA (v6.1#6144)
[jira] [Commented] (HIVE-5577) Remove TestNegativeCliDriver script_broken_pipe1
[ https://issues.apache.org/jira/browse/HIVE-5577?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13806261#comment-13806261 ] Hudson commented on HIVE-5577: -- FAILURE: Integrated in Hive-trunk-hadoop2-ptest #152 (See [https://builds.apache.org/job/Hive-trunk-hadoop2-ptest/152/]) HIVE-5577 : Remove TestNegativeCliDriver script_broken_pipe1 (Brock Noland via Navis) (navis: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1535643) * /hive/trunk/ql/src/test/queries/clientnegative/script_broken_pipe1.q * /hive/trunk/ql/src/test/results/clientnegative/script_broken_pipe1.q.out Remove TestNegativeCliDriver script_broken_pipe1 Key: HIVE-5577 URL: https://issues.apache.org/jira/browse/HIVE-5577 Project: Hive Issue Type: Bug Reporter: Brock Noland Assignee: Brock Noland Fix For: 0.13.0 Attachments: HIVE-5577.1.patch.txt TestNegativeCliDriver script_broken_pipe1 is extremely flaky and not a terribly important test. Let's remove it. Failures https://builds.apache.org/user/brock/my-views/view/hive/job/Hive-trunk-hadoop1-ptest/206/testReport/org.apache.hadoop.hive.cli/TestNegativeCliDriver/testNegativeCliDriver_script_broken_pipe1/ https://builds.apache.org/user/brock/my-views/view/hive/job/Hive-trunk-hadoop1-ptest/206/testReport/junit/org.apache.hadoop.hive.cli/TestNegativeCliDriver/testNegativeCliDriver_script_broken_pipe1/ https://builds.apache.org/user/brock/my-views/view/hive/job/Hive-trunk-hadoop1-ptest/204/testReport/org.apache.hadoop.hive.cli/TestNegativeCliDriver/testNegativeCliDriver_script_broken_pipe1/ -- This message was sent by Atlassian JIRA (v6.1#6144)
[jira] [Commented] (HIVE-5350) Cleanup exception handling around parallel orderby
[ https://issues.apache.org/jira/browse/HIVE-5350?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13806255#comment-13806255 ] Hudson commented on HIVE-5350: -- FAILURE: Integrated in Hive-trunk-hadoop2-ptest #152 (See [https://builds.apache.org/job/Hive-trunk-hadoop2-ptest/152/]) HIVE-5350 - Cleanup exception handling around parallel orderby (Navis via Brock Noland) (brock: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1535045) * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/PartitionKeySampler.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/mr/ExecDriver.java Cleanup exception handling around parallel orderby -- Key: HIVE-5350 URL: https://issues.apache.org/jira/browse/HIVE-5350 Project: Hive Issue Type: Bug Reporter: Brock Noland Assignee: Navis Priority: Minor Fix For: 0.13.0 Attachments: D13617.1.patch I think we should log the message to the console and the full exception to the log: ExecDriver: {noformat} try { handleSampling(driverContext, mWork, job, conf); job.setPartitionerClass(HiveTotalOrderPartitioner.class); } catch (Exception e) { console.printInfo(Not enough sampling data.. Rolling back to single reducer task); rWork.setNumReduceTasks(1); job.setNumReduceTasks(1); } {noformat} -- This message was sent by Atlassian JIRA (v6.1#6144)
[jira] [Commented] (HIVE-5440) HiveServer2 doesn't apply SQL operation's config property
[ https://issues.apache.org/jira/browse/HIVE-5440?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13806265#comment-13806265 ] Hudson commented on HIVE-5440: -- FAILURE: Integrated in Hive-trunk-hadoop2-ptest #152 (See [https://builds.apache.org/job/Hive-trunk-hadoop2-ptest/152/]) HIVE-5440: HiveServer2 doesn't apply SQL operation's config property (Prasad Mujumdar via Thejas Nair) (thejas: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1535889) * /hive/trunk/service/src/java/org/apache/hive/service/cli/operation/ExecuteStatementOperation.java * /hive/trunk/service/src/java/org/apache/hive/service/cli/operation/SQLOperation.java * /hive/trunk/service/src/test/org/apache/hive/service/cli/CLIServiceTest.java HiveServer2 doesn't apply SQL operation's config property -- Key: HIVE-5440 URL: https://issues.apache.org/jira/browse/HIVE-5440 Project: Hive Issue Type: Bug Components: HiveServer2 Affects Versions: 0.12.0 Reporter: Prasad Mujumdar Assignee: Prasad Mujumdar Fix For: 0.13.0 Attachments: HIVE-5440.1.patch, HIVE-5440.2.patch The HiveServer2 thrift IDL includes an optional config overlay map which is currently not used. -- This message was sent by Atlassian JIRA (v6.1#6144)
[jira] [Commented] (HIVE-5552) Merging of QBJoinTrees doesn't handle filter pushdowns correctly
[ https://issues.apache.org/jira/browse/HIVE-5552?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13806267#comment-13806267 ] Hudson commented on HIVE-5552: -- FAILURE: Integrated in Hive-trunk-hadoop2-ptest #152 (See [https://builds.apache.org/job/Hive-trunk-hadoop2-ptest/152/]) HIVE-5552 : Merging of QBJoinTrees doesnt handle filter pushdowns correctly (Harish Butani via Ashutosh Chauhan) (hashutosh: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1535858) * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/parse/ParseUtils.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/parse/QBJoinTree.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/parse/SemanticAnalyzer.java * /hive/trunk/ql/src/test/queries/clientpositive/join_merging.q * /hive/trunk/ql/src/test/results/clientpositive/join_merging.q.out Merging of QBJoinTrees doesn't handle filter pushdowns correctly Key: HIVE-5552 URL: https://issues.apache.org/jira/browse/HIVE-5552 Project: Hive Issue Type: Bug Components: Query Processor Reporter: Harish Butani Assignee: Harish Butani Fix For: 0.13.0 Attachments: HIVE-5552.1.patch, HIVE-5552.2.patch The following query fails: (this based on the schema from auto_join_filters.q) {noformat} explain SELECT sum(hash(a.key,a.value,b.key,b.value)) FROM myinput1 a LEFT OUTER JOIN myinput1 b on a.value = b.value RIGHT OUTER JOIN myinput1 c ON b.value = c.value AND a.key 40 {noformat} Whereas this query succeeds {noformat} explain SELECT sum(hash(a.key,a.value,b.key,b.value)) FROM myinput1 a LEFT OUTER JOIN myinput1 b RIGHT OUTER JOIN myinput1 c ON a.value = b.value and b.value = c.value AND a.key 40 {noformat} Pushing the first condition to the first join, triggers a merge of the 2 QBJoinTrees. During merge all the right side filters identified for pushing are assumed to refer to the merging table (b in this case). But the pushable filters can refer to any left table. -- This message was sent by Atlassian JIRA (v6.1#6144)
[jira] [Commented] (HIVE-5560) Hive produces incorrect results on multi-distinct query
[ https://issues.apache.org/jira/browse/HIVE-5560?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13806268#comment-13806268 ] Hudson commented on HIVE-5560: -- FAILURE: Integrated in Hive-trunk-hadoop2-ptest #152 (See [https://builds.apache.org/job/Hive-trunk-hadoop2-ptest/152/]) HIVE-5560 : Hive produces incorrect results on multi-distinct query (Navis via Ashutosh Chauhan) (hashutosh: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1535059) * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/parse/SemanticAnalyzer.java * /hive/trunk/ql/src/test/queries/clientpositive/groupby2_map_multi_distinct.q * /hive/trunk/ql/src/test/results/clientpositive/auto_join18_multi_distinct.q.out * /hive/trunk/ql/src/test/results/clientpositive/count.q.out * /hive/trunk/ql/src/test/results/clientpositive/groupby2_map_multi_distinct.q.out * /hive/trunk/ql/src/test/results/clientpositive/groupby2_noskew_multi_distinct.q.out * /hive/trunk/ql/src/test/results/clientpositive/groupby3_map_multi_distinct.q.out * /hive/trunk/ql/src/test/results/clientpositive/groupby3_noskew_multi_distinct.q.out * /hive/trunk/ql/src/test/results/clientpositive/groupby_distinct_samekey.q.out * /hive/trunk/ql/src/test/results/clientpositive/groupby_map_ppr_multi_distinct.q.out * /hive/trunk/ql/src/test/results/clientpositive/groupby_multi_single_reducer.q.out * /hive/trunk/ql/src/test/results/clientpositive/groupby_ppr_multi_distinct.q.out * /hive/trunk/ql/src/test/results/clientpositive/groupby_sort_11.q.out * /hive/trunk/ql/src/test/results/clientpositive/join18_multi_distinct.q.out * /hive/trunk/ql/src/test/results/clientpositive/limit_pushdown.q.out Hive produces incorrect results on multi-distinct query --- Key: HIVE-5560 URL: https://issues.apache.org/jira/browse/HIVE-5560 Project: Hive Issue Type: Bug Components: Query Processor Affects Versions: 0.11.0, 0.12.0 Reporter: Vikram Dixit K Assignee: Navis Fix For: 0.13.0 Attachments: D13599.1.patch, D13599.2.patch {noformat} select key, count(distinct key) + count(distinct value) from src tablesample (10 ROWS) group by key POSTHOOK: type: QUERY POSTHOOK: Input: default@src A masked pattern was here 165 1 val_165 1 238 1 val_238 1 255 1 val_255 1 27 1 val_27 1 278 1 val_278 1 311 1 val_311 1 409 1 val_409 1 484 1 val_484 1 86 1 val_86 1 98 1 val_98 1 {noformat} -- This message was sent by Atlassian JIRA (v6.1#6144)
[jira] [Commented] (HIVE-5629) Fix two javadoc failures in HCatalog
[ https://issues.apache.org/jira/browse/HIVE-5629?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13806279#comment-13806279 ] Hudson commented on HIVE-5629: -- SUCCESS: Integrated in Hive-trunk-hadoop1-ptest #215 (See [https://builds.apache.org/job/Hive-trunk-hadoop1-ptest/215/]) HIVE-5629 : Fix two javadoc failures in HCatalog (Brock Noland via Ashutosh Chauhan) (hashutosh: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1535513) * /hive/trunk/hcatalog/core/src/main/java/org/apache/hive/hcatalog/mapreduce/InputJobInfo.java * /hive/trunk/hcatalog/webhcat/svr/src/main/java/org/apache/hive/hcatalog/templeton/PigDelegator.java Fix two javadoc failures in HCatalog Key: HIVE-5629 URL: https://issues.apache.org/jira/browse/HIVE-5629 Project: Hive Issue Type: Bug Reporter: Brock Noland Assignee: Brock Noland Fix For: 0.13.0 Attachments: HIVE-5629.patch I am seeing two javadoc failures on HCatalog. These are not being seen by PTest and indeed I cannot reproduce on my Mac but can on Linux. Regardless they should be fixed. -- This message was sent by Atlassian JIRA (v6.1#6144)
[jira] [Commented] (HIVE-5628) ListBucketingPrunnerTest and DynamicMultiDimeCollectionTest should start with Test not end with it
[ https://issues.apache.org/jira/browse/HIVE-5628?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13806277#comment-13806277 ] Hudson commented on HIVE-5628: -- SUCCESS: Integrated in Hive-trunk-hadoop1-ptest #215 (See [https://builds.apache.org/job/Hive-trunk-hadoop1-ptest/215/]) HIVE-5628 : ListBucketingPrunnerTest and DynamicMultiDimeCollectionTest should start with Test not end with it (Brock Noland via Ashutosh Chauhan) (hashutosh: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1535761) * /hive/trunk/ql/src/test/org/apache/hadoop/hive/ql/optimizer/listbucketingpruner/DynamicMultiDimeCollectionTest.java * /hive/trunk/ql/src/test/org/apache/hadoop/hive/ql/optimizer/listbucketingpruner/ListBucketingPrunnerTest.java * /hive/trunk/ql/src/test/org/apache/hadoop/hive/ql/optimizer/listbucketingpruner/TestDynamicMultiDimeCollection.java * /hive/trunk/ql/src/test/org/apache/hadoop/hive/ql/optimizer/listbucketingpruner/TestListBucketingPrunner.java ListBucketingPrunnerTest and DynamicMultiDimeCollectionTest should start with Test not end with it -- Key: HIVE-5628 URL: https://issues.apache.org/jira/browse/HIVE-5628 Project: Hive Issue Type: Bug Reporter: Brock Noland Assignee: Brock Noland Fix For: 0.13.0 Attachments: HIVE-5628.patch ListBucketingPrunnerTest and DynamicMultiDimeCollectionTest will not be run by PTest because they end with Test and PTest requires tests start with Test. -- This message was sent by Atlassian JIRA (v6.1#6144)
[jira] [Commented] (HIVE-5605) AddResourceOperation, DeleteResourceOperation, DfsOperation, SetOperation should be removed from org.apache.hive.service.cli.operation
[ https://issues.apache.org/jira/browse/HIVE-5605?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13806280#comment-13806280 ] Hudson commented on HIVE-5605: -- SUCCESS: Integrated in Hive-trunk-hadoop1-ptest #215 (See [https://builds.apache.org/job/Hive-trunk-hadoop1-ptest/215/]) HIVE-5605 - AddResourceOperation, DeleteResourceOperation, DfsOperation, SetOperation should be removed from org.apache.hive.service.cli.operation (Vaibhav Gumashta via Brock Noland) (brock: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1535043) * /hive/trunk/service/src/java/org/apache/hive/service/cli/operation/AddResourceOperation.java * /hive/trunk/service/src/java/org/apache/hive/service/cli/operation/DeleteResourceOperation.java * /hive/trunk/service/src/java/org/apache/hive/service/cli/operation/DfsOperation.java * /hive/trunk/service/src/java/org/apache/hive/service/cli/operation/SetOperation.java AddResourceOperation, DeleteResourceOperation, DfsOperation, SetOperation should be removed from org.apache.hive.service.cli.operation --- Key: HIVE-5605 URL: https://issues.apache.org/jira/browse/HIVE-5605 Project: Hive Issue Type: Improvement Components: HiveServer2 Affects Versions: 0.13.0 Reporter: Vaibhav Gumashta Assignee: Vaibhav Gumashta Priority: Minor Fix For: 0.13.0 Attachments: HIVE-5605.1.patch These classes are not used as the processing for Add, Delete, DFS and Set commands is done by HiveCommandOperation -- This message was sent by Atlassian JIRA (v6.1#6144)
[jira] [Commented] (HIVE-5440) HiveServer2 doesn't apply SQL operation's config property
[ https://issues.apache.org/jira/browse/HIVE-5440?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13806292#comment-13806292 ] Hudson commented on HIVE-5440: -- SUCCESS: Integrated in Hive-trunk-hadoop1-ptest #215 (See [https://builds.apache.org/job/Hive-trunk-hadoop1-ptest/215/]) HIVE-5440: HiveServer2 doesn't apply SQL operation's config property (Prasad Mujumdar via Thejas Nair) (thejas: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1535889) * /hive/trunk/service/src/java/org/apache/hive/service/cli/operation/ExecuteStatementOperation.java * /hive/trunk/service/src/java/org/apache/hive/service/cli/operation/SQLOperation.java * /hive/trunk/service/src/test/org/apache/hive/service/cli/CLIServiceTest.java HiveServer2 doesn't apply SQL operation's config property -- Key: HIVE-5440 URL: https://issues.apache.org/jira/browse/HIVE-5440 Project: Hive Issue Type: Bug Components: HiveServer2 Affects Versions: 0.12.0 Reporter: Prasad Mujumdar Assignee: Prasad Mujumdar Fix For: 0.13.0 Attachments: HIVE-5440.1.patch, HIVE-5440.2.patch The HiveServer2 thrift IDL includes an optional config overlay map which is currently not used. -- This message was sent by Atlassian JIRA (v6.1#6144)
[jira] [Commented] (HIVE-5220) Add option for removing intermediate directory for partition, which is empty
[ https://issues.apache.org/jira/browse/HIVE-5220?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13806285#comment-13806285 ] Hudson commented on HIVE-5220: -- SUCCESS: Integrated in Hive-trunk-hadoop1-ptest #215 (See [https://builds.apache.org/job/Hive-trunk-hadoop1-ptest/215/]) HIVE-5220 : Add option for removing intermediate directory for partition, which is empty (Navis via Ashutosh Chauhan) (hashutosh: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1535072) * /hive/trunk/metastore/src/java/org/apache/hadoop/hive/metastore/HiveMetaStore.java * /hive/trunk/metastore/src/java/org/apache/hadoop/hive/metastore/Warehouse.java Add option for removing intermediate directory for partition, which is empty Key: HIVE-5220 URL: https://issues.apache.org/jira/browse/HIVE-5220 Project: Hive Issue Type: Improvement Components: Query Processor Reporter: Navis Assignee: Navis Priority: Minor Fix For: 0.13.0 Attachments: D12729.2.patch, HIVE-5220.D12729.1.patch For deeply nested partitioned table, intermediate directories are not removed even if there is no partitions in it by removing them. {noformat} /deep_part/c=09/d=01 /deep_part/c=09/d=01/e=01 /deep_part/c=09/d=01/e=02 /deep_part/c=09/d=02 /deep_part/c=09/d=02/e=01 /deep_part/c=09/d=02/e=02 {noformat} After removing partition (c='09'), directory remains like this, {noformat} /deep_part/c=09/d=01 /deep_part/c=09/d=02 {noformat} -- This message was sent by Atlassian JIRA (v6.1#6144)
[jira] [Commented] (HIVE-5637) Sporadic minimr test failure
[ https://issues.apache.org/jira/browse/HIVE-5637?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13806293#comment-13806293 ] Hudson commented on HIVE-5637: -- SUCCESS: Integrated in Hive-trunk-hadoop1-ptest #215 (See [https://builds.apache.org/job/Hive-trunk-hadoop1-ptest/215/]) HIVE-5637 : Sporadic minimr test failure (Navis via Ashutosh Chauhan) (hashutosh: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1535510) * /hive/trunk/ql/src/test/queries/clientpositive/import_exported_table.q Sporadic minimr test failure Key: HIVE-5637 URL: https://issues.apache.org/jira/browse/HIVE-5637 Project: Hive Issue Type: Test Components: Tests Reporter: Navis Assignee: Navis Priority: Trivial Fix For: 0.13.0 Attachments: HIVE-5637.1.patch.txt {noformat} ant test -Dtestcase=TestMinimrCliDriver -Dqfile=import_exported_table.q,load_hdfs_file_with_space_in_the_name.q -Dmodule=ql {noformat} Fails with message like this. {noformat} Begin query: load_hdfs_file_with_space_in_the_name.q mkdir: cannot create directory hdfs:///tmp/test/: File exists Exception: Client Execution failed with error code = -1 See build/ql/tmp/hive.log, or try ant test ... -Dtest.silent=false to get more logs. junit.framework.AssertionFailedError: Client Execution failed with error code = -1 See build/ql/tmp/hive.log, or try ant test ... -Dtest.silent=false to get more logs. at junit.framework.Assert.fail(Assert.java:47) at org.apache.hadoop.hive.cli.TestMinimrCliDriver.runTest(TestMinimrCliDriver.java:349) at org.apache.hadoop.hive.cli.TestMinimrCliDriver.testCliDriver_load_hdfs_file_with_space_in_the_name(TestMinimrCliDriver.java:291) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25) at java.lang.reflect.Method.invoke(Method.java:597) at junit.framework.TestCase.runTest(TestCase.java:154) at junit.framework.TestCase.runBare(TestCase.java:127) at junit.framework.TestResult$1.protect(TestResult.java:106) at junit.framework.TestResult.runProtected(TestResult.java:124) at junit.framework.TestResult.run(TestResult.java:109) at junit.framework.TestCase.run(TestCase.java:118) at junit.framework.TestSuite.runTest(TestSuite.java:208) at junit.framework.TestSuite.run(TestSuite.java:203) at org.apache.tools.ant.taskdefs.optional.junit.JUnitTestRunner.run(JUnitTestRunner.java:520) at org.apache.tools.ant.taskdefs.optional.junit.JUnitTestRunner.launch(JUnitTestRunner.java:1060) at org.apache.tools.ant.taskdefs.optional.junit.JUnitTestRunner.main(JUnitTestRunner.java:911) {noformat} -- This message was sent by Atlassian JIRA (v6.1#6144)
[jira] [Commented] (HIVE-5625) Fix issue with metastore version restriction test.
[ https://issues.apache.org/jira/browse/HIVE-5625?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13806281#comment-13806281 ] Hudson commented on HIVE-5625: -- SUCCESS: Integrated in Hive-trunk-hadoop1-ptest #215 (See [https://builds.apache.org/job/Hive-trunk-hadoop1-ptest/215/]) HIVE-5625 - Fix issue with metastore version restriction test. (Vikram Dixit K via Brock Noland) (brock: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1535402) * /hive/trunk/metastore/src/test/org/apache/hadoop/hive/metastore/TestMetastoreVersion.java Fix issue with metastore version restriction test. -- Key: HIVE-5625 URL: https://issues.apache.org/jira/browse/HIVE-5625 Project: Hive Issue Type: Bug Components: Tests Affects Versions: 0.13.0 Reporter: Vikram Dixit K Assignee: Vikram Dixit K Fix For: 0.13.0 Attachments: HIVE-5625.1.patch Based on Brock's comments, the change made in HIVE-5403 change the nature of the test. -- This message was sent by Atlassian JIRA (v6.1#6144)
[jira] [Commented] (HIVE-5577) Remove TestNegativeCliDriver script_broken_pipe1
[ https://issues.apache.org/jira/browse/HIVE-5577?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13806288#comment-13806288 ] Hudson commented on HIVE-5577: -- SUCCESS: Integrated in Hive-trunk-hadoop1-ptest #215 (See [https://builds.apache.org/job/Hive-trunk-hadoop1-ptest/215/]) HIVE-5577 : Remove TestNegativeCliDriver script_broken_pipe1 (Brock Noland via Navis) (navis: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1535643) * /hive/trunk/ql/src/test/queries/clientnegative/script_broken_pipe1.q * /hive/trunk/ql/src/test/results/clientnegative/script_broken_pipe1.q.out Remove TestNegativeCliDriver script_broken_pipe1 Key: HIVE-5577 URL: https://issues.apache.org/jira/browse/HIVE-5577 Project: Hive Issue Type: Bug Reporter: Brock Noland Assignee: Brock Noland Fix For: 0.13.0 Attachments: HIVE-5577.1.patch.txt TestNegativeCliDriver script_broken_pipe1 is extremely flaky and not a terribly important test. Let's remove it. Failures https://builds.apache.org/user/brock/my-views/view/hive/job/Hive-trunk-hadoop1-ptest/206/testReport/org.apache.hadoop.hive.cli/TestNegativeCliDriver/testNegativeCliDriver_script_broken_pipe1/ https://builds.apache.org/user/brock/my-views/view/hive/job/Hive-trunk-hadoop1-ptest/206/testReport/junit/org.apache.hadoop.hive.cli/TestNegativeCliDriver/testNegativeCliDriver_script_broken_pipe1/ https://builds.apache.org/user/brock/my-views/view/hive/job/Hive-trunk-hadoop1-ptest/204/testReport/org.apache.hadoop.hive.cli/TestNegativeCliDriver/testNegativeCliDriver_script_broken_pipe1/ -- This message was sent by Atlassian JIRA (v6.1#6144)
[jira] [Commented] (HIVE-5482) JDBC should depend on httpclient.version and httpcore.version 4.1.3 to be consistent with other modules
[ https://issues.apache.org/jira/browse/HIVE-5482?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13806283#comment-13806283 ] Hudson commented on HIVE-5482: -- SUCCESS: Integrated in Hive-trunk-hadoop1-ptest #215 (See [https://builds.apache.org/job/Hive-trunk-hadoop1-ptest/215/]) HIVE-5482 : JDBC should depend on httpclient.version and httpcore.version 4.1.3 to be consistent with other modules (Vaibhav Gumashta via Thejas Nair) (thejas: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1535620) * /hive/trunk/ivy/libraries.properties JDBC should depend on httpclient.version and httpcore.version 4.1.3 to be consistent with other modules --- Key: HIVE-5482 URL: https://issues.apache.org/jira/browse/HIVE-5482 Project: Hive Issue Type: Bug Components: HiveServer2, JDBC Reporter: Vaibhav Gumashta Assignee: Vaibhav Gumashta Fix For: 0.13.0 Attachments: HIVE-5482.1.patch Currently depends on 4.2.4 and 4.2.5 which conflicts with thrift-0.9 which depends on 4.1.3 -- This message was sent by Atlassian JIRA (v6.1#6144)
[jira] [Commented] (HIVE-5403) Move loading of filesystem, ugi, metastore client to hive session
[ https://issues.apache.org/jira/browse/HIVE-5403?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13806290#comment-13806290 ] Hudson commented on HIVE-5403: -- SUCCESS: Integrated in Hive-trunk-hadoop1-ptest #215 (See [https://builds.apache.org/job/Hive-trunk-hadoop1-ptest/215/]) HIVE-5403: Perflogger broken due to HIVE-5403 (Vikram Dixit K via Gunther Hagleitner) (gunther: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1535598) * /hive/trunk/metastore/src/test/org/apache/hadoop/hive/metastore/TestMetastoreVersion.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/session/SessionState.java HIVE-5403 : Move loading of filesystem, ugi, metastore client to hive session (Vikram Dixit via Ashutosh Chauhan) (hashutosh: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1535039) * /hive/trunk/metastore/src/test/org/apache/hadoop/hive/metastore/TestMetastoreVersion.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/metadata/Hive.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/session/SessionState.java Move loading of filesystem, ugi, metastore client to hive session - Key: HIVE-5403 URL: https://issues.apache.org/jira/browse/HIVE-5403 Project: Hive Issue Type: Bug Affects Versions: 0.13.0 Reporter: Vikram Dixit K Assignee: Vikram Dixit K Fix For: 0.13.0 Attachments: HIVE-5403.1.patch, HIVE-5403.2.patch, HIVE-5403.3.patch, HIVE-5403.4.patch As part of HIVE-5184, the metastore connection, loading filesystem were done as part of the tez session so as to speed up query times while paying a cost at startup. We can do this more generally in hive to apply to both the mapreduce and tez side of things. -- This message was sent by Atlassian JIRA (v6.1#6144)
[jira] [Commented] (HIVE-5454) HCatalog runs a partition listing with an empty filter
[ https://issues.apache.org/jira/browse/HIVE-5454?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13806291#comment-13806291 ] Hudson commented on HIVE-5454: -- SUCCESS: Integrated in Hive-trunk-hadoop1-ptest #215 (See [https://builds.apache.org/job/Hive-trunk-hadoop1-ptest/215/]) HIVE-5454 - HCatalog runs a partition listing with an empty filter (Harsh J via Brock Noland) (brock: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1535051) * /hive/trunk/hcatalog/core/src/main/java/org/apache/hive/hcatalog/data/transfer/impl/HCatInputFormatReader.java * /hive/trunk/hcatalog/core/src/main/java/org/apache/hive/hcatalog/mapreduce/HCatInputFormat.java * /hive/trunk/hcatalog/core/src/test/java/org/apache/hive/hcatalog/mapreduce/HCatMapReduceTest.java * /hive/trunk/hcatalog/hcatalog-pig-adapter/src/main/java/org/apache/hive/hcatalog/pig/HCatLoader.java * /hive/trunk/hcatalog/src/docs/src/documentation/content/xdocs/inputoutput.xml * /hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hcatalog/utils/HBaseReadWrite.java * /hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/GroupByAge.java * /hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/ReadJson.java * /hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/ReadRC.java * /hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/ReadText.java * /hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/ReadWrite.java * /hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/SimpleRead.java * /hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/StoreComplex.java * /hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/StoreDemo.java * /hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/StoreNumbers.java * /hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/SumNumbers.java * /hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/TypeDataCheck.java * /hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/WriteJson.java * /hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/WriteRC.java * /hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/WriteText.java * /hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/WriteTextPartitioned.java * /hive/trunk/hcatalog/storage-handlers/hbase/src/test/org/apache/hive/hcatalog/hbase/TestHBaseInputFormat.java HCatalog runs a partition listing with an empty filter -- Key: HIVE-5454 URL: https://issues.apache.org/jira/browse/HIVE-5454 Project: Hive Issue Type: Bug Components: HCatalog Affects Versions: 0.12.0 Reporter: Harsh J Assignee: Harsh J Fix For: 0.13.0 Attachments: D13317.1.patch, D13317.2.patch, D13317.3.patch This is a HCATALOG-527 caused regression, wherein the HCatLoader's way of calling HCatInputFormat causes it to do 2x partition lookups - once without the filter, and then again with the filter. For tables with large number partitions (10, say), the non-filter lookup proves fatal both to the client (Read timed out errors from ThriftMetaStoreClient cause the server doesn't respond) and to the server (too much data loaded into the cache, OOME, or slowdown). The fix would be to use a single call that also passes a partition filter information, as was in the case of HCatalog 0.4 sources before HCATALOG-527. (HCatalog-release-wise, this affects all 0.5.x users) -- This message was sent by Atlassian JIRA (v6.1#6144)
[jira] [Commented] (HIVE-5350) Cleanup exception handling around parallel orderby
[ https://issues.apache.org/jira/browse/HIVE-5350?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13806282#comment-13806282 ] Hudson commented on HIVE-5350: -- SUCCESS: Integrated in Hive-trunk-hadoop1-ptest #215 (See [https://builds.apache.org/job/Hive-trunk-hadoop1-ptest/215/]) HIVE-5350 - Cleanup exception handling around parallel orderby (Navis via Brock Noland) (brock: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1535045) * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/PartitionKeySampler.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/mr/ExecDriver.java Cleanup exception handling around parallel orderby -- Key: HIVE-5350 URL: https://issues.apache.org/jira/browse/HIVE-5350 Project: Hive Issue Type: Bug Reporter: Brock Noland Assignee: Navis Priority: Minor Fix For: 0.13.0 Attachments: D13617.1.patch I think we should log the message to the console and the full exception to the log: ExecDriver: {noformat} try { handleSampling(driverContext, mWork, job, conf); job.setPartitionerClass(HiveTotalOrderPartitioner.class); } catch (Exception e) { console.printInfo(Not enough sampling data.. Rolling back to single reducer task); rWork.setNumReduceTasks(1); job.setNumReduceTasks(1); } {noformat} -- This message was sent by Atlassian JIRA (v6.1#6144)
[jira] [Commented] (HIVE-5430) Refactor VectorizationContext and handle NOT expression with nulls.
[ https://issues.apache.org/jira/browse/HIVE-5430?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13806289#comment-13806289 ] Hudson commented on HIVE-5430: -- SUCCESS: Integrated in Hive-trunk-hadoop1-ptest #215 (See [https://builds.apache.org/job/Hive-trunk-hadoop1-ptest/215/]) HIVE-5430 : Refactor VectorizationContext and handle NOT expression with nulls. (Jitendra Nath Pandey via Ashutosh Chauhan) (hashutosh: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1535055) * /hive/trunk/ant/src/org/apache/hadoop/hive/ant/GenVectorTestCode.java * /hive/trunk/ql/src/gen/vectorization/ExpressionTemplates/ColumnArithmeticColumn.txt * /hive/trunk/ql/src/gen/vectorization/ExpressionTemplates/ColumnArithmeticScalar.txt * /hive/trunk/ql/src/gen/vectorization/ExpressionTemplates/ColumnCompareColumn.txt * /hive/trunk/ql/src/gen/vectorization/ExpressionTemplates/ColumnCompareScalar.txt * /hive/trunk/ql/src/gen/vectorization/ExpressionTemplates/ColumnUnaryFunc.txt * /hive/trunk/ql/src/gen/vectorization/ExpressionTemplates/ColumnUnaryMinus.txt * /hive/trunk/ql/src/gen/vectorization/ExpressionTemplates/FilterColumnCompareColumn.txt * /hive/trunk/ql/src/gen/vectorization/ExpressionTemplates/FilterColumnCompareScalar.txt * /hive/trunk/ql/src/gen/vectorization/ExpressionTemplates/FilterScalarCompareColumn.txt * /hive/trunk/ql/src/gen/vectorization/ExpressionTemplates/FilterStringColumnCompareColumn.txt * /hive/trunk/ql/src/gen/vectorization/ExpressionTemplates/FilterStringColumnCompareScalar.txt * /hive/trunk/ql/src/gen/vectorization/ExpressionTemplates/FilterStringScalarCompareColumn.txt * /hive/trunk/ql/src/gen/vectorization/ExpressionTemplates/ScalarArithmeticColumn.txt * /hive/trunk/ql/src/gen/vectorization/ExpressionTemplates/ScalarCompareColumn.txt * /hive/trunk/ql/src/gen/vectorization/ExpressionTemplates/StringColumnCompareColumn.txt * /hive/trunk/ql/src/gen/vectorization/ExpressionTemplates/StringColumnCompareScalar.txt * /hive/trunk/ql/src/gen/vectorization/ExpressionTemplates/StringScalarCompareColumn.txt * /hive/trunk/ql/src/gen/vectorization/TestTemplates/TestColumnScalarFilterVectorExpressionEvaluation.txt * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/OperatorFactory.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/VectorExpressionDescriptor.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/VectorFilterOperator.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/VectorGroupByOperator.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/VectorMapJoinOperator.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/VectorReduceSinkOperator.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/VectorSelectOperator.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/VectorizationContext.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/VectorizedExpressions.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/expressions/AbstractFilterStringColLikeStringScalar.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/expressions/ColAndCol.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/expressions/ColOrCol.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/expressions/ConstantVectorExpression.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/expressions/FilterColAndScalar.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/expressions/FilterColOrScalar.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/expressions/FilterExprAndExpr.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/expressions/FilterExprOrExpr.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/expressions/FilterNotExpr.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/expressions/FilterScalarAndColumn.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/expressions/FilterScalarOrColumn.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/expressions/FilterStringColLikeStringScalar.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/expressions/FilterStringColRegExpStringScalar.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/expressions/FuncLogWithBaseDoubleToDouble.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/expressions/FuncLogWithBaseLongToDouble.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/expressions/FuncLongToString.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/expressions/FuncPowerDoubleToDouble.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/expressions/FuncPowerLongToDouble.java *
[jira] [Commented] (HIVE-5552) Merging of QBJoinTrees doesn't handle filter pushdowns correctly
[ https://issues.apache.org/jira/browse/HIVE-5552?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13806294#comment-13806294 ] Hudson commented on HIVE-5552: -- SUCCESS: Integrated in Hive-trunk-hadoop1-ptest #215 (See [https://builds.apache.org/job/Hive-trunk-hadoop1-ptest/215/]) HIVE-5552 : Merging of QBJoinTrees doesnt handle filter pushdowns correctly (Harish Butani via Ashutosh Chauhan) (hashutosh: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1535858) * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/parse/ParseUtils.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/parse/QBJoinTree.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/parse/SemanticAnalyzer.java * /hive/trunk/ql/src/test/queries/clientpositive/join_merging.q * /hive/trunk/ql/src/test/results/clientpositive/join_merging.q.out Merging of QBJoinTrees doesn't handle filter pushdowns correctly Key: HIVE-5552 URL: https://issues.apache.org/jira/browse/HIVE-5552 Project: Hive Issue Type: Bug Components: Query Processor Reporter: Harish Butani Assignee: Harish Butani Fix For: 0.13.0 Attachments: HIVE-5552.1.patch, HIVE-5552.2.patch The following query fails: (this based on the schema from auto_join_filters.q) {noformat} explain SELECT sum(hash(a.key,a.value,b.key,b.value)) FROM myinput1 a LEFT OUTER JOIN myinput1 b on a.value = b.value RIGHT OUTER JOIN myinput1 c ON b.value = c.value AND a.key 40 {noformat} Whereas this query succeeds {noformat} explain SELECT sum(hash(a.key,a.value,b.key,b.value)) FROM myinput1 a LEFT OUTER JOIN myinput1 b RIGHT OUTER JOIN myinput1 c ON a.value = b.value and b.value = c.value AND a.key 40 {noformat} Pushing the first condition to the first join, triggers a merge of the 2 QBJoinTrees. During merge all the right side filters identified for pushing are assumed to refer to the merging table (b in this case). But the pushable filters can refer to any left table. -- This message was sent by Atlassian JIRA (v6.1#6144)
[jira] [Commented] (HIVE-5216) Need to annotate public API in HCatalog
[ https://issues.apache.org/jira/browse/HIVE-5216?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13806286#comment-13806286 ] Hudson commented on HIVE-5216: -- SUCCESS: Integrated in Hive-trunk-hadoop1-ptest #215 (See [https://builds.apache.org/job/Hive-trunk-hadoop1-ptest/215/]) HIVE-5216 : Need to annotate public API in HCatalog (Eugene Koifman via Thejas Nair) (thejas: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1535535) * /hive/trunk/hcatalog/core/src/main/java/org/apache/hive/hcatalog/data/HCatRecord.java * /hive/trunk/hcatalog/core/src/main/java/org/apache/hive/hcatalog/data/schema/HCatSchema.java * /hive/trunk/hcatalog/core/src/main/java/org/apache/hive/hcatalog/data/transfer/HCatReader.java * /hive/trunk/hcatalog/core/src/main/java/org/apache/hive/hcatalog/data/transfer/HCatWriter.java * /hive/trunk/hcatalog/core/src/main/java/org/apache/hive/hcatalog/mapreduce/HCatOutputFormat.java * /hive/trunk/hcatalog/hcatalog-pig-adapter/src/main/java/org/apache/hive/hcatalog/pig/HCatLoader.java * /hive/trunk/hcatalog/hcatalog-pig-adapter/src/main/java/org/apache/hive/hcatalog/pig/HCatStorer.java * /hive/trunk/hcatalog/webhcat/svr/src/main/java/org/apache/hive/hcatalog/templeton/Main.java Need to annotate public API in HCatalog --- Key: HIVE-5216 URL: https://issues.apache.org/jira/browse/HIVE-5216 Project: Hive Issue Type: Bug Components: HCatalog, WebHCat Affects Versions: 0.12.0 Reporter: Eugene Koifman Assignee: Eugene Koifman Fix For: 0.13.0 Attachments: HIVE-5216.2.patch, HIVE-5216.patch need to annotate which API is considered public using something like @InterfaceAudience.Public @InterfaceStability.Evolving Currently this is what is considered (at a minimum) public API HCatLoader HCatStorer HCatInputFormat HCatOutputFormat HCatReader HCatWriter HCatRecord HCatSchema This is needed so that clients/dependent projects know which API they can rely on and which can change w/o notice. -- This message was sent by Atlassian JIRA (v6.1#6144)
[jira] [Commented] (HIVE-784) Support uncorrelated subqueries in the WHERE clause
[ https://issues.apache.org/jira/browse/HIVE-784?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13806284#comment-13806284 ] Hudson commented on HIVE-784: - SUCCESS: Integrated in Hive-trunk-hadoop1-ptest #215 (See [https://builds.apache.org/job/Hive-trunk-hadoop1-ptest/215/]) HIVE-784 : Support uncorrelated subqueries in the WHERE clause (Harish Butani via Ashutosh Chauhan) (hashutosh: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1535040) * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/ErrorMsg.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/parse/HiveParser.g * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/parse/IdentifiersParser.g * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/parse/QBSubQuery.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/parse/SemanticAnalyzer.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/parse/SubQueryUtils.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/parse/TypeCheckProcFactory.java * /hive/trunk/ql/src/test/queries/clientnegative/subquery_exists_implicit_gby.q * /hive/trunk/ql/src/test/queries/clientnegative/subquery_in_groupby.q * /hive/trunk/ql/src/test/queries/clientnegative/subquery_in_select.q * /hive/trunk/ql/src/test/queries/clientnegative/subquery_multiple_cols_in_select.q * /hive/trunk/ql/src/test/queries/clientnegative/subquery_notexists_implicit_gby.q * /hive/trunk/ql/src/test/queries/clientnegative/subquery_subquery_chain.q * /hive/trunk/ql/src/test/queries/clientnegative/subquery_windowing_corr.q * /hive/trunk/ql/src/test/queries/clientnegative/subquery_with_or_cond.q * /hive/trunk/ql/src/test/queries/clientpositive/subquery_exists.q * /hive/trunk/ql/src/test/queries/clientpositive/subquery_in.q * /hive/trunk/ql/src/test/queries/clientpositive/subquery_multiinsert.q * /hive/trunk/ql/src/test/queries/clientpositive/subquery_notexists.q * /hive/trunk/ql/src/test/queries/clientpositive/subquery_notin.q * /hive/trunk/ql/src/test/results/clientnegative/subquery_exists_implicit_gby.q.out * /hive/trunk/ql/src/test/results/clientnegative/subquery_in_groupby.q.out * /hive/trunk/ql/src/test/results/clientnegative/subquery_in_select.q.out * /hive/trunk/ql/src/test/results/clientnegative/subquery_multiple_cols_in_select.q.out * /hive/trunk/ql/src/test/results/clientnegative/subquery_notexists_implicit_gby.q.out * /hive/trunk/ql/src/test/results/clientnegative/subquery_subquery_chain.q.out * /hive/trunk/ql/src/test/results/clientnegative/subquery_windowing_corr.q.out * /hive/trunk/ql/src/test/results/clientnegative/subquery_with_or_cond.q.out * /hive/trunk/ql/src/test/results/clientpositive/subquery_exists.q.out * /hive/trunk/ql/src/test/results/clientpositive/subquery_in.q.out * /hive/trunk/ql/src/test/results/clientpositive/subquery_multiinsert.q.out * /hive/trunk/ql/src/test/results/clientpositive/subquery_notexists.q.out * /hive/trunk/ql/src/test/results/clientpositive/subquery_notin.q.out Support uncorrelated subqueries in the WHERE clause --- Key: HIVE-784 URL: https://issues.apache.org/jira/browse/HIVE-784 Project: Hive Issue Type: New Feature Components: Query Processor Reporter: Ning Zhang Assignee: Harish Butani Fix For: 0.13.0 Attachments: D13443.1.patch, D13443.2.patch, HIVE-784.1.patch.txt, HIVE-784.2.patch, SubQuerySpec.pdf, tpchQueriesUsingSubQueryClauses.sql Hive currently only support views in the FROM-clause, some Facebook use cases suggest that Hive should support subqueries such as those connected by IN/EXISTS in the WHERE-clause. -- This message was sent by Atlassian JIRA (v6.1#6144)
[jira] [Commented] (HIVE-5599) Change default logging level to INFO
[ https://issues.apache.org/jira/browse/HIVE-5599?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13806287#comment-13806287 ] Hudson commented on HIVE-5599: -- SUCCESS: Integrated in Hive-trunk-hadoop1-ptest #215 (See [https://builds.apache.org/job/Hive-trunk-hadoop1-ptest/215/]) HIVE-5599 - Change default logging level to INFO (Brock Noland, Reviewed by Thejas M Nair) (brock: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1535047) * /hive/trunk/common/src/java/conf/hive-log4j.properties Change default logging level to INFO Key: HIVE-5599 URL: https://issues.apache.org/jira/browse/HIVE-5599 Project: Hive Issue Type: Bug Affects Versions: 0.12.0 Reporter: Brock Noland Assignee: Brock Noland Fix For: 0.13.0 Attachments: HIVE-5599.patch The default logging level is warn: https://github.com/apache/hive/blob/trunk/common/src/java/conf/hive-log4j.properties#L19 but hive logs lot's of good information at INFO level. Additionally most hadoop projects log at INFO by default. Let's change the logging level to INFO by default. -- This message was sent by Atlassian JIRA (v6.1#6144)
[jira] [Commented] (HIVE-5619) Allow concat() to accept mixed string/binary args
[ https://issues.apache.org/jira/browse/HIVE-5619?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13806278#comment-13806278 ] Hudson commented on HIVE-5619: -- SUCCESS: Integrated in Hive-trunk-hadoop1-ptest #215 (See [https://builds.apache.org/job/Hive-trunk-hadoop1-ptest/215/]) HIVE-5619 : Allow concat() to accept mixed string/binary args (Jason Dere via Ashutosh Chauhan) (hashutosh: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1535621) * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFConcat.java * /hive/trunk/ql/src/test/queries/clientpositive/udf_concat.q * /hive/trunk/ql/src/test/results/clientpositive/udf_concat.q.out * /hive/trunk/serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/primitive/PrimitiveObjectInspectorUtils.java Allow concat() to accept mixed string/binary args - Key: HIVE-5619 URL: https://issues.apache.org/jira/browse/HIVE-5619 Project: Hive Issue Type: Improvement Reporter: Jason Dere Assignee: Jason Dere Fix For: 0.13.0 Attachments: HIVE-5619.1.patch concat() is currently strict about allowing either all binary or all non-binary arguments. Loosen this to permit mixed params. -- This message was sent by Atlassian JIRA (v6.1#6144)
[jira] [Commented] (HIVE-5511) percentComplete returned by job status from WebHCat is null
[ https://issues.apache.org/jira/browse/HIVE-5511?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13806276#comment-13806276 ] Hudson commented on HIVE-5511: -- SUCCESS: Integrated in Hive-trunk-hadoop1-ptest #215 (See [https://builds.apache.org/job/Hive-trunk-hadoop1-ptest/215/]) HIVE-5511 : percentComplete returned by job status from WebHCat is null (Eugene Koifman via Thejas Nair) (thejas: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1535796) * /hive/trunk/hcatalog/src/test/e2e/templeton/README.txt * /hive/trunk/hcatalog/src/test/e2e/templeton/drivers/TestDriverCurl.pm * /hive/trunk/hcatalog/src/test/e2e/templeton/tests/jobsubmission.conf * /hive/trunk/hcatalog/src/test/e2e/templeton/tests/jobsubmission_streaming.conf * /hive/trunk/hcatalog/webhcat/svr/src/main/bin/webhcat_config.sh * /hive/trunk/hcatalog/webhcat/svr/src/main/config/override-container-log4j.properties * /hive/trunk/hcatalog/webhcat/svr/src/main/java/org/apache/hive/hcatalog/templeton/AppConfig.java * /hive/trunk/hcatalog/webhcat/svr/src/main/java/org/apache/hive/hcatalog/templeton/CompleteDelegator.java * /hive/trunk/hcatalog/webhcat/svr/src/main/java/org/apache/hive/hcatalog/templeton/tool/HDFSStorage.java * /hive/trunk/hcatalog/webhcat/svr/src/main/java/org/apache/hive/hcatalog/templeton/tool/HiveJobIDParser.java * /hive/trunk/hcatalog/webhcat/svr/src/main/java/org/apache/hive/hcatalog/templeton/tool/JarJobIDParser.java * /hive/trunk/hcatalog/webhcat/svr/src/main/java/org/apache/hive/hcatalog/templeton/tool/JobSubmissionConstants.java * /hive/trunk/hcatalog/webhcat/svr/src/main/java/org/apache/hive/hcatalog/templeton/tool/LaunchMapper.java * /hive/trunk/hcatalog/webhcat/svr/src/main/java/org/apache/hive/hcatalog/templeton/tool/PigJobIDParser.java * /hive/trunk/hcatalog/webhcat/svr/src/main/java/org/apache/hive/hcatalog/templeton/tool/TempletonControllerJob.java * /hive/trunk/hcatalog/webhcat/svr/src/main/java/org/apache/hive/hcatalog/templeton/tool/TempletonUtils.java * /hive/trunk/hcatalog/webhcat/svr/src/main/java/org/apache/hive/hcatalog/templeton/tool/TrivialExecService.java * /hive/trunk/hcatalog/webhcat/svr/src/test/java/org/apache/hive/hcatalog/templeton/tool/TestTrivialExecService.java * /hive/trunk/shims/src/0.20S/java/org/apache/hadoop/mapred/WebHCatJTShim20S.java * /hive/trunk/shims/src/0.23/java/org/apache/hadoop/mapred/WebHCatJTShim23.java * /hive/trunk/shims/src/common/java/org/apache/hadoop/hive/shims/HadoopShims.java percentComplete returned by job status from WebHCat is null --- Key: HIVE-5511 URL: https://issues.apache.org/jira/browse/HIVE-5511 Project: Hive Issue Type: Bug Components: WebHCat Affects Versions: 0.12.0 Reporter: Eugene Koifman Assignee: Eugene Koifman Fix For: 0.13.0 Attachments: HIVE-5511.3.patch, HIVE-5511.5.patch In hadoop1 the logging from MR is sent to stderr. In H2, by default, to syslog. templeton.tool.LaunchMapper expects to see the output on stderr to produce 'percentComplete' in job status. -- This message was sent by Atlassian JIRA (v6.1#6144)
[jira] [Commented] (HIVE-5560) Hive produces incorrect results on multi-distinct query
[ https://issues.apache.org/jira/browse/HIVE-5560?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13806295#comment-13806295 ] Hudson commented on HIVE-5560: -- SUCCESS: Integrated in Hive-trunk-hadoop1-ptest #215 (See [https://builds.apache.org/job/Hive-trunk-hadoop1-ptest/215/]) HIVE-5560 : Hive produces incorrect results on multi-distinct query (Navis via Ashutosh Chauhan) (hashutosh: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1535059) * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/parse/SemanticAnalyzer.java * /hive/trunk/ql/src/test/queries/clientpositive/groupby2_map_multi_distinct.q * /hive/trunk/ql/src/test/results/clientpositive/auto_join18_multi_distinct.q.out * /hive/trunk/ql/src/test/results/clientpositive/count.q.out * /hive/trunk/ql/src/test/results/clientpositive/groupby2_map_multi_distinct.q.out * /hive/trunk/ql/src/test/results/clientpositive/groupby2_noskew_multi_distinct.q.out * /hive/trunk/ql/src/test/results/clientpositive/groupby3_map_multi_distinct.q.out * /hive/trunk/ql/src/test/results/clientpositive/groupby3_noskew_multi_distinct.q.out * /hive/trunk/ql/src/test/results/clientpositive/groupby_distinct_samekey.q.out * /hive/trunk/ql/src/test/results/clientpositive/groupby_map_ppr_multi_distinct.q.out * /hive/trunk/ql/src/test/results/clientpositive/groupby_multi_single_reducer.q.out * /hive/trunk/ql/src/test/results/clientpositive/groupby_ppr_multi_distinct.q.out * /hive/trunk/ql/src/test/results/clientpositive/groupby_sort_11.q.out * /hive/trunk/ql/src/test/results/clientpositive/join18_multi_distinct.q.out * /hive/trunk/ql/src/test/results/clientpositive/limit_pushdown.q.out Hive produces incorrect results on multi-distinct query --- Key: HIVE-5560 URL: https://issues.apache.org/jira/browse/HIVE-5560 Project: Hive Issue Type: Bug Components: Query Processor Affects Versions: 0.11.0, 0.12.0 Reporter: Vikram Dixit K Assignee: Navis Fix For: 0.13.0 Attachments: D13599.1.patch, D13599.2.patch {noformat} select key, count(distinct key) + count(distinct value) from src tablesample (10 ROWS) group by key POSTHOOK: type: QUERY POSTHOOK: Input: default@src A masked pattern was here 165 1 val_165 1 238 1 val_238 1 255 1 val_255 1 27 1 val_27 1 278 1 val_278 1 311 1 val_311 1 409 1 val_409 1 484 1 val_484 1 86 1 val_86 1 98 1 val_98 1 {noformat} -- This message was sent by Atlassian JIRA (v6.1#6144)
[jira] [Commented] (HIVE-4974) JDBC2 statements and result sets are not able to return their parents
[ https://issues.apache.org/jira/browse/HIVE-4974?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13806376#comment-13806376 ] Hudson commented on HIVE-4974: -- FAILURE: Integrated in Hive-trunk-hadoop2-ptest #153 (See [https://builds.apache.org/job/Hive-trunk-hadoop2-ptest/153/]) HIVE-4974 - JDBC2 statements and result sets are not able to return their parents (Chris Dome via Brock Noland) (brock: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1536153) * /hive/trunk/jdbc/src/java/org/apache/hive/jdbc/HiveBaseResultSet.java * /hive/trunk/jdbc/src/java/org/apache/hive/jdbc/HiveCallableStatement.java * /hive/trunk/jdbc/src/java/org/apache/hive/jdbc/HiveConnection.java * /hive/trunk/jdbc/src/java/org/apache/hive/jdbc/HiveDatabaseMetaData.java * /hive/trunk/jdbc/src/java/org/apache/hive/jdbc/HivePreparedStatement.java * /hive/trunk/jdbc/src/java/org/apache/hive/jdbc/HiveQueryResultSet.java * /hive/trunk/jdbc/src/java/org/apache/hive/jdbc/HiveStatement.java * /hive/trunk/jdbc/src/test/org/apache/hive/jdbc/TestJdbcDriver2.java JDBC2 statements and result sets are not able to return their parents - Key: HIVE-4974 URL: https://issues.apache.org/jira/browse/HIVE-4974 Project: Hive Issue Type: Bug Components: JDBC Affects Versions: 0.11.0, 0.12.0 Reporter: Chris Drome Assignee: Chris Drome Priority: Minor Fix For: 0.13.0 Attachments: HIVE-4974.2-trunk.patch.txt, HIVE-4974-trunk-1.patch.txt, HIVE-4974-trunk-2.patch.txt, HIVE-4974-trunk.patch.txt The getConnection method of HiveStatement and HivePreparedStatement throw a not supported SQLException. The constructors should take the HiveConnection that creates them as an argument. Similarly, HiveBaseResultSet is not capable of returning the Statement that created it. -- This message was sent by Atlassian JIRA (v6.1#6144)
[jira] [Commented] (HIVE-3976) Support specifying scale and precision with Hive decimal type
[ https://issues.apache.org/jira/browse/HIVE-3976?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13806375#comment-13806375 ] Hudson commented on HIVE-3976: -- FAILURE: Integrated in Hive-trunk-hadoop2-ptest #153 (See [https://builds.apache.org/job/Hive-trunk-hadoop2-ptest/153/]) HIVE-3976 - Support specifying scale and precision with Hive decimal type (Xuefu Zhang via Brock Noland) (brock: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1536151) * /hive/trunk/common/src/java/org/apache/hadoop/hive/common/type/HiveDecimal.java * /hive/trunk/common/src/test/org/apache/hadoop/hive/common/type/TestHiveDecimal.java * /hive/trunk/data/files/kv9.txt * /hive/trunk/jdbc/src/java/org/apache/hadoop/hive/jdbc/HiveResultSetMetaData.java * /hive/trunk/jdbc/src/java/org/apache/hadoop/hive/jdbc/Utils.java * /hive/trunk/jdbc/src/test/org/apache/hadoop/hive/jdbc/TestJdbcDriver.java * /hive/trunk/jdbc/src/test/org/apache/hive/jdbc/TestJdbcDriver2.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/Driver.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/FunctionRegistry.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/NumericOpMethodResolver.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/io/orc/OrcStruct.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/io/orc/WriterImpl.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/parse/DDLSemanticAnalyzer.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/parse/HiveParser.g * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/parse/ParseUtils.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/parse/TypeCheckProcFactory.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/UDFOPDivide.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/UDFOPPlus.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFBridge.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFToDecimal.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFToVarchar.java * /hive/trunk/ql/src/test/org/apache/hadoop/hive/ql/exec/TestFunctionRegistry.java * /hive/trunk/ql/src/test/org/apache/hadoop/hive/ql/io/orc/TestOrcFile.java * /hive/trunk/ql/src/test/org/apache/hadoop/hive/ql/parse/TestHiveDecimalParse.java * /hive/trunk/ql/src/test/queries/clientpositive/decimal_1.q * /hive/trunk/ql/src/test/queries/clientpositive/decimal_2.q * /hive/trunk/ql/src/test/queries/clientpositive/decimal_3.q * /hive/trunk/ql/src/test/queries/clientpositive/decimal_4.q * /hive/trunk/ql/src/test/queries/clientpositive/decimal_5.q * /hive/trunk/ql/src/test/queries/clientpositive/decimal_6.q * /hive/trunk/ql/src/test/queries/clientpositive/decimal_join.q * /hive/trunk/ql/src/test/queries/clientpositive/decimal_precision.q * /hive/trunk/ql/src/test/queries/clientpositive/decimal_udf.q * /hive/trunk/ql/src/test/queries/clientpositive/orc_predicate_pushdown.q * /hive/trunk/ql/src/test/queries/clientpositive/ptf_decimal.q * /hive/trunk/ql/src/test/queries/clientpositive/serde_regex.q * /hive/trunk/ql/src/test/queries/clientpositive/udf_pmod.q * /hive/trunk/ql/src/test/queries/clientpositive/udf_to_double.q * /hive/trunk/ql/src/test/queries/clientpositive/udf_to_float.q * /hive/trunk/ql/src/test/queries/clientpositive/udf_to_string.q * /hive/trunk/ql/src/test/queries/clientpositive/windowing_expressions.q * /hive/trunk/ql/src/test/queries/clientpositive/windowing_multipartitioning.q * /hive/trunk/ql/src/test/queries/clientpositive/windowing_navfn.q * /hive/trunk/ql/src/test/queries/clientpositive/windowing_ntile.q * /hive/trunk/ql/src/test/queries/clientpositive/windowing_rank.q * /hive/trunk/ql/src/test/results/clientnegative/invalid_cast_from_binary_1.q.out * /hive/trunk/ql/src/test/results/clientnegative/invalid_cast_from_binary_2.q.out * /hive/trunk/ql/src/test/results/clientnegative/invalid_cast_from_binary_3.q.out * /hive/trunk/ql/src/test/results/clientnegative/invalid_cast_from_binary_4.q.out * /hive/trunk/ql/src/test/results/clientnegative/invalid_cast_from_binary_5.q.out * /hive/trunk/ql/src/test/results/clientnegative/invalid_cast_from_binary_6.q.out * /hive/trunk/ql/src/test/results/clientnegative/wrong_column_type.q.out * /hive/trunk/ql/src/test/results/clientpositive/decimal_1.q.out * /hive/trunk/ql/src/test/results/clientpositive/decimal_2.q.out * /hive/trunk/ql/src/test/results/clientpositive/decimal_3.q.out * /hive/trunk/ql/src/test/results/clientpositive/decimal_4.q.out * /hive/trunk/ql/src/test/results/clientpositive/decimal_5.q.out * /hive/trunk/ql/src/test/results/clientpositive/decimal_6.q.out * /hive/trunk/ql/src/test/results/clientpositive/decimal_join.q.out * /hive/trunk/ql/src/test/results/clientpositive/decimal_precision.q.out * /hive/trunk/ql/src/test/results/clientpositive/decimal_serde.q.out * /hive/trunk/ql/src/test/results/clientpositive/decimal_udf.q.out
[jira] [Commented] (HIVE-4974) JDBC2 statements and result sets are not able to return their parents
[ https://issues.apache.org/jira/browse/HIVE-4974?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13806432#comment-13806432 ] Hudson commented on HIVE-4974: -- SUCCESS: Integrated in Hive-trunk-hadoop1-ptest #216 (See [https://builds.apache.org/job/Hive-trunk-hadoop1-ptest/216/]) HIVE-4974 - JDBC2 statements and result sets are not able to return their parents (Chris Dome via Brock Noland) (brock: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1536153) * /hive/trunk/jdbc/src/java/org/apache/hive/jdbc/HiveBaseResultSet.java * /hive/trunk/jdbc/src/java/org/apache/hive/jdbc/HiveCallableStatement.java * /hive/trunk/jdbc/src/java/org/apache/hive/jdbc/HiveConnection.java * /hive/trunk/jdbc/src/java/org/apache/hive/jdbc/HiveDatabaseMetaData.java * /hive/trunk/jdbc/src/java/org/apache/hive/jdbc/HivePreparedStatement.java * /hive/trunk/jdbc/src/java/org/apache/hive/jdbc/HiveQueryResultSet.java * /hive/trunk/jdbc/src/java/org/apache/hive/jdbc/HiveStatement.java * /hive/trunk/jdbc/src/test/org/apache/hive/jdbc/TestJdbcDriver2.java JDBC2 statements and result sets are not able to return their parents - Key: HIVE-4974 URL: https://issues.apache.org/jira/browse/HIVE-4974 Project: Hive Issue Type: Bug Components: JDBC Affects Versions: 0.11.0, 0.12.0 Reporter: Chris Drome Assignee: Chris Drome Priority: Minor Fix For: 0.13.0 Attachments: HIVE-4974.2-trunk.patch.txt, HIVE-4974-trunk-1.patch.txt, HIVE-4974-trunk-2.patch.txt, HIVE-4974-trunk.patch.txt The getConnection method of HiveStatement and HivePreparedStatement throw a not supported SQLException. The constructors should take the HiveConnection that creates them as an argument. Similarly, HiveBaseResultSet is not capable of returning the Statement that created it. -- This message was sent by Atlassian JIRA (v6.1#6144)
[jira] [Commented] (HIVE-3976) Support specifying scale and precision with Hive decimal type
[ https://issues.apache.org/jira/browse/HIVE-3976?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13806431#comment-13806431 ] Hudson commented on HIVE-3976: -- SUCCESS: Integrated in Hive-trunk-hadoop1-ptest #216 (See [https://builds.apache.org/job/Hive-trunk-hadoop1-ptest/216/]) HIVE-3976 - Support specifying scale and precision with Hive decimal type (Xuefu Zhang via Brock Noland) (brock: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1536151) * /hive/trunk/common/src/java/org/apache/hadoop/hive/common/type/HiveDecimal.java * /hive/trunk/common/src/test/org/apache/hadoop/hive/common/type/TestHiveDecimal.java * /hive/trunk/data/files/kv9.txt * /hive/trunk/jdbc/src/java/org/apache/hadoop/hive/jdbc/HiveResultSetMetaData.java * /hive/trunk/jdbc/src/java/org/apache/hadoop/hive/jdbc/Utils.java * /hive/trunk/jdbc/src/test/org/apache/hadoop/hive/jdbc/TestJdbcDriver.java * /hive/trunk/jdbc/src/test/org/apache/hive/jdbc/TestJdbcDriver2.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/Driver.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/FunctionRegistry.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/NumericOpMethodResolver.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/io/orc/OrcStruct.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/io/orc/WriterImpl.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/parse/DDLSemanticAnalyzer.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/parse/HiveParser.g * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/parse/ParseUtils.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/parse/TypeCheckProcFactory.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/UDFOPDivide.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/UDFOPPlus.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFBridge.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFToDecimal.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFToVarchar.java * /hive/trunk/ql/src/test/org/apache/hadoop/hive/ql/exec/TestFunctionRegistry.java * /hive/trunk/ql/src/test/org/apache/hadoop/hive/ql/io/orc/TestOrcFile.java * /hive/trunk/ql/src/test/org/apache/hadoop/hive/ql/parse/TestHiveDecimalParse.java * /hive/trunk/ql/src/test/queries/clientpositive/decimal_1.q * /hive/trunk/ql/src/test/queries/clientpositive/decimal_2.q * /hive/trunk/ql/src/test/queries/clientpositive/decimal_3.q * /hive/trunk/ql/src/test/queries/clientpositive/decimal_4.q * /hive/trunk/ql/src/test/queries/clientpositive/decimal_5.q * /hive/trunk/ql/src/test/queries/clientpositive/decimal_6.q * /hive/trunk/ql/src/test/queries/clientpositive/decimal_join.q * /hive/trunk/ql/src/test/queries/clientpositive/decimal_precision.q * /hive/trunk/ql/src/test/queries/clientpositive/decimal_udf.q * /hive/trunk/ql/src/test/queries/clientpositive/orc_predicate_pushdown.q * /hive/trunk/ql/src/test/queries/clientpositive/ptf_decimal.q * /hive/trunk/ql/src/test/queries/clientpositive/serde_regex.q * /hive/trunk/ql/src/test/queries/clientpositive/udf_pmod.q * /hive/trunk/ql/src/test/queries/clientpositive/udf_to_double.q * /hive/trunk/ql/src/test/queries/clientpositive/udf_to_float.q * /hive/trunk/ql/src/test/queries/clientpositive/udf_to_string.q * /hive/trunk/ql/src/test/queries/clientpositive/windowing_expressions.q * /hive/trunk/ql/src/test/queries/clientpositive/windowing_multipartitioning.q * /hive/trunk/ql/src/test/queries/clientpositive/windowing_navfn.q * /hive/trunk/ql/src/test/queries/clientpositive/windowing_ntile.q * /hive/trunk/ql/src/test/queries/clientpositive/windowing_rank.q * /hive/trunk/ql/src/test/results/clientnegative/invalid_cast_from_binary_1.q.out * /hive/trunk/ql/src/test/results/clientnegative/invalid_cast_from_binary_2.q.out * /hive/trunk/ql/src/test/results/clientnegative/invalid_cast_from_binary_3.q.out * /hive/trunk/ql/src/test/results/clientnegative/invalid_cast_from_binary_4.q.out * /hive/trunk/ql/src/test/results/clientnegative/invalid_cast_from_binary_5.q.out * /hive/trunk/ql/src/test/results/clientnegative/invalid_cast_from_binary_6.q.out * /hive/trunk/ql/src/test/results/clientnegative/wrong_column_type.q.out * /hive/trunk/ql/src/test/results/clientpositive/decimal_1.q.out * /hive/trunk/ql/src/test/results/clientpositive/decimal_2.q.out * /hive/trunk/ql/src/test/results/clientpositive/decimal_3.q.out * /hive/trunk/ql/src/test/results/clientpositive/decimal_4.q.out * /hive/trunk/ql/src/test/results/clientpositive/decimal_5.q.out * /hive/trunk/ql/src/test/results/clientpositive/decimal_6.q.out * /hive/trunk/ql/src/test/results/clientpositive/decimal_join.q.out * /hive/trunk/ql/src/test/results/clientpositive/decimal_precision.q.out * /hive/trunk/ql/src/test/results/clientpositive/decimal_serde.q.out * /hive/trunk/ql/src/test/results/clientpositive/decimal_udf.q.out
[jira] [Commented] (HIVE-3976) Support specifying scale and precision with Hive decimal type
[ https://issues.apache.org/jira/browse/HIVE-3976?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13806535#comment-13806535 ] Hudson commented on HIVE-3976: -- FAILURE: Integrated in Hive-trunk-h0.21 #2424 (See [https://builds.apache.org/job/Hive-trunk-h0.21/2424/]) HIVE-3976 - Support specifying scale and precision with Hive decimal type (Xuefu Zhang via Brock Noland) (brock: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1536151) * /hive/trunk/common/src/java/org/apache/hadoop/hive/common/type/HiveDecimal.java * /hive/trunk/common/src/test/org/apache/hadoop/hive/common/type/TestHiveDecimal.java * /hive/trunk/data/files/kv9.txt * /hive/trunk/jdbc/src/java/org/apache/hadoop/hive/jdbc/HiveResultSetMetaData.java * /hive/trunk/jdbc/src/java/org/apache/hadoop/hive/jdbc/Utils.java * /hive/trunk/jdbc/src/test/org/apache/hadoop/hive/jdbc/TestJdbcDriver.java * /hive/trunk/jdbc/src/test/org/apache/hive/jdbc/TestJdbcDriver2.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/Driver.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/FunctionRegistry.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/NumericOpMethodResolver.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/io/orc/OrcStruct.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/io/orc/WriterImpl.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/parse/DDLSemanticAnalyzer.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/parse/HiveParser.g * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/parse/ParseUtils.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/parse/TypeCheckProcFactory.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/UDFOPDivide.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/UDFOPPlus.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFBridge.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFToDecimal.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFToVarchar.java * /hive/trunk/ql/src/test/org/apache/hadoop/hive/ql/exec/TestFunctionRegistry.java * /hive/trunk/ql/src/test/org/apache/hadoop/hive/ql/io/orc/TestOrcFile.java * /hive/trunk/ql/src/test/org/apache/hadoop/hive/ql/parse/TestHiveDecimalParse.java * /hive/trunk/ql/src/test/queries/clientpositive/decimal_1.q * /hive/trunk/ql/src/test/queries/clientpositive/decimal_2.q * /hive/trunk/ql/src/test/queries/clientpositive/decimal_3.q * /hive/trunk/ql/src/test/queries/clientpositive/decimal_4.q * /hive/trunk/ql/src/test/queries/clientpositive/decimal_5.q * /hive/trunk/ql/src/test/queries/clientpositive/decimal_6.q * /hive/trunk/ql/src/test/queries/clientpositive/decimal_join.q * /hive/trunk/ql/src/test/queries/clientpositive/decimal_precision.q * /hive/trunk/ql/src/test/queries/clientpositive/decimal_udf.q * /hive/trunk/ql/src/test/queries/clientpositive/orc_predicate_pushdown.q * /hive/trunk/ql/src/test/queries/clientpositive/ptf_decimal.q * /hive/trunk/ql/src/test/queries/clientpositive/serde_regex.q * /hive/trunk/ql/src/test/queries/clientpositive/udf_pmod.q * /hive/trunk/ql/src/test/queries/clientpositive/udf_to_double.q * /hive/trunk/ql/src/test/queries/clientpositive/udf_to_float.q * /hive/trunk/ql/src/test/queries/clientpositive/udf_to_string.q * /hive/trunk/ql/src/test/queries/clientpositive/windowing_expressions.q * /hive/trunk/ql/src/test/queries/clientpositive/windowing_multipartitioning.q * /hive/trunk/ql/src/test/queries/clientpositive/windowing_navfn.q * /hive/trunk/ql/src/test/queries/clientpositive/windowing_ntile.q * /hive/trunk/ql/src/test/queries/clientpositive/windowing_rank.q * /hive/trunk/ql/src/test/results/clientnegative/invalid_cast_from_binary_1.q.out * /hive/trunk/ql/src/test/results/clientnegative/invalid_cast_from_binary_2.q.out * /hive/trunk/ql/src/test/results/clientnegative/invalid_cast_from_binary_3.q.out * /hive/trunk/ql/src/test/results/clientnegative/invalid_cast_from_binary_4.q.out * /hive/trunk/ql/src/test/results/clientnegative/invalid_cast_from_binary_5.q.out * /hive/trunk/ql/src/test/results/clientnegative/invalid_cast_from_binary_6.q.out * /hive/trunk/ql/src/test/results/clientnegative/wrong_column_type.q.out * /hive/trunk/ql/src/test/results/clientpositive/decimal_1.q.out * /hive/trunk/ql/src/test/results/clientpositive/decimal_2.q.out * /hive/trunk/ql/src/test/results/clientpositive/decimal_3.q.out * /hive/trunk/ql/src/test/results/clientpositive/decimal_4.q.out * /hive/trunk/ql/src/test/results/clientpositive/decimal_5.q.out * /hive/trunk/ql/src/test/results/clientpositive/decimal_6.q.out * /hive/trunk/ql/src/test/results/clientpositive/decimal_join.q.out * /hive/trunk/ql/src/test/results/clientpositive/decimal_precision.q.out * /hive/trunk/ql/src/test/results/clientpositive/decimal_serde.q.out * /hive/trunk/ql/src/test/results/clientpositive/decimal_udf.q.out *
[jira] [Commented] (HIVE-4974) JDBC2 statements and result sets are not able to return their parents
[ https://issues.apache.org/jira/browse/HIVE-4974?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13806536#comment-13806536 ] Hudson commented on HIVE-4974: -- FAILURE: Integrated in Hive-trunk-h0.21 #2424 (See [https://builds.apache.org/job/Hive-trunk-h0.21/2424/]) HIVE-4974 - JDBC2 statements and result sets are not able to return their parents (Chris Dome via Brock Noland) (brock: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1536153) * /hive/trunk/jdbc/src/java/org/apache/hive/jdbc/HiveBaseResultSet.java * /hive/trunk/jdbc/src/java/org/apache/hive/jdbc/HiveCallableStatement.java * /hive/trunk/jdbc/src/java/org/apache/hive/jdbc/HiveConnection.java * /hive/trunk/jdbc/src/java/org/apache/hive/jdbc/HiveDatabaseMetaData.java * /hive/trunk/jdbc/src/java/org/apache/hive/jdbc/HivePreparedStatement.java * /hive/trunk/jdbc/src/java/org/apache/hive/jdbc/HiveQueryResultSet.java * /hive/trunk/jdbc/src/java/org/apache/hive/jdbc/HiveStatement.java * /hive/trunk/jdbc/src/test/org/apache/hive/jdbc/TestJdbcDriver2.java JDBC2 statements and result sets are not able to return their parents - Key: HIVE-4974 URL: https://issues.apache.org/jira/browse/HIVE-4974 Project: Hive Issue Type: Bug Components: JDBC Affects Versions: 0.11.0, 0.12.0 Reporter: Chris Drome Assignee: Chris Drome Priority: Minor Fix For: 0.13.0 Attachments: HIVE-4974.2-trunk.patch.txt, HIVE-4974-trunk-1.patch.txt, HIVE-4974-trunk-2.patch.txt, HIVE-4974-trunk.patch.txt The getConnection method of HiveStatement and HivePreparedStatement throw a not supported SQLException. The constructors should take the HiveConnection that creates them as an argument. Similarly, HiveBaseResultSet is not capable of returning the Statement that created it. -- This message was sent by Atlassian JIRA (v6.1#6144)
[jira] [Commented] (HIVE-3976) Support specifying scale and precision with Hive decimal type
[ https://issues.apache.org/jira/browse/HIVE-3976?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13806578#comment-13806578 ] Hudson commented on HIVE-3976: -- FAILURE: Integrated in Hive-trunk-hadoop2 #524 (See [https://builds.apache.org/job/Hive-trunk-hadoop2/524/]) HIVE-3976 - Support specifying scale and precision with Hive decimal type (Xuefu Zhang via Brock Noland) (brock: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1536151) * /hive/trunk/common/src/java/org/apache/hadoop/hive/common/type/HiveDecimal.java * /hive/trunk/common/src/test/org/apache/hadoop/hive/common/type/TestHiveDecimal.java * /hive/trunk/data/files/kv9.txt * /hive/trunk/jdbc/src/java/org/apache/hadoop/hive/jdbc/HiveResultSetMetaData.java * /hive/trunk/jdbc/src/java/org/apache/hadoop/hive/jdbc/Utils.java * /hive/trunk/jdbc/src/test/org/apache/hadoop/hive/jdbc/TestJdbcDriver.java * /hive/trunk/jdbc/src/test/org/apache/hive/jdbc/TestJdbcDriver2.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/Driver.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/FunctionRegistry.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/NumericOpMethodResolver.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/io/orc/OrcStruct.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/io/orc/WriterImpl.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/parse/DDLSemanticAnalyzer.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/parse/HiveParser.g * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/parse/ParseUtils.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/parse/TypeCheckProcFactory.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/UDFOPDivide.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/UDFOPPlus.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFBridge.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFToDecimal.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFToVarchar.java * /hive/trunk/ql/src/test/org/apache/hadoop/hive/ql/exec/TestFunctionRegistry.java * /hive/trunk/ql/src/test/org/apache/hadoop/hive/ql/io/orc/TestOrcFile.java * /hive/trunk/ql/src/test/org/apache/hadoop/hive/ql/parse/TestHiveDecimalParse.java * /hive/trunk/ql/src/test/queries/clientpositive/decimal_1.q * /hive/trunk/ql/src/test/queries/clientpositive/decimal_2.q * /hive/trunk/ql/src/test/queries/clientpositive/decimal_3.q * /hive/trunk/ql/src/test/queries/clientpositive/decimal_4.q * /hive/trunk/ql/src/test/queries/clientpositive/decimal_5.q * /hive/trunk/ql/src/test/queries/clientpositive/decimal_6.q * /hive/trunk/ql/src/test/queries/clientpositive/decimal_join.q * /hive/trunk/ql/src/test/queries/clientpositive/decimal_precision.q * /hive/trunk/ql/src/test/queries/clientpositive/decimal_udf.q * /hive/trunk/ql/src/test/queries/clientpositive/orc_predicate_pushdown.q * /hive/trunk/ql/src/test/queries/clientpositive/ptf_decimal.q * /hive/trunk/ql/src/test/queries/clientpositive/serde_regex.q * /hive/trunk/ql/src/test/queries/clientpositive/udf_pmod.q * /hive/trunk/ql/src/test/queries/clientpositive/udf_to_double.q * /hive/trunk/ql/src/test/queries/clientpositive/udf_to_float.q * /hive/trunk/ql/src/test/queries/clientpositive/udf_to_string.q * /hive/trunk/ql/src/test/queries/clientpositive/windowing_expressions.q * /hive/trunk/ql/src/test/queries/clientpositive/windowing_multipartitioning.q * /hive/trunk/ql/src/test/queries/clientpositive/windowing_navfn.q * /hive/trunk/ql/src/test/queries/clientpositive/windowing_ntile.q * /hive/trunk/ql/src/test/queries/clientpositive/windowing_rank.q * /hive/trunk/ql/src/test/results/clientnegative/invalid_cast_from_binary_1.q.out * /hive/trunk/ql/src/test/results/clientnegative/invalid_cast_from_binary_2.q.out * /hive/trunk/ql/src/test/results/clientnegative/invalid_cast_from_binary_3.q.out * /hive/trunk/ql/src/test/results/clientnegative/invalid_cast_from_binary_4.q.out * /hive/trunk/ql/src/test/results/clientnegative/invalid_cast_from_binary_5.q.out * /hive/trunk/ql/src/test/results/clientnegative/invalid_cast_from_binary_6.q.out * /hive/trunk/ql/src/test/results/clientnegative/wrong_column_type.q.out * /hive/trunk/ql/src/test/results/clientpositive/decimal_1.q.out * /hive/trunk/ql/src/test/results/clientpositive/decimal_2.q.out * /hive/trunk/ql/src/test/results/clientpositive/decimal_3.q.out * /hive/trunk/ql/src/test/results/clientpositive/decimal_4.q.out * /hive/trunk/ql/src/test/results/clientpositive/decimal_5.q.out * /hive/trunk/ql/src/test/results/clientpositive/decimal_6.q.out * /hive/trunk/ql/src/test/results/clientpositive/decimal_join.q.out * /hive/trunk/ql/src/test/results/clientpositive/decimal_precision.q.out * /hive/trunk/ql/src/test/results/clientpositive/decimal_serde.q.out * /hive/trunk/ql/src/test/results/clientpositive/decimal_udf.q.out *
[jira] [Commented] (HIVE-4974) JDBC2 statements and result sets are not able to return their parents
[ https://issues.apache.org/jira/browse/HIVE-4974?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13806579#comment-13806579 ] Hudson commented on HIVE-4974: -- FAILURE: Integrated in Hive-trunk-hadoop2 #524 (See [https://builds.apache.org/job/Hive-trunk-hadoop2/524/]) HIVE-4974 - JDBC2 statements and result sets are not able to return their parents (Chris Dome via Brock Noland) (brock: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1536153) * /hive/trunk/jdbc/src/java/org/apache/hive/jdbc/HiveBaseResultSet.java * /hive/trunk/jdbc/src/java/org/apache/hive/jdbc/HiveCallableStatement.java * /hive/trunk/jdbc/src/java/org/apache/hive/jdbc/HiveConnection.java * /hive/trunk/jdbc/src/java/org/apache/hive/jdbc/HiveDatabaseMetaData.java * /hive/trunk/jdbc/src/java/org/apache/hive/jdbc/HivePreparedStatement.java * /hive/trunk/jdbc/src/java/org/apache/hive/jdbc/HiveQueryResultSet.java * /hive/trunk/jdbc/src/java/org/apache/hive/jdbc/HiveStatement.java * /hive/trunk/jdbc/src/test/org/apache/hive/jdbc/TestJdbcDriver2.java JDBC2 statements and result sets are not able to return their parents - Key: HIVE-4974 URL: https://issues.apache.org/jira/browse/HIVE-4974 Project: Hive Issue Type: Bug Components: JDBC Affects Versions: 0.11.0, 0.12.0 Reporter: Chris Drome Assignee: Chris Drome Priority: Minor Fix For: 0.13.0 Attachments: HIVE-4974.2-trunk.patch.txt, HIVE-4974-trunk-1.patch.txt, HIVE-4974-trunk-2.patch.txt, HIVE-4974-trunk.patch.txt The getConnection method of HiveStatement and HivePreparedStatement throw a not supported SQLException. The constructors should take the HiveConnection that creates them as an argument. Similarly, HiveBaseResultSet is not capable of returning the Statement that created it. -- This message was sent by Atlassian JIRA (v6.1#6144)
[jira] [Commented] (HIVE-5511) percentComplete returned by job status from WebHCat is null
[ https://issues.apache.org/jira/browse/HIVE-5511?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13806058#comment-13806058 ] Hudson commented on HIVE-5511: -- FAILURE: Integrated in Hive-trunk-h0.21 #2422 (See [https://builds.apache.org/job/Hive-trunk-h0.21/2422/]) HIVE-5511 : percentComplete returned by job status from WebHCat is null (Eugene Koifman via Thejas Nair) (thejas: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1535796) * /hive/trunk/hcatalog/src/test/e2e/templeton/README.txt * /hive/trunk/hcatalog/src/test/e2e/templeton/drivers/TestDriverCurl.pm * /hive/trunk/hcatalog/src/test/e2e/templeton/tests/jobsubmission.conf * /hive/trunk/hcatalog/src/test/e2e/templeton/tests/jobsubmission_streaming.conf * /hive/trunk/hcatalog/webhcat/svr/src/main/bin/webhcat_config.sh * /hive/trunk/hcatalog/webhcat/svr/src/main/config/override-container-log4j.properties * /hive/trunk/hcatalog/webhcat/svr/src/main/java/org/apache/hive/hcatalog/templeton/AppConfig.java * /hive/trunk/hcatalog/webhcat/svr/src/main/java/org/apache/hive/hcatalog/templeton/CompleteDelegator.java * /hive/trunk/hcatalog/webhcat/svr/src/main/java/org/apache/hive/hcatalog/templeton/tool/HDFSStorage.java * /hive/trunk/hcatalog/webhcat/svr/src/main/java/org/apache/hive/hcatalog/templeton/tool/HiveJobIDParser.java * /hive/trunk/hcatalog/webhcat/svr/src/main/java/org/apache/hive/hcatalog/templeton/tool/JarJobIDParser.java * /hive/trunk/hcatalog/webhcat/svr/src/main/java/org/apache/hive/hcatalog/templeton/tool/JobSubmissionConstants.java * /hive/trunk/hcatalog/webhcat/svr/src/main/java/org/apache/hive/hcatalog/templeton/tool/LaunchMapper.java * /hive/trunk/hcatalog/webhcat/svr/src/main/java/org/apache/hive/hcatalog/templeton/tool/PigJobIDParser.java * /hive/trunk/hcatalog/webhcat/svr/src/main/java/org/apache/hive/hcatalog/templeton/tool/TempletonControllerJob.java * /hive/trunk/hcatalog/webhcat/svr/src/main/java/org/apache/hive/hcatalog/templeton/tool/TempletonUtils.java * /hive/trunk/hcatalog/webhcat/svr/src/main/java/org/apache/hive/hcatalog/templeton/tool/TrivialExecService.java * /hive/trunk/hcatalog/webhcat/svr/src/test/java/org/apache/hive/hcatalog/templeton/tool/TestTrivialExecService.java * /hive/trunk/shims/src/0.20S/java/org/apache/hadoop/mapred/WebHCatJTShim20S.java * /hive/trunk/shims/src/0.23/java/org/apache/hadoop/mapred/WebHCatJTShim23.java * /hive/trunk/shims/src/common/java/org/apache/hadoop/hive/shims/HadoopShims.java percentComplete returned by job status from WebHCat is null --- Key: HIVE-5511 URL: https://issues.apache.org/jira/browse/HIVE-5511 Project: Hive Issue Type: Bug Components: WebHCat Affects Versions: 0.12.0 Reporter: Eugene Koifman Assignee: Eugene Koifman Fix For: 0.13.0 Attachments: HIVE-5511.3.patch, HIVE-5511.5.patch In hadoop1 the logging from MR is sent to stderr. In H2, by default, to syslog. templeton.tool.LaunchMapper expects to see the output on stderr to produce 'percentComplete' in job status. -- This message was sent by Atlassian JIRA (v6.1#6144)
[jira] [Commented] (HIVE-5552) Merging of QBJoinTrees doesn't handle filter pushdowns correctly
[ https://issues.apache.org/jira/browse/HIVE-5552?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13806060#comment-13806060 ] Hudson commented on HIVE-5552: -- FAILURE: Integrated in Hive-trunk-h0.21 #2422 (See [https://builds.apache.org/job/Hive-trunk-h0.21/2422/]) HIVE-5552 : Merging of QBJoinTrees doesnt handle filter pushdowns correctly (Harish Butani via Ashutosh Chauhan) (hashutosh: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1535858) * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/parse/ParseUtils.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/parse/QBJoinTree.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/parse/SemanticAnalyzer.java * /hive/trunk/ql/src/test/queries/clientpositive/join_merging.q * /hive/trunk/ql/src/test/results/clientpositive/join_merging.q.out Merging of QBJoinTrees doesn't handle filter pushdowns correctly Key: HIVE-5552 URL: https://issues.apache.org/jira/browse/HIVE-5552 Project: Hive Issue Type: Bug Components: Query Processor Reporter: Harish Butani Assignee: Harish Butani Fix For: 0.13.0 Attachments: HIVE-5552.1.patch, HIVE-5552.2.patch The following query fails: (this based on the schema from auto_join_filters.q) {noformat} explain SELECT sum(hash(a.key,a.value,b.key,b.value)) FROM myinput1 a LEFT OUTER JOIN myinput1 b on a.value = b.value RIGHT OUTER JOIN myinput1 c ON b.value = c.value AND a.key 40 {noformat} Whereas this query succeeds {noformat} explain SELECT sum(hash(a.key,a.value,b.key,b.value)) FROM myinput1 a LEFT OUTER JOIN myinput1 b RIGHT OUTER JOIN myinput1 c ON a.value = b.value and b.value = c.value AND a.key 40 {noformat} Pushing the first condition to the first join, triggers a merge of the 2 QBJoinTrees. During merge all the right side filters identified for pushing are assumed to refer to the merging table (b in this case). But the pushable filters can refer to any left table. -- This message was sent by Atlassian JIRA (v6.1#6144)
[jira] [Commented] (HIVE-5440) HiveServer2 doesn't apply SQL operation's config property
[ https://issues.apache.org/jira/browse/HIVE-5440?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13806059#comment-13806059 ] Hudson commented on HIVE-5440: -- FAILURE: Integrated in Hive-trunk-h0.21 #2422 (See [https://builds.apache.org/job/Hive-trunk-h0.21/2422/]) HIVE-5440: HiveServer2 doesn't apply SQL operation's config property (Prasad Mujumdar via Thejas Nair) (thejas: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1535889) * /hive/trunk/service/src/java/org/apache/hive/service/cli/operation/ExecuteStatementOperation.java * /hive/trunk/service/src/java/org/apache/hive/service/cli/operation/SQLOperation.java * /hive/trunk/service/src/test/org/apache/hive/service/cli/CLIServiceTest.java HiveServer2 doesn't apply SQL operation's config property -- Key: HIVE-5440 URL: https://issues.apache.org/jira/browse/HIVE-5440 Project: Hive Issue Type: Bug Components: HiveServer2 Affects Versions: 0.12.0 Reporter: Prasad Mujumdar Assignee: Prasad Mujumdar Fix For: 0.13.0 Attachments: HIVE-5440.1.patch, HIVE-5440.2.patch The HiveServer2 thrift IDL includes an optional config overlay map which is currently not used. -- This message was sent by Atlassian JIRA (v6.1#6144)
[jira] [Commented] (HIVE-5552) Merging of QBJoinTrees doesn't handle filter pushdowns correctly
[ https://issues.apache.org/jira/browse/HIVE-5552?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13806129#comment-13806129 ] Hudson commented on HIVE-5552: -- ABORTED: Integrated in Hive-trunk-hadoop2 #522 (See [https://builds.apache.org/job/Hive-trunk-hadoop2/522/]) HIVE-5552 : Merging of QBJoinTrees doesnt handle filter pushdowns correctly (Harish Butani via Ashutosh Chauhan) (hashutosh: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1535858) * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/parse/ParseUtils.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/parse/QBJoinTree.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/parse/SemanticAnalyzer.java * /hive/trunk/ql/src/test/queries/clientpositive/join_merging.q * /hive/trunk/ql/src/test/results/clientpositive/join_merging.q.out Merging of QBJoinTrees doesn't handle filter pushdowns correctly Key: HIVE-5552 URL: https://issues.apache.org/jira/browse/HIVE-5552 Project: Hive Issue Type: Bug Components: Query Processor Reporter: Harish Butani Assignee: Harish Butani Fix For: 0.13.0 Attachments: HIVE-5552.1.patch, HIVE-5552.2.patch The following query fails: (this based on the schema from auto_join_filters.q) {noformat} explain SELECT sum(hash(a.key,a.value,b.key,b.value)) FROM myinput1 a LEFT OUTER JOIN myinput1 b on a.value = b.value RIGHT OUTER JOIN myinput1 c ON b.value = c.value AND a.key 40 {noformat} Whereas this query succeeds {noformat} explain SELECT sum(hash(a.key,a.value,b.key,b.value)) FROM myinput1 a LEFT OUTER JOIN myinput1 b RIGHT OUTER JOIN myinput1 c ON a.value = b.value and b.value = c.value AND a.key 40 {noformat} Pushing the first condition to the first join, triggers a merge of the 2 QBJoinTrees. During merge all the right side filters identified for pushing are assumed to refer to the merging table (b in this case). But the pushable filters can refer to any left table. -- This message was sent by Atlassian JIRA (v6.1#6144)
[jira] [Commented] (HIVE-5511) percentComplete returned by job status from WebHCat is null
[ https://issues.apache.org/jira/browse/HIVE-5511?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13806127#comment-13806127 ] Hudson commented on HIVE-5511: -- ABORTED: Integrated in Hive-trunk-hadoop2 #522 (See [https://builds.apache.org/job/Hive-trunk-hadoop2/522/]) HIVE-5511 : percentComplete returned by job status from WebHCat is null (Eugene Koifman via Thejas Nair) (thejas: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1535796) * /hive/trunk/hcatalog/src/test/e2e/templeton/README.txt * /hive/trunk/hcatalog/src/test/e2e/templeton/drivers/TestDriverCurl.pm * /hive/trunk/hcatalog/src/test/e2e/templeton/tests/jobsubmission.conf * /hive/trunk/hcatalog/src/test/e2e/templeton/tests/jobsubmission_streaming.conf * /hive/trunk/hcatalog/webhcat/svr/src/main/bin/webhcat_config.sh * /hive/trunk/hcatalog/webhcat/svr/src/main/config/override-container-log4j.properties * /hive/trunk/hcatalog/webhcat/svr/src/main/java/org/apache/hive/hcatalog/templeton/AppConfig.java * /hive/trunk/hcatalog/webhcat/svr/src/main/java/org/apache/hive/hcatalog/templeton/CompleteDelegator.java * /hive/trunk/hcatalog/webhcat/svr/src/main/java/org/apache/hive/hcatalog/templeton/tool/HDFSStorage.java * /hive/trunk/hcatalog/webhcat/svr/src/main/java/org/apache/hive/hcatalog/templeton/tool/HiveJobIDParser.java * /hive/trunk/hcatalog/webhcat/svr/src/main/java/org/apache/hive/hcatalog/templeton/tool/JarJobIDParser.java * /hive/trunk/hcatalog/webhcat/svr/src/main/java/org/apache/hive/hcatalog/templeton/tool/JobSubmissionConstants.java * /hive/trunk/hcatalog/webhcat/svr/src/main/java/org/apache/hive/hcatalog/templeton/tool/LaunchMapper.java * /hive/trunk/hcatalog/webhcat/svr/src/main/java/org/apache/hive/hcatalog/templeton/tool/PigJobIDParser.java * /hive/trunk/hcatalog/webhcat/svr/src/main/java/org/apache/hive/hcatalog/templeton/tool/TempletonControllerJob.java * /hive/trunk/hcatalog/webhcat/svr/src/main/java/org/apache/hive/hcatalog/templeton/tool/TempletonUtils.java * /hive/trunk/hcatalog/webhcat/svr/src/main/java/org/apache/hive/hcatalog/templeton/tool/TrivialExecService.java * /hive/trunk/hcatalog/webhcat/svr/src/test/java/org/apache/hive/hcatalog/templeton/tool/TestTrivialExecService.java * /hive/trunk/shims/src/0.20S/java/org/apache/hadoop/mapred/WebHCatJTShim20S.java * /hive/trunk/shims/src/0.23/java/org/apache/hadoop/mapred/WebHCatJTShim23.java * /hive/trunk/shims/src/common/java/org/apache/hadoop/hive/shims/HadoopShims.java percentComplete returned by job status from WebHCat is null --- Key: HIVE-5511 URL: https://issues.apache.org/jira/browse/HIVE-5511 Project: Hive Issue Type: Bug Components: WebHCat Affects Versions: 0.12.0 Reporter: Eugene Koifman Assignee: Eugene Koifman Fix For: 0.13.0 Attachments: HIVE-5511.3.patch, HIVE-5511.5.patch In hadoop1 the logging from MR is sent to stderr. In H2, by default, to syslog. templeton.tool.LaunchMapper expects to see the output on stderr to produce 'percentComplete' in job status. -- This message was sent by Atlassian JIRA (v6.1#6144)
[jira] [Commented] (HIVE-5440) HiveServer2 doesn't apply SQL operation's config property
[ https://issues.apache.org/jira/browse/HIVE-5440?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13806128#comment-13806128 ] Hudson commented on HIVE-5440: -- ABORTED: Integrated in Hive-trunk-hadoop2 #522 (See [https://builds.apache.org/job/Hive-trunk-hadoop2/522/]) HIVE-5440: HiveServer2 doesn't apply SQL operation's config property (Prasad Mujumdar via Thejas Nair) (thejas: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1535889) * /hive/trunk/service/src/java/org/apache/hive/service/cli/operation/ExecuteStatementOperation.java * /hive/trunk/service/src/java/org/apache/hive/service/cli/operation/SQLOperation.java * /hive/trunk/service/src/test/org/apache/hive/service/cli/CLIServiceTest.java HiveServer2 doesn't apply SQL operation's config property -- Key: HIVE-5440 URL: https://issues.apache.org/jira/browse/HIVE-5440 Project: Hive Issue Type: Bug Components: HiveServer2 Affects Versions: 0.12.0 Reporter: Prasad Mujumdar Assignee: Prasad Mujumdar Fix For: 0.13.0 Attachments: HIVE-5440.1.patch, HIVE-5440.2.patch The HiveServer2 thrift IDL includes an optional config overlay map which is currently not used. -- This message was sent by Atlassian JIRA (v6.1#6144)
[jira] [Commented] (HIVE-5628) ListBucketingPrunnerTest and DynamicMultiDimeCollectionTest should start with Test not end with it
[ https://issues.apache.org/jira/browse/HIVE-5628?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13806126#comment-13806126 ] Hudson commented on HIVE-5628: -- ABORTED: Integrated in Hive-trunk-hadoop2 #522 (See [https://builds.apache.org/job/Hive-trunk-hadoop2/522/]) HIVE-5628 : ListBucketingPrunnerTest and DynamicMultiDimeCollectionTest should start with Test not end with it (Brock Noland via Ashutosh Chauhan) (hashutosh: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1535761) * /hive/trunk/ql/src/test/org/apache/hadoop/hive/ql/optimizer/listbucketingpruner/DynamicMultiDimeCollectionTest.java * /hive/trunk/ql/src/test/org/apache/hadoop/hive/ql/optimizer/listbucketingpruner/ListBucketingPrunnerTest.java * /hive/trunk/ql/src/test/org/apache/hadoop/hive/ql/optimizer/listbucketingpruner/TestDynamicMultiDimeCollection.java * /hive/trunk/ql/src/test/org/apache/hadoop/hive/ql/optimizer/listbucketingpruner/TestListBucketingPrunner.java ListBucketingPrunnerTest and DynamicMultiDimeCollectionTest should start with Test not end with it -- Key: HIVE-5628 URL: https://issues.apache.org/jira/browse/HIVE-5628 Project: Hive Issue Type: Bug Reporter: Brock Noland Assignee: Brock Noland Fix For: 0.13.0 Attachments: HIVE-5628.patch ListBucketingPrunnerTest and DynamicMultiDimeCollectionTest will not be run by PTest because they end with Test and PTest requires tests start with Test. -- This message was sent by Atlassian JIRA (v6.1#6144)
[jira] [Commented] (HIVE-5625) Fix issue with metastore version restriction test.
[ https://issues.apache.org/jira/browse/HIVE-5625?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13805079#comment-13805079 ] Hudson commented on HIVE-5625: -- ABORTED: Integrated in Hive-trunk-hadoop2 #520 (See [https://builds.apache.org/job/Hive-trunk-hadoop2/520/]) HIVE-5625 - Fix issue with metastore version restriction test. (Vikram Dixit K via Brock Noland) (brock: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1535402) * /hive/trunk/metastore/src/test/org/apache/hadoop/hive/metastore/TestMetastoreVersion.java Fix issue with metastore version restriction test. -- Key: HIVE-5625 URL: https://issues.apache.org/jira/browse/HIVE-5625 Project: Hive Issue Type: Bug Components: Tests Affects Versions: 0.13.0 Reporter: Vikram Dixit K Assignee: Vikram Dixit K Fix For: 0.13.0 Attachments: HIVE-5625.1.patch Based on Brock's comments, the change made in HIVE-5403 change the nature of the test. -- This message was sent by Atlassian JIRA (v6.1#6144)
[jira] [Commented] (HIVE-5216) Need to annotate public API in HCatalog
[ https://issues.apache.org/jira/browse/HIVE-5216?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13805383#comment-13805383 ] Hudson commented on HIVE-5216: -- FAILURE: Integrated in Hive-trunk-h0.21 #2420 (See [https://builds.apache.org/job/Hive-trunk-h0.21/2420/]) HIVE-5216 : Need to annotate public API in HCatalog (Eugene Koifman via Thejas Nair) (thejas: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1535535) * /hive/trunk/hcatalog/core/src/main/java/org/apache/hive/hcatalog/data/HCatRecord.java * /hive/trunk/hcatalog/core/src/main/java/org/apache/hive/hcatalog/data/schema/HCatSchema.java * /hive/trunk/hcatalog/core/src/main/java/org/apache/hive/hcatalog/data/transfer/HCatReader.java * /hive/trunk/hcatalog/core/src/main/java/org/apache/hive/hcatalog/data/transfer/HCatWriter.java * /hive/trunk/hcatalog/core/src/main/java/org/apache/hive/hcatalog/mapreduce/HCatOutputFormat.java * /hive/trunk/hcatalog/hcatalog-pig-adapter/src/main/java/org/apache/hive/hcatalog/pig/HCatLoader.java * /hive/trunk/hcatalog/hcatalog-pig-adapter/src/main/java/org/apache/hive/hcatalog/pig/HCatStorer.java * /hive/trunk/hcatalog/webhcat/svr/src/main/java/org/apache/hive/hcatalog/templeton/Main.java Need to annotate public API in HCatalog --- Key: HIVE-5216 URL: https://issues.apache.org/jira/browse/HIVE-5216 Project: Hive Issue Type: Bug Components: HCatalog, WebHCat Affects Versions: 0.12.0 Reporter: Eugene Koifman Assignee: Eugene Koifman Fix For: 0.13.0 Attachments: HIVE-5216.2.patch, HIVE-5216.patch need to annotate which API is considered public using something like @InterfaceAudience.Public @InterfaceStability.Evolving Currently this is what is considered (at a minimum) public API HCatLoader HCatStorer HCatInputFormat HCatOutputFormat HCatReader HCatWriter HCatRecord HCatSchema This is needed so that clients/dependent projects know which API they can rely on and which can change w/o notice. -- This message was sent by Atlassian JIRA (v6.1#6144)
[jira] [Commented] (HIVE-5629) Fix two javadoc failures in HCatalog
[ https://issues.apache.org/jira/browse/HIVE-5629?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13805378#comment-13805378 ] Hudson commented on HIVE-5629: -- FAILURE: Integrated in Hive-trunk-h0.21 #2420 (See [https://builds.apache.org/job/Hive-trunk-h0.21/2420/]) HIVE-5629 : Fix two javadoc failures in HCatalog (Brock Noland via Ashutosh Chauhan) (hashutosh: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1535513) * /hive/trunk/hcatalog/core/src/main/java/org/apache/hive/hcatalog/mapreduce/InputJobInfo.java * /hive/trunk/hcatalog/webhcat/svr/src/main/java/org/apache/hive/hcatalog/templeton/PigDelegator.java Fix two javadoc failures in HCatalog Key: HIVE-5629 URL: https://issues.apache.org/jira/browse/HIVE-5629 Project: Hive Issue Type: Bug Reporter: Brock Noland Assignee: Brock Noland Fix For: 0.13.0 Attachments: HIVE-5629.patch I am seeing two javadoc failures on HCatalog. These are not being seen by PTest and indeed I cannot reproduce on my Mac but can on Linux. Regardless they should be fixed. -- This message was sent by Atlassian JIRA (v6.1#6144)
[jira] [Commented] (HIVE-5482) JDBC should depend on httpclient.version and httpcore.version 4.1.3 to be consistent with other modules
[ https://issues.apache.org/jira/browse/HIVE-5482?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13805381#comment-13805381 ] Hudson commented on HIVE-5482: -- FAILURE: Integrated in Hive-trunk-h0.21 #2420 (See [https://builds.apache.org/job/Hive-trunk-h0.21/2420/]) HIVE-5482 : JDBC should depend on httpclient.version and httpcore.version 4.1.3 to be consistent with other modules (Vaibhav Gumashta via Thejas Nair) (thejas: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1535620) * /hive/trunk/ivy/libraries.properties JDBC should depend on httpclient.version and httpcore.version 4.1.3 to be consistent with other modules --- Key: HIVE-5482 URL: https://issues.apache.org/jira/browse/HIVE-5482 Project: Hive Issue Type: Bug Components: HiveServer2, JDBC Reporter: Vaibhav Gumashta Assignee: Vaibhav Gumashta Fix For: 0.13.0 Attachments: HIVE-5482.1.patch Currently depends on 4.2.4 and 4.2.5 which conflicts with thrift-0.9 which depends on 4.1.3 -- This message was sent by Atlassian JIRA (v6.1#6144)
[jira] [Commented] (HIVE-5637) Sporadic minimr test failure
[ https://issues.apache.org/jira/browse/HIVE-5637?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13805382#comment-13805382 ] Hudson commented on HIVE-5637: -- FAILURE: Integrated in Hive-trunk-h0.21 #2420 (See [https://builds.apache.org/job/Hive-trunk-h0.21/2420/]) HIVE-5637 : Sporadic minimr test failure (Navis via Ashutosh Chauhan) (hashutosh: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1535510) * /hive/trunk/ql/src/test/queries/clientpositive/import_exported_table.q Sporadic minimr test failure Key: HIVE-5637 URL: https://issues.apache.org/jira/browse/HIVE-5637 Project: Hive Issue Type: Test Components: Tests Reporter: Navis Assignee: Navis Priority: Trivial Fix For: 0.13.0 Attachments: HIVE-5637.1.patch.txt {noformat} ant test -Dtestcase=TestMinimrCliDriver -Dqfile=import_exported_table.q,load_hdfs_file_with_space_in_the_name.q -Dmodule=ql {noformat} Fails with message like this. {noformat} Begin query: load_hdfs_file_with_space_in_the_name.q mkdir: cannot create directory hdfs:///tmp/test/: File exists Exception: Client Execution failed with error code = -1 See build/ql/tmp/hive.log, or try ant test ... -Dtest.silent=false to get more logs. junit.framework.AssertionFailedError: Client Execution failed with error code = -1 See build/ql/tmp/hive.log, or try ant test ... -Dtest.silent=false to get more logs. at junit.framework.Assert.fail(Assert.java:47) at org.apache.hadoop.hive.cli.TestMinimrCliDriver.runTest(TestMinimrCliDriver.java:349) at org.apache.hadoop.hive.cli.TestMinimrCliDriver.testCliDriver_load_hdfs_file_with_space_in_the_name(TestMinimrCliDriver.java:291) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25) at java.lang.reflect.Method.invoke(Method.java:597) at junit.framework.TestCase.runTest(TestCase.java:154) at junit.framework.TestCase.runBare(TestCase.java:127) at junit.framework.TestResult$1.protect(TestResult.java:106) at junit.framework.TestResult.runProtected(TestResult.java:124) at junit.framework.TestResult.run(TestResult.java:109) at junit.framework.TestCase.run(TestCase.java:118) at junit.framework.TestSuite.runTest(TestSuite.java:208) at junit.framework.TestSuite.run(TestSuite.java:203) at org.apache.tools.ant.taskdefs.optional.junit.JUnitTestRunner.run(JUnitTestRunner.java:520) at org.apache.tools.ant.taskdefs.optional.junit.JUnitTestRunner.launch(JUnitTestRunner.java:1060) at org.apache.tools.ant.taskdefs.optional.junit.JUnitTestRunner.main(JUnitTestRunner.java:911) {noformat} -- This message was sent by Atlassian JIRA (v6.1#6144)
[jira] [Commented] (HIVE-5619) Allow concat() to accept mixed string/binary args
[ https://issues.apache.org/jira/browse/HIVE-5619?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13805377#comment-13805377 ] Hudson commented on HIVE-5619: -- FAILURE: Integrated in Hive-trunk-h0.21 #2420 (See [https://builds.apache.org/job/Hive-trunk-h0.21/2420/]) HIVE-5619 : Allow concat() to accept mixed string/binary args (Jason Dere via Ashutosh Chauhan) (hashutosh: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1535621) * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFConcat.java * /hive/trunk/ql/src/test/queries/clientpositive/udf_concat.q * /hive/trunk/ql/src/test/results/clientpositive/udf_concat.q.out * /hive/trunk/serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/primitive/PrimitiveObjectInspectorUtils.java Allow concat() to accept mixed string/binary args - Key: HIVE-5619 URL: https://issues.apache.org/jira/browse/HIVE-5619 Project: Hive Issue Type: Improvement Reporter: Jason Dere Assignee: Jason Dere Fix For: 0.13.0 Attachments: HIVE-5619.1.patch concat() is currently strict about allowing either all binary or all non-binary arguments. Loosen this to permit mixed params. -- This message was sent by Atlassian JIRA (v6.1#6144)
[jira] [Commented] (HIVE-5577) Remove TestNegativeCliDriver script_broken_pipe1
[ https://issues.apache.org/jira/browse/HIVE-5577?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13805379#comment-13805379 ] Hudson commented on HIVE-5577: -- FAILURE: Integrated in Hive-trunk-h0.21 #2420 (See [https://builds.apache.org/job/Hive-trunk-h0.21/2420/]) HIVE-5577 : Remove TestNegativeCliDriver script_broken_pipe1 (Brock Noland via Navis) (navis: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1535643) * /hive/trunk/ql/src/test/queries/clientnegative/script_broken_pipe1.q * /hive/trunk/ql/src/test/results/clientnegative/script_broken_pipe1.q.out Remove TestNegativeCliDriver script_broken_pipe1 Key: HIVE-5577 URL: https://issues.apache.org/jira/browse/HIVE-5577 Project: Hive Issue Type: Bug Reporter: Brock Noland Assignee: Brock Noland Fix For: 0.13.0 Attachments: HIVE-5577.1.patch.txt TestNegativeCliDriver script_broken_pipe1 is extremely flaky and not a terribly important test. Let's remove it. Failures https://builds.apache.org/user/brock/my-views/view/hive/job/Hive-trunk-hadoop1-ptest/206/testReport/org.apache.hadoop.hive.cli/TestNegativeCliDriver/testNegativeCliDriver_script_broken_pipe1/ https://builds.apache.org/user/brock/my-views/view/hive/job/Hive-trunk-hadoop1-ptest/206/testReport/junit/org.apache.hadoop.hive.cli/TestNegativeCliDriver/testNegativeCliDriver_script_broken_pipe1/ https://builds.apache.org/user/brock/my-views/view/hive/job/Hive-trunk-hadoop1-ptest/204/testReport/org.apache.hadoop.hive.cli/TestNegativeCliDriver/testNegativeCliDriver_script_broken_pipe1/ -- This message was sent by Atlassian JIRA (v6.1#6144)
[jira] [Commented] (HIVE-5628) ListBucketingPrunnerTest and DynamicMultiDimeCollectionTest should start with Test not end with it
[ https://issues.apache.org/jira/browse/HIVE-5628?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13805722#comment-13805722 ] Hudson commented on HIVE-5628: -- FAILURE: Integrated in Hive-trunk-h0.21 #2421 (See [https://builds.apache.org/job/Hive-trunk-h0.21/2421/]) HIVE-5628 : ListBucketingPrunnerTest and DynamicMultiDimeCollectionTest should start with Test not end with it (Brock Noland via Ashutosh Chauhan) (hashutosh: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1535761) * /hive/trunk/ql/src/test/org/apache/hadoop/hive/ql/optimizer/listbucketingpruner/DynamicMultiDimeCollectionTest.java * /hive/trunk/ql/src/test/org/apache/hadoop/hive/ql/optimizer/listbucketingpruner/ListBucketingPrunnerTest.java * /hive/trunk/ql/src/test/org/apache/hadoop/hive/ql/optimizer/listbucketingpruner/TestDynamicMultiDimeCollection.java * /hive/trunk/ql/src/test/org/apache/hadoop/hive/ql/optimizer/listbucketingpruner/TestListBucketingPrunner.java ListBucketingPrunnerTest and DynamicMultiDimeCollectionTest should start with Test not end with it -- Key: HIVE-5628 URL: https://issues.apache.org/jira/browse/HIVE-5628 Project: Hive Issue Type: Bug Reporter: Brock Noland Assignee: Brock Noland Fix For: 0.13.0 Attachments: HIVE-5628.patch ListBucketingPrunnerTest and DynamicMultiDimeCollectionTest will not be run by PTest because they end with Test and PTest requires tests start with Test. -- This message was sent by Atlassian JIRA (v6.1#6144)
[jira] [Commented] (HIVE-5577) Remove TestNegativeCliDriver script_broken_pipe1
[ https://issues.apache.org/jira/browse/HIVE-5577?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13805820#comment-13805820 ] Hudson commented on HIVE-5577: -- ABORTED: Integrated in Hive-trunk-hadoop2 #521 (See [https://builds.apache.org/job/Hive-trunk-hadoop2/521/]) HIVE-5577 : Remove TestNegativeCliDriver script_broken_pipe1 (Brock Noland via Navis) (navis: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1535643) * /hive/trunk/ql/src/test/queries/clientnegative/script_broken_pipe1.q * /hive/trunk/ql/src/test/results/clientnegative/script_broken_pipe1.q.out Remove TestNegativeCliDriver script_broken_pipe1 Key: HIVE-5577 URL: https://issues.apache.org/jira/browse/HIVE-5577 Project: Hive Issue Type: Bug Reporter: Brock Noland Assignee: Brock Noland Fix For: 0.13.0 Attachments: HIVE-5577.1.patch.txt TestNegativeCliDriver script_broken_pipe1 is extremely flaky and not a terribly important test. Let's remove it. Failures https://builds.apache.org/user/brock/my-views/view/hive/job/Hive-trunk-hadoop1-ptest/206/testReport/org.apache.hadoop.hive.cli/TestNegativeCliDriver/testNegativeCliDriver_script_broken_pipe1/ https://builds.apache.org/user/brock/my-views/view/hive/job/Hive-trunk-hadoop1-ptest/206/testReport/junit/org.apache.hadoop.hive.cli/TestNegativeCliDriver/testNegativeCliDriver_script_broken_pipe1/ https://builds.apache.org/user/brock/my-views/view/hive/job/Hive-trunk-hadoop1-ptest/204/testReport/org.apache.hadoop.hive.cli/TestNegativeCliDriver/testNegativeCliDriver_script_broken_pipe1/ -- This message was sent by Atlassian JIRA (v6.1#6144)
[jira] [Commented] (HIVE-5403) Move loading of filesystem, ugi, metastore client to hive session
[ https://issues.apache.org/jira/browse/HIVE-5403?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13805821#comment-13805821 ] Hudson commented on HIVE-5403: -- ABORTED: Integrated in Hive-trunk-hadoop2 #521 (See [https://builds.apache.org/job/Hive-trunk-hadoop2/521/]) HIVE-5403: Perflogger broken due to HIVE-5403 (Vikram Dixit K via Gunther Hagleitner) (gunther: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1535598) * /hive/trunk/metastore/src/test/org/apache/hadoop/hive/metastore/TestMetastoreVersion.java * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/session/SessionState.java Move loading of filesystem, ugi, metastore client to hive session - Key: HIVE-5403 URL: https://issues.apache.org/jira/browse/HIVE-5403 Project: Hive Issue Type: Bug Affects Versions: 0.13.0 Reporter: Vikram Dixit K Assignee: Vikram Dixit K Fix For: 0.13.0 Attachments: HIVE-5403.1.patch, HIVE-5403.2.patch, HIVE-5403.3.patch, HIVE-5403.4.patch As part of HIVE-5184, the metastore connection, loading filesystem were done as part of the tez session so as to speed up query times while paying a cost at startup. We can do this more generally in hive to apply to both the mapreduce and tez side of things. -- This message was sent by Atlassian JIRA (v6.1#6144)
[jira] [Commented] (HIVE-5629) Fix two javadoc failures in HCatalog
[ https://issues.apache.org/jira/browse/HIVE-5629?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13805819#comment-13805819 ] Hudson commented on HIVE-5629: -- ABORTED: Integrated in Hive-trunk-hadoop2 #521 (See [https://builds.apache.org/job/Hive-trunk-hadoop2/521/]) HIVE-5629 : Fix two javadoc failures in HCatalog (Brock Noland via Ashutosh Chauhan) (hashutosh: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1535513) * /hive/trunk/hcatalog/core/src/main/java/org/apache/hive/hcatalog/mapreduce/InputJobInfo.java * /hive/trunk/hcatalog/webhcat/svr/src/main/java/org/apache/hive/hcatalog/templeton/PigDelegator.java Fix two javadoc failures in HCatalog Key: HIVE-5629 URL: https://issues.apache.org/jira/browse/HIVE-5629 Project: Hive Issue Type: Bug Reporter: Brock Noland Assignee: Brock Noland Fix For: 0.13.0 Attachments: HIVE-5629.patch I am seeing two javadoc failures on HCatalog. These are not being seen by PTest and indeed I cannot reproduce on my Mac but can on Linux. Regardless they should be fixed. -- This message was sent by Atlassian JIRA (v6.1#6144)
[jira] [Commented] (HIVE-5216) Need to annotate public API in HCatalog
[ https://issues.apache.org/jira/browse/HIVE-5216?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13805824#comment-13805824 ] Hudson commented on HIVE-5216: -- ABORTED: Integrated in Hive-trunk-hadoop2 #521 (See [https://builds.apache.org/job/Hive-trunk-hadoop2/521/]) HIVE-5216 : Need to annotate public API in HCatalog (Eugene Koifman via Thejas Nair) (thejas: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1535535) * /hive/trunk/hcatalog/core/src/main/java/org/apache/hive/hcatalog/data/HCatRecord.java * /hive/trunk/hcatalog/core/src/main/java/org/apache/hive/hcatalog/data/schema/HCatSchema.java * /hive/trunk/hcatalog/core/src/main/java/org/apache/hive/hcatalog/data/transfer/HCatReader.java * /hive/trunk/hcatalog/core/src/main/java/org/apache/hive/hcatalog/data/transfer/HCatWriter.java * /hive/trunk/hcatalog/core/src/main/java/org/apache/hive/hcatalog/mapreduce/HCatOutputFormat.java * /hive/trunk/hcatalog/hcatalog-pig-adapter/src/main/java/org/apache/hive/hcatalog/pig/HCatLoader.java * /hive/trunk/hcatalog/hcatalog-pig-adapter/src/main/java/org/apache/hive/hcatalog/pig/HCatStorer.java * /hive/trunk/hcatalog/webhcat/svr/src/main/java/org/apache/hive/hcatalog/templeton/Main.java Need to annotate public API in HCatalog --- Key: HIVE-5216 URL: https://issues.apache.org/jira/browse/HIVE-5216 Project: Hive Issue Type: Bug Components: HCatalog, WebHCat Affects Versions: 0.12.0 Reporter: Eugene Koifman Assignee: Eugene Koifman Fix For: 0.13.0 Attachments: HIVE-5216.2.patch, HIVE-5216.patch need to annotate which API is considered public using something like @InterfaceAudience.Public @InterfaceStability.Evolving Currently this is what is considered (at a minimum) public API HCatLoader HCatStorer HCatInputFormat HCatOutputFormat HCatReader HCatWriter HCatRecord HCatSchema This is needed so that clients/dependent projects know which API they can rely on and which can change w/o notice. -- This message was sent by Atlassian JIRA (v6.1#6144)
[jira] [Commented] (HIVE-5619) Allow concat() to accept mixed string/binary args
[ https://issues.apache.org/jira/browse/HIVE-5619?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13805818#comment-13805818 ] Hudson commented on HIVE-5619: -- ABORTED: Integrated in Hive-trunk-hadoop2 #521 (See [https://builds.apache.org/job/Hive-trunk-hadoop2/521/]) HIVE-5619 : Allow concat() to accept mixed string/binary args (Jason Dere via Ashutosh Chauhan) (hashutosh: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1535621) * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFConcat.java * /hive/trunk/ql/src/test/queries/clientpositive/udf_concat.q * /hive/trunk/ql/src/test/results/clientpositive/udf_concat.q.out * /hive/trunk/serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/primitive/PrimitiveObjectInspectorUtils.java Allow concat() to accept mixed string/binary args - Key: HIVE-5619 URL: https://issues.apache.org/jira/browse/HIVE-5619 Project: Hive Issue Type: Improvement Reporter: Jason Dere Assignee: Jason Dere Fix For: 0.13.0 Attachments: HIVE-5619.1.patch concat() is currently strict about allowing either all binary or all non-binary arguments. Loosen this to permit mixed params. -- This message was sent by Atlassian JIRA (v6.1#6144)
[jira] [Commented] (HIVE-5637) Sporadic minimr test failure
[ https://issues.apache.org/jira/browse/HIVE-5637?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13805823#comment-13805823 ] Hudson commented on HIVE-5637: -- ABORTED: Integrated in Hive-trunk-hadoop2 #521 (See [https://builds.apache.org/job/Hive-trunk-hadoop2/521/]) HIVE-5637 : Sporadic minimr test failure (Navis via Ashutosh Chauhan) (hashutosh: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1535510) * /hive/trunk/ql/src/test/queries/clientpositive/import_exported_table.q Sporadic minimr test failure Key: HIVE-5637 URL: https://issues.apache.org/jira/browse/HIVE-5637 Project: Hive Issue Type: Test Components: Tests Reporter: Navis Assignee: Navis Priority: Trivial Fix For: 0.13.0 Attachments: HIVE-5637.1.patch.txt {noformat} ant test -Dtestcase=TestMinimrCliDriver -Dqfile=import_exported_table.q,load_hdfs_file_with_space_in_the_name.q -Dmodule=ql {noformat} Fails with message like this. {noformat} Begin query: load_hdfs_file_with_space_in_the_name.q mkdir: cannot create directory hdfs:///tmp/test/: File exists Exception: Client Execution failed with error code = -1 See build/ql/tmp/hive.log, or try ant test ... -Dtest.silent=false to get more logs. junit.framework.AssertionFailedError: Client Execution failed with error code = -1 See build/ql/tmp/hive.log, or try ant test ... -Dtest.silent=false to get more logs. at junit.framework.Assert.fail(Assert.java:47) at org.apache.hadoop.hive.cli.TestMinimrCliDriver.runTest(TestMinimrCliDriver.java:349) at org.apache.hadoop.hive.cli.TestMinimrCliDriver.testCliDriver_load_hdfs_file_with_space_in_the_name(TestMinimrCliDriver.java:291) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25) at java.lang.reflect.Method.invoke(Method.java:597) at junit.framework.TestCase.runTest(TestCase.java:154) at junit.framework.TestCase.runBare(TestCase.java:127) at junit.framework.TestResult$1.protect(TestResult.java:106) at junit.framework.TestResult.runProtected(TestResult.java:124) at junit.framework.TestResult.run(TestResult.java:109) at junit.framework.TestCase.run(TestCase.java:118) at junit.framework.TestSuite.runTest(TestSuite.java:208) at junit.framework.TestSuite.run(TestSuite.java:203) at org.apache.tools.ant.taskdefs.optional.junit.JUnitTestRunner.run(JUnitTestRunner.java:520) at org.apache.tools.ant.taskdefs.optional.junit.JUnitTestRunner.launch(JUnitTestRunner.java:1060) at org.apache.tools.ant.taskdefs.optional.junit.JUnitTestRunner.main(JUnitTestRunner.java:911) {noformat} -- This message was sent by Atlassian JIRA (v6.1#6144)
[jira] [Commented] (HIVE-5482) JDBC should depend on httpclient.version and httpcore.version 4.1.3 to be consistent with other modules
[ https://issues.apache.org/jira/browse/HIVE-5482?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13805822#comment-13805822 ] Hudson commented on HIVE-5482: -- ABORTED: Integrated in Hive-trunk-hadoop2 #521 (See [https://builds.apache.org/job/Hive-trunk-hadoop2/521/]) HIVE-5482 : JDBC should depend on httpclient.version and httpcore.version 4.1.3 to be consistent with other modules (Vaibhav Gumashta via Thejas Nair) (thejas: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1535620) * /hive/trunk/ivy/libraries.properties JDBC should depend on httpclient.version and httpcore.version 4.1.3 to be consistent with other modules --- Key: HIVE-5482 URL: https://issues.apache.org/jira/browse/HIVE-5482 Project: Hive Issue Type: Bug Components: HiveServer2, JDBC Reporter: Vaibhav Gumashta Assignee: Vaibhav Gumashta Fix For: 0.13.0 Attachments: HIVE-5482.1.patch Currently depends on 4.2.4 and 4.2.5 which conflicts with thrift-0.9 which depends on 4.1.3 -- This message was sent by Atlassian JIRA (v6.1#6144)
[jira] [Commented] (HIVE-5220) Add option for removing intermediate directory for partition, which is empty
[ https://issues.apache.org/jira/browse/HIVE-5220?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13804265#comment-13804265 ] Hudson commented on HIVE-5220: -- FAILURE: Integrated in Hive-trunk-h0.21 #2418 (See [https://builds.apache.org/job/Hive-trunk-h0.21/2418/]) HIVE-5220 : Add option for removing intermediate directory for partition, which is empty (Navis via Ashutosh Chauhan) (hashutosh: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1535072) * /hive/trunk/metastore/src/java/org/apache/hadoop/hive/metastore/HiveMetaStore.java * /hive/trunk/metastore/src/java/org/apache/hadoop/hive/metastore/Warehouse.java Add option for removing intermediate directory for partition, which is empty Key: HIVE-5220 URL: https://issues.apache.org/jira/browse/HIVE-5220 Project: Hive Issue Type: Improvement Components: Query Processor Reporter: Navis Assignee: Navis Priority: Minor Fix For: 0.13.0 Attachments: D12729.2.patch, HIVE-5220.D12729.1.patch For deeply nested partitioned table, intermediate directories are not removed even if there is no partitions in it by removing them. {noformat} /deep_part/c=09/d=01 /deep_part/c=09/d=01/e=01 /deep_part/c=09/d=01/e=02 /deep_part/c=09/d=02 /deep_part/c=09/d=02/e=01 /deep_part/c=09/d=02/e=02 {noformat} After removing partition (c='09'), directory remains like this, {noformat} /deep_part/c=09/d=01 /deep_part/c=09/d=02 {noformat} -- This message was sent by Atlassian JIRA (v6.1#6144)
[jira] [Commented] (HIVE-5625) Fix issue with metastore version restriction test.
[ https://issues.apache.org/jira/browse/HIVE-5625?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13805066#comment-13805066 ] Hudson commented on HIVE-5625: -- FAILURE: Integrated in Hive-trunk-h0.21 #2419 (See [https://builds.apache.org/job/Hive-trunk-h0.21/2419/]) HIVE-5625 - Fix issue with metastore version restriction test. (Vikram Dixit K via Brock Noland) (brock: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1535402) * /hive/trunk/metastore/src/test/org/apache/hadoop/hive/metastore/TestMetastoreVersion.java Fix issue with metastore version restriction test. -- Key: HIVE-5625 URL: https://issues.apache.org/jira/browse/HIVE-5625 Project: Hive Issue Type: Bug Components: Tests Affects Versions: 0.13.0 Reporter: Vikram Dixit K Assignee: Vikram Dixit K Fix For: 0.13.0 Attachments: HIVE-5625.1.patch Based on Brock's comments, the change made in HIVE-5403 change the nature of the test. -- This message was sent by Atlassian JIRA (v6.1#6144)
[jira] [Commented] (HIVE-5514) webhcat_server.sh foreground option does not work as expected
[ https://issues.apache.org/jira/browse/HIVE-5514?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13802630#comment-13802630 ] Hudson commented on HIVE-5514: -- FAILURE: Integrated in Hive-trunk-hadoop2 #517 (See [https://builds.apache.org/job/Hive-trunk-hadoop2/517/]) HIVE-5514 - webhcat_server.sh foreground option does not work as expected (brock: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1534662) * /hive/trunk/hcatalog/webhcat/svr/src/main/bin/webhcat_server.sh webhcat_server.sh foreground option does not work as expected - Key: HIVE-5514 URL: https://issues.apache.org/jira/browse/HIVE-5514 Project: Hive Issue Type: Improvement Reporter: Brock Noland Assignee: Brock Noland Priority: Minor Fix For: 0.13.0 Attachments: HIVE-5514.patch Executing webhcat script webhcat_server.sh with the foreground option, it calls calls hadoop without using exec. When you kill the webhcat_server.sh process, it does not kill the real webhcat server. Just need to add the word exec below in webhcat_server.sh: {noformat} function foreground_webhcat() { exec $start_cmd } {noformat} NO PRECOMMIT TESTS -- This message was sent by Atlassian JIRA (v6.1#6144)