[ANNOUNCE] Apache Hive 0.10.0 Released
The Apache Hive team is proud to announce the the release of Apache Hive version 0.10.0. The Apache Hive (TM) data warehouse software facilitates querying and managing large datasets residing in distributed storage. Built on top of Apache Hadoop (TM), it provides: * Tools to enable easy data extract/transform/load (ETL) * A mechanism to impose structure on a variety of data formats * Access to files stored either directly in Apache HDFS (TM) or in other data storage systems such as Apache HBase (TM) * Query execution via MapReduce For Hive release details and downloads, please visit:http://hive.apache.org/releases.html Hive 0.10.0 Release Notes are available here:https://issues.apache.org/jira/secure/ReleaseNote.jspa?version=12320745styleName=TextprojectId=12310843 More than 350 jira issues are fixed in this release. We would like to thank many contributors who made this release possible. Regards, The Apache Hive Team
[jira] [Commented] (HIVE-2693) Add DECIMAL data type
[ https://issues.apache.org/jira/browse/HIVE-2693?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13550950#comment-13550950 ] Carl Steinbach commented on HIVE-2693: -- +1. Changes look good to me. Ashutosh and Namit: let me know if you want more time to look at this. Otherwise I'll plan to commit it in 24 hours. Thanks. Add DECIMAL data type - Key: HIVE-2693 URL: https://issues.apache.org/jira/browse/HIVE-2693 Project: Hive Issue Type: New Feature Components: Query Processor, Types Affects Versions: 0.10.0 Reporter: Carl Steinbach Assignee: Prasad Mujumdar Attachments: 2693_7.patch, 2693_8.patch, 2693_fix_all_tests1.patch, HIVE-2693-10.patch, HIVE-2693-11.patch, HIVE-2693-12-SortableSerDe.patch, HIVE-2693-13.patch, HIVE-2693-14.patch, HIVE-2693-15.patch, HIVE-2693-16.patch, HIVE-2693-17.patch, HIVE-2693-18.patch, HIVE-2693-19.patch, HIVE-2693-1.patch.txt, HIVE-2693-20.patch, HIVE-2693-21.patch, HIVE-2693-all.patch, HIVE-2693.D7683.1.patch, HIVE-2693-fix.patch, HIVE-2693.patch, HIVE-2693-take3.patch, HIVE-2693-take4.patch Add support for the DECIMAL data type. HIVE-2272 (TIMESTAMP) provides a nice template for how to do this. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Updated] (HIVE-3403) user should not specify mapjoin to perform sort-merge bucketed join
[ https://issues.apache.org/jira/browse/HIVE-3403?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Namit Jain updated HIVE-3403: - Attachment: hive.3403.11.patch user should not specify mapjoin to perform sort-merge bucketed join --- Key: HIVE-3403 URL: https://issues.apache.org/jira/browse/HIVE-3403 Project: Hive Issue Type: Bug Reporter: Namit Jain Assignee: Namit Jain Attachments: hive.3403.10.patch, hive.3403.11.patch, hive.3403.1.patch, hive.3403.2.patch, hive.3403.3.patch, hive.3403.4.patch, hive.3403.5.patch, hive.3403.6.patch, hive.3403.7.patch, hive.3403.8.patch, hive.3403.9.patch Currently, in order to perform a sort merge bucketed join, the user needs to set hive.optimize.bucketmapjoin.sortedmerge to true, and also specify the mapjoin hint. The user should not specify any hints. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Updated] (HIVE-3810) HiveHistory.log need to replace '\r' with space before writing Entry.value to historyfile
[ https://issues.apache.org/jira/browse/HIVE-3810?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] qiangwang updated HIVE-3810: Resolution: Fixed Status: Resolved (was: Patch Available) HiveHistory.log need to replace '\r' with space before writing Entry.value to historyfile - Key: HIVE-3810 URL: https://issues.apache.org/jira/browse/HIVE-3810 Project: Hive Issue Type: Bug Components: Logging Reporter: qiangwang Assignee: Mark Grover Priority: Minor Attachments: HIVE-3810.1.patch, HIVE-3810.2.patch HiveHistory.log will replace '\n' with space before writing Entry.value to history file: val = val.replace('\n', ' '); but HiveHistory.parseHiveHistory use BufferedReader.readLine which takes '\n', '\r', '\r\n' as line delimiter to parse history file if val contains '\r', there is a high possibility that HiveHistory.parseLine will fail, in which case usually RecordTypes.valueOf(recType) will throw exception 'java.lang.IllegalArgumentException' HiveHistory.log need to replace '\r' with space as well: val = val.replace('\n', ' '); changed to val = val.replaceAll(\r|\n, ); or val = val.replace('\r', ' ').replace('\n', ' '); -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Updated] (HIVE-3857) HiveHistoryViewer concurrency problem
[ https://issues.apache.org/jira/browse/HIVE-3857?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] qiangwang updated HIVE-3857: Resolution: Fixed Status: Resolved (was: Patch Available) HiveHistoryViewer concurrency problem - Key: HIVE-3857 URL: https://issues.apache.org/jira/browse/HIVE-3857 Project: Hive Issue Type: Bug Components: Logging Reporter: qiangwang Priority: Minor Attachments: HiveHistory.patch new HiveHistoryViewer() throws ConcurrentModificationException when called concurrently by several threads. According to the stacktrace, HiveHistory.parseLine use private static MapString, String parseBuffer to store parsed data and this finally caused the exception. I don't know why a static buffer rather than a local buffer is used! -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HIVE-3874) Create a new Optimized Row Columnar file format for Hive
[ https://issues.apache.org/jira/browse/HIVE-3874?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13550961#comment-13550961 ] Namit Jain commented on HIVE-3874: -- In partition metadata, can I somehow specify - dont create index ? When that partition is over-written, the index would dis-appear. I agree, we can think about that later. Create a new Optimized Row Columnar file format for Hive Key: HIVE-3874 URL: https://issues.apache.org/jira/browse/HIVE-3874 Project: Hive Issue Type: Improvement Components: Serializers/Deserializers Reporter: Owen O'Malley Assignee: Owen O'Malley Attachments: OrcFileIntro.pptx There are several limitations of the current RC File format that I'd like to address by creating a new format: * each column value is stored as a binary blob, which means: ** the entire column value must be read, decompressed, and deserialized ** the file format can't use smarter type-specific compression ** push down filters can't be evaluated * the start of each row group needs to be found by scanning * user metadata can only be added to the file when the file is created * the file doesn't store the number of rows per a file or row group * there is no mechanism for seeking to a particular row number, which is required for external indexes. * there is no mechanism for storing light weight indexes within the file to enable push-down filters to skip entire row groups. * the type of the rows aren't stored in the file -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HIVE-3264) Add support for binary dataype to AvroSerde
[ https://issues.apache.org/jira/browse/HIVE-3264?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13550963#comment-13550963 ] Eli Reisman commented on HIVE-3264: --- The patch isn't stale yet (whew!), I'm working on a test... Add support for binary dataype to AvroSerde --- Key: HIVE-3264 URL: https://issues.apache.org/jira/browse/HIVE-3264 Project: Hive Issue Type: Bug Components: Serializers/Deserializers Affects Versions: 0.9.0 Reporter: Jakob Homan Labels: patch Attachments: HIVE-3264-1.patch, HIVE-3264-2.patch, HIVE-3264-3.patch, HIVE-3264-4.patch, HIVE-3264-5.patch When the AvroSerde was written, Hive didn't have a binary type, so Avro's byte array type is converted an array of small ints. Now that HIVE-2380 is in, this step isn't necessary and we can convert both Avro's bytes type and probably fixed type to Hive's binary type. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HIVE-2693) Add DECIMAL data type
[ https://issues.apache.org/jira/browse/HIVE-2693?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13550965#comment-13550965 ] Ashutosh Chauhan commented on HIVE-2693: Latest patch looks good to me. +1 Add DECIMAL data type - Key: HIVE-2693 URL: https://issues.apache.org/jira/browse/HIVE-2693 Project: Hive Issue Type: New Feature Components: Query Processor, Types Affects Versions: 0.10.0 Reporter: Carl Steinbach Assignee: Prasad Mujumdar Attachments: 2693_7.patch, 2693_8.patch, 2693_fix_all_tests1.patch, HIVE-2693-10.patch, HIVE-2693-11.patch, HIVE-2693-12-SortableSerDe.patch, HIVE-2693-13.patch, HIVE-2693-14.patch, HIVE-2693-15.patch, HIVE-2693-16.patch, HIVE-2693-17.patch, HIVE-2693-18.patch, HIVE-2693-19.patch, HIVE-2693-1.patch.txt, HIVE-2693-20.patch, HIVE-2693-21.patch, HIVE-2693-all.patch, HIVE-2693.D7683.1.patch, HIVE-2693-fix.patch, HIVE-2693.patch, HIVE-2693-take3.patch, HIVE-2693-take4.patch Add support for the DECIMAL data type. HIVE-2272 (TIMESTAMP) provides a nice template for how to do this. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HIVE-2693) Add DECIMAL data type
[ https://issues.apache.org/jira/browse/HIVE-2693?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13550970#comment-13550970 ] Namit Jain commented on HIVE-2693: -- Carl, give me some time. I will try to take a look today/tomorrow. Can you give me ~2 days ? Go ahead, if you dont hear from me. Add DECIMAL data type - Key: HIVE-2693 URL: https://issues.apache.org/jira/browse/HIVE-2693 Project: Hive Issue Type: New Feature Components: Query Processor, Types Affects Versions: 0.10.0 Reporter: Carl Steinbach Assignee: Prasad Mujumdar Attachments: 2693_7.patch, 2693_8.patch, 2693_fix_all_tests1.patch, HIVE-2693-10.patch, HIVE-2693-11.patch, HIVE-2693-12-SortableSerDe.patch, HIVE-2693-13.patch, HIVE-2693-14.patch, HIVE-2693-15.patch, HIVE-2693-16.patch, HIVE-2693-17.patch, HIVE-2693-18.patch, HIVE-2693-19.patch, HIVE-2693-1.patch.txt, HIVE-2693-20.patch, HIVE-2693-21.patch, HIVE-2693-all.patch, HIVE-2693.D7683.1.patch, HIVE-2693-fix.patch, HIVE-2693.patch, HIVE-2693-take3.patch, HIVE-2693-take4.patch Add support for the DECIMAL data type. HIVE-2272 (TIMESTAMP) provides a nice template for how to do this. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Updated] (HIVE-3888) wrong mapside groupby if no partition is being selected
[ https://issues.apache.org/jira/browse/HIVE-3888?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Namit Jain updated HIVE-3888: - Status: Patch Available (was: Open) wrong mapside groupby if no partition is being selected --- Key: HIVE-3888 URL: https://issues.apache.org/jira/browse/HIVE-3888 Project: Hive Issue Type: Bug Components: Query Processor Reporter: Namit Jain Assignee: Namit Jain Attachments: hive.3888.1.patch -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Updated] (HIVE-3888) wrong mapside groupby if no partition is being selected
[ https://issues.apache.org/jira/browse/HIVE-3888?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Namit Jain updated HIVE-3888: - Resolution: Fixed Fix Version/s: 0.11.0 Hadoop Flags: Reviewed Status: Resolved (was: Patch Available) Committed. Thanks Ashutosh wrong mapside groupby if no partition is being selected --- Key: HIVE-3888 URL: https://issues.apache.org/jira/browse/HIVE-3888 Project: Hive Issue Type: Bug Components: Query Processor Reporter: Namit Jain Assignee: Namit Jain Fix For: 0.11.0 Attachments: hive.3888.1.patch -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
Re: [ANNOUNCE] Apache Hive 0.10.0 Released
Congratulations! – Lefty On Thu, Jan 10, 2013 at 11:59 PM, Ashutosh Chauhan hashut...@apache.orgwrote: The Apache Hive team is proud to announce the the release of Apache Hive version 0.10.0. The Apache Hive (TM) data warehouse software facilitates querying and managing large datasets residing in distributed storage. Built on top of Apache Hadoop (TM), it provides: * Tools to enable easy data extract/transform/load (ETL) * A mechanism to impose structure on a variety of data formats * Access to files stored either directly in Apache HDFS (TM) or in other data storage systems such as Apache HBase (TM) * Query execution via MapReduce For Hive release details and downloads, please visit:http://hive.apache.org/releases.html Hive 0.10.0 Release Notes are available here: https://issues.apache.org/jira/secure/ReleaseNote.jspa?version=12320745styleName=TextprojectId=12310843 More than 350 jira issues are fixed in this release. We would like to thank many contributors who made this release possible. Regards, The Apache Hive Team
[jira] [Commented] (HIVE-3888) wrong mapside groupby if no partition is being selected
[ https://issues.apache.org/jira/browse/HIVE-3888?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13551021#comment-13551021 ] Hudson commented on HIVE-3888: -- Integrated in hive-trunk-hadoop1 #7 (See [https://builds.apache.org/job/hive-trunk-hadoop1/7/]) HIVE-3888 wrong mapside groupby if no partition is being selected (Namit Jain via Ashutosh and namit) (Revision 1431936) Result = ABORTED namit : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1431936 Files : * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/optimizer/GroupByOptimizer.java * /hive/trunk/ql/src/test/queries/clientpositive/groupby_sort_6.q * /hive/trunk/ql/src/test/queries/clientpositive/groupby_sort_7.q * /hive/trunk/ql/src/test/results/clientpositive/groupby_sort_6.q.out * /hive/trunk/ql/src/test/results/clientpositive/groupby_sort_7.q.out * /hive/trunk/ql/src/test/results/clientpositive/metadataonly1.q.out wrong mapside groupby if no partition is being selected --- Key: HIVE-3888 URL: https://issues.apache.org/jira/browse/HIVE-3888 Project: Hive Issue Type: Bug Components: Query Processor Reporter: Namit Jain Assignee: Namit Jain Fix For: 0.11.0 Attachments: hive.3888.1.patch -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HIVE-3875) negative value for hive.stats.ndv.error should be disallowed
[ https://issues.apache.org/jira/browse/HIVE-3875?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13551044#comment-13551044 ] Hudson commented on HIVE-3875: -- Integrated in Hive-trunk-hadoop2 #58 (See [https://builds.apache.org/job/Hive-trunk-hadoop2/58/]) HIVE-3875. Negative value for hive.stats.ndv.error should be disallowed (Shreepadma Venugopalan via cws) (Revision 1431793) Result = FAILURE cws : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1431793 Files : * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/parse/ColumnStatsSemanticAnalyzer.java negative value for hive.stats.ndv.error should be disallowed - Key: HIVE-3875 URL: https://issues.apache.org/jira/browse/HIVE-3875 Project: Hive Issue Type: Bug Components: Statistics Affects Versions: 0.10.0 Reporter: Shreepadma Venugopalan Assignee: Shreepadma Venugopalan Fix For: 0.11.0 Attachments: HIVE-3875.1.patch.txt Currently, if a negative value is specified for hive.stats.ndv.error in hive-site.xml, it is treated as 0. We should instead throw an exception. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
Hive-trunk-hadoop2 - Build # 58 - Failure
36 tests failed. FAILED: org.apache.hadoop.hive.cli.TestHBaseCliDriver.testCliDriver_external_table_ppd Error Message: Unexpected exception in setup Stack Trace: junit.framework.AssertionFailedError: Unexpected exception in setup at junit.framework.Assert.fail(Assert.java:50) at org.apache.hadoop.hive.cli.TestHBaseCliDriver.setUp(TestHBaseCliDriver.java:59) at junit.framework.TestCase.runBare(TestCase.java:132) at junit.framework.TestResult$1.protect(TestResult.java:110) at junit.framework.TestResult.runProtected(TestResult.java:128) at junit.framework.TestResult.run(TestResult.java:113) at junit.framework.TestCase.run(TestCase.java:124) at junit.framework.TestSuite.runTest(TestSuite.java:243) at junit.framework.TestSuite.run(TestSuite.java:238) at junit.extensions.TestDecorator.basicRun(TestDecorator.java:24) at junit.extensions.TestSetup$1.protect(TestSetup.java:23) at junit.framework.TestResult.runProtected(TestResult.java:128) at junit.extensions.TestSetup.run(TestSetup.java:27) at org.apache.tools.ant.taskdefs.optional.junit.JUnitTestRunner.run(JUnitTestRunner.java:422) at org.apache.tools.ant.taskdefs.optional.junit.JUnitTestRunner.launch(JUnitTestRunner.java:931) at org.apache.tools.ant.taskdefs.optional.junit.JUnitTestRunner.main(JUnitTestRunner.java:785) FAILED: org.apache.hadoop.hive.cli.TestHBaseCliDriver.testCliDriver_hbase_binary_external_table_queries Error Message: Unexpected exception in setup Stack Trace: junit.framework.AssertionFailedError: Unexpected exception in setup at junit.framework.Assert.fail(Assert.java:50) at org.apache.hadoop.hive.cli.TestHBaseCliDriver.setUp(TestHBaseCliDriver.java:59) at junit.framework.TestCase.runBare(TestCase.java:132) at junit.framework.TestResult$1.protect(TestResult.java:110) at junit.framework.TestResult.runProtected(TestResult.java:128) at junit.framework.TestResult.run(TestResult.java:113) at junit.framework.TestCase.run(TestCase.java:124) at junit.framework.TestSuite.runTest(TestSuite.java:243) at junit.framework.TestSuite.run(TestSuite.java:238) at junit.extensions.TestDecorator.basicRun(TestDecorator.java:24) at junit.extensions.TestSetup$1.protect(TestSetup.java:23) at junit.framework.TestResult.runProtected(TestResult.java:128) at junit.extensions.TestSetup.run(TestSetup.java:27) at org.apache.tools.ant.taskdefs.optional.junit.JUnitTestRunner.run(JUnitTestRunner.java:422) at org.apache.tools.ant.taskdefs.optional.junit.JUnitTestRunner.launch(JUnitTestRunner.java:931) at org.apache.tools.ant.taskdefs.optional.junit.JUnitTestRunner.main(JUnitTestRunner.java:785) FAILED: org.apache.hadoop.hive.cli.TestHBaseCliDriver.testCliDriver_hbase_binary_map_queries Error Message: Unexpected exception in setup Stack Trace: junit.framework.AssertionFailedError: Unexpected exception in setup at junit.framework.Assert.fail(Assert.java:50) at org.apache.hadoop.hive.cli.TestHBaseCliDriver.setUp(TestHBaseCliDriver.java:59) at junit.framework.TestCase.runBare(TestCase.java:132) at junit.framework.TestResult$1.protect(TestResult.java:110) at junit.framework.TestResult.runProtected(TestResult.java:128) at junit.framework.TestResult.run(TestResult.java:113) at junit.framework.TestCase.run(TestCase.java:124) at junit.framework.TestSuite.runTest(TestSuite.java:243) at junit.framework.TestSuite.run(TestSuite.java:238) at junit.extensions.TestDecorator.basicRun(TestDecorator.java:24) at junit.extensions.TestSetup$1.protect(TestSetup.java:23) at junit.framework.TestResult.runProtected(TestResult.java:128) at junit.extensions.TestSetup.run(TestSetup.java:27) at org.apache.tools.ant.taskdefs.optional.junit.JUnitTestRunner.run(JUnitTestRunner.java:422) at org.apache.tools.ant.taskdefs.optional.junit.JUnitTestRunner.launch(JUnitTestRunner.java:931) at org.apache.tools.ant.taskdefs.optional.junit.JUnitTestRunner.main(JUnitTestRunner.java:785) FAILED: org.apache.hadoop.hive.cli.TestHBaseCliDriver.testCliDriver_hbase_binary_storage_queries Error Message: Unexpected exception in setup Stack Trace: junit.framework.AssertionFailedError: Unexpected exception in setup at junit.framework.Assert.fail(Assert.java:50) at org.apache.hadoop.hive.cli.TestHBaseCliDriver.setUp(TestHBaseCliDriver.java:59) at junit.framework.TestCase.runBare(TestCase.java:132) at junit.framework.TestResult$1.protect(TestResult.java:110) at junit.framework.TestResult.runProtected(TestResult.java:128) at junit.framework.TestResult.run(TestResult.java:113) at junit.framework.TestCase.run(TestCase.java:124) at
[jira] [Updated] (HIVE-3403) user should not specify mapjoin to perform sort-merge bucketed join
[ https://issues.apache.org/jira/browse/HIVE-3403?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Namit Jain updated HIVE-3403: - Attachment: hive.3403.12.patch user should not specify mapjoin to perform sort-merge bucketed join --- Key: HIVE-3403 URL: https://issues.apache.org/jira/browse/HIVE-3403 Project: Hive Issue Type: Bug Reporter: Namit Jain Assignee: Namit Jain Attachments: hive.3403.10.patch, hive.3403.11.patch, hive.3403.12.patch, hive.3403.1.patch, hive.3403.2.patch, hive.3403.3.patch, hive.3403.4.patch, hive.3403.5.patch, hive.3403.6.patch, hive.3403.7.patch, hive.3403.8.patch, hive.3403.9.patch Currently, in order to perform a sort merge bucketed join, the user needs to set hive.optimize.bucketmapjoin.sortedmerge to true, and also specify the mapjoin hint. The user should not specify any hints. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HIVE-3875) negative value for hive.stats.ndv.error should be disallowed
[ https://issues.apache.org/jira/browse/HIVE-3875?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13551092#comment-13551092 ] Hudson commented on HIVE-3875: -- Integrated in Hive-trunk-h0.21 #1906 (See [https://builds.apache.org/job/Hive-trunk-h0.21/1906/]) HIVE-3875. Negative value for hive.stats.ndv.error should be disallowed (Shreepadma Venugopalan via cws) (Revision 1431793) Result = SUCCESS cws : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1431793 Files : * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/parse/ColumnStatsSemanticAnalyzer.java negative value for hive.stats.ndv.error should be disallowed - Key: HIVE-3875 URL: https://issues.apache.org/jira/browse/HIVE-3875 Project: Hive Issue Type: Bug Components: Statistics Affects Versions: 0.10.0 Reporter: Shreepadma Venugopalan Assignee: Shreepadma Venugopalan Fix For: 0.11.0 Attachments: HIVE-3875.1.patch.txt Currently, if a negative value is specified for hive.stats.ndv.error in hive-site.xml, it is treated as 0. We should instead throw an exception. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
Hive-trunk-h0.21 - Build # 1906 - Fixed
Changes for Build #1905 Changes for Build #1906 [cws] HIVE-3875. Negative value for hive.stats.ndv.error should be disallowed (Shreepadma Venugopalan via cws) All tests passed The Apache Jenkins build system has built Hive-trunk-h0.21 (build #1906) Status: Fixed Check console output at https://builds.apache.org/job/Hive-trunk-h0.21/1906/ to view the results.
[jira] [Commented] (HIVE-3810) HiveHistory.log need to replace '\r' with space before writing Entry.value to historyfile
[ https://issues.apache.org/jira/browse/HIVE-3810?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13551149#comment-13551149 ] Mark Grover commented on HIVE-3810: --- [~wsxy...@gmail.com] I see the status got resolved to fix. Did this get committed? HiveHistory.log need to replace '\r' with space before writing Entry.value to historyfile - Key: HIVE-3810 URL: https://issues.apache.org/jira/browse/HIVE-3810 Project: Hive Issue Type: Bug Components: Logging Reporter: qiangwang Assignee: Mark Grover Priority: Minor Attachments: HIVE-3810.1.patch, HIVE-3810.2.patch HiveHistory.log will replace '\n' with space before writing Entry.value to history file: val = val.replace('\n', ' '); but HiveHistory.parseHiveHistory use BufferedReader.readLine which takes '\n', '\r', '\r\n' as line delimiter to parse history file if val contains '\r', there is a high possibility that HiveHistory.parseLine will fail, in which case usually RecordTypes.valueOf(recType) will throw exception 'java.lang.IllegalArgumentException' HiveHistory.log need to replace '\r' with space as well: val = val.replace('\n', ' '); changed to val = val.replaceAll(\r|\n, ); or val = val.replace('\r', ' ').replace('\n', ' '); -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Updated] (HIVE-3876) call resetValid instead of ensureCapacity in the constructor of BytesRefArrayWritable
[ https://issues.apache.org/jira/browse/HIVE-3876?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Yin Huai updated HIVE-3876: --- Status: Patch Available (was: Open) call resetValid instead of ensureCapacity in the constructor of BytesRefArrayWritable - Key: HIVE-3876 URL: https://issues.apache.org/jira/browse/HIVE-3876 Project: Hive Issue Type: Improvement Components: Serializers/Deserializers Affects Versions: 0.10.0 Reporter: Yin Huai Assignee: Yin Huai Priority: Minor Attachments: HIVE-3876.1.patch.txt In the constructor of BytesRefArrayWritable, ensureCapacity(capacity) is called, but valid has not been adjusted accordingly. After a new BytesRefArrayWritable has been created with a initial capacity of x, if resetValid() has not been called explicitly, the size returned is still 0. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Updated] (HIVE-2439) Upgrade antlr version to 3.4
[ https://issues.apache.org/jira/browse/HIVE-2439?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Thiruvel Thirumoolan updated HIVE-2439: --- Attachment: HIVE-2439_branch10.patch HIVE-2439_trunk_1.patch Upgrade antlr version to 3.4 Key: HIVE-2439 URL: https://issues.apache.org/jira/browse/HIVE-2439 Project: Hive Issue Type: Improvement Affects Versions: 0.8.0 Reporter: Ashutosh Chauhan Assignee: Thiruvel Thirumoolan Attachments: HIVE-2439_branch10.patch, HIVE-2439_branch9_2.patch, HIVE-2439_branch9_3.patch, HIVE-2439_branch9.patch, hive-2439_incomplete.patch, HIVE-2439_trunk_1.patch, HIVE-2439_trunk.patch Upgrade antlr version to 3.4 -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Created] (HIVE-3889) Add floating point compression to ORC file
Owen O'Malley created HIVE-3889: --- Summary: Add floating point compression to ORC file Key: HIVE-3889 URL: https://issues.apache.org/jira/browse/HIVE-3889 Project: Hive Issue Type: New Feature Components: Serializers/Deserializers Reporter: Owen O'Malley Assignee: Owen O'Malley Karol Wegrzycki, a CS student at University of Warsaw, has implemented an FPC compressor for doubles. It would be great to hook this up to the ORC file format so that we can get better compression for doubles. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Updated] (HIVE-3889) Add floating point compression to ORC file
[ https://issues.apache.org/jira/browse/HIVE-3889?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Owen O'Malley updated HIVE-3889: Attachment: fpc-impl.tar This is the file that Karol emailed to me for me to submit to Apache. Add floating point compression to ORC file -- Key: HIVE-3889 URL: https://issues.apache.org/jira/browse/HIVE-3889 Project: Hive Issue Type: New Feature Components: Serializers/Deserializers Reporter: Owen O'Malley Assignee: Owen O'Malley Attachments: fpc-impl.tar Karol Wegrzycki, a CS student at University of Warsaw, has implemented an FPC compressor for doubles. It would be great to hook this up to the ORC file format so that we can get better compression for doubles. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HIVE-3874) Create a new Optimized Row Columnar file format for Hive
[ https://issues.apache.org/jira/browse/HIVE-3874?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13551231#comment-13551231 ] Owen O'Malley commented on HIVE-3874: - Namit, I'm using the table properties to manage the other features like compression, so I would probably make a table property like 'orc.create.index' or something. Would that make sense? I should note that the indexes are very light. In a sample file: * uncompressed text: 370MB * compress orc: 86MB * row index in orc: 140k Create a new Optimized Row Columnar file format for Hive Key: HIVE-3874 URL: https://issues.apache.org/jira/browse/HIVE-3874 Project: Hive Issue Type: Improvement Components: Serializers/Deserializers Reporter: Owen O'Malley Assignee: Owen O'Malley Attachments: OrcFileIntro.pptx There are several limitations of the current RC File format that I'd like to address by creating a new format: * each column value is stored as a binary blob, which means: ** the entire column value must be read, decompressed, and deserialized ** the file format can't use smarter type-specific compression ** push down filters can't be evaluated * the start of each row group needs to be found by scanning * user metadata can only be added to the file when the file is created * the file doesn't store the number of rows per a file or row group * there is no mechanism for seeking to a particular row number, which is required for external indexes. * there is no mechanism for storing light weight indexes within the file to enable push-down filters to skip entire row groups. * the type of the rows aren't stored in the file -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HIVE-3874) Create a new Optimized Row Columnar file format for Hive
[ https://issues.apache.org/jira/browse/HIVE-3874?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13551248#comment-13551248 ] Owen O'Malley commented on HIVE-3874: - Doug, of course Trevni could be modified arbitrarily to match the needs of Hive. But Hive will benefit more if there is a deep integration between the file format and the query engine. Both HBase and Accumulo have file formats that were originally based on Hadoop's TFile. But the need for integration with the query engine was such that their projects were better served by having the file format in their project rather than an upstream project. Of course the Avro project is free to copy any of the ORC code into Trevni, but Hive has the need to innovate in this area without asking Avro to make changes and waiting for them to be released. Create a new Optimized Row Columnar file format for Hive Key: HIVE-3874 URL: https://issues.apache.org/jira/browse/HIVE-3874 Project: Hive Issue Type: Improvement Components: Serializers/Deserializers Reporter: Owen O'Malley Assignee: Owen O'Malley Attachments: OrcFileIntro.pptx There are several limitations of the current RC File format that I'd like to address by creating a new format: * each column value is stored as a binary blob, which means: ** the entire column value must be read, decompressed, and deserialized ** the file format can't use smarter type-specific compression ** push down filters can't be evaluated * the start of each row group needs to be found by scanning * user metadata can only be added to the file when the file is created * the file doesn't store the number of rows per a file or row group * there is no mechanism for seeking to a particular row number, which is required for external indexes. * there is no mechanism for storing light weight indexes within the file to enable push-down filters to skip entire row groups. * the type of the rows aren't stored in the file -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HIVE-3874) Create a new Optimized Row Columnar file format for Hive
[ https://issues.apache.org/jira/browse/HIVE-3874?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13551275#comment-13551275 ] Sambavi Muthukrishnan commented on HIVE-3874: - Owen: is the row group expected to be about 250 MB (per stripe size)? Does your implementation attempt to make every row group align with HDFS block size so a split = 1 block? Also: do you have an ETA for an initial patch? We would really like to try this out - we have some additional ideas that we would like to try out on top of this. Create a new Optimized Row Columnar file format for Hive Key: HIVE-3874 URL: https://issues.apache.org/jira/browse/HIVE-3874 Project: Hive Issue Type: Improvement Components: Serializers/Deserializers Reporter: Owen O'Malley Assignee: Owen O'Malley Attachments: OrcFileIntro.pptx There are several limitations of the current RC File format that I'd like to address by creating a new format: * each column value is stored as a binary blob, which means: ** the entire column value must be read, decompressed, and deserialized ** the file format can't use smarter type-specific compression ** push down filters can't be evaluated * the start of each row group needs to be found by scanning * user metadata can only be added to the file when the file is created * the file doesn't store the number of rows per a file or row group * there is no mechanism for seeking to a particular row number, which is required for external indexes. * there is no mechanism for storing light weight indexes within the file to enable push-down filters to skip entire row groups. * the type of the rows aren't stored in the file -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HIVE-3874) Create a new Optimized Row Columnar file format for Hive
[ https://issues.apache.org/jira/browse/HIVE-3874?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13551286#comment-13551286 ] Owen O'Malley commented on HIVE-3874: - Sambavi, I should have a patch ready next week. Yes, the row groups (stripes) are 250MB by default. I currently set the HDFS block size for the files to 2 times the stripe size, but I don't try to align them other than that. Create a new Optimized Row Columnar file format for Hive Key: HIVE-3874 URL: https://issues.apache.org/jira/browse/HIVE-3874 Project: Hive Issue Type: Improvement Components: Serializers/Deserializers Reporter: Owen O'Malley Assignee: Owen O'Malley Attachments: OrcFileIntro.pptx There are several limitations of the current RC File format that I'd like to address by creating a new format: * each column value is stored as a binary blob, which means: ** the entire column value must be read, decompressed, and deserialized ** the file format can't use smarter type-specific compression ** push down filters can't be evaluated * the start of each row group needs to be found by scanning * user metadata can only be added to the file when the file is created * the file doesn't store the number of rows per a file or row group * there is no mechanism for seeking to a particular row number, which is required for external indexes. * there is no mechanism for storing light weight indexes within the file to enable push-down filters to skip entire row groups. * the type of the rows aren't stored in the file -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Updated] (HIVE-3803) explain dependency should show the dependencies hierarchically in presence of views
[ https://issues.apache.org/jira/browse/HIVE-3803?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Namit Jain updated HIVE-3803: - Attachment: hive.3803.12.patch explain dependency should show the dependencies hierarchically in presence of views --- Key: HIVE-3803 URL: https://issues.apache.org/jira/browse/HIVE-3803 Project: Hive Issue Type: Bug Components: Query Processor Reporter: Namit Jain Assignee: Namit Jain Attachments: hive.3803.10.patch, hive.3803.11.patch, hive.3803.12.patch, hive.3803.1.patch, hive.3803.2.patch, hive.3803.3.patch, hive.3803.4.patch, hive.3803.5.patch, hive.3803.6.patch, hive.3803.7.patch, hive.3803.8.patch, hive.3803.9.patch It should also include tables whose partitions are being accessed -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HIVE-3874) Create a new Optimized Row Columnar file format for Hive
[ https://issues.apache.org/jira/browse/HIVE-3874?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13551347#comment-13551347 ] Namit Jain commented on HIVE-3874: -- Owen, that would be good 'a table property like 'orc.create.index' or something' Create a new Optimized Row Columnar file format for Hive Key: HIVE-3874 URL: https://issues.apache.org/jira/browse/HIVE-3874 Project: Hive Issue Type: Improvement Components: Serializers/Deserializers Reporter: Owen O'Malley Assignee: Owen O'Malley Attachments: OrcFileIntro.pptx There are several limitations of the current RC File format that I'd like to address by creating a new format: * each column value is stored as a binary blob, which means: ** the entire column value must be read, decompressed, and deserialized ** the file format can't use smarter type-specific compression ** push down filters can't be evaluated * the start of each row group needs to be found by scanning * user metadata can only be added to the file when the file is created * the file doesn't store the number of rows per a file or row group * there is no mechanism for seeking to a particular row number, which is required for external indexes. * there is no mechanism for storing light weight indexes within the file to enable push-down filters to skip entire row groups. * the type of the rows aren't stored in the file -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HIVE-3888) wrong mapside groupby if no partition is being selected
[ https://issues.apache.org/jira/browse/HIVE-3888?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13551368#comment-13551368 ] Hudson commented on HIVE-3888: -- Integrated in Hive-trunk-hadoop2 #59 (See [https://builds.apache.org/job/Hive-trunk-hadoop2/59/]) HIVE-3888 wrong mapside groupby if no partition is being selected (Namit Jain via Ashutosh and namit) (Revision 1431936) Result = FAILURE namit : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1431936 Files : * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/optimizer/GroupByOptimizer.java * /hive/trunk/ql/src/test/queries/clientpositive/groupby_sort_6.q * /hive/trunk/ql/src/test/queries/clientpositive/groupby_sort_7.q * /hive/trunk/ql/src/test/results/clientpositive/groupby_sort_6.q.out * /hive/trunk/ql/src/test/results/clientpositive/groupby_sort_7.q.out * /hive/trunk/ql/src/test/results/clientpositive/metadataonly1.q.out wrong mapside groupby if no partition is being selected --- Key: HIVE-3888 URL: https://issues.apache.org/jira/browse/HIVE-3888 Project: Hive Issue Type: Bug Components: Query Processor Reporter: Namit Jain Assignee: Namit Jain Fix For: 0.11.0 Attachments: hive.3888.1.patch -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
Hive-trunk-hadoop2 - Build # 59 - Failure
38 tests failed. FAILED: org.apache.hadoop.hive.cli.TestHBaseCliDriver.testCliDriver_external_table_ppd Error Message: Unexpected exception in setup Stack Trace: junit.framework.AssertionFailedError: Unexpected exception in setup at junit.framework.Assert.fail(Assert.java:50) at org.apache.hadoop.hive.cli.TestHBaseCliDriver.setUp(TestHBaseCliDriver.java:59) at junit.framework.TestCase.runBare(TestCase.java:132) at junit.framework.TestResult$1.protect(TestResult.java:110) at junit.framework.TestResult.runProtected(TestResult.java:128) at junit.framework.TestResult.run(TestResult.java:113) at junit.framework.TestCase.run(TestCase.java:124) at junit.framework.TestSuite.runTest(TestSuite.java:243) at junit.framework.TestSuite.run(TestSuite.java:238) at junit.extensions.TestDecorator.basicRun(TestDecorator.java:24) at junit.extensions.TestSetup$1.protect(TestSetup.java:23) at junit.framework.TestResult.runProtected(TestResult.java:128) at junit.extensions.TestSetup.run(TestSetup.java:27) at org.apache.tools.ant.taskdefs.optional.junit.JUnitTestRunner.run(JUnitTestRunner.java:422) at org.apache.tools.ant.taskdefs.optional.junit.JUnitTestRunner.launch(JUnitTestRunner.java:931) at org.apache.tools.ant.taskdefs.optional.junit.JUnitTestRunner.main(JUnitTestRunner.java:785) FAILED: org.apache.hadoop.hive.cli.TestHBaseCliDriver.testCliDriver_hbase_binary_external_table_queries Error Message: Unexpected exception in setup Stack Trace: junit.framework.AssertionFailedError: Unexpected exception in setup at junit.framework.Assert.fail(Assert.java:50) at org.apache.hadoop.hive.cli.TestHBaseCliDriver.setUp(TestHBaseCliDriver.java:59) at junit.framework.TestCase.runBare(TestCase.java:132) at junit.framework.TestResult$1.protect(TestResult.java:110) at junit.framework.TestResult.runProtected(TestResult.java:128) at junit.framework.TestResult.run(TestResult.java:113) at junit.framework.TestCase.run(TestCase.java:124) at junit.framework.TestSuite.runTest(TestSuite.java:243) at junit.framework.TestSuite.run(TestSuite.java:238) at junit.extensions.TestDecorator.basicRun(TestDecorator.java:24) at junit.extensions.TestSetup$1.protect(TestSetup.java:23) at junit.framework.TestResult.runProtected(TestResult.java:128) at junit.extensions.TestSetup.run(TestSetup.java:27) at org.apache.tools.ant.taskdefs.optional.junit.JUnitTestRunner.run(JUnitTestRunner.java:422) at org.apache.tools.ant.taskdefs.optional.junit.JUnitTestRunner.launch(JUnitTestRunner.java:931) at org.apache.tools.ant.taskdefs.optional.junit.JUnitTestRunner.main(JUnitTestRunner.java:785) FAILED: org.apache.hadoop.hive.cli.TestHBaseCliDriver.testCliDriver_hbase_binary_map_queries Error Message: Unexpected exception in setup Stack Trace: junit.framework.AssertionFailedError: Unexpected exception in setup at junit.framework.Assert.fail(Assert.java:50) at org.apache.hadoop.hive.cli.TestHBaseCliDriver.setUp(TestHBaseCliDriver.java:59) at junit.framework.TestCase.runBare(TestCase.java:132) at junit.framework.TestResult$1.protect(TestResult.java:110) at junit.framework.TestResult.runProtected(TestResult.java:128) at junit.framework.TestResult.run(TestResult.java:113) at junit.framework.TestCase.run(TestCase.java:124) at junit.framework.TestSuite.runTest(TestSuite.java:243) at junit.framework.TestSuite.run(TestSuite.java:238) at junit.extensions.TestDecorator.basicRun(TestDecorator.java:24) at junit.extensions.TestSetup$1.protect(TestSetup.java:23) at junit.framework.TestResult.runProtected(TestResult.java:128) at junit.extensions.TestSetup.run(TestSetup.java:27) at org.apache.tools.ant.taskdefs.optional.junit.JUnitTestRunner.run(JUnitTestRunner.java:422) at org.apache.tools.ant.taskdefs.optional.junit.JUnitTestRunner.launch(JUnitTestRunner.java:931) at org.apache.tools.ant.taskdefs.optional.junit.JUnitTestRunner.main(JUnitTestRunner.java:785) FAILED: org.apache.hadoop.hive.cli.TestHBaseCliDriver.testCliDriver_hbase_binary_storage_queries Error Message: Unexpected exception in setup Stack Trace: junit.framework.AssertionFailedError: Unexpected exception in setup at junit.framework.Assert.fail(Assert.java:50) at org.apache.hadoop.hive.cli.TestHBaseCliDriver.setUp(TestHBaseCliDriver.java:59) at junit.framework.TestCase.runBare(TestCase.java:132) at junit.framework.TestResult$1.protect(TestResult.java:110) at junit.framework.TestResult.runProtected(TestResult.java:128) at junit.framework.TestResult.run(TestResult.java:113) at junit.framework.TestCase.run(TestCase.java:124) at
[jira] [Commented] (HIVE-2439) Upgrade antlr version to 3.4
[ https://issues.apache.org/jira/browse/HIVE-2439?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13551439#comment-13551439 ] Ashutosh Chauhan commented on HIVE-2439: Thanks Thriuvel for addressing comments. Latest patch looks good. +1. Running tests now. Upgrade antlr version to 3.4 Key: HIVE-2439 URL: https://issues.apache.org/jira/browse/HIVE-2439 Project: Hive Issue Type: Improvement Affects Versions: 0.8.0 Reporter: Ashutosh Chauhan Assignee: Thiruvel Thirumoolan Attachments: HIVE-2439_branch10.patch, HIVE-2439_branch9_2.patch, HIVE-2439_branch9_3.patch, HIVE-2439_branch9.patch, hive-2439_incomplete.patch, HIVE-2439_trunk_1.patch, HIVE-2439_trunk.patch Upgrade antlr version to 3.4 -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Reopened] (HIVE-3810) HiveHistory.log need to replace '\r' with space before writing Entry.value to historyfile
[ https://issues.apache.org/jira/browse/HIVE-3810?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Ashutosh Chauhan reopened HIVE-3810: I dont see this in svn commit log. Reopening for review. HiveHistory.log need to replace '\r' with space before writing Entry.value to historyfile - Key: HIVE-3810 URL: https://issues.apache.org/jira/browse/HIVE-3810 Project: Hive Issue Type: Bug Components: Logging Reporter: qiangwang Assignee: Mark Grover Priority: Minor Attachments: HIVE-3810.1.patch, HIVE-3810.2.patch HiveHistory.log will replace '\n' with space before writing Entry.value to history file: val = val.replace('\n', ' '); but HiveHistory.parseHiveHistory use BufferedReader.readLine which takes '\n', '\r', '\r\n' as line delimiter to parse history file if val contains '\r', there is a high possibility that HiveHistory.parseLine will fail, in which case usually RecordTypes.valueOf(recType) will throw exception 'java.lang.IllegalArgumentException' HiveHistory.log need to replace '\r' with space as well: val = val.replace('\n', ' '); changed to val = val.replaceAll(\r|\n, ); or val = val.replace('\r', ' ').replace('\n', ' '); -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Reopened] (HIVE-3857) HiveHistoryViewer concurrency problem
[ https://issues.apache.org/jira/browse/HIVE-3857?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Ashutosh Chauhan reopened HIVE-3857: Assignee: qiangwang This is not committed yet. Reopening for review. HiveHistoryViewer concurrency problem - Key: HIVE-3857 URL: https://issues.apache.org/jira/browse/HIVE-3857 Project: Hive Issue Type: Bug Components: Logging Reporter: qiangwang Assignee: qiangwang Priority: Minor Attachments: HiveHistory.patch new HiveHistoryViewer() throws ConcurrentModificationException when called concurrently by several threads. According to the stacktrace, HiveHistory.parseLine use private static MapString, String parseBuffer to store parsed data and this finally caused the exception. I don't know why a static buffer rather than a local buffer is used! -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HIVE-3824) bug if different serdes are used for different partitions
[ https://issues.apache.org/jira/browse/HIVE-3824?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13551479#comment-13551479 ] Ashutosh Chauhan commented on HIVE-3824: Namit, Feel free to commit this if tests pass. bug if different serdes are used for different partitions - Key: HIVE-3824 URL: https://issues.apache.org/jira/browse/HIVE-3824 Project: Hive Issue Type: Bug Components: Query Processor Reporter: Namit Jain Assignee: Namit Jain Attachments: hive.3824.1.patch, hive.3824.3.patch Consider the following testcase: create table tst5 (key string, value string) partitioned by (ds string) stored as rcfile; insert overwrite table tst5 partition (ds='1') select * from src; insert overwrite table tst5 partition (ds='2') select * from src; insert overwrite table tst5 partition (ds='3') select * from src; alter table tst5 stored as sequencefile; insert overwrite table tst5 partition (ds='4') select * from src; insert overwrite table tst5 partition (ds='5') select * from src; insert overwrite table tst5 partition (ds='6') select * from src; alter table tst5 set serde 'org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'; insert overwrite table tst5 partition (ds='7') select * from src; insert overwrite table tst5 partition (ds='8') select * from src; insert overwrite table tst5 partition (ds='9') select * from src; The following query works fine: select key + key, value from tst5 where ((ds = '4') or (ds = '1')); since both the partitions use ColumnarSerDe But the following query fails: select key + key, value from tst5 where ((ds = '4') or (ds = '1') or (ds='7')); since different serdes are used. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HIVE-3888) wrong mapside groupby if no partition is being selected
[ https://issues.apache.org/jira/browse/HIVE-3888?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13551498#comment-13551498 ] Hudson commented on HIVE-3888: -- Integrated in Hive-trunk-h0.21 #1907 (See [https://builds.apache.org/job/Hive-trunk-h0.21/1907/]) HIVE-3888 wrong mapside groupby if no partition is being selected (Namit Jain via Ashutosh and namit) (Revision 1431936) Result = ABORTED namit : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1431936 Files : * /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/optimizer/GroupByOptimizer.java * /hive/trunk/ql/src/test/queries/clientpositive/groupby_sort_6.q * /hive/trunk/ql/src/test/queries/clientpositive/groupby_sort_7.q * /hive/trunk/ql/src/test/results/clientpositive/groupby_sort_6.q.out * /hive/trunk/ql/src/test/results/clientpositive/groupby_sort_7.q.out * /hive/trunk/ql/src/test/results/clientpositive/metadataonly1.q.out wrong mapside groupby if no partition is being selected --- Key: HIVE-3888 URL: https://issues.apache.org/jira/browse/HIVE-3888 Project: Hive Issue Type: Bug Components: Query Processor Reporter: Namit Jain Assignee: Namit Jain Fix For: 0.11.0 Attachments: hive.3888.1.patch -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Updated] (HIVE-3567) Hive List Bucketing - Query logic supports , =, , = and in
[ https://issues.apache.org/jira/browse/HIVE-3567?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Gang Tim Liu updated HIVE-3567: --- Description: Query logic supports equal. It should handle , =, , = and in also. (was: Query logic supports equal. It should handle , =, and = also.) Summary: Hive List Bucketing - Query logic supports , =, , = and in (was: Hive List Bucketing - Query logic supports , =, and =) Hive List Bucketing - Query logic supports , =, , = and in --- Key: HIVE-3567 URL: https://issues.apache.org/jira/browse/HIVE-3567 Project: Hive Issue Type: New Feature Affects Versions: 0.10.0 Reporter: Gang Tim Liu Assignee: Gang Tim Liu Priority: Minor Query logic supports equal. It should handle , =, , = and in also. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HIVE-3426) union with same source should be optimized
[ https://issues.apache.org/jira/browse/HIVE-3426?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13551621#comment-13551621 ] Ashutosh Chauhan commented on HIVE-3426: What Zhenxiao is proposing a good first step. Lets try to optimize that query first and worry about GBY in subquery later. union with same source should be optimized -- Key: HIVE-3426 URL: https://issues.apache.org/jira/browse/HIVE-3426 Project: Hive Issue Type: Improvement Components: Query Processor Reporter: Namit Jain Assignee: Zhenxiao Luo -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HIVE-942) use bucketing for group by
[ https://issues.apache.org/jira/browse/HIVE-942?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13551625#comment-13551625 ] Ashutosh Chauhan commented on HIVE-942: --- I think if BucketizedHiveInputFormat is used with map-side aggregation on, this should already work, no? Probably reducer phase will also get created, which is not required and can be optimized away. Is that the optimization we are talking about? use bucketing for group by -- Key: HIVE-942 URL: https://issues.apache.org/jira/browse/HIVE-942 Project: Hive Issue Type: New Feature Components: Query Processor Reporter: Namit Jain Group by on a bucketed column can be completely performed on the mapper if the split can be adjusted to span the key boundary. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
Re: [ANNOUNCE] Apache Hive 0.10.0 Released
Great work everyone! On Fri, Jan 11, 2013 at 12:37 AM, Lefty Leverenz le...@hortonworks.comwrote: Congratulations! – Lefty On Thu, Jan 10, 2013 at 11:59 PM, Ashutosh Chauhan hashut...@apache.org wrote: The Apache Hive team is proud to announce the the release of Apache Hive version 0.10.0. The Apache Hive (TM) data warehouse software facilitates querying and managing large datasets residing in distributed storage. Built on top of Apache Hadoop (TM), it provides: * Tools to enable easy data extract/transform/load (ETL) * A mechanism to impose structure on a variety of data formats * Access to files stored either directly in Apache HDFS (TM) or in other data storage systems such as Apache HBase (TM) * Query execution via MapReduce For Hive release details and downloads, please visit:http://hive.apache.org/releases.html Hive 0.10.0 Release Notes are available here: https://issues.apache.org/jira/secure/ReleaseNote.jspa?version=12320745styleName=TextprojectId=12310843 More than 350 jira issues are fixed in this release. We would like to thank many contributors who made this release possible. Regards, The Apache Hive Team
[jira] [Assigned] (HIVE-1649) Ability to update counters and status from TRANSFORM scripts
[ https://issues.apache.org/jira/browse/HIVE-1649?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Carl Steinbach reassigned HIVE-1649: Assignee: Guo Hongjie @Guo: Can you post a review request on reviewboard or phabricator? Directions explaining how to do this are on the wiki. Thanks. Ability to update counters and status from TRANSFORM scripts Key: HIVE-1649 URL: https://issues.apache.org/jira/browse/HIVE-1649 Project: Hive Issue Type: New Feature Components: Query Processor Affects Versions: 0.6.0 Reporter: Carl Steinbach Assignee: Guo Hongjie Attachments: HIVE-1649.1.patch Hadoop Streaming supports the ability to update counters and status by writing specially coded messages to the script's stderr stream. A streaming process can use the stderr to emit counter information. {{reporter:counter:group,counter,amount}} should be sent to stderr to update the counter. A streaming process can use the stderr to emit status information. To set a status, {{reporter:status:message}} should be sent to stderr. Hive should support these same features with its TRANSFORM mechanism. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Updated] (HIVE-1649) Ability to update counters and status from TRANSFORM scripts
[ https://issues.apache.org/jira/browse/HIVE-1649?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Carl Steinbach updated HIVE-1649: - Resolution: Duplicate Status: Resolved (was: Patch Available) Ability to update counters and status from TRANSFORM scripts Key: HIVE-1649 URL: https://issues.apache.org/jira/browse/HIVE-1649 Project: Hive Issue Type: New Feature Components: Query Processor Affects Versions: 0.6.0 Reporter: Carl Steinbach Assignee: Guo Hongjie Attachments: HIVE-1649.1.patch Hadoop Streaming supports the ability to update counters and status by writing specially coded messages to the script's stderr stream. A streaming process can use the stderr to emit counter information. {{reporter:counter:group,counter,amount}} should be sent to stderr to update the counter. A streaming process can use the stderr to emit status information. To set a status, {{reporter:status:message}} should be sent to stderr. Hive should support these same features with its TRANSFORM mechanism. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Assigned] (HIVE-305) Port Hadoop streaming's counters/status reporters to Hive Transforms
[ https://issues.apache.org/jira/browse/HIVE-305?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Carl Steinbach reassigned HIVE-305: --- Assignee: Guo Hongjie @Guo: please submit a review request on reviewboard or phabricator. Thanks. Port Hadoop streaming's counters/status reporters to Hive Transforms Key: HIVE-305 URL: https://issues.apache.org/jira/browse/HIVE-305 Project: Hive Issue Type: New Feature Components: Query Processor Reporter: Venky Iyer Assignee: Guo Hongjie Attachments: HIVE-305.1.patch https://issues.apache.org/jira/browse/HADOOP-1328 Introduced a way for a streaming process to update global counters and status using stderr stream to emit information. Use reporter:counter:group,counter,amount to update a counter. Use reporter:status:message to update status. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Updated] (HIVE-3004) RegexSerDe should support other column types in addition to STRING
[ https://issues.apache.org/jira/browse/HIVE-3004?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Shreepadma Venugopalan updated HIVE-3004: - Status: Patch Available (was: Open) RegexSerDe should support other column types in addition to STRING -- Key: HIVE-3004 URL: https://issues.apache.org/jira/browse/HIVE-3004 Project: Hive Issue Type: Improvement Components: Serializers/Deserializers Reporter: Carl Steinbach Assignee: Shreepadma Venugopalan Attachments: HIVE-3004-1.patch, HIVE-3004.2.patch -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Updated] (HIVE-3004) RegexSerDe should support other column types in addition to STRING
[ https://issues.apache.org/jira/browse/HIVE-3004?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Shreepadma Venugopalan updated HIVE-3004: - Attachment: HIVE-3004.3.patch.txt RegexSerDe should support other column types in addition to STRING -- Key: HIVE-3004 URL: https://issues.apache.org/jira/browse/HIVE-3004 Project: Hive Issue Type: Improvement Components: Serializers/Deserializers Reporter: Carl Steinbach Assignee: Shreepadma Venugopalan Attachments: HIVE-3004-1.patch, HIVE-3004.2.patch, HIVE-3004.3.patch.txt -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HIVE-3004) RegexSerDe should support other column types in addition to STRING
[ https://issues.apache.org/jira/browse/HIVE-3004?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13551697#comment-13551697 ] Shreepadma Venugopalan commented on HIVE-3004: -- Thanks Ashutosh. I've attached the new patch to the JIRA. RegexSerDe should support other column types in addition to STRING -- Key: HIVE-3004 URL: https://issues.apache.org/jira/browse/HIVE-3004 Project: Hive Issue Type: Improvement Components: Serializers/Deserializers Reporter: Carl Steinbach Assignee: Shreepadma Venugopalan Attachments: HIVE-3004-1.patch, HIVE-3004.2.patch, HIVE-3004.3.patch.txt -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
Review Request: HIVE-3004: RegexSerDe should support other column types in addition to STRING
--- This is an automatically generated e-mail. To reply, visit: https://reviews.apache.org/r/8931/ --- Review request for hive. Description --- This patch enhances regex serde to parse column types other than STRING. Only primitive types are supported. Diffs - ql/src/test/queries/clientnegative/serde_regex.q 6603b91 ql/src/test/queries/clientpositive/serde_regex.q c6809cb ql/src/test/results/clientnegative/serde_regex.q.out 03fe907 ql/src/test/results/clientpositive/serde_regex.q.out a8ce604 serde/src/java/org/apache/hadoop/hive/serde2/RegexSerDe.java e728244 Diff: https://reviews.apache.org/r/8931/diff/ Testing --- New test cases have been added and they pass. Thanks, Shreepadma Venugopalan
Re: Review Request: HIVE-3004: RegexSerDe should support other column types in addition to STRING
--- This is an automatically generated e-mail. To reply, visit: https://reviews.apache.org/r/8931/ --- (Updated Jan. 12, 2013, 12:28 a.m.) Review request for hive and Ashutosh Chauhan. Description --- This patch enhances regex serde to parse column types other than STRING. Only primitive types are supported. This addresses bug HIVE-3004. https://issues.apache.org/jira/browse/HIVE-3004 Diffs - ql/src/test/queries/clientnegative/serde_regex.q 6603b91 ql/src/test/queries/clientpositive/serde_regex.q c6809cb ql/src/test/results/clientnegative/serde_regex.q.out 03fe907 ql/src/test/results/clientpositive/serde_regex.q.out a8ce604 serde/src/java/org/apache/hadoop/hive/serde2/RegexSerDe.java e728244 Diff: https://reviews.apache.org/r/8931/diff/ Testing --- New test cases have been added and they pass. Thanks, Shreepadma Venugopalan
[jira] [Commented] (HIVE-3004) RegexSerDe should support other column types in addition to STRING
[ https://issues.apache.org/jira/browse/HIVE-3004?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13551698#comment-13551698 ] Shreepadma Venugopalan commented on HIVE-3004: -- Review board : https://reviews.apache.org/r/8931/ RegexSerDe should support other column types in addition to STRING -- Key: HIVE-3004 URL: https://issues.apache.org/jira/browse/HIVE-3004 Project: Hive Issue Type: Improvement Components: Serializers/Deserializers Reporter: Carl Steinbach Assignee: Shreepadma Venugopalan Attachments: HIVE-3004-1.patch, HIVE-3004.2.patch, HIVE-3004.3.patch.txt -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HIVE-3426) union with same source should be optimized
[ https://issues.apache.org/jira/browse/HIVE-3426?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13551701#comment-13551701 ] Shreepadma Venugopalan commented on HIVE-3426: -- Yup, let's try to optimize the simple case first. Optimizing subqueries with GBY can be the next step. union with same source should be optimized -- Key: HIVE-3426 URL: https://issues.apache.org/jira/browse/HIVE-3426 Project: Hive Issue Type: Improvement Components: Query Processor Reporter: Namit Jain Assignee: Zhenxiao Luo -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
Re: [DISCUSS] HCatalog becoming a subproject of Hive
Hi Alan, I agree that submitting this for a vote is the best option. If anyone has additional proposed modifications please make them. Otherwise I propose that the Hive PMC vote on this proposal. In order for the Hive PMC to be able to vote on these changes they need to be expressed in terms of one or more of the actions listed at the end of the Hive project bylaws: https://cwiki.apache.org/confluence/display/Hive/Bylaws So I think we first need to amend to the bylaws in order to define the rights and privileges of a submodule committer, and then separately vote the HCatalog committers in as Hive submodule committers. Does this make sense? Thanks. Carl
[jira] [Assigned] (HIVE-3653) Failure in a counter poller run should not be considered as a job failure
[ https://issues.apache.org/jira/browse/HIVE-3653?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Shreepadma Venugopalan reassigned HIVE-3653: Assignee: Shreepadma Venugopalan Failure in a counter poller run should not be considered as a job failure - Key: HIVE-3653 URL: https://issues.apache.org/jira/browse/HIVE-3653 Project: Hive Issue Type: Bug Components: Clients Affects Versions: 0.7.1 Reporter: Harsh J Assignee: Shreepadma Venugopalan A client had a simple transient failure in polling the JT for job status (which it does for HIVECOUNTERSPULLINTERVAL for each currently running job). {code} java.io.IOException: Call to HOST/IP:PORT failed on local exception: java.io.IOException: Connection reset by peer at org.apache.hadoop.ipc.Client.wrapException(Client.java:1142) at org.apache.hadoop.ipc.Client.call(Client.java:1110) at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:226) at org.apache.hadoop.mapred.$Proxy10.getJobStatus(Unknown Source) at org.apache.hadoop.mapred.JobClient.getJob(JobClient.java:1053) at org.apache.hadoop.mapred.JobClient.getJob(JobClient.java:1065) at org.apache.hadoop.hive.ql.exec.ExecDriver.progress(ExecDriver.java:351) at org.apache.hadoop.hive.ql.exec.ExecDriver.execute(ExecDriver.java:686) at org.apache.hadoop.hive.ql.exec.MapRedTask.execute(MapRedTask.java:123) at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:131) at org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:57) at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:1063) at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:900) at org.apache.hadoop.hive.ql.Driver.run(Driver.java:748) at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:209) at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:286) at org.apache.hadoop.hive.cli.CliDriver.processReader(CliDriver.java:310) at org.apache.hadoop.hive.cli.CliDriver.processFile(CliDriver.java:317) at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:490) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25) at java.lang.reflect.Method.invoke(Method.java:597) at org.apache.hadoop.util.RunJar.main(RunJar.java:197) {code} This lead to Hive thinking the running job itself has failed, and it failed the query run, although the running job progressed to completion in the background. We should not let transient IOExceptions in counter polling cause query termination, and should instead just retry. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Updated] (HIVE-3803) explain dependency should show the dependencies hierarchically in presence of views
[ https://issues.apache.org/jira/browse/HIVE-3803?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Sambavi Muthukrishnan updated HIVE-3803: Attachment: hive.3803.13.patch explain dependency should show the dependencies hierarchically in presence of views --- Key: HIVE-3803 URL: https://issues.apache.org/jira/browse/HIVE-3803 Project: Hive Issue Type: Bug Components: Query Processor Reporter: Namit Jain Assignee: Namit Jain Attachments: hive.3803.10.patch, hive.3803.11.patch, hive.3803.12.patch, hive.3803.13.patch, hive.3803.1.patch, hive.3803.2.patch, hive.3803.3.patch, hive.3803.4.patch, hive.3803.5.patch, hive.3803.6.patch, hive.3803.7.patch, hive.3803.8.patch, hive.3803.9.patch It should also include tables whose partitions are being accessed -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HIVE-3803) explain dependency should show the dependencies hierarchically in presence of views
[ https://issues.apache.org/jira/browse/HIVE-3803?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13551747#comment-13551747 ] Sambavi Muthukrishnan commented on HIVE-3803: - Uploaded new patch with the updates to some test files. explain dependency should show the dependencies hierarchically in presence of views --- Key: HIVE-3803 URL: https://issues.apache.org/jira/browse/HIVE-3803 Project: Hive Issue Type: Bug Components: Query Processor Reporter: Namit Jain Assignee: Namit Jain Attachments: hive.3803.10.patch, hive.3803.11.patch, hive.3803.12.patch, hive.3803.13.patch, hive.3803.1.patch, hive.3803.2.patch, hive.3803.3.patch, hive.3803.4.patch, hive.3803.5.patch, hive.3803.6.patch, hive.3803.7.patch, hive.3803.8.patch, hive.3803.9.patch It should also include tables whose partitions are being accessed -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Updated] (HIVE-305) Port Hadoop streaming's counters/status reporters to Hive Transforms
[ https://issues.apache.org/jira/browse/HIVE-305?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Guo Hongjie updated HIVE-305: - Attachment: HIVE-305.2.patch Port Hadoop streaming's counters/status reporters to Hive Transforms Key: HIVE-305 URL: https://issues.apache.org/jira/browse/HIVE-305 Project: Hive Issue Type: New Feature Components: Query Processor Reporter: Venky Iyer Assignee: Guo Hongjie Attachments: HIVE-305.1.patch, HIVE-305.2.patch https://issues.apache.org/jira/browse/HADOOP-1328 Introduced a way for a streaming process to update global counters and status using stderr stream to emit information. Use reporter:counter:group,counter,amount to update a counter. Use reporter:status:message to update status. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HIVE-3803) explain dependency should show the dependencies hierarchically in presence of views
[ https://issues.apache.org/jira/browse/HIVE-3803?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13551783#comment-13551783 ] Kevin Wilfong commented on HIVE-3803: - +1 explain dependency should show the dependencies hierarchically in presence of views --- Key: HIVE-3803 URL: https://issues.apache.org/jira/browse/HIVE-3803 Project: Hive Issue Type: Bug Components: Query Processor Reporter: Namit Jain Assignee: Namit Jain Attachments: hive.3803.10.patch, hive.3803.11.patch, hive.3803.12.patch, hive.3803.13.patch, hive.3803.1.patch, hive.3803.2.patch, hive.3803.3.patch, hive.3803.4.patch, hive.3803.5.patch, hive.3803.6.patch, hive.3803.7.patch, hive.3803.8.patch, hive.3803.9.patch It should also include tables whose partitions are being accessed -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HIVE-305) Port Hadoop streaming's counters/status reporters to Hive Transforms
[ https://issues.apache.org/jira/browse/HIVE-305?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13551787#comment-13551787 ] Guo Hongjie commented on HIVE-305: -- thanks @Carl,do you mean submit patch button? I cannot find any button about review request. Port Hadoop streaming's counters/status reporters to Hive Transforms Key: HIVE-305 URL: https://issues.apache.org/jira/browse/HIVE-305 Project: Hive Issue Type: New Feature Components: Query Processor Reporter: Venky Iyer Assignee: Guo Hongjie Attachments: HIVE-305.1.patch, HIVE-305.2.patch https://issues.apache.org/jira/browse/HADOOP-1328 Introduced a way for a streaming process to update global counters and status using stderr stream to emit information. Use reporter:counter:group,counter,amount to update a counter. Use reporter:status:message to update status. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
Review Request: HIVE-305: Port Hadoop streaming's counters/status reporters to Hive Transforms
--- This is an automatically generated e-mail. To reply, visit: https://reviews.apache.org/r/8933/ --- Review request for hive. Description --- Port Hadoop streaming's counters/status reporters to Hive Transforms. https://issues.apache.org/jira/browse/HADOOP-1328 Introduced a way for a streaming process to update global counters and status using stderr stream to emit information. Use reporter:counter:group,counter,amount to update a counter. Use reporter:status:message to update status. This addresses bug HIVE-305. https://issues.apache.org/jira/browse/HIVE-305 Diffs - http://svn.apache.org/repos/asf/hive/trunk/common/src/java/org/apache/hadoop/hive/conf/HiveConf.java 1431196 http://svn.apache.org/repos/asf/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/ScriptOperator.java 1431196 Diff: https://reviews.apache.org/r/8933/diff/ Testing --- just simple test it. No new test case because I couldn't find the good way to write it . Any suggestions are welcome. Thanks, Guo Hongjie
[jira] [Commented] (HIVE-305) Port Hadoop streaming's counters/status reporters to Hive Transforms
[ https://issues.apache.org/jira/browse/HIVE-305?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13551801#comment-13551801 ] Guo Hongjie commented on HIVE-305: -- i found that in the mail list.thanks Port Hadoop streaming's counters/status reporters to Hive Transforms Key: HIVE-305 URL: https://issues.apache.org/jira/browse/HIVE-305 Project: Hive Issue Type: New Feature Components: Query Processor Reporter: Venky Iyer Assignee: Guo Hongjie Attachments: HIVE-305.1.patch, HIVE-305.2.patch https://issues.apache.org/jira/browse/HADOOP-1328 Introduced a way for a streaming process to update global counters and status using stderr stream to emit information. Use reporter:counter:group,counter,amount to update a counter. Use reporter:status:message to update status. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira