[ANNOUNCE] Apache Hive 0.10.0 Released

2013-01-11 Thread Ashutosh Chauhan
The Apache Hive team is proud to announce the the release of Apache
Hive version 0.10.0.

The Apache Hive (TM) data warehouse software facilitates querying and
managing large datasets residing in distributed storage. Built on top
of Apache Hadoop (TM), it provides:

* Tools to enable easy data extract/transform/load (ETL)

* A mechanism to impose structure on a variety of data formats

* Access to files stored either directly in Apache HDFS (TM) or in other
  data storage systems such as Apache HBase (TM)

* Query execution via MapReduce

For Hive release details and downloads, please
visit:http://hive.apache.org/releases.html

Hive 0.10.0 Release Notes are available
here:https://issues.apache.org/jira/secure/ReleaseNote.jspa?version=12320745styleName=TextprojectId=12310843

More than 350 jira issues are fixed in this release. We would like to
thank many contributors who made this release possible.

Regards,

The Apache Hive Team


[jira] [Commented] (HIVE-2693) Add DECIMAL data type

2013-01-11 Thread Carl Steinbach (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-2693?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13550950#comment-13550950
 ] 

Carl Steinbach commented on HIVE-2693:
--

+1. Changes look good to me.

Ashutosh and Namit: let me know if you want more time to look at this. 
Otherwise I'll plan to commit it in 24 hours. Thanks.

 Add DECIMAL data type
 -

 Key: HIVE-2693
 URL: https://issues.apache.org/jira/browse/HIVE-2693
 Project: Hive
  Issue Type: New Feature
  Components: Query Processor, Types
Affects Versions: 0.10.0
Reporter: Carl Steinbach
Assignee: Prasad Mujumdar
 Attachments: 2693_7.patch, 2693_8.patch, 2693_fix_all_tests1.patch, 
 HIVE-2693-10.patch, HIVE-2693-11.patch, HIVE-2693-12-SortableSerDe.patch, 
 HIVE-2693-13.patch, HIVE-2693-14.patch, HIVE-2693-15.patch, 
 HIVE-2693-16.patch, HIVE-2693-17.patch, HIVE-2693-18.patch, 
 HIVE-2693-19.patch, HIVE-2693-1.patch.txt, HIVE-2693-20.patch, 
 HIVE-2693-21.patch, HIVE-2693-all.patch, HIVE-2693.D7683.1.patch, 
 HIVE-2693-fix.patch, HIVE-2693.patch, HIVE-2693-take3.patch, 
 HIVE-2693-take4.patch


 Add support for the DECIMAL data type. HIVE-2272 (TIMESTAMP) provides a nice 
 template for how to do this.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-3403) user should not specify mapjoin to perform sort-merge bucketed join

2013-01-11 Thread Namit Jain (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-3403?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Namit Jain updated HIVE-3403:
-

Attachment: hive.3403.11.patch

 user should not specify mapjoin to perform sort-merge bucketed join
 ---

 Key: HIVE-3403
 URL: https://issues.apache.org/jira/browse/HIVE-3403
 Project: Hive
  Issue Type: Bug
Reporter: Namit Jain
Assignee: Namit Jain
 Attachments: hive.3403.10.patch, hive.3403.11.patch, 
 hive.3403.1.patch, hive.3403.2.patch, hive.3403.3.patch, hive.3403.4.patch, 
 hive.3403.5.patch, hive.3403.6.patch, hive.3403.7.patch, hive.3403.8.patch, 
 hive.3403.9.patch


 Currently, in order to perform a sort merge bucketed join, the user needs
 to set hive.optimize.bucketmapjoin.sortedmerge to true, and also specify the 
 mapjoin hint.
 The user should not specify any hints.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-3810) HiveHistory.log need to replace '\r' with space before writing Entry.value to historyfile

2013-01-11 Thread qiangwang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-3810?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

qiangwang updated HIVE-3810:


Resolution: Fixed
Status: Resolved  (was: Patch Available)

 HiveHistory.log need to replace '\r' with space before writing Entry.value to 
 historyfile
 -

 Key: HIVE-3810
 URL: https://issues.apache.org/jira/browse/HIVE-3810
 Project: Hive
  Issue Type: Bug
  Components: Logging
Reporter: qiangwang
Assignee: Mark Grover
Priority: Minor
 Attachments: HIVE-3810.1.patch, HIVE-3810.2.patch


 HiveHistory.log will replace '\n' with space before writing Entry.value to 
 history file:
 val = val.replace('\n', ' ');
 but HiveHistory.parseHiveHistory use BufferedReader.readLine which takes 
 '\n', '\r', '\r\n'  as line delimiter to parse history file
 if val contains '\r', there is a high possibility that HiveHistory.parseLine 
 will fail, in which case usually RecordTypes.valueOf(recType) will throw 
 exception 'java.lang.IllegalArgumentException'
 HiveHistory.log need to replace '\r' with space as well:
 val = val.replace('\n', ' ');
 changed to
 val = val.replaceAll(\r|\n,  );
 or
 val = val.replace('\r', ' ').replace('\n', ' ');

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-3857) HiveHistoryViewer concurrency problem

2013-01-11 Thread qiangwang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-3857?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

qiangwang updated HIVE-3857:


Resolution: Fixed
Status: Resolved  (was: Patch Available)

 HiveHistoryViewer concurrency problem
 -

 Key: HIVE-3857
 URL: https://issues.apache.org/jira/browse/HIVE-3857
 Project: Hive
  Issue Type: Bug
  Components: Logging
Reporter: qiangwang
Priority: Minor
 Attachments: HiveHistory.patch


 new HiveHistoryViewer() throws ConcurrentModificationException when called 
 concurrently by several threads.
 According to the stacktrace, HiveHistory.parseLine use private static 
 MapString, String parseBuffer to store parsed data and this finally caused 
 the exception.
 I don't know why a static buffer rather than a local buffer is used!

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-3874) Create a new Optimized Row Columnar file format for Hive

2013-01-11 Thread Namit Jain (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-3874?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13550961#comment-13550961
 ] 

Namit Jain commented on HIVE-3874:
--

In partition metadata, can I somehow specify - dont create index ?
When that partition is over-written, the index would dis-appear.

I agree, we can think about that later.

 Create a new Optimized Row Columnar file format for Hive
 

 Key: HIVE-3874
 URL: https://issues.apache.org/jira/browse/HIVE-3874
 Project: Hive
  Issue Type: Improvement
  Components: Serializers/Deserializers
Reporter: Owen O'Malley
Assignee: Owen O'Malley
 Attachments: OrcFileIntro.pptx


 There are several limitations of the current RC File format that I'd like to 
 address by creating a new format:
 * each column value is stored as a binary blob, which means:
 ** the entire column value must be read, decompressed, and deserialized
 ** the file format can't use smarter type-specific compression
 ** push down filters can't be evaluated
 * the start of each row group needs to be found by scanning
 * user metadata can only be added to the file when the file is created
 * the file doesn't store the number of rows per a file or row group
 * there is no mechanism for seeking to a particular row number, which is 
 required for external indexes.
 * there is no mechanism for storing light weight indexes within the file to 
 enable push-down filters to skip entire row groups.
 * the type of the rows aren't stored in the file

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-3264) Add support for binary dataype to AvroSerde

2013-01-11 Thread Eli Reisman (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-3264?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13550963#comment-13550963
 ] 

Eli Reisman commented on HIVE-3264:
---

The patch isn't stale yet (whew!), I'm working on a test...


 Add support for binary dataype to AvroSerde
 ---

 Key: HIVE-3264
 URL: https://issues.apache.org/jira/browse/HIVE-3264
 Project: Hive
  Issue Type: Bug
  Components: Serializers/Deserializers
Affects Versions: 0.9.0
Reporter: Jakob Homan
  Labels: patch
 Attachments: HIVE-3264-1.patch, HIVE-3264-2.patch, HIVE-3264-3.patch, 
 HIVE-3264-4.patch, HIVE-3264-5.patch


 When the AvroSerde was written, Hive didn't have a binary type, so Avro's 
 byte array type is converted an array of small ints.  Now that HIVE-2380 is 
 in, this step isn't necessary and we can convert both Avro's bytes type and 
 probably fixed type to Hive's binary type.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-2693) Add DECIMAL data type

2013-01-11 Thread Ashutosh Chauhan (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-2693?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13550965#comment-13550965
 ] 

Ashutosh Chauhan commented on HIVE-2693:


Latest patch looks good to me. +1

 Add DECIMAL data type
 -

 Key: HIVE-2693
 URL: https://issues.apache.org/jira/browse/HIVE-2693
 Project: Hive
  Issue Type: New Feature
  Components: Query Processor, Types
Affects Versions: 0.10.0
Reporter: Carl Steinbach
Assignee: Prasad Mujumdar
 Attachments: 2693_7.patch, 2693_8.patch, 2693_fix_all_tests1.patch, 
 HIVE-2693-10.patch, HIVE-2693-11.patch, HIVE-2693-12-SortableSerDe.patch, 
 HIVE-2693-13.patch, HIVE-2693-14.patch, HIVE-2693-15.patch, 
 HIVE-2693-16.patch, HIVE-2693-17.patch, HIVE-2693-18.patch, 
 HIVE-2693-19.patch, HIVE-2693-1.patch.txt, HIVE-2693-20.patch, 
 HIVE-2693-21.patch, HIVE-2693-all.patch, HIVE-2693.D7683.1.patch, 
 HIVE-2693-fix.patch, HIVE-2693.patch, HIVE-2693-take3.patch, 
 HIVE-2693-take4.patch


 Add support for the DECIMAL data type. HIVE-2272 (TIMESTAMP) provides a nice 
 template for how to do this.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-2693) Add DECIMAL data type

2013-01-11 Thread Namit Jain (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-2693?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13550970#comment-13550970
 ] 

Namit Jain commented on HIVE-2693:
--

Carl, give me some time. I will try to take a look today/tomorrow.
Can you give me ~2 days ? Go ahead, if you dont hear from me.

 Add DECIMAL data type
 -

 Key: HIVE-2693
 URL: https://issues.apache.org/jira/browse/HIVE-2693
 Project: Hive
  Issue Type: New Feature
  Components: Query Processor, Types
Affects Versions: 0.10.0
Reporter: Carl Steinbach
Assignee: Prasad Mujumdar
 Attachments: 2693_7.patch, 2693_8.patch, 2693_fix_all_tests1.patch, 
 HIVE-2693-10.patch, HIVE-2693-11.patch, HIVE-2693-12-SortableSerDe.patch, 
 HIVE-2693-13.patch, HIVE-2693-14.patch, HIVE-2693-15.patch, 
 HIVE-2693-16.patch, HIVE-2693-17.patch, HIVE-2693-18.patch, 
 HIVE-2693-19.patch, HIVE-2693-1.patch.txt, HIVE-2693-20.patch, 
 HIVE-2693-21.patch, HIVE-2693-all.patch, HIVE-2693.D7683.1.patch, 
 HIVE-2693-fix.patch, HIVE-2693.patch, HIVE-2693-take3.patch, 
 HIVE-2693-take4.patch


 Add support for the DECIMAL data type. HIVE-2272 (TIMESTAMP) provides a nice 
 template for how to do this.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-3888) wrong mapside groupby if no partition is being selected

2013-01-11 Thread Namit Jain (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-3888?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Namit Jain updated HIVE-3888:
-

Status: Patch Available  (was: Open)

 wrong mapside groupby if no partition is being selected
 ---

 Key: HIVE-3888
 URL: https://issues.apache.org/jira/browse/HIVE-3888
 Project: Hive
  Issue Type: Bug
  Components: Query Processor
Reporter: Namit Jain
Assignee: Namit Jain
 Attachments: hive.3888.1.patch




--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-3888) wrong mapside groupby if no partition is being selected

2013-01-11 Thread Namit Jain (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-3888?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Namit Jain updated HIVE-3888:
-

   Resolution: Fixed
Fix Version/s: 0.11.0
 Hadoop Flags: Reviewed
   Status: Resolved  (was: Patch Available)

Committed. Thanks Ashutosh

 wrong mapside groupby if no partition is being selected
 ---

 Key: HIVE-3888
 URL: https://issues.apache.org/jira/browse/HIVE-3888
 Project: Hive
  Issue Type: Bug
  Components: Query Processor
Reporter: Namit Jain
Assignee: Namit Jain
 Fix For: 0.11.0

 Attachments: hive.3888.1.patch




--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


Re: [ANNOUNCE] Apache Hive 0.10.0 Released

2013-01-11 Thread Lefty Leverenz
Congratulations!

– Lefty



On Thu, Jan 10, 2013 at 11:59 PM, Ashutosh Chauhan hashut...@apache.orgwrote:

 The Apache Hive team is proud to announce the the release of Apache
 Hive version 0.10.0.

 The Apache Hive (TM) data warehouse software facilitates querying and
 managing large datasets residing in distributed storage. Built on top
 of Apache Hadoop (TM), it provides:

 * Tools to enable easy data extract/transform/load (ETL)

 * A mechanism to impose structure on a variety of data formats

 * Access to files stored either directly in Apache HDFS (TM) or in other
   data storage systems such as Apache HBase (TM)

 * Query execution via MapReduce

 For Hive release details and downloads, please
 visit:http://hive.apache.org/releases.html

 Hive 0.10.0 Release Notes are available
 here:
 https://issues.apache.org/jira/secure/ReleaseNote.jspa?version=12320745styleName=TextprojectId=12310843

 More than 350 jira issues are fixed in this release. We would like to
 thank many contributors who made this release possible.

 Regards,

 The Apache Hive Team



[jira] [Commented] (HIVE-3888) wrong mapside groupby if no partition is being selected

2013-01-11 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-3888?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13551021#comment-13551021
 ] 

Hudson commented on HIVE-3888:
--

Integrated in hive-trunk-hadoop1 #7 (See 
[https://builds.apache.org/job/hive-trunk-hadoop1/7/])
HIVE-3888 wrong mapside groupby if no partition is being selected
(Namit Jain via Ashutosh and namit) (Revision 1431936)

 Result = ABORTED
namit : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1431936
Files : 
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/optimizer/GroupByOptimizer.java
* /hive/trunk/ql/src/test/queries/clientpositive/groupby_sort_6.q
* /hive/trunk/ql/src/test/queries/clientpositive/groupby_sort_7.q
* /hive/trunk/ql/src/test/results/clientpositive/groupby_sort_6.q.out
* /hive/trunk/ql/src/test/results/clientpositive/groupby_sort_7.q.out
* /hive/trunk/ql/src/test/results/clientpositive/metadataonly1.q.out


 wrong mapside groupby if no partition is being selected
 ---

 Key: HIVE-3888
 URL: https://issues.apache.org/jira/browse/HIVE-3888
 Project: Hive
  Issue Type: Bug
  Components: Query Processor
Reporter: Namit Jain
Assignee: Namit Jain
 Fix For: 0.11.0

 Attachments: hive.3888.1.patch




--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-3875) negative value for hive.stats.ndv.error should be disallowed

2013-01-11 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-3875?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13551044#comment-13551044
 ] 

Hudson commented on HIVE-3875:
--

Integrated in Hive-trunk-hadoop2 #58 (See 
[https://builds.apache.org/job/Hive-trunk-hadoop2/58/])
HIVE-3875. Negative value for hive.stats.ndv.error should be disallowed 
(Shreepadma Venugopalan via cws) (Revision 1431793)

 Result = FAILURE
cws : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1431793
Files : 
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/parse/ColumnStatsSemanticAnalyzer.java


 negative value for hive.stats.ndv.error should be disallowed 
 -

 Key: HIVE-3875
 URL: https://issues.apache.org/jira/browse/HIVE-3875
 Project: Hive
  Issue Type: Bug
  Components: Statistics
Affects Versions: 0.10.0
Reporter: Shreepadma Venugopalan
Assignee: Shreepadma Venugopalan
 Fix For: 0.11.0

 Attachments: HIVE-3875.1.patch.txt


 Currently, if a negative value is specified for hive.stats.ndv.error in 
 hive-site.xml, it is treated as 0. We should instead throw an exception.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


Hive-trunk-hadoop2 - Build # 58 - Failure

2013-01-11 Thread Apache Jenkins Server


36 tests failed.
FAILED:  
org.apache.hadoop.hive.cli.TestHBaseCliDriver.testCliDriver_external_table_ppd

Error Message:
Unexpected exception in setup

Stack Trace:
junit.framework.AssertionFailedError: Unexpected exception in setup
at junit.framework.Assert.fail(Assert.java:50)
at 
org.apache.hadoop.hive.cli.TestHBaseCliDriver.setUp(TestHBaseCliDriver.java:59)
at junit.framework.TestCase.runBare(TestCase.java:132)
at junit.framework.TestResult$1.protect(TestResult.java:110)
at junit.framework.TestResult.runProtected(TestResult.java:128)
at junit.framework.TestResult.run(TestResult.java:113)
at junit.framework.TestCase.run(TestCase.java:124)
at junit.framework.TestSuite.runTest(TestSuite.java:243)
at junit.framework.TestSuite.run(TestSuite.java:238)
at junit.extensions.TestDecorator.basicRun(TestDecorator.java:24)
at junit.extensions.TestSetup$1.protect(TestSetup.java:23)
at junit.framework.TestResult.runProtected(TestResult.java:128)
at junit.extensions.TestSetup.run(TestSetup.java:27)
at 
org.apache.tools.ant.taskdefs.optional.junit.JUnitTestRunner.run(JUnitTestRunner.java:422)
at 
org.apache.tools.ant.taskdefs.optional.junit.JUnitTestRunner.launch(JUnitTestRunner.java:931)
at 
org.apache.tools.ant.taskdefs.optional.junit.JUnitTestRunner.main(JUnitTestRunner.java:785)


FAILED:  
org.apache.hadoop.hive.cli.TestHBaseCliDriver.testCliDriver_hbase_binary_external_table_queries

Error Message:
Unexpected exception in setup

Stack Trace:
junit.framework.AssertionFailedError: Unexpected exception in setup
at junit.framework.Assert.fail(Assert.java:50)
at 
org.apache.hadoop.hive.cli.TestHBaseCliDriver.setUp(TestHBaseCliDriver.java:59)
at junit.framework.TestCase.runBare(TestCase.java:132)
at junit.framework.TestResult$1.protect(TestResult.java:110)
at junit.framework.TestResult.runProtected(TestResult.java:128)
at junit.framework.TestResult.run(TestResult.java:113)
at junit.framework.TestCase.run(TestCase.java:124)
at junit.framework.TestSuite.runTest(TestSuite.java:243)
at junit.framework.TestSuite.run(TestSuite.java:238)
at junit.extensions.TestDecorator.basicRun(TestDecorator.java:24)
at junit.extensions.TestSetup$1.protect(TestSetup.java:23)
at junit.framework.TestResult.runProtected(TestResult.java:128)
at junit.extensions.TestSetup.run(TestSetup.java:27)
at 
org.apache.tools.ant.taskdefs.optional.junit.JUnitTestRunner.run(JUnitTestRunner.java:422)
at 
org.apache.tools.ant.taskdefs.optional.junit.JUnitTestRunner.launch(JUnitTestRunner.java:931)
at 
org.apache.tools.ant.taskdefs.optional.junit.JUnitTestRunner.main(JUnitTestRunner.java:785)


FAILED:  
org.apache.hadoop.hive.cli.TestHBaseCliDriver.testCliDriver_hbase_binary_map_queries

Error Message:
Unexpected exception in setup

Stack Trace:
junit.framework.AssertionFailedError: Unexpected exception in setup
at junit.framework.Assert.fail(Assert.java:50)
at 
org.apache.hadoop.hive.cli.TestHBaseCliDriver.setUp(TestHBaseCliDriver.java:59)
at junit.framework.TestCase.runBare(TestCase.java:132)
at junit.framework.TestResult$1.protect(TestResult.java:110)
at junit.framework.TestResult.runProtected(TestResult.java:128)
at junit.framework.TestResult.run(TestResult.java:113)
at junit.framework.TestCase.run(TestCase.java:124)
at junit.framework.TestSuite.runTest(TestSuite.java:243)
at junit.framework.TestSuite.run(TestSuite.java:238)
at junit.extensions.TestDecorator.basicRun(TestDecorator.java:24)
at junit.extensions.TestSetup$1.protect(TestSetup.java:23)
at junit.framework.TestResult.runProtected(TestResult.java:128)
at junit.extensions.TestSetup.run(TestSetup.java:27)
at 
org.apache.tools.ant.taskdefs.optional.junit.JUnitTestRunner.run(JUnitTestRunner.java:422)
at 
org.apache.tools.ant.taskdefs.optional.junit.JUnitTestRunner.launch(JUnitTestRunner.java:931)
at 
org.apache.tools.ant.taskdefs.optional.junit.JUnitTestRunner.main(JUnitTestRunner.java:785)


FAILED:  
org.apache.hadoop.hive.cli.TestHBaseCliDriver.testCliDriver_hbase_binary_storage_queries

Error Message:
Unexpected exception in setup

Stack Trace:
junit.framework.AssertionFailedError: Unexpected exception in setup
at junit.framework.Assert.fail(Assert.java:50)
at 
org.apache.hadoop.hive.cli.TestHBaseCliDriver.setUp(TestHBaseCliDriver.java:59)
at junit.framework.TestCase.runBare(TestCase.java:132)
at junit.framework.TestResult$1.protect(TestResult.java:110)
at junit.framework.TestResult.runProtected(TestResult.java:128)
at junit.framework.TestResult.run(TestResult.java:113)
at junit.framework.TestCase.run(TestCase.java:124)
at 

[jira] [Updated] (HIVE-3403) user should not specify mapjoin to perform sort-merge bucketed join

2013-01-11 Thread Namit Jain (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-3403?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Namit Jain updated HIVE-3403:
-

Attachment: hive.3403.12.patch

 user should not specify mapjoin to perform sort-merge bucketed join
 ---

 Key: HIVE-3403
 URL: https://issues.apache.org/jira/browse/HIVE-3403
 Project: Hive
  Issue Type: Bug
Reporter: Namit Jain
Assignee: Namit Jain
 Attachments: hive.3403.10.patch, hive.3403.11.patch, 
 hive.3403.12.patch, hive.3403.1.patch, hive.3403.2.patch, hive.3403.3.patch, 
 hive.3403.4.patch, hive.3403.5.patch, hive.3403.6.patch, hive.3403.7.patch, 
 hive.3403.8.patch, hive.3403.9.patch


 Currently, in order to perform a sort merge bucketed join, the user needs
 to set hive.optimize.bucketmapjoin.sortedmerge to true, and also specify the 
 mapjoin hint.
 The user should not specify any hints.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-3875) negative value for hive.stats.ndv.error should be disallowed

2013-01-11 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-3875?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13551092#comment-13551092
 ] 

Hudson commented on HIVE-3875:
--

Integrated in Hive-trunk-h0.21 #1906 (See 
[https://builds.apache.org/job/Hive-trunk-h0.21/1906/])
HIVE-3875. Negative value for hive.stats.ndv.error should be disallowed 
(Shreepadma Venugopalan via cws) (Revision 1431793)

 Result = SUCCESS
cws : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1431793
Files : 
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/parse/ColumnStatsSemanticAnalyzer.java


 negative value for hive.stats.ndv.error should be disallowed 
 -

 Key: HIVE-3875
 URL: https://issues.apache.org/jira/browse/HIVE-3875
 Project: Hive
  Issue Type: Bug
  Components: Statistics
Affects Versions: 0.10.0
Reporter: Shreepadma Venugopalan
Assignee: Shreepadma Venugopalan
 Fix For: 0.11.0

 Attachments: HIVE-3875.1.patch.txt


 Currently, if a negative value is specified for hive.stats.ndv.error in 
 hive-site.xml, it is treated as 0. We should instead throw an exception.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


Hive-trunk-h0.21 - Build # 1906 - Fixed

2013-01-11 Thread Apache Jenkins Server
Changes for Build #1905

Changes for Build #1906
[cws] HIVE-3875. Negative value for hive.stats.ndv.error should be disallowed 
(Shreepadma Venugopalan via cws)




All tests passed

The Apache Jenkins build system has built Hive-trunk-h0.21 (build #1906)

Status: Fixed

Check console output at https://builds.apache.org/job/Hive-trunk-h0.21/1906/ to 
view the results.

[jira] [Commented] (HIVE-3810) HiveHistory.log need to replace '\r' with space before writing Entry.value to historyfile

2013-01-11 Thread Mark Grover (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-3810?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13551149#comment-13551149
 ] 

Mark Grover commented on HIVE-3810:
---

[~wsxy...@gmail.com] I see the status got resolved to fix. Did this get 
committed?

 HiveHistory.log need to replace '\r' with space before writing Entry.value to 
 historyfile
 -

 Key: HIVE-3810
 URL: https://issues.apache.org/jira/browse/HIVE-3810
 Project: Hive
  Issue Type: Bug
  Components: Logging
Reporter: qiangwang
Assignee: Mark Grover
Priority: Minor
 Attachments: HIVE-3810.1.patch, HIVE-3810.2.patch


 HiveHistory.log will replace '\n' with space before writing Entry.value to 
 history file:
 val = val.replace('\n', ' ');
 but HiveHistory.parseHiveHistory use BufferedReader.readLine which takes 
 '\n', '\r', '\r\n'  as line delimiter to parse history file
 if val contains '\r', there is a high possibility that HiveHistory.parseLine 
 will fail, in which case usually RecordTypes.valueOf(recType) will throw 
 exception 'java.lang.IllegalArgumentException'
 HiveHistory.log need to replace '\r' with space as well:
 val = val.replace('\n', ' ');
 changed to
 val = val.replaceAll(\r|\n,  );
 or
 val = val.replace('\r', ' ').replace('\n', ' ');

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-3876) call resetValid instead of ensureCapacity in the constructor of BytesRefArrayWritable

2013-01-11 Thread Yin Huai (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-3876?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yin Huai updated HIVE-3876:
---

Status: Patch Available  (was: Open)

 call resetValid instead of ensureCapacity in the constructor of 
 BytesRefArrayWritable
 -

 Key: HIVE-3876
 URL: https://issues.apache.org/jira/browse/HIVE-3876
 Project: Hive
  Issue Type: Improvement
  Components: Serializers/Deserializers
Affects Versions: 0.10.0
Reporter: Yin Huai
Assignee: Yin Huai
Priority: Minor
 Attachments: HIVE-3876.1.patch.txt


 In the constructor of BytesRefArrayWritable, ensureCapacity(capacity) is 
 called, but valid has not been adjusted accordingly. After a new 
 BytesRefArrayWritable has been created with a initial capacity of x, if 
 resetValid() has not been called explicitly, the size returned is still 0.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-2439) Upgrade antlr version to 3.4

2013-01-11 Thread Thiruvel Thirumoolan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-2439?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Thiruvel Thirumoolan updated HIVE-2439:
---

Attachment: HIVE-2439_branch10.patch
HIVE-2439_trunk_1.patch

 Upgrade antlr version to 3.4
 

 Key: HIVE-2439
 URL: https://issues.apache.org/jira/browse/HIVE-2439
 Project: Hive
  Issue Type: Improvement
Affects Versions: 0.8.0
Reporter: Ashutosh Chauhan
Assignee: Thiruvel Thirumoolan
 Attachments: HIVE-2439_branch10.patch, HIVE-2439_branch9_2.patch, 
 HIVE-2439_branch9_3.patch, HIVE-2439_branch9.patch, 
 hive-2439_incomplete.patch, HIVE-2439_trunk_1.patch, HIVE-2439_trunk.patch


 Upgrade antlr version to 3.4

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (HIVE-3889) Add floating point compression to ORC file

2013-01-11 Thread Owen O'Malley (JIRA)
Owen O'Malley created HIVE-3889:
---

 Summary: Add floating point compression to ORC file
 Key: HIVE-3889
 URL: https://issues.apache.org/jira/browse/HIVE-3889
 Project: Hive
  Issue Type: New Feature
  Components: Serializers/Deserializers
Reporter: Owen O'Malley
Assignee: Owen O'Malley


Karol Wegrzycki, a CS student at University of Warsaw, has implemented an FPC 
compressor for doubles. It would be great to hook this up to the ORC file 
format so that we can get better compression for doubles.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-3889) Add floating point compression to ORC file

2013-01-11 Thread Owen O'Malley (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-3889?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Owen O'Malley updated HIVE-3889:


Attachment: fpc-impl.tar

This is the file that Karol emailed to me for me to submit to Apache.

 Add floating point compression to ORC file
 --

 Key: HIVE-3889
 URL: https://issues.apache.org/jira/browse/HIVE-3889
 Project: Hive
  Issue Type: New Feature
  Components: Serializers/Deserializers
Reporter: Owen O'Malley
Assignee: Owen O'Malley
 Attachments: fpc-impl.tar


 Karol Wegrzycki, a CS student at University of Warsaw, has implemented an FPC 
 compressor for doubles. It would be great to hook this up to the ORC file 
 format so that we can get better compression for doubles.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-3874) Create a new Optimized Row Columnar file format for Hive

2013-01-11 Thread Owen O'Malley (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-3874?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13551231#comment-13551231
 ] 

Owen O'Malley commented on HIVE-3874:
-

Namit, I'm using the table properties to manage the other features like 
compression, so I would probably make a table property like 'orc.create.index' 
or something. Would that make sense?

I should note that the indexes are very light. In a sample file:

* uncompressed text: 370MB
* compress orc: 86MB
* row index in orc: 140k

 Create a new Optimized Row Columnar file format for Hive
 

 Key: HIVE-3874
 URL: https://issues.apache.org/jira/browse/HIVE-3874
 Project: Hive
  Issue Type: Improvement
  Components: Serializers/Deserializers
Reporter: Owen O'Malley
Assignee: Owen O'Malley
 Attachments: OrcFileIntro.pptx


 There are several limitations of the current RC File format that I'd like to 
 address by creating a new format:
 * each column value is stored as a binary blob, which means:
 ** the entire column value must be read, decompressed, and deserialized
 ** the file format can't use smarter type-specific compression
 ** push down filters can't be evaluated
 * the start of each row group needs to be found by scanning
 * user metadata can only be added to the file when the file is created
 * the file doesn't store the number of rows per a file or row group
 * there is no mechanism for seeking to a particular row number, which is 
 required for external indexes.
 * there is no mechanism for storing light weight indexes within the file to 
 enable push-down filters to skip entire row groups.
 * the type of the rows aren't stored in the file

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-3874) Create a new Optimized Row Columnar file format for Hive

2013-01-11 Thread Owen O'Malley (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-3874?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13551248#comment-13551248
 ] 

Owen O'Malley commented on HIVE-3874:
-

Doug, of course Trevni could be modified arbitrarily to match the needs of 
Hive. But Hive will benefit more if there is a deep integration between the 
file format and the query engine. Both HBase and Accumulo have file formats 
that were originally based on Hadoop's TFile. But the need for integration with 
the query engine was such that their projects were better served by having the 
file format in their project rather than an upstream project. 

Of course the Avro project is free to copy any of the ORC code into Trevni, but 
Hive has the need to innovate in this area without asking Avro to make changes 
and waiting for them to be released. 

 Create a new Optimized Row Columnar file format for Hive
 

 Key: HIVE-3874
 URL: https://issues.apache.org/jira/browse/HIVE-3874
 Project: Hive
  Issue Type: Improvement
  Components: Serializers/Deserializers
Reporter: Owen O'Malley
Assignee: Owen O'Malley
 Attachments: OrcFileIntro.pptx


 There are several limitations of the current RC File format that I'd like to 
 address by creating a new format:
 * each column value is stored as a binary blob, which means:
 ** the entire column value must be read, decompressed, and deserialized
 ** the file format can't use smarter type-specific compression
 ** push down filters can't be evaluated
 * the start of each row group needs to be found by scanning
 * user metadata can only be added to the file when the file is created
 * the file doesn't store the number of rows per a file or row group
 * there is no mechanism for seeking to a particular row number, which is 
 required for external indexes.
 * there is no mechanism for storing light weight indexes within the file to 
 enable push-down filters to skip entire row groups.
 * the type of the rows aren't stored in the file

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-3874) Create a new Optimized Row Columnar file format for Hive

2013-01-11 Thread Sambavi Muthukrishnan (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-3874?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13551275#comment-13551275
 ] 

Sambavi Muthukrishnan commented on HIVE-3874:
-

Owen: is the row group expected to be about 250 MB (per stripe size)? Does your 
implementation attempt to make every row group align with HDFS block size so a 
split = 1 block?

Also: do you have an ETA for an initial patch? We would really like to try this 
out - we have some additional ideas that we would like to try out on top of 
this.

 Create a new Optimized Row Columnar file format for Hive
 

 Key: HIVE-3874
 URL: https://issues.apache.org/jira/browse/HIVE-3874
 Project: Hive
  Issue Type: Improvement
  Components: Serializers/Deserializers
Reporter: Owen O'Malley
Assignee: Owen O'Malley
 Attachments: OrcFileIntro.pptx


 There are several limitations of the current RC File format that I'd like to 
 address by creating a new format:
 * each column value is stored as a binary blob, which means:
 ** the entire column value must be read, decompressed, and deserialized
 ** the file format can't use smarter type-specific compression
 ** push down filters can't be evaluated
 * the start of each row group needs to be found by scanning
 * user metadata can only be added to the file when the file is created
 * the file doesn't store the number of rows per a file or row group
 * there is no mechanism for seeking to a particular row number, which is 
 required for external indexes.
 * there is no mechanism for storing light weight indexes within the file to 
 enable push-down filters to skip entire row groups.
 * the type of the rows aren't stored in the file

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-3874) Create a new Optimized Row Columnar file format for Hive

2013-01-11 Thread Owen O'Malley (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-3874?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13551286#comment-13551286
 ] 

Owen O'Malley commented on HIVE-3874:
-

Sambavi, I should have a patch ready next week. Yes, the row groups (stripes) 
are 250MB by default. I currently set the HDFS block size for the files to 2 
times the stripe size, but I don't try to align them other than that.

 Create a new Optimized Row Columnar file format for Hive
 

 Key: HIVE-3874
 URL: https://issues.apache.org/jira/browse/HIVE-3874
 Project: Hive
  Issue Type: Improvement
  Components: Serializers/Deserializers
Reporter: Owen O'Malley
Assignee: Owen O'Malley
 Attachments: OrcFileIntro.pptx


 There are several limitations of the current RC File format that I'd like to 
 address by creating a new format:
 * each column value is stored as a binary blob, which means:
 ** the entire column value must be read, decompressed, and deserialized
 ** the file format can't use smarter type-specific compression
 ** push down filters can't be evaluated
 * the start of each row group needs to be found by scanning
 * user metadata can only be added to the file when the file is created
 * the file doesn't store the number of rows per a file or row group
 * there is no mechanism for seeking to a particular row number, which is 
 required for external indexes.
 * there is no mechanism for storing light weight indexes within the file to 
 enable push-down filters to skip entire row groups.
 * the type of the rows aren't stored in the file

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-3803) explain dependency should show the dependencies hierarchically in presence of views

2013-01-11 Thread Namit Jain (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-3803?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Namit Jain updated HIVE-3803:
-

Attachment: hive.3803.12.patch

 explain dependency should show the dependencies hierarchically in presence of 
 views
 ---

 Key: HIVE-3803
 URL: https://issues.apache.org/jira/browse/HIVE-3803
 Project: Hive
  Issue Type: Bug
  Components: Query Processor
Reporter: Namit Jain
Assignee: Namit Jain
 Attachments: hive.3803.10.patch, hive.3803.11.patch, 
 hive.3803.12.patch, hive.3803.1.patch, hive.3803.2.patch, hive.3803.3.patch, 
 hive.3803.4.patch, hive.3803.5.patch, hive.3803.6.patch, hive.3803.7.patch, 
 hive.3803.8.patch, hive.3803.9.patch


 It should also include tables whose partitions are being accessed

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-3874) Create a new Optimized Row Columnar file format for Hive

2013-01-11 Thread Namit Jain (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-3874?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13551347#comment-13551347
 ] 

Namit Jain commented on HIVE-3874:
--

Owen, that would be good 'a table property like 'orc.create.index' or something'

 Create a new Optimized Row Columnar file format for Hive
 

 Key: HIVE-3874
 URL: https://issues.apache.org/jira/browse/HIVE-3874
 Project: Hive
  Issue Type: Improvement
  Components: Serializers/Deserializers
Reporter: Owen O'Malley
Assignee: Owen O'Malley
 Attachments: OrcFileIntro.pptx


 There are several limitations of the current RC File format that I'd like to 
 address by creating a new format:
 * each column value is stored as a binary blob, which means:
 ** the entire column value must be read, decompressed, and deserialized
 ** the file format can't use smarter type-specific compression
 ** push down filters can't be evaluated
 * the start of each row group needs to be found by scanning
 * user metadata can only be added to the file when the file is created
 * the file doesn't store the number of rows per a file or row group
 * there is no mechanism for seeking to a particular row number, which is 
 required for external indexes.
 * there is no mechanism for storing light weight indexes within the file to 
 enable push-down filters to skip entire row groups.
 * the type of the rows aren't stored in the file

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-3888) wrong mapside groupby if no partition is being selected

2013-01-11 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-3888?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13551368#comment-13551368
 ] 

Hudson commented on HIVE-3888:
--

Integrated in Hive-trunk-hadoop2 #59 (See 
[https://builds.apache.org/job/Hive-trunk-hadoop2/59/])
HIVE-3888 wrong mapside groupby if no partition is being selected
(Namit Jain via Ashutosh and namit) (Revision 1431936)

 Result = FAILURE
namit : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1431936
Files : 
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/optimizer/GroupByOptimizer.java
* /hive/trunk/ql/src/test/queries/clientpositive/groupby_sort_6.q
* /hive/trunk/ql/src/test/queries/clientpositive/groupby_sort_7.q
* /hive/trunk/ql/src/test/results/clientpositive/groupby_sort_6.q.out
* /hive/trunk/ql/src/test/results/clientpositive/groupby_sort_7.q.out
* /hive/trunk/ql/src/test/results/clientpositive/metadataonly1.q.out


 wrong mapside groupby if no partition is being selected
 ---

 Key: HIVE-3888
 URL: https://issues.apache.org/jira/browse/HIVE-3888
 Project: Hive
  Issue Type: Bug
  Components: Query Processor
Reporter: Namit Jain
Assignee: Namit Jain
 Fix For: 0.11.0

 Attachments: hive.3888.1.patch




--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


Hive-trunk-hadoop2 - Build # 59 - Failure

2013-01-11 Thread Apache Jenkins Server


38 tests failed.
FAILED:  
org.apache.hadoop.hive.cli.TestHBaseCliDriver.testCliDriver_external_table_ppd

Error Message:
Unexpected exception in setup

Stack Trace:
junit.framework.AssertionFailedError: Unexpected exception in setup
at junit.framework.Assert.fail(Assert.java:50)
at 
org.apache.hadoop.hive.cli.TestHBaseCliDriver.setUp(TestHBaseCliDriver.java:59)
at junit.framework.TestCase.runBare(TestCase.java:132)
at junit.framework.TestResult$1.protect(TestResult.java:110)
at junit.framework.TestResult.runProtected(TestResult.java:128)
at junit.framework.TestResult.run(TestResult.java:113)
at junit.framework.TestCase.run(TestCase.java:124)
at junit.framework.TestSuite.runTest(TestSuite.java:243)
at junit.framework.TestSuite.run(TestSuite.java:238)
at junit.extensions.TestDecorator.basicRun(TestDecorator.java:24)
at junit.extensions.TestSetup$1.protect(TestSetup.java:23)
at junit.framework.TestResult.runProtected(TestResult.java:128)
at junit.extensions.TestSetup.run(TestSetup.java:27)
at 
org.apache.tools.ant.taskdefs.optional.junit.JUnitTestRunner.run(JUnitTestRunner.java:422)
at 
org.apache.tools.ant.taskdefs.optional.junit.JUnitTestRunner.launch(JUnitTestRunner.java:931)
at 
org.apache.tools.ant.taskdefs.optional.junit.JUnitTestRunner.main(JUnitTestRunner.java:785)


FAILED:  
org.apache.hadoop.hive.cli.TestHBaseCliDriver.testCliDriver_hbase_binary_external_table_queries

Error Message:
Unexpected exception in setup

Stack Trace:
junit.framework.AssertionFailedError: Unexpected exception in setup
at junit.framework.Assert.fail(Assert.java:50)
at 
org.apache.hadoop.hive.cli.TestHBaseCliDriver.setUp(TestHBaseCliDriver.java:59)
at junit.framework.TestCase.runBare(TestCase.java:132)
at junit.framework.TestResult$1.protect(TestResult.java:110)
at junit.framework.TestResult.runProtected(TestResult.java:128)
at junit.framework.TestResult.run(TestResult.java:113)
at junit.framework.TestCase.run(TestCase.java:124)
at junit.framework.TestSuite.runTest(TestSuite.java:243)
at junit.framework.TestSuite.run(TestSuite.java:238)
at junit.extensions.TestDecorator.basicRun(TestDecorator.java:24)
at junit.extensions.TestSetup$1.protect(TestSetup.java:23)
at junit.framework.TestResult.runProtected(TestResult.java:128)
at junit.extensions.TestSetup.run(TestSetup.java:27)
at 
org.apache.tools.ant.taskdefs.optional.junit.JUnitTestRunner.run(JUnitTestRunner.java:422)
at 
org.apache.tools.ant.taskdefs.optional.junit.JUnitTestRunner.launch(JUnitTestRunner.java:931)
at 
org.apache.tools.ant.taskdefs.optional.junit.JUnitTestRunner.main(JUnitTestRunner.java:785)


FAILED:  
org.apache.hadoop.hive.cli.TestHBaseCliDriver.testCliDriver_hbase_binary_map_queries

Error Message:
Unexpected exception in setup

Stack Trace:
junit.framework.AssertionFailedError: Unexpected exception in setup
at junit.framework.Assert.fail(Assert.java:50)
at 
org.apache.hadoop.hive.cli.TestHBaseCliDriver.setUp(TestHBaseCliDriver.java:59)
at junit.framework.TestCase.runBare(TestCase.java:132)
at junit.framework.TestResult$1.protect(TestResult.java:110)
at junit.framework.TestResult.runProtected(TestResult.java:128)
at junit.framework.TestResult.run(TestResult.java:113)
at junit.framework.TestCase.run(TestCase.java:124)
at junit.framework.TestSuite.runTest(TestSuite.java:243)
at junit.framework.TestSuite.run(TestSuite.java:238)
at junit.extensions.TestDecorator.basicRun(TestDecorator.java:24)
at junit.extensions.TestSetup$1.protect(TestSetup.java:23)
at junit.framework.TestResult.runProtected(TestResult.java:128)
at junit.extensions.TestSetup.run(TestSetup.java:27)
at 
org.apache.tools.ant.taskdefs.optional.junit.JUnitTestRunner.run(JUnitTestRunner.java:422)
at 
org.apache.tools.ant.taskdefs.optional.junit.JUnitTestRunner.launch(JUnitTestRunner.java:931)
at 
org.apache.tools.ant.taskdefs.optional.junit.JUnitTestRunner.main(JUnitTestRunner.java:785)


FAILED:  
org.apache.hadoop.hive.cli.TestHBaseCliDriver.testCliDriver_hbase_binary_storage_queries

Error Message:
Unexpected exception in setup

Stack Trace:
junit.framework.AssertionFailedError: Unexpected exception in setup
at junit.framework.Assert.fail(Assert.java:50)
at 
org.apache.hadoop.hive.cli.TestHBaseCliDriver.setUp(TestHBaseCliDriver.java:59)
at junit.framework.TestCase.runBare(TestCase.java:132)
at junit.framework.TestResult$1.protect(TestResult.java:110)
at junit.framework.TestResult.runProtected(TestResult.java:128)
at junit.framework.TestResult.run(TestResult.java:113)
at junit.framework.TestCase.run(TestCase.java:124)
at 

[jira] [Commented] (HIVE-2439) Upgrade antlr version to 3.4

2013-01-11 Thread Ashutosh Chauhan (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-2439?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13551439#comment-13551439
 ] 

Ashutosh Chauhan commented on HIVE-2439:


Thanks Thriuvel for addressing comments. Latest patch looks good. +1. Running 
tests now.

 Upgrade antlr version to 3.4
 

 Key: HIVE-2439
 URL: https://issues.apache.org/jira/browse/HIVE-2439
 Project: Hive
  Issue Type: Improvement
Affects Versions: 0.8.0
Reporter: Ashutosh Chauhan
Assignee: Thiruvel Thirumoolan
 Attachments: HIVE-2439_branch10.patch, HIVE-2439_branch9_2.patch, 
 HIVE-2439_branch9_3.patch, HIVE-2439_branch9.patch, 
 hive-2439_incomplete.patch, HIVE-2439_trunk_1.patch, HIVE-2439_trunk.patch


 Upgrade antlr version to 3.4

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Reopened] (HIVE-3810) HiveHistory.log need to replace '\r' with space before writing Entry.value to historyfile

2013-01-11 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-3810?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan reopened HIVE-3810:



I dont see this in svn commit log. Reopening for review.

 HiveHistory.log need to replace '\r' with space before writing Entry.value to 
 historyfile
 -

 Key: HIVE-3810
 URL: https://issues.apache.org/jira/browse/HIVE-3810
 Project: Hive
  Issue Type: Bug
  Components: Logging
Reporter: qiangwang
Assignee: Mark Grover
Priority: Minor
 Attachments: HIVE-3810.1.patch, HIVE-3810.2.patch


 HiveHistory.log will replace '\n' with space before writing Entry.value to 
 history file:
 val = val.replace('\n', ' ');
 but HiveHistory.parseHiveHistory use BufferedReader.readLine which takes 
 '\n', '\r', '\r\n'  as line delimiter to parse history file
 if val contains '\r', there is a high possibility that HiveHistory.parseLine 
 will fail, in which case usually RecordTypes.valueOf(recType) will throw 
 exception 'java.lang.IllegalArgumentException'
 HiveHistory.log need to replace '\r' with space as well:
 val = val.replace('\n', ' ');
 changed to
 val = val.replaceAll(\r|\n,  );
 or
 val = val.replace('\r', ' ').replace('\n', ' ');

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Reopened] (HIVE-3857) HiveHistoryViewer concurrency problem

2013-01-11 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-3857?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan reopened HIVE-3857:


  Assignee: qiangwang

This is not committed yet. Reopening for review.

 HiveHistoryViewer concurrency problem
 -

 Key: HIVE-3857
 URL: https://issues.apache.org/jira/browse/HIVE-3857
 Project: Hive
  Issue Type: Bug
  Components: Logging
Reporter: qiangwang
Assignee: qiangwang
Priority: Minor
 Attachments: HiveHistory.patch


 new HiveHistoryViewer() throws ConcurrentModificationException when called 
 concurrently by several threads.
 According to the stacktrace, HiveHistory.parseLine use private static 
 MapString, String parseBuffer to store parsed data and this finally caused 
 the exception.
 I don't know why a static buffer rather than a local buffer is used!

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-3824) bug if different serdes are used for different partitions

2013-01-11 Thread Ashutosh Chauhan (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-3824?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13551479#comment-13551479
 ] 

Ashutosh Chauhan commented on HIVE-3824:


Namit, Feel free to commit this if tests pass.

 bug if different serdes are used for different partitions
 -

 Key: HIVE-3824
 URL: https://issues.apache.org/jira/browse/HIVE-3824
 Project: Hive
  Issue Type: Bug
  Components: Query Processor
Reporter: Namit Jain
Assignee: Namit Jain
 Attachments: hive.3824.1.patch, hive.3824.3.patch


 Consider the following testcase:
 create table tst5 (key string, value string) partitioned by (ds string) 
 stored as rcfile;
 insert overwrite table tst5 partition (ds='1') select * from src;
 insert overwrite table tst5 partition (ds='2') select * from src;
 insert overwrite table tst5 partition (ds='3') select * from src;
 alter table tst5 stored as sequencefile; 
 insert overwrite table tst5 partition (ds='4') select * from src;
 insert overwrite table tst5 partition (ds='5') select * from src;
 insert overwrite table tst5 partition (ds='6') select * from src;  
 alter table tst5 set serde 
 'org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'; 
 insert overwrite table tst5 partition (ds='7') select * from src;
 insert overwrite table tst5 partition (ds='8') select * from src;
 insert overwrite table tst5 partition (ds='9') select * from src;  
 The following query works fine:
  select key + key, value from tst5 where ((ds = '4') or (ds = '1'));   
 since both the partitions use ColumnarSerDe
 But the following query fails:
 select key + key, value from tst5 where ((ds = '4') or (ds = '1') or 
 (ds='7'));
 since different serdes are used.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-3888) wrong mapside groupby if no partition is being selected

2013-01-11 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-3888?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13551498#comment-13551498
 ] 

Hudson commented on HIVE-3888:
--

Integrated in Hive-trunk-h0.21 #1907 (See 
[https://builds.apache.org/job/Hive-trunk-h0.21/1907/])
HIVE-3888 wrong mapside groupby if no partition is being selected
(Namit Jain via Ashutosh and namit) (Revision 1431936)

 Result = ABORTED
namit : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1431936
Files : 
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/optimizer/GroupByOptimizer.java
* /hive/trunk/ql/src/test/queries/clientpositive/groupby_sort_6.q
* /hive/trunk/ql/src/test/queries/clientpositive/groupby_sort_7.q
* /hive/trunk/ql/src/test/results/clientpositive/groupby_sort_6.q.out
* /hive/trunk/ql/src/test/results/clientpositive/groupby_sort_7.q.out
* /hive/trunk/ql/src/test/results/clientpositive/metadataonly1.q.out


 wrong mapside groupby if no partition is being selected
 ---

 Key: HIVE-3888
 URL: https://issues.apache.org/jira/browse/HIVE-3888
 Project: Hive
  Issue Type: Bug
  Components: Query Processor
Reporter: Namit Jain
Assignee: Namit Jain
 Fix For: 0.11.0

 Attachments: hive.3888.1.patch




--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-3567) Hive List Bucketing - Query logic supports , =, , = and in

2013-01-11 Thread Gang Tim Liu (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-3567?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gang Tim Liu updated HIVE-3567:
---

Description: Query logic supports equal. It should handle , =,  , =  
and in also.  (was: Query logic supports equal. It should handle , =,  and 
= also.)
Summary: Hive List Bucketing - Query logic supports , =,  , = and 
in  (was: Hive List Bucketing - Query logic supports , =,  and =)

 Hive List Bucketing - Query logic supports , =,  , = and in
 ---

 Key: HIVE-3567
 URL: https://issues.apache.org/jira/browse/HIVE-3567
 Project: Hive
  Issue Type: New Feature
Affects Versions: 0.10.0
Reporter: Gang Tim Liu
Assignee: Gang Tim Liu
Priority: Minor

 Query logic supports equal. It should handle , =,  , =  and in also.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-3426) union with same source should be optimized

2013-01-11 Thread Ashutosh Chauhan (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-3426?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13551621#comment-13551621
 ] 

Ashutosh Chauhan commented on HIVE-3426:


What Zhenxiao is proposing a good first step. Lets try to optimize that query 
first and worry about GBY in subquery later.

 union with same source should be optimized
 --

 Key: HIVE-3426
 URL: https://issues.apache.org/jira/browse/HIVE-3426
 Project: Hive
  Issue Type: Improvement
  Components: Query Processor
Reporter: Namit Jain
Assignee: Zhenxiao Luo



--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-942) use bucketing for group by

2013-01-11 Thread Ashutosh Chauhan (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-942?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13551625#comment-13551625
 ] 

Ashutosh Chauhan commented on HIVE-942:
---

I think if BucketizedHiveInputFormat is used with map-side aggregation on, this 
should already work, no? Probably reducer phase will also get created, which is 
not required and can be optimized away. Is that the optimization we are talking 
about? 

 use bucketing for group by
 --

 Key: HIVE-942
 URL: https://issues.apache.org/jira/browse/HIVE-942
 Project: Hive
  Issue Type: New Feature
  Components: Query Processor
Reporter: Namit Jain

 Group by on a bucketed column can be completely performed on the mapper if 
 the split can be adjusted to span the key boundary.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


Re: [ANNOUNCE] Apache Hive 0.10.0 Released

2013-01-11 Thread Carl Steinbach
Great work everyone!

On Fri, Jan 11, 2013 at 12:37 AM, Lefty Leverenz le...@hortonworks.comwrote:

 Congratulations!

 – Lefty



 On Thu, Jan 10, 2013 at 11:59 PM, Ashutosh Chauhan hashut...@apache.org
 wrote:

  The Apache Hive team is proud to announce the the release of Apache
  Hive version 0.10.0.
 
  The Apache Hive (TM) data warehouse software facilitates querying and
  managing large datasets residing in distributed storage. Built on top
  of Apache Hadoop (TM), it provides:
 
  * Tools to enable easy data extract/transform/load (ETL)
 
  * A mechanism to impose structure on a variety of data formats
 
  * Access to files stored either directly in Apache HDFS (TM) or in other
data storage systems such as Apache HBase (TM)
 
  * Query execution via MapReduce
 
  For Hive release details and downloads, please
  visit:http://hive.apache.org/releases.html
 
  Hive 0.10.0 Release Notes are available
  here:
 
 https://issues.apache.org/jira/secure/ReleaseNote.jspa?version=12320745styleName=TextprojectId=12310843
 
  More than 350 jira issues are fixed in this release. We would like to
  thank many contributors who made this release possible.
 
  Regards,
 
  The Apache Hive Team
 



[jira] [Assigned] (HIVE-1649) Ability to update counters and status from TRANSFORM scripts

2013-01-11 Thread Carl Steinbach (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-1649?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Carl Steinbach reassigned HIVE-1649:


Assignee: Guo Hongjie

@Guo: Can you post a review request on reviewboard or phabricator? Directions 
explaining how to do this are on the wiki. Thanks.

 Ability to update counters and status from TRANSFORM scripts
 

 Key: HIVE-1649
 URL: https://issues.apache.org/jira/browse/HIVE-1649
 Project: Hive
  Issue Type: New Feature
  Components: Query Processor
Affects Versions: 0.6.0
Reporter: Carl Steinbach
Assignee: Guo Hongjie
 Attachments: HIVE-1649.1.patch


 Hadoop Streaming supports the ability to update counters and status by 
 writing specially coded messages to the script's stderr stream.
 A streaming process can use the stderr to emit counter information. 
 {{reporter:counter:group,counter,amount}} should be sent to stderr to 
 update the counter.
 A streaming process can use the stderr to emit status information. To set a 
 status, {{reporter:status:message}} should be sent to stderr.
 Hive should support these same features with its TRANSFORM mechanism.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-1649) Ability to update counters and status from TRANSFORM scripts

2013-01-11 Thread Carl Steinbach (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-1649?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Carl Steinbach updated HIVE-1649:
-

Resolution: Duplicate
Status: Resolved  (was: Patch Available)

 Ability to update counters and status from TRANSFORM scripts
 

 Key: HIVE-1649
 URL: https://issues.apache.org/jira/browse/HIVE-1649
 Project: Hive
  Issue Type: New Feature
  Components: Query Processor
Affects Versions: 0.6.0
Reporter: Carl Steinbach
Assignee: Guo Hongjie
 Attachments: HIVE-1649.1.patch


 Hadoop Streaming supports the ability to update counters and status by 
 writing specially coded messages to the script's stderr stream.
 A streaming process can use the stderr to emit counter information. 
 {{reporter:counter:group,counter,amount}} should be sent to stderr to 
 update the counter.
 A streaming process can use the stderr to emit status information. To set a 
 status, {{reporter:status:message}} should be sent to stderr.
 Hive should support these same features with its TRANSFORM mechanism.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Assigned] (HIVE-305) Port Hadoop streaming's counters/status reporters to Hive Transforms

2013-01-11 Thread Carl Steinbach (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-305?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Carl Steinbach reassigned HIVE-305:
---

Assignee: Guo Hongjie

@Guo: please submit a review request on reviewboard or phabricator. Thanks.

 Port Hadoop streaming's counters/status reporters to Hive Transforms
 

 Key: HIVE-305
 URL: https://issues.apache.org/jira/browse/HIVE-305
 Project: Hive
  Issue Type: New Feature
  Components: Query Processor
Reporter: Venky Iyer
Assignee: Guo Hongjie
 Attachments: HIVE-305.1.patch


 https://issues.apache.org/jira/browse/HADOOP-1328
  Introduced a way for a streaming process to update global counters and 
 status using stderr stream to emit information. Use 
 reporter:counter:group,counter,amount  to update  a counter. Use 
 reporter:status:message to update status. 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-3004) RegexSerDe should support other column types in addition to STRING

2013-01-11 Thread Shreepadma Venugopalan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-3004?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Shreepadma Venugopalan updated HIVE-3004:
-

Status: Patch Available  (was: Open)

 RegexSerDe should support other column types in addition to STRING
 --

 Key: HIVE-3004
 URL: https://issues.apache.org/jira/browse/HIVE-3004
 Project: Hive
  Issue Type: Improvement
  Components: Serializers/Deserializers
Reporter: Carl Steinbach
Assignee: Shreepadma Venugopalan
 Attachments: HIVE-3004-1.patch, HIVE-3004.2.patch




--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-3004) RegexSerDe should support other column types in addition to STRING

2013-01-11 Thread Shreepadma Venugopalan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-3004?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Shreepadma Venugopalan updated HIVE-3004:
-

Attachment: HIVE-3004.3.patch.txt

 RegexSerDe should support other column types in addition to STRING
 --

 Key: HIVE-3004
 URL: https://issues.apache.org/jira/browse/HIVE-3004
 Project: Hive
  Issue Type: Improvement
  Components: Serializers/Deserializers
Reporter: Carl Steinbach
Assignee: Shreepadma Venugopalan
 Attachments: HIVE-3004-1.patch, HIVE-3004.2.patch, 
 HIVE-3004.3.patch.txt




--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-3004) RegexSerDe should support other column types in addition to STRING

2013-01-11 Thread Shreepadma Venugopalan (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-3004?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13551697#comment-13551697
 ] 

Shreepadma Venugopalan commented on HIVE-3004:
--

Thanks Ashutosh. I've attached the new patch to the JIRA. 

 RegexSerDe should support other column types in addition to STRING
 --

 Key: HIVE-3004
 URL: https://issues.apache.org/jira/browse/HIVE-3004
 Project: Hive
  Issue Type: Improvement
  Components: Serializers/Deserializers
Reporter: Carl Steinbach
Assignee: Shreepadma Venugopalan
 Attachments: HIVE-3004-1.patch, HIVE-3004.2.patch, 
 HIVE-3004.3.patch.txt




--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


Review Request: HIVE-3004: RegexSerDe should support other column types in addition to STRING

2013-01-11 Thread Shreepadma Venugopalan

---
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/8931/
---

Review request for hive.


Description
---

This patch enhances regex serde to parse column types other than STRING. Only 
primitive types are supported.


Diffs
-

  ql/src/test/queries/clientnegative/serde_regex.q 6603b91 
  ql/src/test/queries/clientpositive/serde_regex.q c6809cb 
  ql/src/test/results/clientnegative/serde_regex.q.out 03fe907 
  ql/src/test/results/clientpositive/serde_regex.q.out a8ce604 
  serde/src/java/org/apache/hadoop/hive/serde2/RegexSerDe.java e728244 

Diff: https://reviews.apache.org/r/8931/diff/


Testing
---

New test cases have been added and they pass.


Thanks,

Shreepadma Venugopalan



Re: Review Request: HIVE-3004: RegexSerDe should support other column types in addition to STRING

2013-01-11 Thread Shreepadma Venugopalan

---
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/8931/
---

(Updated Jan. 12, 2013, 12:28 a.m.)


Review request for hive and Ashutosh Chauhan.


Description
---

This patch enhances regex serde to parse column types other than STRING. Only 
primitive types are supported.


This addresses bug HIVE-3004.
https://issues.apache.org/jira/browse/HIVE-3004


Diffs
-

  ql/src/test/queries/clientnegative/serde_regex.q 6603b91 
  ql/src/test/queries/clientpositive/serde_regex.q c6809cb 
  ql/src/test/results/clientnegative/serde_regex.q.out 03fe907 
  ql/src/test/results/clientpositive/serde_regex.q.out a8ce604 
  serde/src/java/org/apache/hadoop/hive/serde2/RegexSerDe.java e728244 

Diff: https://reviews.apache.org/r/8931/diff/


Testing
---

New test cases have been added and they pass.


Thanks,

Shreepadma Venugopalan



[jira] [Commented] (HIVE-3004) RegexSerDe should support other column types in addition to STRING

2013-01-11 Thread Shreepadma Venugopalan (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-3004?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13551698#comment-13551698
 ] 

Shreepadma Venugopalan commented on HIVE-3004:
--

Review board : https://reviews.apache.org/r/8931/

 RegexSerDe should support other column types in addition to STRING
 --

 Key: HIVE-3004
 URL: https://issues.apache.org/jira/browse/HIVE-3004
 Project: Hive
  Issue Type: Improvement
  Components: Serializers/Deserializers
Reporter: Carl Steinbach
Assignee: Shreepadma Venugopalan
 Attachments: HIVE-3004-1.patch, HIVE-3004.2.patch, 
 HIVE-3004.3.patch.txt




--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-3426) union with same source should be optimized

2013-01-11 Thread Shreepadma Venugopalan (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-3426?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13551701#comment-13551701
 ] 

Shreepadma Venugopalan commented on HIVE-3426:
--

Yup, let's try to optimize the simple case first. Optimizing subqueries with 
GBY can be the next step. 

 union with same source should be optimized
 --

 Key: HIVE-3426
 URL: https://issues.apache.org/jira/browse/HIVE-3426
 Project: Hive
  Issue Type: Improvement
  Components: Query Processor
Reporter: Namit Jain
Assignee: Zhenxiao Luo



--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


Re: [DISCUSS] HCatalog becoming a subproject of Hive

2013-01-11 Thread Carl Steinbach
Hi Alan,

I agree that submitting this for a vote is the best option.


 If anyone has additional proposed modifications please make them.
  Otherwise I propose that the Hive PMC vote on this proposal.


In order for the Hive PMC to be able to vote on these changes they need to
be expressed in terms of one or more of the actions listed at the end of
the Hive project bylaws:

https://cwiki.apache.org/confluence/display/Hive/Bylaws

So I think we first need to amend to the bylaws in order to define the
rights and privileges of a submodule committer, and then separately vote
the HCatalog committers in as Hive submodule committers. Does this make
sense?

Thanks.

Carl


[jira] [Assigned] (HIVE-3653) Failure in a counter poller run should not be considered as a job failure

2013-01-11 Thread Shreepadma Venugopalan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-3653?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Shreepadma Venugopalan reassigned HIVE-3653:


Assignee: Shreepadma Venugopalan

 Failure in a counter poller run should not be considered as a job failure
 -

 Key: HIVE-3653
 URL: https://issues.apache.org/jira/browse/HIVE-3653
 Project: Hive
  Issue Type: Bug
  Components: Clients
Affects Versions: 0.7.1
Reporter: Harsh J
Assignee: Shreepadma Venugopalan

 A client had a simple transient failure in polling the JT for job status 
 (which it does for HIVECOUNTERSPULLINTERVAL for each currently running job).
 {code}
 java.io.IOException: Call to HOST/IP:PORT failed on local exception: 
 java.io.IOException: Connection reset by peer 
 at org.apache.hadoop.ipc.Client.wrapException(Client.java:1142) 
 at org.apache.hadoop.ipc.Client.call(Client.java:1110) 
 at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:226) 
 at org.apache.hadoop.mapred.$Proxy10.getJobStatus(Unknown Source) 
 at org.apache.hadoop.mapred.JobClient.getJob(JobClient.java:1053) 
 at org.apache.hadoop.mapred.JobClient.getJob(JobClient.java:1065) 
 at org.apache.hadoop.hive.ql.exec.ExecDriver.progress(ExecDriver.java:351) 
 at org.apache.hadoop.hive.ql.exec.ExecDriver.execute(ExecDriver.java:686) 
 at org.apache.hadoop.hive.ql.exec.MapRedTask.execute(MapRedTask.java:123) 
 at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:131) 
 at 
 org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:57) 
 at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:1063) 
 at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:900) 
 at org.apache.hadoop.hive.ql.Driver.run(Driver.java:748) 
 at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:209) 
 at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:286) 
 at org.apache.hadoop.hive.cli.CliDriver.processReader(CliDriver.java:310) 
 at org.apache.hadoop.hive.cli.CliDriver.processFile(CliDriver.java:317) 
 at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:490) 
 at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) 
 at 
 sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39) 
 at 
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
  
 at java.lang.reflect.Method.invoke(Method.java:597) 
 at org.apache.hadoop.util.RunJar.main(RunJar.java:197) 
 {code}
 This lead to Hive thinking the running job itself has failed, and it failed 
 the query run, although the running job progressed to completion in the 
 background.
 We should not let transient IOExceptions in counter polling cause query 
 termination, and should instead just retry.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-3803) explain dependency should show the dependencies hierarchically in presence of views

2013-01-11 Thread Sambavi Muthukrishnan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-3803?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sambavi Muthukrishnan updated HIVE-3803:


Attachment: hive.3803.13.patch

 explain dependency should show the dependencies hierarchically in presence of 
 views
 ---

 Key: HIVE-3803
 URL: https://issues.apache.org/jira/browse/HIVE-3803
 Project: Hive
  Issue Type: Bug
  Components: Query Processor
Reporter: Namit Jain
Assignee: Namit Jain
 Attachments: hive.3803.10.patch, hive.3803.11.patch, 
 hive.3803.12.patch, hive.3803.13.patch, hive.3803.1.patch, hive.3803.2.patch, 
 hive.3803.3.patch, hive.3803.4.patch, hive.3803.5.patch, hive.3803.6.patch, 
 hive.3803.7.patch, hive.3803.8.patch, hive.3803.9.patch


 It should also include tables whose partitions are being accessed

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-3803) explain dependency should show the dependencies hierarchically in presence of views

2013-01-11 Thread Sambavi Muthukrishnan (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-3803?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13551747#comment-13551747
 ] 

Sambavi Muthukrishnan commented on HIVE-3803:
-

Uploaded new patch with the updates to some test files.

 explain dependency should show the dependencies hierarchically in presence of 
 views
 ---

 Key: HIVE-3803
 URL: https://issues.apache.org/jira/browse/HIVE-3803
 Project: Hive
  Issue Type: Bug
  Components: Query Processor
Reporter: Namit Jain
Assignee: Namit Jain
 Attachments: hive.3803.10.patch, hive.3803.11.patch, 
 hive.3803.12.patch, hive.3803.13.patch, hive.3803.1.patch, hive.3803.2.patch, 
 hive.3803.3.patch, hive.3803.4.patch, hive.3803.5.patch, hive.3803.6.patch, 
 hive.3803.7.patch, hive.3803.8.patch, hive.3803.9.patch


 It should also include tables whose partitions are being accessed

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-305) Port Hadoop streaming's counters/status reporters to Hive Transforms

2013-01-11 Thread Guo Hongjie (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-305?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Guo Hongjie updated HIVE-305:
-

Attachment: HIVE-305.2.patch

 Port Hadoop streaming's counters/status reporters to Hive Transforms
 

 Key: HIVE-305
 URL: https://issues.apache.org/jira/browse/HIVE-305
 Project: Hive
  Issue Type: New Feature
  Components: Query Processor
Reporter: Venky Iyer
Assignee: Guo Hongjie
 Attachments: HIVE-305.1.patch, HIVE-305.2.patch


 https://issues.apache.org/jira/browse/HADOOP-1328
  Introduced a way for a streaming process to update global counters and 
 status using stderr stream to emit information. Use 
 reporter:counter:group,counter,amount  to update  a counter. Use 
 reporter:status:message to update status. 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-3803) explain dependency should show the dependencies hierarchically in presence of views

2013-01-11 Thread Kevin Wilfong (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-3803?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13551783#comment-13551783
 ] 

Kevin Wilfong commented on HIVE-3803:
-

+1

 explain dependency should show the dependencies hierarchically in presence of 
 views
 ---

 Key: HIVE-3803
 URL: https://issues.apache.org/jira/browse/HIVE-3803
 Project: Hive
  Issue Type: Bug
  Components: Query Processor
Reporter: Namit Jain
Assignee: Namit Jain
 Attachments: hive.3803.10.patch, hive.3803.11.patch, 
 hive.3803.12.patch, hive.3803.13.patch, hive.3803.1.patch, hive.3803.2.patch, 
 hive.3803.3.patch, hive.3803.4.patch, hive.3803.5.patch, hive.3803.6.patch, 
 hive.3803.7.patch, hive.3803.8.patch, hive.3803.9.patch


 It should also include tables whose partitions are being accessed

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-305) Port Hadoop streaming's counters/status reporters to Hive Transforms

2013-01-11 Thread Guo Hongjie (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-305?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13551787#comment-13551787
 ] 

Guo Hongjie commented on HIVE-305:
--

thanks @Carl,do you mean submit patch button? I cannot find any button about 
review request. 

 Port Hadoop streaming's counters/status reporters to Hive Transforms
 

 Key: HIVE-305
 URL: https://issues.apache.org/jira/browse/HIVE-305
 Project: Hive
  Issue Type: New Feature
  Components: Query Processor
Reporter: Venky Iyer
Assignee: Guo Hongjie
 Attachments: HIVE-305.1.patch, HIVE-305.2.patch


 https://issues.apache.org/jira/browse/HADOOP-1328
  Introduced a way for a streaming process to update global counters and 
 status using stderr stream to emit information. Use 
 reporter:counter:group,counter,amount  to update  a counter. Use 
 reporter:status:message to update status. 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


Review Request: HIVE-305: Port Hadoop streaming's counters/status reporters to Hive Transforms

2013-01-11 Thread Guo Hongjie

---
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/8933/
---

Review request for hive.


Description
---


Port Hadoop streaming's counters/status reporters to Hive Transforms.
https://issues.apache.org/jira/browse/HADOOP-1328
 Introduced a way for a streaming process to update global counters and status 
using stderr stream to emit information. Use 
reporter:counter:group,counter,amount  to update a counter. Use 
reporter:status:message to update status. 


This addresses bug HIVE-305.
https://issues.apache.org/jira/browse/HIVE-305


Diffs
-

  
http://svn.apache.org/repos/asf/hive/trunk/common/src/java/org/apache/hadoop/hive/conf/HiveConf.java
 1431196 
  
http://svn.apache.org/repos/asf/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/ScriptOperator.java
 1431196 

Diff: https://reviews.apache.org/r/8933/diff/


Testing
---

just simple test it. No new test case because I couldn't find the good way to 
write it . Any suggestions are welcome.


Thanks,

Guo Hongjie



[jira] [Commented] (HIVE-305) Port Hadoop streaming's counters/status reporters to Hive Transforms

2013-01-11 Thread Guo Hongjie (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-305?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13551801#comment-13551801
 ] 

Guo Hongjie commented on HIVE-305:
--

i found that in the mail list.thanks

 Port Hadoop streaming's counters/status reporters to Hive Transforms
 

 Key: HIVE-305
 URL: https://issues.apache.org/jira/browse/HIVE-305
 Project: Hive
  Issue Type: New Feature
  Components: Query Processor
Reporter: Venky Iyer
Assignee: Guo Hongjie
 Attachments: HIVE-305.1.patch, HIVE-305.2.patch


 https://issues.apache.org/jira/browse/HADOOP-1328
  Introduced a way for a streaming process to update global counters and 
 status using stderr stream to emit information. Use 
 reporter:counter:group,counter,amount  to update  a counter. Use 
 reporter:status:message to update status. 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira