[jira] [Assigned] (HIVE-3566) Hive List Bucketing - Query Logic Optimization

2012-10-11 Thread Gang Tim Liu (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-3566?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gang Tim Liu reassigned HIVE-3566:
--

Assignee: Gang Tim Liu

 Hive List Bucketing - Query Logic Optimization
 --

 Key: HIVE-3566
 URL: https://issues.apache.org/jira/browse/HIVE-3566
 Project: Hive
  Issue Type: Improvement
Reporter: Gang Tim Liu
Assignee: Gang Tim Liu
Priority: Minor

 There are a few places to optimize list bucketing query:
 1. evaluateExprOnCell() need not be called for all cells for 'Other'. Stop 
 the first time, any 'Other' is false.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (HIVE-3566) Hive List Bucketing - Query Logic Optimization

2012-10-11 Thread Gang Tim Liu (JIRA)
Gang Tim Liu created HIVE-3566:
--

 Summary: Hive List Bucketing - Query Logic Optimization
 Key: HIVE-3566
 URL: https://issues.apache.org/jira/browse/HIVE-3566
 Project: Hive
  Issue Type: Improvement
Reporter: Gang Tim Liu
Priority: Minor


There are a few places to optimize list bucketing query:
1. evaluateExprOnCell() need not be called for all cells for 'Other'. Stop the 
first time, any 'Other' is false.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (HIVE-3567) Hive List Bucketing - Query logic supports more operator thank equal

2012-10-11 Thread Gang Tim Liu (JIRA)
Gang Tim Liu created HIVE-3567:
--

 Summary: Hive List Bucketing - Query logic supports more operator 
thank equal
 Key: HIVE-3567
 URL: https://issues.apache.org/jira/browse/HIVE-3567
 Project: Hive
  Issue Type: New Feature
Reporter: Gang Tim Liu


Query logic supports equal. It should handle , =,  and = also.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Assigned] (HIVE-3567) Hive List Bucketing - Query logic supports more operator thank equal

2012-10-11 Thread Gang Tim Liu (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-3567?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gang Tim Liu reassigned HIVE-3567:
--

Assignee: Gang Tim Liu

 Hive List Bucketing - Query logic supports more operator thank equal
 

 Key: HIVE-3567
 URL: https://issues.apache.org/jira/browse/HIVE-3567
 Project: Hive
  Issue Type: New Feature
Reporter: Gang Tim Liu
Assignee: Gang Tim Liu

 Query logic supports equal. It should handle , =,  and = also.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-3567) Hive List Bucketing - Query logic supports more operator than equal

2012-10-11 Thread Gang Tim Liu (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-3567?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gang Tim Liu updated HIVE-3567:
---

Summary: Hive List Bucketing - Query logic supports more operator than 
equal  (was: Hive List Bucketing - Query logic supports more operator thank 
equal)

 Hive List Bucketing - Query logic supports more operator than equal
 ---

 Key: HIVE-3567
 URL: https://issues.apache.org/jira/browse/HIVE-3567
 Project: Hive
  Issue Type: New Feature
Reporter: Gang Tim Liu
Assignee: Gang Tim Liu

 Query logic supports equal. It should handle , =,  and = also.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-3276) optimize union sub-queries

2012-10-11 Thread Carl Steinbach (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-3276?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13473935#comment-13473935
 ] 

Carl Steinbach commented on HIVE-3276:
--

@Namit: I added two comments on phabricator. I'm looking at this pretty late so 
feel free to ignore them.

 optimize union sub-queries
 --

 Key: HIVE-3276
 URL: https://issues.apache.org/jira/browse/HIVE-3276
 Project: Hive
  Issue Type: Bug
Reporter: Namit Jain
Assignee: Namit Jain
 Attachments: hive.3276.10.patch, hive.3276.11.patch, 
 HIVE-3276.1.patch, hive.3276.2.patch, hive.3276.3.patch, hive.3276.4.patch, 
 hive.3276.5.patch, hive.3276.6.patch, hive.3276.7.patch, hive.3276.8.patch, 
 hive.3276.9.patch


 It might be a good idea to optimize simple union queries containing 
 map-reduce jobs in at least one of the sub-qeuries.
 For eg:
 a query like:
 insert overwrite table T1 partition P1
 select * from 
 (
   subq1
 union all
   subq2
 ) u;
 today creates 3 map-reduce jobs, one for subq1, another for subq2 and 
 the final one for the union. 
 It might be a good idea to optimize this. Instead of creating the union 
 task, it might be simpler to create a move task (or something like a move
 task), where the outputs of the two sub-queries will be moved to the final 
 directory. This can easily extend to more than 2 sub-queries in the union.
 This is very useful if there is a select * followed by filesink after the
 union. This can be independently useful, and also be used to optimize the
 skewed joins https://cwiki.apache.org/Hive/skewed-join-optimization.html.
 If there is a select, filter between the union and the filesink, the select
 and the filter can be moved before the union, and the follow-up job can
 still be removed.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (HIVE-3568) No pushdown subquery's group by key in outside filter

2012-10-11 Thread Gang Tim Liu (JIRA)
Gang Tim Liu created HIVE-3568:
--

 Summary: No pushdown subquery's group by key in outside filter
 Key: HIVE-3568
 URL: https://issues.apache.org/jira/browse/HIVE-3568
 Project: Hive
  Issue Type: Bug
  Components: Query Processor
Reporter: Gang Tim Liu
Priority: Minor


pushdown doesn't happen with group by in subquery and group by key in outside 
filter.

select * from (select x, count(1) from fact_daily where ds ='1' group by x) 
subq where x = 484;

subquery still scans the partition.

expect only scan x = 484.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-3554) Hive List Bucketing - Query logic

2012-10-11 Thread Gang Tim Liu (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-3554?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gang Tim Liu updated HIVE-3554:
---

Status: Patch Available  (was: In Progress)

Yes, link other issues.

Yes, patch is available.

 Hive List Bucketing - Query logic
 -

 Key: HIVE-3554
 URL: https://issues.apache.org/jira/browse/HIVE-3554
 Project: Hive
  Issue Type: New Feature
Reporter: Gang Tim Liu
Assignee: Gang Tim Liu
 Attachments: HIVE-3554.patch.1, HIVE-3554.patch.2, HIVE-3554.patch.3


 This is part of efforts for list bucketing feature: 
 https://cwiki.apache.org/Hive/listbucketing.html
 This patch includes:
 1. Query logic: hive chooses right sub-directory instead of partition 
 directory.
 2. alter table grammar which is required to support query logic
 This patch doesn't include list bucketing DML. Main reasons:
 1. risk. w/o DML, this patch won't impact any existing hive regression 
 features since no touch on any data manipulation so that very low risk.
 2. manageability. w/ DML, patch is getting bigger and hard to review. 
 Removing DML, it's easy to review.
 We still disable hive feature by default since DML is not in yet.
 DML will be in follow-up patch. 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-3554) Hive List Bucketing - Query logic

2012-10-11 Thread Gang Tim Liu (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-3554?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gang Tim Liu updated HIVE-3554:
---

Attachment: HIVE-3554.patch.3

 Hive List Bucketing - Query logic
 -

 Key: HIVE-3554
 URL: https://issues.apache.org/jira/browse/HIVE-3554
 Project: Hive
  Issue Type: New Feature
Reporter: Gang Tim Liu
Assignee: Gang Tim Liu
 Attachments: HIVE-3554.patch.1, HIVE-3554.patch.2, HIVE-3554.patch.3


 This is part of efforts for list bucketing feature: 
 https://cwiki.apache.org/Hive/listbucketing.html
 This patch includes:
 1. Query logic: hive chooses right sub-directory instead of partition 
 directory.
 2. alter table grammar which is required to support query logic
 This patch doesn't include list bucketing DML. Main reasons:
 1. risk. w/o DML, this patch won't impact any existing hive regression 
 features since no touch on any data manipulation so that very low risk.
 2. manageability. w/ DML, patch is getting bigger and hard to review. 
 Removing DML, it's easy to review.
 We still disable hive feature by default since DML is not in yet.
 DML will be in follow-up patch. 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Work started] (HIVE-3554) Hive List Bucketing - Query logic

2012-10-11 Thread Gang Tim Liu (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-3554?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Work on HIVE-3554 started by Gang Tim Liu.

 Hive List Bucketing - Query logic
 -

 Key: HIVE-3554
 URL: https://issues.apache.org/jira/browse/HIVE-3554
 Project: Hive
  Issue Type: New Feature
Reporter: Gang Tim Liu
Assignee: Gang Tim Liu
 Attachments: HIVE-3554.patch.1, HIVE-3554.patch.2, HIVE-3554.patch.3


 This is part of efforts for list bucketing feature: 
 https://cwiki.apache.org/Hive/listbucketing.html
 This patch includes:
 1. Query logic: hive chooses right sub-directory instead of partition 
 directory.
 2. alter table grammar which is required to support query logic
 This patch doesn't include list bucketing DML. Main reasons:
 1. risk. w/o DML, this patch won't impact any existing hive regression 
 features since no touch on any data manipulation so that very low risk.
 2. manageability. w/ DML, patch is getting bigger and hard to review. 
 Removing DML, it's easy to review.
 We still disable hive feature by default since DML is not in yet.
 DML will be in follow-up patch. 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-3213) ODBC API enhancements

2012-10-11 Thread Carl Steinbach (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-3213?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13473969#comment-13473969
 ] 

Carl Steinbach commented on HIVE-3213:
--

@Prasad: Please attach the patch, provide a link to the review request, and 
change the status to 'patch available' when you're ready to have this reviewed. 
Thanks.

 ODBC API enhancements
 -

 Key: HIVE-3213
 URL: https://issues.apache.org/jira/browse/HIVE-3213
 Project: Hive
  Issue Type: Sub-task
  Components: ODBC
Affects Versions: 0.10.0
Reporter: Prasad Mujumdar
Assignee: Prasad Mujumdar
 Fix For: 0.10.0




--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


Build failed in Jenkins: Hive-0.9.1-SNAPSHOT-h0.21-keepgoing=false #165

2012-10-11 Thread Apache Jenkins Server
See 
https://builds.apache.org/job/Hive-0.9.1-SNAPSHOT-h0.21-keepgoing=false/165/

--
[...truncated 10125 lines...]
 [echo] Project: odbc
 [copy] Warning: 
https://builds.apache.org/job/Hive-0.9.1-SNAPSHOT-h0.21-keepgoing=false/ws/hive/odbc/src/conf
 does not exist.

ivy-resolve-test:
 [echo] Project: odbc

ivy-retrieve-test:
 [echo] Project: odbc

compile-test:
 [echo] Project: odbc

create-dirs:
 [echo] Project: serde
 [copy] Warning: 
https://builds.apache.org/job/Hive-0.9.1-SNAPSHOT-h0.21-keepgoing=false/ws/hive/serde/src/test/resources
 does not exist.

init:
 [echo] Project: serde

ivy-init-settings:
 [echo] Project: serde

ivy-resolve:
 [echo] Project: serde
[ivy:resolve] :: loading settings :: file = 
https://builds.apache.org/job/Hive-0.9.1-SNAPSHOT-h0.21-keepgoing=false/ws/hive/ivy/ivysettings.xml
[ivy:report] Processing 
https://builds.apache.org/job/Hive-0.9.1-SNAPSHOT-h0.21-keepgoing=false/165/artifact/hive/build/ivy/resolution-cache/org.apache.hive-hive-serde-default.xml
 to 
https://builds.apache.org/job/Hive-0.9.1-SNAPSHOT-h0.21-keepgoing=false/165/artifact/hive/build/ivy/report/org.apache.hive-hive-serde-default.html

ivy-retrieve:
 [echo] Project: serde

dynamic-serde:

compile:
 [echo] Project: serde

ivy-resolve-test:
 [echo] Project: serde

ivy-retrieve-test:
 [echo] Project: serde

compile-test:
 [echo] Project: serde
[javac] Compiling 26 source files to 
https://builds.apache.org/job/Hive-0.9.1-SNAPSHOT-h0.21-keepgoing=false/165/artifact/hive/build/serde/test/classes
[javac] Note: Some input files use or override a deprecated API.
[javac] Note: Recompile with -Xlint:deprecation for details.
[javac] Note: Some input files use unchecked or unsafe operations.
[javac] Note: Recompile with -Xlint:unchecked for details.

create-dirs:
 [echo] Project: service
 [copy] Warning: 
https://builds.apache.org/job/Hive-0.9.1-SNAPSHOT-h0.21-keepgoing=false/ws/hive/service/src/test/resources
 does not exist.

init:
 [echo] Project: service

ivy-init-settings:
 [echo] Project: service

ivy-resolve:
 [echo] Project: service
[ivy:resolve] :: loading settings :: file = 
https://builds.apache.org/job/Hive-0.9.1-SNAPSHOT-h0.21-keepgoing=false/ws/hive/ivy/ivysettings.xml
[ivy:report] Processing 
https://builds.apache.org/job/Hive-0.9.1-SNAPSHOT-h0.21-keepgoing=false/165/artifact/hive/build/ivy/resolution-cache/org.apache.hive-hive-service-default.xml
 to 
https://builds.apache.org/job/Hive-0.9.1-SNAPSHOT-h0.21-keepgoing=false/165/artifact/hive/build/ivy/report/org.apache.hive-hive-service-default.html

ivy-retrieve:
 [echo] Project: service

compile:
 [echo] Project: service

ivy-resolve-test:
 [echo] Project: service

ivy-retrieve-test:
 [echo] Project: service

compile-test:
 [echo] Project: service
[javac] Compiling 2 source files to 
https://builds.apache.org/job/Hive-0.9.1-SNAPSHOT-h0.21-keepgoing=false/165/artifact/hive/build/service/test/classes

test:
 [echo] Project: hive

test-shims:
 [echo] Project: hive

test-conditions:
 [echo] Project: shims

gen-test:
 [echo] Project: shims

create-dirs:
 [echo] Project: shims
 [copy] Warning: 
https://builds.apache.org/job/Hive-0.9.1-SNAPSHOT-h0.21-keepgoing=false/ws/hive/shims/src/test/resources
 does not exist.

init:
 [echo] Project: shims

ivy-init-settings:
 [echo] Project: shims

ivy-resolve:
 [echo] Project: shims
[ivy:resolve] :: loading settings :: file = 
https://builds.apache.org/job/Hive-0.9.1-SNAPSHOT-h0.21-keepgoing=false/ws/hive/ivy/ivysettings.xml
[ivy:report] Processing 
https://builds.apache.org/job/Hive-0.9.1-SNAPSHOT-h0.21-keepgoing=false/165/artifact/hive/build/ivy/resolution-cache/org.apache.hive-hive-shims-default.xml
 to 
https://builds.apache.org/job/Hive-0.9.1-SNAPSHOT-h0.21-keepgoing=false/165/artifact/hive/build/ivy/report/org.apache.hive-hive-shims-default.html

ivy-retrieve:
 [echo] Project: shims

compile:
 [echo] Project: shims
 [echo] Building shims 0.20

build_shims:
 [echo] Project: shims
 [echo] Compiling 
https://builds.apache.org/job/Hive-0.9.1-SNAPSHOT-h0.21-keepgoing=false/ws/hive/shims/src/common/java;/home/jenkins/jenkins-slave/workspace/Hive-0.9.1-SNAPSHOT-h0.21-keepgoing=false/hive/shims/src/0.20/java
 against hadoop 0.20.2 
(https://builds.apache.org/job/Hive-0.9.1-SNAPSHOT-h0.21-keepgoing=false/165/artifact/hive/build/hadoopcore/hadoop-0.20.2)

ivy-init-settings:
 [echo] Project: shims

ivy-resolve-hadoop-shim:
 [echo] Project: shims
[ivy:resolve] :: loading settings :: file = 
https://builds.apache.org/job/Hive-0.9.1-SNAPSHOT-h0.21-keepgoing=false/ws/hive/ivy/ivysettings.xml

ivy-retrieve-hadoop-shim:
 [echo] Project: shims
 [echo] Building shims 0.20S

build_shims:
 [echo] Project: shims
 [echo] Compiling 

[jira] [Updated] (HIVE-3554) Hive List Bucketing - Query logic

2012-10-11 Thread Gang Tim Liu (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-3554?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gang Tim Liu updated HIVE-3554:
---

Attachment: HIVE-3554.patch.4

 Hive List Bucketing - Query logic
 -

 Key: HIVE-3554
 URL: https://issues.apache.org/jira/browse/HIVE-3554
 Project: Hive
  Issue Type: New Feature
Reporter: Gang Tim Liu
Assignee: Gang Tim Liu
 Attachments: HIVE-3554.patch.1, HIVE-3554.patch.2, HIVE-3554.patch.3, 
 HIVE-3554.patch.4


 This is part of efforts for list bucketing feature: 
 https://cwiki.apache.org/Hive/listbucketing.html
 This patch includes:
 1. Query logic: hive chooses right sub-directory instead of partition 
 directory.
 2. alter table grammar which is required to support query logic
 This patch doesn't include list bucketing DML. Main reasons:
 1. risk. w/o DML, this patch won't impact any existing hive regression 
 features since no touch on any data manipulation so that very low risk.
 2. manageability. w/ DML, patch is getting bigger and hard to review. 
 Removing DML, it's easy to review.
 We still disable hive feature by default since DML is not in yet.
 DML will be in follow-up patch. 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-3569) RCFile requires native Hadoop library

2012-10-11 Thread David Phillips (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-3569?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

David Phillips updated HIVE-3569:
-

Description: 
RCFile requires the native Hadoop library. It does not work when using the Java 
{{GzipCodec}}.

The root cause is that the two versions of {{GzipCodec.createInputStream()}} 
work differently. The native version simply saves a reference to the supplied 
input stream. The Java version wraps the stream in a Java {{GZIPInputStream}}, 
which immediately tries to read the header.

The problem occurs because the stream passed by the {{RCFile.ValueBuffer}} 
constructor is empty (the buffer backing the stream is still empty at that 
point).


{noformat}
12/10/11 10:37:25 WARN util.NativeCodeLoader: Unable to load native-hadoop 
library for your platform... using builtin-java classes where applicable
12/10/11 10:37:25 INFO io.CodecPool: Got brand-new decompressor
12/10/11 10:37:25 INFO io.CodecPool: Got brand-new decompressor
Exception in thread main java.io.EOFException
at java.util.zip.GZIPInputStream.readUByte(GZIPInputStream.java:264)
at java.util.zip.GZIPInputStream.readUShort(GZIPInputStream.java:254)
at java.util.zip.GZIPInputStream.readHeader(GZIPInputStream.java:163)
at java.util.zip.GZIPInputStream.init(GZIPInputStream.java:78)
at java.util.zip.GZIPInputStream.init(GZIPInputStream.java:90)
at 
org.apache.hadoop.io.compress.GzipCodec$GzipInputStream$ResetableGZIPInputStream.init(GzipCodec.java:92)
at 
org.apache.hadoop.io.compress.GzipCodec$GzipInputStream.init(GzipCodec.java:101)
at 
org.apache.hadoop.io.compress.GzipCodec.createInputStream(GzipCodec.java:169)
at 
org.apache.hadoop.io.compress.GzipCodec.createInputStream(GzipCodec.java:179)
at 
org.apache.hadoop.hive.ql.io.RCFile$ValueBuffer.init(RCFile.java:451)
at org.apache.hadoop.hive.ql.io.RCFile$Reader.init(RCFile.java:1205)
at org.apache.hadoop.hive.ql.io.RCFile$Reader.init(RCFile.java:)
at 
org.apache.hadoop.hive.ql.io.RCFileRecordReader.init(RCFileRecordReader.java:52)
{noformat}

  was:
RCFile requires the native Hadoop library. It does not work when using the Java 
{{GzipCodec}}.

The root cause is that the two versions of {{GzipCodec.createInputStream()}} 
work differently. The native version simply saves a reference to the supplied 
input stream. The Java version wraps the stream in a Java {{GZIPInputStream}}, 
which immediately tries to read the header.

The problem occurs because the stream passed by the {{RCFile.ValueBuffer}} 
constructor is empty (the buffer backing the stream is still empty at that 
point).

{noformat}
12/10/11 10:37:25 WARN util.NativeCodeLoader: Unable to load native-hadoop 
library for your platform... using builtin-java classes where applicable
12/10/11 10:37:25 INFO io.CodecPool: Got brand-new decompressor
12/10/11 10:37:25 INFO io.CodecPool: Got brand-new decompressor
Exception in thread main java.io.EOFException
at java.util.zip.GZIPInputStream.readUByte(GZIPInputStream.java:264)
at java.util.zip.GZIPInputStream.readUShort(GZIPInputStream.java:254)
at java.util.zip.GZIPInputStream.readHeader(GZIPInputStream.java:163)
at java.util.zip.GZIPInputStream.init(GZIPInputStream.java:78)
at java.util.zip.GZIPInputStream.init(GZIPInputStream.java:90)
at 
org.apache.hadoop.io.compress.GzipCodec$GzipInputStream$ResetableGZIPInputStream.init(GzipCodec.java:92)
at 
org.apache.hadoop.io.compress.GzipCodec$GzipInputStream.init(GzipCodec.java:101)
at 
org.apache.hadoop.io.compress.GzipCodec.createInputStream(GzipCodec.java:169)
at 
org.apache.hadoop.io.compress.GzipCodec.createInputStream(GzipCodec.java:179)
at 
org.apache.hadoop.hive.ql.io.RCFile$ValueBuffer.init(RCFile.java:451)
at org.apache.hadoop.hive.ql.io.RCFile$Reader.init(RCFile.java:1205)
at org.apache.hadoop.hive.ql.io.RCFile$Reader.init(RCFile.java:)
at 
org.apache.hadoop.hive.ql.io.RCFileRecordReader.init(RCFileRecordReader.java:52)
{noformat}


 RCFile requires native Hadoop library
 -

 Key: HIVE-3569
 URL: https://issues.apache.org/jira/browse/HIVE-3569
 Project: Hive
  Issue Type: Bug
Affects Versions: 0.10.0
Reporter: David Phillips

 RCFile requires the native Hadoop library. It does not work when using the 
 Java {{GzipCodec}}.
 The root cause is that the two versions of {{GzipCodec.createInputStream()}} 
 work differently. The native version simply saves a reference to the supplied 
 input stream. The Java version wraps the stream in a Java 
 {{GZIPInputStream}}, which immediately tries to read the header.
 The problem occurs because the stream passed by the {{RCFile.ValueBuffer}} 
 constructor is empty (the buffer backing the 

[jira] [Updated] (HIVE-3569) RCFile requires native Hadoop library

2012-10-11 Thread David Phillips (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-3569?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

David Phillips updated HIVE-3569:
-

Description: 
RCFile requires the native Hadoop library. It does not work when using the Java 
{{GzipCodec}}.

The root cause is that the two versions of {{GzipCodec.createInputStream()}} 
work differently. The native version simply saves a reference to the supplied 
input stream. The Java version wraps the stream in a Java {{GZIPInputStream}}, 
which immediately tries to read the header.

The problem occurs because the stream passed by the {{RCFile.ValueBuffer}} 
constructor is empty (the buffer backing the stream is still empty at that 
point).



{noformat}
12/10/11 10:37:25 WARN util.NativeCodeLoader: Unable to load native-hadoop 
library for your platform... using builtin-java classes where applicable
12/10/11 10:37:25 INFO io.CodecPool: Got brand-new decompressor
12/10/11 10:37:25 INFO io.CodecPool: Got brand-new decompressor
Exception in thread main java.io.EOFException
at java.util.zip.GZIPInputStream.readUByte(GZIPInputStream.java:264)
at java.util.zip.GZIPInputStream.readUShort(GZIPInputStream.java:254)
at java.util.zip.GZIPInputStream.readHeader(GZIPInputStream.java:163)
at java.util.zip.GZIPInputStream.init(GZIPInputStream.java:78)
at java.util.zip.GZIPInputStream.init(GZIPInputStream.java:90)
at 
org.apache.hadoop.io.compress.GzipCodec$GzipInputStream$ResetableGZIPInputStream.init(GzipCodec.java:92)
at 
org.apache.hadoop.io.compress.GzipCodec$GzipInputStream.init(GzipCodec.java:101)
at 
org.apache.hadoop.io.compress.GzipCodec.createInputStream(GzipCodec.java:169)
at 
org.apache.hadoop.io.compress.GzipCodec.createInputStream(GzipCodec.java:179)
at 
org.apache.hadoop.hive.ql.io.RCFile$ValueBuffer.init(RCFile.java:451)
at org.apache.hadoop.hive.ql.io.RCFile$Reader.init(RCFile.java:1205)
at org.apache.hadoop.hive.ql.io.RCFile$Reader.init(RCFile.java:)
at 
org.apache.hadoop.hive.ql.io.RCFileRecordReader.init(RCFileRecordReader.java:52)
{noformat}

  was:
RCFile requires the native Hadoop library. It does not work when using the Java 
{{GzipCodec}}.

The root cause is that the two versions of {{GzipCodec.createInputStream()}} 
work differently. The native version simply saves a reference to the supplied 
input stream. The Java version wraps the stream in a Java {{GZIPInputStream}}, 
which immediately tries to read the header.

The problem occurs because the stream passed by the {{RCFile.ValueBuffer}} 
constructor is empty (the buffer backing the stream is still empty at that 
point).


{noformat}
12/10/11 10:37:25 WARN util.NativeCodeLoader: Unable to load native-hadoop 
library for your platform... using builtin-java classes where applicable
12/10/11 10:37:25 INFO io.CodecPool: Got brand-new decompressor
12/10/11 10:37:25 INFO io.CodecPool: Got brand-new decompressor
Exception in thread main java.io.EOFException
at java.util.zip.GZIPInputStream.readUByte(GZIPInputStream.java:264)
at java.util.zip.GZIPInputStream.readUShort(GZIPInputStream.java:254)
at java.util.zip.GZIPInputStream.readHeader(GZIPInputStream.java:163)
at java.util.zip.GZIPInputStream.init(GZIPInputStream.java:78)
at java.util.zip.GZIPInputStream.init(GZIPInputStream.java:90)
at 
org.apache.hadoop.io.compress.GzipCodec$GzipInputStream$ResetableGZIPInputStream.init(GzipCodec.java:92)
at 
org.apache.hadoop.io.compress.GzipCodec$GzipInputStream.init(GzipCodec.java:101)
at 
org.apache.hadoop.io.compress.GzipCodec.createInputStream(GzipCodec.java:169)
at 
org.apache.hadoop.io.compress.GzipCodec.createInputStream(GzipCodec.java:179)
at 
org.apache.hadoop.hive.ql.io.RCFile$ValueBuffer.init(RCFile.java:451)
at org.apache.hadoop.hive.ql.io.RCFile$Reader.init(RCFile.java:1205)
at org.apache.hadoop.hive.ql.io.RCFile$Reader.init(RCFile.java:)
at 
org.apache.hadoop.hive.ql.io.RCFileRecordReader.init(RCFileRecordReader.java:52)
{noformat}


 RCFile requires native Hadoop library
 -

 Key: HIVE-3569
 URL: https://issues.apache.org/jira/browse/HIVE-3569
 Project: Hive
  Issue Type: Bug
Affects Versions: 0.10.0
Reporter: David Phillips

 RCFile requires the native Hadoop library. It does not work when using the 
 Java {{GzipCodec}}.
 The root cause is that the two versions of {{GzipCodec.createInputStream()}} 
 work differently. The native version simply saves a reference to the supplied 
 input stream. The Java version wraps the stream in a Java 
 {{GZIPInputStream}}, which immediately tries to read the header.
 The problem occurs because the stream passed by the {{RCFile.ValueBuffer}} 
 constructor is empty (the buffer backing the 

[jira] [Updated] (HIVE-3569) RCFile requires native Hadoop library

2012-10-11 Thread David Phillips (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-3569?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

David Phillips updated HIVE-3569:
-

Description: 
RCFile requires the native Hadoop library. It does not work when using the Java 
{{GzipCodec}}.

The root cause is that the two versions of {{GzipCodec.createInputStream()}} 
work differently. The native version simply saves a reference to the supplied 
input stream. The Java version wraps the stream in a Java {{GZIPInputStream}}, 
which immediately tries to read the header.

The problem occurs because the stream passed by the {{RCFile.ValueBuffer}} 
constructor is empty (the buffer backing the stream is still empty at that 
point).

{noformat}
12/10/11 10:37:25 WARN util.NativeCodeLoader: Unable to load native-hadoop 
library for your platform... using builtin-java classes where applicable
12/10/11 10:37:25 INFO io.CodecPool: Got brand-new decompressor
12/10/11 10:37:25 INFO io.CodecPool: Got brand-new decompressor
Exception in thread main java.io.EOFException
at java.util.zip.GZIPInputStream.readUByte(GZIPInputStream.java:264)
at java.util.zip.GZIPInputStream.readUShort(GZIPInputStream.java:254)
at java.util.zip.GZIPInputStream.readHeader(GZIPInputStream.java:163)
at java.util.zip.GZIPInputStream.init(GZIPInputStream.java:78)
at java.util.zip.GZIPInputStream.init(GZIPInputStream.java:90)
at 
org.apache.hadoop.io.compress.GzipCodec$GzipInputStream$ResetableGZIPInputStream.init(GzipCodec.java:92)
at 
org.apache.hadoop.io.compress.GzipCodec$GzipInputStream.init(GzipCodec.java:101)
at 
org.apache.hadoop.io.compress.GzipCodec.createInputStream(GzipCodec.java:169)
at 
org.apache.hadoop.io.compress.GzipCodec.createInputStream(GzipCodec.java:179)
at 
org.apache.hadoop.hive.ql.io.RCFile$ValueBuffer.init(RCFile.java:451)
at org.apache.hadoop.hive.ql.io.RCFile$Reader.init(RCFile.java:1205)
at org.apache.hadoop.hive.ql.io.RCFile$Reader.init(RCFile.java:)
at 
org.apache.hadoop.hive.ql.io.RCFileRecordReader.init(RCFileRecordReader.java:52)
{noformat}

  was:
RCFile requires the native Hadoop library. It does not work when using the Java 
{{GzipCodec}}.

The root cause is that the two versions of {{GzipCodec.createInputStream()}} 
work differently. The native version simply saves a reference to the supplied 
input stream. The Java version wraps the stream in a Java {{GZIPInputStream}}, 
which immediately tries to read the header.

The problem occurs because the stream passed by the {{RCFile.ValueBuffer}} 
constructor is empty (the buffer backing the stream is still empty at that 
point).



{noformat}
12/10/11 10:37:25 WARN util.NativeCodeLoader: Unable to load native-hadoop 
library for your platform... using builtin-java classes where applicable
12/10/11 10:37:25 INFO io.CodecPool: Got brand-new decompressor
12/10/11 10:37:25 INFO io.CodecPool: Got brand-new decompressor
Exception in thread main java.io.EOFException
at java.util.zip.GZIPInputStream.readUByte(GZIPInputStream.java:264)
at java.util.zip.GZIPInputStream.readUShort(GZIPInputStream.java:254)
at java.util.zip.GZIPInputStream.readHeader(GZIPInputStream.java:163)
at java.util.zip.GZIPInputStream.init(GZIPInputStream.java:78)
at java.util.zip.GZIPInputStream.init(GZIPInputStream.java:90)
at 
org.apache.hadoop.io.compress.GzipCodec$GzipInputStream$ResetableGZIPInputStream.init(GzipCodec.java:92)
at 
org.apache.hadoop.io.compress.GzipCodec$GzipInputStream.init(GzipCodec.java:101)
at 
org.apache.hadoop.io.compress.GzipCodec.createInputStream(GzipCodec.java:169)
at 
org.apache.hadoop.io.compress.GzipCodec.createInputStream(GzipCodec.java:179)
at 
org.apache.hadoop.hive.ql.io.RCFile$ValueBuffer.init(RCFile.java:451)
at org.apache.hadoop.hive.ql.io.RCFile$Reader.init(RCFile.java:1205)
at org.apache.hadoop.hive.ql.io.RCFile$Reader.init(RCFile.java:)
at 
org.apache.hadoop.hive.ql.io.RCFileRecordReader.init(RCFileRecordReader.java:52)
{noformat}


 RCFile requires native Hadoop library
 -

 Key: HIVE-3569
 URL: https://issues.apache.org/jira/browse/HIVE-3569
 Project: Hive
  Issue Type: Bug
Affects Versions: 0.10.0
Reporter: David Phillips

 RCFile requires the native Hadoop library. It does not work when using the 
 Java {{GzipCodec}}.
 The root cause is that the two versions of {{GzipCodec.createInputStream()}} 
 work differently. The native version simply saves a reference to the supplied 
 input stream. The Java version wraps the stream in a Java 
 {{GZIPInputStream}}, which immediately tries to read the header.
 The problem occurs because the stream passed by the {{RCFile.ValueBuffer}} 
 constructor is empty (the buffer backing the 

[jira] [Created] (HIVE-3570) Add/fix facility to collect operator specific statisticsin hive + add hash-in/hash-out counter for GroupBy Optr

2012-10-11 Thread Satadru Pan (JIRA)
Satadru Pan created HIVE-3570:
-

 Summary: Add/fix facility to collect operator specific 
statisticsin hive + add hash-in/hash-out counter for GroupBy Optr
 Key: HIVE-3570
 URL: https://issues.apache.org/jira/browse/HIVE-3570
 Project: Hive
  Issue Type: Improvement
  Components: Statistics
Affects Versions: 0.9.0
Reporter: Satadru Pan
Priority: Minor


Requirement: Collect Operator specific stats for hive queries. Use the counter 
framework available in Hive Operator.java to accomplish that.


--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-3570) Add/fix facility to collect operator specific statisticsin hive + add hash-in/hash-out counter for GroupBy Optr

2012-10-11 Thread Satadru Pan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-3570?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Satadru Pan updated HIVE-3570:
--

Attachment: HIVE-3570.1.patch.txt

Patch for the task HIVE-3570

 Add/fix facility to collect operator specific statisticsin hive + add 
 hash-in/hash-out counter for GroupBy Optr
 ---

 Key: HIVE-3570
 URL: https://issues.apache.org/jira/browse/HIVE-3570
 Project: Hive
  Issue Type: Improvement
  Components: Statistics
Affects Versions: 0.9.0
Reporter: Satadru Pan
Priority: Minor
 Attachments: HIVE-3570.1.patch.txt


 Requirement: Collect Operator specific stats for hive queries. Use the 
 counter framework available in Hive Operator.java to accomplish that.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-3564) hivetest.py: revision number and applied patch

2012-10-11 Thread Ivan Gorbachev (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-3564?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ivan Gorbachev updated HIVE-3564:
-

Status: Patch Available  (was: Open)

 hivetest.py: revision number and applied patch
 --

 Key: HIVE-3564
 URL: https://issues.apache.org/jira/browse/HIVE-3564
 Project: Hive
  Issue Type: Improvement
  Components: Testing Infrastructure
Reporter: Ivan Gorbachev
Assignee: Ivan Gorbachev

 It's required to add new option for hivetest.py which will allow to show base 
 revision number and applied patch.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-3564) hivetest.py: revision number and applied patch

2012-10-11 Thread Ivan Gorbachev (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-3564?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13474475#comment-13474475
 ] 

Ivan Gorbachev commented on HIVE-3564:
--

https://reviews.facebook.net/D5973

 hivetest.py: revision number and applied patch
 --

 Key: HIVE-3564
 URL: https://issues.apache.org/jira/browse/HIVE-3564
 Project: Hive
  Issue Type: Improvement
  Components: Testing Infrastructure
Reporter: Ivan Gorbachev
Assignee: Ivan Gorbachev
 Attachments: hive-3564.0.patch.txt


 It's required to add new option for hivetest.py which will allow to show base 
 revision number and applied patch.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-3564) hivetest.py: revision number and applied patch

2012-10-11 Thread Ivan Gorbachev (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-3564?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ivan Gorbachev updated HIVE-3564:
-

Attachment: hive-3564.0.patch.txt

 hivetest.py: revision number and applied patch
 --

 Key: HIVE-3564
 URL: https://issues.apache.org/jira/browse/HIVE-3564
 Project: Hive
  Issue Type: Improvement
  Components: Testing Infrastructure
Reporter: Ivan Gorbachev
Assignee: Ivan Gorbachev
 Attachments: hive-3564.0.patch.txt


 It's required to add new option for hivetest.py which will allow to show base 
 revision number and applied patch.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Assigned] (HIVE-3570) Add/fix facility to collect operator specific statisticsin hive + add hash-in/hash-out counter for GroupBy Optr

2012-10-11 Thread Namit Jain (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-3570?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Namit Jain reassigned HIVE-3570:


Assignee: Satadru Pan

 Add/fix facility to collect operator specific statisticsin hive + add 
 hash-in/hash-out counter for GroupBy Optr
 ---

 Key: HIVE-3570
 URL: https://issues.apache.org/jira/browse/HIVE-3570
 Project: Hive
  Issue Type: Improvement
  Components: Statistics
Affects Versions: 0.9.0
Reporter: Satadru Pan
Assignee: Satadru Pan
Priority: Minor
 Attachments: HIVE-3570.1.patch.txt


 Requirement: Collect Operator specific stats for hive queries. Use the 
 counter framework available in Hive Operator.java to accomplish that.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-3564) hivetest.py: revision number and applied patch

2012-10-11 Thread Namit Jain (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-3564?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13474520#comment-13474520
 ] 

Namit Jain commented on HIVE-3564:
--

+1

 hivetest.py: revision number and applied patch
 --

 Key: HIVE-3564
 URL: https://issues.apache.org/jira/browse/HIVE-3564
 Project: Hive
  Issue Type: Improvement
  Components: Testing Infrastructure
Reporter: Ivan Gorbachev
Assignee: Ivan Gorbachev
 Attachments: hive-3564.0.patch.txt


 It's required to add new option for hivetest.py which will allow to show base 
 revision number and applied patch.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Assigned] (HIVE-446) Implement TRUNCATE

2012-10-11 Thread Andrew Chalfant (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-446?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Andrew Chalfant reassigned HIVE-446:


Assignee: Andrew Chalfant

 Implement TRUNCATE
 --

 Key: HIVE-446
 URL: https://issues.apache.org/jira/browse/HIVE-446
 Project: Hive
  Issue Type: New Feature
  Components: Query Processor
Reporter: Prasad Chakka
Assignee: Andrew Chalfant

 truncate the data but leave the table and metadata intact.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


Hive-trunk-h0.21 - Build # 1733 - Still Failing

2012-10-11 Thread Apache Jenkins Server
Changes for Build #1728

Changes for Build #1729

Changes for Build #1730
[namit] HIVE-3550 Early skipping for limit operator at reduce stage
(Navis via namit)

[namit] HIVE-3536 Output of sort merge join is no longer bucketed
(Kevin Wilfong) 

For some reason, I missed merging the test in my previous merge

[kevinwilfong] HIVE-3522. Make separator for Entity name configurable. 
(Raghotham Murthy via kevinwilfong)


Changes for Build #1731

Changes for Build #1732

Changes for Build #1733



1 tests failed.
FAILED:  
org.apache.hadoop.hive.cli.TestNegativeCliDriver.testNegativeCliDriver_stats_aggregator_error_1

Error Message:
Forked Java VM exited abnormally. Please note the time in the report does not 
reflect the time until the VM exit.

Stack Trace:
junit.framework.AssertionFailedError: Forked Java VM exited abnormally. Please 
note the time in the report does not reflect the time until the VM exit.
at 
net.sf.antcontrib.logic.ForTask.doSequentialIteration(ForTask.java:259)
at net.sf.antcontrib.logic.ForTask.doToken(ForTask.java:268)
at net.sf.antcontrib.logic.ForTask.doTheTasks(ForTask.java:324)
at net.sf.antcontrib.logic.ForTask.execute(ForTask.java:244)




The Apache Jenkins build system has built Hive-trunk-h0.21 (build #1733)

Status: Still Failing

Check console output at https://builds.apache.org/job/Hive-trunk-h0.21/1733/ to 
view the results.

[jira] [Created] (HIVE-3571) add a way to run a small unit quickly

2012-10-11 Thread Namit Jain (JIRA)
Namit Jain created HIVE-3571:


 Summary: add a way to run a small unit quickly
 Key: HIVE-3571
 URL: https://issues.apache.org/jira/browse/HIVE-3571
 Project: Hive
  Issue Type: Test
  Components: Testing Infrastructure
Reporter: Namit Jain


A simple unit test:

ant test -Dtestcase=TestCliDriver -Dqfile=groupby2.q

takes a long time.

There should be a quick way to achieve that for debugging.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-3518) QTestUtil side-effects

2012-10-11 Thread Navis (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-3518?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Navis updated HIVE-3518:


Assignee: Navis
  Status: Patch Available  (was: Open)

 QTestUtil side-effects
 --

 Key: HIVE-3518
 URL: https://issues.apache.org/jira/browse/HIVE-3518
 Project: Hive
  Issue Type: Bug
  Components: Testing Infrastructure, Tests
Reporter: Ivan Gorbachev
Assignee: Navis
 Attachments: HIVE-3518.D5865.1.patch, HIVE-3518.D5865.2.patch, 
 metadata_export_drop.q


 It seems that QTestUtil has side-effects. This test 
 ([^metadata_export_drop.q]) causes failure of other tests on cleanup stage:
 {quote}
 Exception: java.lang.IllegalArgumentException: java.net.URISyntaxException: 
 Relative path in absolute URI: 
 file:../build/ql/test/data/exports/HIVE-3427/src.2012-09-28-11-38-17
 org.apache.hadoop.hive.ql.metadata.HiveException: 
 java.lang.IllegalArgumentException: java.net.URISyntaxException: Relative 
 path in absolute URI: 
 file:../build/ql/test/data/exports/HIVE-3427/src.2012-09-28-11-38-17
 at org.apache.hadoop.hive.ql.metadata.Hive.dropTable(Hive.java:845)
 at org.apache.hadoop.hive.ql.metadata.Hive.dropTable(Hive.java:821)
 at org.apache.hadoop.hive.ql.QTestUtil.cleanUp(QTestUtil.java:445)
 at org.apache.hadoop.hive.ql.QTestUtil.shutdown(QTestUtil.java:300)
 at org.apache.hadoop.hive.cli.TestCliDriver.tearDown(TestCliDriver.java:87)
 at junit.framework.TestCase.runBare(TestCase.java:140)
 at junit.framework.TestResult$1.protect(TestResult.java:110)
 at junit.framework.TestResult.runProtected(TestResult.java:128)
 at junit.framework.TestResult.run(TestResult.java:113)
 at junit.framework.TestCase.run(TestCase.java:124)
 at junit.framework.TestSuite.runTest(TestSuite.java:232)
 at junit.framework.TestSuite.run(TestSuite.java:227)
 at 
 org.eclipse.jdt.internal.junit.runner.junit3.JUnit3TestReference.run(JUnit3TestReference.java:130)
 at 
 org.eclipse.jdt.internal.junit.runner.TestExecution.run(TestExecution.java:38)
 at 
 org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:460)
 at 
 org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:673)
 at 
 org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.run(RemoteTestRunner.java:386)
 at 
 org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.main(RemoteTestRunner.java:196)
 Caused by: java.lang.IllegalArgumentException: java.net.URISyntaxException: 
 Relative path in absolute URI: 
 file:../build/ql/test/data/exports/HIVE-3427/src.2012-09-28-11-38-17
 at org.apache.hadoop.fs.Path.initialize(Path.java:140)
 at org.apache.hadoop.fs.Path.init(Path.java:132)
 at 
 org.apache.hadoop.fs.ProxyFileSystem.swizzleParamPath(ProxyFileSystem.java:56)
 at org.apache.hadoop.fs.ProxyFileSystem.mkdirs(ProxyFileSystem.java:214)
 at org.apache.hadoop.fs.FilterFileSystem.mkdirs(FilterFileSystem.java:183)
 at org.apache.hadoop.fs.FileSystem.mkdirs(FileSystem.java:1120)
 at 
 org.apache.hadoop.hive.ql.parse.MetaDataExportListener.export_meta_data(MetaDataExportListener.java:81)
 at 
 org.apache.hadoop.hive.ql.parse.MetaDataExportListener.onEvent(MetaDataExportListener.java:106)
 at 
 org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.drop_table_core(HiveMetaStore.java:1024)
 at 
 org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.drop_table(HiveMetaStore.java:1185)
 at 
 org.apache.hadoop.hive.metastore.HiveMetaStoreClient.dropTable(HiveMetaStoreClient.java:566)
 at org.apache.hadoop.hive.ql.metadata.Hive.dropTable(Hive.java:839)
 ... 17 more
 Caused by: java.net.URISyntaxException: Relative path in absolute URI: 
 file:../build/ql/test/data/exports/HIVE-3427/src.2012-09-28-11-38-17
 at java.net.URI.checkPath(URI.java:1787)
 at java.net.URI.init(URI.java:735)
 at org.apache.hadoop.fs.Path.initialize(Path.java:137)
 ... 28 more
 {quote}
 Flushing 'hive.metastore.pre.event.listeners' into empty string solves the 
 issue. During debugging I figured out this property wan't cleaned for other 
 tests after it was set in metadata_export_drop.q.
 How to reproduce:
 {code} ant test -Dtestcase=TestCliDriver -Dqfile=metadata_export_drop.q,some 
 test.q{code}
 where some test.q means any test which contains CREATE statement. For 
 example, sample10.q

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-3518) QTestUtil side-effects

2012-10-11 Thread Phabricator (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-3518?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Phabricator updated HIVE-3518:
--

Attachment: HIVE-3518.D5865.2.patch

navis updated the revision HIVE-3518 [jira] QTestUtil side-effects.
Reviewers: JIRA

  Resetting conf made other issues, so just sure new conf propagated to meta 
store.


REVISION DETAIL
  https://reviews.facebook.net/D5865

AFFECTED FILES
  ql/src/test/org/apache/hadoop/hive/ql/QTestUtil.java

To: JIRA, navis


 QTestUtil side-effects
 --

 Key: HIVE-3518
 URL: https://issues.apache.org/jira/browse/HIVE-3518
 Project: Hive
  Issue Type: Bug
  Components: Testing Infrastructure, Tests
Reporter: Ivan Gorbachev
Assignee: Navis
 Attachments: HIVE-3518.D5865.1.patch, HIVE-3518.D5865.2.patch, 
 metadata_export_drop.q


 It seems that QTestUtil has side-effects. This test 
 ([^metadata_export_drop.q]) causes failure of other tests on cleanup stage:
 {quote}
 Exception: java.lang.IllegalArgumentException: java.net.URISyntaxException: 
 Relative path in absolute URI: 
 file:../build/ql/test/data/exports/HIVE-3427/src.2012-09-28-11-38-17
 org.apache.hadoop.hive.ql.metadata.HiveException: 
 java.lang.IllegalArgumentException: java.net.URISyntaxException: Relative 
 path in absolute URI: 
 file:../build/ql/test/data/exports/HIVE-3427/src.2012-09-28-11-38-17
 at org.apache.hadoop.hive.ql.metadata.Hive.dropTable(Hive.java:845)
 at org.apache.hadoop.hive.ql.metadata.Hive.dropTable(Hive.java:821)
 at org.apache.hadoop.hive.ql.QTestUtil.cleanUp(QTestUtil.java:445)
 at org.apache.hadoop.hive.ql.QTestUtil.shutdown(QTestUtil.java:300)
 at org.apache.hadoop.hive.cli.TestCliDriver.tearDown(TestCliDriver.java:87)
 at junit.framework.TestCase.runBare(TestCase.java:140)
 at junit.framework.TestResult$1.protect(TestResult.java:110)
 at junit.framework.TestResult.runProtected(TestResult.java:128)
 at junit.framework.TestResult.run(TestResult.java:113)
 at junit.framework.TestCase.run(TestCase.java:124)
 at junit.framework.TestSuite.runTest(TestSuite.java:232)
 at junit.framework.TestSuite.run(TestSuite.java:227)
 at 
 org.eclipse.jdt.internal.junit.runner.junit3.JUnit3TestReference.run(JUnit3TestReference.java:130)
 at 
 org.eclipse.jdt.internal.junit.runner.TestExecution.run(TestExecution.java:38)
 at 
 org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:460)
 at 
 org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:673)
 at 
 org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.run(RemoteTestRunner.java:386)
 at 
 org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.main(RemoteTestRunner.java:196)
 Caused by: java.lang.IllegalArgumentException: java.net.URISyntaxException: 
 Relative path in absolute URI: 
 file:../build/ql/test/data/exports/HIVE-3427/src.2012-09-28-11-38-17
 at org.apache.hadoop.fs.Path.initialize(Path.java:140)
 at org.apache.hadoop.fs.Path.init(Path.java:132)
 at 
 org.apache.hadoop.fs.ProxyFileSystem.swizzleParamPath(ProxyFileSystem.java:56)
 at org.apache.hadoop.fs.ProxyFileSystem.mkdirs(ProxyFileSystem.java:214)
 at org.apache.hadoop.fs.FilterFileSystem.mkdirs(FilterFileSystem.java:183)
 at org.apache.hadoop.fs.FileSystem.mkdirs(FileSystem.java:1120)
 at 
 org.apache.hadoop.hive.ql.parse.MetaDataExportListener.export_meta_data(MetaDataExportListener.java:81)
 at 
 org.apache.hadoop.hive.ql.parse.MetaDataExportListener.onEvent(MetaDataExportListener.java:106)
 at 
 org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.drop_table_core(HiveMetaStore.java:1024)
 at 
 org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.drop_table(HiveMetaStore.java:1185)
 at 
 org.apache.hadoop.hive.metastore.HiveMetaStoreClient.dropTable(HiveMetaStoreClient.java:566)
 at org.apache.hadoop.hive.ql.metadata.Hive.dropTable(Hive.java:839)
 ... 17 more
 Caused by: java.net.URISyntaxException: Relative path in absolute URI: 
 file:../build/ql/test/data/exports/HIVE-3427/src.2012-09-28-11-38-17
 at java.net.URI.checkPath(URI.java:1787)
 at java.net.URI.init(URI.java:735)
 at org.apache.hadoop.fs.Path.initialize(Path.java:137)
 ... 28 more
 {quote}
 Flushing 'hive.metastore.pre.event.listeners' into empty string solves the 
 issue. During debugging I figured out this property wan't cleaned for other 
 tests after it was set in metadata_export_drop.q.
 How to reproduce:
 {code} ant test -Dtestcase=TestCliDriver -Dqfile=metadata_export_drop.q,some 
 test.q{code}
 where some test.q means any test which contains CREATE statement. For 
 example, sample10.q

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-3554) Hive List Bucketing - Query logic

2012-10-11 Thread Gang Tim Liu (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-3554?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gang Tim Liu updated HIVE-3554:
---

Attachment: HIVE-3554.patch.5

 Hive List Bucketing - Query logic
 -

 Key: HIVE-3554
 URL: https://issues.apache.org/jira/browse/HIVE-3554
 Project: Hive
  Issue Type: New Feature
Reporter: Gang Tim Liu
Assignee: Gang Tim Liu
 Attachments: HIVE-3554.patch.1, HIVE-3554.patch.2, HIVE-3554.patch.3, 
 HIVE-3554.patch.4, HIVE-3554.patch.5


 This is part of efforts for list bucketing feature: 
 https://cwiki.apache.org/Hive/listbucketing.html
 This patch includes:
 1. Query logic: hive chooses right sub-directory instead of partition 
 directory.
 2. alter table grammar which is required to support query logic
 This patch doesn't include list bucketing DML. Main reasons:
 1. risk. w/o DML, this patch won't impact any existing hive regression 
 features since no touch on any data manipulation so that very low risk.
 2. manageability. w/ DML, patch is getting bigger and hard to review. 
 Removing DML, it's easy to review.
 We still disable hive feature by default since DML is not in yet.
 DML will be in follow-up patch. 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-3377) ant model-jar command fails in metastore

2012-10-11 Thread Kevin Wilfong (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-3377?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13474763#comment-13474763
 ] 

Kevin Wilfong commented on HIVE-3377:
-

Deleting line 112 in build-common.xml seems to help.  I don't know if the build 
will work in all operating systems without it though.

 ant model-jar command fails in metastore
 

 Key: HIVE-3377
 URL: https://issues.apache.org/jira/browse/HIVE-3377
 Project: Hive
  Issue Type: Bug
  Components: Metastore
Affects Versions: 0.10.0
Reporter: Vandana Ayyalasomayajula
Priority: Minor
  Labels: build

 Running ant model-jar command to set up eclipse dev environment from the 
 following wiki:
 https://cwiki.apache.org/Hive/gettingstarted-eclipsesetup.html
 fails with the following message:
 BUILD FAILED
 **/workspace/hive-trunk/metastore/build.xml:22: The following error occurred 
 while executing this line:
 **/workspace/hive-trunk/build-common.xml:112: Problem: failed to create task 
 or type osfamily
 Cause: The name is undefined.
 Action: Check the spelling.
 Action: Check that any custom tasks/types have been declared.
 Action: Check that any presetdef/macrodef declarations have taken place.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-3377) ant model-jar command fails in metastore

2012-10-11 Thread Krish (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-3377?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13474768#comment-13474768
 ] 

Krish commented on HIVE-3377:
-

Line # 112 is blank; here are line numbers 105 - 120. please let me know which 
line to delete.

105   !-- include contrib on local classpath, but not on cluster --
106   !-- https://reviews.facebook.net/D2133#comment-47 --
107   path id=test.local.classpath
108 path refid=${test.classpath.id}/
109 fileset dir=${hive.root}/build/ivy/lib/test 
includes=hive-contrib*.jar erroronmissingdir=false/
110   /path
111 
112 
113   loadproperties srcfile=${ivy.conf.dir}/libraries.properties/
114 
115   osfamily property=os.family/
116 
117   condition property=offline
118 istrue value=${is-offline}/
119   /condition
120   import file=build-offline.xml/


 ant model-jar command fails in metastore
 

 Key: HIVE-3377
 URL: https://issues.apache.org/jira/browse/HIVE-3377
 Project: Hive
  Issue Type: Bug
  Components: Metastore
Affects Versions: 0.10.0
Reporter: Vandana Ayyalasomayajula
Priority: Minor
  Labels: build

 Running ant model-jar command to set up eclipse dev environment from the 
 following wiki:
 https://cwiki.apache.org/Hive/gettingstarted-eclipsesetup.html
 fails with the following message:
 BUILD FAILED
 **/workspace/hive-trunk/metastore/build.xml:22: The following error occurred 
 while executing this line:
 **/workspace/hive-trunk/build-common.xml:112: Problem: failed to create task 
 or type osfamily
 Cause: The name is undefined.
 Action: Check the spelling.
 Action: Check that any custom tasks/types have been declared.
 Action: Check that any presetdef/macrodef declarations have taken place.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-3377) ant model-jar command fails in metastore

2012-10-11 Thread Kevin Wilfong (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-3377?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13474773#comment-13474773
 ] 

Kevin Wilfong commented on HIVE-3377:
-

115 then

 ant model-jar command fails in metastore
 

 Key: HIVE-3377
 URL: https://issues.apache.org/jira/browse/HIVE-3377
 Project: Hive
  Issue Type: Bug
  Components: Metastore
Affects Versions: 0.10.0
Reporter: Vandana Ayyalasomayajula
Priority: Minor
  Labels: build

 Running ant model-jar command to set up eclipse dev environment from the 
 following wiki:
 https://cwiki.apache.org/Hive/gettingstarted-eclipsesetup.html
 fails with the following message:
 BUILD FAILED
 **/workspace/hive-trunk/metastore/build.xml:22: The following error occurred 
 while executing this line:
 **/workspace/hive-trunk/build-common.xml:112: Problem: failed to create task 
 or type osfamily
 Cause: The name is undefined.
 Action: Check the spelling.
 Action: Check that any custom tasks/types have been declared.
 Action: Check that any presetdef/macrodef declarations have taken place.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-3377) ant model-jar command fails in metastore

2012-10-11 Thread Krish (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-3377?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13474782#comment-13474782
 ] 

Krish commented on HIVE-3377:
-

Thanks, that worked but next command ant gen-test failed.
Here is the error message, any thoughts?

/opt/hive-trunk$ ant get-test
Buildfile: /opt/hive-trunk/build.xml

BUILD FAILED
Target get-test does not exist in the project hive. 

Total time: 1 second


 ant model-jar command fails in metastore
 

 Key: HIVE-3377
 URL: https://issues.apache.org/jira/browse/HIVE-3377
 Project: Hive
  Issue Type: Bug
  Components: Metastore
Affects Versions: 0.10.0
Reporter: Vandana Ayyalasomayajula
Priority: Minor
  Labels: build

 Running ant model-jar command to set up eclipse dev environment from the 
 following wiki:
 https://cwiki.apache.org/Hive/gettingstarted-eclipsesetup.html
 fails with the following message:
 BUILD FAILED
 **/workspace/hive-trunk/metastore/build.xml:22: The following error occurred 
 while executing this line:
 **/workspace/hive-trunk/build-common.xml:112: Problem: failed to create task 
 or type osfamily
 Cause: The name is undefined.
 Action: Check the spelling.
 Action: Check that any custom tasks/types have been declared.
 Action: Check that any presetdef/macrodef declarations have taken place.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-3377) ant model-jar command fails in metastore

2012-10-11 Thread Kevin Wilfong (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-3377?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13474787#comment-13474787
 ] 

Kevin Wilfong commented on HIVE-3377:
-

Sorry, that one's new to me.

 ant model-jar command fails in metastore
 

 Key: HIVE-3377
 URL: https://issues.apache.org/jira/browse/HIVE-3377
 Project: Hive
  Issue Type: Bug
  Components: Metastore
Affects Versions: 0.10.0
Reporter: Vandana Ayyalasomayajula
Priority: Minor
  Labels: build

 Running ant model-jar command to set up eclipse dev environment from the 
 following wiki:
 https://cwiki.apache.org/Hive/gettingstarted-eclipsesetup.html
 fails with the following message:
 BUILD FAILED
 **/workspace/hive-trunk/metastore/build.xml:22: The following error occurred 
 while executing this line:
 **/workspace/hive-trunk/build-common.xml:112: Problem: failed to create task 
 or type osfamily
 Cause: The name is undefined.
 Action: Check the spelling.
 Action: Check that any custom tasks/types have been declared.
 Action: Check that any presetdef/macrodef declarations have taken place.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira