[ 
https://issues.apache.org/jira/browse/HIVE-3218?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13548255#comment-13548255
 ] 

Hudson commented on HIVE-3218:
------------------------------

Integrated in Hive-trunk-hadoop2 #54 (See 
[https://builds.apache.org/job/Hive-trunk-hadoop2/54/])
    HIVE-3218 Stream table of SMBJoin/BucketMapJoin with two or more 
          partitions is not handled properly (Navis via namit) (Revision 
1367012)

     Result = ABORTED
namit : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1367012
Files : 
* /hive/trunk/data/files/srcsortbucket1outof4.txt
* /hive/trunk/data/files/srcsortbucket2outof4.txt
* /hive/trunk/data/files/srcsortbucket3outof4.txt
* /hive/trunk/data/files/srcsortbucket4outof4.txt
* /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/BucketMatcher.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/DefaultBucketMatcher.java
* /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/ExecMapperContext.java
* /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/FileSinkOperator.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/HashTableSinkOperator.java
* /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/MapJoinOperator.java
* /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/MapredLocalTask.java
* /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/SMBMapJoinOperator.java
* /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/Utilities.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/optimizer/BucketMapJoinOptimizer.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/optimizer/GenMapRedUtils.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/optimizer/SortedMergeBucketMapJoinOptimizer.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/MapJoinResolver.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/BucketMapJoinContext.java
* /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/HashTableSinkDesc.java
* /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/MapJoinDesc.java
* /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/plan/MapredLocalWork.java
* /hive/trunk/ql/src/test/queries/clientpositive/bucketcontext_1.q
* /hive/trunk/ql/src/test/queries/clientpositive/bucketcontext_2.q
* /hive/trunk/ql/src/test/queries/clientpositive/bucketcontext_3.q
* /hive/trunk/ql/src/test/queries/clientpositive/bucketcontext_4.q
* /hive/trunk/ql/src/test/results/clientpositive/bucketcontext_1.q.out
* /hive/trunk/ql/src/test/results/clientpositive/bucketcontext_2.q.out
* /hive/trunk/ql/src/test/results/clientpositive/bucketcontext_3.q.out
* /hive/trunk/ql/src/test/results/clientpositive/bucketcontext_4.q.out
* /hive/trunk/ql/src/test/results/clientpositive/bucketmapjoin1.q.out
* /hive/trunk/ql/src/test/results/clientpositive/bucketmapjoin2.q.out
* /hive/trunk/ql/src/test/results/clientpositive/bucketmapjoin3.q.out
* /hive/trunk/ql/src/test/results/clientpositive/bucketmapjoin5.q.out
* /hive/trunk/ql/src/test/results/clientpositive/bucketmapjoin_negative2.q.out
* /hive/trunk/ql/src/test/results/clientpositive/stats11.q.out

                
> Stream table of SMBJoin/BucketMapJoin with two or more partitions is not 
> handled properly
> -----------------------------------------------------------------------------------------
>
>                 Key: HIVE-3218
>                 URL: https://issues.apache.org/jira/browse/HIVE-3218
>             Project: Hive
>          Issue Type: Bug
>          Components: Query Processor
>    Affects Versions: 0.10.0
>            Reporter: Navis
>            Assignee: Navis
>            Priority: Critical
>             Fix For: 0.10.0
>
>         Attachments: HIVE-3218.1.patch.txt, hive.3218.2.patch
>
>
> {noformat}
> drop table hive_test_smb_bucket1;
> drop table hive_test_smb_bucket2;
> create table hive_test_smb_bucket1 (key int, value string) partitioned by (ds 
> string) clustered by (key) sorted by (key) into 2 buckets;
> create table hive_test_smb_bucket2 (key int, value string) partitioned by (ds 
> string) clustered by (key) sorted by (key) into 2 buckets;
> set hive.enforce.bucketing = true;
> set hive.enforce.sorting = true;
> insert overwrite table hive_test_smb_bucket1 partition (ds='2010-10-14') 
> select key, value from src;
> insert overwrite table hive_test_smb_bucket1 partition (ds='2010-10-15') 
> select key, value from src;
> insert overwrite table hive_test_smb_bucket2 partition (ds='2010-10-15') 
> select key, value from src;
> set hive.optimize.bucketmapjoin = true;
> set hive.optimize.bucketmapjoin.sortedmerge = true;
> set hive.input.format = 
> org.apache.hadoop.hive.ql.io.BucketizedHiveInputFormat;
> SELECT /* + MAPJOIN(b) */ * FROM hive_test_smb_bucket1 a JOIN 
> hive_test_smb_bucket2 b ON a.key = b.key;
> {noformat}
> which make bucket join context..
> {noformat}
> Alias Bucket Output File Name Mapping:
>         
> hdfs://localhost:9000/user/hive/warehouse/hive_test_smb_bucket1/ds=2010-10-14/000000_0
>  0
>         
> hdfs://localhost:9000/user/hive/warehouse/hive_test_smb_bucket1/ds=2010-10-14/000001_0
>  1
>         
> hdfs://localhost:9000/user/hive/warehouse/hive_test_smb_bucket1/ds=2010-10-15/000000_0
>  0
>         
> hdfs://localhost:9000/user/hive/warehouse/hive_test_smb_bucket1/ds=2010-10-15/000001_0
>  1
> {noformat}
> fails with exception
> {noformat}
> java.lang.RuntimeException: Hive Runtime Error while closing operators
>       at org.apache.hadoop.hive.ql.exec.ExecMapper.close(ExecMapper.java:226)
>       at org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:57)
>       at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:391)
>       at org.apache.hadoop.mapred.MapTask.run(MapTask.java:325)
>       at org.apache.hadoop.mapred.Child$4.run(Child.java:270)
>       at java.security.AccessController.doPrivileged(Native Method)
>       at javax.security.auth.Subject.doAs(Subject.java:416)
>       at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1127)
>       at org.apache.hadoop.mapred.Child.main(Child.java:264)
> Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: Unable to rename 
> output from: 
> hdfs://localhost:9000/tmp/hive-navis/hive_2012-06-29_22-17-49_574_6018646381714861925/_task_tmp.-ext-10001/_tmp.000001_0
>  to: 
> hdfs://localhost:9000/tmp/hive-navis/hive_2012-06-29_22-17-49_574_6018646381714861925/_tmp.-ext-10001/000001_0
>       at 
> org.apache.hadoop.hive.ql.exec.FileSinkOperator$FSPaths.commit(FileSinkOperator.java:198)
>       at 
> org.apache.hadoop.hive.ql.exec.FileSinkOperator$FSPaths.access$300(FileSinkOperator.java:100)
>       at 
> org.apache.hadoop.hive.ql.exec.FileSinkOperator.closeOp(FileSinkOperator.java:717)
>       at org.apache.hadoop.hive.ql.exec.Operator.close(Operator.java:557)
>       at org.apache.hadoop.hive.ql.exec.Operator.close(Operator.java:566)
>       at org.apache.hadoop.hive.ql.exec.Operator.close(Operator.java:566)
>       at org.apache.hadoop.hive.ql.exec.Operator.close(Operator.java:566)
>       at org.apache.hadoop.hive.ql.exec.Operator.close(Operator.java:566)
>       at org.apache.hadoop.hive.ql.exec.Operator.close(Operator.java:566)
>       at org.apache.hadoop.hive.ql.exec.ExecMapper.close(ExecMapper.java:193)
>       ... 8 more
> {noformat}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

Reply via email to