difin commented on code in PR #5792:
URL: https://github.com/apache/hive/pull/5792#discussion_r2087292173
##########
iceberg/iceberg-handler/src/main/java/org/apache/iceberg/mr/hive/compaction/IcebergQueryCompactor.java:
##########
@@ -96,16 +108,39 @@ public boolean run(CompactorContext context) throws
IOException, HiveException,
throw new HiveException(ErrorMsg.COMPACTION_NO_PARTITION);
}
} else {
- long partitionHash = IcebergTableUtil.getPartitionHash(icebergTable,
partSpec);
+ Pair<Integer, StructProjection> partSpecPair =
+ IcebergTableUtil.getPartitionStructWithSpecId(icebergTable,
partSpec);
+ int specId = partSpecPair.getKey();
+ StructProjection partition = partSpecPair.getValue();
+
+ HiveConf.setBoolVar(conf, ConfVars.HIVE_CONVERT_JOIN, false);
Review Comment:
When enabled, compaction query is failing with the following exception in
`convertJoinOpMapJoinOp`:
```
INSERT OVERWRITE TABLE default.ice_orc SELECT * FROM default.ice_orc
WHERE FILE__PATH IN (SELECT FILE_PATH FROM default.ice_orc.FILES WHERE
partition.event_src_trunc='AAA' AND SPEC_ID=0) AND PARTITION__SPEC__ID = 0
INFO : Compiling
command(queryId=hive_20250507190138_8a53bfd0-9ed8-4951-9aa3-548e57b2ba93):
INSERT OVERWRITE TABLE default.ice_orc SELECT d.* FROM default.ice_orc d,
default.ice_orc.files f
WHERE d.FILE__PATH = f.FILE_PATH and f.partition.event_src_trunc='AAA' AND
f.SPEC_ID=0 AND d.PARTITION__SPEC__ID = 0
INFO : No Stats for default@ice_orc, Columns: event_id, event_src,
event_time
INFO : No Stats for default@ice_orc, Columns: file_path, partition, spec_id
INFO : No Stats for default@ice_orc, Columns: partition
ERROR : FAILED: NullPointerException null
java.lang.NullPointerException
at
org.apache.hadoop.hive.ql.plan.ExprNodeDescUtils.indexOf(ExprNodeDescUtils.java:77)
at
org.apache.hadoop.hive.ql.plan.ExprNodeDescUtils.indexOf(ExprNodeDescUtils.java:72)
at
org.apache.hadoop.hive.ql.optimizer.MapJoinProcessor.getMapJoinDesc(MapJoinProcessor.java:1311)
at
org.apache.hadoop.hive.ql.optimizer.MapJoinProcessor.convertJoinOpMapJoinOp(MapJoinProcessor.java:584)
at
org.apache.hadoop.hive.ql.optimizer.ConvertJoinMapJoin.convertJoinMapJoin(ConvertJoinMapJoin.java:1348)
at
org.apache.hadoop.hive.ql.optimizer.ConvertJoinMapJoin.process(ConvertJoinMapJoin.java:208)
at
org.apache.hadoop.hive.ql.lib.DefaultRuleDispatcher.dispatch(DefaultRuleDispatcher.java:90)
at
org.apache.hadoop.hive.ql.lib.DefaultGraphWalker.dispatchAndReturn(DefaultGraphWalker.java:105)
at
org.apache.hadoop.hive.ql.lib.DefaultGraphWalker.dispatch(DefaultGraphWalker.java:89)
at
org.apache.hadoop.hive.ql.lib.ForwardWalker.walk(ForwardWalker.java:74)
at
org.apache.hadoop.hive.ql.lib.DefaultGraphWalker.startWalking(DefaultGraphWalker.java:120)
at
org.apache.hadoop.hive.ql.parse.TezCompiler.runStatsDependentOptimizations(TezCompiler.java:485)
at
org.apache.hadoop.hive.ql.parse.TezCompiler.optimizeOperatorPlan(TezCompiler.java:218)
at
org.apache.hadoop.hive.ql.parse.TaskCompiler.compile(TaskCompiler.java:178)
at
org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.analyzeInternal(SemanticAnalyzer.java:13763)
at
org.apache.hadoop.hive.ql.parse.CalcitePlanner.analyzeInternal(CalcitePlanner.java:489)
at
org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer.analyze(BaseSemanticAnalyzer.java:327)
at org.apache.hadoop.hive.ql.Compiler.analyze(Compiler.java:227)
at org.apache.hadoop.hive.ql.Compiler.compile(Compiler.java:108)
at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:203)
at org.apache.hadoop.hive.ql.Driver.compileInternal(Driver.java:651)
at org.apache.hadoop.hive.ql.Driver.compileAndRespond(Driver.java:596)
at org.apache.hadoop.hive.ql.Driver.compileAndRespond(Driver.java:590)
at
org.apache.hadoop.hive.ql.reexec.ReExecDriver.compileAndRespond(ReExecDriver.java:126)
at
org.apache.hive.service.cli.operation.SQLOperation.prepare(SQLOperation.java:209)
at
org.apache.hive.service.cli.operation.SQLOperation$BackgroundWork$1.run(SQLOperation.java:340)
at java.base/java.security.AccessController.doPrivileged(Native Method)
at java.base/javax.security.auth.Subject.doAs(Subject.java:423)
at
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1899)
at
org.apache.hive.service.cli.operation.SQLOperation$BackgroundWork.run(SQLOperation.java:362)
at
java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515)
at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)
at
java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515)
at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)
at
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)
at
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)
at java.base/java.lang.Thread.run(Thread.java:829)
```
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
To unsubscribe, e-mail: [email protected]
For queries about this service, please contact Infrastructure at:
[email protected]
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]