deniskuzZ commented on code in PR #5792: URL: https://github.com/apache/hive/pull/5792#discussion_r2096052636
########## iceberg/iceberg-handler/src/main/java/org/apache/iceberg/mr/hive/compaction/IcebergQueryCompactor.java: ########## @@ -96,31 +127,46 @@ public boolean run(CompactorContext context) throws IOException, HiveException, throw new HiveException(ErrorMsg.COMPACTION_NO_PARTITION); } } else { - long partitionHash = IcebergTableUtil.getPartitionHash(icebergTable, partSpec); + HiveConf.setBoolVar(conf, ConfVars.HIVE_CONVERT_JOIN, false); + conf.setBoolVar(ConfVars.HIVE_VECTORIZATION_ENABLED, false); HiveConf.setVar(conf, ConfVars.REWRITE_POLICY, RewritePolicy.PARTITION.name()); - conf.set(IcebergCompactionService.PARTITION_PATH, new Path(partSpec).toString()); + conf.set(IcebergCompactionService.PARTITION_PATH, new Path(ci.partName).toString()); - Map<String, String> partSpecMap = new LinkedHashMap<>(); - Warehouse.makeSpecFromName(partSpecMap, new Path(partSpec), null); + int specId = IcebergTableUtil.getPartitionSpecId(icebergTable, ci.partName); + String partitionPredicate = buildPartitionPredicate(ci, icebergTable); - compactionQuery = String.format("insert overwrite table %1$s select * from %1$s where %2$s=%3$d " + - "and %4$s is not null %5$s %6$s", compactTableName, VirtualColumn.PARTITION_HASH.getName(), partitionHash, - VirtualColumn.FILE_PATH.getName(), fileSizePredicate == null ? "" : "and " + fileSizePredicate, orderBy); + compactionQuery = String.format("INSERT OVERWRITE TABLE %1$s SELECT * FROM %1$s WHERE %2$s IN " + + "(SELECT FILE_PATH FROM %1$s.FILES WHERE %3$s AND SPEC_ID = %7$d) AND %6$s = %7$d %4$s %5$s", + compactTableName, VirtualColumn.FILE_PATH.getName(), partitionPredicate, + fileSizePredicate == null ? "" : "AND " + fileSizePredicate, orderBy, + VirtualColumn.PARTITION_SPEC_ID.getName(), specId); } + return compactionQuery; + } - SessionState sessionState = setupQueryCompactionSession(conf, ci, tblProperties); - String compactionTarget = "table " + HiveUtils.unparseIdentifier(compactTableName) + - (partSpec != null ? ", partition " + HiveUtils.unparseIdentifier(partSpec) : ""); + private String buildPartitionPredicate(CompactionInfo ci, Table icebergTable) { + Map<String, String> partSpecMap = new LinkedHashMap<>(); + Warehouse.makeSpecFromName(partSpecMap, new Path(ci.partName), null); Review Comment: why not use simpler method? ```` Map<String, String> partSpecMap = Warehouse.makeSpecFromName(ci.partName)) ```` -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: gitbox-unsubscr...@hive.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org --------------------------------------------------------------------- To unsubscribe, e-mail: gitbox-unsubscr...@hive.apache.org For additional commands, e-mail: gitbox-h...@hive.apache.org