[
https://issues.apache.org/jira/browse/HIVE-22369?focusedWorklogId=346092&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-346092
]
ASF GitHub Bot logged work on HIVE-22369:
-----------------------------------------
Author: ASF GitHub Bot
Created on: 19/Nov/19 17:02
Start Date: 19/Nov/19 17:02
Worklog Time Spent: 10m
Work Description: miklosgergely commented on pull request #845:
HIVE-22369 Handle HiveTableFunctionScan at return path
URL: https://github.com/apache/hive/pull/845#discussion_r348049562
##########
File path:
ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/translator/HiveOpConverter.java
##########
@@ -186,12 +193,67 @@ OpAttr dispatch(RelNode rn) throws SemanticException {
return visit((HiveSortExchange) rn);
} else if (rn instanceof HiveAggregate) {
return visit((HiveAggregate) rn);
+ } else if (rn instanceof HiveTableFunctionScan) {
+ return visit((HiveTableFunctionScan) rn);
}
LOG.error(rn.getClass().getCanonicalName() + "operator translation not
supported"
+ " yet in return path.");
return null;
}
+ private OpAttr visit(HiveTableFunctionScan scanRel) throws SemanticException
{
+ if (LOG.isDebugEnabled()) {
+ LOG.debug("Translating operator rel#" + scanRel.getId() + ":"
+ + scanRel.getRelTypeName() + " with row type: [" +
scanRel.getRowType() + "]");
+ }
+
+ RexCall call = (RexCall)scanRel.getCall();
+
+ String functionName = call.getOperator().getName();
+ FunctionInfo fi = FunctionRegistry.getFunctionInfo(functionName);
+ GenericUDTF genericUDTF = fi.getGenericUDTF();
+
+ RowResolver rowResolver = new RowResolver();
+ List<String> fieldNames = new
ArrayList<>(scanRel.getRowType().getFieldNames());
+ List<String> exprNames = new ArrayList<>(fieldNames);
+ List<ExprNodeDesc> exprCols = new ArrayList<>();
+ Map<String, ExprNodeDesc> colExprMap = new HashMap<>();
+ for (int pos = 0; pos < call.getOperands().size(); pos++) {
+ ExprNodeConverter converter = new
ExprNodeConverter(SemanticAnalyzer.DUMMY_TABLE, fieldNames.get(pos),
+ scanRel.getRowType(), scanRel.getRowType(),
((HiveTableScan)scanRel.getInput(0)).getPartOrVirtualCols(),
+ scanRel.getCluster().getTypeFactory(), true);
+ ExprNodeDesc exprCol = call.getOperands().get(pos).accept(converter);
+ colExprMap.put(exprNames.get(pos), exprCol);
+ exprCols.add(exprCol);
+
+ ColumnInfo columnInfo = new ColumnInfo(fieldNames.get(pos),
exprCol.getWritableObjectInspector(), null, false);
+ rowResolver.put(columnInfo.getTabAlias(), columnInfo.getAlias(),
columnInfo);
+ }
+
+ QB qb = new QB(semanticAnalyzer.getQB().getId(), nextAlias(), true);
+ qb.getMetaData().setSrcForAlias(SemanticAnalyzer.DUMMY_TABLE,
semanticAnalyzer.getDummyTable());
+ TableScanOperator op = (TableScanOperator)
semanticAnalyzer.genTablePlan(SemanticAnalyzer.DUMMY_TABLE, qb);
+ op.getConf().setRowLimit(1);
+ qb.addAlias(SemanticAnalyzer.DUMMY_TABLE);
+ qb.setTabAlias(SemanticAnalyzer.DUMMY_TABLE, SemanticAnalyzer.DUMMY_TABLE);
+
+ Operator<?> output = OperatorFactory.getAndMakeChild(new
SelectDesc(exprCols, fieldNames, false),
+ new RowSchema(rowResolver.getRowSchema()), op);
+ output.setColumnExprMap(colExprMap);
+ semanticAnalyzer.putOpInsertMap(output, rowResolver);
+
+ Operator<?> funcOp = semanticAnalyzer.genUDTFPlan(genericUDTF, null,
fieldNames, qb, output, false);
Review comment:
Moving `genUDTFPlan` to `HiveOpConverter` would mean that it should be
copied from the `SemanticAnalyzer`, and having a duplication before the old one
get's removed? It's ok for me, but we should also consider that
`HiveOpConverter` is already 1300 lines long, and it wouldn't be wise to create
another monster class. So I'm ok with moving it here, but then soon we should
have a jira for creating a better design here too, and cutting
`HiveOpConverter` to pieces.
----------------------------------------------------------------
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
For queries about this service, please contact Infrastructure at:
[email protected]
Issue Time Tracking
-------------------
Worklog Id: (was: 346092)
Time Spent: 40m (was: 0.5h)
> Handle HiveTableFunctionScan at return path
> -------------------------------------------
>
> Key: HIVE-22369
> URL: https://issues.apache.org/jira/browse/HIVE-22369
> Project: Hive
> Issue Type: Sub-task
> Components: Hive
> Reporter: Miklos Gergely
> Assignee: Miklos Gergely
> Priority: Major
> Labels: pull-request-available
> Fix For: 4.0.0
>
> Attachments: HIVE-22369.01.patch
>
> Time Spent: 40m
> Remaining Estimate: 0h
>
> The
> [optimizedOptiqPlan|https://github.com/apache/hive/blob/5c91d324f22c2ae47e234e76a9bc5ee1a71e6a70/ql/src/java/org/apache/hadoop/hive/ql/parse/CalcitePlanner.java#L1573]
> at CalcitePlanner.getOptimizedHiveOPDag is ultimately generated by
> CalcitePlanner.internalGenSelectLogicalPlan, which may either provide a
> [HiveProject|https://github.com/apache/hive/blob/5c91d324f22c2ae47e234e76a9bc5ee1a71e6a70/ql/src/java/org/apache/hadoop/hive/ql/parse/CalcitePlanner.java#L4831]
> or a
> [HiveTableFunctionScan|https://github.com/apache/hive/blob/5c91d324f22c2ae47e234e76a9bc5ee1a71e6a70/ql/src/java/org/apache/hadoop/hive/ql/parse/CalcitePlanner.java#L4776].
> When HiveCalciteUtil.getTopLevelSelect is invoked on this it is looking for
> a
> [HiveProject|https://github.com/apache/hive/blob/5c91d324f22c2ae47e234e76a9bc5ee1a71e6a70/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/HiveCalciteUtil.java#L633]
> node in the tree, which if won't find in case of a HiveTableFunctionScan was
> returned. This is why TestNewGetSplitsFormat is failing with return path.
--
This message was sent by Atlassian Jira
(v8.3.4#803005)