[
https://issues.apache.org/jira/browse/KYLIN-5121?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17524008#comment-17524008
]
ASF subversion and git services commented on KYLIN-5121:
--------------------------------------------------------
Commit dc3fd9ab25ecd03ec53f8855f5366db8de542ce5 in kylin's branch
refs/heads/kylin-soft-affinity-local-cache from hujiahua
[ https://gitbox.apache.org/repos/asf?p=kylin.git;h=dc3fd9ab25 ]
[KYLIN-5121] Make JobMetricsUtils.collectMetrics be working again
> Make JobMetricsUtils.collectMetrics be working again
> ----------------------------------------------------
>
> Key: KYLIN-5121
> URL: https://issues.apache.org/jira/browse/KYLIN-5121
> Project: Kylin
> Issue Type: Improvement
> Reporter: hujiahua
> Priority: Major
>
> At present, the rowCount needs to be eval after the cube built every time,
> and spark `QueryExecution` metric have `numOutputRows` metric for this
> purpose. But, after patch KYLIN-4662 (Migrate from third-party Spark to
> offical Apache Spark), the util function `JobMetricsUtils.collectMetrics`
> becomes out of working. Each rowCount needs to call `Dataset.count()`, which
> wastes resources and affects the cube build time.
> Here is my solution: Get the QueryExecution object based on custom
> QueryExecutionListener, and match the corresponding QueryExecution by
> comparing the output path. (BWT, The output path of cube id is always unique)
--
This message was sent by Atlassian Jira
(v8.20.1#820001)