[
https://issues.apache.org/jira/browse/HIVE-8029?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Brock Noland updated HIVE-8029:
-------------------------------
Resolution: Fixed
Fix Version/s: spark-branch
Status: Resolved (was: Patch Available)
Thank you Chengxiang! I have committed this to spark. Note in the commit I
actually said Rui since I was just reviewing HIVE-8017. I apologize for this
mistake, but since the JIRA is assigned to you, you will still get the
appropriate accreditation for the patch.
> Remove reducers number configure in SparkTask [Spark Branch]
> ------------------------------------------------------------
>
> Key: HIVE-8029
> URL: https://issues.apache.org/jira/browse/HIVE-8029
> Project: Hive
> Issue Type: Improvement
> Components: Spark
> Reporter: Chengxiang Li
> Assignee: Chengxiang Li
> Labels: Spark-M4
> Fix For: spark-branch
>
> Attachments: HIVE-8029.1-spark.patch
>
>
> We do not need duplicated logic to configure reducers number in SparkTask, as
> SetSparkReduceParallelism would always set reducers number in compiler phase.
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)