[
https://issues.apache.org/jira/browse/SPARK-30446?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17009830#comment-17009830
]
Thomas Graves commented on SPARK-30446:
---------------------------------------
Yeah so running on standalone if you set the spark.task.cpus=2 (or anything >
1) and you don't set executor cores it fails even though it shouldn't because
executor cores are all the cores of the worker by default:
20/01/07 09:34:02 ERROR Main: Failed to initialize Spark session.
org.apache.spark.SparkException: The number of cores per executor (=1) has to
be >= the task config: spark.task.cpus = 2 when run on spark://tomg-x299:7077.
> Accelerator aware scheduling checkResourcesPerTask doesn't cover all cases
> --------------------------------------------------------------------------
>
> Key: SPARK-30446
> URL: https://issues.apache.org/jira/browse/SPARK-30446
> Project: Spark
> Issue Type: Bug
> Components: Spark Core
> Affects Versions: 3.0.0
> Reporter: Thomas Graves
> Priority: Major
>
> with accelerator aware scheduling SparkContext.checkResourcesPerTask
> Tries to make sure that users have configured things properly and warn or
> error if not.
> It doesn't properly handle all cases like warning if cpu resources are being
> wasted. We should test this better and handle those.
> I fixed these in the stage level scheduling but not sure the timeline on
> getting that in so we may want to fix this separately as well.
--
This message was sent by Atlassian Jira
(v8.3.4#803005)
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]