tgravescs commented on a change in pull request #25047: [WIP][SPARK-27371][CORE] Support GPU-aware resources scheduling in Standalone URL: https://github.com/apache/spark/pull/25047#discussion_r303929005
########## File path: core/src/main/scala/org/apache/spark/internal/config/package.scala ########## @@ -35,6 +35,16 @@ package object config { private[spark] val SPARK_EXECUTOR_PREFIX = "spark.executor" private[spark] val SPARK_TASK_PREFIX = "spark.task" + private[spark] val SPARK_RESOURCES_DIR = + ConfigBuilder("spark.resources.dir") + .doc("Directory used to create spark-resources directory under it. If user does not" + + "config it explicitly, SPARK_HOME will be used by default. And since spark-resources" + Review comment: Also would we want the driver config to be different from the worker config? I would expect those to be different hosts so someone could specify it different per host, but if people use one configuration file for everything having them separate spark.worker.. and spark.driver.. might be convenient/flexible. ---------------------------------------------------------------- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services --------------------------------------------------------------------- To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org For additional commands, e-mail: reviews-h...@spark.apache.org