You have multiple workers in one Spark (standalone) app? this wouldn't prevent N apps from each having a worker on a machine.
On Fri, Mar 13, 2020 at 11:51 AM Andrew Melo <andrew.m...@gmail.com> wrote: > > Hello, > > On Fri, Feb 28, 2020 at 13:21 Xingbo Jiang <jiangxb1...@gmail.com> wrote: >> >> Hi all, >> >> Based on my experience, there is no scenario that necessarily requires >> deploying multiple Workers on the same node with Standalone backend. A >> worker should book all the resources reserved to Spark on the host it is >> launched, then it can allocate those resources to one or more executors >> launched by this worker. Since each executor runs in a separated JVM, we can >> limit the memory of each executor to avoid long GC pause. >> >> The remaining concern is the local-cluster mode is implemented by launching >> multiple workers on the local host, we might need to re-implement >> LocalSparkCluster to launch only one Worker and multiple executors. It >> should be fine because local-cluster mode is only used in running Spark unit >> test cases, thus end users should not be affected by this change. >> >> Removing multiple workers on the same host support could simplify the deploy >> model of Standalone backend, and also reduce the burden to support legacy >> deploy pattern in the future feature developments. (There is an example in >> https://issues.apache.org/jira/browse/SPARK-27371 , where we designed a >> complex approach to coordinate resource requirements from different workers >> launched on the same host). >> >> The proposal is to update the document to deprecate the support of system >> environment `SPARK_WORKER_INSTANCES` in Spark 3.0, and remove the support in >> the next major version (Spark 3.1). >> >> Please kindly let me know if you have use cases relying on this feature. > > > When deploying spark on batch systems (by wrapping the standalone deployment > in scripts that can be consumed by the batch scheduler), we typically end up > with >1 worker per host. If I understand correctly, this proposal would make > our use case unsupported. > > Thanks, > Andrew > > > >> >> Thanks! >> >> Xingbo > > -- > It's dark in this basement. --------------------------------------------------------------------- To unsubscribe e-mail: dev-unsubscr...@spark.apache.org