[ https://issues.apache.org/jira/browse/SPARK-2764?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Josh Rosen reassigned SPARK-2764: --------------------------------- Assignee: Josh Rosen > Simplify process structure of PySpark daemon / worker launching process > ----------------------------------------------------------------------- > > Key: SPARK-2764 > URL: https://issues.apache.org/jira/browse/SPARK-2764 > Project: Spark > Issue Type: Improvement > Components: PySpark > Reporter: Josh Rosen > Assignee: Josh Rosen > > PySpark's daemon-based worker factory has a very complicated process > structure that I've always found confusing. The per-java-worker daemon.py > process launches a numCores-sized pool of subprocesses, and those > subprocesses launching the actual worker processes that process data. > I think we can simplify this by having daemon.py launch the workers directly > without this extra layer of indirection. See my comments on the pull request > that introduced daemon.py: https://github.com/mesos/spark/pull/563 -- This message was sent by Atlassian JIRA (v6.2#6252)