[ https://issues.apache.org/jira/browse/SPARK-9559?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14651882#comment-14651882 ]
partha bishnu commented on SPARK-9559: -------------------------------------- Hi I am running some tests on spark in stand-alone mode with 3 nodes cluster. spark master is running on n-1, and slaves are on n-2 and n-3. Each machine is with 8G RAM and 4 core cpu. I am trying to test worker redundancy. I wanted to set up the cluster such a way so that there will be two worker JVM, one on each slave (n-2 and n-3) after I start up the cluster. Then one of the slave's worker JVM will launch the executor jvm to process the tasks when I submit the job with the following flags: ---total-executor-cores 1 and --executor-memory 1G (1) Job submitted successfully in client mode. n-2 had worker jvm launched a executor jvm. So now n-2 had one worker jvm and one executor jvm running and n-3 just had the worker jvm running as before. (2) I killed the worker jvm and the executor jvm on n-2 (3) I expected spark master on n-1 will then ask the worker jvm on n-3 to launch a new executor to start processing jobs but that did not happen. driver just got hung on the screen. n-2 disappeared from spark cluster as expected. n-3 just had the worker jvm running as before and no new executor was launched as expected after n-2 disappeared. > Worker redundancy/failover in spark stand-alone mode > ---------------------------------------------------- > > Key: SPARK-9559 > URL: https://issues.apache.org/jira/browse/SPARK-9559 > Project: Spark > Issue Type: Bug > Components: Spark Core > Affects Versions: 1.3.0 > Reporter: partha bishnu > -- This message was sent by Atlassian JIRA (v6.3.4#6332) --------------------------------------------------------------------- To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org