[ 
https://issues.apache.org/jira/browse/SPARK-2641?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kanwaljit Singh updated SPARK-2641:
-----------------------------------

    Description: 
When running spark-submit in Yarn cluster mode, we provide properties file 
using --properties-file option.

spark.executor.instances=5
spark.executor.memory=2120m
spark.executor.cores=3

The running job picks up the cores and memory, but not the correct instances.

I think the issue is here in org.apache.spark.deploy.SparkSubmitArguments:

// Use properties file as fallback for values which have a direct analog to
    // arguments in this script.
    master = 
Option(master).getOrElse(defaultProperties.get("spark.master").orNull)
    executorMemory = Option(executorMemory)
      .getOrElse(defaultProperties.get("spark.executor.memory").orNull)
    executorCores = Option(executorCores)
      .getOrElse(defaultProperties.get("spark.executor.cores").orNull)
    totalExecutorCores = Option(totalExecutorCores)
      .getOrElse(defaultProperties.get("spark.cores.max").orNull)
    name = 
Option(name).getOrElse(defaultProperties.get("spark.app.name").orNull)
    jars = Option(jars).getOrElse(defaultProperties.get("spark.jars").orNull)

Along with these defaults, we should also set default for instances:

numExecutors=Option(numExecutors).getOrElse(defaultProperties.get("spark.executor.instances").orNull)

> Spark submit doesn't pick up executor instances from properties file
> --------------------------------------------------------------------
>
>                 Key: SPARK-2641
>                 URL: https://issues.apache.org/jira/browse/SPARK-2641
>             Project: Spark
>          Issue Type: Bug
>          Components: Spark Core
>    Affects Versions: 1.0.0
>            Reporter: Kanwaljit Singh
>
> When running spark-submit in Yarn cluster mode, we provide properties file 
> using --properties-file option.
> spark.executor.instances=5
> spark.executor.memory=2120m
> spark.executor.cores=3
> The running job picks up the cores and memory, but not the correct instances.
> I think the issue is here in org.apache.spark.deploy.SparkSubmitArguments:
> // Use properties file as fallback for values which have a direct analog to
>     // arguments in this script.
>     master = 
> Option(master).getOrElse(defaultProperties.get("spark.master").orNull)
>     executorMemory = Option(executorMemory)
>       .getOrElse(defaultProperties.get("spark.executor.memory").orNull)
>     executorCores = Option(executorCores)
>       .getOrElse(defaultProperties.get("spark.executor.cores").orNull)
>     totalExecutorCores = Option(totalExecutorCores)
>       .getOrElse(defaultProperties.get("spark.cores.max").orNull)
>     name = 
> Option(name).getOrElse(defaultProperties.get("spark.app.name").orNull)
>     jars = Option(jars).getOrElse(defaultProperties.get("spark.jars").orNull)
> Along with these defaults, we should also set default for instances:
> numExecutors=Option(numExecutors).getOrElse(defaultProperties.get("spark.executor.instances").orNull)



--
This message was sent by Atlassian JIRA
(v6.2#6252)

Reply via email to