[ 
https://issues.apache.org/jira/browse/SPARK-17125?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15426110#comment-15426110
 ] 

Apache Spark commented on SPARK-17125:
--------------------------------------

User 'zjffdu' has created a pull request for this issue:
https://github.com/apache/spark/pull/14699

> Allow to specify spark config using non-string type in SparkR
> -------------------------------------------------------------
>
>                 Key: SPARK-17125
>                 URL: https://issues.apache.org/jira/browse/SPARK-17125
>             Project: Spark
>          Issue Type: Improvement
>          Components: SparkR
>    Affects Versions: 2.0.0
>            Reporter: Jeff Zhang
>            Priority: Minor
>
> I try to specify spark conf spark.executor.instances as following in SparkR, 
> but fails. Since list supports any kind of data type, it is natural for user 
> to specify int type for configuration like spark.exeucotr.instances. 
> {code}
> sparkR.session(master="yarn-client", sparkConfig = 
> list(spark.executor.instances=1))
> {code}
> {noformat}
> Error in invokeJava(isStatic = TRUE, className, methodName, ...) : 
>   java.lang.IllegalArgumentException: spark.executor.instances should be int, 
> but was 1.0
>       at 
> org.apache.spark.internal.config.ConfigHelpers$.toNumber(ConfigBuilder.scala:31)
>       at 
> org.apache.spark.internal.config.ConfigBuilder$$anonfun$intConf$1.apply(ConfigBuilder.scala:178)
>       at 
> org.apache.spark.internal.config.ConfigBuilder$$anonfun$intConf$1.apply(ConfigBuilder.scala:178)
>       at scala.Option.map(Option.scala:146)
>       at 
> org.apache.spark.internal.config.OptionalConfigEntry.readFrom(ConfigEntry.scala:150)
>       at 
> org.apache.spark.internal.config.OptionalConfigEntry.readFrom(ConfigEntry.scala:138)
>       at org.apache.spark.SparkConf.get(SparkConf.scala:251)
>       at 
> org.apache.spark.deploy.yarn.YarnSparkHadoopUtil$.getInitialTargetExecutorNumber(YarnSparkHadoopUtil.scala:313)
>       at 
> org.apache.spark.scheduler.cluster.YarnClientSchedulerBackend.start(YarnClientSchedulerBackend.scala:54)
>       at 
> org.apache.spark.scheduler.TaskSchedulerImpl.start(TaskSchedulerImpl.scala:154)
>       at org.apache.spark.SparkContext
> {noformat}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to