[
https://issues.apache.org/jira/browse/SPARK-6680?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14660123#comment-14660123
]
Cyril Lakech commented on SPARK-6680:
-------------------------------------
Hi,
After a while, I think I find a solution to this problem using --conf
spark.driver.host=${SPARK_LOCAL_IP}
and with export SPARK_LOCAL_IP=`awk 'NR==1 {print $1}' /etc/hosts`
cf: https://github.com/epahomov/docker-spark/pull/2
> Be able to specifie IP for spark-shell(spark driver) blocker for Docker
> integration
> -----------------------------------------------------------------------------------
>
> Key: SPARK-6680
> URL: https://issues.apache.org/jira/browse/SPARK-6680
> Project: Spark
> Issue Type: New Feature
> Components: Deploy
> Affects Versions: 1.3.0
> Environment: Docker.
> Reporter: Egor Pakhomov
> Priority: Minor
> Labels: core, deploy, docker
>
> Suppose I have 3 docker containers - spark_master, spark_worker and
> spark_shell. In docker for public IP of this container there is an alias like
> "fgsdfg454534". It only visible in this container. When spark use it for
> communication other containers receive this alias and don't know what to do
> with it. Thats why I used SPARK_LOCAL_IP for master and worker. But it
> doesn't work for spark driver(for spark shell - other types of drivers I
> haven't try). Spark driver sent everyone "fgsdfg454534" alias about itself
> and then nobody can address it. I've overcome it in
> https://github.com/epahomov/docker-spark, but it would be better if it would
> be solved on spark code level.
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]