[ 
https://issues.apache.org/jira/browse/SPARK-6680?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15476947#comment-15476947
 ] 

YSMAL Vincent commented on SPARK-6680:
--------------------------------------

HI, using docker you can get rid of this alias on hostname, by using the 
{code}--hostname spark-master{code} option in docker container.


> Be able to specifie IP for spark-shell(spark driver) blocker for Docker 
> integration
> -----------------------------------------------------------------------------------
>
>                 Key: SPARK-6680
>                 URL: https://issues.apache.org/jira/browse/SPARK-6680
>             Project: Spark
>          Issue Type: New Feature
>          Components: Deploy
>    Affects Versions: 1.3.0
>         Environment: Docker.
>            Reporter: Egor Pakhomov
>            Priority: Minor
>              Labels: core, deploy, docker
>
> Suppose I have 3 docker containers - spark_master, spark_worker and 
> spark_shell. In docker for public IP of this container there is an alias like 
> "fgsdfg454534". It only visible in this container. When spark use it for 
> communication other containers receive this alias and don't know what to do 
> with it. Thats why I used SPARK_LOCAL_IP for master and worker. But it 
> doesn't work for spark driver(for spark shell - other types of drivers I 
> haven't try). Spark driver sent everyone "fgsdfg454534" alias about itself 
> and then nobody can address it. I've overcome it in 
> https://github.com/epahomov/docker-spark, but it would be better if it would 
> be solved on spark code level.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org

Reply via email to