Egor Pakhomov created SPARK-6680:
------------------------------------

             Summary: Be able to specifie IP for spark-shell(spark driver) 
blocker for Docker integration
                 Key: SPARK-6680
                 URL: https://issues.apache.org/jira/browse/SPARK-6680
             Project: Spark
          Issue Type: New Feature
          Components: Deploy
    Affects Versions: 1.3.0
         Environment: Docker.
            Reporter: Egor Pakhomov
            Priority: Blocker


Suppose I have 3 docker containers - spark_master, spark_worker and 
spark_shell. In docker for public IP of this container there is an alias like 
"fgsdfg454534". It only visible in this container. When spark use it for 
communication other containers receive this alias and don't know what to do 
with it. Thats why I used SPARK_LOCAL_IP for master and worker. But it doesn't 
work for spark driver(for spark shell - other types of drivers I haven't try). 
Spark driver sent everyone "fgsdfg454534" alias about itself and then nobody 
can address it. I've overcome it in https://github.com/epahomov/docker-spark, 
but it would be better if it would be solved on spark code level.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org

Reply via email to