When you say driver running on mesos can you explain how are you doing that...??
> On Mar 10, 2016, at 4:44 PM, Eran Chinthaka Withana > <eran.chinth...@gmail.com> wrote: > > Yanling I'm already running the driver on mesos (through docker). FYI, I'm > running this on cluster mode with MesosClusterDispatcher. > > Mac (client) ----> MesosClusterDispatcher ----> Driver running on Mesos --> > Workers running on Mesos > > My next step is to run MesosClusterDispatcher in mesos through marathon. > > Thanks, > Eran Chinthaka Withana > >> On Thu, Mar 10, 2016 at 11:11 AM, yanlin wang <yanl...@me.com> wrote: >> How you guys make driver docker within container to be reachable from spark >> worker ? >> >> Would you share your driver docker? i am trying to put only driver in docker >> and spark running with yarn outside of container and i don’t want to use >> —net=host >> >> Thx >> Yanlin >> >>> On Mar 10, 2016, at 11:06 AM, Guillaume Eynard Bontemps >>> <g.eynard.bonte...@gmail.com> wrote: >>> >>> Glad to hear it. Thanks all for sharing your solutions. >>> >>> >>> Le jeu. 10 mars 2016 19:19, Eran Chinthaka Withana >>> <eran.chinth...@gmail.com> a écrit : >>>> Phew, it worked. All I had to do was to add export >>>> SPARK_JAVA_OPTS="-Dspark.mesos.executor.docker.image=echinthaka/mesos-spark:0.23.1-1.6.0-2.6" >>>> before calling spark-submit. Guillaume, thanks for the pointer. >>>> >>>> Timothy, thanks for looking into this. Looking forward to see a fix soon. >>>> >>>> Thanks, >>>> Eran Chinthaka Withana >>>> >>>>> On Thu, Mar 10, 2016 at 10:10 AM, Tim Chen <t...@mesosphere.io> wrote: >>>>> Hi Eran, >>>>> >>>>> I need to investigate but perhaps that's true, we're using >>>>> SPARK_JAVA_OPTS to pass all the options and not --conf. >>>>> >>>>> I'll take a look at the bug, but if you can try the workaround and see if >>>>> that fixes your problem. >>>>> >>>>> Tim >>>>> >>>>>> On Thu, Mar 10, 2016 at 10:08 AM, Eran Chinthaka Withana >>>>>> <eran.chinth...@gmail.com> wrote: >>>>>> Hi Timothy >>>>>> >>>>>>> What version of spark are you guys running? >>>>>> >>>>>> I'm using Spark 1.6.0. You can see the Dockerfile I used here: >>>>>> https://github.com/echinthaka/spark-mesos-docker/blob/master/docker/mesos-spark/Dockerfile >>>>>> >>>>>> >>>>>>> And also did you set the working dir in your image to be spark home? >>>>>> >>>>>> Yes I did. You can see it here: https://goo.gl/8PxtV8 >>>>>> >>>>>> Can it be because of this: >>>>>> https://issues.apache.org/jira/browse/SPARK-13258 as Guillaume pointed >>>>>> out above? As you can see, I'm passing in the docker image URI through >>>>>> spark-submit (--conf >>>>>> spark.mesos.executor.docker.image=echinthaka/mesos-spark:0.23.1-1.6.0-2.6) >>>>>> >>>>>> Thanks, >>>>>> Eran >