There's a JIRA here: https://issues.apache.org/jira/browse/SPARK-11638
I haven't had time to look at it. On Thu, Feb 16, 2017 at 11:00 AM, cherryii <cherr...@adobe.com> wrote: > I'm getting errors when I try to run my docker container in bridge > networking > mode on mesos. > Here is my spark submit script > > /spark/bin/spark-submit \ > --class com.package.MySparkJob \ > --name My-Spark-Job \ > --files /path/config.cfg, ${JAR} \ > --master ${SPARK_MASTER_HOST} \ > --deploy-mode client \ > --supervise \ > --total-executor-cores ${SPARK_EXECUTOR_TOTAL_CORES} \ > --driver-cores ${SPARK_DRIVER_CORES} \ > --driver-memory ${SPARK_DRIVER_MEMORY} \ > --num-executors ${SPARK_NUM_EXECUTORS} \ > --executor-cores ${SPARK_EXECUTOR_CORES} \ > --executor-memory ${SPARK_EXECUTOR_MEMORY} \ > --driver-class-path ${JAR} \ > --conf > "spark.mesos.executor.docker.image=${SPARK_MESOS_EXECUTOR_DOCKER_IMAGE}" \ > --conf > "spark.mesos.executor.docker.volumes=${SPARK_MESOS_ > EXECUTOR_DOCKER_VOLUMES}" > \ > --conf "spark.mesos.uris=${SPARK_MESOS_URIS}" \ > --conf "spark.executorEnv.OBERON_DB_PASS=${OBERON_DB_PASS}" \ > --conf "spark.executorEnv.S3_SECRET_ACCESS_KEY=${S3_SECRET_ACCESS_KEY}" \ > --conf "spark.executorEnv.S3_ACCESS_KEY=${S3_ACCESS_KEY}" \ > --conf "spark.mesos.executor.home=${SPARK_HOME}" \ > --conf "spark.executorEnv.MESOS_NATIVE_JAVA_LIBRARY=${SPARK_MESOS_LIB}" \ > --conf "spark.files.overwrite=true" \ > --conf "spark.shuffle.service.enabled=false" \ > --conf "spark.dynamicAllocation.enabled=false" \ > --conf "spark.ui.port=${PORT_SPARKUI}" \ > --conf "spark.driver.host=${SPARK_PUBLIC_DNS}" \ > --conf "spark.driver.port=${PORT_SPARKDRIVER}" \ > --conf "spark.driver.blockManager.port=${PORT_SPARKBLOCKMANAGER}" \ > --conf "spark.jars=${JAR}" \ > --conf "spark.executor.extraClassPath=${JAR}" \ > ${JAR} > > Here is the error I'm seeing: > java.net.BindException: Cannot assign requested address: Service > 'sparkDriver' failed after 16 retries! Consider explicitly setting the > appropriate port for the service 'sparkDriver' (for example spark.ui.port > for SparkUI) to an available port or increasing spark.port.maxRetries. > at sun.nio.ch.Net.bind0(Native Method) > at sun.nio.ch.Net.bind(Net.java:433) > at sun.nio.ch.Net.bind(Net.java:425) > at sun.nio.ch.ServerSocketChannelImpl.bind(ServerSocketChannelImpl.java: > 223) > at sun.nio.ch.ServerSocketAdaptor.bind(ServerSocketAdaptor.java:74) > at > io.netty.channel.socket.nio.NioServerSocketChannel.doBind( > NioServerSocketChannel.java:125) > at > io.netty.channel.AbstractChannel$AbstractUnsafe.bind( > AbstractChannel.java:485) > at > io.netty.channel.DefaultChannelPipeline$HeadContext.bind( > DefaultChannelPipeline.java:1089) > at > io.netty.channel.AbstractChannelHandlerContext.invokeBind( > AbstractChannelHandlerContext.java:430) > at > io.netty.channel.AbstractChannelHandlerContext.bind( > AbstractChannelHandlerContext.java:415) > at > io.netty.channel.DefaultChannelPipeline.bind(DefaultChannelPipeline.java: > 903) > at io.netty.channel.AbstractChannel.bind(AbstractChannel.java:198) > at io.netty.bootstrap.AbstractBootstrap$2.run(AbstractBootstrap.java:348) > at > io.netty.util.concurrent.SingleThreadEventExecutor.runAllTasks( > SingleThreadEventExecutor.java:357) > at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:357) > at > io.netty.util.concurrent.SingleThreadEventExecutor$2. > run(SingleThreadEventExecutor.java:111) > at java.lang.Thread.run(Thread.java:745) > > I was trying to follow instructions here: > https://github.com/apache/spark/pull/15120 > So in my Marathon json I'm defining the ports to use for the spark driver, > spark ui and block manager. > > Can anyone help me get this running in bridge networking mode? > > > > -- > View this message in context: http://apache-spark-user-list. > 1001560.n3.nabble.com/Spark-on-Mesos-with-Docker-in- > bridge-networking-mode-tp28397.html > Sent from the Apache Spark User List mailing list archive at Nabble.com. > > --------------------------------------------------------------------- > To unsubscribe e-mail: user-unsubscr...@spark.apache.org > > -- Michael Gummelt Software Engineer Mesosphere