I'm getting errors when I try to run my docker container in bridge networking
mode on mesos. 
Here is my spark submit script

/spark/bin/spark-submit \
 --class com.package.MySparkJob \
 --name My-Spark-Job \
 --files /path/config.cfg, ${JAR} \
 --master ${SPARK_MASTER_HOST} \
 --deploy-mode client \
 --supervise \
 --total-executor-cores ${SPARK_EXECUTOR_TOTAL_CORES} \
 --driver-cores ${SPARK_DRIVER_CORES} \
 --driver-memory ${SPARK_DRIVER_MEMORY} \
 --num-executors ${SPARK_NUM_EXECUTORS} \
 --executor-cores ${SPARK_EXECUTOR_CORES} \
 --executor-memory ${SPARK_EXECUTOR_MEMORY} \
 --driver-class-path ${JAR} \
 --conf
"spark.mesos.executor.docker.image=${SPARK_MESOS_EXECUTOR_DOCKER_IMAGE}" \
 --conf
"spark.mesos.executor.docker.volumes=${SPARK_MESOS_EXECUTOR_DOCKER_VOLUMES}"
\
 --conf "spark.mesos.uris=${SPARK_MESOS_URIS}" \
 --conf "spark.executorEnv.OBERON_DB_PASS=${OBERON_DB_PASS}" \
 --conf "spark.executorEnv.S3_SECRET_ACCESS_KEY=${S3_SECRET_ACCESS_KEY}" \
 --conf "spark.executorEnv.S3_ACCESS_KEY=${S3_ACCESS_KEY}" \
 --conf "spark.mesos.executor.home=${SPARK_HOME}" \
 --conf "spark.executorEnv.MESOS_NATIVE_JAVA_LIBRARY=${SPARK_MESOS_LIB}" \
 --conf "spark.files.overwrite=true" \
 --conf "spark.shuffle.service.enabled=false" \
 --conf "spark.dynamicAllocation.enabled=false" \
 --conf "spark.ui.port=${PORT_SPARKUI}" \
 --conf "spark.driver.host=${SPARK_PUBLIC_DNS}" \
 --conf "spark.driver.port=${PORT_SPARKDRIVER}" \
 --conf "spark.driver.blockManager.port=${PORT_SPARKBLOCKMANAGER}" \
 --conf "spark.jars=${JAR}" \
 --conf "spark.executor.extraClassPath=${JAR}" \
 ${JAR} 

Here is the error I'm seeing: 
java.net.BindException: Cannot assign requested address: Service
'sparkDriver' failed after 16 retries! Consider explicitly setting the
appropriate port for the service 'sparkDriver' (for example spark.ui.port
for SparkUI) to an available port or increasing spark.port.maxRetries.
at sun.nio.ch.Net.bind0(Native Method)
at sun.nio.ch.Net.bind(Net.java:433)
at sun.nio.ch.Net.bind(Net.java:425)
at sun.nio.ch.ServerSocketChannelImpl.bind(ServerSocketChannelImpl.java:223)
at sun.nio.ch.ServerSocketAdaptor.bind(ServerSocketAdaptor.java:74)
at
io.netty.channel.socket.nio.NioServerSocketChannel.doBind(NioServerSocketChannel.java:125)
at
io.netty.channel.AbstractChannel$AbstractUnsafe.bind(AbstractChannel.java:485)
at
io.netty.channel.DefaultChannelPipeline$HeadContext.bind(DefaultChannelPipeline.java:1089)
at
io.netty.channel.AbstractChannelHandlerContext.invokeBind(AbstractChannelHandlerContext.java:430)
at
io.netty.channel.AbstractChannelHandlerContext.bind(AbstractChannelHandlerContext.java:415)
at
io.netty.channel.DefaultChannelPipeline.bind(DefaultChannelPipeline.java:903)
at io.netty.channel.AbstractChannel.bind(AbstractChannel.java:198)
at io.netty.bootstrap.AbstractBootstrap$2.run(AbstractBootstrap.java:348)
at
io.netty.util.concurrent.SingleThreadEventExecutor.runAllTasks(SingleThreadEventExecutor.java:357)
at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:357)
at
io.netty.util.concurrent.SingleThreadEventExecutor$2.run(SingleThreadEventExecutor.java:111)
at java.lang.Thread.run(Thread.java:745)

I was trying to follow instructions here:
https://github.com/apache/spark/pull/15120
So in my Marathon json I'm defining the ports to use for the spark driver,
spark ui and block manager.

Can anyone help me get this running in bridge networking mode?



--
View this message in context: 
http://apache-spark-user-list.1001560.n3.nabble.com/Spark-on-Mesos-with-Docker-in-bridge-networking-mode-tp28397.html
Sent from the Apache Spark User List mailing list archive at Nabble.com.

---------------------------------------------------------------------
To unsubscribe e-mail: user-unsubscr...@spark.apache.org

Reply via email to