Based on Jacob's suggestion, I started using --net=host which is a new option in latest version of docker. I also set SPARK_LOCAL_IP to the host's IP address and then AKKA does not use the hostname and I don't need the Spark driver's hostname to be resolvable.
Thanks guys for your help! On Tue, Jun 17, 2014 at 7:49 PM, Aaron Davidson <ilike...@gmail.com> wrote: > Yup, alright, same solution then :) > > > On Tue, Jun 17, 2014 at 7:39 PM, Mohit Jaggi <mohitja...@gmail.com> wrote: > >> I used --privileged to start the container and then unmounted /etc/hosts. >> Then I created a new /etc/hosts file >> >> >> On Tue, Jun 17, 2014 at 4:58 PM, Aaron Davidson <ilike...@gmail.com> >> wrote: >> >>> I remember having to do a similar thing in the spark docker scripts for >>> testing purposes. Were you able to modify the /etc/hosts directly? I >>> remember issues with that as docker apparently mounts it as part of its >>> read-only filesystem. >>> >>> >>> On Tue, Jun 17, 2014 at 4:36 PM, Mohit Jaggi <mohitja...@gmail.com> >>> wrote: >>> >>>> It was a DNS issue. AKKA apparently uses the hostname of the endpoints >>>> and hence they need to be resolvable. In my case the hostname of the docker >>>> container was a randomly generated string and was not resolvable. I added a >>>> workaround (entry in etc/hosts file of spark master) for now. If anyone can >>>> point to a more elegant solution, that would be awesome! >>>> >>>> >>>> On Tue, Jun 17, 2014 at 3:48 PM, Mohit Jaggi <mohitja...@gmail.com> >>>> wrote: >>>> >>>>> I am using cutting edge code from git but doing my own sbt assembly. >>>>> >>>>> >>>>> On Mon, Jun 16, 2014 at 10:28 PM, Andre Schumacher < >>>>> schum...@icsi.berkeley.edu> wrote: >>>>> >>>>>> >>>>>> Hi, >>>>>> >>>>>> are you using the amplab/spark-1.0.0 images from the global registry? >>>>>> >>>>>> Andre >>>>>> >>>>>> On 06/17/2014 01:36 AM, Mohit Jaggi wrote: >>>>>> > Hi Folks, >>>>>> > >>>>>> > I am having trouble getting spark driver running in docker. If I >>>>>> run a >>>>>> > pyspark example on my mac it works but the same example on a docker >>>>>> image >>>>>> > (Via boot2docker) fails with following logs. I am pointing the >>>>>> spark driver >>>>>> > (which is running the example) to a spark cluster (driver is not >>>>>> part of >>>>>> > the cluster). I guess this has something to do with docker's >>>>>> networking >>>>>> > stack (it may be getting NAT'd) but I am not sure why (if at all) >>>>>> the >>>>>> > spark-worker or spark-master is trying to create a new TCP >>>>>> connection to >>>>>> > the driver, instead of responding on the connection initiated by >>>>>> the driver. >>>>>> > >>>>>> > I would appreciate any help in figuring this out. >>>>>> > >>>>>> > Thanks, >>>>>> > >>>>>> > Mohit. >>>>>> > >>>>>> > --------logs-------- >>>>>> > >>>>>> > Spark Executor Command: "java" "-cp" >>>>>> > >>>>>> "::/home/ayasdi/spark/conf:/home/xxxx/spark/assembly/target/scala-2.10/spark-assembly-1.0.0-hadoop1.0.4.jar" >>>>>> > "-Xms2g" "-Xmx2g" "-Xms512M" "-Xmx512M" >>>>>> > "org.apache.spark.executor.CoarseGrainedExecutorBackend" >>>>>> > "akka.tcp://spark@fc31887475e3:43921/user/CoarseGrainedScheduler" >>>>>> "1" >>>>>> > "cobalt" "24" "akka.tcp://sparkWorker@aaaa:33952/user/Worker" >>>>>> > "app-20140616152201-0021" >>>>>> > >>>>>> > ======================================== >>>>>> > >>>>>> > >>>>>> > log4j:WARN No appenders could be found for logger >>>>>> > (org.apache.hadoop.conf.Configuration). >>>>>> > >>>>>> > log4j:WARN Please initialize the log4j system properly. >>>>>> > >>>>>> > log4j:WARN See >>>>>> http://logging.apache.org/log4j/1.2/faq.html#noconfig for >>>>>> > more info. >>>>>> > >>>>>> > 14/06/16 15:22:05 INFO SparkHadoopUtil: Using Spark's default log4j >>>>>> > profile: org/apache/spark/log4j-defaults.properties >>>>>> > >>>>>> > 14/06/16 15:22:05 INFO SecurityManager: Changing view acls to: >>>>>> ayasdi,root >>>>>> > >>>>>> > 14/06/16 15:22:05 INFO SecurityManager: SecurityManager: >>>>>> authentication >>>>>> > disabled; ui acls disabled; users with view permissions: Set(xxx, >>>>>> xxx) >>>>>> > >>>>>> > 14/06/16 15:22:05 INFO Slf4jLogger: Slf4jLogger started >>>>>> > >>>>>> > 14/06/16 15:22:05 INFO Remoting: Starting remoting >>>>>> > >>>>>> > 14/06/16 15:22:06 INFO Remoting: Remoting started; listening on >>>>>> addresses >>>>>> > :[akka.tcp://sparkExecutor@aaaa:33536] >>>>>> > >>>>>> > 14/06/16 15:22:06 INFO Remoting: Remoting now listens on addresses: >>>>>> > [akka.tcp://sparkExecutor@aaaa:33536] >>>>>> > >>>>>> > 14/06/16 15:22:06 INFO CoarseGrainedExecutorBackend: Connecting to >>>>>> driver: >>>>>> > akka.tcp://spark@fc31887475e3:43921/user/CoarseGrainedScheduler >>>>>> > >>>>>> > 14/06/16 15:22:06 INFO WorkerWatcher: Connecting to worker >>>>>> > akka.tcp://sparkWorker@aaaa:33952/user/Worker >>>>>> > >>>>>> > 14/06/16 15:22:06 WARN Remoting: Tried to associate with >>>>>> unreachable remote >>>>>> > address [akka.tcp://spark@fc31887475e3:43921]. Address is now >>>>>> gated for >>>>>> > 60000 ms, all messages to this address will be delivered to dead >>>>>> letters. >>>>>> > >>>>>> > 14/06/16 15:22:06 ERROR CoarseGrainedExecutorBackend: Driver >>>>>> Disassociated >>>>>> > [akka.tcp://sparkExecutor@aaaa:33536] -> >>>>>> [akka.tcp://spark@fc31887475e3:43921] >>>>>> > disassociated! Shutting down. >>>>>> > >>>>>> >>>>>> >>>>> >>>> >>> >> >