Yup, alright, same solution then :)

On Tue, Jun 17, 2014 at 7:39 PM, Mohit Jaggi <mohitja...@gmail.com> wrote:

> I used --privileged to start the container and then unmounted /etc/hosts.
> Then I created a new /etc/hosts file
>
>
> On Tue, Jun 17, 2014 at 4:58 PM, Aaron Davidson <ilike...@gmail.com>
> wrote:
>
>> I remember having to do a similar thing in the spark docker scripts for
>> testing purposes. Were you able to modify the /etc/hosts directly? I
>> remember issues with that as docker apparently mounts it as part of its
>> read-only filesystem.
>>
>>
>> On Tue, Jun 17, 2014 at 4:36 PM, Mohit Jaggi <mohitja...@gmail.com>
>> wrote:
>>
>>> It was a DNS issue. AKKA apparently uses the hostname of the endpoints
>>> and hence they need to be resolvable. In my case the hostname of the docker
>>> container was a randomly generated string and was not resolvable. I added a
>>> workaround (entry in etc/hosts file of spark master) for now. If anyone can
>>> point to a more elegant solution, that would be awesome!
>>>
>>>
>>> On Tue, Jun 17, 2014 at 3:48 PM, Mohit Jaggi <mohitja...@gmail.com>
>>> wrote:
>>>
>>>> I am using cutting edge code from git but doing my own sbt assembly.
>>>>
>>>>
>>>> On Mon, Jun 16, 2014 at 10:28 PM, Andre Schumacher <
>>>> schum...@icsi.berkeley.edu> wrote:
>>>>
>>>>>
>>>>> Hi,
>>>>>
>>>>> are you using the amplab/spark-1.0.0 images from the global registry?
>>>>>
>>>>> Andre
>>>>>
>>>>> On 06/17/2014 01:36 AM, Mohit Jaggi wrote:
>>>>> > Hi Folks,
>>>>> >
>>>>> > I am having trouble getting spark driver running in docker. If I run
>>>>> a
>>>>> > pyspark example on my mac it works but the same example on a docker
>>>>> image
>>>>> > (Via boot2docker) fails with following logs. I am pointing the spark
>>>>> driver
>>>>> > (which is running the example) to a spark cluster (driver is not
>>>>> part of
>>>>> > the cluster). I guess this has something to do with docker's
>>>>> networking
>>>>> > stack (it may be getting NAT'd) but I am not sure why (if at all) the
>>>>> > spark-worker or spark-master is trying to create a new TCP
>>>>> connection to
>>>>> > the driver, instead of responding on the connection initiated by the
>>>>> driver.
>>>>> >
>>>>> > I would appreciate any help in figuring this out.
>>>>> >
>>>>> > Thanks,
>>>>> >
>>>>> > Mohit.
>>>>> >
>>>>> > --------logs--------
>>>>> >
>>>>> > Spark Executor Command: "java" "-cp"
>>>>> >
>>>>> "::/home/ayasdi/spark/conf:/home/xxxx/spark/assembly/target/scala-2.10/spark-assembly-1.0.0-hadoop1.0.4.jar"
>>>>> > "-Xms2g" "-Xmx2g" "-Xms512M" "-Xmx512M"
>>>>> > "org.apache.spark.executor.CoarseGrainedExecutorBackend"
>>>>> > "akka.tcp://spark@fc31887475e3:43921/user/CoarseGrainedScheduler"
>>>>> "1"
>>>>> > "cobalt" "24" "akka.tcp://sparkWorker@aaaa:33952/user/Worker"
>>>>> > "app-20140616152201-0021"
>>>>> >
>>>>> > ========================================
>>>>> >
>>>>> >
>>>>> > log4j:WARN No appenders could be found for logger
>>>>> > (org.apache.hadoop.conf.Configuration).
>>>>> >
>>>>> > log4j:WARN Please initialize the log4j system properly.
>>>>> >
>>>>> > log4j:WARN See http://logging.apache.org/log4j/1.2/faq.html#noconfig
>>>>> for
>>>>> > more info.
>>>>> >
>>>>> > 14/06/16 15:22:05 INFO SparkHadoopUtil: Using Spark's default log4j
>>>>> > profile: org/apache/spark/log4j-defaults.properties
>>>>> >
>>>>> > 14/06/16 15:22:05 INFO SecurityManager: Changing view acls to:
>>>>> ayasdi,root
>>>>> >
>>>>> > 14/06/16 15:22:05 INFO SecurityManager: SecurityManager:
>>>>> authentication
>>>>> > disabled; ui acls disabled; users with view permissions: Set(xxx,
>>>>> xxx)
>>>>> >
>>>>> > 14/06/16 15:22:05 INFO Slf4jLogger: Slf4jLogger started
>>>>> >
>>>>> > 14/06/16 15:22:05 INFO Remoting: Starting remoting
>>>>> >
>>>>> > 14/06/16 15:22:06 INFO Remoting: Remoting started; listening on
>>>>> addresses
>>>>> > :[akka.tcp://sparkExecutor@aaaa:33536]
>>>>> >
>>>>> > 14/06/16 15:22:06 INFO Remoting: Remoting now listens on addresses:
>>>>> > [akka.tcp://sparkExecutor@aaaa:33536]
>>>>> >
>>>>> > 14/06/16 15:22:06 INFO CoarseGrainedExecutorBackend: Connecting to
>>>>> driver:
>>>>> > akka.tcp://spark@fc31887475e3:43921/user/CoarseGrainedScheduler
>>>>> >
>>>>> > 14/06/16 15:22:06 INFO WorkerWatcher: Connecting to worker
>>>>> > akka.tcp://sparkWorker@aaaa:33952/user/Worker
>>>>> >
>>>>> > 14/06/16 15:22:06 WARN Remoting: Tried to associate with unreachable
>>>>> remote
>>>>> > address [akka.tcp://spark@fc31887475e3:43921]. Address is now gated
>>>>> for
>>>>> > 60000 ms, all messages to this address will be delivered to dead
>>>>> letters.
>>>>> >
>>>>> > 14/06/16 15:22:06 ERROR CoarseGrainedExecutorBackend: Driver
>>>>> Disassociated
>>>>> > [akka.tcp://sparkExecutor@aaaa:33536] ->
>>>>> [akka.tcp://spark@fc31887475e3:43921]
>>>>> > disassociated! Shutting down.
>>>>> >
>>>>>
>>>>>
>>>>
>>>
>>
>

Reply via email to