Take a look at
https://github.com/lresende/docker-systemml-notebook

I am not using yarn, but have spark, hdfs, zeppelin working all together.

On Wednesday, April 6, 2016, John Omernik <j...@omernik.com> wrote:

> So, I am running into this again, net=host isn't an option for me, because
> then my hadoop client conflicts with the nodes hadoop client.  So I am
> trying to work out net=bridge, however, even specifying the driver port and
> host in spark doesn't work because spark tries to listen on only the
> hostname specified in spark.driver.host, not on all interfaces, thus it
> doesn't work inside the container.  Anyone else have luck with running
> Zeppelin in a Docker container, and using Yarn?
>
> On Wed, Jun 17, 2015 at 12:20 PM, John Omernik <j...@omernik.com
> <javascript:_e(%7B%7D,'cvml','j...@omernik.com');>> wrote:
>
>> My challenge is I am running this in mesos, therefore I have a limited
>> amount of ports and I would need to know them ahead of time.  Even with
>> that ,the IP provided to the Application Master by the is internal Docker
>> IP, so that proves challenging as well.  Any other thoughts on this?
>>
>> On Wed, Jun 17, 2015 at 12:03 PM, Jong Wook Kim <jongw...@nyu.edu
>> <javascript:_e(%7B%7D,'cvml','jongw...@nyu.edu');>> wrote:
>>
>>> I just used the --net=host option to make the container just use the
>>> host’s network stack which is probably the easiest way.
>>>
>>> if you are not allowed to do that, change the container’s 
>>> /proc/sys/net/ipv4/ip_local_port_range
>>> to have a narrower port range, like 59000-59999 and forward the entire
>>> range.
>>>
>>>
>>> On Jun 18, 2015, at 1:15 AM, John Omernik <j...@omernik.com
>>> <javascript:_e(%7B%7D,'cvml','j...@omernik.com');>> wrote:
>>>
>>> So I think I ran across an issue with Zeppelin/Docker but I wanted to
>>> describe here and see what people thought.
>>>
>>> Basically I have a YARN cluster (Yarn... well Myriad on Mesos, but for
>>> pyspark, it works like Yarn).
>>>
>>> I have setup a docker container with all I need including Spark etc.
>>>
>>> When I try to run pyspark, my Application Master in Yarn tries to
>>> connect back to the driver which is inconveniently located in a docker
>>> container, and the only ports exposed are those for Zeppelin. (My yarn
>>> spark applicaiton is accepted, but I see in the logs, and it never "runs"
>>>
>>> 15/06/17 11:04:29 ERROR yarn.ApplicationMaster: Failed to connect to driver 
>>> at 172.17.0.16:59601, retrying ...
>>> 15/06/17 11:04:32 ERROR yarn.ApplicationMaster: Failed to connect to driver 
>>> at 172.17.0.16:59601, retrying .
>>>
>>>
>>> Ok, so yarn-client mode won't work (easily), however, yarn-cluster mode
>>> doesn't work either because you can't run a shell in yarn-cluster mode.
>>>
>>> So this is all just trying to run $SPARK_HOME/bin/pyspark --master
>>> yarn-client from inside of my Zeppelin Docker container (not even running
>>> Zeppelin yet)
>>>
>>> I am guessing that the issue is part of my issues I am having getting
>>> the spark interpreter running in Zeppelin because Zeppelin is doing just
>>> that, running pyspark in yarn-client mode.
>>>
>>> Ok, so how have people running Zeppelin in Docker dealt with this?  What
>>> other ideas should I look into here?
>>>
>>>
>>>
>>>
>>
>

-- 
Sent from my Mobile device

Reply via email to