TIm, sorry I am wrong above.

The above config is from spark dispatcher container and those configuration
is not being propagated to the driver.

I will try the workaround you told and let you know how it goes.

Thanks



On Tue, Mar 15, 2016 at 4:42 PM, Pradeep Chhetri <
pradeep.chhetr...@gmail.com> wrote:

> Hello Tim,
>
> Here is my conf/spark-defaults.conf which is inside the docker image:
>
> $ cat conf/spark-defaults.conf
>
> spark.mesos.coarse: false
> spark.mesos.executor.docker.image: docker-registry/mesos-spark:master-12
> spark.mesos.mesosExecutor.cores: 0.25
> spark.mesos.executor.home: /opt/spark
> spark.mesos.uris: file:///etc/docker.tar.gz
>
> I am already setting it inside the docker image.
>
> Am I missing something ?
>
> Regards,
>
> On Tue, Mar 15, 2016 at 4:37 PM, Tim Chen <t...@mesosphere.io> wrote:
>
>> Hi Pradeep,
>>
>> Yes we still have a pending PR that will start propagating these settings
>> down to the executors, right now it's only applied on the driver. As a work
>> around you can download or set spark.mesos.executor.docker.image in the
>> spark-default.conf file in the docker image you use to launch the driver
>> and it should automatically get this setting when the driver is launched.
>>
>> Tim
>>
>> On Tue, Mar 15, 2016 at 9:26 AM, Pradeep Chhetri <
>> pradeep.chhetr...@gmail.com> wrote:
>>
>>> Hello Timothy,
>>>
>>> I am setting spark.mesos.executor.docker.image. In my case, the driver
>>> is actually started as a docker container (SparkPi in screenshot) but the
>>> tasks which are spawned by driver are not starting as containers but plain
>>> java processes. Is this expected ?
>>>
>>> Thanks
>>>
>>> On Tue, Mar 15, 2016 at 4:19 PM, Timothy Chen <t...@mesosphere.io> wrote:
>>>
>>>> You can launch the driver and executor in docker containers as well by
>>>> setting spark.mesos.executor.docker.image to the image you want to use to
>>>> launch them.
>>>>
>>>> Tim
>>>>
>>>> On Mar 15, 2016, at 8:49 AM, Radoslaw Gruchalski <ra...@gruchalski.com>
>>>> wrote:
>>>>
>>>> Pradeep,
>>>>
>>>> You can mount a spark directory as a volume. This means you have to
>>>> have spark deployed on every agent.
>>>>
>>>> Another thing you can do, place spark in hdfs, assuming that you have
>>>> hdfs available but that too will download a copy to the sandbox.
>>>>
>>>> I'd prefer the former.
>>>>
>>>> Sent from Outlook Mobile <https://aka.ms/qtex0l>
>>>>
>>>> _____________________________
>>>> From: Pradeep Chhetri <pradeep.chhetr...@gmail.com>
>>>> Sent: Tuesday, March 15, 2016 4:41 pm
>>>> Subject: Apache Spark Over Mesos
>>>> To: <user@mesos.apache.org>
>>>>
>>>>
>>>> Hello,
>>>>
>>>> I am able to run Apache Spark over Mesos. Its quite simple to run Spark
>>>> Dispatcher over marathon and ask it to run Spark Executor (I guess also can
>>>> be called as Spark Driver) as docker container.
>>>>
>>>> I have a query regarding this:
>>>>
>>>> All spark tasks are spawned directly by first downloading the spark
>>>> artifacts. I was thinking if there is some way I can start them too as
>>>> docker containers. This will save the time for downloading the spark
>>>> artifacts. I am running spark in fine-grained mode.
>>>>
>>>> I have attached a screenshot of a sample job
>>>>
>>>> <Screen Shot 2016-03-15 at 15.15.06.png>
>>>> ​
>>>> Thanks,
>>>>
>>>> --
>>>> Pradeep Chhetri
>>>>
>>>>
>>>>
>>>
>>>
>>> --
>>> Pradeep Chhetri
>>>
>>
>>
>
>
> --
> Pradeep Chhetri
>



-- 
Pradeep Chhetri

Reply via email to