We do support running on Apache Mesos via docker images - so this
would not be restricted to k8s.
But unlike mesos support, which has other modes of running, I believe
k8s support more heavily depends on availability of docker images.


Regards,
Mridul


On Wed, Nov 29, 2017 at 8:56 AM, Sean Owen <so...@cloudera.com> wrote:
> Would it be logical to provide Docker-based distributions of other pieces of
> Spark? or is this specific to K8S?
> The problem is we wouldn't generally also provide a distribution of Spark
> for the reasons you give, because if that, then why not RPMs and so on.
>
> On Wed, Nov 29, 2017 at 10:41 AM Anirudh Ramanathan <ramanath...@google.com>
> wrote:
>>
>> In this context, I think the docker images are similar to the binaries
>> rather than an extension.
>> It's packaging the compiled distribution to save people the effort of
>> building one themselves, akin to binaries or the python package.
>>
>> For reference, this is the base dockerfile for the main image that we
>> intend to publish. It's not particularly complicated.
>> The driver and executor images are based on said base image and only
>> customize the CMD (any file/directory inclusions are extraneous and will be
>> removed).
>>
>> Is there only one way to build it? That's a bit harder to reason about.
>> The base image I'd argue is likely going to always be built that way. The
>> driver and executor images, there may be cases where people want to
>> customize it - (like putting all dependencies into it for example).
>> In those cases, as long as our images are bare bones, they can use the
>> spark-driver/spark-executor images we publish as the base, and build their
>> customization as a layer on top of it.
>>
>> I think the composability of docker images, makes this a bit different
>> from say - debian packages.
>> We can publish canonical images that serve as both - a complete image for
>> most Spark applications, as well as a stable substrate to build
>> customization upon.
>>
>> On Wed, Nov 29, 2017 at 7:38 AM, Mark Hamstra <m...@clearstorydata.com>
>> wrote:
>>>
>>> It's probably also worth considering whether there is only one,
>>> well-defined, correct way to create such an image or whether this is a
>>> reasonable avenue for customization. Part of why we don't do something like
>>> maintain and publish canonical Debian packages for Spark is because
>>> different organizations doing packaging and distribution of infrastructures
>>> or operating systems can reasonably want to do this in a custom (or
>>> non-customary) way. If there is really only one reasonable way to do a
>>> docker image, then my bias starts to tend more toward the Spark PMC taking
>>> on the responsibility to maintain and publish that image. If there is more
>>> than one way to do it and publishing a particular image is more just a
>>> convenience, then my bias tends more away from maintaining and publish it.
>>>
>>> On Wed, Nov 29, 2017 at 5:14 AM, Sean Owen <so...@cloudera.com> wrote:
>>>>
>>>> Source code is the primary release; compiled binary releases are
>>>> conveniences that are also released. A docker image sounds fairly different
>>>> though. To the extent it's the standard delivery mechanism for some 
>>>> artifact
>>>> (think: pyspark on PyPI as well) that makes sense, but is that the
>>>> situation? if it's more of an extension or alternate presentation of Spark
>>>> components, that typically wouldn't be part of a Spark release. The ones 
>>>> the
>>>> PMC takes responsibility for maintaining ought to be the core, critical
>>>> means of distribution alone.
>>>>
>>>> On Wed, Nov 29, 2017 at 2:52 AM Anirudh Ramanathan
>>>> <ramanath...@google.com.invalid> wrote:
>>>>>
>>>>> Hi all,
>>>>>
>>>>> We're all working towards the Kubernetes scheduler backend (full steam
>>>>> ahead!) that's targeted towards Spark 2.3. One of the questions that comes
>>>>> up often is docker images.
>>>>>
>>>>> While we're making available dockerfiles to allow people to create
>>>>> their own docker images from source, ideally, we'd want to publish 
>>>>> official
>>>>> docker images as part of the release process.
>>>>>
>>>>> I understand that the ASF has procedure around this, and we would want
>>>>> to get that started to help us get these artifacts published by 2.3. I'd
>>>>> love to get a discussion around this started, and the thoughts of the
>>>>> community regarding this.
>>>>>
>>>>> --
>>>>> Thanks,
>>>>> Anirudh Ramanathan
>>>
>>>
>>
>>
>>
>> --
>> Anirudh Ramanathan

---------------------------------------------------------------------
To unsubscribe e-mail: dev-unsubscr...@spark.apache.org

Reply via email to