FYI, I am getting 404 for https://hub.docker.com/apache/spark

On Mon, Feb 21, 2022 at 5:51 PM Holden Karau <hol...@pigscanfly.ca> wrote:

> Yeah I think we should still adopt that naming convention, however no one
> has taken the time submit write a script to do it yet so until we get that
> script merged I think we'll just have one build. I can try and do that for
> the next release but it would be a great 2nd issue for someone getting more
> familiar with the release tooling.
>
> On Mon, Feb 21, 2022 at 2:18 PM Mich Talebzadeh <mich.talebza...@gmail.com>
> wrote:
>
>> Ok thanks for the correction.
>>
>> The docker pull line shows as follows:
>>
>> docker pull apache/spark:v3.2.1
>>
>>
>> So this only tells me the version of Spark 3.2.1
>>
>>
>> I thought we discussed deciding on the docker naming conventions in
>> detail, and broadly agreed on what needs to be in the naming convention.
>> For example, in this thread:
>>
>>
>> Time to start publishing Spark Docker Images? - mich.talebza...@gmail.com
>> - Gmail (google.com)
>> <https://mail.google.com/mail/u/0/?hl=en-GB#search/publishing/FMfcgzGkZQSzbXWQDWfddGDNRDQfPCpg>
>>  dated
>> 22nd July 2021
>>
>>
>> Referring to that, I think the broad agreement was that the docker image
>> name should be of the form:
>>
>>
>> The name of the file provides:
>>
>>    - Built for spark or spark-py (PySpark) spark-r
>>    - Spark version: 3.1.1, 3.1.2, 3.2.1 etc.
>>    - Scala version; 2.1.2
>>    - The OS version based on JAVA: 8-jre-slim-buster, 11-jre-slim-buster
>>    meaning JAVA 8 and JAVA 11 respectively
>>
>> I believe it is a good thing and we ought to adopt that convention. For
>> example:
>>
>>
>> spark-py-3.2.1-scala_2.12-11-jre-slim-buster
>>
>>
>> HTH
>>
>>
>>
>>    view my Linkedin profile
>> <https://www.linkedin.com/in/mich-talebzadeh-ph-d-5205b2/>
>>
>>
>>  https://en.everybodywiki.com/Mich_Talebzadeh
>>
>>
>>
>> *Disclaimer:* Use it at your own risk. Any and all responsibility for
>> any loss, damage or destruction of data or any other property which may
>> arise from relying on this email's technical content is explicitly
>> disclaimed. The author will in no case be liable for any monetary damages
>> arising from such loss, damage or destruction.
>>
>>
>>
>>
>> On Mon, 21 Feb 2022 at 21:58, Holden Karau <hol...@pigscanfly.ca> wrote:
>>
>>> My bad, the correct link is:
>>>
>>> https://hub.docker.com/r/apache/spark/tags
>>>
>>> On Mon, Feb 21, 2022 at 1:17 PM Mich Talebzadeh <
>>> mich.talebza...@gmail.com> wrote:
>>>
>>>> well that docker link is not found! may be permission issue
>>>>
>>>> [image: image.png]
>>>>
>>>>
>>>>
>>>>
>>>>    view my Linkedin profile
>>>> <https://www.linkedin.com/in/mich-talebzadeh-ph-d-5205b2/>
>>>>
>>>>
>>>>  https://en.everybodywiki.com/Mich_Talebzadeh
>>>>
>>>>
>>>>
>>>> *Disclaimer:* Use it at your own risk. Any and all responsibility for
>>>> any loss, damage or destruction of data or any other property which may
>>>> arise from relying on this email's technical content is explicitly
>>>> disclaimed. The author will in no case be liable for any monetary damages
>>>> arising from such loss, damage or destruction.
>>>>
>>>>
>>>>
>>>>
>>>> On Mon, 21 Feb 2022 at 21:09, Holden Karau <hol...@pigscanfly.ca>
>>>> wrote:
>>>>
>>>>> We are happy to announce the availability of Spark 3.1.3!
>>>>>
>>>>> Spark 3.1.3 is a maintenance release containing stability fixes. This
>>>>> release is based on the branch-3.1 maintenance branch of Spark. We
>>>>> strongly
>>>>> recommend all 3.1 users to upgrade to this stable release.
>>>>>
>>>>> To download Spark 3.1.3, head over to the download page:
>>>>> https://spark.apache.org/downloads.html
>>>>>
>>>>> To view the release notes:
>>>>> https://spark.apache.org/releases/spark-release-3-1-3.html
>>>>>
>>>>> We would like to acknowledge all community members for contributing to
>>>>> this
>>>>> release. This release would not have been possible without you.
>>>>>
>>>>> *New Dockerhub magic in this release:*
>>>>>
>>>>> We've also started publishing docker containers to the Apache
>>>>> Dockerhub,
>>>>> these contain non-ASF artifacts that are subject to different license
>>>>> terms than the
>>>>> Spark release. The docker containers are built for Linux x86 and ARM64
>>>>> since that's
>>>>> what I have access to (thanks to NV for the ARM64 machines).
>>>>>
>>>>> You can get them from https://hub.docker.com/apache/spark (and
>>>>> spark-r and spark-py) :)
>>>>> (And version 3.2.1 is also now published on Dockerhub).
>>>>>
>>>>> Holden
>>>>>
>>>>> --
>>>>> Twitter: https://twitter.com/holdenkarau
>>>>> Books (Learning Spark, High Performance Spark, etc.):
>>>>> https://amzn.to/2MaRAG9  <https://amzn.to/2MaRAG9>
>>>>> YouTube Live Streams: https://www.youtube.com/user/holdenkarau
>>>>>
>>>>
>>>
>>> --
>>> Twitter: https://twitter.com/holdenkarau
>>> Books (Learning Spark, High Performance Spark, etc.):
>>> https://amzn.to/2MaRAG9  <https://amzn.to/2MaRAG9>
>>> YouTube Live Streams: https://www.youtube.com/user/holdenkarau
>>>
>>
>
> --
> Twitter: https://twitter.com/holdenkarau
> Books (Learning Spark, High Performance Spark, etc.):
> https://amzn.to/2MaRAG9  <https://amzn.to/2MaRAG9>
> YouTube Live Streams: https://www.youtube.com/user/holdenkarau
>


-- 
Regards,
Prasad Paravatha

Reply via email to