At least we should upgrade hadoop to the latest version
https://hadoop.apache.org/release/2.10.2.html

Are there some spesial reasons why we have a hadoop version that is 7
years old?

ons. 14. sep. 2022, 20:25 skrev Dongjoon Hyun <dongjoon.h...@gmail.com>:

> Ya, +1 for Sean's comment.
>
> In addition, all Apache Spark's Maven artifacts are depending on Hadoop
> 3.3.x already.
>
>
> https://mvnrepository.com/artifact/org.apache.spark/spark-core_2.12/3.3.0
>
> https://mvnrepository.com/artifact/org.apache.spark/spark-core_2.13/3.3.0
>
> Apache Spark has been moving away from Hadoop 2 due to many many reasons.
>
> Dongjoon.
>
>
> On Wed, Sep 14, 2022 at 10:54 AM Sean Owen <sro...@gmail.com> wrote:
>
>> Yeah we're not going to make convenience binaries for all possible
>> combinations. It's a pretty good assumption that anyone moving to later
>> Scala versions is also off old Hadoop versions.
>> You can of course build the combo you like.
>>
>> On Wed, Sep 14, 2022 at 11:26 AM Denis Bolshakov <
>> bolshakov.de...@gmail.com> wrote:
>>
>>> Unfortunately it's for hadoop 3 only.
>>>
>>> ср, 14 сент. 2022 г., 19:04 Dongjoon Hyun <dongjoon.h...@gmail.com>:
>>>
>>>> Hi, Denis.
>>>>
>>>> Apache Spark community already provides both Scala 2.12 and 2.13
>>>> pre-built distributions.
>>>> Please check the distribution site and Apache Spark download page.
>>>>
>>>> https://dlcdn.apache.org/spark/spark-3.3.0/
>>>>
>>>>     spark-3.3.0-bin-hadoop3-scala2.13.tgz
>>>>     spark-3.3.0-bin-hadoop3.tgz
>>>>
>>>> [image: Screenshot 2022-09-14 at 9.03.27 AM.png]
>>>>
>>>> Dongjoon.
>>>>
>>>> On Wed, Sep 14, 2022 at 12:31 AM Denis Bolshakov <
>>>> bolshakov.de...@gmail.com> wrote:
>>>>
>>>>> Hello,
>>>>>
>>>>> It would be great if it's possible to provide a spark distro for both
>>>>> scala 2.12 and scala 2.13.
>>>>>
>>>>> It will encourage spark users to switch to scala 2.13.
>>>>>
>>>>> I know that spark jar artifacts available for both scala versions, but
>>>>> it does not make sense to migrate to scala 2.13 while there is no spark
>>>>> distro for this version.
>>>>>
>>>>> Kind regards,
>>>>> Denis
>>>>>
>>>>> On Tue, 13 Sept 2022 at 17:38, Yuming Wang <wgy...@gmail.com> wrote:
>>>>>
>>>>>> Thank you all.
>>>>>>
>>>>>> I will be preparing 3.3.1 RC1 soon.
>>>>>>
>>>>>> On Tue, Sep 13, 2022 at 12:09 PM John Zhuge <jzh...@apache.org>
>>>>>> wrote:
>>>>>>
>>>>>>> +1
>>>>>>>
>>>>>>> On Mon, Sep 12, 2022 at 9:08 PM Yang,Jie(INF) <yangji...@baidu.com>
>>>>>>> wrote:
>>>>>>>
>>>>>>>> +1
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>> Thanks Yuming ~
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>> *发件人**: *Hyukjin Kwon <gurwls...@gmail.com>
>>>>>>>> *日期**: *2022年9月13日 星期二 08:19
>>>>>>>> *收件人**: *Gengliang Wang <ltn...@gmail.com>
>>>>>>>> *抄送**: *"L. C. Hsieh" <vii...@gmail.com>, Dongjoon Hyun <
>>>>>>>> dongjoon.h...@gmail.com>, Yuming Wang <wgy...@gmail.com>, dev <
>>>>>>>> dev@spark.apache.org>
>>>>>>>> *主题**: *Re: Time for Spark 3.3.1 release?
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>> +1
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>> On Tue, 13 Sept 2022 at 06:45, Gengliang Wang <ltn...@gmail.com>
>>>>>>>> wrote:
>>>>>>>>
>>>>>>>> +1.
>>>>>>>>
>>>>>>>> Thank you, Yuming!
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>> On Mon, Sep 12, 2022 at 12:10 PM L. C. Hsieh <vii...@gmail.com>
>>>>>>>> wrote:
>>>>>>>>
>>>>>>>> +1
>>>>>>>>
>>>>>>>> Thanks Yuming!
>>>>>>>>
>>>>>>>> On Mon, Sep 12, 2022 at 11:50 AM Dongjoon Hyun <
>>>>>>>> dongjoon.h...@gmail.com> wrote:
>>>>>>>> >
>>>>>>>> > +1
>>>>>>>> >
>>>>>>>> > Thanks,
>>>>>>>> > Dongjoon.
>>>>>>>> >
>>>>>>>> > On Mon, Sep 12, 2022 at 6:38 AM Yuming Wang <wgy...@gmail.com>
>>>>>>>> wrote:
>>>>>>>> >>
>>>>>>>> >> Hi, All.
>>>>>>>> >>
>>>>>>>> >>
>>>>>>>> >>
>>>>>>>> >> Since Apache Spark 3.3.0 tag creation (Jun 10), new 138 patches
>>>>>>>> including 7 correctness patches arrived at branch-3.3.
>>>>>>>> >>
>>>>>>>> >>
>>>>>>>> >>
>>>>>>>> >> Shall we make a new release, Apache Spark 3.3.1, as the second
>>>>>>>> release at branch-3.3? I'd like to volunteer as the release manager for
>>>>>>>> Apache Spark 3.3.1.
>>>>>>>> >>
>>>>>>>> >>
>>>>>>>> >>
>>>>>>>> >> All changes:
>>>>>>>> >>
>>>>>>>> >> https://github.com/apache/spark/compare/v3.3.0...branch-3.3
>>>>>>>> <https://mailshield.baidu.com/check?q=WzRnV6InLAPdBDRyJZecGtPwF02t%2bnFNwOI8oAyGcb60kX%2bRCS6N3SUnFjTdf47bb94KmZHTTKE%2bBHbIT27Rog%3d%3d>
>>>>>>>> >>
>>>>>>>> >>
>>>>>>>> >>
>>>>>>>> >> Correctness issues:
>>>>>>>> >>
>>>>>>>> >> SPARK-40149: Propagate metadata columns through Project
>>>>>>>> >>
>>>>>>>> >> SPARK-40002: Don't push down limit through window using ntile
>>>>>>>> >>
>>>>>>>> >> SPARK-39976: ArrayIntersect should handle null in left
>>>>>>>> expression correctly
>>>>>>>> >>
>>>>>>>> >> SPARK-39833: Disable Parquet column index in DSv1 to fix a
>>>>>>>> correctness issue in the case of overlapping partition and data columns
>>>>>>>> >>
>>>>>>>> >> SPARK-39061: Set nullable correctly for Inline output attributes
>>>>>>>> >>
>>>>>>>> >> SPARK-39887: RemoveRedundantAliases should keep aliases that
>>>>>>>> make the output of projection nodes unique
>>>>>>>> >>
>>>>>>>> >> SPARK-38614: Don't push down limit through window that's using
>>>>>>>> percent_rank
>>>>>>>>
>>>>>>>>
>>>>>>>> ---------------------------------------------------------------------
>>>>>>>> To unsubscribe e-mail: dev-unsubscr...@spark.apache.org
>>>>>>>>
>>>>>>>> --
>>>>>>> John Zhuge
>>>>>>>
>>>>>>
>>>>>
>>>>> --
>>>>> //with Best Regards
>>>>> --Denis Bolshakov
>>>>> e-mail: bolshakov.de...@gmail.com
>>>>>
>>>>

Reply via email to