Thanks for your answer Matei. I agree that a more explicit maintenance
policy is needed (even for the 2.x releases). I did not immediately
find anything about this in the website, so I ended up assuming the
information of the wikipedia article that says that the 1.6.x line is
still maintained.

I see that Spark as an open source project can get updates if the
community brings them in, but it is probably also a good idea to be
clear about the expectations for the end users. I suppose some users
who can migrate to version 2 won’t do it if there is still support
(notice that ‘support’ can be tricky considering how different
companies re-package/maintain Spark but this is a different
discussion). Anyway it would be great to have this defined somewhere.
Maybe worth a discussion on dev@.

On Thu, Oct 19, 2017 at 11:20 PM, Matei Zaharia <matei.zaha...@gmail.com> wrote:
> Hi Ismael,
>
> It depends on what you mean by “support”. In general, there won’t be new 
> feature releases for 1.X (e.g. Spark 1.7) because all the new features are 
> being added to the master branch. However, there is always room for bug fix 
> releases if there is a catastrophic bug, and committers can make those at any 
> time. In general though, I’d recommend moving workloads to Spark 2.x. We 
> tried to make the migration as easy as possible (a few APIs changed, but not 
> many), and 2.x has been out for a long time now and is widely used.
>
> We should perhaps write a more explicit maintenance policy, but all of this 
> is run based on what committers want to work on; if someone thinks that 
> there’s a serious enough issue in 1.6 to update it, they can put together a 
> new release. It does help to hear from users about this though, e.g. if you 
> think there’s a significant issue that people are missing.
>
> Matei
>
>> On Oct 19, 2017, at 5:20 AM, Ismaël Mejía <ieme...@gmail.com> wrote:
>>
>> Hello,
>>
>> I noticed that some of the (Big Data / Cloud Managed) Hadoop
>> distributions are starting to (phase out / deprecate) Spark 1.x and I
>> was wondering if the Spark community has already decided when will it
>> end the support for Spark 1.x. I ask this also considering that the
>> latest release in the series is already almost one year old. Any idea
>> on this ?
>>
>> Thanks,
>> Ismaël
>>
>> ---------------------------------------------------------------------
>> To unsubscribe e-mail: user-unsubscr...@spark.apache.org
>>
>

---------------------------------------------------------------------
To unsubscribe e-mail: user-unsubscr...@spark.apache.org

Reply via email to