I'm still not fully convinced that we need to drop Spark 1.x support,
instead of supporting both 1.x and 2.x. I would appreciate if we could
first conclude the discussion around migrating to Spark 2.0.

Furthermore, I think that creating a dependency to Spark versioning would
unnecessarily complicate our own release process. I would rather use major
releases as an opportunity to cleanup APIs and drop certain language
features. And this is unlikely to coincide with Spark's releases. From my
perspective it would be even more confusing for a user to release a major
version for a relatively minor change as support for a new Spark version.

Regards,
Matthias




From:   Luciano Resende <luckbr1...@gmail.com>
To:     dev@systemml.incubator.apache.org
Date:   08/25/2016 07:15 PM
Subject:        Re: [DISCUSS] Apache SystemML Release 1.0.0



On Wed, Aug 24, 2016 at 6:11 PM, <dusenberr...@gmail.com> wrote:

> Yes I'm also in favor of moving to a 1.0 version for our upcoming release
> targeting the Spark 1.x series. Since we'll also be subsequently
releasing
> a version targeting the Spark 2.x series, I would also like to suggest
that
> we name that version 2.0. This version naming scheme would allow us to
> easily associate a SystemML version with the Spark series that it
targets,
> thus reducing confusion for a user. Rather than view a 2.0 version as a
> successor to 1.0, let's view it instead as simply a naming scheme that
> corresponds to the targeted version of Spark.
>
> So, 1.0 would be our upcoming release targeting Spark 1.x, and 2.0 would
> be our upcoming release targeting Spark 2.x.
>
>
+ 1



--
Luciano Resende
http://twitter.com/lresende1975
http://lresende.blogspot.com/

Reply via email to