A related question. Are the plans to move the default Spark builds to Scala 2.11 with Spark 2.0?
Regards Deenar On 27 January 2016 at 19:55, Michael Armbrust <mich...@databricks.com> wrote: > We do maintenance releases on demand when there is enough to justify doing > one. I'm hoping to cut 1.6.1 soon, but have not had time yet. > > On Wed, Jan 27, 2016 at 8:12 AM, Daniel Siegmann < > daniel.siegm...@teamaol.com> wrote: > >> Will there continue to be monthly releases on the 1.6.x branch during the >> additional time for bug fixes and such? >> >> On Tue, Jan 26, 2016 at 11:28 PM, Koert Kuipers <ko...@tresata.com> >> wrote: >> >>> thanks thats all i needed >>> >>> On Tue, Jan 26, 2016 at 6:19 PM, Sean Owen <so...@cloudera.com> wrote: >>> >>>> I think it will come significantly later -- or else we'd be at code >>>> freeze for 2.x in a few days. I haven't heard anyone discuss this >>>> officially but had batted around May or so instead informally in >>>> conversation. Does anyone have a particularly strong opinion on that? >>>> That's basically an extra 3 month period. >>>> >>>> https://cwiki.apache.org/confluence/display/SPARK/Wiki+Homepage >>>> >>>> On Tue, Jan 26, 2016 at 10:00 PM, Koert Kuipers <ko...@tresata.com> >>>> wrote: >>>> > Is the idea that spark 2.0 comes out roughly 3 months after 1.6? So >>>> > quarterly release as usual? >>>> > Thanks >>>> >>> >>> >> >