I started playing with Scala 2.12.0-M3 but the compilation didn't pass (as
expected)

Planning to get back to 2.12 once it is released.

FYI

On Wed, Nov 11, 2015 at 4:34 PM, Jakob Odersky <joder...@gmail.com> wrote:

> Hi Sukant,
>
> Regarding the first point: when building spark during my daily work, I
> always use Scala 2.11 and have only run into build problems once. Assuming
> a working build I have never had any issues with the resulting artifacts.
>
> More generally however, I would advise you to go with Scala 2.11 under all
> circumstances. Scala 2.10 has reached end-of-life and, from what I make out
> of your question, you have the opportunity to switch to a newer technology,
> so why stay with legacy? Furthermore, Scala 2.12 will be coming out early
> next year, so I reckon that Spark will switch to Scala 2.11 by default
> pretty soon*.
>
> regards,
> --Jakob
>
> *I'm myself pretty new to the Spark community so please don't take my
> words on it as gospel
>
>
> On 11 November 2015 at 15:25, Ted Yu <yuzhih...@gmail.com> wrote:
>
>> For #1, the published jars are usable.
>> However, you should build from source for your specific combination of
>> profiles.
>>
>> Cheers
>>
>> On Wed, Nov 11, 2015 at 3:22 PM, shajra-cogscale <
>> sha...@cognitivescale.com> wrote:
>>
>>> Hi,
>>>
>>> My company isn't using Spark in production yet, but we are using a bit of
>>> Scala.  There's a few people who have wanted to be conservative and keep
>>> our
>>> Scala at 2.10 in the event we start using Spark.  There are others who
>>> want
>>> to move to 2.11 with the idea that by the time we're using Spark it will
>>> be
>>> more or less 2.11-ready.
>>>
>>> It's hard to make a strong judgement on these kinds of things without
>>> getting some community feedback.
>>>
>>> Looking through the internet I saw:
>>>
>>> 1) There's advice to build 2.11 packages from source -- but also
>>> published
>>> jars to Maven Central for 2.11.  Are these jars on Maven Central usable
>>> and
>>> the advice to build from source outdated?
>>>
>>> 2)  There's a note that the JDBC RDD isn't 2.11-compliant.  This is okay
>>> for
>>> us, but is there anything else to worry about?
>>>
>>> It would be nice to get some answers to those questions as well as any
>>> other
>>> feedback from maintainers or anyone that's used Spark with Scala 2.11
>>> beyond
>>> simple examples.
>>>
>>> Thanks,
>>> Sukant
>>>
>>>
>>>
>>> --
>>> View this message in context:
>>> http://apache-spark-user-list.1001560.n3.nabble.com/Status-of-2-11-support-tp25362.html
>>> Sent from the Apache Spark User List mailing list archive at Nabble.com.
>>>
>>> ---------------------------------------------------------------------
>>> To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
>>> For additional commands, e-mail: user-h...@spark.apache.org
>>>
>>>
>>
>

Reply via email to