Hi, My company isn't using Spark in production yet, but we are using a bit of Scala. There's a few people who have wanted to be conservative and keep our Scala at 2.10 in the event we start using Spark. There are others who want to move to 2.11 with the idea that by the time we're using Spark it will be more or less 2.11-ready.
It's hard to make a strong judgement on these kinds of things without getting some community feedback. Looking through the internet I saw: 1) There's advice to build 2.11 packages from source -- but also published jars to Maven Central for 2.11. Are these jars on Maven Central usable and the advice to build from source outdated? 2) There's a note that the JDBC RDD isn't 2.11-compliant. This is okay for us, but is there anything else to worry about? It would be nice to get some answers to those questions as well as any other feedback from maintainers or anyone that's used Spark with Scala 2.11 beyond simple examples. Thanks, Sukant -- View this message in context: http://apache-spark-user-list.1001560.n3.nabble.com/Status-of-2-11-support-tp25362.html Sent from the Apache Spark User List mailing list archive at Nabble.com. --------------------------------------------------------------------- To unsubscribe, e-mail: user-unsubscr...@spark.apache.org For additional commands, e-mail: user-h...@spark.apache.org