The pre-built binary distros should use 2.11 in 2.4.x. Artifacts for
both Scala versions are available, yes.
Yeah I think it should really say you can use 2.11 or 2.12.

On Thu, Sep 17, 2020 at 9:12 AM Jacek Laskowski <ja...@japila.pl> wrote:
>
> Hi,
>
> Just found this paragraph in 
> http://spark.apache.org/docs/2.4.6/index.html#downloading:
>
> "Spark runs on Java 8, Python 2.7+/3.4+ and R 3.1+. For the Scala API, Spark 
> 2.4.6 uses Scala 2.12. You will need to use a compatible Scala version 
> (2.12.x)."
>
> That seems to contradict the version of Scala in the pom.xml [1] which is 
> 2.11.12. I think this says that Spark 2.4.6 uses Scala 2.12 by default which 
> is incorrect to me. Am I missing something?
>
> My question is what's the official Scala version of Spark 2.4.6 (and others 
> in 2.4.x release line)?
>
> (I do know that Spark 2.4.x could be compiled with Scala 2.12, but that 
> requires scala-2.12 profile [2] to be enabled)
>
> [1] https://github.com/apache/spark/blob/v2.4.6/pom.xml#L158
> [2] https://github.com/apache/spark/blob/v2.4.6/pom.xml#L2830
>
> Pozdrawiam,
> Jacek Laskowski
> ----
> https://about.me/JacekLaskowski
> "The Internals Of" Online Books
> Follow me on https://twitter.com/jaceklaskowski
>

---------------------------------------------------------------------
To unsubscribe e-mail: dev-unsubscr...@spark.apache.org

Reply via email to