On Thu, May 12, 2016 at 9:23 PM, Luciano Resende <luckbr1...@gmail.com> wrote:
> Spark has moved to build using Scala 2.11 by default in master/trunk. > Does this mean that the pre-built binaries for download will also move to 2.11 as well? > > > As for the 2.0.0-SNAPSHOT, it is actually the version of master/trunk and > you might be missing some modules/profiles for your build. What command did > you use to build ? > > On Thu, May 12, 2016 at 9:01 PM, Raghava Mutharaju < > m.vijayaragh...@gmail.com> wrote: > >> Hello All, >> >> I built Spark from the source code available at >> https://github.com/apache/spark/. Although I haven't specified the >> "-Dscala-2.11" option (to build with Scala 2.11), from the build messages I >> see that it ended up using Scala 2.11. Now, for my application sbt, what >> should be the spark version? I tried the following >> >> val spark = "org.apache.spark" %% "spark-core" % "2.0.0-SNAPSHOT" >> val sparksql = "org.apache.spark" % "spark-sql_2.11" % "2.0.0-SNAPSHOT" >> >> and scalaVersion := "2.11.8" >> >> But this setting of spark version gives sbt error >> >> unresolved dependency: org.apache.spark#spark-core_2.11;2.0.0-SNAPSHOT >> >> I guess this is because the repository doesn't contain 2.0.0-SNAPSHOT. >> Does this mean, the only option is to put all the required jars in the lib >> folder (unmanaged dependencies)? >> >> Regards, >> Raghava. >> > > > > -- > Luciano Resende > http://twitter.com/lresende1975 > http://lresende.blogspot.com/ >