Sorry, Should have sent this to user…
However… it looks like the docs page may need some editing? Thx -Mike > Begin forwarded message: > > From: Michael Segel <msegel_had...@hotmail.com> > Subject: Silly question about building Spark 1.4.1 > Date: July 20, 2015 at 12:26:40 PM MST > To: d...@spark.apache.org > > Hi, > > I’m looking at the online docs for building spark 1.4.1 … > > http://spark.apache.org/docs/latest/building-spark.html > <http://spark.apache.org/docs/latest/building-spark.html> > > I was interested in building spark for Scala 2.11 (latest scala) and also for > Hive and JDBC support. > > The docs say: > “ > To produce a Spark package compiled with Scala 2.11, use the -Dscala-2.11 > property: > dev/change-version-to-2.11.sh > mvn -Pyarn -Phadoop-2.4 -Dscala-2.11 -DskipTests clean package > “ > So… > Is there a reason I shouldn’t build against hadoop-2.6 ? > > If I want to add the Thirft and Hive support, is it possible? > Looking at the Scala build, it looks like hive support is being built? > (Looking at the stdout messages…) > Should the docs be updated? Am I missing something? > (Dean W. can confirm, I am completely brain dead. ;-) > > Thx > > -Mike > PS. Yes I can probably download a prebuilt image, but I’m a glutton for > punishment. ;-) >