Well, this is very strange. My only change is to add -X to make-distribution and it succeeds:
% git diff (spark/spark) *diff --git a/make-distribution.sh b/make-distribution.sh* *index a2b0c43..351fac2 100755* *--- a/make-distribution.sh* *+++ b/make-distribution.sh* @@ -183,7 +183,7 @@ export MAVEN_OPTS="-Xmx2g -XX:MaxPermSize=512M -XX:ReservedCodeCacheSize=512m" # Store the command as an array because $MVN variable might have spaces in it. # Normal quoting tricks don't work. # See: http://mywiki.wooledge.org/BashFAQ/050 -BUILD_COMMAND=("$MVN" clean package -DskipTests $@) +BUILD_COMMAND=("$MVN" -X clean package -DskipTests $@) # Actually build the jar echo -e "\nBuilding with..." export JAVA_HOME=/Library/Java/Home % which javac /usr/bin/javac % javac -version javac 1.7.0_79 On Mon, Aug 24, 2015 at 11:30 PM, Sean Owen <so...@cloudera.com> wrote: > -cdh-user > > This suggests that Maven is still using Java 6. I think this is indeed > controlled by JAVA_HOME. Use 'mvn -X ...' to see a lot more about what > is being used and why. I still suspect JAVA_HOME is not visible to the > Maven process. Or maybe you have JRE 7 installed but not JDK 7 and > it's somehow still finding the Java 6 javac. > > On Tue, Aug 25, 2015 at 3:45 AM, Eric Friedman > <eric.d.fried...@gmail.com> wrote: > > I'm trying to build Spark 1.4 with Java 7 and despite having that as my > > JAVA_HOME, I get > > > > [INFO] --- scala-maven-plugin:3.2.2:compile (scala-compile-first) @ > > spark-launcher_2.10 --- > > > > [INFO] Using zinc server for incremental compilation > > > > [info] Compiling 8 Java sources to > > /Users/eric/spark/spark/launcher/target/scala-2.10/classes... > > > > [error] javac: invalid source release: 1.7 > > > > [error] Usage: javac <options> <source files> > > > > [error] use -help for a list of possible options > > > > [error] Compile failed at Aug 24, 2015 7:44:40 PM [0.020s] > > > > [INFO] > > ------------------------------------------------------------------------ > > > > [INFO] Reactor Summary: > > > > [INFO] > > > > [INFO] Spark Project Parent POM ........................... SUCCESS [ > 3.109 > > s] > > > > [INFO] Spark Project Launcher ............................. FAILURE [ > 4.493 > > s] > > > > > > On Fri, Aug 21, 2015 at 9:43 AM, Marcelo Vanzin <van...@cloudera.com> > wrote: > >> > >> That was only true until Spark 1.3. Spark 1.4 can be built with JDK7 > >> and pyspark will still work. > >> > >> On Fri, Aug 21, 2015 at 8:29 AM, Chen Song <chen.song...@gmail.com> > wrote: > >> > Thanks Sean. > >> > > >> > So how PySpark is supported. I thought PySpark needs jdk 1.6. > >> > > >> > Chen > >> > > >> > On Fri, Aug 21, 2015 at 11:16 AM, Sean Owen <so...@cloudera.com> > wrote: > >> >> > >> >> Spark 1.4 requires Java 7. > >> >> > >> >> > >> >> On Fri, Aug 21, 2015, 3:12 PM Chen Song <chen.song...@gmail.com> > wrote: > >> >>> > >> >>> I tried to build Spark 1.4.1 on cdh 5.4.0. Because we need to > support > >> >>> PySpark, I used JDK 1.6. > >> >>> > >> >>> I got the following error, > >> >>> > >> >>> [INFO] --- scala-maven-plugin:3.2.0:testCompile > >> >>> (scala-test-compile-first) @ spark-streaming_2.10 --- > >> >>> > >> >>> java.lang.UnsupportedClassVersionError: > >> >>> org/apache/hadoop/io/LongWritable > >> >>> : Unsupported major.minor version 51.0 > >> >>> at java.lang.ClassLoader.defineClass1(Native Method) > >> >>> at java.lang.ClassLoader.defineClassCond(ClassLoader.java:637) > >> >>> at java.lang.ClassLoader.defineClass(ClassLoader.java:621) > >> >>> at > >> >>> > >> >>> > java.security.SecureClassLoader.defineClass(SecureClassLoader.java:141) > >> >>> > >> >>> I know that is due to the hadoop jar for cdh5.4.0 is built with JDK > 7. > >> >>> Anyone has done this before? > >> >>> > >> >>> Thanks, > >> >>> > >> >>> -- > >> >>> Chen Song > >> >>> > >> > > >> > > >> > > >> > -- > >> > Chen Song > >> > > >> > -- > >> > > >> > --- > >> > You received this message because you are subscribed to the Google > >> > Groups > >> > "CDH Users" group. > >> > To unsubscribe from this group and stop receiving emails from it, send > >> > an > >> > email to cdh-user+unsubscr...@cloudera.org. > >> > For more options, visit > >> > https://groups.google.com/a/cloudera.org/d/optout. > >> > >> > >> > >> -- > >> Marcelo > >> > >> --------------------------------------------------------------------- > >> To unsubscribe, e-mail: user-unsubscr...@spark.apache.org > >> For additional commands, e-mail: user-h...@spark.apache.org > >> > > >