Ruslan, Quick observation. I see that you put Scala 2.10. How come you didn't use Scala 2.11?
Just curious. Cheers, Ben On Fri, Aug 4, 2017 at 8:36 AM Ruslan Dautkhanov <dautkha...@gmail.com> wrote: > This should do: > > >> export SPARK_HOME=/opt/cloudera/parcels/SPARK2/lib/spark2 >> export HIVE_HOME=/opt/cloudera/parcels/CDH/lib/hive >> export HADOOP_HOME=/opt/cloudera/parcels/CDH/lib/hadoop >> export HADOOP_CONF_DIR=/etc/hadoop/conf >> export HIVE_CONF_DIR=/etc/hive/conf > > > >> >> mvn clean package -DskipTests -Pspark-2.1 >> -Dhadoop.version=2.6.0-cdh5.10.1 -Phadoop-2.6 -Pvendor-repo -Pscala-2.10 >> -Psparkr -pl '!alluxio,!flink,!ignite,!lens,!cassandra,!bigquery,!scio' -e > > > You may needs additional steps depending which interpreters you use (like > R etc). > > > -- > Ruslan Dautkhanov > > On Fri, Aug 4, 2017 at 8:31 AM, Benjamin Kim <bbuil...@gmail.com> wrote: > >> Hi Ruslan, >> >> Can you send me the steps you used to build it, especially the Maven >> command with the arguments? I will try to build it also. >> >> I do believe that the binaries are for official releases. >> >> Cheers, >> Ben >> >> >> On Wed, Aug 2, 2017 at 3:44 PM Ruslan Dautkhanov <dautkha...@gmail.com> >> wrote: >> >>> It was built. I think binaries are only available for official releases? >>> >>> >>> >>> -- >>> Ruslan Dautkhanov >>> >>> On Wed, Aug 2, 2017 at 4:41 PM, Benjamin Kim <bbuil...@gmail.com> wrote: >>> >>>> Did you build Zeppelin or download the binary? >>>> >>>> On Wed, Aug 2, 2017 at 3:40 PM Ruslan Dautkhanov <dautkha...@gmail.com> >>>> wrote: >>>> >>>>> We're using an ~April snapshot of Zeppelin, so not sure about 0.7.1. >>>>> >>>>> Yes, we have that spark home in zeppelin-env.sh >>>>> >>>>> >>>>> >>>>> -- >>>>> Ruslan Dautkhanov >>>>> >>>>> On Wed, Aug 2, 2017 at 4:31 PM, Benjamin Kim <bbuil...@gmail.com> >>>>> wrote: >>>>> >>>>>> Does this work with Zeppelin 0.7.1? We an error when setting >>>>>> SPARK_HOME in zeppelin-env.sh to what you have below. >>>>>> >>>>>> On Wed, Aug 2, 2017 at 3:24 PM Ruslan Dautkhanov < >>>>>> dautkha...@gmail.com> wrote: >>>>>> >>>>>>> You don't have to use spark2-shell and spark2-submit to use Spark 2. >>>>>>> That can be controled by setting SPARK_HOME using regular >>>>>>> spark-submit/spark-shell. >>>>>>> >>>>>>> $ which spark-submit >>>>>>> /usr/bin/spark-submit >>>>>>> $ which spark-shell >>>>>>> /usr/bin/spark-shell >>>>>>> >>>>>>> $ spark-shell >>>>>>> Welcome to >>>>>>> ____ __ >>>>>>> / __/__ ___ _____/ /__ >>>>>>> _\ \/ _ \/ _ `/ __/ '_/ >>>>>>> /___/ .__/\_,_/_/ /_/\_\ version 1.6.0 >>>>>>> /_/ >>>>>>> >>>>>>> >>>>>>> >>>>>>> $ export SPARK_HOME=/opt/cloudera/parcels/SPARK2/lib/spark2 >>>>>>> >>>>>>> $ spark-shell >>>>>>> Welcome to >>>>>>> ____ __ >>>>>>> / __/__ ___ _____/ /__ >>>>>>> _\ \/ _ \/ _ `/ __/ '_/ >>>>>>> /___/ .__/\_,_/_/ /_/\_\ version 2.1.0.cloudera1 >>>>>>> /_/ >>>>>>> >>>>>>> >>>>>>> spark-submit and spark-shell are just shell script wrappers. >>>>>>> >>>>>>> >>>>>>> >>>>>>> -- >>>>>>> Ruslan Dautkhanov >>>>>>> >>>>>>> On Wed, Aug 2, 2017 at 10:22 AM, Benjamin Kim <bbuil...@gmail.com> >>>>>>> wrote: >>>>>>> >>>>>>>> According to the Zeppelin documentation, Zeppelin 0.7.1 supports >>>>>>>> Spark 2.1. But, I don't know if it supports Spark 2.2 or even 2.1 from >>>>>>>> Cloudera. For some reason, Cloudera defaults to Spark 1.6 and so does >>>>>>>> the >>>>>>>> calls to spark-shell and spark-submit. To force the use of Spark 2.x, >>>>>>>> the >>>>>>>> calls need to be spark2-shell and spark2-submit. I wonder if this is >>>>>>>> causing the problem. By the way, we are using Java8 corporate wide, and >>>>>>>> there seems to be no problems using Zeppelin. >>>>>>>> >>>>>>>> Cheers, >>>>>>>> Ben >>>>>>>> >>>>>>>> On Tue, Aug 1, 2017 at 7:05 PM Ruslan Dautkhanov < >>>>>>>> dautkha...@gmail.com> wrote: >>>>>>>> >>>>>>>>> Might need to recompile Zeppelin with Scala 2.11? >>>>>>>>> Also Spark 2.2 now requires JDK8 I believe. >>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> -- >>>>>>>>> Ruslan Dautkhanov >>>>>>>>> >>>>>>>>> On Tue, Aug 1, 2017 at 6:26 PM, Benjamin Kim <bbuil...@gmail.com> >>>>>>>>> wrote: >>>>>>>>> >>>>>>>>>> Here is more. >>>>>>>>>> >>>>>>>>>> org.apache.zeppelin.interpreter.InterpreterException: WARNING: >>>>>>>>>> User-defined SPARK_HOME >>>>>>>>>> (/opt/cloudera/parcels/SPARK2-2.2.0.cloudera1-1.cdh5.12.0.p0.142354/lib/spark2) >>>>>>>>>> overrides detected (/opt/cloudera/parcels/SPARK2/lib/spark2). >>>>>>>>>> WARNING: Running spark-class from user-defined location. >>>>>>>>>> Exception in thread "main" java.lang.NoSuchMethodError: >>>>>>>>>> scala.Predef$.$conforms()Lscala/Predef$$less$colon$less; >>>>>>>>>> at >>>>>>>>>> org.apache.spark.util.Utils$.getDefaultPropertiesFile(Utils.scala:2103) >>>>>>>>>> at >>>>>>>>>> org.apache.spark.deploy.SparkSubmitArguments$$anonfun$mergeDefaultSparkProperties$1.apply(SparkSubmitArguments.scala:124) >>>>>>>>>> at >>>>>>>>>> org.apache.spark.deploy.SparkSubmitArguments$$anonfun$mergeDefaultSparkProperties$1.apply(SparkSubmitArguments.scala:124) >>>>>>>>>> at scala.Option.getOrElse(Option.scala:120) >>>>>>>>>> at >>>>>>>>>> org.apache.spark.deploy.SparkSubmitArguments.mergeDefaultSparkProperties(SparkSubmitArguments.scala:124) >>>>>>>>>> at >>>>>>>>>> org.apache.spark.deploy.SparkSubmitArguments.<init>(SparkSubmitArguments.scala:110) >>>>>>>>>> at >>>>>>>>>> org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:112) >>>>>>>>>> at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala) >>>>>>>>>> >>>>>>>>>> Cheers, >>>>>>>>>> Ben >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> On Tue, Aug 1, 2017 at 5:24 PM Jeff Zhang <zjf...@gmail.com> >>>>>>>>>> wrote: >>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> Then it is due to some classpath issue. I am not sure familiar >>>>>>>>>>> with CDH, please check whether spark of CDH include hadoop jar with >>>>>>>>>>> it. >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> Benjamin Kim <bbuil...@gmail.com>于2017年8月2日周三 上午8:22写道: >>>>>>>>>>> >>>>>>>>>>>> Here is the error that was sent to me. >>>>>>>>>>>> >>>>>>>>>>>> org.apache.zeppelin.interpreter.InterpreterException: Exception >>>>>>>>>>>> in thread "main" java.lang.NoClassDefFoundError: >>>>>>>>>>>> org/apache/hadoop/fs/FSDataInputStream >>>>>>>>>>>> Caused by: java.lang.ClassNotFoundException: >>>>>>>>>>>> org.apache.hadoop.fs.FSDataInputStream >>>>>>>>>>>> >>>>>>>>>>>> Cheers, >>>>>>>>>>>> Ben >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> On Tue, Aug 1, 2017 at 5:20 PM Jeff Zhang <zjf...@gmail.com> >>>>>>>>>>>> wrote: >>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> By default, 0.7.1 doesn't support spark 2.2. But you can set >>>>>>>>>>>>> zeppelin.spark.enableSupportedVersionCheck >>>>>>>>>>>>> in interpreter setting to disable the supported version check. >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> Jeff Zhang <zjf...@gmail.com>于2017年8月2日周三 上午8:18写道: >>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> What's the error you see in log ? >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> Benjamin Kim <bbuil...@gmail.com>于2017年8月2日周三 上午8:18写道: >>>>>>>>>>>>>> >>>>>>>>>>>>>>> Has anyone configured Zeppelin 0.7.1 for Cloudera's release >>>>>>>>>>>>>>> of Spark 2.2? I can't get it to work. I downloaded the binary >>>>>>>>>>>>>>> and set >>>>>>>>>>>>>>> SPARK_HOME to /opt/cloudera/parcels/SPARK2/lib/spark2. I must >>>>>>>>>>>>>>> be missing >>>>>>>>>>>>>>> something. >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> Cheers, >>>>>>>>>>>>>>> Ben >>>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>> >>>>>>> >>>>> >>> >