This should do:

> export SPARK_HOME=/opt/cloudera/parcels/SPARK2/lib/spark2
> export HIVE_HOME=/opt/cloudera/parcels/CDH/lib/hive
> export HADOOP_HOME=/opt/cloudera/parcels/CDH/lib/hadoop
> export HADOOP_CONF_DIR=/etc/hadoop/conf
> export HIVE_CONF_DIR=/etc/hive/conf



>
> mvn clean package -DskipTests -Pspark-2.1 -Dhadoop.version=2.6.0-cdh5.10.1
> -Phadoop-2.6 -Pvendor-repo -Pscala-2.10 -Psparkr -pl
> '!alluxio,!flink,!ignite,!lens,!cassandra,!bigquery,!scio' -e


You may needs additional steps depending which interpreters you use (like R
etc).


-- 
Ruslan Dautkhanov

On Fri, Aug 4, 2017 at 8:31 AM, Benjamin Kim <bbuil...@gmail.com> wrote:

> Hi Ruslan,
>
> Can you send me the steps you used to build it, especially the Maven
> command with the arguments? I will try to build it also.
>
> I do believe that the binaries are for official releases.
>
> Cheers,
> Ben
>
>
> On Wed, Aug 2, 2017 at 3:44 PM Ruslan Dautkhanov <dautkha...@gmail.com>
> wrote:
>
>> It was built. I think binaries are only available for official releases?
>>
>>
>>
>> --
>> Ruslan Dautkhanov
>>
>> On Wed, Aug 2, 2017 at 4:41 PM, Benjamin Kim <bbuil...@gmail.com> wrote:
>>
>>> Did you build Zeppelin or download the binary?
>>>
>>> On Wed, Aug 2, 2017 at 3:40 PM Ruslan Dautkhanov <dautkha...@gmail.com>
>>> wrote:
>>>
>>>> We're using an ~April snapshot of Zeppelin, so not sure about 0.7.1.
>>>>
>>>> Yes, we have that spark home in zeppelin-env.sh
>>>>
>>>>
>>>>
>>>> --
>>>> Ruslan Dautkhanov
>>>>
>>>> On Wed, Aug 2, 2017 at 4:31 PM, Benjamin Kim <bbuil...@gmail.com>
>>>> wrote:
>>>>
>>>>> Does this work with Zeppelin 0.7.1? We an error when setting
>>>>> SPARK_HOME in zeppelin-env.sh to what you have below.
>>>>>
>>>>> On Wed, Aug 2, 2017 at 3:24 PM Ruslan Dautkhanov <dautkha...@gmail.com>
>>>>> wrote:
>>>>>
>>>>>> You don't have to use spark2-shell and spark2-submit to use Spark 2.
>>>>>> That can be controled by setting SPARK_HOME using regular
>>>>>> spark-submit/spark-shell.
>>>>>>
>>>>>> $ which spark-submit
>>>>>> /usr/bin/spark-submit
>>>>>> $ which spark-shell
>>>>>> /usr/bin/spark-shell
>>>>>>
>>>>>> $ spark-shell
>>>>>> Welcome to
>>>>>>       ____              __
>>>>>>      / __/__  ___ _____/ /__
>>>>>>     _\ \/ _ \/ _ `/ __/  '_/
>>>>>>    /___/ .__/\_,_/_/ /_/\_\   version 1.6.0
>>>>>>       /_/
>>>>>>
>>>>>>
>>>>>>
>>>>>> $ export SPARK_HOME=/opt/cloudera/parcels/SPARK2/lib/spark2
>>>>>>
>>>>>> $ spark-shell
>>>>>> Welcome to
>>>>>>       ____              __
>>>>>>      / __/__  ___ _____/ /__
>>>>>>     _\ \/ _ \/ _ `/ __/  '_/
>>>>>>    /___/ .__/\_,_/_/ /_/\_\   version 2.1.0.cloudera1
>>>>>>       /_/
>>>>>>
>>>>>>
>>>>>> spark-submit and spark-shell are just shell script wrappers.
>>>>>>
>>>>>>
>>>>>>
>>>>>> --
>>>>>> Ruslan Dautkhanov
>>>>>>
>>>>>> On Wed, Aug 2, 2017 at 10:22 AM, Benjamin Kim <bbuil...@gmail.com>
>>>>>> wrote:
>>>>>>
>>>>>>> According to the Zeppelin documentation, Zeppelin 0.7.1 supports
>>>>>>> Spark 2.1. But, I don't know if it supports Spark 2.2 or even 2.1 from
>>>>>>> Cloudera. For some reason, Cloudera defaults to Spark 1.6 and so does 
>>>>>>> the
>>>>>>> calls to spark-shell and spark-submit. To force the use of Spark 2.x, 
>>>>>>> the
>>>>>>> calls need to be spark2-shell and spark2-submit. I wonder if this is
>>>>>>> causing the problem. By the way, we are using Java8 corporate wide, and
>>>>>>> there seems to be no problems using Zeppelin.
>>>>>>>
>>>>>>> Cheers,
>>>>>>> Ben
>>>>>>>
>>>>>>> On Tue, Aug 1, 2017 at 7:05 PM Ruslan Dautkhanov <
>>>>>>> dautkha...@gmail.com> wrote:
>>>>>>>
>>>>>>>> Might need to recompile Zeppelin with Scala 2.11?
>>>>>>>> Also Spark 2.2 now requires JDK8 I believe.
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>> --
>>>>>>>> Ruslan Dautkhanov
>>>>>>>>
>>>>>>>> On Tue, Aug 1, 2017 at 6:26 PM, Benjamin Kim <bbuil...@gmail.com>
>>>>>>>> wrote:
>>>>>>>>
>>>>>>>>> Here is more.
>>>>>>>>>
>>>>>>>>> org.apache.zeppelin.interpreter.InterpreterException: WARNING:
>>>>>>>>> User-defined SPARK_HOME (/opt/cloudera/parcels/SPARK2-
>>>>>>>>> 2.2.0.cloudera1-1.cdh5.12.0.p0.142354/lib/spark2) overrides
>>>>>>>>> detected (/opt/cloudera/parcels/SPARK2/lib/spark2).
>>>>>>>>> WARNING: Running spark-class from user-defined location.
>>>>>>>>> Exception in thread "main" java.lang.NoSuchMethodError:
>>>>>>>>> scala.Predef$.$conforms()Lscala/Predef$$less$colon$less;
>>>>>>>>> at org.apache.spark.util.Utils$.getDefaultPropertiesFile(
>>>>>>>>> Utils.scala:2103)
>>>>>>>>> at org.apache.spark.deploy.SparkSubmitArguments$$anonfun$
>>>>>>>>> mergeDefaultSparkProperties$1.apply(SparkSubmitArguments.
>>>>>>>>> scala:124)
>>>>>>>>> at org.apache.spark.deploy.SparkSubmitArguments$$anonfun$
>>>>>>>>> mergeDefaultSparkProperties$1.apply(SparkSubmitArguments.
>>>>>>>>> scala:124)
>>>>>>>>> at scala.Option.getOrElse(Option.scala:120)
>>>>>>>>> at org.apache.spark.deploy.SparkSubmitArguments.
>>>>>>>>> mergeDefaultSparkProperties(SparkSubmitArguments.scala:124)
>>>>>>>>> at org.apache.spark.deploy.SparkSubmitArguments.<init>(
>>>>>>>>> SparkSubmitArguments.scala:110)
>>>>>>>>> at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.
>>>>>>>>> scala:112)
>>>>>>>>> at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
>>>>>>>>>
>>>>>>>>> Cheers,
>>>>>>>>> Ben
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> On Tue, Aug 1, 2017 at 5:24 PM Jeff Zhang <zjf...@gmail.com>
>>>>>>>>> wrote:
>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> Then it is due to some classpath issue. I am not sure familiar
>>>>>>>>>> with CDH, please check whether spark of CDH include hadoop jar with 
>>>>>>>>>> it.
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> Benjamin Kim <bbuil...@gmail.com>于2017年8月2日周三 上午8:22写道:
>>>>>>>>>>
>>>>>>>>>>> Here is the error that was sent to me.
>>>>>>>>>>>
>>>>>>>>>>> org.apache.zeppelin.interpreter.InterpreterException: Exception
>>>>>>>>>>> in thread "main" java.lang.NoClassDefFoundError:
>>>>>>>>>>> org/apache/hadoop/fs/FSDataInputStream
>>>>>>>>>>> Caused by: java.lang.ClassNotFoundException:
>>>>>>>>>>> org.apache.hadoop.fs.FSDataInputStream
>>>>>>>>>>>
>>>>>>>>>>> Cheers,
>>>>>>>>>>> Ben
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> On Tue, Aug 1, 2017 at 5:20 PM Jeff Zhang <zjf...@gmail.com>
>>>>>>>>>>> wrote:
>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> By default, 0.7.1 doesn't support spark 2.2. But you can set
>>>>>>>>>>>> zeppelin.spark.enableSupportedVersionCheck in interpreter
>>>>>>>>>>>> setting to disable the supported version check.
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> Jeff Zhang <zjf...@gmail.com>于2017年8月2日周三 上午8:18写道:
>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>> What's the error you see in log ?
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>> Benjamin Kim <bbuil...@gmail.com>于2017年8月2日周三 上午8:18写道:
>>>>>>>>>>>>>
>>>>>>>>>>>>>> Has anyone configured Zeppelin 0.7.1 for Cloudera's release
>>>>>>>>>>>>>> of Spark 2.2? I can't get it to work. I downloaded the binary 
>>>>>>>>>>>>>> and set
>>>>>>>>>>>>>> SPARK_HOME to /opt/cloudera/parcels/SPARK2/lib/spark2. I
>>>>>>>>>>>>>> must be missing something.
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> Cheers,
>>>>>>>>>>>>>> Ben
>>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>
>>>>>>
>>>>
>>

Reply via email to