We're using an ~April snapshot of Zeppelin, so not sure about 0.7.1.

Yes, we have that spark home in zeppelin-env.sh



-- 
Ruslan Dautkhanov

On Wed, Aug 2, 2017 at 4:31 PM, Benjamin Kim <bbuil...@gmail.com> wrote:

> Does this work with Zeppelin 0.7.1? We an error when setting SPARK_HOME in
> zeppelin-env.sh to what you have below.
>
> On Wed, Aug 2, 2017 at 3:24 PM Ruslan Dautkhanov <dautkha...@gmail.com>
> wrote:
>
>> You don't have to use spark2-shell and spark2-submit to use Spark 2.
>> That can be controled by setting SPARK_HOME using regular
>> spark-submit/spark-shell.
>>
>> $ which spark-submit
>> /usr/bin/spark-submit
>> $ which spark-shell
>> /usr/bin/spark-shell
>>
>> $ spark-shell
>> Welcome to
>>       ____              __
>>      / __/__  ___ _____/ /__
>>     _\ \/ _ \/ _ `/ __/  '_/
>>    /___/ .__/\_,_/_/ /_/\_\   version 1.6.0
>>       /_/
>>
>>
>>
>> $ export SPARK_HOME=/opt/cloudera/parcels/SPARK2/lib/spark2
>>
>> $ spark-shell
>> Welcome to
>>       ____              __
>>      / __/__  ___ _____/ /__
>>     _\ \/ _ \/ _ `/ __/  '_/
>>    /___/ .__/\_,_/_/ /_/\_\   version 2.1.0.cloudera1
>>       /_/
>>
>>
>> spark-submit and spark-shell are just shell script wrappers.
>>
>>
>>
>> --
>> Ruslan Dautkhanov
>>
>> On Wed, Aug 2, 2017 at 10:22 AM, Benjamin Kim <bbuil...@gmail.com> wrote:
>>
>>> According to the Zeppelin documentation, Zeppelin 0.7.1 supports Spark
>>> 2.1. But, I don't know if it supports Spark 2.2 or even 2.1 from Cloudera.
>>> For some reason, Cloudera defaults to Spark 1.6 and so does the calls to
>>> spark-shell and spark-submit. To force the use of Spark 2.x, the calls need
>>> to be spark2-shell and spark2-submit. I wonder if this is causing the
>>> problem. By the way, we are using Java8 corporate wide, and there seems to
>>> be no problems using Zeppelin.
>>>
>>> Cheers,
>>> Ben
>>>
>>> On Tue, Aug 1, 2017 at 7:05 PM Ruslan Dautkhanov <dautkha...@gmail.com>
>>> wrote:
>>>
>>>> Might need to recompile Zeppelin with Scala 2.11?
>>>> Also Spark 2.2 now requires JDK8 I believe.
>>>>
>>>>
>>>>
>>>> --
>>>> Ruslan Dautkhanov
>>>>
>>>> On Tue, Aug 1, 2017 at 6:26 PM, Benjamin Kim <bbuil...@gmail.com>
>>>> wrote:
>>>>
>>>>> Here is more.
>>>>>
>>>>> org.apache.zeppelin.interpreter.InterpreterException: WARNING:
>>>>> User-defined SPARK_HOME (/opt/cloudera/parcels/SPARK2-
>>>>> 2.2.0.cloudera1-1.cdh5.12.0.p0.142354/lib/spark2) overrides detected
>>>>> (/opt/cloudera/parcels/SPARK2/lib/spark2).
>>>>> WARNING: Running spark-class from user-defined location.
>>>>> Exception in thread "main" java.lang.NoSuchMethodError:
>>>>> scala.Predef$.$conforms()Lscala/Predef$$less$colon$less;
>>>>> at org.apache.spark.util.Utils$.getDefaultPropertiesFile(
>>>>> Utils.scala:2103)
>>>>> at org.apache.spark.deploy.SparkSubmitArguments$$anonfun$
>>>>> mergeDefaultSparkProperties$1.apply(SparkSubmitArguments.scala:124)
>>>>> at org.apache.spark.deploy.SparkSubmitArguments$$anonfun$
>>>>> mergeDefaultSparkProperties$1.apply(SparkSubmitArguments.scala:124)
>>>>> at scala.Option.getOrElse(Option.scala:120)
>>>>> at org.apache.spark.deploy.SparkSubmitArguments.
>>>>> mergeDefaultSparkProperties(SparkSubmitArguments.scala:124)
>>>>> at org.apache.spark.deploy.SparkSubmitArguments.<init>(
>>>>> SparkSubmitArguments.scala:110)
>>>>> at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:112)
>>>>> at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
>>>>>
>>>>> Cheers,
>>>>> Ben
>>>>>
>>>>>
>>>>> On Tue, Aug 1, 2017 at 5:24 PM Jeff Zhang <zjf...@gmail.com> wrote:
>>>>>
>>>>>>
>>>>>> Then it is due to some classpath issue. I am not sure familiar with
>>>>>> CDH, please check whether spark of CDH include hadoop jar with it.
>>>>>>
>>>>>>
>>>>>> Benjamin Kim <bbuil...@gmail.com>于2017年8月2日周三 上午8:22写道:
>>>>>>
>>>>>>> Here is the error that was sent to me.
>>>>>>>
>>>>>>> org.apache.zeppelin.interpreter.InterpreterException: Exception in
>>>>>>> thread "main" java.lang.NoClassDefFoundError: org/apache/hadoop/fs/
>>>>>>> FSDataInputStream
>>>>>>> Caused by: java.lang.ClassNotFoundException: org.apache.hadoop.fs.
>>>>>>> FSDataInputStream
>>>>>>>
>>>>>>> Cheers,
>>>>>>> Ben
>>>>>>>
>>>>>>>
>>>>>>> On Tue, Aug 1, 2017 at 5:20 PM Jeff Zhang <zjf...@gmail.com> wrote:
>>>>>>>
>>>>>>>>
>>>>>>>> By default, 0.7.1 doesn't support spark 2.2. But you can set
>>>>>>>> zeppelin.spark.enableSupportedVersionCheck in interpreter setting
>>>>>>>> to disable the supported version check.
>>>>>>>>
>>>>>>>>
>>>>>>>> Jeff Zhang <zjf...@gmail.com>于2017年8月2日周三 上午8:18写道:
>>>>>>>>
>>>>>>>>>
>>>>>>>>> What's the error you see in log ?
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> Benjamin Kim <bbuil...@gmail.com>于2017年8月2日周三 上午8:18写道:
>>>>>>>>>
>>>>>>>>>> Has anyone configured Zeppelin 0.7.1 for Cloudera's release of
>>>>>>>>>> Spark 2.2? I can't get it to work. I downloaded the binary and set
>>>>>>>>>> SPARK_HOME to /opt/cloudera/parcels/SPARK2/lib/spark2. I must be
>>>>>>>>>> missing something.
>>>>>>>>>>
>>>>>>>>>> Cheers,
>>>>>>>>>> Ben
>>>>>>>>>>
>>>>>>>>>
>>>>
>>

Reply via email to