I think it is fixed in https://github.com/apache/zeppelin/pull/1237, please
try that.

On Thu, Jul 28, 2016 at 1:58 AM, Jonathan Esterhazy <
jonathan.esterh...@gmail.com> wrote:

> I just tried this with using zeppelin master, and the classnotfound error
> is still there.
>
>
> On Thu, Jul 7, 2016 at 7:07 AM, Hyung Sung Shim <hss...@nflabs.com> wrote:
>
>> Hi. Jonathan.
>> If it is possible, could you use the zeppelin master on the github?
>> Zeppelin is working well with spark authentication when i tested.
>>
>>
>>
>> 2016-07-02 3:05 GMT+09:00 Jonathan Esterhazy <
>> jonathan.esterh...@gmail.com>:
>>
>>> Hyung, thx for your help. I opened these:
>>>
>>> https://issues.apache.org/jira/browse/ZEPPELIN-1096 (this scala problem)
>>> https://issues.apache.org/jira/browse/ZEPPELIN-1097 (similar looking
>>> python problem)
>>>
>>> LMK if I can provide more info or help in some way.
>>>
>>> On Fri, Jul 1, 2016 at 5:08 AM, Hyung Sung Shim <hss...@nflabs.com>
>>> wrote:
>>>
>>>> Hi Jonathan.
>>>> Unfortunately I got same error in my test bed.
>>>> Do you mind create an jira issue for this?
>>>>
>>>> 2016-07-01 3:19 GMT+09:00 Jonathan Esterhazy <
>>>> jonathan.esterh...@gmail.com>:
>>>>
>>>>> I added this and still get the same exception. The same property is
>>>>> also set in spark-defaults.conf.
>>>>>
>>>>> After that didn't work, I also tried adding --conf
>>>>> spark.authenticate=true and --conf
>>>>> spark.authenticate.enableSaslEncryption=true, to match the other related
>>>>> settings in spark-defaults.conf. Still get the same 
>>>>> classnotfoundexception.
>>>>>
>>>>>
>>>>> On Thu, Jun 30, 2016 at 10:45 AM, Hyung Sung Shim <hss...@nflabs.com>
>>>>> wrote:
>>>>>
>>>>>> Please add the *export SPARK_SUBMIT_OPTIONS="--conf
>>>>>> spark.authenticate.secret=secret"* in conf/zeppelin-env.sh and
>>>>>> restart zeppelin and retry your code.
>>>>>>
>>>>>>
>>>>>> 2016-06-30 23:34 GMT+09:00 Jonathan Esterhazy <
>>>>>> jonathan.esterh...@gmail.com>:
>>>>>>
>>>>>>> yes it does. i only see this problem in zeppelin.
>>>>>>>
>>>>>>> On Thu, Jun 30, 2016 at 7:05 AM, Hyung Sung Shim <hss...@nflabs.com>
>>>>>>> wrote:
>>>>>>>
>>>>>>>> Hi Jonathan.
>>>>>>>> It's not easy to build the test environments but I am working on
>>>>>>>> this.
>>>>>>>> I have question for you.
>>>>>>>> Does your code working well on spark-shell in the
>>>>>>>> spark.authenticate mode?
>>>>>>>>
>>>>>>>> 2016-06-30 22:47 GMT+09:00 Jonathan Esterhazy <
>>>>>>>> jonathan.esterh...@gmail.com>:
>>>>>>>>
>>>>>>>>> Hyung, did you have any luck w/ zeppelin + spark authentication?
>>>>>>>>> I'm quite stumped.
>>>>>>>>>
>>>>>>>>> thx.
>>>>>>>>>
>>>>>>>>> On Tue, Jun 28, 2016 at 9:11 PM, Hyung Sung Shim <
>>>>>>>>> hss...@nflabs.com> wrote:
>>>>>>>>>
>>>>>>>>>> Thank you.
>>>>>>>>>> Let me try.
>>>>>>>>>>
>>>>>>>>>> 2016-06-28 22:18 GMT+09:00 Jonathan Esterhazy <
>>>>>>>>>> jonathan.esterh...@gmail.com>:
>>>>>>>>>>
>>>>>>>>>>> Hyung,
>>>>>>>>>>>
>>>>>>>>>>> Yes, here they are.
>>>>>>>>>>>
>>>>>>>>>>> zeppelin-env.sh:
>>>>>>>>>>>
>>>>>>>>>>> export ZEPPELIN_PORT=8890
>>>>>>>>>>> export ZEPPELIN_CONF_DIR=/etc/zeppelin/conf
>>>>>>>>>>> export ZEPPELIN_LOG_DIR=/var/log/zeppelin
>>>>>>>>>>> export ZEPPELIN_PID_DIR=/var/run/zeppelin
>>>>>>>>>>> export ZEPPELIN_PID=$ZEPPELIN_PID_DIR/zeppelin.pid
>>>>>>>>>>> export ZEPPELIN_NOTEBOOK_DIR=/var/lib/zeppelin/notebook
>>>>>>>>>>> export ZEPPELIN_WAR_TEMPDIR=/var/run/zeppelin/webapps
>>>>>>>>>>> export MASTER=yarn-client
>>>>>>>>>>> export SPARK_HOME=/usr/lib/spark
>>>>>>>>>>> export HADOOP_CONF_DIR=/etc/hadoop/conf
>>>>>>>>>>> export
>>>>>>>>>>> CLASSPATH=":/etc/hive/conf:/usr/lib/hadoop-lzo/lib/*:/usr/lib/hadoop/hadoop-aws.jar:/usr/share/aws/aws-java-sdk/*:/usr/share/aws/emr/emrfs/conf:/usr/share/aws/emr/emrfs/lib/*
>>>>>>>>>>> :/usr/share/aws/emr/emrfs/auxlib/*"
>>>>>>>>>>> export JAVA_HOME=/usr/lib/jvm/java-1.8.0
>>>>>>>>>>> export ZEPPELIN_NOTEBOOK_S3_BUCKET=mybucket
>>>>>>>>>>> export ZEPPELIN_NOTEBOOK_S3_USER=zeppelin
>>>>>>>>>>> export
>>>>>>>>>>> ZEPPELIN_NOTEBOOK_STORAGE=org.apache.zeppelin.notebook.repo.S3NotebookRepo
>>>>>>>>>>>
>>>>>>>>>>> spark-defaults.conf:
>>>>>>>>>>>
>>>>>>>>>>> spark.master                     yarn
>>>>>>>>>>> spark.driver.extraClassPath
>>>>>>>>>>>  
>>>>>>>>>>> /etc/hadoop/conf:/etc/hive/conf:/usr/lib/hadoop-lzo/lib/*:/usr/lib/hadoop/hadoop-aws.jar:/usr/share/aws/aws-java-sdk/*:/usr/share/aws/emr/emrfs/conf
>>>>>>>>>>> :/usr/share/aws/emr/emrfs/lib/*:/usr/share/aws/emr/emrfs/auxlib/*
>>>>>>>>>>> spark.driver.extraLibraryPath
>>>>>>>>>>>  /usr/lib/hadoop/lib/native:/usr/lib/hadoop-lzo/lib/native
>>>>>>>>>>> spark.executor.extraClassPath
>>>>>>>>>>>  
>>>>>>>>>>> /etc/hadoop/conf:/etc/hive/conf:/usr/lib/hadoop-lzo/lib/*:/usr/lib/hadoop/hadoop-aws.jar:/usr/share/aws/aws-java-sdk/*:/usr/share/aws/emr/emrfs/conf
>>>>>>>>>>> :/usr/share/aws/emr/emrfs/lib/*:/usr/share/aws/emr/emrfs/auxlib/*
>>>>>>>>>>> spark.executor.extraLibraryPath
>>>>>>>>>>>  /usr/lib/hadoop/lib/native:/usr/lib/hadoop-lzo/lib/native
>>>>>>>>>>> spark.eventLog.enabled           true
>>>>>>>>>>> spark.eventLog.dir               hdfs:///var/log/spark/apps
>>>>>>>>>>> spark.history.fs.logDirectory    hdfs:///var/log/spark/apps
>>>>>>>>>>> spark.yarn.historyServer.address
>>>>>>>>>>> ip-172-30-54-30.ec2.internal:18080
>>>>>>>>>>> spark.history.ui.port            18080
>>>>>>>>>>> spark.shuffle.service.enabled    true
>>>>>>>>>>> spark.driver.extraJavaOptions
>>>>>>>>>>>  -Dlog4j.configuration=file:///etc/spark/conf/log4j.properties
>>>>>>>>>>> -XX:+UseConcMarkSweepGC -XX:CMSInitiatingOccupancyFraction=70
>>>>>>>>>>> -XX:MaxHeapFreeRatio=70
>>>>>>>>>>> -XX:+CMSClassUnloadingEnabled -XX:MaxPermSize=512M
>>>>>>>>>>> -XX:OnOutOfMemoryError='kill -9 %p'
>>>>>>>>>>> spark.dynamicAllocation.enabled  true
>>>>>>>>>>> spark.executor.extraJavaOptions  -verbose:gc -XX:+PrintGCDetails
>>>>>>>>>>> -XX:+PrintGCDateStamps -XX:+UseConcMarkSweepGC
>>>>>>>>>>> -XX:CMSInitiatingOccupancyFraction=70 -XX:MaxHeapFreeRatio=70 
>>>>>>>>>>> -XX:+CM
>>>>>>>>>>> SClassUnloadingEnabled -XX:OnOutOfMemoryError='kill -9 %p'
>>>>>>>>>>> spark.executor.memory            8640m
>>>>>>>>>>> spark.executor.cores             7
>>>>>>>>>>> spark.authenticate.enableSaslEncryption true
>>>>>>>>>>> spark.driver.memory              1g
>>>>>>>>>>> spark.network.sasl.serverAlwaysEncrypt true
>>>>>>>>>>> spark.driver.cores               1
>>>>>>>>>>> spark.ssl.protocol               TLSv1.2
>>>>>>>>>>> spark.ssl.keyStorePassword       password
>>>>>>>>>>> spark.yarn.maxAppAttempts        1
>>>>>>>>>>> spark.ssl.keyStore               /etc/emr/security/keystore.jks
>>>>>>>>>>> spark.authenticate               true
>>>>>>>>>>> spark.ssl.keyPassword            password
>>>>>>>>>>> spark.ssl.enabled                true
>>>>>>>>>>> spark.ssl.enabledAlgorithms      TLS_RSA_WITH_AES_256_CBC_SHA
>>>>>>>>>>> spark.ssl.trustStore             /etc/emr/security/truststore.jks
>>>>>>>>>>> spark.authenticate.secret        secret
>>>>>>>>>>> spark.ssl.trustStorePassword     password
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> On Mon, Jun 27, 2016 at 7:33 PM, Hyung Sung Shim <
>>>>>>>>>>> hss...@nflabs.com> wrote:
>>>>>>>>>>>
>>>>>>>>>>>> Hi.
>>>>>>>>>>>> Could you share your conf/zeppelin-env.sh
>>>>>>>>>>>> and spark-defaults.conf ?
>>>>>>>>>>>>
>>>>>>>>>>>> 2016-06-28 8:52 GMT+09:00 Jonathan Esterhazy <
>>>>>>>>>>>> jonathan.esterh...@gmail.com>:
>>>>>>>>>>>>
>>>>>>>>>>>>> I am having trouble using zeppelin in a spark cluster that has
>>>>>>>>>>>>> spark node authentication turned on (e.g. with 
>>>>>>>>>>>>> spark.authenticate=true,
>>>>>>>>>>>>> spark.authenticate.secret=...)
>>>>>>>>>>>>>
>>>>>>>>>>>>> Notebook code that calls built-in spark functions (or other
>>>>>>>>>>>>> things on executor classpath) work fine, but functions defined in 
>>>>>>>>>>>>> the
>>>>>>>>>>>>> notebook (anonymous or named) throw ClassNotFoundExceptions when 
>>>>>>>>>>>>> called
>>>>>>>>>>>>> from an executor.
>>>>>>>>>>>>>
>>>>>>>>>>>>> For example, this code works:
>>>>>>>>>>>>>
>>>>>>>>>>>>> val rdd = sc.textFile("hdfs://my-text-file")
>>>>>>>>>>>>> rdd.take(1).foreach(println)
>>>>>>>>>>>>>
>>>>>>>>>>>>> rdd.saveAsTextFile("hdfs:///my-other-text-file")
>>>>>>>>>>>>>
>>>>>>>>>>>>> but code like this...
>>>>>>>>>>>>>
>>>>>>>>>>>>> rdd.filter(_.contains("my data"))
>>>>>>>>>>>>>
>>>>>>>>>>>>> fails with
>>>>>>>>>>>>>
>>>>>>>>>>>>> Caused by: java.lang.ClassNotFoundException:
>>>>>>>>>>>>> $iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$anonfun$1
>>>>>>>>>>>>> at
>>>>>>>>>>>>> org.apache.spark.repl.ExecutorClassLoader.findClass(ExecutorClassLoader.scala:84)
>>>>>>>>>>>>> at java.lang.ClassLoader.loadClass(ClassLoader.java:424)
>>>>>>>>>>>>> at java.lang.ClassLoader.loadClass(ClassLoader.java:357)
>>>>>>>>>>>>> at java.lang.Class.forName0(Native Method)
>>>>>>>>>>>>> at java.lang.Class.forName(Class.java:348)
>>>>>>>>>>>>>         ...
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>> I get the same kind of error if the filter function is defined
>>>>>>>>>>>>> as a named function in the notebook, or as a member of singleton 
>>>>>>>>>>>>> object
>>>>>>>>>>>>> defined in the notebook.
>>>>>>>>>>>>>
>>>>>>>>>>>>> When I look at the executor's log output, I see this error:
>>>>>>>>>>>>>
>>>>>>>>>>>>> 16/06/27 21:36:23 ERROR repl.ExecutorClassLoader: Failed to
>>>>>>>>>>>>> check existence of class
>>>>>>>>>>>>> $line31.$read$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$anonfun$1 
>>>>>>>>>>>>> on REPL
>>>>>>>>>>>>> class server at https://172.30.54.30:34980
>>>>>>>>>>>>> java.lang.NullPointerException
>>>>>>>>>>>>> at
>>>>>>>>>>>>> org.apache.spark.repl.ExecutorClassLoader.getClassFileInputStreamFromHttpServer(ExecutorClassLoader.scala:113)
>>>>>>>>>>>>> at
>>>>>>>>>>>>> org.apache.spark.repl.ExecutorClassLoader.findClassLocally(ExecutorClassLoader.scala:146)
>>>>>>>>>>>>> at
>>>>>>>>>>>>> org.apache.spark.repl.ExecutorClassLoader.findClass(ExecutorClassLoader.scala:76)
>>>>>>>>>>>>> at java.lang.ClassLoader.loadClass(ClassLoader.java:424)
>>>>>>>>>>>>> at java.lang.ClassLoader.loadClass(ClassLoader.java:357)
>>>>>>>>>>>>> at java.lang.Class.forName0(Native Method)
>>>>>>>>>>>>> at java.lang.Class.forName(Class.java:348)
>>>>>>>>>>>>>
>>>>>>>>>>>>>         ...
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>> If I disable spark authentication, everything works as
>>>>>>>>>>>>> expected. I am running zeppelin 0.5.6 on spark 1.6.1 with yarn.
>>>>>>>>>>>>>
>>>>>>>>>>>>> Has anyone been able to get zeppelin working with spark
>>>>>>>>>>>>> authentication?
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>
>>>>>>>>
>>>>>>>
>>>>>>
>>>>>
>>>>
>>>
>>
>


-- 
Best Regards

Jeff Zhang

Reply via email to