I think it is fixed in https://github.com/apache/zeppelin/pull/1237, please try that.
On Thu, Jul 28, 2016 at 1:58 AM, Jonathan Esterhazy < jonathan.esterh...@gmail.com> wrote: > I just tried this with using zeppelin master, and the classnotfound error > is still there. > > > On Thu, Jul 7, 2016 at 7:07 AM, Hyung Sung Shim <hss...@nflabs.com> wrote: > >> Hi. Jonathan. >> If it is possible, could you use the zeppelin master on the github? >> Zeppelin is working well with spark authentication when i tested. >> >> >> >> 2016-07-02 3:05 GMT+09:00 Jonathan Esterhazy < >> jonathan.esterh...@gmail.com>: >> >>> Hyung, thx for your help. I opened these: >>> >>> https://issues.apache.org/jira/browse/ZEPPELIN-1096 (this scala problem) >>> https://issues.apache.org/jira/browse/ZEPPELIN-1097 (similar looking >>> python problem) >>> >>> LMK if I can provide more info or help in some way. >>> >>> On Fri, Jul 1, 2016 at 5:08 AM, Hyung Sung Shim <hss...@nflabs.com> >>> wrote: >>> >>>> Hi Jonathan. >>>> Unfortunately I got same error in my test bed. >>>> Do you mind create an jira issue for this? >>>> >>>> 2016-07-01 3:19 GMT+09:00 Jonathan Esterhazy < >>>> jonathan.esterh...@gmail.com>: >>>> >>>>> I added this and still get the same exception. The same property is >>>>> also set in spark-defaults.conf. >>>>> >>>>> After that didn't work, I also tried adding --conf >>>>> spark.authenticate=true and --conf >>>>> spark.authenticate.enableSaslEncryption=true, to match the other related >>>>> settings in spark-defaults.conf. Still get the same >>>>> classnotfoundexception. >>>>> >>>>> >>>>> On Thu, Jun 30, 2016 at 10:45 AM, Hyung Sung Shim <hss...@nflabs.com> >>>>> wrote: >>>>> >>>>>> Please add the *export SPARK_SUBMIT_OPTIONS="--conf >>>>>> spark.authenticate.secret=secret"* in conf/zeppelin-env.sh and >>>>>> restart zeppelin and retry your code. >>>>>> >>>>>> >>>>>> 2016-06-30 23:34 GMT+09:00 Jonathan Esterhazy < >>>>>> jonathan.esterh...@gmail.com>: >>>>>> >>>>>>> yes it does. i only see this problem in zeppelin. >>>>>>> >>>>>>> On Thu, Jun 30, 2016 at 7:05 AM, Hyung Sung Shim <hss...@nflabs.com> >>>>>>> wrote: >>>>>>> >>>>>>>> Hi Jonathan. >>>>>>>> It's not easy to build the test environments but I am working on >>>>>>>> this. >>>>>>>> I have question for you. >>>>>>>> Does your code working well on spark-shell in the >>>>>>>> spark.authenticate mode? >>>>>>>> >>>>>>>> 2016-06-30 22:47 GMT+09:00 Jonathan Esterhazy < >>>>>>>> jonathan.esterh...@gmail.com>: >>>>>>>> >>>>>>>>> Hyung, did you have any luck w/ zeppelin + spark authentication? >>>>>>>>> I'm quite stumped. >>>>>>>>> >>>>>>>>> thx. >>>>>>>>> >>>>>>>>> On Tue, Jun 28, 2016 at 9:11 PM, Hyung Sung Shim < >>>>>>>>> hss...@nflabs.com> wrote: >>>>>>>>> >>>>>>>>>> Thank you. >>>>>>>>>> Let me try. >>>>>>>>>> >>>>>>>>>> 2016-06-28 22:18 GMT+09:00 Jonathan Esterhazy < >>>>>>>>>> jonathan.esterh...@gmail.com>: >>>>>>>>>> >>>>>>>>>>> Hyung, >>>>>>>>>>> >>>>>>>>>>> Yes, here they are. >>>>>>>>>>> >>>>>>>>>>> zeppelin-env.sh: >>>>>>>>>>> >>>>>>>>>>> export ZEPPELIN_PORT=8890 >>>>>>>>>>> export ZEPPELIN_CONF_DIR=/etc/zeppelin/conf >>>>>>>>>>> export ZEPPELIN_LOG_DIR=/var/log/zeppelin >>>>>>>>>>> export ZEPPELIN_PID_DIR=/var/run/zeppelin >>>>>>>>>>> export ZEPPELIN_PID=$ZEPPELIN_PID_DIR/zeppelin.pid >>>>>>>>>>> export ZEPPELIN_NOTEBOOK_DIR=/var/lib/zeppelin/notebook >>>>>>>>>>> export ZEPPELIN_WAR_TEMPDIR=/var/run/zeppelin/webapps >>>>>>>>>>> export MASTER=yarn-client >>>>>>>>>>> export SPARK_HOME=/usr/lib/spark >>>>>>>>>>> export HADOOP_CONF_DIR=/etc/hadoop/conf >>>>>>>>>>> export >>>>>>>>>>> CLASSPATH=":/etc/hive/conf:/usr/lib/hadoop-lzo/lib/*:/usr/lib/hadoop/hadoop-aws.jar:/usr/share/aws/aws-java-sdk/*:/usr/share/aws/emr/emrfs/conf:/usr/share/aws/emr/emrfs/lib/* >>>>>>>>>>> :/usr/share/aws/emr/emrfs/auxlib/*" >>>>>>>>>>> export JAVA_HOME=/usr/lib/jvm/java-1.8.0 >>>>>>>>>>> export ZEPPELIN_NOTEBOOK_S3_BUCKET=mybucket >>>>>>>>>>> export ZEPPELIN_NOTEBOOK_S3_USER=zeppelin >>>>>>>>>>> export >>>>>>>>>>> ZEPPELIN_NOTEBOOK_STORAGE=org.apache.zeppelin.notebook.repo.S3NotebookRepo >>>>>>>>>>> >>>>>>>>>>> spark-defaults.conf: >>>>>>>>>>> >>>>>>>>>>> spark.master yarn >>>>>>>>>>> spark.driver.extraClassPath >>>>>>>>>>> >>>>>>>>>>> /etc/hadoop/conf:/etc/hive/conf:/usr/lib/hadoop-lzo/lib/*:/usr/lib/hadoop/hadoop-aws.jar:/usr/share/aws/aws-java-sdk/*:/usr/share/aws/emr/emrfs/conf >>>>>>>>>>> :/usr/share/aws/emr/emrfs/lib/*:/usr/share/aws/emr/emrfs/auxlib/* >>>>>>>>>>> spark.driver.extraLibraryPath >>>>>>>>>>> /usr/lib/hadoop/lib/native:/usr/lib/hadoop-lzo/lib/native >>>>>>>>>>> spark.executor.extraClassPath >>>>>>>>>>> >>>>>>>>>>> /etc/hadoop/conf:/etc/hive/conf:/usr/lib/hadoop-lzo/lib/*:/usr/lib/hadoop/hadoop-aws.jar:/usr/share/aws/aws-java-sdk/*:/usr/share/aws/emr/emrfs/conf >>>>>>>>>>> :/usr/share/aws/emr/emrfs/lib/*:/usr/share/aws/emr/emrfs/auxlib/* >>>>>>>>>>> spark.executor.extraLibraryPath >>>>>>>>>>> /usr/lib/hadoop/lib/native:/usr/lib/hadoop-lzo/lib/native >>>>>>>>>>> spark.eventLog.enabled true >>>>>>>>>>> spark.eventLog.dir hdfs:///var/log/spark/apps >>>>>>>>>>> spark.history.fs.logDirectory hdfs:///var/log/spark/apps >>>>>>>>>>> spark.yarn.historyServer.address >>>>>>>>>>> ip-172-30-54-30.ec2.internal:18080 >>>>>>>>>>> spark.history.ui.port 18080 >>>>>>>>>>> spark.shuffle.service.enabled true >>>>>>>>>>> spark.driver.extraJavaOptions >>>>>>>>>>> -Dlog4j.configuration=file:///etc/spark/conf/log4j.properties >>>>>>>>>>> -XX:+UseConcMarkSweepGC -XX:CMSInitiatingOccupancyFraction=70 >>>>>>>>>>> -XX:MaxHeapFreeRatio=70 >>>>>>>>>>> -XX:+CMSClassUnloadingEnabled -XX:MaxPermSize=512M >>>>>>>>>>> -XX:OnOutOfMemoryError='kill -9 %p' >>>>>>>>>>> spark.dynamicAllocation.enabled true >>>>>>>>>>> spark.executor.extraJavaOptions -verbose:gc -XX:+PrintGCDetails >>>>>>>>>>> -XX:+PrintGCDateStamps -XX:+UseConcMarkSweepGC >>>>>>>>>>> -XX:CMSInitiatingOccupancyFraction=70 -XX:MaxHeapFreeRatio=70 >>>>>>>>>>> -XX:+CM >>>>>>>>>>> SClassUnloadingEnabled -XX:OnOutOfMemoryError='kill -9 %p' >>>>>>>>>>> spark.executor.memory 8640m >>>>>>>>>>> spark.executor.cores 7 >>>>>>>>>>> spark.authenticate.enableSaslEncryption true >>>>>>>>>>> spark.driver.memory 1g >>>>>>>>>>> spark.network.sasl.serverAlwaysEncrypt true >>>>>>>>>>> spark.driver.cores 1 >>>>>>>>>>> spark.ssl.protocol TLSv1.2 >>>>>>>>>>> spark.ssl.keyStorePassword password >>>>>>>>>>> spark.yarn.maxAppAttempts 1 >>>>>>>>>>> spark.ssl.keyStore /etc/emr/security/keystore.jks >>>>>>>>>>> spark.authenticate true >>>>>>>>>>> spark.ssl.keyPassword password >>>>>>>>>>> spark.ssl.enabled true >>>>>>>>>>> spark.ssl.enabledAlgorithms TLS_RSA_WITH_AES_256_CBC_SHA >>>>>>>>>>> spark.ssl.trustStore /etc/emr/security/truststore.jks >>>>>>>>>>> spark.authenticate.secret secret >>>>>>>>>>> spark.ssl.trustStorePassword password >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> On Mon, Jun 27, 2016 at 7:33 PM, Hyung Sung Shim < >>>>>>>>>>> hss...@nflabs.com> wrote: >>>>>>>>>>> >>>>>>>>>>>> Hi. >>>>>>>>>>>> Could you share your conf/zeppelin-env.sh >>>>>>>>>>>> and spark-defaults.conf ? >>>>>>>>>>>> >>>>>>>>>>>> 2016-06-28 8:52 GMT+09:00 Jonathan Esterhazy < >>>>>>>>>>>> jonathan.esterh...@gmail.com>: >>>>>>>>>>>> >>>>>>>>>>>>> I am having trouble using zeppelin in a spark cluster that has >>>>>>>>>>>>> spark node authentication turned on (e.g. with >>>>>>>>>>>>> spark.authenticate=true, >>>>>>>>>>>>> spark.authenticate.secret=...) >>>>>>>>>>>>> >>>>>>>>>>>>> Notebook code that calls built-in spark functions (or other >>>>>>>>>>>>> things on executor classpath) work fine, but functions defined in >>>>>>>>>>>>> the >>>>>>>>>>>>> notebook (anonymous or named) throw ClassNotFoundExceptions when >>>>>>>>>>>>> called >>>>>>>>>>>>> from an executor. >>>>>>>>>>>>> >>>>>>>>>>>>> For example, this code works: >>>>>>>>>>>>> >>>>>>>>>>>>> val rdd = sc.textFile("hdfs://my-text-file") >>>>>>>>>>>>> rdd.take(1).foreach(println) >>>>>>>>>>>>> >>>>>>>>>>>>> rdd.saveAsTextFile("hdfs:///my-other-text-file") >>>>>>>>>>>>> >>>>>>>>>>>>> but code like this... >>>>>>>>>>>>> >>>>>>>>>>>>> rdd.filter(_.contains("my data")) >>>>>>>>>>>>> >>>>>>>>>>>>> fails with >>>>>>>>>>>>> >>>>>>>>>>>>> Caused by: java.lang.ClassNotFoundException: >>>>>>>>>>>>> $iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$anonfun$1 >>>>>>>>>>>>> at >>>>>>>>>>>>> org.apache.spark.repl.ExecutorClassLoader.findClass(ExecutorClassLoader.scala:84) >>>>>>>>>>>>> at java.lang.ClassLoader.loadClass(ClassLoader.java:424) >>>>>>>>>>>>> at java.lang.ClassLoader.loadClass(ClassLoader.java:357) >>>>>>>>>>>>> at java.lang.Class.forName0(Native Method) >>>>>>>>>>>>> at java.lang.Class.forName(Class.java:348) >>>>>>>>>>>>> ... >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> I get the same kind of error if the filter function is defined >>>>>>>>>>>>> as a named function in the notebook, or as a member of singleton >>>>>>>>>>>>> object >>>>>>>>>>>>> defined in the notebook. >>>>>>>>>>>>> >>>>>>>>>>>>> When I look at the executor's log output, I see this error: >>>>>>>>>>>>> >>>>>>>>>>>>> 16/06/27 21:36:23 ERROR repl.ExecutorClassLoader: Failed to >>>>>>>>>>>>> check existence of class >>>>>>>>>>>>> $line31.$read$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$anonfun$1 >>>>>>>>>>>>> on REPL >>>>>>>>>>>>> class server at https://172.30.54.30:34980 >>>>>>>>>>>>> java.lang.NullPointerException >>>>>>>>>>>>> at >>>>>>>>>>>>> org.apache.spark.repl.ExecutorClassLoader.getClassFileInputStreamFromHttpServer(ExecutorClassLoader.scala:113) >>>>>>>>>>>>> at >>>>>>>>>>>>> org.apache.spark.repl.ExecutorClassLoader.findClassLocally(ExecutorClassLoader.scala:146) >>>>>>>>>>>>> at >>>>>>>>>>>>> org.apache.spark.repl.ExecutorClassLoader.findClass(ExecutorClassLoader.scala:76) >>>>>>>>>>>>> at java.lang.ClassLoader.loadClass(ClassLoader.java:424) >>>>>>>>>>>>> at java.lang.ClassLoader.loadClass(ClassLoader.java:357) >>>>>>>>>>>>> at java.lang.Class.forName0(Native Method) >>>>>>>>>>>>> at java.lang.Class.forName(Class.java:348) >>>>>>>>>>>>> >>>>>>>>>>>>> ... >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> If I disable spark authentication, everything works as >>>>>>>>>>>>> expected. I am running zeppelin 0.5.6 on spark 1.6.1 with yarn. >>>>>>>>>>>>> >>>>>>>>>>>>> Has anyone been able to get zeppelin working with spark >>>>>>>>>>>>> authentication? >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>> >>>>>>>>> >>>>>>>> >>>>>>> >>>>>> >>>>> >>>> >>> >> > -- Best Regards Jeff Zhang