Hi Jonathan. Unfortunately I got same error in my test bed. Do you mind create an jira issue for this?
2016-07-01 3:19 GMT+09:00 Jonathan Esterhazy <jonathan.esterh...@gmail.com>: > I added this and still get the same exception. The same property is also > set in spark-defaults.conf. > > After that didn't work, I also tried adding --conf spark.authenticate=true > and --conf spark.authenticate.enableSaslEncryption=true, to match the other > related settings in spark-defaults.conf. Still get the same > classnotfoundexception. > > > On Thu, Jun 30, 2016 at 10:45 AM, Hyung Sung Shim <hss...@nflabs.com> > wrote: > >> Please add the *export SPARK_SUBMIT_OPTIONS="--conf >> spark.authenticate.secret=secret"* in conf/zeppelin-env.sh and restart >> zeppelin and retry your code. >> >> >> 2016-06-30 23:34 GMT+09:00 Jonathan Esterhazy < >> jonathan.esterh...@gmail.com>: >> >>> yes it does. i only see this problem in zeppelin. >>> >>> On Thu, Jun 30, 2016 at 7:05 AM, Hyung Sung Shim <hss...@nflabs.com> >>> wrote: >>> >>>> Hi Jonathan. >>>> It's not easy to build the test environments but I am working on this. >>>> I have question for you. >>>> Does your code working well on spark-shell in the spark.authenticate >>>> mode? >>>> >>>> 2016-06-30 22:47 GMT+09:00 Jonathan Esterhazy < >>>> jonathan.esterh...@gmail.com>: >>>> >>>>> Hyung, did you have any luck w/ zeppelin + spark authentication? I'm >>>>> quite stumped. >>>>> >>>>> thx. >>>>> >>>>> On Tue, Jun 28, 2016 at 9:11 PM, Hyung Sung Shim <hss...@nflabs.com> >>>>> wrote: >>>>> >>>>>> Thank you. >>>>>> Let me try. >>>>>> >>>>>> 2016-06-28 22:18 GMT+09:00 Jonathan Esterhazy < >>>>>> jonathan.esterh...@gmail.com>: >>>>>> >>>>>>> Hyung, >>>>>>> >>>>>>> Yes, here they are. >>>>>>> >>>>>>> zeppelin-env.sh: >>>>>>> >>>>>>> export ZEPPELIN_PORT=8890 >>>>>>> export ZEPPELIN_CONF_DIR=/etc/zeppelin/conf >>>>>>> export ZEPPELIN_LOG_DIR=/var/log/zeppelin >>>>>>> export ZEPPELIN_PID_DIR=/var/run/zeppelin >>>>>>> export ZEPPELIN_PID=$ZEPPELIN_PID_DIR/zeppelin.pid >>>>>>> export ZEPPELIN_NOTEBOOK_DIR=/var/lib/zeppelin/notebook >>>>>>> export ZEPPELIN_WAR_TEMPDIR=/var/run/zeppelin/webapps >>>>>>> export MASTER=yarn-client >>>>>>> export SPARK_HOME=/usr/lib/spark >>>>>>> export HADOOP_CONF_DIR=/etc/hadoop/conf >>>>>>> export >>>>>>> CLASSPATH=":/etc/hive/conf:/usr/lib/hadoop-lzo/lib/*:/usr/lib/hadoop/hadoop-aws.jar:/usr/share/aws/aws-java-sdk/*:/usr/share/aws/emr/emrfs/conf:/usr/share/aws/emr/emrfs/lib/* >>>>>>> :/usr/share/aws/emr/emrfs/auxlib/*" >>>>>>> export JAVA_HOME=/usr/lib/jvm/java-1.8.0 >>>>>>> export ZEPPELIN_NOTEBOOK_S3_BUCKET=mybucket >>>>>>> export ZEPPELIN_NOTEBOOK_S3_USER=zeppelin >>>>>>> export >>>>>>> ZEPPELIN_NOTEBOOK_STORAGE=org.apache.zeppelin.notebook.repo.S3NotebookRepo >>>>>>> >>>>>>> spark-defaults.conf: >>>>>>> >>>>>>> spark.master yarn >>>>>>> spark.driver.extraClassPath >>>>>>> >>>>>>> /etc/hadoop/conf:/etc/hive/conf:/usr/lib/hadoop-lzo/lib/*:/usr/lib/hadoop/hadoop-aws.jar:/usr/share/aws/aws-java-sdk/*:/usr/share/aws/emr/emrfs/conf >>>>>>> :/usr/share/aws/emr/emrfs/lib/*:/usr/share/aws/emr/emrfs/auxlib/* >>>>>>> spark.driver.extraLibraryPath >>>>>>> /usr/lib/hadoop/lib/native:/usr/lib/hadoop-lzo/lib/native >>>>>>> spark.executor.extraClassPath >>>>>>> >>>>>>> /etc/hadoop/conf:/etc/hive/conf:/usr/lib/hadoop-lzo/lib/*:/usr/lib/hadoop/hadoop-aws.jar:/usr/share/aws/aws-java-sdk/*:/usr/share/aws/emr/emrfs/conf >>>>>>> :/usr/share/aws/emr/emrfs/lib/*:/usr/share/aws/emr/emrfs/auxlib/* >>>>>>> spark.executor.extraLibraryPath >>>>>>> /usr/lib/hadoop/lib/native:/usr/lib/hadoop-lzo/lib/native >>>>>>> spark.eventLog.enabled true >>>>>>> spark.eventLog.dir hdfs:///var/log/spark/apps >>>>>>> spark.history.fs.logDirectory hdfs:///var/log/spark/apps >>>>>>> spark.yarn.historyServer.address ip-172-30-54-30.ec2.internal:18080 >>>>>>> spark.history.ui.port 18080 >>>>>>> spark.shuffle.service.enabled true >>>>>>> spark.driver.extraJavaOptions >>>>>>> -Dlog4j.configuration=file:///etc/spark/conf/log4j.properties >>>>>>> -XX:+UseConcMarkSweepGC -XX:CMSInitiatingOccupancyFraction=70 >>>>>>> -XX:MaxHeapFreeRatio=70 >>>>>>> -XX:+CMSClassUnloadingEnabled -XX:MaxPermSize=512M >>>>>>> -XX:OnOutOfMemoryError='kill -9 %p' >>>>>>> spark.dynamicAllocation.enabled true >>>>>>> spark.executor.extraJavaOptions -verbose:gc -XX:+PrintGCDetails >>>>>>> -XX:+PrintGCDateStamps -XX:+UseConcMarkSweepGC >>>>>>> -XX:CMSInitiatingOccupancyFraction=70 -XX:MaxHeapFreeRatio=70 -XX:+CM >>>>>>> SClassUnloadingEnabled -XX:OnOutOfMemoryError='kill -9 %p' >>>>>>> spark.executor.memory 8640m >>>>>>> spark.executor.cores 7 >>>>>>> spark.authenticate.enableSaslEncryption true >>>>>>> spark.driver.memory 1g >>>>>>> spark.network.sasl.serverAlwaysEncrypt true >>>>>>> spark.driver.cores 1 >>>>>>> spark.ssl.protocol TLSv1.2 >>>>>>> spark.ssl.keyStorePassword password >>>>>>> spark.yarn.maxAppAttempts 1 >>>>>>> spark.ssl.keyStore /etc/emr/security/keystore.jks >>>>>>> spark.authenticate true >>>>>>> spark.ssl.keyPassword password >>>>>>> spark.ssl.enabled true >>>>>>> spark.ssl.enabledAlgorithms TLS_RSA_WITH_AES_256_CBC_SHA >>>>>>> spark.ssl.trustStore /etc/emr/security/truststore.jks >>>>>>> spark.authenticate.secret secret >>>>>>> spark.ssl.trustStorePassword password >>>>>>> >>>>>>> >>>>>>> >>>>>>> On Mon, Jun 27, 2016 at 7:33 PM, Hyung Sung Shim <hss...@nflabs.com> >>>>>>> wrote: >>>>>>> >>>>>>>> Hi. >>>>>>>> Could you share your conf/zeppelin-env.sh and spark-defaults.conf ? >>>>>>>> >>>>>>>> 2016-06-28 8:52 GMT+09:00 Jonathan Esterhazy < >>>>>>>> jonathan.esterh...@gmail.com>: >>>>>>>> >>>>>>>>> I am having trouble using zeppelin in a spark cluster that has >>>>>>>>> spark node authentication turned on (e.g. with >>>>>>>>> spark.authenticate=true, >>>>>>>>> spark.authenticate.secret=...) >>>>>>>>> >>>>>>>>> Notebook code that calls built-in spark functions (or other things >>>>>>>>> on executor classpath) work fine, but functions defined in the >>>>>>>>> notebook >>>>>>>>> (anonymous or named) throw ClassNotFoundExceptions when called from an >>>>>>>>> executor. >>>>>>>>> >>>>>>>>> For example, this code works: >>>>>>>>> >>>>>>>>> val rdd = sc.textFile("hdfs://my-text-file") >>>>>>>>> rdd.take(1).foreach(println) >>>>>>>>> >>>>>>>>> rdd.saveAsTextFile("hdfs:///my-other-text-file") >>>>>>>>> >>>>>>>>> but code like this... >>>>>>>>> >>>>>>>>> rdd.filter(_.contains("my data")) >>>>>>>>> >>>>>>>>> fails with >>>>>>>>> >>>>>>>>> Caused by: java.lang.ClassNotFoundException: >>>>>>>>> $iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$anonfun$1 >>>>>>>>> at >>>>>>>>> org.apache.spark.repl.ExecutorClassLoader.findClass(ExecutorClassLoader.scala:84) >>>>>>>>> at java.lang.ClassLoader.loadClass(ClassLoader.java:424) >>>>>>>>> at java.lang.ClassLoader.loadClass(ClassLoader.java:357) >>>>>>>>> at java.lang.Class.forName0(Native Method) >>>>>>>>> at java.lang.Class.forName(Class.java:348) >>>>>>>>> ... >>>>>>>>> >>>>>>>>> >>>>>>>>> I get the same kind of error if the filter function is defined as >>>>>>>>> a named function in the notebook, or as a member of singleton object >>>>>>>>> defined in the notebook. >>>>>>>>> >>>>>>>>> When I look at the executor's log output, I see this error: >>>>>>>>> >>>>>>>>> 16/06/27 21:36:23 ERROR repl.ExecutorClassLoader: Failed to check >>>>>>>>> existence of class >>>>>>>>> $line31.$read$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$anonfun$1 on >>>>>>>>> REPL >>>>>>>>> class server at https://172.30.54.30:34980 >>>>>>>>> java.lang.NullPointerException >>>>>>>>> at >>>>>>>>> org.apache.spark.repl.ExecutorClassLoader.getClassFileInputStreamFromHttpServer(ExecutorClassLoader.scala:113) >>>>>>>>> at >>>>>>>>> org.apache.spark.repl.ExecutorClassLoader.findClassLocally(ExecutorClassLoader.scala:146) >>>>>>>>> at >>>>>>>>> org.apache.spark.repl.ExecutorClassLoader.findClass(ExecutorClassLoader.scala:76) >>>>>>>>> at java.lang.ClassLoader.loadClass(ClassLoader.java:424) >>>>>>>>> at java.lang.ClassLoader.loadClass(ClassLoader.java:357) >>>>>>>>> at java.lang.Class.forName0(Native Method) >>>>>>>>> at java.lang.Class.forName(Class.java:348) >>>>>>>>> >>>>>>>>> ... >>>>>>>>> >>>>>>>>> >>>>>>>>> If I disable spark authentication, everything works as expected. I >>>>>>>>> am running zeppelin 0.5.6 on spark 1.6.1 with yarn. >>>>>>>>> >>>>>>>>> Has anyone been able to get zeppelin working with spark >>>>>>>>> authentication? >>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>> >>>>>>> >>>>>> >>>>> >>>> >>> >> >