yes, this PR fixes my classnotfoundexceptions. awesome!

any idea when it might make it into a release?

On Wed, Jul 27, 2016 at 11:25 PM, Jeff Zhang <zjf...@gmail.com> wrote:

> I think it is fixed in https://github.com/apache/zeppelin/pull/1237,
> please try that.
>
> On Thu, Jul 28, 2016 at 1:58 AM, Jonathan Esterhazy <
> jonathan.esterh...@gmail.com> wrote:
>
>> I just tried this with using zeppelin master, and the classnotfound error
>> is still there.
>>
>>
>> On Thu, Jul 7, 2016 at 7:07 AM, Hyung Sung Shim <hss...@nflabs.com>
>> wrote:
>>
>>> Hi. Jonathan.
>>> If it is possible, could you use the zeppelin master on the github?
>>> Zeppelin is working well with spark authentication when i tested.
>>>
>>>
>>>
>>> 2016-07-02 3:05 GMT+09:00 Jonathan Esterhazy <
>>> jonathan.esterh...@gmail.com>:
>>>
>>>> Hyung, thx for your help. I opened these:
>>>>
>>>> https://issues.apache.org/jira/browse/ZEPPELIN-1096 (this scala
>>>> problem)
>>>> https://issues.apache.org/jira/browse/ZEPPELIN-1097 (similar looking
>>>> python problem)
>>>>
>>>> LMK if I can provide more info or help in some way.
>>>>
>>>> On Fri, Jul 1, 2016 at 5:08 AM, Hyung Sung Shim <hss...@nflabs.com>
>>>> wrote:
>>>>
>>>>> Hi Jonathan.
>>>>> Unfortunately I got same error in my test bed.
>>>>> Do you mind create an jira issue for this?
>>>>>
>>>>> 2016-07-01 3:19 GMT+09:00 Jonathan Esterhazy <
>>>>> jonathan.esterh...@gmail.com>:
>>>>>
>>>>>> I added this and still get the same exception. The same property is
>>>>>> also set in spark-defaults.conf.
>>>>>>
>>>>>> After that didn't work, I also tried adding --conf
>>>>>> spark.authenticate=true and --conf
>>>>>> spark.authenticate.enableSaslEncryption=true, to match the other related
>>>>>> settings in spark-defaults.conf. Still get the same 
>>>>>> classnotfoundexception.
>>>>>>
>>>>>>
>>>>>> On Thu, Jun 30, 2016 at 10:45 AM, Hyung Sung Shim <hss...@nflabs.com>
>>>>>> wrote:
>>>>>>
>>>>>>> Please add the *export SPARK_SUBMIT_OPTIONS="--conf
>>>>>>> spark.authenticate.secret=secret"* in conf/zeppelin-env.sh and
>>>>>>> restart zeppelin and retry your code.
>>>>>>>
>>>>>>>
>>>>>>> 2016-06-30 23:34 GMT+09:00 Jonathan Esterhazy <
>>>>>>> jonathan.esterh...@gmail.com>:
>>>>>>>
>>>>>>>> yes it does. i only see this problem in zeppelin.
>>>>>>>>
>>>>>>>> On Thu, Jun 30, 2016 at 7:05 AM, Hyung Sung Shim <hss...@nflabs.com
>>>>>>>> > wrote:
>>>>>>>>
>>>>>>>>> Hi Jonathan.
>>>>>>>>> It's not easy to build the test environments but I am working on
>>>>>>>>> this.
>>>>>>>>> I have question for you.
>>>>>>>>> Does your code working well on spark-shell in the
>>>>>>>>> spark.authenticate mode?
>>>>>>>>>
>>>>>>>>> 2016-06-30 22:47 GMT+09:00 Jonathan Esterhazy <
>>>>>>>>> jonathan.esterh...@gmail.com>:
>>>>>>>>>
>>>>>>>>>> Hyung, did you have any luck w/ zeppelin + spark authentication?
>>>>>>>>>> I'm quite stumped.
>>>>>>>>>>
>>>>>>>>>> thx.
>>>>>>>>>>
>>>>>>>>>> On Tue, Jun 28, 2016 at 9:11 PM, Hyung Sung Shim <
>>>>>>>>>> hss...@nflabs.com> wrote:
>>>>>>>>>>
>>>>>>>>>>> Thank you.
>>>>>>>>>>> Let me try.
>>>>>>>>>>>
>>>>>>>>>>> 2016-06-28 22:18 GMT+09:00 Jonathan Esterhazy <
>>>>>>>>>>> jonathan.esterh...@gmail.com>:
>>>>>>>>>>>
>>>>>>>>>>>> Hyung,
>>>>>>>>>>>>
>>>>>>>>>>>> Yes, here they are.
>>>>>>>>>>>>
>>>>>>>>>>>> zeppelin-env.sh:
>>>>>>>>>>>>
>>>>>>>>>>>> export ZEPPELIN_PORT=8890
>>>>>>>>>>>> export ZEPPELIN_CONF_DIR=/etc/zeppelin/conf
>>>>>>>>>>>> export ZEPPELIN_LOG_DIR=/var/log/zeppelin
>>>>>>>>>>>> export ZEPPELIN_PID_DIR=/var/run/zeppelin
>>>>>>>>>>>> export ZEPPELIN_PID=$ZEPPELIN_PID_DIR/zeppelin.pid
>>>>>>>>>>>> export ZEPPELIN_NOTEBOOK_DIR=/var/lib/zeppelin/notebook
>>>>>>>>>>>> export ZEPPELIN_WAR_TEMPDIR=/var/run/zeppelin/webapps
>>>>>>>>>>>> export MASTER=yarn-client
>>>>>>>>>>>> export SPARK_HOME=/usr/lib/spark
>>>>>>>>>>>> export HADOOP_CONF_DIR=/etc/hadoop/conf
>>>>>>>>>>>> export
>>>>>>>>>>>> CLASSPATH=":/etc/hive/conf:/usr/lib/hadoop-lzo/lib/*:/usr/lib/hadoop/hadoop-aws.jar:/usr/share/aws/aws-java-sdk/*:/usr/share/aws/emr/emrfs/conf:/usr/share/aws/emr/emrfs/lib/*
>>>>>>>>>>>> :/usr/share/aws/emr/emrfs/auxlib/*"
>>>>>>>>>>>> export JAVA_HOME=/usr/lib/jvm/java-1.8.0
>>>>>>>>>>>> export ZEPPELIN_NOTEBOOK_S3_BUCKET=mybucket
>>>>>>>>>>>> export ZEPPELIN_NOTEBOOK_S3_USER=zeppelin
>>>>>>>>>>>> export
>>>>>>>>>>>> ZEPPELIN_NOTEBOOK_STORAGE=org.apache.zeppelin.notebook.repo.S3NotebookRepo
>>>>>>>>>>>>
>>>>>>>>>>>> spark-defaults.conf:
>>>>>>>>>>>>
>>>>>>>>>>>> spark.master                     yarn
>>>>>>>>>>>> spark.driver.extraClassPath
>>>>>>>>>>>>  
>>>>>>>>>>>> /etc/hadoop/conf:/etc/hive/conf:/usr/lib/hadoop-lzo/lib/*:/usr/lib/hadoop/hadoop-aws.jar:/usr/share/aws/aws-java-sdk/*:/usr/share/aws/emr/emrfs/conf
>>>>>>>>>>>>
>>>>>>>>>>>> :/usr/share/aws/emr/emrfs/lib/*:/usr/share/aws/emr/emrfs/auxlib/*
>>>>>>>>>>>> spark.driver.extraLibraryPath
>>>>>>>>>>>>  /usr/lib/hadoop/lib/native:/usr/lib/hadoop-lzo/lib/native
>>>>>>>>>>>> spark.executor.extraClassPath
>>>>>>>>>>>>  
>>>>>>>>>>>> /etc/hadoop/conf:/etc/hive/conf:/usr/lib/hadoop-lzo/lib/*:/usr/lib/hadoop/hadoop-aws.jar:/usr/share/aws/aws-java-sdk/*:/usr/share/aws/emr/emrfs/conf
>>>>>>>>>>>>
>>>>>>>>>>>> :/usr/share/aws/emr/emrfs/lib/*:/usr/share/aws/emr/emrfs/auxlib/*
>>>>>>>>>>>> spark.executor.extraLibraryPath
>>>>>>>>>>>>  /usr/lib/hadoop/lib/native:/usr/lib/hadoop-lzo/lib/native
>>>>>>>>>>>> spark.eventLog.enabled           true
>>>>>>>>>>>> spark.eventLog.dir               hdfs:///var/log/spark/apps
>>>>>>>>>>>> spark.history.fs.logDirectory    hdfs:///var/log/spark/apps
>>>>>>>>>>>> spark.yarn.historyServer.address
>>>>>>>>>>>> ip-172-30-54-30.ec2.internal:18080
>>>>>>>>>>>> spark.history.ui.port            18080
>>>>>>>>>>>> spark.shuffle.service.enabled    true
>>>>>>>>>>>> spark.driver.extraJavaOptions
>>>>>>>>>>>>  -Dlog4j.configuration=file:///etc/spark/conf/log4j.properties
>>>>>>>>>>>> -XX:+UseConcMarkSweepGC -XX:CMSInitiatingOccupancyFraction=70
>>>>>>>>>>>> -XX:MaxHeapFreeRatio=70
>>>>>>>>>>>> -XX:+CMSClassUnloadingEnabled -XX:MaxPermSize=512M
>>>>>>>>>>>> -XX:OnOutOfMemoryError='kill -9 %p'
>>>>>>>>>>>> spark.dynamicAllocation.enabled  true
>>>>>>>>>>>> spark.executor.extraJavaOptions  -verbose:gc
>>>>>>>>>>>> -XX:+PrintGCDetails -XX:+PrintGCDateStamps -XX:+UseConcMarkSweepGC
>>>>>>>>>>>> -XX:CMSInitiatingOccupancyFraction=70 -XX:MaxHeapFreeRatio=70 
>>>>>>>>>>>> -XX:+CM
>>>>>>>>>>>> SClassUnloadingEnabled -XX:OnOutOfMemoryError='kill -9 %p'
>>>>>>>>>>>> spark.executor.memory            8640m
>>>>>>>>>>>> spark.executor.cores             7
>>>>>>>>>>>> spark.authenticate.enableSaslEncryption true
>>>>>>>>>>>> spark.driver.memory              1g
>>>>>>>>>>>> spark.network.sasl.serverAlwaysEncrypt true
>>>>>>>>>>>> spark.driver.cores               1
>>>>>>>>>>>> spark.ssl.protocol               TLSv1.2
>>>>>>>>>>>> spark.ssl.keyStorePassword       password
>>>>>>>>>>>> spark.yarn.maxAppAttempts        1
>>>>>>>>>>>> spark.ssl.keyStore               /etc/emr/security/keystore.jks
>>>>>>>>>>>> spark.authenticate               true
>>>>>>>>>>>> spark.ssl.keyPassword            password
>>>>>>>>>>>> spark.ssl.enabled                true
>>>>>>>>>>>> spark.ssl.enabledAlgorithms      TLS_RSA_WITH_AES_256_CBC_SHA
>>>>>>>>>>>> spark.ssl.trustStore
>>>>>>>>>>>> /etc/emr/security/truststore.jks
>>>>>>>>>>>> spark.authenticate.secret        secret
>>>>>>>>>>>> spark.ssl.trustStorePassword     password
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> On Mon, Jun 27, 2016 at 7:33 PM, Hyung Sung Shim <
>>>>>>>>>>>> hss...@nflabs.com> wrote:
>>>>>>>>>>>>
>>>>>>>>>>>>> Hi.
>>>>>>>>>>>>> Could you share your conf/zeppelin-env.sh
>>>>>>>>>>>>> and spark-defaults.conf ?
>>>>>>>>>>>>>
>>>>>>>>>>>>> 2016-06-28 8:52 GMT+09:00 Jonathan Esterhazy <
>>>>>>>>>>>>> jonathan.esterh...@gmail.com>:
>>>>>>>>>>>>>
>>>>>>>>>>>>>> I am having trouble using zeppelin in a spark cluster that
>>>>>>>>>>>>>> has spark node authentication turned on (e.g. with 
>>>>>>>>>>>>>> spark.authenticate=true,
>>>>>>>>>>>>>> spark.authenticate.secret=...)
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> Notebook code that calls built-in spark functions (or other
>>>>>>>>>>>>>> things on executor classpath) work fine, but functions defined 
>>>>>>>>>>>>>> in the
>>>>>>>>>>>>>> notebook (anonymous or named) throw ClassNotFoundExceptions when 
>>>>>>>>>>>>>> called
>>>>>>>>>>>>>> from an executor.
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> For example, this code works:
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> val rdd = sc.textFile("hdfs://my-text-file")
>>>>>>>>>>>>>> rdd.take(1).foreach(println)
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> rdd.saveAsTextFile("hdfs:///my-other-text-file")
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> but code like this...
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> rdd.filter(_.contains("my data"))
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> fails with
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> Caused by: java.lang.ClassNotFoundException:
>>>>>>>>>>>>>> $iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$anonfun$1
>>>>>>>>>>>>>> at
>>>>>>>>>>>>>> org.apache.spark.repl.ExecutorClassLoader.findClass(ExecutorClassLoader.scala:84)
>>>>>>>>>>>>>> at java.lang.ClassLoader.loadClass(ClassLoader.java:424)
>>>>>>>>>>>>>> at java.lang.ClassLoader.loadClass(ClassLoader.java:357)
>>>>>>>>>>>>>> at java.lang.Class.forName0(Native Method)
>>>>>>>>>>>>>> at java.lang.Class.forName(Class.java:348)
>>>>>>>>>>>>>>         ...
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> I get the same kind of error if the filter function is
>>>>>>>>>>>>>> defined as a named function in the notebook, or as a member of 
>>>>>>>>>>>>>> singleton
>>>>>>>>>>>>>> object defined in the notebook.
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> When I look at the executor's log output, I see this error:
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> 16/06/27 21:36:23 ERROR repl.ExecutorClassLoader: Failed to
>>>>>>>>>>>>>> check existence of class
>>>>>>>>>>>>>> $line31.$read$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$anonfun$1 
>>>>>>>>>>>>>> on REPL
>>>>>>>>>>>>>> class server at https://172.30.54.30:34980
>>>>>>>>>>>>>> java.lang.NullPointerException
>>>>>>>>>>>>>> at
>>>>>>>>>>>>>> org.apache.spark.repl.ExecutorClassLoader.getClassFileInputStreamFromHttpServer(ExecutorClassLoader.scala:113)
>>>>>>>>>>>>>> at
>>>>>>>>>>>>>> org.apache.spark.repl.ExecutorClassLoader.findClassLocally(ExecutorClassLoader.scala:146)
>>>>>>>>>>>>>> at
>>>>>>>>>>>>>> org.apache.spark.repl.ExecutorClassLoader.findClass(ExecutorClassLoader.scala:76)
>>>>>>>>>>>>>> at java.lang.ClassLoader.loadClass(ClassLoader.java:424)
>>>>>>>>>>>>>> at java.lang.ClassLoader.loadClass(ClassLoader.java:357)
>>>>>>>>>>>>>> at java.lang.Class.forName0(Native Method)
>>>>>>>>>>>>>> at java.lang.Class.forName(Class.java:348)
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>         ...
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> If I disable spark authentication, everything works as
>>>>>>>>>>>>>> expected. I am running zeppelin 0.5.6 on spark 1.6.1 with yarn.
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> Has anyone been able to get zeppelin working with spark
>>>>>>>>>>>>>> authentication?
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>
>>>>>>>>
>>>>>>>
>>>>>>
>>>>>
>>>>
>>>
>>
>
>
> --
> Best Regards
>
> Jeff Zhang
>

Reply via email to