I think it is safe to merge into 0.6.1

On Fri, Jul 29, 2016 at 1:08 AM, Jonathan Esterhazy <
jonathan.esterh...@gmail.com> wrote:

> yes, this PR fixes my classnotfoundexceptions. awesome!
>
> any idea when it might make it into a release?
>
> On Wed, Jul 27, 2016 at 11:25 PM, Jeff Zhang <zjf...@gmail.com> wrote:
>
>> I think it is fixed in https://github.com/apache/zeppelin/pull/1237,
>> please try that.
>>
>> On Thu, Jul 28, 2016 at 1:58 AM, Jonathan Esterhazy <
>> jonathan.esterh...@gmail.com> wrote:
>>
>>> I just tried this with using zeppelin master, and the classnotfound
>>> error is still there.
>>>
>>>
>>> On Thu, Jul 7, 2016 at 7:07 AM, Hyung Sung Shim <hss...@nflabs.com>
>>> wrote:
>>>
>>>> Hi. Jonathan.
>>>> If it is possible, could you use the zeppelin master on the github?
>>>> Zeppelin is working well with spark authentication when i tested.
>>>>
>>>>
>>>>
>>>> 2016-07-02 3:05 GMT+09:00 Jonathan Esterhazy <
>>>> jonathan.esterh...@gmail.com>:
>>>>
>>>>> Hyung, thx for your help. I opened these:
>>>>>
>>>>> https://issues.apache.org/jira/browse/ZEPPELIN-1096 (this scala
>>>>> problem)
>>>>> https://issues.apache.org/jira/browse/ZEPPELIN-1097 (similar looking
>>>>> python problem)
>>>>>
>>>>> LMK if I can provide more info or help in some way.
>>>>>
>>>>> On Fri, Jul 1, 2016 at 5:08 AM, Hyung Sung Shim <hss...@nflabs.com>
>>>>> wrote:
>>>>>
>>>>>> Hi Jonathan.
>>>>>> Unfortunately I got same error in my test bed.
>>>>>> Do you mind create an jira issue for this?
>>>>>>
>>>>>> 2016-07-01 3:19 GMT+09:00 Jonathan Esterhazy <
>>>>>> jonathan.esterh...@gmail.com>:
>>>>>>
>>>>>>> I added this and still get the same exception. The same property is
>>>>>>> also set in spark-defaults.conf.
>>>>>>>
>>>>>>> After that didn't work, I also tried adding --conf
>>>>>>> spark.authenticate=true and --conf
>>>>>>> spark.authenticate.enableSaslEncryption=true, to match the other related
>>>>>>> settings in spark-defaults.conf. Still get the same 
>>>>>>> classnotfoundexception.
>>>>>>>
>>>>>>>
>>>>>>> On Thu, Jun 30, 2016 at 10:45 AM, Hyung Sung Shim <hss...@nflabs.com
>>>>>>> > wrote:
>>>>>>>
>>>>>>>> Please add the *export SPARK_SUBMIT_OPTIONS="--conf
>>>>>>>> spark.authenticate.secret=secret"* in conf/zeppelin-env.sh and
>>>>>>>> restart zeppelin and retry your code.
>>>>>>>>
>>>>>>>>
>>>>>>>> 2016-06-30 23:34 GMT+09:00 Jonathan Esterhazy <
>>>>>>>> jonathan.esterh...@gmail.com>:
>>>>>>>>
>>>>>>>>> yes it does. i only see this problem in zeppelin.
>>>>>>>>>
>>>>>>>>> On Thu, Jun 30, 2016 at 7:05 AM, Hyung Sung Shim <
>>>>>>>>> hss...@nflabs.com> wrote:
>>>>>>>>>
>>>>>>>>>> Hi Jonathan.
>>>>>>>>>> It's not easy to build the test environments but I am working on
>>>>>>>>>> this.
>>>>>>>>>> I have question for you.
>>>>>>>>>> Does your code working well on spark-shell in the
>>>>>>>>>> spark.authenticate mode?
>>>>>>>>>>
>>>>>>>>>> 2016-06-30 22:47 GMT+09:00 Jonathan Esterhazy <
>>>>>>>>>> jonathan.esterh...@gmail.com>:
>>>>>>>>>>
>>>>>>>>>>> Hyung, did you have any luck w/ zeppelin + spark authentication?
>>>>>>>>>>> I'm quite stumped.
>>>>>>>>>>>
>>>>>>>>>>> thx.
>>>>>>>>>>>
>>>>>>>>>>> On Tue, Jun 28, 2016 at 9:11 PM, Hyung Sung Shim <
>>>>>>>>>>> hss...@nflabs.com> wrote:
>>>>>>>>>>>
>>>>>>>>>>>> Thank you.
>>>>>>>>>>>> Let me try.
>>>>>>>>>>>>
>>>>>>>>>>>> 2016-06-28 22:18 GMT+09:00 Jonathan Esterhazy <
>>>>>>>>>>>> jonathan.esterh...@gmail.com>:
>>>>>>>>>>>>
>>>>>>>>>>>>> Hyung,
>>>>>>>>>>>>>
>>>>>>>>>>>>> Yes, here they are.
>>>>>>>>>>>>>
>>>>>>>>>>>>> zeppelin-env.sh:
>>>>>>>>>>>>>
>>>>>>>>>>>>> export ZEPPELIN_PORT=8890
>>>>>>>>>>>>> export ZEPPELIN_CONF_DIR=/etc/zeppelin/conf
>>>>>>>>>>>>> export ZEPPELIN_LOG_DIR=/var/log/zeppelin
>>>>>>>>>>>>> export ZEPPELIN_PID_DIR=/var/run/zeppelin
>>>>>>>>>>>>> export ZEPPELIN_PID=$ZEPPELIN_PID_DIR/zeppelin.pid
>>>>>>>>>>>>> export ZEPPELIN_NOTEBOOK_DIR=/var/lib/zeppelin/notebook
>>>>>>>>>>>>> export ZEPPELIN_WAR_TEMPDIR=/var/run/zeppelin/webapps
>>>>>>>>>>>>> export MASTER=yarn-client
>>>>>>>>>>>>> export SPARK_HOME=/usr/lib/spark
>>>>>>>>>>>>> export HADOOP_CONF_DIR=/etc/hadoop/conf
>>>>>>>>>>>>> export
>>>>>>>>>>>>> CLASSPATH=":/etc/hive/conf:/usr/lib/hadoop-lzo/lib/*:/usr/lib/hadoop/hadoop-aws.jar:/usr/share/aws/aws-java-sdk/*:/usr/share/aws/emr/emrfs/conf:/usr/share/aws/emr/emrfs/lib/*
>>>>>>>>>>>>> :/usr/share/aws/emr/emrfs/auxlib/*"
>>>>>>>>>>>>> export JAVA_HOME=/usr/lib/jvm/java-1.8.0
>>>>>>>>>>>>> export ZEPPELIN_NOTEBOOK_S3_BUCKET=mybucket
>>>>>>>>>>>>> export ZEPPELIN_NOTEBOOK_S3_USER=zeppelin
>>>>>>>>>>>>> export
>>>>>>>>>>>>> ZEPPELIN_NOTEBOOK_STORAGE=org.apache.zeppelin.notebook.repo.S3NotebookRepo
>>>>>>>>>>>>>
>>>>>>>>>>>>> spark-defaults.conf:
>>>>>>>>>>>>>
>>>>>>>>>>>>> spark.master                     yarn
>>>>>>>>>>>>> spark.driver.extraClassPath
>>>>>>>>>>>>>  
>>>>>>>>>>>>> /etc/hadoop/conf:/etc/hive/conf:/usr/lib/hadoop-lzo/lib/*:/usr/lib/hadoop/hadoop-aws.jar:/usr/share/aws/aws-java-sdk/*:/usr/share/aws/emr/emrfs/conf
>>>>>>>>>>>>>
>>>>>>>>>>>>> :/usr/share/aws/emr/emrfs/lib/*:/usr/share/aws/emr/emrfs/auxlib/*
>>>>>>>>>>>>> spark.driver.extraLibraryPath
>>>>>>>>>>>>>  /usr/lib/hadoop/lib/native:/usr/lib/hadoop-lzo/lib/native
>>>>>>>>>>>>> spark.executor.extraClassPath
>>>>>>>>>>>>>  
>>>>>>>>>>>>> /etc/hadoop/conf:/etc/hive/conf:/usr/lib/hadoop-lzo/lib/*:/usr/lib/hadoop/hadoop-aws.jar:/usr/share/aws/aws-java-sdk/*:/usr/share/aws/emr/emrfs/conf
>>>>>>>>>>>>>
>>>>>>>>>>>>> :/usr/share/aws/emr/emrfs/lib/*:/usr/share/aws/emr/emrfs/auxlib/*
>>>>>>>>>>>>> spark.executor.extraLibraryPath
>>>>>>>>>>>>>  /usr/lib/hadoop/lib/native:/usr/lib/hadoop-lzo/lib/native
>>>>>>>>>>>>> spark.eventLog.enabled           true
>>>>>>>>>>>>> spark.eventLog.dir               hdfs:///var/log/spark/apps
>>>>>>>>>>>>> spark.history.fs.logDirectory    hdfs:///var/log/spark/apps
>>>>>>>>>>>>> spark.yarn.historyServer.address
>>>>>>>>>>>>> ip-172-30-54-30.ec2.internal:18080
>>>>>>>>>>>>> spark.history.ui.port            18080
>>>>>>>>>>>>> spark.shuffle.service.enabled    true
>>>>>>>>>>>>> spark.driver.extraJavaOptions
>>>>>>>>>>>>>  -Dlog4j.configuration=file:///etc/spark/conf/log4j.properties
>>>>>>>>>>>>> -XX:+UseConcMarkSweepGC -XX:CMSInitiatingOccupancyFraction=70
>>>>>>>>>>>>> -XX:MaxHeapFreeRatio=70
>>>>>>>>>>>>> -XX:+CMSClassUnloadingEnabled -XX:MaxPermSize=512M
>>>>>>>>>>>>> -XX:OnOutOfMemoryError='kill -9 %p'
>>>>>>>>>>>>> spark.dynamicAllocation.enabled  true
>>>>>>>>>>>>> spark.executor.extraJavaOptions  -verbose:gc
>>>>>>>>>>>>> -XX:+PrintGCDetails -XX:+PrintGCDateStamps -XX:+UseConcMarkSweepGC
>>>>>>>>>>>>> -XX:CMSInitiatingOccupancyFraction=70 -XX:MaxHeapFreeRatio=70 
>>>>>>>>>>>>> -XX:+CM
>>>>>>>>>>>>> SClassUnloadingEnabled -XX:OnOutOfMemoryError='kill -9 %p'
>>>>>>>>>>>>> spark.executor.memory            8640m
>>>>>>>>>>>>> spark.executor.cores             7
>>>>>>>>>>>>> spark.authenticate.enableSaslEncryption true
>>>>>>>>>>>>> spark.driver.memory              1g
>>>>>>>>>>>>> spark.network.sasl.serverAlwaysEncrypt true
>>>>>>>>>>>>> spark.driver.cores               1
>>>>>>>>>>>>> spark.ssl.protocol               TLSv1.2
>>>>>>>>>>>>> spark.ssl.keyStorePassword       password
>>>>>>>>>>>>> spark.yarn.maxAppAttempts        1
>>>>>>>>>>>>> spark.ssl.keyStore               /etc/emr/security/keystore.jks
>>>>>>>>>>>>> spark.authenticate               true
>>>>>>>>>>>>> spark.ssl.keyPassword            password
>>>>>>>>>>>>> spark.ssl.enabled                true
>>>>>>>>>>>>> spark.ssl.enabledAlgorithms      TLS_RSA_WITH_AES_256_CBC_SHA
>>>>>>>>>>>>> spark.ssl.trustStore
>>>>>>>>>>>>> /etc/emr/security/truststore.jks
>>>>>>>>>>>>> spark.authenticate.secret        secret
>>>>>>>>>>>>> spark.ssl.trustStorePassword     password
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>> On Mon, Jun 27, 2016 at 7:33 PM, Hyung Sung Shim <
>>>>>>>>>>>>> hss...@nflabs.com> wrote:
>>>>>>>>>>>>>
>>>>>>>>>>>>>> Hi.
>>>>>>>>>>>>>> Could you share your conf/zeppelin-env.sh
>>>>>>>>>>>>>> and spark-defaults.conf ?
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> 2016-06-28 8:52 GMT+09:00 Jonathan Esterhazy <
>>>>>>>>>>>>>> jonathan.esterh...@gmail.com>:
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> I am having trouble using zeppelin in a spark cluster that
>>>>>>>>>>>>>>> has spark node authentication turned on (e.g. with 
>>>>>>>>>>>>>>> spark.authenticate=true,
>>>>>>>>>>>>>>> spark.authenticate.secret=...)
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> Notebook code that calls built-in spark functions (or other
>>>>>>>>>>>>>>> things on executor classpath) work fine, but functions defined 
>>>>>>>>>>>>>>> in the
>>>>>>>>>>>>>>> notebook (anonymous or named) throw ClassNotFoundExceptions 
>>>>>>>>>>>>>>> when called
>>>>>>>>>>>>>>> from an executor.
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> For example, this code works:
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> val rdd = sc.textFile("hdfs://my-text-file")
>>>>>>>>>>>>>>> rdd.take(1).foreach(println)
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> rdd.saveAsTextFile("hdfs:///my-other-text-file")
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> but code like this...
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> rdd.filter(_.contains("my data"))
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> fails with
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> Caused by: java.lang.ClassNotFoundException:
>>>>>>>>>>>>>>> $iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$anonfun$1
>>>>>>>>>>>>>>> at
>>>>>>>>>>>>>>> org.apache.spark.repl.ExecutorClassLoader.findClass(ExecutorClassLoader.scala:84)
>>>>>>>>>>>>>>> at java.lang.ClassLoader.loadClass(ClassLoader.java:424)
>>>>>>>>>>>>>>> at java.lang.ClassLoader.loadClass(ClassLoader.java:357)
>>>>>>>>>>>>>>> at java.lang.Class.forName0(Native Method)
>>>>>>>>>>>>>>> at java.lang.Class.forName(Class.java:348)
>>>>>>>>>>>>>>>         ...
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> I get the same kind of error if the filter function is
>>>>>>>>>>>>>>> defined as a named function in the notebook, or as a member of 
>>>>>>>>>>>>>>> singleton
>>>>>>>>>>>>>>> object defined in the notebook.
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> When I look at the executor's log output, I see this error:
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> 16/06/27 21:36:23 ERROR repl.ExecutorClassLoader: Failed to
>>>>>>>>>>>>>>> check existence of class
>>>>>>>>>>>>>>> $line31.$read$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$anonfun$1
>>>>>>>>>>>>>>>  on REPL
>>>>>>>>>>>>>>> class server at https://172.30.54.30:34980
>>>>>>>>>>>>>>> java.lang.NullPointerException
>>>>>>>>>>>>>>> at
>>>>>>>>>>>>>>> org.apache.spark.repl.ExecutorClassLoader.getClassFileInputStreamFromHttpServer(ExecutorClassLoader.scala:113)
>>>>>>>>>>>>>>> at
>>>>>>>>>>>>>>> org.apache.spark.repl.ExecutorClassLoader.findClassLocally(ExecutorClassLoader.scala:146)
>>>>>>>>>>>>>>> at
>>>>>>>>>>>>>>> org.apache.spark.repl.ExecutorClassLoader.findClass(ExecutorClassLoader.scala:76)
>>>>>>>>>>>>>>> at java.lang.ClassLoader.loadClass(ClassLoader.java:424)
>>>>>>>>>>>>>>> at java.lang.ClassLoader.loadClass(ClassLoader.java:357)
>>>>>>>>>>>>>>> at java.lang.Class.forName0(Native Method)
>>>>>>>>>>>>>>> at java.lang.Class.forName(Class.java:348)
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>         ...
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> If I disable spark authentication, everything works as
>>>>>>>>>>>>>>> expected. I am running zeppelin 0.5.6 on spark 1.6.1 with yarn.
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> Has anyone been able to get zeppelin working with spark
>>>>>>>>>>>>>>> authentication?
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>
>>>>>>>>
>>>>>>>
>>>>>>
>>>>>
>>>>
>>>
>>
>>
>> --
>> Best Regards
>>
>> Jeff Zhang
>>
>
>


-- 
Best Regards

Jeff Zhang

Reply via email to