Rose Aysina created ZEPPELIN-5928: ------------------------------------- Summary: Spark Interpreter fails with "has no attribute '_wrapped'" in Kubernetes with Zeppelin Key: ZEPPELIN-5928 URL: https://issues.apache.org/jira/browse/ZEPPELIN-5928 Project: Zeppelin Issue Type: Bug Components: Kubernetes, pySpark, spark Affects Versions: 0.10.1 Reporter: Rose Aysina Fix For: 0.11.0 Attachments: Dockerfile, Screenshot 2023-06-09 at 11.39.01 AM.png, Screenshot 2023-06-09 at 11.39.34 AM.png, log.txt
Hola! We discovered bug while deploying Zeppelin on Kubernetes with Spark Interpreter also in Kubernetes mode (also attached full log to this issue): {noformat} INFO [2023-06-09 07:59:49,053] ({FIFOScheduler-interpreter_537057292-Worker-1} PythonInterpreter.java[bootstrapInterpreter]:562) - Bootstrap interpreter via python/zeppelin_pyspark.py ERROR [2023-06-09 07:59:50,416] ({FIFOScheduler-interpreter_537057292-Worker-1} PySparkInterpreter.java[open]:104) - Fail to bootstrap pyspark java.io.IOException: Fail to run bootstrap script: python/zeppelin_pyspark.py %text Fail to execute line 54: sqlc = __zSqlc__ = __zSpark__._wrapped Traceback (most recent call last): File "/tmp/python129188975973677791/zeppelin_python.py", line 162, in <module> exec(code, _zcUserQueryNameSpace) File "<stdin>", line 54, in <module> AttributeError: 'SparkSession' object has no attribute '_wrapped' at org.apache.zeppelin.python.PythonInterpreter.bootstrapInterpreter(PythonInterpreter.java:579) at org.apache.zeppelin.spark.PySparkInterpreter.open(PySparkInterpreter.java:102) at org.apache.zeppelin.interpreter.LazyOpenInterpreter.open(LazyOpenInterpreter.java:70) at org.apache.zeppelin.interpreter.remote.RemoteInterpreterServer$InterpretJob.jobRun(RemoteInterpreterServer.java:844) at org.apache.zeppelin.interpreter.remote.RemoteInterpreterServer$InterpretJob.jobRun(RemoteInterpreterServer.java:752) at org.apache.zeppelin.scheduler.Job.run(Job.java:172) at org.apache.zeppelin.scheduler.AbstractScheduler.runJob(AbstractScheduler.java:132) at org.apache.zeppelin.scheduler.FIFOScheduler.lambda$runJobInScheduler$0(FIFOScheduler.java:42) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:750) INFO [2023-06-09 07:59:50,418] ({FIFOScheduler-interpreter_537057292-Worker-1} PySparkInterpreter.java[close]:112) - Close PySparkInterpreter INFO [2023-06-09 07:59:50,418] ({FIFOScheduler-interpreter_537057292-Worker-1} PythonInterpreter.java[close]:259) - Kill python process INFO [2023-06-09 07:59:50,423] ({FIFOScheduler-interpreter_537057292-Worker-1} AbstractScheduler.java[runJob]:154) - Job 20230605-095507_490419065 finished by scheduler interpreter_537057292 with status ERROR WARN [2023-06-09 07:59:50,425] ({Exec Default Executor} ProcessLauncher.java[onProcessFailed]:134) - Process with cmd [python, /tmp/python129188975973677791/zeppelin_python.py, 10.165.178.231, 41565] is failed due to org.apache.commons.exec.ExecuteException: Process exited with an error: 143 (Exit value: 143) at org.apache.commons.exec.DefaultExecutor.executeInternal(DefaultExecutor.java:404) at org.apache.commons.exec.DefaultExecutor.access$200(DefaultExecutor.java:48) at org.apache.commons.exec.DefaultExecutor$1.run(DefaultExecutor.java:200) at java.lang.Thread.run(Thread.java:750) INFO [2023-06-09 07:59:50,426] ({Exec Default Executor} ProcessLauncher.java[transition]:109) - Process state is transitioned to TERMINATED {noformat} !Screenshot 2023-06-09 at 11.39.01 AM.png|width=383,height=209! ** *Exact problem on Stack Overflow from another user:* [https://stackoverflow.com/q/75949679] How I configure Spark in the notebook in Zeppelin Server (also see screenshots): {code:java} %spark.conf spark.executor.instances 5 spark.kubernetes.container.image.pullSecrets docker-registry spark.app.name SparkTESTSparkTESTSparkTESTSparkTESTSparkTESTSparkTESTSparkTESTSparkTESTSparkTEST spark.jars.ivy /tmp {code} Then Spark Interpreter starts, then it starts spark Executors (all of them in K8S), then raises the error above. Executors don't fail after the error, it's just Zeppelin Server can't connect to Interpreter and prints the same error in the notebook cell. !Screenshot 2023-06-09 at 11.39.34 AM.png|width=383,height=209! *Environment:* * Zeppelin version: *0.10.1* * Spark version: *3.4.0* * Python version: *3.11* * Dockerfile for Zeppelin Interpreter is attached. *What do we do now:* we use version *0.11.0-SNAPSHOT* and build docker images from Zeppelin sources on latest master in GitHub. {color:#de350b}In the newest version there is no such bug!{color} *So the questions are:* - Will patch to similar issues be released? If so, then when? - Can the issue be fixed in the current version? Thank you! Any help will be appreciated. -- This message was sent by Atlassian Jira (v8.20.10#820010)