TitusFong created ZEPPELIN-2692:
-----------------------------------
Summary: pyspark die on second run error
Key: ZEPPELIN-2692
URL: https://issues.apache.org/jira/browse/ZEPPELIN-2692
Project: Zeppelin
Issue Type: Bug
Components: pySpark, zeppelin-interpreter
Affects Versions: 0.7.1
Environment: mac, spark 2.1.1
Reporter: TitusFong
I ran pyspark code the first time it was fine, the second time it dies and show
this on every single cell of my zeppelin notebook, and also other notebook that
I am running with pyspark, I have to turn on Zeppelin code and restart in order
to fix this.
Traceback (most recent call last):
File
"/var/folders/zh/dvdnf74d1t9cq78hjjm3xft80000gn/T/zeppelin_pyspark-1462033700144752464.py",
line 343, in <module>
sc.setJobGroup(jobGroup, "Zeppelin")
File "/Users/titusfong/spark/python/pyspark/context.py", line 902, in
setJobGroup
self._jsc.setJobGroup(groupId, description, interruptOnCancel)
AttributeError: 'NoneType' object has no attribute 'setJobGroup'
this error should be similar to this issue which I don't think is solved yet:
http://apache-zeppelin-users-incubating-mailing-list.75479.x6.nabble.com/Error-about-PySpark-td4988.html
https://issues.apache.org/jira/browse/ZEPPELIN-2449?jql=project%20%3D%20ZEPPELIN%20AND%20text%20~%20%22sc.setJobGroup%22
--
This message was sent by Atlassian JIRA
(v6.4.14#64029)