Ah! Thanks Ruslan! I'm still using 0.7.0 - Let me update to 0.8.0 and I'll come back update this thread with the results.
On Mon, Mar 20, 2017 at 3:10 PM, William Markito Oliveira < william.mark...@gmail.com> wrote: > Hi moon, thanks for the tip. Here to summarize my current settings are the > following > > conf/zeppelin-env.sh has only SPARK_HOME setting: > > export SPARK_HOME=/opt/spark-2.1.0-bin-hadoop2.7/ > > Then on the configuration of the interpreter through the web interface I > have: > > PYSPARK_PYTHON=/opt/miniconda2/envs/myenv/bin/python > zeppelin.pyspark.python=python > > But when I submit from the notebook I'm receiving: pyspark is not > responding > > And the log file outputs: > > Traceback (most recent call last): File > "/tmp/zeppelin_pyspark-6480867511995958556.py", > line 22, in <module> from pyspark.conf import SparkConf ImportError: No > module named pyspark.conf > > Any thoughts ? Thanks a lot! > > On Mon, Mar 20, 2017 at 2:27 PM, moon soo Lee <m...@apache.org> wrote: > >> When property key in interpreter configuration screen matches certain >> condition [1], it'll be treated as a environment variable. >> >> You can remove PYSPARK_PYTHON from conf/zeppelin-env.sh and place it in >> interpreter configuration. >> >> Thanks, >> moon >> >> [1] https://github.com/apache/zeppelin/blob/master/zeppelin- >> interpreter/src/main/java/org/apache/zeppelin/interpreter/re >> mote/RemoteInterpreter.java#L152 >> >> >> On Mon, Mar 20, 2017 at 12:21 PM William Markito Oliveira < >> william.mark...@gmail.com> wrote: >> >>> Thanks for the quick response Ruslan. >>> >>> But given that it's an environment variable, I can't quickly change that >>> value and point to a different python environment without restarting the >>> Zeppelin process, can I ? I mean is there a way to set the value for >>> PYSPARK_PYTHON from the Interpreter configuration screen ? >>> >>> Thanks, >>> >>> >>> On Mon, Mar 20, 2017 at 2:15 PM, Ruslan Dautkhanov <dautkha...@gmail.com >>> > wrote: >>> >>> You can set PYSPARK_PYTHON environment variable for that. >>> >>> Not sure about zeppelin.pyspark.python. I think it does not work >>> See comments in https://issues.apache.org/jira/browse/ZEPPELIN-1265 >>> >>> Eventually, i think we can remove zeppelin.pyspark.python and use only >>> PYSPARK_PYTHON instead to avoid confusion. >>> >>> >>> -- >>> Ruslan Dautkhanov >>> >>> On Mon, Mar 20, 2017 at 12:59 PM, William Markito Oliveira < >>> mark...@apache.org> wrote: >>> >>> I'm trying to use zeppelin.pyspark.python as the variable to set the >>> python that Spark worker nodes should use for my job, but it doesn't seem >>> to be working. >>> >>> Am I missing something or this variable does not do that ? >>> >>> My goal is to change that variable to point to different conda >>> environments. These environments are available in all worker nodes since >>> it's on a shared location and ideally all nodes then would have access to >>> the same libraries and dependencies. >>> >>> Thanks, >>> >>> ~/William >>> >>> >>> >>> >>> >>> -- >>> ~/William >>> >> > > > -- > ~/William > -- ~/William