> from pyspark.conf import SparkConf
> ImportError: No module named *pyspark.conf*


William, you probably meant

from pyspark import SparkConf


?


-- 
Ruslan Dautkhanov

On Mon, Mar 20, 2017 at 2:12 PM, William Markito Oliveira <
william.mark...@gmail.com> wrote:

> Ah! Thanks Ruslan! I'm still using 0.7.0 - Let me update to 0.8.0 and
> I'll come back update this thread with the results.
>
> On Mon, Mar 20, 2017 at 3:10 PM, William Markito Oliveira <
> william.mark...@gmail.com> wrote:
>
>> Hi moon, thanks for the tip. Here to summarize my current settings are
>> the following
>>
>> conf/zeppelin-env.sh has only SPARK_HOME setting:
>>
>> export SPARK_HOME=/opt/spark-2.1.0-bin-hadoop2.7/
>>
>> Then on the configuration of the interpreter through the web interface I
>> have:
>>
>> PYSPARK_PYTHON=/opt/miniconda2/envs/myenv/bin/python
>> zeppelin.pyspark.python=python
>>
>> But when I submit from the notebook I'm receiving:  pyspark is not
>> responding
>>
>> And the log file outputs:
>>
>> Traceback (most recent call last): File 
>> "/tmp/zeppelin_pyspark-6480867511995958556.py",
>> line 22, in <module> from pyspark.conf import SparkConf ImportError: No
>> module named pyspark.conf
>>
>> Any thoughts ?  Thanks a lot!
>>
>> On Mon, Mar 20, 2017 at 2:27 PM, moon soo Lee <m...@apache.org> wrote:
>>
>>> When property key in interpreter configuration screen matches certain
>>> condition [1], it'll be treated as a environment variable.
>>>
>>> You can remove PYSPARK_PYTHON from conf/zeppelin-env.sh and place it in
>>> interpreter configuration.
>>>
>>> Thanks,
>>> moon
>>>
>>> [1] https://github.com/apache/zeppelin/blob/master/zeppelin-
>>> interpreter/src/main/java/org/apache/zeppelin/interpreter/re
>>> mote/RemoteInterpreter.java#L152
>>>
>>>
>>> On Mon, Mar 20, 2017 at 12:21 PM William Markito Oliveira <
>>> william.mark...@gmail.com> wrote:
>>>
>>>> Thanks for the quick response Ruslan.
>>>>
>>>> But given that it's an environment variable, I can't quickly change
>>>> that value and point to a different python environment without restarting
>>>> the Zeppelin process, can I ? I mean is there a way to set the value for
>>>> PYSPARK_PYTHON from the Interpreter configuration screen ?
>>>>
>>>> Thanks,
>>>>
>>>>
>>>> On Mon, Mar 20, 2017 at 2:15 PM, Ruslan Dautkhanov <
>>>> dautkha...@gmail.com> wrote:
>>>>
>>>> You can set PYSPARK_PYTHON environment variable for that.
>>>>
>>>> Not sure about zeppelin.pyspark.python. I think it does not work
>>>> See comments in https://issues.apache.org/jira/browse/ZEPPELIN-1265
>>>>
>>>> Eventually, i think we can remove zeppelin.pyspark.python and use only
>>>> PYSPARK_PYTHON instead to avoid confusion.
>>>>
>>>>
>>>> --
>>>> Ruslan Dautkhanov
>>>>
>>>> On Mon, Mar 20, 2017 at 12:59 PM, William Markito Oliveira <
>>>> mark...@apache.org> wrote:
>>>>
>>>> I'm trying to use zeppelin.pyspark.python as the variable to set the
>>>> python that Spark worker nodes should use for my job, but it doesn't seem
>>>> to be working.
>>>>
>>>> Am I missing something or this variable does not do that ?
>>>>
>>>> My goal is to change that variable to point to different conda
>>>> environments.  These environments are available in all worker nodes since
>>>> it's on a shared location and ideally all nodes then would have access to
>>>> the same libraries and dependencies.
>>>>
>>>> Thanks,
>>>>
>>>> ~/William
>>>>
>>>>
>>>>
>>>>
>>>>
>>>> --
>>>> ~/William
>>>>
>>>
>>
>>
>> --
>> ~/William
>>
>
>
>
> --
> ~/William
>

Reply via email to