I didn't done this.I try and let you know
Regards,
 

     Le Mardi 4 août 2015 18h32, "Vadla, Karthik" <karthik.va...@intel.com> a 
écrit :
   

 #yiv4390282639 #yiv4390282639 -- _filtered #yiv4390282639 {panose-1:2 4 5 3 5 
4 6 3 2 4;} _filtered #yiv4390282639 {font-family:Calibri;panose-1:2 15 5 2 2 2 
4 3 2 4;} _filtered #yiv4390282639 {font-family:inherit;panose-1:0 0 0 0 0 0 0 
0 0 0;}#yiv4390282639 #yiv4390282639 p.yiv4390282639MsoNormal, #yiv4390282639 
li.yiv4390282639MsoNormal, #yiv4390282639 div.yiv4390282639MsoNormal 
{margin:0in;margin-bottom:.0001pt;font-size:12.0pt;}#yiv4390282639 a:link, 
#yiv4390282639 span.yiv4390282639MsoHyperlink 
{color:#0563C1;text-decoration:underline;}#yiv4390282639 a:visited, 
#yiv4390282639 span.yiv4390282639MsoHyperlinkFollowed 
{color:#954F72;text-decoration:underline;}#yiv4390282639 
span.yiv4390282639EmailStyle17 {color:#1F497D;}#yiv4390282639 
span.yiv4390282639crayon-i {}#yiv4390282639 span.yiv4390282639crayon-h 
{}#yiv4390282639 span.yiv4390282639crayon-e {}#yiv4390282639 
span.yiv4390282639crayon-t {}#yiv4390282639 span.yiv4390282639crayon-o 
{}#yiv4390282639 span.yiv4390282639crayon-v {}#yiv4390282639 
span.yiv4390282639crayon-cn {}#yiv4390282639 span.yiv4390282639crayon-sy 
{}#yiv4390282639 .yiv4390282639MsoChpDefault {font-size:10.0pt;} _filtered 
#yiv4390282639 {margin:1.0in 1.0in 1.0in 1.0in;}#yiv4390282639 
div.yiv4390282639WordSection1 {}#yiv4390282639 Hi Clark,    How did you build 
your zeppelin binaries? Did you configure pyspark interpreter manually?       
To configure pyspark automatically while building binaries use below mvn  clean 
package -Pspark-1.3  -Ppyspark  -Dhadoop.version=2.6.0-cdh5.4.2 -Phadoop-2.6 
–DskipTests       try above, still if you face same issue. Let me know.       
Thanks Karthik       From: clark djilo kuissu [mailto:djilokui...@yahoo.fr]
Sent: Tuesday, August 4, 2015 4:49 AM
To: Users; Moon Soo Lee
Subject: Launch Pyspark Interpretor    Hi,         I try to launch the pyspark 
interpretor without success.    I ran the server:

$ bin/zeppelin-daemon.sh start

Running a simple notebook beginning with %pyspark, I got an error about py4j 
not being found. Just did pip install py4j (ref).

Now I'm getting this error:

pyspark is not responding Traceback (most recent call last):
  File "/tmp/zeppelin_pyspark.py", line 22, in <module>
    from pyspark.conf import SparkConf
ImportError: No module named pyspark.conf    So I tried this in my bashrc file 
took from stackoverflow 

SPARK_HOME=/spark
PYTHONPATH=$SPARK_HOME/python:$SPARK_HOME/python/lib/py4j-0.8.2.1-src.zip:$PYTHONPATH
    It didn't work.    What I am suppose to do ?    Regards,    Clark 

  

Reply via email to