Re: hadoop2.6.0 + spark1.4.1 + python2.7.10

2015-09-09 Thread Ashish Dutt
Dear Sasha, What I did was that I installed the parcels on all the nodes of the cluster. Typically the location was /opt/cloudera/parcels/CDH5.4.2-1.cdh5.4.2.p0.2 Hope this helps you. With regards, Ashish On Tue, Sep 8, 2015 at 10:18 PM, Sasha Kacanski wrote: > Hi

Re: hadoop2.6.0 + spark1.4.1 + python2.7.10

2015-09-08 Thread Sasha Kacanski
Hi Ashish, Thanks for the update. I tried all of it, but what I don't get it is that I run cluster with one node so presumably I should have PYspark binaries there as I am developing on same host. Could you tell me where you placed parcels or whatever cloudera is using. My understanding of yarn

Re: hadoop2.6.0 + spark1.4.1 + python2.7.10

2015-09-07 Thread Ashish Dutt
Hello Sasha, I have no answer for debian. My cluster is on Linux and I'm using CDH 5.4 Your question- "Error from python worker: /cube/PY/Python27/bin/python: No module named pyspark" On a single node (ie one server/machine/computer) I installed pyspark binaries and it worked. Connected it to

Re: hadoop2.6.0 + spark1.4.1 + python2.7.10

2015-09-06 Thread Ashish Dutt
Hi Aleksandar, Quite some time ago, I faced the same problem and I found a solution which I have posted here on my blog . See if that can help you and if it does not then you can check out these questions & solution on stackoverflow

hadoop2.6.0 + spark1.4.1 + python2.7.10

2015-09-06 Thread Sasha Kacanski
Hi, I am successfully running python app via pyCharm in local mode setMaster("local[*]") When I turn on SparkConf().setMaster("yarn-client") and run via park-submit PysparkPandas.py I run into issue: Error from python worker: /cube/PY/Python27/bin/python: No module named pyspark PYTHONPATH