Mohit, This doesn't seems to be working can you please provide more details? when I use "from pyspark import SparkContext" it is disabled in pycharm. I use pycharm community edition. Where should I set the environment variables in same python script or different python script?
Also, should I run any Spark local cluster so Spark program runs on top of that? Appreciate your help -Sathish On Wed, Aug 6, 2014 at 6:22 PM, Mohit Singh <mohit1...@gmail.com> wrote: > My naive set up.. > Adding > os.environ['SPARK_HOME'] = "/path/to/spark" > sys.path.append("/path/to/spark/python") > on top of my script. > from pyspark import SparkContext > from pyspark import SparkConf > Execution works from within pycharm... > > Though my next step is to figure out autocompletion and I bet there are > better ways to develop apps for spark.. > > > > On Wed, Aug 6, 2014 at 4:16 PM, Sathish Kumaran Vairavelu < > vsathishkuma...@gmail.com> wrote: > >> Hello, >> >> I am trying to use the python IDE PyCharm for Spark application >> development. How can I use pyspark with Python IDE? Can anyone help me with >> this? >> >> >> Thanks >> >> Sathish >> >> >> > > > -- > Mohit > > "When you want success as badly as you want the air, then you will get it. > There is no other secret of success." > -Socrates >