Re: problem initiating spark context with pyspark

2017-06-11 Thread Gourav Sengupta
; *Cc:* user@spark.apache.org > *Subject:* Re: problem initiating spark context with pyspark > > Thanks Doc I saw this on another board yesterday so I've tried this by > first going to the directory where I've stored the wintutils.exe and then > as an admin running the command that y

Re: problem initiating spark context with pyspark

2017-06-10 Thread Felix Cheung
: Curtis Burkhalter <curtisburkhal...@gmail.com> Sent: Wednesday, June 7, 2017 7:46:56 AM To: Doc Dwarf Cc: user@spark.apache.org Subject: Re: problem initiating spark context with pyspark Thanks Doc I saw this on another board yesterday so I've tried this by first going to the directory wher

Re: problem initiating spark context with pyspark

2017-06-10 Thread Marco Mistroni
Ha...it's a 1 off.I run spk on Ubuntu and docker on windows...I don't think spark and windows are best friends.  On Jun 10, 2017 6:36 PM, "Gourav Sengupta" wrote: > seeing for the very first time someone try SPARK on Windows :) > > On Thu, Jun 8, 2017 at

Re: problem initiating spark context with pyspark

2017-06-10 Thread Gourav Sengupta
seeing for the very first time someone try SPARK on Windows :) On Thu, Jun 8, 2017 at 8:38 PM, Marco Mistroni wrote: > try this link > > http://letstalkspark.blogspot.co.uk/2016/02/getting-started- > with-spark-on-window-64.html > > it helped me when i had similar problems

Re: problem initiating spark context with pyspark

2017-06-08 Thread Marco Mistroni
try this link http://letstalkspark.blogspot.co.uk/2016/02/getting-started-with-spark-on-window-64.html it helped me when i had similar problems with windows... hth On Wed, Jun 7, 2017 at 3:46 PM, Curtis Burkhalter < curtisburkhal...@gmail.com> wrote: > Thanks Doc I saw this on another

Re: problem initiating spark context with pyspark

2017-06-07 Thread Curtis Burkhalter
Thanks Doc I saw this on another board yesterday so I've tried this by first going to the directory where I've stored the wintutils.exe and then as an admin running the command that you suggested and I get this exception when checking the permissions: C:\winutils\bin>winutils.exe ls -F

Re: problem initiating spark context with pyspark

2017-06-07 Thread Doc Dwarf
Hi Curtis, I believe in windows, the following command needs to be executed: (will need winutils installed) D:\winutils\bin\winutils.exe chmod 777 D:\tmp\hive On 6 June 2017 at 09:45, Curtis Burkhalter wrote: > Hello all, > > I'm new to Spark and I'm trying to

problem initiating spark context with pyspark

2017-06-06 Thread Curtis Burkhalter
Hello all, I'm new to Spark and I'm trying to interact with it using Pyspark. I'm using the prebuilt version of spark v. 2.1.1 and when I go to the command line and use the command 'bin\pyspark' I have initialization problems and get the following message: C:\spark\spark-2.1.1-bin-hadoop2.7>