That’s very strange. I just un-set my SPARK_HOME env param, downloaded a fresh 1.6.0 tarball, unzipped it to local dir (~/Downloads), and it ran just fine - the driver port is some randomly generated large number. So SPARK_HOME is definitely not needed to run this.
Aida, you are not running this as the super-user, are you? What versions of Java & Scala do you have installed? > On Mar 9, 2016, at 3:53 PM, Aida Tefera <aida1.tef...@gmail.com> wrote: > > Hi Jakob, > > Tried running the command env|grep SPARK; nothing comes back > > Tried env|grep Spark; which is the directory I created for Spark once I > downloaded the tgz file; comes back with PWD=/Users/aidatefera/Spark > > Tried running ./bin/spark-shell ; comes back with same error as below; i.e > could not bind to port 0 etc. > > Sent from my iPhone > >> On 9 Mar 2016, at 21:42, Jakob Odersky <ja...@odersky.com> wrote: >> >> As Tristan mentioned, it looks as though Spark is trying to bind on >> port 0 and then 1 (which is not allowed). Could it be that some >> environment variables from you previous installation attempts are >> polluting your configuration? >> What does running "env | grep SPARK" show you? >> >> Also, try running just "/bin/spark-shell" (without the --master >> argument), maybe your shell is doing some funky stuff with the >> brackets. > > --------------------------------------------------------------------- > To unsubscribe, e-mail: user-unsubscr...@spark.apache.org > For additional commands, e-mail: user-h...@spark.apache.org > --------------------------------------------------------------------- To unsubscribe, e-mail: user-unsubscr...@spark.apache.org For additional commands, e-mail: user-h...@spark.apache.org