Spark master by default binds to whatever ip address your current host resolves to. You have a few options to change that: - override the ip by setting the environment variable SPARK_LOCAL_IP - change the ip in your local "hosts" file (/etc/hosts on linux, not sure on windows) - specify a different hostname such as "localhost" when starting spark master by passing the "--host HOSTNAME" command-line parameter (the ip address will be resolved from the supplied HOSTNAME)
best, --Jakob On Mon, Feb 22, 2016 at 5:09 PM, Arko Provo Mukherjee <arkoprovomukher...@gmail.com> wrote: > Hello, > > I am running Spark on Windows. > > I start up master as follows: > .\spark-class.cmd org.apache.spark.deploy.master.Master > > I see that the SparkMaster doesn't start on 127.0.0.1 but starts on my > "actual" IP. This is troublesome for me as I use it in my code and > need to change every time I restart. > > Is there a way to make SparkMaster listen to 127.0.0.1:7077? > > Thanks much in advace! > Warm regards > Arko > > --------------------------------------------------------------------- > To unsubscribe, e-mail: user-unsubscr...@spark.apache.org > For additional commands, e-mail: user-h...@spark.apache.org > --------------------------------------------------------------------- To unsubscribe, e-mail: user-unsubscr...@spark.apache.org For additional commands, e-mail: user-h...@spark.apache.org