On starting spark-shell I see this just before the scala prompt: WARN : Your hostname, BloomBear-SSD resolves to a loopback/non-reachable address: fe80:0:0:0:0:5efe:c0a8:317%net10, but we couldn't find any external IP address!
I get this error even when firewall is disabled. I also tried setting the environment variable SPARK_IP_LOCAL to various choices listed below: SPARK_LOCAL_IP=localhost SPARK_LOCAL_IP=127.0.0.1 SPARK_LOCAL_IP=192.168.1.88 (my local machine's IPv4 address) SPARK_LOCAL_IP=fe80::eda5:a1a7:be1e:13cb%14 (my local machine's IPv6 address) I still get this annoying error! How can I resolve this? See below for my environment Environment windows 7 64 bit Spark 1.5.2 Scala 2.10.6 Python 2.7.10 (from Anaconda) PATH includes: C:\Users\Stefan\spark-1.5.2-bin-hadoop2.6\bin C:\ProgramData\Oracle\Java\javapath C:\Users\Stefan\scala\bin C:\Users\Stefan\hadoop-2.6.0\bin (where the bin\winutils resides) C:\ProgramData\Oracle\Java\javapath SYSTEM variables set are: SPARK_HOME=C:\Users\Stefan\spark-1.5.2-bin-hadoop2.6 JAVA_HOME=C:\Program Files\Java\jre1.8.0_65 HADOOP_HOME=C:\Users\Stefan\hadoop-2.6.0 \tmp\hive directory at root on C; drive with full permissions, e.g. >winutils ls \tmp\hive drwxrwxrwx 1 BloomBear-SSD\Stefan BloomBear-SSD\None 0 Dec 8 2015 \tmp\hive