It just a binding issue with the hostnames in your /etc/hosts file. You can set SPARK_LOCAL_IP and SPARK_MASTER_IP in your conf/spark-env.sh file and restart your cluster. (in that case the spark://myworkstation:7077 will change to the ip address that you provided eg: spark://10.211.55.3).
Thanks Best Regards On Tue, Jan 13, 2015 at 11:15 PM, jeremy p <athomewithagroove...@gmail.com> wrote: > Hello all, > > I wrote some Java code that uses Spark, but for some reason I can't run it > from the command line. I am running Spark on a single node (my > workstation). The program stops running after this line is executed : > > SparkContext sparkContext = new SparkContext("spark://myworkstation:7077", > "sparkbase"); > > When that line is executed, this is printed to the screen : > 15/01/12 15:56:19 WARN util.Utils: Your hostname, myworkstation resolves > to a loopback address: 127.0.1.1; using 10.211.55.3 instead (on interface > eth0) > 15/01/12 15:56:19 WARN util.Utils: Set SPARK_LOCAL_IP if you need to bind > to another address > 15/01/12 15:56:19 INFO spark.SecurityManager: Changing view acls to: > myusername > 15/01/12 15:56:19 INFO spark.SecurityManager: Changing modify acls to: > myusername > 15/01/12 15:56:19 INFO spark.SecurityManager: SecurityManager: > authentication disabled; ui acls disabled; users with view permissions: > Set(myusername); users with modify permissions: Set(myusername) > > After it writes this to the screen, the program stops executing without > reporting an exception. > > What's odd is that when I run this code from Eclipse, the same lines are > printed to the screen, but the program keeps executing. > > Don't know if it matters, but I'm using the maven assembly plugin, which > includes the dependencies in the JAR. > > Here are the versions I'm using : > Cloudera : 2.5.0-cdh5.2.1 > Hadoop : 2.5.0-cdh5.2.1 > HBase : HBase 0.98.6-cdh5.2.1 > Java : 1.7.0_65 > Ubuntu : 14.04.1 LTS > Spark : 1.2 >