Try using the --ip parameter <http://spark.apache.org/docs/latest/spark-standalone.html#starting-a-cluster-manually> while starting the worker. like:
spark-1.0.1/bin/spark-class org.apache.spark.deploy.worker.Worker --ip 1.2.3.4 spark://1.2.3.4:7077 Thanks Best Regards On Fri, Oct 24, 2014 at 12:34 PM, Theodore Si <sjyz...@gmail.com> wrote: > Hi all, > > I have two network interface card on one node, one is a Eithernet card, > the other Infiniband HCA. > The master has two IP addresses, lets say 1.2.3.4 (for Eithernet card) and > 2.3.4.5 (for HCA). > I can start the master by > export SPARK_MASTER_IP='1.2.3.4';sbin/start-master.sh > to let master listen on 1.2.3.4:7077 > > But when I connect the worker to the master by using > spark-1.0.1/bin/spark-class org.apache.spark.deploy.worker.Worker spark:// > 1.2.3.4:7077 > > I will get errors, since it is using its HCA card. How can I let the > worker use its Eithernet card? > > Thanks >