Is this command working?? java -cp ::/usr/local/spark-1.0.0/conf:/usr/local/spark-1.0.0/ assembly/target/scala-2.10/spark-assembly-1.0.0-hadoop1.2.1.jar -XX:MaxPermSize=128m -Dspark.akka.logLifecycleEvents=true -Xms512m -Xmx512m org.apache.spark.deploy.worker.Worker spark://x.x.x.174:7077
Thanks Best Regards On Tue, Jul 1, 2014 at 6:08 PM, MEETHU MATHEW <meethu2...@yahoo.co.in> wrote: > > Hi , > > I am using Spark Standalone mode with one master and 2 slaves.I am not > able to start the workers and connect it to the master using > > ./bin/spark-class org.apache.spark.deploy.worker.Worker > spark://x.x.x.174:7077 > > The log says > > Exception in thread "main" org.jboss.netty.channel.ChannelException: > Failed to bind to: master/x.x.x.174:0 > at > org.jboss.netty.bootstrap.ServerBootstrap.bind(ServerBootstrap.java:272) > ... > Caused by: java.net.BindException: Cannot assign requested address > > When I try to start the worker from the slaves using the following java > command,its running without any exception > > java -cp > ::/usr/local/spark-1.0.0/conf:/usr/local/spark-1.0.0/assembly/target/scala-2.10/spark-assembly-1.0.0-hadoop1.2.1.jar > -XX:MaxPermSize=128m -Dspark.akka.logLifecycleEvents=true -Xms512m -Xmx512m > org.apache.spark.deploy.worker.Worker spark://:master:7077 > > > > > Thanks & Regards, > Meethu M >