The problem is resolved.I have added SPARK_LOCAL_IP=master in both slaves also.When i changed this my slaves are working. Thank you all for your suggestions Thanks & Regards, Meethu M
On Wednesday, 2 July 2014 10:22 AM, Aaron Davidson <ilike...@gmail.com> wrote: In your spark-env.sh, do you happen to set SPARK_PUBLIC_DNS or something of that kin? This error suggests the worker is trying to bind a server on the master's IP, which clearly doesn't make sense On Mon, Jun 30, 2014 at 11:59 PM, MEETHU MATHEW <meethu2...@yahoo.co.in> wrote: Hi, > > >I did netstat -na | grep 192.168.125.174 and its showing 192.168.125.174:7077 >LISTEN(after starting master) > > >I tried to execute the following script from the slaves manually but it ends >up with the same exception and log.This script is internally executing the >java command. > /usr/local/spark-1.0.0/sbin/start-slave.sh 1 spark://192.168.125.174:7077 >In this case netstat is showing any connection established to master:7077. > > >When we manually execute the java command,the connection is getting >established to master. > > >Thanks & Regards, >Meethu M > > > >On Monday, 30 June 2014 6:38 PM, Akhil Das <ak...@sigmoidanalytics.com> wrote: > > > >Are you sure you have this ip 192.168.125.174 bind for that machine? (netstat >-na | grep 192.168.125.174) > > >Thanks >Best Regards > > >On Mon, Jun 30, 2014 at 5:34 PM, MEETHU MATHEW <meethu2...@yahoo.co.in> wrote: > >Hi all, >> >> >>I reinstalled spark,reboot the system,but still I am not able to start the >>workers.Its throwing the following exception: >> >> >>Exception in thread "main" org.jboss.netty.channel.ChannelException: Failed >>to bind to: master/192.168.125.174:0 >> >> >>I doubt the problem is with 192.168.125.174:0. Eventhough the command >>contains master:7077,why its showing 0 in the log. >> >> >>java -cp >>::/usr/local/spark-1.0.0/conf:/usr/local/spark-1.0.0/assembly/target/scala-2.10/spark-assembly-1.0.0-hadoop1.2.1.jar >> -XX:MaxPermSize=128m -Dspark.akka.logLifecycleEvents=true -Xms512m -Xmx512m >>org.apache.spark.deploy.worker.Worker spark://master:7077 >> >> >>Can somebody tell me a solution. >> >>Thanks & Regards, >>Meethu M >> >> >> >>On Friday, 27 June 2014 4:28 PM, MEETHU MATHEW <meethu2...@yahoo.co.in> wrote: >> >> >> >>Hi, >>ya I tried setting another PORT also,but the same problem.. >>master is set in etc/hosts >> >>Thanks & Regards, >>Meethu M >> >> >> >>On Friday, 27 June 2014 3:23 PM, Akhil Das <ak...@sigmoidanalytics.com> wrote: >> >> >> >>tha's strange, did you try setting the master port to something else (use >>SPARK_MASTER_PORT). >> >> >>Also you said you are able to start it from the java commandline >> >> >>java -cp >>::/usr/local/spark-1.0.0/conf:/usr/local/spark-1.0.0/assembly/target/scala-2.10/spark-assembly-1.0.0-hadoop1.2.1.jar >> -XX:MaxPermSize=128m -Dspark.akka.logLifecycleEvents=true -Xms512m -Xmx512m >>org.apache.spark.deploy.worker.Worker spark://:master:7077 >> >> >> >>What is the master ip specified here? is it like you have entry for master in >>the /etc/hosts? >> >> >>Thanks >>Best Regards >> >> >>On Fri, Jun 27, 2014 at 3:09 PM, MEETHU MATHEW <meethu2...@yahoo.co.in> wrote: >> >>Hi Akhil, >>> >>> >>>I am running it in a LAN itself..The IP of the master is given correctly. >>> >>>Thanks & Regards, >>>Meethu M >>> >>> >>> >>>On Friday, 27 June 2014 2:51 PM, Akhil Das <ak...@sigmoidanalytics.com> >>>wrote: >>> >>> >>> >>>why is it binding to port 0? 192.168.125.174:0 :/ >>> >>> >>>Check the ip address of that master machine (ifconfig) looks like the ip >>>address has been changed (hoping you are running this machines on a LAN) >>> >>> >>>Thanks >>>Best Regards >>> >>> >>>On Fri, Jun 27, 2014 at 12:00 PM, MEETHU MATHEW <meethu2...@yahoo.co.in> >>>wrote: >>> >>>Hi all, >>>> >>>> >>>>My Spark(Standalone mode) was running fine till yesterday.But now I am >>>>getting the following exeception when I am running start-slaves.sh or >>>>start-all.sh >>>> >>>> >>>>slave3: failed to launch org.apache.spark.deploy.worker.Worker: >>>>slave3: at >>>>java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:918) >>>>slave3: at java.lang.Thread.run(Thread.java:662) >>>> >>>> >>>>The log files has the following lines. >>>> >>>> >>>>14/06/27 11:06:30 INFO SecurityManager: Using Spark's default log4j >>>>profile: org/apache/spark/log4j-defaults.properties >>>>14/06/27 11:06:30 INFO SecurityManager: Changing view acls to: hduser >>>>14/06/27 11:06:30 INFO SecurityManager: SecurityManager: authentication >>>>disabled; ui acls disabled; users with view permissions: Set(hduser) >>>>14/06/27 11:06:30 INFO Slf4jLogger: Slf4jLogger started >>>>14/06/27 11:06:30 INFO Remoting: Starting remoting >>>>Exception in thread "main" org.jboss.netty.channel.ChannelException: Failed >>>>to bind to: master/192.168.125.174:0 >>>>at org.jboss.netty.bootstrap.ServerBootstrap.bind(ServerBootstrap.java:272) >>>>... >>>>Caused by: java.net.BindException: Cannot assign requested address >>>>... >>>>I saw the same error reported before and have tried the following solutions. >>>> >>>> >>>>Set the variable SPARK_LOCAL_IP ,Changed the SPARK_MASTER_PORT to a >>>>different number..But nothing is working. >>>> >>>> >>>>When I try to start the worker from the respective machines using the >>>>following java command,its running without any exception >>>> >>>> >>>>java -cp >>>>::/usr/local/spark-1.0.0/conf:/usr/local/spark-1.0.0/assembly/target/scala-2.10/spark-assembly-1.0.0-hadoop1.2.1.jar >>>> -XX:MaxPermSize=128m -Dspark.akka.logLifecycleEvents=true -Xms512m >>>>-Xmx512m org.apache.spark.deploy.worker.Worker spark://:master:7077 >>>> >>>> >>>> >>>>Somebody please give a solution >>>> >>>>Thanks & Regards, >>>>Meethu M >>> >>> >>> >> >> >> >> >> > > >