Looks like that port is not available because another app is using that port. Can you take a look at netstat -a and use a port that is free?
Thanks, Hari On Fri, Nov 7, 2014 at 2:05 PM, Jeniba Johnson <jeniba.john...@lntinfotech.com> wrote: > Hi, > I have installed spark-1.1.0 and apache flume 1.4 for running streaming > example FlumeEventCount. Previously the code was working fine. Now Iam facing > with the below mentioned issues. My flume is running properly it is able to > write the file. > The command I use is > bin/run-example org.apache.spark.examples.streaming.FlumeEventCount > 172.29.17.178 65001 > 14/11/07 23:19:23 INFO receiver.ReceiverSupervisorImpl: Stopping receiver > with message: Error starting receiver 0: > org.jboss.netty.channel.ChannelException: Failed to bind to: > /172.29.17.178:65001 > 14/11/07 23:19:23 INFO flume.FlumeReceiver: Flume receiver stopped > 14/11/07 23:19:23 INFO receiver.ReceiverSupervisorImpl: Called receiver onStop > 14/11/07 23:19:23 INFO receiver.ReceiverSupervisorImpl: Deregistering > receiver 0 > 14/11/07 23:19:23 ERROR scheduler.ReceiverTracker: Deregistered receiver for > stream 0: Error starting receiver 0 - > org.jboss.netty.channel.ChannelException: Failed to bind to: > /172.29.17.178:65001 > at > org.jboss.netty.bootstrap.ServerBootstrap.bind(ServerBootstrap.java:272) > at org.apache.avro.ipc.NettyServer.<init>(NettyServer.java:106) > at org.apache.avro.ipc.NettyServer.<init>(NettyServer.java:119) > at org.apache.avro.ipc.NettyServer.<init>(NettyServer.java:74) > at org.apache.avro.ipc.NettyServer.<init>(NettyServer.java:68) > at > org.apache.spark.streaming.flume.FlumeReceiver.initServer(FlumeInputDStream.scala:164) > at > org.apache.spark.streaming.flume.FlumeReceiver.onStart(FlumeInputDStream.scala:171) > at > org.apache.spark.streaming.receiver.ReceiverSupervisor.startReceiver(ReceiverSupervisor.scala:121) > at > org.apache.spark.streaming.receiver.ReceiverSupervisor.start(ReceiverSupervisor.scala:106) > at > org.apache.spark.streaming.scheduler.ReceiverTracker$ReceiverLauncher$$anonfun$9.apply(ReceiverTracker.scala:264) > at > org.apache.spark.streaming.scheduler.ReceiverTracker$ReceiverLauncher$$anonfun$9.apply(ReceiverTracker.scala:257) > at > org.apache.spark.SparkContext$$anonfun$runJob$4.apply(SparkContext.scala:1121) > at > org.apache.spark.SparkContext$$anonfun$runJob$4.apply(SparkContext.scala:1121) > at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:62) > at org.apache.spark.scheduler.Task.run(Task.scala:54) > at > org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:177) > at > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) > at > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) > at java.lang.Thread.run(Thread.java:722) > Caused by: java.net.BindException: Address already in use > at sun.nio.ch.Net.bind0(Native Method) > at sun.nio.ch.Net.bind(Net.java:344) > at sun.nio.ch.Net.bind(Net.java:336) > at > sun.nio.ch.ServerSocketChannelImpl.bind(ServerSocketChannelImpl.java:199) > at sun.nio.ch.ServerSocketAdaptor.bind(ServerSocketAdaptor.java:74) > at > org.jboss.netty.channel.socket.nio.NioServerBoss$RegisterTask.run(NioServerBoss.java:193) > at > org.jboss.netty.channel.socket.nio.AbstractNioSelector.processTaskQueue(AbstractNioSelector.java:366) > at > org.jboss.netty.channel.socket.nio.AbstractNioSelector.run(AbstractNioSelector.java:290) > at > org.jboss.netty.channel.socket.nio.NioServerBoss.run(NioServerBoss.java:42) > ... 3 more > 14/11/07 23:19:23 INFO receiver.ReceiverSupervisorImpl: Stopped receiver 0 > 14/11/07 23:19:23 INFO receiver.BlockGenerator: Stopping BlockGenerator > 14/11/07 23:19:23 INFO util.RecurringTimer: Stopped timer for BlockGenerator > after time 1415382563200 > 14/11/07 23:19:23 INFO receiver.BlockGenerator: Waiting for block pushing > thread > 14/11/07 23:19:23 INFO receiver.BlockGenerator: Pushing out the last 0 blocks > 14/11/07 23:19:23 INFO receiver.BlockGenerator: Stopped block pushing thread > 14/11/07 23:19:23 INFO receiver.BlockGenerator: Stopped BlockGenerator > 14/11/07 23:19:23 INFO receiver.ReceiverSupervisorImpl: Waiting for executor > stop is over > 14/11/07 23:19:23 ERROR receiver.ReceiverSupervisorImpl: Stopped executor > with error: org.jboss.netty.channel.ChannelException: Failed to bind to: > /172.29.17.178:65001 > 14/11/07 23:19:23 ERROR executor.Executor: Exception in task 0.0 in stage 0.0 > (TID 0) > org.jboss.netty.channel.ChannelException: Failed to bind to: > /172.29.17.178:65001 > at > org.jboss.netty.bootstrap.ServerBootstrap.bind(ServerBootstrap.java:272) > at org.apache.avro.ipc.NettyServer.<init>(NettyServer.java:106) > at org.apache.avro.ipc.NettyServer.<init>(NettyServer.java:119) > at org.apache.avro.ipc.NettyServer.<init>(NettyServer.java:74) > at org.apache.avro.ipc.NettyServer.<init>(NettyServer.java:68) > at > org.apache.spark.streaming.flume.FlumeReceiver.initServer(FlumeInputDStream.scala:164) > at > org.apache.spark.streaming.flume.FlumeReceiver.onStart(FlumeInputDStream.scala:171) > at > org.apache.spark.streaming.receiver.ReceiverSupervisor.startReceiver(ReceiverSupervisor.scala:121) > at > org.apache.spark.streaming.receiver.ReceiverSupervisor.start(ReceiverSupervisor.scala:106) > at > org.apache.spark.streaming.scheduler.ReceiverTracker$ReceiverLauncher$$anonfun$9.apply(ReceiverTracker.scala:264) > at > org.apache.spark.streaming.scheduler.ReceiverTracker$ReceiverLauncher$$anonfun$9.apply(ReceiverTracker.scala:257) > at > org.apache.spark.SparkContext$$anonfun$runJob$4.apply(SparkContext.scala:1121) > at > org.apache.spark.SparkContext$$anonfun$runJob$4.apply(SparkContext.scala:1121) > at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:62) > at org.apache.spark.scheduler.Task.run(Task.scala:54) > at > org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:177) > at > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) > at > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) > at java.lang.Thread.run(Thread.java:722) > Caused by: java.net.BindException: Address already in use > at sun.nio.ch.Net.bind0(Native Method) > at sun.nio.ch.Net.bind(Net.java:344) > at sun.nio.ch.Net.bind(Net.java:336) > at > sun.nio.ch.ServerSocketChannelImpl.bind(ServerSocketChannelImpl.java:199) > at sun.nio.ch.ServerSocketAdaptor.bind(ServerSocketAdaptor.java:74) > at > org.jboss.netty.channel.socket.nio.NioServerBoss$RegisterTask.run(NioServerBoss.java:193) > at > org.jboss.netty.channel.socket.nio.AbstractNioSelector.processTaskQueue(AbstractNioSelector.java:366) > at > org.jboss.netty.channel.socket.nio.AbstractNioSelector.run(AbstractNioSelector.java:290) > at > org.jboss.netty.channel.socket.nio.NioServerBoss.run(NioServerBoss.java:42) > ... 3 more > 14/11/07 23:19:23 WARN scheduler.TaskSetManager: Lost task 0.0 in stage 0.0 > (TID 0, localhost): org.jboss.netty.channel.ChannelException: Failed to bind > to: /172.29.17.178:65001 > > org.jboss.netty.bootstrap.ServerBootstrap.bind(ServerBootstrap.java:272) > org.apache.avro.ipc.NettyServer.<init>(NettyServer.java:106) > org.apache.avro.ipc.NettyServer.<init>(NettyServer.java:119) > org.apache.avro.ipc.NettyServer.<init>(NettyServer.java:74) > org.apache.avro.ipc.NettyServer.<init>(NettyServer.java:68) > > org.apache.spark.streaming.flume.FlumeReceiver.initServer(FlumeInputDStream.scala:164) > > org.apache.spark.streaming.flume.FlumeReceiver.onStart(FlumeInputDStream.scala:171) > > org.apache.spark.streaming.receiver.ReceiverSupervisor.startReceiver(ReceiverSupervisor.scala:121) > > org.apache.spark.streaming.receiver.ReceiverSupervisor.start(ReceiverSupervisor.scala:106) > > org.apache.spark.streaming.scheduler.ReceiverTracker$ReceiverLauncher$$anonfun$9.apply(ReceiverTracker.scala:264) > > org.apache.spark.streaming.scheduler.ReceiverTracker$ReceiverLauncher$$anonfun$9.apply(ReceiverTracker.scala:257) > > org.apache.spark.SparkContext$$anonfun$runJob$4.apply(SparkContext.scala:1121) > > org.apache.spark.SparkContext$$anonfun$runJob$4.apply(SparkContext.scala:1121) > org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:62) > org.apache.spark.scheduler.Task.run(Task.scala:54) > org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:177) > > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) > > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) > java.lang.Thread.run(Thread.java:722) > 14/11/07 23:19:23 ERROR scheduler.TaskSetManager: Task 0 in stage 0.0 failed > 1 times; aborting job > 14/11/07 23:19:23 INFO scheduler.TaskSchedulerImpl: Removed TaskSet 0.0, > whose tasks have all completed, from pool > 14/11/07 23:19:23 INFO scheduler.TaskSchedulerImpl: Cancelling stage 0 > 14/11/07 23:19:23 INFO scheduler.DAGScheduler: Failed to run runJob at > ReceiverTracker.scala:275 > Exception in thread "Thread-28" org.apache.spark.SparkException: Job aborted > due to stage failure: Task 0 in stage 0.0 failed 1 times, most recent > failure: Lost task 0.0 in stage 0.0 (TID 0, localhost): > org.jboss.netty.channel.ChannelException: Failed to bind to: > /172.29.17.178:65001 > > org.jboss.netty.bootstrap.ServerBootstrap.bind(ServerBootstrap.java:272) > org.apache.avro.ipc.NettyServer.<init>(NettyServer.java:106) > org.apache.avro.ipc.NettyServer.<init>(NettyServer.java:119) > org.apache.avro.ipc.NettyServer.<init>(NettyServer.java:74) > org.apache.avro.ipc.NettyServer.<init>(NettyServer.java:68) > > org.apache.spark.streaming.flume.FlumeReceiver.initServer(FlumeInputDStream.scala:164) > > org.apache.spark.streaming.flume.FlumeReceiver.onStart(FlumeInputDStream.scala:171) > > org.apache.spark.streaming.receiver.ReceiverSupervisor.startReceiver(ReceiverSupervisor.scala:121) > > org.apache.spark.streaming.receiver.ReceiverSupervisor.start(ReceiverSupervisor.scala:106) > > org.apache.spark.streaming.scheduler.ReceiverTracker$ReceiverLauncher$$anonfun$9.apply(ReceiverTracker.scala:264) > > org.apache.spark.streaming.scheduler.ReceiverTracker$ReceiverLauncher$$anonfun$9.apply(ReceiverTracker.scala:257) > > org.apache.spark.SparkContext$$anonfun$runJob$4.apply(SparkContext.scala:1121) > > org.apache.spark.SparkContext$$anonfun$runJob$4.apply(SparkContext.scala:1121) > org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:62) > org.apache.spark.scheduler.Task.run(Task.scala:54) > org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:177) > > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) > > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) > java.lang.Thread.run(Thread.java:722) > Driver stacktrace: > at > org.apache.spark.scheduler.DAGScheduler.org$apache$spark$scheduler$DAGScheduler$$failJobAndIndependentStages(DAGScheduler.scala:1185) > at > org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1174) > at > org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1173) > at > scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59) > at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:47) > at > org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:1173) > at > org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:688) > at > org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:688) > at scala.Option.foreach(Option.scala:236) > at > org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:688) > at > org.apache.spark.scheduler.DAGSchedulerEventProcessActor$$anonfun$receive$2.applyOrElse(DAGScheduler.scala:1391) > at akka.actor.ActorCell.receiveMessage(ActorCell.scala:498) > at akka.actor.ActorCell.invoke(ActorCell.scala:456) > at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:237) > at akka.dispatch.Mailbox.run(Mailbox.scala:219) > at > akka.dispatch.ForkJoinExecutorConfigurator$AkkaForkJoinTask.exec(AbstractDispatcher.scala:386) > at > scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260) > at > scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339) > at > scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979) > at > scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107) > Regards, > Jeniba Johnson > ________________________________ > The contents of this e-mail and any attachment(s) may contain confidential or > privileged information for the intended recipient(s). Unintended recipients > are prohibited from taking action on the basis of information in this e-mail > and using or disseminating the information, and must notify the sender and > delete it from their system. L&T Infotech will not accept responsibility or > liability for the accuracy or completeness of, or the presence of any virus > or disabling code in this e-mail"