Hello Neha, This is the result of a known bug in 0.9. Can you try running the latest Spark master branch to see if this problem is resolved?
TD On Tue, Apr 22, 2014 at 2:48 AM, NehaS Singh <nehas.si...@lntinfotech.com>wrote: > Hi, > > I have installed > spark-0.9.0-incubating-bin-cdh4 and I am using apache flume for streaming. > I have used the streaming.examples.FlumeEventCount. Also I have written > Avro conf file for flume.When I try to do streamin ing spark and I run > the following command it throws error > > > > ./bin/run-example org.apache.spark.streaming.examples.FlumeEventCount > local[2] ip 10001 > > > > org.jboss.netty.channel.ChannelException: Failed to bind to: /ip:9988 > > at > org.jboss.netty.bootstrap.ServerBootstrap.bind(ServerBootstrap.java:272) > > at org.apache.avro.ipc.NettyServer.<init>(NettyServer.java:106) > > at org.apache.avro.ipc.NettyServer.<init>(NettyServer.java:119) > > at org.apache.avro.ipc.NettyServer.<init>(NettyServer.java:74) > > at org.apache.avro.ipc.NettyServer.<init>(NettyServer.java:68) > > at > org.apache.spark.streaming.flume.FlumeReceiver.onStart(FlumeInputDStream.scala:143) > > at > org.apache.spark.streaming.dstream.NetworkReceiver.start(NetworkInputDStream.scala:126) > > at > org.apache.spark.streaming.scheduler.NetworkInputTracker$ReceiverExecutor$$anonfun$8.apply(NetworkInputTracker.scala:173) > > at > org.apache.spark.streaming.scheduler.NetworkInputTracker$ReceiverExecutor$$anonfun$8.apply(NetworkInputTracker.scala:169) > > at > org.apache.spark.SparkContext$$anonfun$runJob$4.apply(SparkContext.scala:884) > > at > org.apache.spark.SparkContext$$anonfun$runJob$4.apply(SparkContext.scala:884) > > at > org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:109) > > at org.apache.spark.scheduler.Task.run(Task.scala:53) > > at > org.apache.spark.executor.Executor$TaskRunner$$anonfun$run$1.apply$mcV$sp(Executor.scala:213) > > at > org.apache.spark.deploy.SparkHadoopUtil.runAsUser(SparkHadoopUtil.scala:49) > > at > org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:178) > > at > java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) > > at > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) > > at java.lang.Thread.run(Thread.java:662) > > Caused by: java.net.BindException: Cannot assign requested address > > at sun.nio.ch.Net.bind(Native Method) > > at > sun.nio.ch.ServerSocketChannelImpl.bind(ServerSocketChannelImpl.java:124) > > at sun.nio.ch.ServerSocketAdaptor.bind(ServerSocketAdaptor.java:59) > > at > org.jboss.netty.channel.socket.nio.NioServerBoss$RegisterTask.run(NioServerBoss.java:193) > > at > org.jboss.netty.channel.socket.nio.AbstractNioSelector.processTaskQueue(AbstractNioSelector.java:366) > > at > org.jboss.netty.channel.socket.nio.AbstractNioSelector.run(AbstractNioSelector.java:290) > > at > org.jboss.netty.channel.socket.nio.NioServerBoss.run(NioServerBoss.java:42) > > ... 3 more > > > > Can you please help > > > > This is my flumetest.conf > > > > a1.sources = tail-file > > a1.channels = c1 > > a1.sinks=avro-sink > > > > # define the flow > > a1.sources.tail-file.channels = c1 > > a1.sinks.avro-sink.channels = c1 > > a1.channels.c1.type = memory > > a1.channels.c1.capacity = 1000 > > > > # define source and sink > > a1.sources.tail-file.type = exec > > a1.sources.tail-file.command = tail -F /home/hduser/Flume/test.log > > #a1.sources.tail-file.channels = c1 > > a1.sinks.avro-sink.type = avro > > a1.sinks.avro-sink.hostname = ip // agent a2.s ip address or host name > > a1.sinks.avro-sink.port = 10001 > > # Use a channel which buffers events in memory > > a1.channels.c1.type = memory > > a1.channels.c1.capacity = 10000 > > a1.channels.c1.transactionCapacity = 10000 > > > > # Bind the source and sink to the channel > > a1.sources.tail-file.channels = c1 > > a1.sinks.avro-sink.channel = c1 > > > > Flumeslavetest.conf:- > > > > a2.sources = avro-collection-source > > a2.sinks = hdfs-sink > > a2.channels = mem-channel > > > > # define the flow > > a2.sources.avro-collection-source.channels = mem-channel > > a2.sinks.hdfs-sink.channel = mem-channel > > a2.channels.mem-channel.type = memory > > a2.channels.mem-channel.capacity = 1000 > > > > # avro source properties > > a2.sources.avro-collection-source.type = avro > > a2.sources.avro-collection-source.bind = ip // agent a2.s ip address or > host name > > a2.sources.avro-collection-source.port = 10001 > > > > # hdfs sink properties > > a2.sinks.hdfs-sink.type = hdfs > > a2.sinks.hdfs-sink.hdfs.writeFormat = Text > > a2.sinks.hdfs-sink.hdfs.filePrefix = testing > > a2.sinks.hdfs-sink.hdfs.path = hdfs://ip:8020/testingData > > > > > > My flume is running properly its is able to write file on hdfs. > > > > Please help as to how to resolve the error > > > > Regards, > > Neha Singh > > > > ------------------------------ > The contents of this e-mail and any attachment(s) may contain confidential > or privileged information for the intended recipient(s). Unintended > recipients are prohibited from taking action on the basis of information in > this e-mail and using or disseminating the information, and must notify the > sender and delete it from their system. L&T Infotech will not accept > responsibility or liability for the accuracy or completeness of, or the > presence of any virus or disabling code in this e-mail" >