[ https://issues.apache.org/jira/browse/SPARK-9476?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16079148#comment-16079148 ]
Dharmas commented on SPARK-9476: -------------------------------- Even I am getting the same issue !. any update on this. > Kafka stream loses leader after 2h of operation > ------------------------------------------------ > > Key: SPARK-9476 > URL: https://issues.apache.org/jira/browse/SPARK-9476 > Project: Spark > Issue Type: Bug > Components: DStreams > Affects Versions: 1.4.0 > Environment: Docker, Centos, Spark standalone, core i7, 8Gb > Reporter: Ruben Ramalho > > This seems to happen every 2h, it happens both with the direct stream and > regular stream, I'm doing window operations over a 1h period (if that can > help). > Here's part of the error message: > 2015-07-30 13:27:23 WARN ClientUtils$:89 - Fetching topic metadata with > correlation id 10 for topics [Set(updates)] from broker > [id:0,host:192.168.3.23,port:3000] failed > java.nio.channels.ClosedChannelException > at kafka.network.BlockingChannel.send(BlockingChannel.scala:100) > at kafka.producer.SyncProducer.liftedTree1$1(SyncProducer.scala:73) > at > kafka.producer.SyncProducer.kafka$producer$SyncProducer$$doSend(SyncProducer.scala:72) > at kafka.producer.SyncProducer.send(SyncProducer.scala:113) > at kafka.client.ClientUtils$.fetchTopicMetadata(ClientUtils.scala:58) > at kafka.client.ClientUtils$.fetchTopicMetadata(ClientUtils.scala:93) > at > kafka.consumer.ConsumerFetcherManager$LeaderFinderThread.doWork(ConsumerFetcherManager.scala:66) > at kafka.utils.ShutdownableThread.run(ShutdownableThread.scala:60) > 2015-07-30 13:27:23 INFO SyncProducer:68 - Disconnecting from > 192.168.3.23:3000 > 2015-07-30 13:27:23 WARN ConsumerFetcherManager$LeaderFinderThread:89 - > [spark-group_81563e123e9f-1438259236988-fc3d82bf-leader-finder-thread], > Failed to find leader for Set([updates,0]) > kafka.common.KafkaException: fetching topic metadata for topics > [Set(oversight-updates)] from broker > [ArrayBuffer(id:0,host:192.168.3.23,port:3000)] failed > at kafka.client.ClientUtils$.fetchTopicMetadata(ClientUtils.scala:72) > at kafka.client.ClientUtils$.fetchTopicMetadata(ClientUtils.scala:93) > at > kafka.consumer.ConsumerFetcherManager$LeaderFinderThread.doWork(ConsumerFetcherManager.scala:66) > at kafka.utils.ShutdownableThread.run(ShutdownableThread.scala:60) > Caused by: java.nio.channels.ClosedChannelException > at kafka.network.BlockingChannel.send(BlockingChannel.scala:100) > at kafka.producer.SyncProducer.liftedTree1$1(SyncProducer.scala:73) > at > kafka.producer.SyncProducer.kafka$producer$SyncProducer$$doSend(SyncProducer.scala:72) > at kafka.producer.SyncProducer.send(SyncProducer.scala:113) > at kafka.client.ClientUtils$.fetchTopicMetadata(ClientUtils.scala:58) > After the crash I tried to communicate with kafka with a simple scala > consumer and producer and have no problem at all. Spark tough needs a kafka > container restart to start normal operaiton. There are no errors on the kafka > log, apart from an improper closed connection. > I have been trying to solve this problem for days, I suspect this has > something to do with spark. -- This message was sent by Atlassian JIRA (v6.4.14#64029) --------------------------------------------------------------------- To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org