I think the __consumer_offsets configuration looks fine, I just wanted to be sure there wasn't only one replica. How many Zookeepers do you have, and were they restarted as well?
I would suspect that the spout isn't committing properly for some reason. The default behavior is to get the committed offset from Kafka when starting, and if it is present it is used. If it isn't there the spout starts over from the beginning of the partitions. You can check if the spout is committing by enabling debug logging for the classes in storm-kafka-client, to check logs like this one https://github.com/apache/storm/blob/v1.1.0/external/storm-kafka-client/src/main/java/org/apache/storm/kafka/spout/internal/OffsetManager.java#L118 . 2017-08-21 22:20 GMT+02:00 Elyahou Ittah <elyaho...@fiverr.com>: > Hi Stig, > > I don't have this kind of errors normally. It just occured at the rolling > restart of kafka. > > Also the __consumer_offsets configuration is: > Topic:__consumer_offsets PartitionCount:50 ReplicationFactor:3 > Configs:segment.bytes=104857600,cleanup.policy=compact,compression.type= > producer > Topic: __consumer_offsets Partition: 0 Leader: 0 Replicas: 2,1,0 Isr: 0,1 > ... > > The fact that the replication factor is 3 even if there is only two broker > can cause an issue ? > > > On Mon, Aug 21, 2017 at 6:51 PM, Stig Rohde Døssing <s...@apache.org> > wrote: > >> The spout will reemit some messages if it fails to commit offsets to >> Kafka. Are these CommitFailedExceptions occuring in your logs normally? >> >> Also since the spout stores offsets in Kafka, you may want to check the >> replication factor on that topic by running `./kafka-topics.sh --zookeeper >> localhost:2181 --describe --topic __consumer_offsets` in one of your Kafka >> /bin directories. >> >> 2017-08-21 17:17 GMT+02:00 Elyahou Ittah <elyaho...@fiverr.com>: >> >>> The config is the default one, I just set the bootstrap server. >>> >>> Kafka version is 0.11 >>> >>> Storm-kafka-client is 1.1.0 >>> >>> On Mon, Aug 21, 2017 at 5:48 PM, Stig Rohde Døssing <s...@apache.org> >>> wrote: >>> >>>> Hi Elyahou, >>>> >>>> Could you post your spout configuration, Kafka version and >>>> storm-kafka-client version? The logs imply that your spouts are not polling >>>> often enough. >>>> >>>> 2017-08-21 <20%2017%2008%2021> 9:44 GMT+02:00 Elyahou Ittah < >>>> elyaho...@fiverr.com>: >>>> >>>>> I noticed that storm kafka spout reconsume all kafka message after a >>>>> rolling restart of kafka cluster. >>>>> >>>>> This issue occured only with kafkaSpout consumer and not for my other >>>>> consumers (ruby based using the kafka consumer API like kafkaSpout) >>>>> >>>>> Attached logs of the spout. >>>>> >>>>> Do you know what can cause this kind of behavior ? >>>>> >>>> >>>> >>> >> >