use different topology name and spout id and submit again. Thanking you. With Regards Sree
On Thursday, March 2, 2017 10:54 AM, Igor Kuzmenko <f1she...@gmail.com> wrote: I'm using storm-kafka-client 1.1.1-SNAPSHOT build. After topology start kafka spout read all partitions from kafka exept one: | Id | Topic | Partition | Latest Offset | Spout Committed Offset | Lag | | Kafka Spout | gtp | 0 | 5726714188 | 5726700216 | 13972 | | Kafka Spout | gtp | 1 | 5716936379 | 5716922137 | 14242 | | Kafka Spout | gtp | 2 | 5725709217 | 5484094447 | 241614770 | | Kafka Spout | gtp | 3 | 5713385013 | 5713370624 | 14389 | | Kafka Spout | gtp | 4 | 5721077118 | 5721062942 | 14176 | | Kafka Spout | gtp | 5 | 5717492246 | 5717478013 | 14233 | | Kafka Spout | gtp | 6 | 5716438459 | 5716424263 | 14196 | | Kafka Spout | gtp | 7 | 5719165064 | 5719150543 | 14521 | Reading partition #2 fails with: 2017-03-02 23:06:32.013 o.a.k.c.c.i.Fetcher [INFO] [51] [Thread-8-Kafka Spout-executor[9 9]] Fetch offset 5484094448 is out of range for partition gtp-2, resetting offset This is common problem for kafka consumer and can be solved by setting auto.offset.reset to 'earliest' or 'latest' value. So did I: props.put(ConsumerConfig.VALUE_DESERIALIZER_CLASS_CONFIG, getValueDeserializer().getCanonicalName()); After redeploy topology nothing has changed. I still get this info message 'offset is out of range' and same commited offset for 2nd partition.