If you open up an issue on the project, I'd be happy to dig into this in more detail if needed. Excluding the ZK offset checking, Burrow doesn't enumerate consumer groups - it learns about them from offset commits. It sounds like maybe your consumer had not committed offsets for the first partition (at least not after Burrow was started).
-Todd On Friday, July 8, 2016, Tom Dearman <tom.dear...@gmail.com> wrote: > Todd, > > Thanks for that I am taking a look. > > Is there a bug whereby if you only have a couple of messages on a topic, > both with the same key, that burrow doesn’t return correct info. I was > finding that http://localhost:8100/v2/kafka/betwave/consumer < > http://localhost:8100/v2/kafka/betwave/consumer> was returning a message > with empty consumers until I put on another message with a different key, > i.e. a minimum of 2 partitions with something in them. I know this is not > very like production, but on my local this I was only testing with one user > so get just one partition filled. > > Tom > > On 6 Jul 2016, at 18:08, Todd Palino <tpal...@gmail.com <javascript:;>> > wrote: > > > > Yeah, I've written dissertations at this point on why MaxLag is flawed. > We > > also used to use the offset checker tool, and later something similar > that > > was a little easier to slot into our monitoring systems. Problems with > all > > of these is why I wrote Burrow (https://github.com/linkedin/Burrow) > > > > For more details, you can also check out my blog post on the release: > > > https://engineering.linkedin.com/apache-kafka/burrow-kafka-consumer-monitoring-reinvented > > > > -Todd > > > > On Wednesday, July 6, 2016, Tom Dearman <tom.dear...@gmail.com > <javascript:;>> wrote: > > > >> I recently had a problem on my production which I believe was a > >> manifestation of the issue kafka-2978 (Topic partition is not sometimes > >> consumed after rebalancing of consumer group), this is fixed in 0.9.0.1 > and > >> we will upgrade our client soon. However, it made me realise that I > didn’t > >> have any monitoring set up on this. The only thing I can find as a > metric > >> is the > >> > kafka.consumer:type=ConsumerFetcherManager,name=MaxLag,clientId=([-.\w]+), > >> which, if I understand correctly, is the max lag of any partition that > that > >> particular consumer is consuming. > >> 1. If I had been monitoring this, and if my consumer was suffering from > >> the issue in kafka-2978, would I actually have been alerted, i.e. since > the > >> consumer would think it is consuming correctly would it not have updated > >> the metric. > >> 2. There is another way to see offset lag using the command > >> /usr/bin/kafka-consumer-groups --new-consumer --bootstrap-server > >> 10.10.1.61:9092 --describe —group consumer_group_name and parsing the > >> response. Is it safe or advisable to do this? I like the fact that it > >> tells me each partition lag, although it is also not available if no > >> consumer from the group is currently consuming. > >> 3. Is there a better way of doing this? > > > > > > > > -- > > *Todd Palino* > > Staff Site Reliability Engineer > > Data Infrastructure Streaming > > > > > > > > linkedin.com/in/toddpalino > > -- *Todd Palino* Staff Site Reliability Engineer Data Infrastructure Streaming linkedin.com/in/toddpalino