Hi can anyone help with this?

On Fri, Jan 29, 2016 at 11:50 PM, allen chan <allen.michael.c...@gmail.com>
wrote:

> Use case: We are using kafka as broker in one of our elasticsearch
> clusters. Kafka caches the logs if elasticsearch has any performance
> issues.  I have Kafka set to delete logs pretty quickly to keep things in
> the file cache to limit IO.
>
> Questions:
> 1. in 0.9 it seems like consumer offers are stored only in Kafka. Is there
> a way to configure Kafka to delete my production logs pretty quickly but
> have a different retention behavior for the consumer offsets?
>
> 2. Our consumer lag monitoring show us that a lot of times our consumers
> are behind somewhere between 500 to 1000 messages. Looking at the JMX
> metrics requestSizeAvg and requestSizeMax, it shows our average request
> size is 500 bytes and max request size is 800,000 bytes. I assume the lag
> is because that batch could only hold one message given the max is 1000000
> bytes. I plan to enable compression and increase the max.bytes to 10mb to
> fix this short term. In a few blogs, people mentioned the ultimate fix
> should be splitting the message into smaller chunks in the producer and
> then having the consumer put it back together. Is that handled in the kafka
> producer/consumer natively or has to be handled outside of it?
>
> Thanks for the attention.
> Allen Chan
>
>
>


-- 
Allen Michael Chan

Reply via email to