Hi,

We are experiencing an issue where the RecordAccumulator size is growing
beyond the configured buffer.memory size .

Below is the sequence of events when this issue occured -

1. One of the broker out of 10 brokers  was brought down at 15:10 PM
2.Till 17:11 PM (same date) the kafka producer instance was in healthy state
3. The same down broker was brought up at ~17:11PM ( same date)
4. Observed UnknownTopicOrPartitionException for some messages
5. HeapDump on  17:14 shows the RecordAccumulator size around ~2.9 GB way
above the configured buffer.size =120MB


Other Details  -

kafka version = 0.8.2.1

Below is the kafka producer properties configured  -

kafka.bootstrap.servers=qa-vip-url
kafka.acks=1
kafka.buffer.memory=125829120
kafka.compression.type=none
kafka.retries=3
kafka.batch.size=200
kafka.client.id=service-instance-id
kafka.linger.ms=0
kafka.max.request.size=5242880
kafka.receive.buffer.bytes=32768
kafka.send.buffer.bytes=131072
kafka.timeout.ms=3000
kafka.block.on.buffer.full=false
kafka.metadata.fetch.timeout.ms=60000
kafka.metadata.max.age.ms=300000
kafka.reconnect.backoff.ms=30
kafka.retry.backoff.ms=3000
kafka.key.serializer=org.apache.kafka.common.serialization.ByteArraySerializer
kafka.value.serializer=org.apache.kafka.common.serialization.ByteArraySerializer


Please help in identifing the root cause which is causing this memory leak
for RecordAccumulator.

Attached is the HeapDump overview for reference

Let me know if you need any other information

Regards,
Liju John

Reply via email to