Hi Team, We are trying to configure Kafka to produce larger messages,
Below are the configs:- Server.properties message.max.bytes=100000000 replica.fetch.max.bytes=100000001 Producer.properties max.request.size=100000000 compression.type=gzip Consumer.properties fetch.message.max.bytes=100000000 When trying to execute a larger file with below command bin/kafka-console-producer.sh --broker-list localhost:9092 --topic metadata_upload < /export/home/Userid/metadata/TopUBP-1_metadata.json Getting below error:- >[2018-09-19 11:20:03,307] ERROR Error when sending message to topic metadata_upload with key: null, value: 2170060 bytes with error: (org.apache.kafka.clients.producer.internals.ErrorLoggingCallback) org.apache.kafka.common.errors.RecordTooLargeException: The message is 2170148 bytes when serialized which is larger than the maximum request size you have configured with the max.request.size configuration. Thanks, Sarath Reddy.