Hi All,

What determines the batch size while reading from a file from HDFS?

I am trying to read files from HDFS and ingest into Kafka using Spark
Structured Streaming 2.3.1. I get an error sayiKafkafka batch size is too
big and that I need to increase max.request.size. Sure I can increase it
but I would like to know what other parameters I can change such that I
don't have to change the default max.request.size?

The default max.request.size of a Kafka producer docs says it is set to
1MB.
And each file I have in HDFS is < 12MB.

Any suggestions will be great.

Thanks!

Reply via email to