adixitconfluent commented on code in PR #17870:
URL: https://github.com/apache/kafka/pull/17870#discussion_r1912168874
##########
core/src/test/java/kafka/test/api/ShareConsumerTest.java:
##########
@@ -902,7 +902,7 @@ public void
testFetchRecordLargerThanMaxPartitionFetchBytes(String persister) th
shareConsumer.subscribe(Collections.singleton(tp.topic()));
ConsumerRecords<byte[], byte[]> records =
shareConsumer.poll(Duration.ofMillis(5000));
- assertEquals(1, records.count());
+ assertEquals(2, records.count());
Review Comment:
So, earlier this test produced 2 records -
1. A small record whose size was lesser than partitionMaxBytes (1MB)
2. A big record whose size was equal to partitionMaxBytes (1MB)
due to the strict restriction in trunk code, we were able to only fetch 1
record in the first poll, since fetching the second record as well would have
violated the partitionMaxBytes limit. With my changes, partitionMaxBytes is no
longer a criteria while doing fetch, it just has to be within the
requestMaxBytes limit. Hence, fetching both first and second record do not
violate the limit requestMaxBytes in single fetch, so 2 records are returned
now.
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
To unsubscribe, e-mail: [email protected]
For queries about this service, please contact Infrastructure at:
[email protected]