ableegoldman commented on code in PR #12235:
URL: https://github.com/apache/kafka/pull/12235#discussion_r888670820


##########
streams/src/main/java/org/apache/kafka/streams/processor/internals/RecordCollectorImpl.java:
##########
@@ -199,6 +201,7 @@ public <K, V> void send(final String topic,
                 log.trace("Failed record: (key {} value {} timestamp {}) 
topic=[{}] partition=[{}]", key, value, timestamp, topic, partition);
             }
         });
+        return recordSizeInBytes(keyBytes == null ? 0 : keyBytes.length, 
valBytes == null ? 0 : valBytes.length, topic, headers);

Review Comment:
   I did it like this to avoid the extra/unnecessary null check for consumer 
records specifically, which already track the serialized size in bytes unlike 
the producer record. And unfortunately they don't inherit from a common 
interface/class -- but I added separate middle-man methods to handle them and 
moved the null check for the producer case there, should be addressed now



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: jira-unsubscr...@kafka.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org

Reply via email to