This is an automated email from the ASF dual-hosted git repository.

chia7712 pushed a commit to branch 4.0
in repository https://gitbox.apache.org/repos/asf/kafka.git


The following commit(s) were added to refs/heads/4.0 by this push:
     new 41e4743be97 MINOR: Replace scala $var by java string (#18393)
41e4743be97 is described below

commit 41e4743be9700947ecff2cf13ed49edf5d4b364f
Author: Xuan-Zhang Gong <[email protected]>
AuthorDate: Tue Jan 7 03:38:31 2025 +0800

    MINOR: Replace scala $var by java string (#18393)
    
    Reviewers: Luke Chen <[email protected]>, TengYao Chi 
<[email protected]>, Chia-Ping Tsai <[email protected]>
---
 .../java/org/apache/kafka/storage/internals/log/LocalLog.java     | 8 ++++----
 1 file changed, 4 insertions(+), 4 deletions(-)

diff --git 
a/storage/src/main/java/org/apache/kafka/storage/internals/log/LocalLog.java 
b/storage/src/main/java/org/apache/kafka/storage/internals/log/LocalLog.java
index e48f63eafd8..027196a5de7 100644
--- a/storage/src/main/java/org/apache/kafka/storage/internals/log/LocalLog.java
+++ b/storage/src/main/java/org/apache/kafka/storage/internals/log/LocalLog.java
@@ -358,7 +358,7 @@ public class LocalLog {
      */
     public List<LogSegment> deleteAllSegments() {
         return maybeHandleIOException(
-            () -> "Error while deleting all segments for $topicPartition in 
dir ${dir.getParent}",
+            () -> String.format("Error while deleting all segments for %s in 
dir %s", topicPartition, dir.getParent()),
             () -> {
                 List<LogSegment> deletableSegments = new 
ArrayList<>(segments.values());
                 removeAndDeleteSegments(
@@ -470,8 +470,8 @@ public class LocalLog {
         return maybeHandleIOException(
                 () -> "Exception while reading from " + topicPartition + " in 
dir " + dir.getParent(),
                 () -> {
-                    logger.trace("Reading maximum $maxLength bytes at offset 
{} from log with total length {} bytes",
-                            startOffset, segments.sizeInBytes());
+                    logger.trace("Reading maximum {} bytes at offset {} from 
log with total length {} bytes",
+                            maxLength, startOffset, segments.sizeInBytes());
 
                     LogOffsetMetadata endOffsetMetadata = nextOffsetMetadata;
                     long endOffset = endOffsetMetadata.messageOffset;
@@ -943,7 +943,7 @@ public class LocalLog {
                 throw new IllegalStateException("Inconsistent segment sizes 
after split before: " + segment.log().sizeInBytes() + " after: " + 
totalSizeOfNewSegments);
             }
             // replace old segment with new ones
-            LOG.info("{}Replacing overflowed segment $segment with split 
segments {}", logPrefix, newSegments);
+            LOG.info("{}Replacing overflowed segment {} with split segments 
{}", logPrefix, segment, newSegments);
             List<LogSegment> deletedSegments = 
replaceSegments(existingSegments, newSegments, singletonList(segment),
                     dir, topicPartition, config, scheduler, 
logDirFailureChannel, logPrefix, false);
             return new SplitSegmentResult(deletedSegments, newSegments);

Reply via email to