This is an automated email from the ASF dual-hosted git repository.

chia7712 pushed a commit to branch trunk
in repository https://gitbox.apache.org/repos/asf/kafka.git


The following commit(s) were added to refs/heads/trunk by this push:
     new 0ed104c3dc8 MINOR: Cleanup the storage module unit tests (#16202)
0ed104c3dc8 is described below

commit 0ed104c3dc87ba8842b1a46e3e16ff21700dfed5
Author: Kamal Chandraprakash <kchandraprak...@uber.com>
AuthorDate: Thu Jun 6 21:26:08 2024 +0530

    MINOR: Cleanup the storage module unit tests (#16202)
    
    - Use SystemTime instead of MockTime when time is not mocked
    - Use static assertions to reduce the line length
    - Fold the lines if it exceeds the limit
    - rename tp0 to tpId0 when it refers to TopicIdPartition
    
    Reviewers: Kuan-Po (Cooper) Tseng <brandb...@gmail.com>, Chia-Ping Tsai 
<chia7...@gmail.com>
---
 .../storage/RemoteLogMetadataFormatterTest.java    | 18 ++---
 .../storage/RemoteLogMetadataSerdeTest.java        | 38 +++++-----
 .../storage/RemoteLogMetadataTransformTest.java    | 40 +++++------
 ...picBasedRemoteLogMetadataManagerConfigTest.java | 53 +++++---------
 ...ogMetadataManagerMultipleSubscriptionsTest.java | 41 +++++------
 ...icBasedRemoteLogMetadataManagerRestartTest.java | 25 +++----
 .../TopicBasedRemoteLogMetadataManagerTest.java    | 56 ++++++++-------
 .../storage/RemoteLogMetadataManagerTest.java      | 83 ++++++++++++----------
 8 files changed, 169 insertions(+), 185 deletions(-)

diff --git 
a/storage/src/test/java/org/apache/kafka/server/log/remote/metadata/storage/RemoteLogMetadataFormatterTest.java
 
b/storage/src/test/java/org/apache/kafka/server/log/remote/metadata/storage/RemoteLogMetadataFormatterTest.java
index e3d1a2aee0c..1380a735fba 100644
--- 
a/storage/src/test/java/org/apache/kafka/server/log/remote/metadata/storage/RemoteLogMetadataFormatterTest.java
+++ 
b/storage/src/test/java/org/apache/kafka/server/log/remote/metadata/storage/RemoteLogMetadataFormatterTest.java
@@ -24,7 +24,6 @@ import 
org.apache.kafka.server.log.remote.metadata.storage.serialization.RemoteL
 import org.apache.kafka.server.log.remote.storage.RemoteLogSegmentId;
 import org.apache.kafka.server.log.remote.storage.RemoteLogSegmentMetadata;
 import 
org.apache.kafka.server.log.remote.storage.RemoteLogSegmentMetadata.CustomMetadata;
-import org.apache.kafka.server.log.remote.storage.RemoteLogSegmentState;
 import org.junit.jupiter.api.Test;
 
 import java.io.ByteArrayOutputStream;
@@ -35,6 +34,7 @@ import java.util.Map;
 import java.util.Optional;
 
 import static org.junit.jupiter.api.Assertions.assertEquals;
+import static 
org.apache.kafka.server.log.remote.storage.RemoteLogSegmentState.COPY_SEGMENT_STARTED;
 
 public class RemoteLogMetadataFormatterTest {
     private static final Uuid TOPIC_ID = Uuid.randomUuid();
@@ -51,12 +51,12 @@ public class RemoteLogMetadataFormatterTest {
         RemoteLogSegmentId remoteLogSegmentId = new RemoteLogSegmentId(TP0, 
SEGMENT_ID);
         Optional<CustomMetadata> customMetadata = Optional.of(new 
CustomMetadata(new byte[10]));
         RemoteLogSegmentMetadata remoteLogMetadata = new 
RemoteLogSegmentMetadata(
-                remoteLogSegmentId, 0L, 100L, -1L, 1,
-                123L, 1024, customMetadata,
-                RemoteLogSegmentState.COPY_SEGMENT_STARTED, segLeaderEpochs);
+                remoteLogSegmentId, 0L, 100L, -1L, 1, 123L, 1024, 
customMetadata, COPY_SEGMENT_STARTED,
+                segLeaderEpochs);
 
         byte[] metadataBytes = new 
RemoteLogMetadataSerde().serialize(remoteLogMetadata);
-        ConsumerRecord<byte[], byte[]> metadataRecord = new 
ConsumerRecord<>("__remote_log_metadata", 0, 0, null, metadataBytes);
+        ConsumerRecord<byte[], byte[]> metadataRecord = new ConsumerRecord<>(
+                "__remote_log_metadata", 0, 0, null, metadataBytes);
 
         String expected = String.format(
                 "partition: 0, offset: 0, value: " +
@@ -68,9 +68,11 @@ public class RemoteLogMetadataFormatterTest {
                 TOPIC_ID, SEGMENT_ID);
         try (ByteArrayOutputStream baos = new ByteArrayOutputStream();
              PrintStream ps = new PrintStream(baos)) {
-            RemoteLogMetadataSerde.RemoteLogMetadataFormatter formatter = new 
RemoteLogMetadataSerde.RemoteLogMetadataFormatter();
-            formatter.writeTo(metadataRecord, ps);
-            assertEquals(expected, baos.toString());
+            try (RemoteLogMetadataSerde.RemoteLogMetadataFormatter formatter =
+                         new 
RemoteLogMetadataSerde.RemoteLogMetadataFormatter()) {
+                formatter.writeTo(metadataRecord, ps);
+                assertEquals(expected, baos.toString());
+            }
         }
     }
 }
diff --git 
a/storage/src/test/java/org/apache/kafka/server/log/remote/metadata/storage/RemoteLogMetadataSerdeTest.java
 
b/storage/src/test/java/org/apache/kafka/server/log/remote/metadata/storage/RemoteLogMetadataSerdeTest.java
index 5b48790c7fd..b1b91dacf23 100644
--- 
a/storage/src/test/java/org/apache/kafka/server/log/remote/metadata/storage/RemoteLogMetadataSerdeTest.java
+++ 
b/storage/src/test/java/org/apache/kafka/server/log/remote/metadata/storage/RemoteLogMetadataSerdeTest.java
@@ -19,48 +19,48 @@ package org.apache.kafka.server.log.remote.metadata.storage;
 import org.apache.kafka.common.TopicIdPartition;
 import org.apache.kafka.common.TopicPartition;
 import org.apache.kafka.common.Uuid;
-import org.apache.kafka.common.utils.MockTime;
+import org.apache.kafka.common.utils.SystemTime;
 import org.apache.kafka.common.utils.Time;
 import 
org.apache.kafka.server.log.remote.metadata.storage.serialization.RemoteLogMetadataSerde;
 import org.apache.kafka.server.log.remote.storage.RemoteLogSegmentId;
 import org.apache.kafka.server.log.remote.storage.RemoteLogSegmentMetadata;
 import 
org.apache.kafka.server.log.remote.storage.RemoteLogSegmentMetadata.CustomMetadata;
 import 
org.apache.kafka.server.log.remote.storage.RemoteLogSegmentMetadataUpdate;
-import org.apache.kafka.server.log.remote.storage.RemoteLogSegmentState;
 import 
org.apache.kafka.server.log.remote.storage.RemotePartitionDeleteMetadata;
-import org.apache.kafka.server.log.remote.storage.RemotePartitionDeleteState;
 import org.apache.kafka.server.log.remote.storage.RemoteLogMetadata;
-import org.junit.jupiter.api.Assertions;
 import org.junit.jupiter.api.Test;
 
 import java.util.HashMap;
 import java.util.Map;
 import java.util.Optional;
 
+import static 
org.apache.kafka.server.log.remote.storage.RemoteLogSegmentState.COPY_SEGMENT_STARTED;
+import static 
org.apache.kafka.server.log.remote.storage.RemoteLogSegmentState.COPY_SEGMENT_FINISHED;
+import static 
org.apache.kafka.server.log.remote.storage.RemotePartitionDeleteState.DELETE_PARTITION_MARKED;
+import static org.junit.jupiter.api.Assertions.assertEquals;
+import static org.junit.jupiter.api.Assertions.assertThrows;
+
 public class RemoteLogMetadataSerdeTest {
 
     public static final String TOPIC = "foo";
     private static final TopicIdPartition TP0 = new 
TopicIdPartition(Uuid.randomUuid(), new TopicPartition(TOPIC, 0));
-    private final Time time = new MockTime(1);
+    private final Time time = new SystemTime();
 
     @Test
     public void testRemoteLogSegmentMetadataSerde() {
         RemoteLogSegmentMetadata remoteLogSegmentMetadata = 
createRemoteLogSegmentMetadata();
-
         doTestRemoteLogMetadataSerde(remoteLogSegmentMetadata);
     }
 
     @Test
     public void testRemoteLogSegmentMetadataUpdateSerde() {
         RemoteLogSegmentMetadataUpdate remoteLogSegmentMetadataUpdate = 
createRemoteLogSegmentMetadataUpdate();
-
         doTestRemoteLogMetadataSerde(remoteLogSegmentMetadataUpdate);
     }
 
     @Test
     public void testRemotePartitionDeleteMetadataSerde() {
         RemotePartitionDeleteMetadata remotePartitionDeleteMetadata = 
createRemotePartitionDeleteMetadata();
-
         doTestRemoteLogMetadataSerde(remotePartitionDeleteMetadata);
     }
 
@@ -70,24 +70,19 @@ public class RemoteLogMetadataSerdeTest {
         segLeaderEpochs.put(1, 20L);
         segLeaderEpochs.put(2, 80L);
         RemoteLogSegmentId remoteLogSegmentId = new RemoteLogSegmentId(TP0, 
Uuid.randomUuid());
-        return new RemoteLogSegmentMetadata(remoteLogSegmentId, 0L, 100L, -1L, 
1,
-                                            time.milliseconds(), 1024,
-                                            Optional.of(new CustomMetadata(new 
byte[] {0, 1, 2, 3})),
-                                            
RemoteLogSegmentState.COPY_SEGMENT_STARTED,
-                                            segLeaderEpochs
-        );
+        return new RemoteLogSegmentMetadata(remoteLogSegmentId, 0L, 100L, -1L, 
1, 
+                time.milliseconds(), 1024, Optional.of(new CustomMetadata(new 
byte[] {0, 1, 2, 3})),
+                COPY_SEGMENT_STARTED, segLeaderEpochs);
     }
 
     private RemoteLogSegmentMetadataUpdate 
createRemoteLogSegmentMetadataUpdate() {
         RemoteLogSegmentId remoteLogSegmentId = new RemoteLogSegmentId(TP0, 
Uuid.randomUuid());
         return new RemoteLogSegmentMetadataUpdate(remoteLogSegmentId, 
time.milliseconds(),
-                                                  Optional.of(new 
CustomMetadata(new byte[] {0, 1, 2, 3})),
-                                                  
RemoteLogSegmentState.COPY_SEGMENT_FINISHED, 2);
+                Optional.of(new CustomMetadata(new byte[] {0, 1, 2, 3})), 
COPY_SEGMENT_FINISHED, 2);
     }
 
     private RemotePartitionDeleteMetadata 
createRemotePartitionDeleteMetadata() {
-        return new RemotePartitionDeleteMetadata(TP0, 
RemotePartitionDeleteState.DELETE_PARTITION_MARKED,
-                                                 time.milliseconds(), 0);
+        return new RemotePartitionDeleteMetadata(TP0, DELETE_PARTITION_MARKED, 
time.milliseconds(), 0);
     }
 
     private void doTestRemoteLogMetadataSerde(RemoteLogMetadata 
remoteLogMetadata) {
@@ -96,16 +91,17 @@ public class RemoteLogMetadataSerdeTest {
         byte[] metadataBytes = serializer.serialize(remoteLogMetadata);
 
         // Deserialize the bytes and check the RemoteLogMetadata object is as 
expected.
-        // Created another RemoteLogMetadataSerde instance to depict the real 
usecase of serializer and deserializer having their own instances.
+        // Created another RemoteLogMetadataSerde instance to depict the real 
usecase of serializer and 
+        // deserializer having their own instances.
         RemoteLogMetadataSerde deserializer = new RemoteLogMetadataSerde();
         RemoteLogMetadata deserializedRemoteLogMetadata = 
deserializer.deserialize(metadataBytes);
-        Assertions.assertEquals(remoteLogMetadata, 
deserializedRemoteLogMetadata);
+        assertEquals(remoteLogMetadata, deserializedRemoteLogMetadata);
     }
 
     @Test
     public void testInvalidRemoteStorageMetadata() {
         // Serializing receives an exception as it does not have the expected 
RemoteLogMetadata registered in serdes.
-        Assertions.assertThrows(IllegalArgumentException.class,
+        assertThrows(IllegalArgumentException.class,
             () -> new RemoteLogMetadataSerde().serialize(new 
InvalidRemoteLogMetadata(1, time.milliseconds())));
     }
 
diff --git 
a/storage/src/test/java/org/apache/kafka/server/log/remote/metadata/storage/RemoteLogMetadataTransformTest.java
 
b/storage/src/test/java/org/apache/kafka/server/log/remote/metadata/storage/RemoteLogMetadataTransformTest.java
index 504f47e17a5..70770542eb9 100644
--- 
a/storage/src/test/java/org/apache/kafka/server/log/remote/metadata/storage/RemoteLogMetadataTransformTest.java
+++ 
b/storage/src/test/java/org/apache/kafka/server/log/remote/metadata/storage/RemoteLogMetadataTransformTest.java
@@ -19,7 +19,7 @@ package org.apache.kafka.server.log.remote.metadata.storage;
 import org.apache.kafka.common.TopicIdPartition;
 import org.apache.kafka.common.TopicPartition;
 import org.apache.kafka.common.Uuid;
-import org.apache.kafka.common.utils.MockTime;
+import org.apache.kafka.common.utils.SystemTime;
 import org.apache.kafka.common.utils.Time;
 import org.apache.kafka.server.common.ApiMessageAndVersion;
 import 
org.apache.kafka.server.log.remote.metadata.storage.serialization.RemoteLogSegmentMetadataTransform;
@@ -29,60 +29,56 @@ import 
org.apache.kafka.server.log.remote.storage.RemoteLogSegmentId;
 import org.apache.kafka.server.log.remote.storage.RemoteLogSegmentMetadata;
 import 
org.apache.kafka.server.log.remote.storage.RemoteLogSegmentMetadata.CustomMetadata;
 import 
org.apache.kafka.server.log.remote.storage.RemoteLogSegmentMetadataUpdate;
-import org.apache.kafka.server.log.remote.storage.RemoteLogSegmentState;
 import 
org.apache.kafka.server.log.remote.storage.RemotePartitionDeleteMetadata;
-import org.apache.kafka.server.log.remote.storage.RemotePartitionDeleteState;
-import org.junit.jupiter.api.Assertions;
 import org.junit.jupiter.api.Test;
 
 import java.util.Collections;
 import java.util.Optional;
 
+import static 
org.apache.kafka.server.log.remote.storage.RemoteLogSegmentState.COPY_SEGMENT_FINISHED;
+import static 
org.apache.kafka.server.log.remote.storage.RemotePartitionDeleteState.DELETE_PARTITION_STARTED;
+import static org.junit.jupiter.api.Assertions.assertEquals;
+
 public class RemoteLogMetadataTransformTest {
     private static final TopicIdPartition TP0 = new 
TopicIdPartition(Uuid.randomUuid(), new TopicPartition("foo", 0));
-    private final Time time = new MockTime(1);
+    private final Time time = new SystemTime();
 
     @Test
     public void testRemoteLogSegmentMetadataTransform() {
         RemoteLogSegmentMetadataTransform metadataTransform = new 
RemoteLogSegmentMetadataTransform();
-
         RemoteLogSegmentMetadata metadata = createRemoteLogSegmentMetadata();
         ApiMessageAndVersion apiMessageAndVersion = 
metadataTransform.toApiMessageAndVersion(metadata);
         RemoteLogSegmentMetadata remoteLogSegmentMetadataFromRecord = 
metadataTransform
                 .fromApiMessageAndVersion(apiMessageAndVersion);
-
-        Assertions.assertEquals(metadata, remoteLogSegmentMetadataFromRecord);
+        assertEquals(metadata, remoteLogSegmentMetadataFromRecord);
     }
 
     @Test
     public void testRemoteLogSegmentMetadataUpdateTransform() {
         RemoteLogSegmentMetadataUpdateTransform metadataUpdateTransform = new 
RemoteLogSegmentMetadataUpdateTransform();
-
-        RemoteLogSegmentMetadataUpdate metadataUpdate =
-                new RemoteLogSegmentMetadataUpdate(new RemoteLogSegmentId(TP0, 
Uuid.randomUuid()), time.milliseconds(),
-                                                   Optional.of(new 
CustomMetadata(new byte[]{0, 1, 2, 3})),
-                                                   
RemoteLogSegmentState.COPY_SEGMENT_FINISHED, 1);
+        RemoteLogSegmentMetadataUpdate metadataUpdate = new 
RemoteLogSegmentMetadataUpdate(
+                new RemoteLogSegmentId(TP0, Uuid.randomUuid()), 
time.milliseconds(),
+                Optional.of(new CustomMetadata(new byte[]{0, 1, 2, 3})), 
COPY_SEGMENT_FINISHED, 1);
         ApiMessageAndVersion apiMessageAndVersion = 
metadataUpdateTransform.toApiMessageAndVersion(metadataUpdate);
-        RemoteLogSegmentMetadataUpdate metadataUpdateFromRecord = 
metadataUpdateTransform.fromApiMessageAndVersion(apiMessageAndVersion);
-
-        Assertions.assertEquals(metadataUpdate, metadataUpdateFromRecord);
+        RemoteLogSegmentMetadataUpdate metadataUpdateFromRecord =
+                
metadataUpdateTransform.fromApiMessageAndVersion(apiMessageAndVersion);
+        assertEquals(metadataUpdate, metadataUpdateFromRecord);
     }
 
     private RemoteLogSegmentMetadata createRemoteLogSegmentMetadata() {
         RemoteLogSegmentId remoteLogSegmentId = new RemoteLogSegmentId(TP0, 
Uuid.randomUuid());
         return new RemoteLogSegmentMetadata(remoteLogSegmentId, 0L, 100L, -1L, 
1,
-                                            time.milliseconds(), 1024, 
Collections.singletonMap(0, 0L));
+                time.milliseconds(), 1024, Collections.singletonMap(0, 0L));
     }
 
     @Test
     public void testRemoteLogPartitionMetadataTransform() {
         RemotePartitionDeleteMetadataTransform transform = new 
RemotePartitionDeleteMetadataTransform();
-
         RemotePartitionDeleteMetadata partitionDeleteMetadata
-                = new RemotePartitionDeleteMetadata(TP0, 
RemotePartitionDeleteState.DELETE_PARTITION_STARTED, time.milliseconds(), 1);
+                = new RemotePartitionDeleteMetadata(TP0, 
DELETE_PARTITION_STARTED, time.milliseconds(), 1);
         ApiMessageAndVersion apiMessageAndVersion = 
transform.toApiMessageAndVersion(partitionDeleteMetadata);
-        RemotePartitionDeleteMetadata partitionDeleteMetadataFromRecord = 
transform.fromApiMessageAndVersion(apiMessageAndVersion);
-
-        Assertions.assertEquals(partitionDeleteMetadata, 
partitionDeleteMetadataFromRecord);
+        RemotePartitionDeleteMetadata partitionDeleteMetadataFromRecord =
+                transform.fromApiMessageAndVersion(apiMessageAndVersion);
+        assertEquals(partitionDeleteMetadata, 
partitionDeleteMetadataFromRecord);
     }
 }
diff --git 
a/storage/src/test/java/org/apache/kafka/server/log/remote/metadata/storage/TopicBasedRemoteLogMetadataManagerConfigTest.java
 
b/storage/src/test/java/org/apache/kafka/server/log/remote/metadata/storage/TopicBasedRemoteLogMetadataManagerConfigTest.java
index 8e3985d0d5f..34f1fb08366 100644
--- 
a/storage/src/test/java/org/apache/kafka/server/log/remote/metadata/storage/TopicBasedRemoteLogMetadataManagerConfigTest.java
+++ 
b/storage/src/test/java/org/apache/kafka/server/log/remote/metadata/storage/TopicBasedRemoteLogMetadataManagerConfigTest.java
@@ -20,10 +20,7 @@ import org.apache.kafka.clients.CommonClientConfigs;
 import org.apache.kafka.clients.consumer.ConsumerConfig;
 import org.apache.kafka.clients.producer.ProducerConfig;
 import org.apache.kafka.test.TestUtils;
-import org.junit.jupiter.api.Assertions;
 import org.junit.jupiter.api.Test;
-import org.slf4j.Logger;
-import org.slf4j.LoggerFactory;
 
 import java.util.AbstractMap;
 import java.util.HashMap;
@@ -38,10 +35,10 @@ import static 
org.apache.kafka.server.log.remote.metadata.storage.TopicBasedRemo
 import static 
org.apache.kafka.server.log.remote.metadata.storage.TopicBasedRemoteLogMetadataManagerConfig.REMOTE_LOG_METADATA_TOPIC_REPLICATION_FACTOR_PROP;
 import static 
org.apache.kafka.server.log.remote.metadata.storage.TopicBasedRemoteLogMetadataManagerConfig.REMOTE_LOG_METADATA_TOPIC_RETENTION_MS_PROP;
 
-public class TopicBasedRemoteLogMetadataManagerConfigTest {
-    private static final Logger log = 
LoggerFactory.getLogger(TopicBasedRemoteLogMetadataManagerConfigTest.class);
+import static org.junit.jupiter.api.Assertions.assertEquals;
 
-    private static final String BOOTSTRAP_SERVERS = "localhost:9091";
+public class TopicBasedRemoteLogMetadataManagerConfigTest {
+    private static final String BOOTSTRAP_SERVERS = "localhost:2222";
 
     @Test
     public void testValidConfig() {
@@ -60,41 +57,32 @@ public class TopicBasedRemoteLogMetadataManagerConfigTest {
 
         // Check for topic properties
         TopicBasedRemoteLogMetadataManagerConfig rlmmConfig = new 
TopicBasedRemoteLogMetadataManagerConfig(props);
-        
Assertions.assertEquals(props.get(REMOTE_LOG_METADATA_TOPIC_PARTITIONS_PROP), 
rlmmConfig.metadataTopicPartitionsCount());
+        assertEquals(props.get(REMOTE_LOG_METADATA_TOPIC_PARTITIONS_PROP), 
rlmmConfig.metadataTopicPartitionsCount());
 
         // Check for common client configs.
-        Assertions.assertEquals(BOOTSTRAP_SERVERS, 
rlmmConfig.commonProperties().get(CommonClientConfigs.BOOTSTRAP_SERVERS_CONFIG));
-        Assertions.assertEquals(BOOTSTRAP_SERVERS, 
rlmmConfig.producerProperties().get(CommonClientConfigs.BOOTSTRAP_SERVERS_CONFIG));
-        Assertions.assertEquals(BOOTSTRAP_SERVERS, 
rlmmConfig.consumerProperties().get(CommonClientConfigs.BOOTSTRAP_SERVERS_CONFIG));
+        assertEquals(BOOTSTRAP_SERVERS, 
rlmmConfig.commonProperties().get(CommonClientConfigs.BOOTSTRAP_SERVERS_CONFIG));
+        assertEquals(BOOTSTRAP_SERVERS, 
rlmmConfig.producerProperties().get(CommonClientConfigs.BOOTSTRAP_SERVERS_CONFIG));
+        assertEquals(BOOTSTRAP_SERVERS, 
rlmmConfig.consumerProperties().get(CommonClientConfigs.BOOTSTRAP_SERVERS_CONFIG));
 
         for (Map.Entry<String, Object> entry : commonClientConfig.entrySet()) {
-            log.info("Checking config: " + entry.getKey());
-            Assertions.assertEquals(entry.getValue(),
-                                    
rlmmConfig.commonProperties().get(entry.getKey()));
-            Assertions.assertEquals(entry.getValue(),
-                                    
rlmmConfig.producerProperties().get(entry.getKey()));
-            Assertions.assertEquals(entry.getValue(),
-                                    
rlmmConfig.consumerProperties().get(entry.getKey()));
+            assertEquals(entry.getValue(), 
rlmmConfig.commonProperties().get(entry.getKey()));
+            assertEquals(entry.getValue(), 
rlmmConfig.producerProperties().get(entry.getKey()));
+            assertEquals(entry.getValue(), 
rlmmConfig.consumerProperties().get(entry.getKey()));
         }
-
         // Check for producer configs.
         for (Map.Entry<String, Object> entry : producerConfig.entrySet()) {
-            log.info("Checking config: " + entry.getKey());
-            Assertions.assertEquals(entry.getValue(),
-                                    
rlmmConfig.producerProperties().get(entry.getKey()));
+            assertEquals(entry.getValue(), 
rlmmConfig.producerProperties().get(entry.getKey()));
         }
-
         // Check for consumer configs.
         for (Map.Entry<String, Object> entry : consumerConfig.entrySet()) {
-            log.info("Checking config: " + entry.getKey());
-            Assertions.assertEquals(entry.getValue(),
-                                    
rlmmConfig.consumerProperties().get(entry.getKey()));
+            assertEquals(entry.getValue(), 
rlmmConfig.consumerProperties().get(entry.getKey()));
         }
     }
 
     @Test
     public void testCommonProducerConsumerOverridesConfig() {
-        Map.Entry<String, Long> overrideEntry = new 
AbstractMap.SimpleImmutableEntry<>(CommonClientConfigs.METADATA_MAX_AGE_CONFIG, 
60000L);
+        Map.Entry<String, Long> overrideEntry =
+                new 
AbstractMap.SimpleImmutableEntry<>(CommonClientConfigs.METADATA_MAX_AGE_CONFIG, 
60000L);
         Map<String, Object> commonClientConfig = new HashMap<>();
         commonClientConfig.put(CommonClientConfigs.RETRIES_CONFIG, 10);
         commonClientConfig.put(CommonClientConfigs.RETRY_BACKOFF_MS_CONFIG, 
1000L);
@@ -114,12 +102,9 @@ public class TopicBasedRemoteLogMetadataManagerConfigTest {
         Map<String, Object> props = createValidConfigProps(commonClientConfig, 
producerConfig, consumerConfig);
         TopicBasedRemoteLogMetadataManagerConfig rlmmConfig = new 
TopicBasedRemoteLogMetadataManagerConfig(props);
 
-        Assertions.assertEquals(overrideCommonPropValue,
-                                
rlmmConfig.commonProperties().get(overrideEntry.getKey()));
-        Assertions.assertEquals(overriddenProducerPropValue,
-                                
rlmmConfig.producerProperties().get(overrideEntry.getKey()));
-        Assertions.assertEquals(overriddenConsumerPropValue,
-                                
rlmmConfig.consumerProperties().get(overrideEntry.getKey()));
+        assertEquals(overrideCommonPropValue, 
rlmmConfig.commonProperties().get(overrideEntry.getKey()));
+        assertEquals(overriddenProducerPropValue, 
rlmmConfig.producerProperties().get(overrideEntry.getKey()));
+        assertEquals(overriddenConsumerPropValue, 
rlmmConfig.consumerProperties().get(overrideEntry.getKey()));
     }
 
     private Map<String, Object> createValidConfigProps(Map<String, Object> 
commonClientConfig,
@@ -129,7 +114,6 @@ public class TopicBasedRemoteLogMetadataManagerConfigTest {
         props.put(REMOTE_LOG_METADATA_COMMON_CLIENT_PREFIX + 
CommonClientConfigs.BOOTSTRAP_SERVERS_CONFIG, BOOTSTRAP_SERVERS);
         props.put(BROKER_ID, 1);
         props.put(LOG_DIR, TestUtils.tempDirectory().getAbsolutePath());
-
         props.put(REMOTE_LOG_METADATA_TOPIC_REPLICATION_FACTOR_PROP, (short) 
3);
         props.put(REMOTE_LOG_METADATA_TOPIC_PARTITIONS_PROP, 10);
         props.put(REMOTE_LOG_METADATA_TOPIC_RETENTION_MS_PROP, 60 * 60 * 
1000L);
@@ -138,17 +122,14 @@ public class TopicBasedRemoteLogMetadataManagerConfigTest 
{
         for (Map.Entry<String, Object> entry : commonClientConfig.entrySet()) {
             props.put(REMOTE_LOG_METADATA_COMMON_CLIENT_PREFIX + 
entry.getKey(), entry.getValue());
         }
-
         // producer configs
         for (Map.Entry<String, Object> entry : producerConfig.entrySet()) {
             props.put(REMOTE_LOG_METADATA_PRODUCER_PREFIX + entry.getKey(), 
entry.getValue());
         }
-
         //consumer configs
         for (Map.Entry<String, Object> entry : consumerConfig.entrySet()) {
             props.put(REMOTE_LOG_METADATA_CONSUMER_PREFIX + entry.getKey(), 
entry.getValue());
         }
-
         return props;
     }
 }
\ No newline at end of file
diff --git 
a/storage/src/test/java/org/apache/kafka/server/log/remote/metadata/storage/TopicBasedRemoteLogMetadataManagerMultipleSubscriptionsTest.java
 
b/storage/src/test/java/org/apache/kafka/server/log/remote/metadata/storage/TopicBasedRemoteLogMetadataManagerMultipleSubscriptionsTest.java
index 916b475c143..d02da098040 100644
--- 
a/storage/src/test/java/org/apache/kafka/server/log/remote/metadata/storage/TopicBasedRemoteLogMetadataManagerMultipleSubscriptionsTest.java
+++ 
b/storage/src/test/java/org/apache/kafka/server/log/remote/metadata/storage/TopicBasedRemoteLogMetadataManagerMultipleSubscriptionsTest.java
@@ -27,12 +27,11 @@ import org.apache.kafka.clients.admin.NewTopic;
 import org.apache.kafka.common.TopicIdPartition;
 import org.apache.kafka.common.TopicPartition;
 import org.apache.kafka.common.Uuid;
-import org.apache.kafka.common.utils.MockTime;
+import org.apache.kafka.common.utils.SystemTime;
 import org.apache.kafka.common.utils.Time;
 import org.apache.kafka.server.log.remote.storage.RemoteLogSegmentId;
 import org.apache.kafka.server.log.remote.storage.RemoteLogSegmentMetadata;
 import org.apache.kafka.server.log.remote.storage.RemoteStorageException;
-import org.junit.jupiter.api.Assertions;
 import org.junit.jupiter.api.Tag;
 import org.junit.jupiter.api.extension.ExtendWith;
 
@@ -44,6 +43,10 @@ import java.util.concurrent.ExecutionException;
 import java.util.concurrent.Phaser;
 import java.util.concurrent.TimeUnit;
 
+import static org.junit.jupiter.api.Assertions.assertDoesNotThrow;
+import static org.junit.jupiter.api.Assertions.assertEquals;
+import static org.junit.jupiter.api.Assertions.assertThrows;
+import static org.junit.jupiter.api.Assertions.assertTrue;
 import static org.mockito.ArgumentMatchers.any;
 import static org.mockito.Mockito.clearInvocations;
 import static org.mockito.Mockito.doAnswer;
@@ -55,10 +58,7 @@ import static org.mockito.Mockito.verify;
 @ClusterTestDefaults(brokers = 3)
 public class TopicBasedRemoteLogMetadataManagerMultipleSubscriptionsTest {
     private final ClusterInstance clusterInstance;
-
-    private static final int SEG_SIZE = 1024 * 1024;
-
-    private final Time time = new MockTime(1);
+    private final Time time = new SystemTime();
 
     
TopicBasedRemoteLogMetadataManagerMultipleSubscriptionsTest(ClusterInstance 
clusterInstance) {
         this.clusterInstance = clusterInstance;
@@ -125,24 +125,25 @@ public class 
TopicBasedRemoteLogMetadataManagerMultipleSubscriptionsTest {
             // Add segments for these partitions but an exception is received 
as they have not yet been subscribed.
             // These messages would have been published to the respective 
metadata topic partitions but the ConsumerManager
             // has not yet been subscribing as they are not yet registered.
+            int segSize = 1048576;
             RemoteLogSegmentMetadata leaderSegmentMetadata = new 
RemoteLogSegmentMetadata(new RemoteLogSegmentId(leaderTopicIdPartition, 
Uuid.randomUuid()),
                 0, 100, -1L, 0,
-                time.milliseconds(), SEG_SIZE, Collections.singletonMap(0, 
0L));
-            ExecutionException exception = 
Assertions.assertThrows(ExecutionException.class, () -> 
remoteLogMetadataManager.addRemoteLogSegmentMetadata(leaderSegmentMetadata).get());
-            Assertions.assertEquals("org.apache.kafka.common.KafkaException: 
This consumer is not assigned to the target partition 0. Currently assigned 
partitions: []",
+                time.milliseconds(), segSize, Collections.singletonMap(0, 0L));
+            ExecutionException exception = 
assertThrows(ExecutionException.class,
+                    () -> 
remoteLogMetadataManager.addRemoteLogSegmentMetadata(leaderSegmentMetadata).get());
+            assertEquals("org.apache.kafka.common.KafkaException: This 
consumer is not assigned to the target partition 0. Currently assigned 
partitions: []",
                 exception.getMessage());
 
             RemoteLogSegmentMetadata followerSegmentMetadata = new 
RemoteLogSegmentMetadata(new RemoteLogSegmentId(followerTopicIdPartition, 
Uuid.randomUuid()),
                 0, 100, -1L, 0,
-                time.milliseconds(), SEG_SIZE, Collections.singletonMap(0, 
0L));
-            exception = Assertions.assertThrows(ExecutionException.class, () 
-> 
remoteLogMetadataManager.addRemoteLogSegmentMetadata(followerSegmentMetadata).get());
-            Assertions.assertEquals("org.apache.kafka.common.KafkaException: 
This consumer is not assigned to the target partition 0. Currently assigned 
partitions: []",
+                time.milliseconds(), segSize, Collections.singletonMap(0, 0L));
+            exception = assertThrows(ExecutionException.class, () -> 
remoteLogMetadataManager.addRemoteLogSegmentMetadata(followerSegmentMetadata).get());
+            assertEquals("org.apache.kafka.common.KafkaException: This 
consumer is not assigned to the target partition 0. Currently assigned 
partitions: []",
                 exception.getMessage());
 
             // `listRemoteLogSegments` will receive an exception as these 
topic partitions are not yet registered.
-            Assertions.assertThrows(RemoteStorageException.class, () -> 
remoteLogMetadataManager.listRemoteLogSegments(leaderTopicIdPartition));
-            Assertions.assertThrows(RemoteStorageException.class, () -> 
remoteLogMetadataManager.listRemoteLogSegments(followerTopicIdPartition));
-
+            assertThrows(RemoteStorageException.class, () -> 
remoteLogMetadataManager.listRemoteLogSegments(leaderTopicIdPartition));
+            assertThrows(RemoteStorageException.class, () -> 
remoteLogMetadataManager.listRemoteLogSegments(followerTopicIdPartition));
 
             
remoteLogMetadataManager.onPartitionLeadershipChanges(Collections.singleton(leaderTopicIdPartition),
                     Collections.emptySet());
@@ -156,8 +157,8 @@ public class 
TopicBasedRemoteLogMetadataManagerMultipleSubscriptionsTest {
 
             // leader partitions would have received as it is registered, but 
follower partition is not yet registered,
             // hence it throws an exception.
-            
Assertions.assertTrue(remoteLogMetadataManager.listRemoteLogSegments(leaderTopicIdPartition).hasNext());
-            Assertions.assertThrows(RemoteStorageException.class, () -> 
remoteLogMetadataManager.listRemoteLogSegments(followerTopicIdPartition));
+            
assertTrue(remoteLogMetadataManager.listRemoteLogSegments(leaderTopicIdPartition).hasNext());
+            assertThrows(RemoteStorageException.class, () -> 
remoteLogMetadataManager.listRemoteLogSegments(followerTopicIdPartition));
 
             // Register follower partition
             // Phaser::bulkRegister and Phaser::register provide the "countUp" 
feature
@@ -172,15 +173,15 @@ public class 
TopicBasedRemoteLogMetadataManagerMultipleSubscriptionsTest {
             
verify(spyRemotePartitionMetadataStore).markInitialized(followerTopicIdPartition);
             
verify(spyRemotePartitionMetadataStore).handleRemoteLogSegmentMetadata(followerSegmentMetadata);
             // In this state, all the metadata should be available in RLMM for 
both leader and follower partitions.
-            
Assertions.assertTrue(remoteLogMetadataManager.listRemoteLogSegments(leaderTopicIdPartition).hasNext(),
 "No segments found");
-            
Assertions.assertTrue(remoteLogMetadataManager.listRemoteLogSegments(followerTopicIdPartition).hasNext(),
 "No segments found");
+            
assertTrue(remoteLogMetadataManager.listRemoteLogSegments(leaderTopicIdPartition).hasNext(),
 "No segments found");
+            
assertTrue(remoteLogMetadataManager.listRemoteLogSegments(followerTopicIdPartition).hasNext(),
 "No segments found");
         }
     }
 
     private void createTopic(String topic, Map<Integer, List<Integer>> 
replicasAssignments) {
         try (Admin admin = 
Admin.create(Collections.singletonMap(CommonClientConfigs.BOOTSTRAP_SERVERS_CONFIG,
 clusterInstance.bootstrapServers()))) {
             admin.createTopics(Collections.singletonList(new NewTopic(topic, 
replicasAssignments)));
-            Assertions.assertDoesNotThrow(() -> 
clusterInstance.waitForTopic(topic, replicasAssignments.size()));
+            assertDoesNotThrow(() -> clusterInstance.waitForTopic(topic, 
replicasAssignments.size()));
         }
     }
 }
diff --git 
a/storage/src/test/java/org/apache/kafka/server/log/remote/metadata/storage/TopicBasedRemoteLogMetadataManagerRestartTest.java
 
b/storage/src/test/java/org/apache/kafka/server/log/remote/metadata/storage/TopicBasedRemoteLogMetadataManagerRestartTest.java
index dee686dc036..a08f12d9ae2 100644
--- 
a/storage/src/test/java/org/apache/kafka/server/log/remote/metadata/storage/TopicBasedRemoteLogMetadataManagerRestartTest.java
+++ 
b/storage/src/test/java/org/apache/kafka/server/log/remote/metadata/storage/TopicBasedRemoteLogMetadataManagerRestartTest.java
@@ -24,12 +24,11 @@ import org.apache.kafka.clients.admin.NewTopic;
 import org.apache.kafka.common.TopicIdPartition;
 import org.apache.kafka.common.TopicPartition;
 import org.apache.kafka.common.Uuid;
-import org.apache.kafka.common.utils.MockTime;
+import org.apache.kafka.common.utils.SystemTime;
 import org.apache.kafka.common.utils.Time;
 import org.apache.kafka.server.log.remote.storage.RemoteLogSegmentId;
 import org.apache.kafka.server.log.remote.storage.RemoteLogSegmentMetadata;
 import org.apache.kafka.test.TestUtils;
-import org.junit.jupiter.api.Assertions;
 import org.junit.jupiter.api.Tag;
 import org.junit.jupiter.api.extension.ExtendWith;
 
@@ -37,14 +36,13 @@ import java.util.Arrays;
 import java.util.Collections;
 
 import static 
org.apache.kafka.server.log.remote.metadata.storage.TopicBasedRemoteLogMetadataManagerConfig.LOG_DIR;
+import static org.junit.jupiter.api.Assertions.assertTrue;
 
 @ExtendWith(value = ClusterTestExtensions.class)
 @Tag("integration")
 public class TopicBasedRemoteLogMetadataManagerRestartTest {
 
-    private static final int SEG_SIZE = 1024 * 1024;
-
-    private final Time time = new MockTime(1);
+    private final Time time = new SystemTime();
     private final String logDir = 
TestUtils.tempDirectory("_rlmm_segs_").getAbsolutePath();
     private final ClusterInstance clusterInstance;
 
@@ -76,16 +74,17 @@ public class TopicBasedRemoteLogMetadataManagerRestartTest {
         clusterInstance.waitForTopic(leaderTopic, 1);
         clusterInstance.waitForTopic(followerTopic, 1);
 
-        final TopicIdPartition leaderTopicIdPartition = new 
TopicIdPartition(Uuid.randomUuid(), new TopicPartition(leaderTopic, 0));
-        final TopicIdPartition followerTopicIdPartition = new 
TopicIdPartition(Uuid.randomUuid(), new TopicPartition(followerTopic, 0));
+        TopicIdPartition leaderTopicIdPartition = new 
TopicIdPartition(Uuid.randomUuid(), new TopicPartition(leaderTopic, 0));
+        TopicIdPartition followerTopicIdPartition = new 
TopicIdPartition(Uuid.randomUuid(), new TopicPartition(followerTopic, 0));
+        int segSize = 1048576;
         RemoteLogSegmentMetadata leaderSegmentMetadata = new 
RemoteLogSegmentMetadata(
                 new RemoteLogSegmentId(leaderTopicIdPartition, 
Uuid.randomUuid()),
                 0, 100, -1L, 0,
-                time.milliseconds(), SEG_SIZE, Collections.singletonMap(0, 
0L));
+                time.milliseconds(), segSize, Collections.singletonMap(0, 0L));
         RemoteLogSegmentMetadata followerSegmentMetadata = new 
RemoteLogSegmentMetadata(
                 new RemoteLogSegmentId(followerTopicIdPartition, 
Uuid.randomUuid()),
                 0, 100, -1L, 0,
-                time.milliseconds(), SEG_SIZE, Collections.singletonMap(0, 
0L));
+                time.milliseconds(), segSize, Collections.singletonMap(0, 0L));
 
         try (TopicBasedRemoteLogMetadataManager 
topicBasedRemoteLogMetadataManager = 
createTopicBasedRemoteLogMetadataManager()) {
             // Register these partitions to RemoteLogMetadataManager.
@@ -115,12 +114,14 @@ public class 
TopicBasedRemoteLogMetadataManagerRestartTest {
             RemoteLogSegmentMetadata leaderSegmentMetadata2 = new 
RemoteLogSegmentMetadata(
                     new RemoteLogSegmentId(leaderTopicIdPartition, 
Uuid.randomUuid()),
                     101, 200, -1L, 0,
-                    time.milliseconds(), SEG_SIZE, Collections.singletonMap(0, 
101L));
+                    time.milliseconds(), segSize, Collections.singletonMap(0, 
101L));
             
topicBasedRemoteLogMetadataManager.addRemoteLogSegmentMetadata(leaderSegmentMetadata2).get();
 
             // Check that both the stored segment and recently added segment 
are available.
-            
Assertions.assertTrue(TestUtils.sameElementsWithoutOrder(Arrays.asList(leaderSegmentMetadata,
 leaderSegmentMetadata2).iterator(),
-                    
topicBasedRemoteLogMetadataManager.listRemoteLogSegments(leaderTopicIdPartition)));
+            assertTrue(TestUtils.sameElementsWithoutOrder(
+                    Arrays.asList(leaderSegmentMetadata, 
leaderSegmentMetadata2).iterator(),
+                    
topicBasedRemoteLogMetadataManager.listRemoteLogSegments(leaderTopicIdPartition))
+            );
         }
     }
 }
diff --git 
a/storage/src/test/java/org/apache/kafka/server/log/remote/metadata/storage/TopicBasedRemoteLogMetadataManagerTest.java
 
b/storage/src/test/java/org/apache/kafka/server/log/remote/metadata/storage/TopicBasedRemoteLogMetadataManagerTest.java
index 183a8493412..8b9cfd0700c 100644
--- 
a/storage/src/test/java/org/apache/kafka/server/log/remote/metadata/storage/TopicBasedRemoteLogMetadataManagerTest.java
+++ 
b/storage/src/test/java/org/apache/kafka/server/log/remote/metadata/storage/TopicBasedRemoteLogMetadataManagerTest.java
@@ -25,14 +25,13 @@ import org.apache.kafka.clients.admin.NewTopic;
 import org.apache.kafka.common.TopicIdPartition;
 import org.apache.kafka.common.TopicPartition;
 import org.apache.kafka.common.Uuid;
-import org.apache.kafka.common.utils.MockTime;
+import org.apache.kafka.common.utils.SystemTime;
 import org.apache.kafka.common.utils.Time;
 import org.apache.kafka.server.log.remote.storage.RemoteLogSegmentId;
 import org.apache.kafka.server.log.remote.storage.RemoteLogSegmentMetadata;
 import 
org.apache.kafka.server.log.remote.storage.RemoteResourceNotFoundException;
 import org.apache.kafka.server.log.remote.storage.RemoteStorageException;
 import org.junit.jupiter.api.AfterEach;
-import org.junit.jupiter.api.Assertions;
 import org.junit.jupiter.api.extension.ExtendWith;
 
 import java.io.IOException;
@@ -42,6 +41,10 @@ import java.util.concurrent.CountDownLatch;
 import java.util.concurrent.ExecutionException;
 import java.util.concurrent.TimeUnit;
 
+import static org.junit.jupiter.api.Assertions.assertEquals;
+import static org.junit.jupiter.api.Assertions.assertFalse;
+import static org.junit.jupiter.api.Assertions.assertThrows;
+import static org.junit.jupiter.api.Assertions.assertTrue;
 import static org.mockito.ArgumentMatchers.any;
 import static org.mockito.Mockito.doAnswer;
 import static org.mockito.Mockito.spy;
@@ -50,11 +53,10 @@ import static org.mockito.Mockito.verify;
 @ExtendWith(ClusterTestExtensions.class)
 @ClusterTestDefaults(brokers = 3)
 public class TopicBasedRemoteLogMetadataManagerTest {
-    private static final int SEG_SIZE = 1024 * 1024;
-
+    private static final int SEG_SIZE = 1048576;
     private final ClusterInstance clusterInstance;
     private final RemotePartitionMetadataStore 
spyRemotePartitionMetadataEventHandler = spy(new 
RemotePartitionMetadataStore());
-    private final Time time = new MockTime(1);
+    private final Time time = new SystemTime();
     private TopicBasedRemoteLogMetadataManager remoteLogMetadataManager;
 
     TopicBasedRemoteLogMetadataManagerTest(ClusterInstance clusterInstance) {
@@ -83,7 +85,7 @@ public class TopicBasedRemoteLogMetadataManagerTest {
             admin.createTopics(Collections.singletonList(new NewTopic(topic, 
1, (short) 1))).all().get();
             clusterInstance.waitForTopic(topic, 1);
             boolean doesTopicExist = topicBasedRlmm().doesTopicExist(admin, 
topic);
-            Assertions.assertTrue(doesTopicExist);
+            assertTrue(doesTopicExist);
         }
     }
 
@@ -92,7 +94,7 @@ public class TopicBasedRemoteLogMetadataManagerTest {
         try (Admin admin = clusterInstance.createAdminClient()) {
             String topic = "dummy-test-topic";
             boolean doesTopicExist = topicBasedRlmm().doesTopicExist(admin, 
topic);
-            Assertions.assertFalse(doesTopicExist);
+            assertFalse(doesTopicExist);
         }
     }
 
@@ -139,37 +141,37 @@ public class TopicBasedRemoteLogMetadataManagerTest {
         RemoteLogSegmentMetadata leaderSegmentMetadata = new 
RemoteLogSegmentMetadata(new RemoteLogSegmentId(newLeaderTopicIdPartition, 
Uuid.randomUuid()),
                                                                                
 0, 100, -1L, 0,
                                                                                
 time.milliseconds(), SEG_SIZE, Collections.singletonMap(0, 0L));
-        Assertions.assertThrows(Exception.class, () -> 
topicBasedRlmm().addRemoteLogSegmentMetadata(leaderSegmentMetadata).get());
+        assertThrows(Exception.class, () -> 
topicBasedRlmm().addRemoteLogSegmentMetadata(leaderSegmentMetadata).get());
 
         RemoteLogSegmentMetadata followerSegmentMetadata = new 
RemoteLogSegmentMetadata(new RemoteLogSegmentId(newFollowerTopicIdPartition, 
Uuid.randomUuid()),
                                                                                
 0, 100, -1L, 0,
                                                                                
 time.milliseconds(), SEG_SIZE, Collections.singletonMap(0, 0L));
-        Assertions.assertThrows(Exception.class, () -> 
topicBasedRlmm().addRemoteLogSegmentMetadata(followerSegmentMetadata).get());
+        assertThrows(Exception.class, () -> 
topicBasedRlmm().addRemoteLogSegmentMetadata(followerSegmentMetadata).get());
 
         // `listRemoteLogSegments` will receive an exception as these topic 
partitions are not yet registered.
-        Assertions.assertThrows(RemoteResourceNotFoundException.class, () -> 
topicBasedRlmm().listRemoteLogSegments(newLeaderTopicIdPartition));
-        Assertions.assertThrows(RemoteResourceNotFoundException.class, () -> 
topicBasedRlmm().listRemoteLogSegments(newFollowerTopicIdPartition));
+        assertThrows(RemoteResourceNotFoundException.class, () -> 
topicBasedRlmm().listRemoteLogSegments(newLeaderTopicIdPartition));
+        assertThrows(RemoteResourceNotFoundException.class, () -> 
topicBasedRlmm().listRemoteLogSegments(newFollowerTopicIdPartition));
 
         
topicBasedRlmm().onPartitionLeadershipChanges(Collections.singleton(newLeaderTopicIdPartition),
                                                       
Collections.singleton(newFollowerTopicIdPartition));
 
         // RemoteLogSegmentMetadata events are already published, and 
topicBasedRlmm's consumer manager will start
         // fetching those events and build the cache.
-        Assertions.assertTrue(initializationLatch.await(30_000, 
TimeUnit.MILLISECONDS));
-        
Assertions.assertTrue(handleRemoteLogSegmentMetadataLatch.await(30_000, 
TimeUnit.MILLISECONDS));
+        assertTrue(initializationLatch.await(30_000, TimeUnit.MILLISECONDS));
+        assertTrue(handleRemoteLogSegmentMetadataLatch.await(30_000, 
TimeUnit.MILLISECONDS));
 
         
verify(spyRemotePartitionMetadataEventHandler).markInitialized(newLeaderTopicIdPartition);
         
verify(spyRemotePartitionMetadataEventHandler).markInitialized(newFollowerTopicIdPartition);
         
verify(spyRemotePartitionMetadataEventHandler).handleRemoteLogSegmentMetadata(leaderSegmentMetadata);
         
verify(spyRemotePartitionMetadataEventHandler).handleRemoteLogSegmentMetadata(followerSegmentMetadata);
-        
Assertions.assertTrue(topicBasedRlmm().listRemoteLogSegments(newLeaderTopicIdPartition).hasNext());
-        
Assertions.assertTrue(topicBasedRlmm().listRemoteLogSegments(newFollowerTopicIdPartition).hasNext());
+        
assertTrue(topicBasedRlmm().listRemoteLogSegments(newLeaderTopicIdPartition).hasNext());
+        
assertTrue(topicBasedRlmm().listRemoteLogSegments(newFollowerTopicIdPartition).hasNext());
     }
 
     @ClusterTest
     public void testRemoteLogSizeCalculationForUnknownTopicIdPartitionThrows() 
{
         TopicIdPartition topicIdPartition = new 
TopicIdPartition(Uuid.randomUuid(), new TopicPartition("singleton", 0));
-        Assertions.assertThrows(RemoteResourceNotFoundException.class, () -> 
topicBasedRlmm().remoteLogSize(topicIdPartition, 0));
+        assertThrows(RemoteResourceNotFoundException.class, () -> 
topicBasedRlmm().remoteLogSize(topicIdPartition, 0));
     }
 
     @ClusterTest
@@ -206,8 +208,8 @@ public class TopicBasedRemoteLogMetadataManagerTest {
 
         // RemoteLogSegmentMetadata events are already published, and 
topicBasedRlmm's consumer manager will start
         // fetching those events and build the cache.
-        Assertions.assertTrue(initializationLatch.await(30_000, 
TimeUnit.MILLISECONDS));
-        
Assertions.assertTrue(handleRemoteLogSegmentMetadataLatch.await(30_000, 
TimeUnit.MILLISECONDS));
+        assertTrue(initializationLatch.await(30_000, TimeUnit.MILLISECONDS));
+        assertTrue(handleRemoteLogSegmentMetadataLatch.await(30_000, 
TimeUnit.MILLISECONDS));
 
         
verify(spyRemotePartitionMetadataEventHandler).markInitialized(topicIdPartition);
         
verify(spyRemotePartitionMetadataEventHandler).handleRemoteLogSegmentMetadata(firstSegmentMetadata);
@@ -215,7 +217,7 @@ public class TopicBasedRemoteLogMetadataManagerTest {
         
verify(spyRemotePartitionMetadataEventHandler).handleRemoteLogSegmentMetadata(thirdSegmentMetadata);
         Long remoteLogSize = 
topicBasedRemoteLogMetadataManager.remoteLogSize(topicIdPartition, 0);
 
-        Assertions.assertEquals(SEG_SIZE * 6, remoteLogSize);
+        assertEquals(SEG_SIZE * 6, remoteLogSize);
     }
 
     @ClusterTest
@@ -251,16 +253,16 @@ public class TopicBasedRemoteLogMetadataManagerTest {
 
         // RemoteLogSegmentMetadata events are already published, and 
topicBasedRlmm's consumer manager will start
         // fetching those events and build the cache.
-        Assertions.assertTrue(initializationLatch.await(30_000, 
TimeUnit.MILLISECONDS));
-        
Assertions.assertTrue(handleRemoteLogSegmentMetadataLatch.await(30_000, 
TimeUnit.MILLISECONDS));
+        assertTrue(initializationLatch.await(30_000, TimeUnit.MILLISECONDS));
+        assertTrue(handleRemoteLogSegmentMetadataLatch.await(30_000, 
TimeUnit.MILLISECONDS));
 
         
verify(spyRemotePartitionMetadataEventHandler).markInitialized(topicIdPartition);
         
verify(spyRemotePartitionMetadataEventHandler).handleRemoteLogSegmentMetadata(firstSegmentMetadata);
         
verify(spyRemotePartitionMetadataEventHandler).handleRemoteLogSegmentMetadata(secondSegmentMetadata);
         
verify(spyRemotePartitionMetadataEventHandler).handleRemoteLogSegmentMetadata(thirdSegmentMetadata);
-        Assertions.assertEquals(SEG_SIZE, 
topicBasedRemoteLogMetadataManager.remoteLogSize(topicIdPartition, 0));
-        Assertions.assertEquals(SEG_SIZE * 2, 
topicBasedRemoteLogMetadataManager.remoteLogSize(topicIdPartition, 1));
-        Assertions.assertEquals(SEG_SIZE * 3, 
topicBasedRemoteLogMetadataManager.remoteLogSize(topicIdPartition, 2));
+        assertEquals(SEG_SIZE, 
topicBasedRemoteLogMetadataManager.remoteLogSize(topicIdPartition, 0));
+        assertEquals(SEG_SIZE * 2, 
topicBasedRemoteLogMetadataManager.remoteLogSize(topicIdPartition, 1));
+        assertEquals(SEG_SIZE * 3, 
topicBasedRemoteLogMetadataManager.remoteLogSize(topicIdPartition, 2));
     }
 
     @ClusterTest
@@ -293,13 +295,13 @@ public class TopicBasedRemoteLogMetadataManagerTest {
 
         // RemoteLogSegmentMetadata events are already published, and 
topicBasedRlmm's consumer manager will start
         // fetching those events and build the cache.
-        Assertions.assertTrue(initializationLatch.await(30_000, 
TimeUnit.MILLISECONDS));
-        
Assertions.assertTrue(handleRemoteLogSegmentMetadataLatch.await(30_000, 
TimeUnit.MILLISECONDS));
+        assertTrue(initializationLatch.await(30_000, TimeUnit.MILLISECONDS));
+        assertTrue(handleRemoteLogSegmentMetadataLatch.await(30_000, 
TimeUnit.MILLISECONDS));
 
         
verify(spyRemotePartitionMetadataEventHandler).markInitialized(topicIdPartition);
         
verify(spyRemotePartitionMetadataEventHandler).handleRemoteLogSegmentMetadata(firstSegmentMetadata);
         
verify(spyRemotePartitionMetadataEventHandler).handleRemoteLogSegmentMetadata(secondSegmentMetadata);
-        Assertions.assertEquals(0, 
topicBasedRemoteLogMetadataManager.remoteLogSize(topicIdPartition, 9001));
+        assertEquals(0, 
topicBasedRemoteLogMetadataManager.remoteLogSize(topicIdPartition, 9001));
     }
 
 }
diff --git 
a/storage/src/test/java/org/apache/kafka/server/log/remote/storage/RemoteLogMetadataManagerTest.java
 
b/storage/src/test/java/org/apache/kafka/server/log/remote/storage/RemoteLogMetadataManagerTest.java
index b67c316e5e4..d970dd8b68c 100644
--- 
a/storage/src/test/java/org/apache/kafka/server/log/remote/storage/RemoteLogMetadataManagerTest.java
+++ 
b/storage/src/test/java/org/apache/kafka/server/log/remote/storage/RemoteLogMetadataManagerTest.java
@@ -23,33 +23,39 @@ import kafka.test.junit.ClusterTestExtensions;
 import org.apache.kafka.common.TopicIdPartition;
 import org.apache.kafka.common.TopicPartition;
 import org.apache.kafka.common.Uuid;
-import org.apache.kafka.common.utils.MockTime;
+import org.apache.kafka.common.utils.SystemTime;
 import org.apache.kafka.common.utils.Time;
 import 
org.apache.kafka.server.log.remote.metadata.storage.RemoteLogMetadataManagerTestUtils;
 import 
org.apache.kafka.server.log.remote.metadata.storage.RemotePartitionMetadataStore;
 import 
org.apache.kafka.server.log.remote.metadata.storage.TopicBasedRemoteLogMetadataManager;
-import org.junit.jupiter.api.Assertions;
 import org.junit.jupiter.api.Tag;
 import org.junit.jupiter.api.extension.ExtendWith;
 
-//import java.io.IOException;
 import java.util.Collections;
 import java.util.HashMap;
 import java.util.Map;
 import java.util.Optional;
 
+import static org.junit.jupiter.api.Assertions.assertDoesNotThrow;
+import static org.junit.jupiter.api.Assertions.assertEquals;
+import static org.junit.jupiter.api.Assertions.assertFalse;
+import static org.junit.jupiter.api.Assertions.assertThrows;
+
+import static 
org.apache.kafka.server.log.remote.storage.RemoteLogSegmentState.COPY_SEGMENT_FINISHED;
+import static 
org.apache.kafka.server.log.remote.storage.RemotePartitionDeleteState.DELETE_PARTITION_FINISHED;
+import static 
org.apache.kafka.server.log.remote.storage.RemotePartitionDeleteState.DELETE_PARTITION_MARKED;
+import static 
org.apache.kafka.server.log.remote.storage.RemotePartitionDeleteState.DELETE_PARTITION_STARTED;
+
 @Tag("integration")
 @ExtendWith(ClusterTestExtensions.class)
 @ClusterTestDefaults(brokers = 3)
 public class RemoteLogMetadataManagerTest {
     private final ClusterInstance clusterInstance;
-
     private static final TopicIdPartition TP0 = new 
TopicIdPartition(Uuid.randomUuid(), new TopicPartition("foo", 0));
-    private static final int SEG_SIZE = 1024 * 1024;
+    private static final int SEG_SIZE = 1048576;
     private static final int BROKER_ID_0 = 0;
     private static final int BROKER_ID_1 = 1;
-
-    private final Time time = new MockTime(1);
+    private final Time time = new SystemTime();
 
     RemoteLogMetadataManagerTest(ClusterInstance clusterInstance) {
         this.clusterInstance = clusterInstance;
@@ -71,26 +77,25 @@ public class RemoteLogMetadataManagerTest {
             // 1.Create a segment with state COPY_SEGMENT_STARTED, and this 
segment should not be available.
             Map<Integer, Long> segmentLeaderEpochs = 
Collections.singletonMap(0, 101L);
             RemoteLogSegmentId segmentId = new RemoteLogSegmentId(TP0, 
Uuid.randomUuid());
-            RemoteLogSegmentMetadata segmentMetadata = new 
RemoteLogSegmentMetadata(segmentId, 101L, 200L, -1L, BROKER_ID_0,
-                                                                               
     time.milliseconds(), SEG_SIZE, segmentLeaderEpochs);
+            RemoteLogSegmentMetadata segmentMetadata = new 
RemoteLogSegmentMetadata(
+                    segmentId, 101L, 200L, -1L, BROKER_ID_0, 
time.milliseconds(), SEG_SIZE, segmentLeaderEpochs);
             // Wait until the segment is added successfully.
-            Assertions.assertDoesNotThrow(() -> 
remoteLogMetadataManager.addRemoteLogSegmentMetadata(segmentMetadata).get());
+            assertDoesNotThrow(() -> 
remoteLogMetadataManager.addRemoteLogSegmentMetadata(segmentMetadata).get());
 
             // Search should not return the above segment.
-            
Assertions.assertFalse(remoteLogMetadataManager.remoteLogSegmentMetadata(TP0, 
0, 150).isPresent());
+            assertFalse(remoteLogMetadataManager.remoteLogSegmentMetadata(TP0, 
0, 150).isPresent());
 
             // 2.Move that segment to COPY_SEGMENT_FINISHED state and this 
segment should be available.
-            RemoteLogSegmentMetadataUpdate segmentMetadataUpdate = new 
RemoteLogSegmentMetadataUpdate(segmentId, time.milliseconds(),
-                    Optional.empty(),
-                    RemoteLogSegmentState.COPY_SEGMENT_FINISHED,
-                    BROKER_ID_1);
+            RemoteLogSegmentMetadataUpdate segmentMetadataUpdate = new 
RemoteLogSegmentMetadataUpdate(
+                    segmentId, time.milliseconds(), Optional.empty(), 
COPY_SEGMENT_FINISHED, BROKER_ID_1);
             // Wait until the segment is updated successfully.
-            Assertions.assertDoesNotThrow(() -> 
remoteLogMetadataManager.updateRemoteLogSegmentMetadata(segmentMetadataUpdate).get());
+            assertDoesNotThrow(() -> 
remoteLogMetadataManager.updateRemoteLogSegmentMetadata(segmentMetadataUpdate).get());
             RemoteLogSegmentMetadata expectedSegmentMetadata = 
segmentMetadata.createWithUpdates(segmentMetadataUpdate);
 
             // Search should return the above segment.
-            Optional<RemoteLogSegmentMetadata> segmentMetadataForOffset150 = 
remoteLogMetadataManager.remoteLogSegmentMetadata(TP0, 0, 150);
-            Assertions.assertEquals(Optional.of(expectedSegmentMetadata), 
segmentMetadataForOffset150);
+            Optional<RemoteLogSegmentMetadata> segmentMetadataForOffset150 =
+                    remoteLogMetadataManager.remoteLogSegmentMetadata(TP0, 0, 
150);
+            assertEquals(Optional.of(expectedSegmentMetadata), 
segmentMetadataForOffset150);
         }
     }
 
@@ -111,46 +116,46 @@ public class RemoteLogMetadataManagerTest {
             segmentLeaderEpochs.put(2, 50L);
             segmentLeaderEpochs.put(3, 80L);
             RemoteLogSegmentId segmentId = new RemoteLogSegmentId(TP0, 
Uuid.randomUuid());
-            RemoteLogSegmentMetadata segmentMetadata = new 
RemoteLogSegmentMetadata(segmentId, 0L, 100L,
-                                                                               
     -1L, BROKER_ID_0, time.milliseconds(), SEG_SIZE,
-                                                                               
     segmentLeaderEpochs);
+            RemoteLogSegmentMetadata segmentMetadata = new 
RemoteLogSegmentMetadata(
+                    segmentId, 0L, 100L, -1L, BROKER_ID_0, 
time.milliseconds(), SEG_SIZE, segmentLeaderEpochs);
             // Wait until the segment is added successfully.
-            Assertions.assertDoesNotThrow(() -> 
remoteLogMetadataManager.addRemoteLogSegmentMetadata(segmentMetadata).get());
+            assertDoesNotThrow(() -> 
remoteLogMetadataManager.addRemoteLogSegmentMetadata(segmentMetadata).get());
 
             RemoteLogSegmentMetadataUpdate segmentMetadataUpdate = new 
RemoteLogSegmentMetadataUpdate(
-                    segmentId, time.milliseconds(), Optional.empty(), 
RemoteLogSegmentState.COPY_SEGMENT_FINISHED, BROKER_ID_1);
+                    segmentId, time.milliseconds(), Optional.empty(), 
COPY_SEGMENT_FINISHED, BROKER_ID_1);
             // Wait until the segment is updated successfully.
-            Assertions.assertDoesNotThrow(() -> 
remoteLogMetadataManager.updateRemoteLogSegmentMetadata(segmentMetadataUpdate).get());
+            assertDoesNotThrow(() -> 
remoteLogMetadataManager.updateRemoteLogSegmentMetadata(segmentMetadataUpdate).get());
 
             RemoteLogSegmentMetadata expectedSegMetadata = 
segmentMetadata.createWithUpdates(segmentMetadataUpdate);
 
             // Check that the segment exists in RLMM.
-            Optional<RemoteLogSegmentMetadata> segMetadataForOffset30Epoch1 = 
remoteLogMetadataManager.remoteLogSegmentMetadata(TP0, 1, 30L);
-            Assertions.assertEquals(Optional.of(expectedSegMetadata), 
segMetadataForOffset30Epoch1);
+            Optional<RemoteLogSegmentMetadata> segMetadataForOffset30Epoch1 =
+                    remoteLogMetadataManager.remoteLogSegmentMetadata(TP0, 1, 
30L);
+            assertEquals(Optional.of(expectedSegMetadata), 
segMetadataForOffset30Epoch1);
 
             // Mark the partition for deletion and wait for it to be updated 
successfully.
-            Assertions.assertDoesNotThrow(() -> 
remoteLogMetadataManager.putRemotePartitionDeleteMetadata(
-                    
createRemotePartitionDeleteMetadata(RemotePartitionDeleteState.DELETE_PARTITION_MARKED)).get());
+            assertDoesNotThrow(() -> 
remoteLogMetadataManager.putRemotePartitionDeleteMetadata(
+                    
createRemotePartitionDeleteMetadata(DELETE_PARTITION_MARKED)).get());
 
-            Optional<RemoteLogSegmentMetadata> segmentMetadataAfterDelMark = 
remoteLogMetadataManager.remoteLogSegmentMetadata(TP0,
-                                                                               
                                                1, 30L);
-            Assertions.assertEquals(Optional.of(expectedSegMetadata), 
segmentMetadataAfterDelMark);
+            Optional<RemoteLogSegmentMetadata> segmentMetadataAfterDelMark =
+                    remoteLogMetadataManager.remoteLogSegmentMetadata(TP0, 1, 
30L);
+            assertEquals(Optional.of(expectedSegMetadata), 
segmentMetadataAfterDelMark);
 
             // Set the partition deletion state as started. Partition and 
segments should still be accessible as they are not
             // yet deleted. Wait until the segment state is updated 
successfully.
-            Assertions.assertDoesNotThrow(() -> 
remoteLogMetadataManager.putRemotePartitionDeleteMetadata(
-                    
createRemotePartitionDeleteMetadata(RemotePartitionDeleteState.DELETE_PARTITION_STARTED)).get());
+            assertDoesNotThrow(() -> 
remoteLogMetadataManager.putRemotePartitionDeleteMetadata(
+                    
createRemotePartitionDeleteMetadata(DELETE_PARTITION_STARTED)).get());
 
-            Optional<RemoteLogSegmentMetadata> segmentMetadataAfterDelStart = 
remoteLogMetadataManager.remoteLogSegmentMetadata(TP0,
-                                                                               
                                                 1, 30L);
-            Assertions.assertEquals(Optional.of(expectedSegMetadata), 
segmentMetadataAfterDelStart);
+            Optional<RemoteLogSegmentMetadata> segmentMetadataAfterDelStart =
+                    remoteLogMetadataManager.remoteLogSegmentMetadata(TP0, 1, 
30L);
+            assertEquals(Optional.of(expectedSegMetadata), 
segmentMetadataAfterDelStart);
 
             // Set the partition deletion state as finished. RLMM should clear 
all its internal state for that partition.
             // Wait until the segment state is updated successfully.
-            Assertions.assertDoesNotThrow(() -> 
remoteLogMetadataManager.putRemotePartitionDeleteMetadata(
-                    
createRemotePartitionDeleteMetadata(RemotePartitionDeleteState.DELETE_PARTITION_FINISHED)).get());
+            assertDoesNotThrow(() -> 
remoteLogMetadataManager.putRemotePartitionDeleteMetadata(
+                    
createRemotePartitionDeleteMetadata(DELETE_PARTITION_FINISHED)).get());
 
-            Assertions.assertThrows(RemoteResourceNotFoundException.class,
+            assertThrows(RemoteResourceNotFoundException.class,
                 () -> remoteLogMetadataManager.remoteLogSegmentMetadata(TP0, 
1, 30L));
         }
     }

Reply via email to