This is an automated email from the ASF dual-hosted git repository.

github-bot pushed a commit to branch update/scalafmt-core-3.9.10
in repository https://gitbox.apache.org/repos/asf/pekko-connectors-kafka.git

commit 505a5761842906fff940554785dc94d98c89ad04
Author: scala-steward-asf[bot] 
<147768647+scala-steward-asf[bot]@users.noreply.github.com>
AuthorDate: Wed Dec 3 13:58:14 2025 +0000

    Reformat with scalafmt 3.9.10
    
    Executed command: scalafmt --non-interactive
---
 .../scala/org/apache/pekko/kafka/benchmarks/InflightMetrics.scala     | 2 +-
 core/src/main/scala/org/apache/pekko/kafka/CommitterSettings.scala    | 2 +-
 .../scala/org/apache/pekko/kafka/internal/DefaultProducerStage.scala  | 2 +-
 .../main/scala/org/apache/pekko/kafka/internal/DeferredProducer.scala | 2 +-
 .../scala/org/apache/pekko/kafka/internal/KafkaConsumerActor.scala    | 4 ++--
 .../main/scala/org/apache/pekko/kafka/internal/MessageBuilder.scala   | 2 +-
 tests/src/test/scala/docs/scaladsl/proto/Order.scala                  | 2 +-
 .../test/scala/org/apache/pekko/kafka/scaladsl/IntegrationSpec.scala  | 4 ++--
 .../test/scala/org/apache/pekko/kafka/tests/CapturingAppender.scala   | 2 +-
 tests/src/test/scala/org/apache/pekko/kafka/tests/LogbackUtil.scala   | 4 ++--
 10 files changed, 13 insertions(+), 13 deletions(-)

diff --git 
a/benchmarks/src/main/scala/org/apache/pekko/kafka/benchmarks/InflightMetrics.scala
 
b/benchmarks/src/main/scala/org/apache/pekko/kafka/benchmarks/InflightMetrics.scala
index 3d5e5f98..0ff50627 100644
--- 
a/benchmarks/src/main/scala/org/apache/pekko/kafka/benchmarks/InflightMetrics.scala
+++ 
b/benchmarks/src/main/scala/org/apache/pekko/kafka/benchmarks/InflightMetrics.scala
@@ -263,7 +263,7 @@ private[benchmarks] object InflightMetrics {
 
   def reset(measurements: List[Measurement], registry: MetricRegistry): 
List[Metric] = measurements.map {
     case measurement @ Measurement(_, _, CounterMetricType, _) => 
BaseCounter(measurement, measurement)
-    case measurement @ Measurement(_, _, GaugeMetricType, _) =>
+    case measurement @ Measurement(_, _, GaugeMetricType, _)   =>
       HistogramGauge(measurement, registry.histogram(measurement.name))
   }
 
diff --git a/core/src/main/scala/org/apache/pekko/kafka/CommitterSettings.scala 
b/core/src/main/scala/org/apache/pekko/kafka/CommitterSettings.scala
index 08edde4d..9ba19bd8 100644
--- a/core/src/main/scala/org/apache/pekko/kafka/CommitterSettings.scala
+++ b/core/src/main/scala/org/apache/pekko/kafka/CommitterSettings.scala
@@ -101,7 +101,7 @@ object CommitWhen {
   def valueOf(s: String): CommitWhen = s match {
     case "OffsetFirstObserved" => OffsetFirstObserved
     case "NextOffsetObserved"  => NextOffsetObserved
-    case other =>
+    case other                 =>
       throw new IllegalArgumentException(
         s"allowed values are: OffsetFirstObserved, NextOffsetObserved. 
Received: $other")
   }
diff --git 
a/core/src/main/scala/org/apache/pekko/kafka/internal/DefaultProducerStage.scala
 
b/core/src/main/scala/org/apache/pekko/kafka/internal/DefaultProducerStage.scala
index ac1e1895..38912d69 100644
--- 
a/core/src/main/scala/org/apache/pekko/kafka/internal/DefaultProducerStage.scala
+++ 
b/core/src/main/scala/org/apache/pekko/kafka/internal/DefaultProducerStage.scala
@@ -190,7 +190,7 @@ private class DefaultProducerStageLogic[K, V, P, IN <: 
Envelope[K, V, P], OUT <:
       } else
         decider(exception) match {
           case Supervision.Stop => closeAndFailStageCb.invoke(exception)
-          case _ =>
+          case _                =>
             promise.failure(exception)
             confirmAndCheckForCompletionCB.invoke(())
         }
diff --git 
a/core/src/main/scala/org/apache/pekko/kafka/internal/DeferredProducer.scala 
b/core/src/main/scala/org/apache/pekko/kafka/internal/DeferredProducer.scala
index 746690f8..6e44d72d 100644
--- a/core/src/main/scala/org/apache/pekko/kafka/internal/DeferredProducer.scala
+++ b/core/src/main/scala/org/apache/pekko/kafka/internal/DeferredProducer.scala
@@ -79,7 +79,7 @@ private[kafka] trait DeferredProducer[K, V] {
     producerFuture.value match {
       case Some(Success(p)) => assignProducer(p)
       case Some(Failure(e)) => failStage(e)
-      case None =>
+      case None             =>
         val assign = getAsyncCallback(assignProducer)
         producerFuture
           .transform(
diff --git 
a/core/src/main/scala/org/apache/pekko/kafka/internal/KafkaConsumerActor.scala 
b/core/src/main/scala/org/apache/pekko/kafka/internal/KafkaConsumerActor.scala
index 848f5b36..fa1356a5 100644
--- 
a/core/src/main/scala/org/apache/pekko/kafka/internal/KafkaConsumerActor.scala
+++ 
b/core/src/main/scala/org/apache/pekko/kafka/internal/KafkaConsumerActor.scala
@@ -389,7 +389,7 @@ import scala.util.control.NonFatal
   def stopping: Receive = LoggingReceive.withLabel("stopping") {
     case p: Poll[_, _] =>
       receivePoll(p)
-    case _: StopLike =>
+    case _: StopLike     =>
     case Terminated(ref) =>
       stageActorsMap = stageActorsMap.filterNot(_._2 == ref)
     case _ @(_: Commit | _: RequestMessages) =>
@@ -403,7 +403,7 @@ import scala.util.control.NonFatal
     log.debug("Starting {}", self)
     val updateSettings: Future[ConsumerSettings[K, V]] = _settings.enriched
     updateSettings.value match {
-      case Some(Success(s)) => applySettings(s)
+      case Some(Success(s))            => applySettings(s)
       case Some(scala.util.Failure(e)) =>
         owner.foreach(_ ! Failure(e))
         throw e
diff --git 
a/core/src/main/scala/org/apache/pekko/kafka/internal/MessageBuilder.scala 
b/core/src/main/scala/org/apache/pekko/kafka/internal/MessageBuilder.scala
index 4ca02f29..f985faac 100644
--- a/core/src/main/scala/org/apache/pekko/kafka/internal/MessageBuilder.scala
+++ b/core/src/main/scala/org/apache/pekko/kafka/internal/MessageBuilder.scala
@@ -192,7 +192,7 @@ private[kafka] final class CommittableOffsetBatchImpl(
 
     val newCommitter = newOffset match {
       case c: CommittableOffsetImpl => c.committer
-      case _ =>
+      case _                        =>
         throw new IllegalArgumentException(
           s"Unknown CommittableOffset, got [${newOffset.getClass.getName}], " +
           s"expected [${classOf[CommittableOffsetImpl].getName}]")
diff --git a/tests/src/test/scala/docs/scaladsl/proto/Order.scala 
b/tests/src/test/scala/docs/scaladsl/proto/Order.scala
index a18627b5..0c897461 100644
--- a/tests/src/test/scala/docs/scaladsl/proto/Order.scala
+++ b/tests/src/test/scala/docs/scaladsl/proto/Order.scala
@@ -89,7 +89,7 @@ object Order extends 
scalapb.GeneratedMessageCompanion[docs.scaladsl.proto.Order
     while (!_done__) {
       val _tag__ = _input__.readTag()
       _tag__ match {
-        case 0 => _done__ = true
+        case 0  => _done__ = true
         case 10 =>
           __id = _input__.readStringRequireUtf8()
         case tag =>
diff --git 
a/tests/src/test/scala/org/apache/pekko/kafka/scaladsl/IntegrationSpec.scala 
b/tests/src/test/scala/org/apache/pekko/kafka/scaladsl/IntegrationSpec.scala
index d8b20fae..1058face 100644
--- a/tests/src/test/scala/org/apache/pekko/kafka/scaladsl/IntegrationSpec.scala
+++ b/tests/src/test/scala/org/apache/pekko/kafka/scaladsl/IntegrationSpec.scala
@@ -126,8 +126,8 @@ class IntegrationSpec extends SpecBase with 
TestcontainersKafkaLike with Inside
       val Partitions1 = Set(allTps(0), allTps(1))
       val Partitions2 = Set(allTps(2), allTps(3))
       (assigned1.topicPartitions, assigned2.topicPartitions) match {
-        case (Partitions1, Partitions2) =>
-        case (Partitions2, Partitions1) =>
+        case (Partitions1, Partitions2)                =>
+        case (Partitions2, Partitions1)                =>
         case (receivePartitions1, receivedPartitions2) =>
           fail(
             s"The `TopicPartitionsAssigned` contained different topic 
partitions than expected:\nrebalanceActor1: 
$receivePartitions1\nrebalanceActor2: $receivedPartitions2")
diff --git 
a/tests/src/test/scala/org/apache/pekko/kafka/tests/CapturingAppender.scala 
b/tests/src/test/scala/org/apache/pekko/kafka/tests/CapturingAppender.scala
index ffa4a2b3..d6817aaa 100644
--- a/tests/src/test/scala/org/apache/pekko/kafka/tests/CapturingAppender.scala
+++ b/tests/src/test/scala/org/apache/pekko/kafka/tests/CapturingAppender.scala
@@ -36,7 +36,7 @@ import ch.qos.logback.core.AppenderBase
         throw new IllegalStateException(
           s"$CapturingAppenderName not defined for 
[${loggerNameOrRoot(loggerName)}] in logback-test.xml")
       case appender: CapturingAppender => appender
-      case other =>
+      case other                       =>
         throw new IllegalStateException(s"Unexpected $CapturingAppender: 
$other")
     }
   }
diff --git 
a/tests/src/test/scala/org/apache/pekko/kafka/tests/LogbackUtil.scala 
b/tests/src/test/scala/org/apache/pekko/kafka/tests/LogbackUtil.scala
index 1b181541..40eed960 100644
--- a/tests/src/test/scala/org/apache/pekko/kafka/tests/LogbackUtil.scala
+++ b/tests/src/test/scala/org/apache/pekko/kafka/tests/LogbackUtil.scala
@@ -30,7 +30,7 @@ import org.slf4j.LoggerFactory
   def getLogbackLogger(loggerName: String): ch.qos.logback.classic.Logger = {
     LoggerFactory.getLogger(loggerNameOrRoot(loggerName)) match {
       case logger: ch.qos.logback.classic.Logger => logger
-      case null =>
+      case null                                  =>
         throw new IllegalArgumentException(s"Couldn't find logger for 
[$loggerName].")
       case other =>
         throw new IllegalArgumentException(
@@ -45,7 +45,7 @@ import org.slf4j.LoggerFactory
       case ch.qos.logback.classic.Level.INFO_INT  => Level.INFO
       case ch.qos.logback.classic.Level.WARN_INT  => Level.WARN
       case ch.qos.logback.classic.Level.ERROR_INT => Level.ERROR
-      case _ =>
+      case _                                      =>
         throw new IllegalArgumentException("Level " + level.levelStr + ", " + 
level.levelInt + " is unknown.")
     }
   }


---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to