This is an automated email from the ASF dual-hosted git repository.

srowen pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/spark.git


The following commit(s) were added to refs/heads/master by this push:
     new 3b46cc81614 [SPARK-45512][CORE][SQL][SS][DSTREAM] Fix compilation 
warnings related to `other-nullary-override`
3b46cc81614 is described below

commit 3b46cc816143d5bb553e86e8b716c28982cb5748
Author: YangJie <yangji...@baidu.com>
AuthorDate: Tue Oct 17 07:34:06 2023 -0500

    [SPARK-45512][CORE][SQL][SS][DSTREAM] Fix compilation warnings related to 
`other-nullary-override`
    
    ### What changes were proposed in this pull request?
    This PR fixes two compilation warnings related to `other-nullary-override`
    
    ```
    [error] 
/Users/yangjie01/SourceCode/git/spark-mine-sbt/connector/connect/common/src/main/scala/org/apache/spark/sql/connect/client/CloseableIterator.scala:36:16:
 method with a single empty parameter list overrides method hasNext in trait 
Iterator defined without a parameter list [quickfixable]
    [error] Applicable -Wconf / nowarn filters for this fatal warning: 
msg=<part of the message>, cat=other-nullary-override, 
site=org.apache.spark.sql.connect.client.WrappedCloseableIterator
    [error]   override def hasNext(): Boolean = innerIterator.hasNext
    [error]                ^
    [error] 
/Users/yangjie01/SourceCode/git/spark-mine-sbt/connector/connect/common/src/main/scala/org/apache/spark/sql/connect/client/ExecutePlanResponseReattachableIterator.scala:136:16:
 method without a parameter list overrides method hasNext in class 
WrappedCloseableIterator defined with a single empty parameter list 
[quickfixable]
    [error] Applicable -Wconf / nowarn filters for this fatal warning: 
msg=<part of the message>, cat=other-nullary-override, 
site=org.apache.spark.sql.connect.client.ExecutePlanResponseReattachableIterator
    [error]   override def hasNext: Boolean = synchronized {
    [error]                ^
    [error] 
/Users/yangjie01/SourceCode/git/spark-mine-sbt/connector/connect/common/src/main/scala/org/apache/spark/sql/connect/client/GrpcExceptionConverter.scala:73:20:
 method without a parameter list overrides method hasNext in class 
WrappedCloseableIterator defined with a single empty parameter list 
[quickfixable]
    [error] Applicable -Wconf / nowarn filters for this fatal warning: 
msg=<part of the message>, cat=other-nullary-override, 
site=org.apache.spark.sql.connect.client.GrpcExceptionConverter.convertIterator
    [error]       override def hasNext: Boolean = {
    [error]                    ^
    [error] 
/Users/yangjie01/SourceCode/git/spark-mine-sbt/connector/connect/common/src/main/scala/org/apache/spark/sql/connect/client/GrpcRetryHandler.scala:77:18:
 method without a parameter list overrides method next in class 
WrappedCloseableIterator defined with a single empty parameter list 
[quickfixable]
    [error] Applicable -Wconf / nowarn filters for this fatal warning: 
msg=<part of the message>, cat=other-nullary-override, 
site=org.apache.spark.sql.connect.client.GrpcRetryHandler.RetryIterator
    [error]     override def next: U = {
    [error]                  ^
    [error] 
/Users/yangjie01/SourceCode/git/spark-mine-sbt/connector/connect/common/src/main/scala/org/apache/spark/sql/connect/client/GrpcRetryHandler.scala:81:18:
 method without a parameter list overrides method hasNext in class 
WrappedCloseableIterator defined with a single empty parameter list 
[quickfixable]
    [error] Applicable -Wconf / nowarn filters for this fatal warning: 
msg=<part of the message>, cat=other-nullary-override, 
site=org.apache.spark.sql.connect.client.GrpcRetryHandler.RetryIterator
    [error]     override def hasNext: Boolean = {
    [error]
    ```
    
    and removes the corresponding suppression rules from the compilation options
    
    ```
    "-Wconf:cat=other-nullary-override:wv",
    ```
    
    On the other hand, the code corresponding to the following three 
suppression rules no longer exists, so the corresponding suppression rules were 
also cleaned up in this pr.
    
    ```
    "-Wconf:cat=lint-multiarg-infix:wv",
    "-Wconf:msg=method with a single empty parameter list overrides method 
without any parameter list:s",
    "-Wconf:msg=method without a parameter list overrides a method with a 
single empty one:s",
    ```
    
    ### Why are the changes needed?
    Code clean up.
    
    ### Does this PR introduce _any_ user-facing change?
    No
    
    ### How was this patch tested?
    Pass GitHub Actions
    
    ### Was this patch authored or co-authored using generative AI tooling?
    No
    
    Closes #43332 from LuciferYang/other-nullary-override.
    
    Lead-authored-by: YangJie <yangji...@baidu.com>
    Co-authored-by: yangjie01 <yangji...@baidu.com>
    Signed-off-by: Sean Owen <sro...@gmail.com>
---
 .../org/apache/spark/sql/avro/AvroRowReaderSuite.scala | 10 +++++-----
 .../spark/sql/connect/client/CloseableIterator.scala   |  2 +-
 .../ExecutePlanResponseReattachableIterator.scala      |  4 ++--
 .../spark/sql/connect/client/GrpcRetryHandler.scala    |  2 +-
 .../sql/kafka010/KafkaDontFailOnDataLossSuite.scala    |  2 +-
 .../spark/streaming/kafka010/KafkaRDDSuite.scala       |  4 ++--
 .../main/scala/org/apache/spark/TaskContextImpl.scala  |  2 +-
 .../spark/deploy/history/FsHistoryProvider.scala       |  2 +-
 .../src/main/scala/org/apache/spark/rdd/PipedRDD.scala |  4 ++--
 .../org/apache/spark/scheduler/TaskSetManager.scala    |  2 +-
 .../org/apache/spark/storage/memory/MemoryStore.scala  |  2 +-
 .../scala/org/apache/spark/util/HadoopFSUtils.scala    |  4 ++--
 .../spark/util/collection/ExternalAppendOnlyMap.scala  |  2 +-
 .../apache/spark/util/collection/ExternalSorter.scala  |  2 +-
 .../src/test/scala/org/apache/spark/ShuffleSuite.scala |  2 +-
 pom.xml                                                |  7 -------
 project/SparkBuild.scala                               |  4 ----
 .../sql/catalyst/expressions/DynamicPruning.scala      |  2 +-
 .../FunctionTableSubqueryArgumentExpression.scala      |  2 +-
 .../catalyst/expressions/higherOrderFunctions.scala    |  2 +-
 .../spark/sql/catalyst/expressions/subquery.scala      |  8 ++++----
 .../execution/ExternalAppendOnlyUnsafeRowArray.scala   |  4 ++--
 .../execution/WholeStageCodegenEvaluatorFactory.scala  |  2 +-
 .../aggregate/ObjectAggregationIterator.scala          |  2 +-
 .../execution/joins/BroadcastNestedLoopJoinExec.scala  |  2 +-
 .../apache/spark/sql/execution/metric/SQLMetrics.scala |  2 +-
 .../streaming/state/HDFSBackedStateStoreMap.scala      |  4 ++--
 .../org/apache/spark/sql/execution/subquery.scala      |  2 +-
 .../scala/org/apache/spark/sql/SSBQuerySuite.scala     |  2 +-
 .../apache/spark/sql/SparkSessionExtensionSuite.scala  |  4 ++--
 .../sql/execution/command/v1/DropNamespaceSuite.scala  |  2 +-
 .../spark/sql/execution/metric/SQLMetricsSuite.scala   | 18 +++++++++---------
 .../sql/execution/ui/AllExecutionsPageSuite.scala      |  2 +-
 .../spark/sql/hive/execution/Hive_2_1_DDLSuite.scala   |  2 +-
 .../hive/execution/PruneHiveTablePartitionsSuite.scala |  2 +-
 .../org/apache/spark/streaming/InputStreamsSuite.scala |  2 +-
 36 files changed, 56 insertions(+), 67 deletions(-)

diff --git 
a/connector/avro/src/test/scala/org/apache/spark/sql/avro/AvroRowReaderSuite.scala
 
b/connector/avro/src/test/scala/org/apache/spark/sql/avro/AvroRowReaderSuite.scala
index 70d0bc6c0ad..1c450d4f166 100644
--- 
a/connector/avro/src/test/scala/org/apache/spark/sql/avro/AvroRowReaderSuite.scala
+++ 
b/connector/avro/src/test/scala/org/apache/spark/sql/avro/AvroRowReaderSuite.scala
@@ -80,19 +80,19 @@ class AvroRowReaderSuite
 
         override def hasNext: Boolean = hasNextRow
 
-        override def next: InternalRow = nextRow
+        override def next(): InternalRow = nextRow
       }
       assert(it.hasNext == true)
-      assert(it.next.getInt(0) == 1)
+      assert(it.next().getInt(0) == 1)
       // test no intervening next
       assert(it.hasNext == true)
       assert(it.hasNext == true)
       // test no intervening hasNext
-      assert(it.next.getInt(0) == 2)
-      assert(it.next.getInt(0) == 3)
+      assert(it.next().getInt(0) == 2)
+      assert(it.next().getInt(0) == 3)
       assert(it.hasNext == false)
       assertThrows[NoSuchElementException] {
-        it.next
+        it.next()
       }
     }
   }
diff --git 
a/connector/connect/common/src/main/scala/org/apache/spark/sql/connect/client/CloseableIterator.scala
 
b/connector/connect/common/src/main/scala/org/apache/spark/sql/connect/client/CloseableIterator.scala
index 810158b2ac8..4ec6828d885 100644
--- 
a/connector/connect/common/src/main/scala/org/apache/spark/sql/connect/client/CloseableIterator.scala
+++ 
b/connector/connect/common/src/main/scala/org/apache/spark/sql/connect/client/CloseableIterator.scala
@@ -33,7 +33,7 @@ private[sql] abstract class WrappedCloseableIterator[E] 
extends CloseableIterato
 
   override def next(): E = innerIterator.next()
 
-  override def hasNext(): Boolean = innerIterator.hasNext
+  override def hasNext: Boolean = innerIterator.hasNext
 
   override def close(): Unit = innerIterator match {
     case it: CloseableIterator[E] => it.close()
diff --git 
a/connector/connect/common/src/main/scala/org/apache/spark/sql/connect/client/ExecutePlanResponseReattachableIterator.scala
 
b/connector/connect/common/src/main/scala/org/apache/spark/sql/connect/client/ExecutePlanResponseReattachableIterator.scala
index 4c0c1d1f390..9d134f59354 100644
--- 
a/connector/connect/common/src/main/scala/org/apache/spark/sql/connect/client/ExecutePlanResponseReattachableIterator.scala
+++ 
b/connector/connect/common/src/main/scala/org/apache/spark/sql/connect/client/ExecutePlanResponseReattachableIterator.scala
@@ -108,7 +108,7 @@ class ExecutePlanResponseReattachableIterator(
 
   override def next(): proto.ExecutePlanResponse = synchronized {
     // hasNext will trigger reattach in case the stream completed without 
resultComplete
-    if (!hasNext()) {
+    if (!hasNext) {
       throw new java.util.NoSuchElementException()
     }
 
@@ -133,7 +133,7 @@ class ExecutePlanResponseReattachableIterator(
     }
   }
 
-  override def hasNext(): Boolean = synchronized {
+  override def hasNext: Boolean = synchronized {
     if (resultComplete) {
       // After response complete response
       return false
diff --git 
a/connector/connect/common/src/main/scala/org/apache/spark/sql/connect/client/GrpcRetryHandler.scala
 
b/connector/connect/common/src/main/scala/org/apache/spark/sql/connect/client/GrpcRetryHandler.scala
index 3c0b750fd46..74c8423ce43 100644
--- 
a/connector/connect/common/src/main/scala/org/apache/spark/sql/connect/client/GrpcRetryHandler.scala
+++ 
b/connector/connect/common/src/main/scala/org/apache/spark/sql/connect/client/GrpcRetryHandler.scala
@@ -74,7 +74,7 @@ private[sql] class GrpcRetryHandler(
       }
     }
 
-    override def next: U = {
+    override def next(): U = {
       retryIter(_.next)
     }
 
diff --git 
a/connector/kafka-0-10-sql/src/test/scala/org/apache/spark/sql/kafka010/KafkaDontFailOnDataLossSuite.scala
 
b/connector/kafka-0-10-sql/src/test/scala/org/apache/spark/sql/kafka010/KafkaDontFailOnDataLossSuite.scala
index 54ce1717acc..2e0690f2288 100644
--- 
a/connector/kafka-0-10-sql/src/test/scala/org/apache/spark/sql/kafka010/KafkaDontFailOnDataLossSuite.scala
+++ 
b/connector/kafka-0-10-sql/src/test/scala/org/apache/spark/sql/kafka010/KafkaDontFailOnDataLossSuite.scala
@@ -47,7 +47,7 @@ trait KafkaMissingOffsetsTest extends SharedSparkSession {
 
   protected var testUtils: KafkaTestUtils = _
 
-  override def createSparkSession(): TestSparkSession = {
+  override def createSparkSession: TestSparkSession = {
     // Set maxRetries to 3 to handle NPE from `poll` when deleting a topic
     new TestSparkSession(new SparkContext("local[2,3]", "test-sql-context", 
sparkConf))
   }
diff --git 
a/connector/kafka-0-10/src/test/scala/org/apache/spark/streaming/kafka010/KafkaRDDSuite.scala
 
b/connector/kafka-0-10/src/test/scala/org/apache/spark/streaming/kafka010/KafkaRDDSuite.scala
index 735ec2f7b44..986943ac6a1 100644
--- 
a/connector/kafka-0-10/src/test/scala/org/apache/spark/streaming/kafka010/KafkaRDDSuite.scala
+++ 
b/connector/kafka-0-10/src/test/scala/org/apache/spark/streaming/kafka010/KafkaRDDSuite.scala
@@ -48,14 +48,14 @@ class KafkaRDDSuite extends SparkFunSuite {
 
   private var sc: SparkContext = _
 
-  override def beforeAll: Unit = {
+  override def beforeAll(): Unit = {
     super.beforeAll()
     sc = new SparkContext(sparkConf)
     kafkaTestUtils = new KafkaTestUtils
     kafkaTestUtils.setup()
   }
 
-  override def afterAll: Unit = {
+  override def afterAll(): Unit = {
     try {
       try {
         if (sc != null) {
diff --git a/core/src/main/scala/org/apache/spark/TaskContextImpl.scala 
b/core/src/main/scala/org/apache/spark/TaskContextImpl.scala
index 4cc5f165794..8d2c2ab9bc4 100644
--- a/core/src/main/scala/org/apache/spark/TaskContextImpl.scala
+++ b/core/src/main/scala/org/apache/spark/TaskContextImpl.scala
@@ -292,5 +292,5 @@ private[spark] class TaskContextImpl(
 
   private[spark] override def fetchFailed: Option[FetchFailedException] = 
_fetchFailedException
 
-  private[spark] override def getLocalProperties(): Properties = 
localProperties
+  private[spark] override def getLocalProperties: Properties = localProperties
 }
diff --git 
a/core/src/main/scala/org/apache/spark/deploy/history/FsHistoryProvider.scala 
b/core/src/main/scala/org/apache/spark/deploy/history/FsHistoryProvider.scala
index abb65c7311e..b5afa86180b 100644
--- 
a/core/src/main/scala/org/apache/spark/deploy/history/FsHistoryProvider.scala
+++ 
b/core/src/main/scala/org/apache/spark/deploy/history/FsHistoryProvider.scala
@@ -1288,7 +1288,7 @@ private[history] class FsHistoryProvider(conf: SparkConf, 
clock: Clock)
       val diskStore = KVUtils.open(lease.tmpPath, metadata, conf, live = false)
       hybridStore.setDiskStore(diskStore)
       hybridStore.switchToDiskStore(new HybridStore.SwitchToDiskStoreListener {
-        override def onSwitchToDiskStoreSuccess: Unit = {
+        override def onSwitchToDiskStoreSuccess(): Unit = {
           logInfo(s"Completely switched to diskStore for app $appId / 
${attempt.info.attemptId}.")
           diskStore.close()
           val newStorePath = lease.commit(appId, attempt.info.attemptId)
diff --git a/core/src/main/scala/org/apache/spark/rdd/PipedRDD.scala 
b/core/src/main/scala/org/apache/spark/rdd/PipedRDD.scala
index 0359f6c6d38..4f3683ef2df 100644
--- a/core/src/main/scala/org/apache/spark/rdd/PipedRDD.scala
+++ b/core/src/main/scala/org/apache/spark/rdd/PipedRDD.scala
@@ -185,13 +185,13 @@ private[spark] class PipedRDD[T: ClassTag](
     val lines = Source.fromInputStream(proc.getInputStream)(encoding).getLines
     new Iterator[String] {
       def next(): String = {
-        if (!hasNext()) {
+        if (!hasNext) {
           throw SparkCoreErrors.noSuchElementError()
         }
         lines.next()
       }
 
-      def hasNext(): Boolean = {
+      def hasNext: Boolean = {
         val result = if (lines.hasNext) {
           true
         } else {
diff --git 
a/core/src/main/scala/org/apache/spark/scheduler/TaskSetManager.scala 
b/core/src/main/scala/org/apache/spark/scheduler/TaskSetManager.scala
index 4d5f32ac67b..de9c58e8d90 100644
--- a/core/src/main/scala/org/apache/spark/scheduler/TaskSetManager.scala
+++ b/core/src/main/scala/org/apache/spark/scheduler/TaskSetManager.scala
@@ -1032,7 +1032,7 @@ private[spark] class TaskSetManager(
 
   override def removeSchedulable(schedulable: Schedulable): Unit = {}
 
-  override def getSortedTaskSetQueue(): ArrayBuffer[TaskSetManager] = {
+  override def getSortedTaskSetQueue: ArrayBuffer[TaskSetManager] = {
     val sortedTaskSetQueue = new ArrayBuffer[TaskSetManager]()
     sortedTaskSetQueue += this
     sortedTaskSetQueue
diff --git 
a/core/src/main/scala/org/apache/spark/storage/memory/MemoryStore.scala 
b/core/src/main/scala/org/apache/spark/storage/memory/MemoryStore.scala
index dafa67b091d..88bb3eb7105 100644
--- a/core/src/main/scala/org/apache/spark/storage/memory/MemoryStore.scala
+++ b/core/src/main/scala/org/apache/spark/storage/memory/MemoryStore.scala
@@ -746,7 +746,7 @@ private class SerializedValuesHolder[T](
     // We successfully unrolled the entirety of this block
     serializationStream.close()
 
-    override def preciseSize(): Long = bbos.size
+    override def preciseSize: Long = bbos.size
 
     override def build(): MemoryEntry[T] =
       SerializedMemoryEntry[T](bbos.toChunkedByteBuffer, memoryMode, classTag)
diff --git a/core/src/main/scala/org/apache/spark/util/HadoopFSUtils.scala 
b/core/src/main/scala/org/apache/spark/util/HadoopFSUtils.scala
index a033b817f5a..5cd93dfae35 100644
--- a/core/src/main/scala/org/apache/spark/util/HadoopFSUtils.scala
+++ b/core/src/main/scala/org/apache/spark/util/HadoopFSUtils.scala
@@ -91,7 +91,7 @@ private[spark] object HadoopFSUtils extends Logging {
       val remoteIter = path.getFileSystem(hadoopConf).listFiles(path, true)
       val statues = new Iterator[LocatedFileStatus]() {
         def next(): LocatedFileStatus = remoteIter.next
-        def hasNext(): Boolean = remoteIter.hasNext
+        def hasNext: Boolean = remoteIter.hasNext
       }.filterNot(status => 
shouldFilterOutPath(status.getPath.toString.substring(prefixLength)))
         .filter(f => filter.accept(f.getPath))
         .toArray
@@ -209,7 +209,7 @@ private[spark] object HadoopFSUtils extends Logging {
           val remoteIter = fs.listLocatedStatus(path)
           new Iterator[LocatedFileStatus]() {
             def next(): LocatedFileStatus = remoteIter.next
-            def hasNext(): Boolean = remoteIter.hasNext
+            def hasNext: Boolean = remoteIter.hasNext
           }.toArray
         case _ => fs.listStatus(path)
       }
diff --git 
a/core/src/main/scala/org/apache/spark/util/collection/ExternalAppendOnlyMap.scala
 
b/core/src/main/scala/org/apache/spark/util/collection/ExternalAppendOnlyMap.scala
index 93efaafa43b..71f3b0b4b4d 100644
--- 
a/core/src/main/scala/org/apache/spark/util/collection/ExternalAppendOnlyMap.scala
+++ 
b/core/src/main/scala/org/apache/spark/util/collection/ExternalAppendOnlyMap.scala
@@ -592,7 +592,7 @@ class ExternalAppendOnlyMap[K, V, C](
       }
     }
 
-    override def hasNext(): Boolean = cur != null
+    override def hasNext: Boolean = cur != null
 
     override def next(): (K, C) = {
       val r = cur
diff --git 
a/core/src/main/scala/org/apache/spark/util/collection/ExternalSorter.scala 
b/core/src/main/scala/org/apache/spark/util/collection/ExternalSorter.scala
index a68e0de70c5..42501729e1b 100644
--- a/core/src/main/scala/org/apache/spark/util/collection/ExternalSorter.scala
+++ b/core/src/main/scala/org/apache/spark/util/collection/ExternalSorter.scala
@@ -842,7 +842,7 @@ private[spark] class ExternalSorter[K, V, C](
       }
     }
 
-    override def hasNext(): Boolean = cur != null
+    override def hasNext: Boolean = cur != null
 
     override def next(): ((Int, K), C) = {
       val r = cur
diff --git a/core/src/test/scala/org/apache/spark/ShuffleSuite.scala 
b/core/src/test/scala/org/apache/spark/ShuffleSuite.scala
index c1b9af37ce7..0b6fce57727 100644
--- a/core/src/test/scala/org/apache/spark/ShuffleSuite.scala
+++ b/core/src/test/scala/org/apache/spark/ShuffleSuite.scala
@@ -502,7 +502,7 @@ class InterleaveIterators[T, R](
   class BarrierIterator[E](id: Int, sub: Iterator[E]) extends Iterator[E] {
     def hasNext: Boolean = sub.hasNext
 
-    def next: E = {
+    def next(): E = {
       barrier.await()
       sub.next()
     }
diff --git a/pom.xml b/pom.xml
index b6804dfb75f..824ae49f6da 100644
--- a/pom.xml
+++ b/pom.xml
@@ -2965,12 +2965,7 @@
               <arg>-target:17</arg>
               <arg>-Wconf:cat=deprecation:wv,any:e</arg>
               <arg>-Wunused:imports</arg>
-              <!--
-                TODO(SPARK-33805): Undo the corresponding deprecated usage 
suppression rule after fixed
-              -->
               <arg>-Wconf:cat=scaladoc:wv</arg>
-              <arg>-Wconf:cat=lint-multiarg-infix:wv</arg>
-              <arg>-Wconf:cat=other-nullary-override:wv</arg>
               <!--
                 SPARK-33775 Suppress compilation warnings that contain the 
following contents.
                 TODO(SPARK-33805): Undo the corresponding deprecated usage 
suppression rule after fixed.
@@ -2978,8 +2973,6 @@
               
<arg>-Wconf:msg=^(?=.*?method|value|type|object|trait|inheritance)(?=.*?deprecated)(?=.*?since
 2.13).+$:s</arg>
               <arg>-Wconf:msg=^(?=.*?Widening conversion from)(?=.*?is 
deprecated because it loses precision).+$:s</arg>
               <arg>-Wconf:msg=Auto-application to \`\(\)\` is 
deprecated:s</arg>
-              <arg>-Wconf:msg=method with a single empty parameter list 
overrides method without any parameter list:s</arg>
-              <arg>-Wconf:msg=method without a parameter list overrides a 
method with a single empty one:s</arg>
               <!--
                 SPARK-35574 Prevent the recurrence of compilation warnings 
related to
                 `procedure syntax is deprecated`
diff --git a/project/SparkBuild.scala b/project/SparkBuild.scala
index f433fbc78a2..d2c9771b0f8 100644
--- a/project/SparkBuild.scala
+++ b/project/SparkBuild.scala
@@ -232,16 +232,12 @@ object SparkBuild extends PomBuild {
         "-Wconf:cat=deprecation:wv,any:e",
         // 2.13-specific warning hits to be muted (as narrowly as possible) 
and addressed separately
         "-Wunused:imports",
-        "-Wconf:cat=lint-multiarg-infix:wv",
-        "-Wconf:cat=other-nullary-override:wv",
         // SPARK-33775 Suppress compilation warnings that contain the 
following contents.
         // TODO(SPARK-33805): Undo the corresponding deprecated usage 
suppression rule after
         //  fixed.
         
"-Wconf:msg=^(?=.*?method|value|type|object|trait|inheritance)(?=.*?deprecated)(?=.*?since
 2.13).+$:s",
         "-Wconf:msg=^(?=.*?Widening conversion from)(?=.*?is deprecated 
because it loses precision).+$:s",
         "-Wconf:msg=Auto-application to \\`\\(\\)\\` is deprecated:s",
-        "-Wconf:msg=method with a single empty parameter list overrides method 
without any parameter list:s",
-        "-Wconf:msg=method without a parameter list overrides a method with a 
single empty one:s",
         // SPARK-35574 Prevent the recurrence of compilation warnings related 
to `procedure syntax is deprecated`
         "-Wconf:cat=deprecation&msg=procedure syntax is deprecated:e",
         // SPARK-35496 Upgrade Scala to 2.13.7 and suppress:
diff --git 
a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/DynamicPruning.scala
 
b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/DynamicPruning.scala
index 1e94188bd18..ec6925eaa98 100644
--- 
a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/DynamicPruning.scala
+++ 
b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/DynamicPruning.scala
@@ -73,7 +73,7 @@ case class DynamicPruningSubquery(
       pruningKey.dataType == buildKeys(broadcastKeyIndex).dataType
   }
 
-  final override def nodePatternsInternal: Seq[TreePattern] = 
Seq(DYNAMIC_PRUNING_SUBQUERY)
+  final override def nodePatternsInternal(): Seq[TreePattern] = 
Seq(DYNAMIC_PRUNING_SUBQUERY)
 
   override def toString: String = s"dynamicpruning#${exprId.id} 
$conditionString"
 
diff --git 
a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/FunctionTableSubqueryArgumentExpression.scala
 
b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/FunctionTableSubqueryArgumentExpression.scala
index 209eb02e3b3..da2dd9b1e25 100644
--- 
a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/FunctionTableSubqueryArgumentExpression.scala
+++ 
b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/FunctionTableSubqueryArgumentExpression.scala
@@ -93,7 +93,7 @@ case class FunctionTableSubqueryArgumentExpression(
       newChildren: IndexedSeq[Expression]): 
FunctionTableSubqueryArgumentExpression =
     copy(outerAttrs = newChildren)
 
-  final override def nodePatternsInternal: Seq[TreePattern] =
+  final override def nodePatternsInternal(): Seq[TreePattern] =
     Seq(FUNCTION_TABLE_RELATION_ARGUMENT_EXPRESSION)
 
   def hasRepartitioning: Boolean = withSinglePartition || 
partitionByExpressions.nonEmpty
diff --git 
a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/higherOrderFunctions.scala
 
b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/higherOrderFunctions.scala
index fec1df108bc..4df6a5e7666 100644
--- 
a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/higherOrderFunctions.scala
+++ 
b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/higherOrderFunctions.scala
@@ -248,7 +248,7 @@ trait SimpleHigherOrderFunction extends HigherOrderFunction 
with BinaryLike[Expr
 
   def argumentType: AbstractDataType
 
-  override def argumentTypes(): Seq[AbstractDataType] = argumentType :: Nil
+  override def argumentTypes: Seq[AbstractDataType] = argumentType :: Nil
 
   def function: Expression
 
diff --git 
a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/subquery.scala
 
b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/subquery.scala
index bbcd9e336b4..9d8b5de271e 100644
--- 
a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/subquery.scala
+++ 
b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/subquery.scala
@@ -295,7 +295,7 @@ case class ScalarSubquery(
       outerAttrs = newChildren.take(outerAttrs.size),
       joinCond = newChildren.drop(outerAttrs.size))
 
-  final override def nodePatternsInternal: Seq[TreePattern] = 
Seq(SCALAR_SUBQUERY)
+  final override def nodePatternsInternal(): Seq[TreePattern] = 
Seq(SCALAR_SUBQUERY)
 }
 
 object ScalarSubquery {
@@ -339,7 +339,7 @@ case class LateralSubquery(
       outerAttrs = newChildren.take(outerAttrs.size),
       joinCond = newChildren.drop(outerAttrs.size))
 
-  final override def nodePatternsInternal: Seq[TreePattern] = 
Seq(LATERAL_SUBQUERY)
+  final override def nodePatternsInternal(): Seq[TreePattern] = 
Seq(LATERAL_SUBQUERY)
 }
 
 /**
@@ -397,7 +397,7 @@ case class ListQuery(
       outerAttrs = newChildren.take(outerAttrs.size),
       joinCond = newChildren.drop(outerAttrs.size))
 
-  final override def nodePatternsInternal: Seq[TreePattern] = 
Seq(LIST_SUBQUERY)
+  final override def nodePatternsInternal(): Seq[TreePattern] = 
Seq(LIST_SUBQUERY)
 }
 
 /**
@@ -452,5 +452,5 @@ case class Exists(
       outerAttrs = newChildren.take(outerAttrs.size),
       joinCond = newChildren.drop(outerAttrs.size))
 
-  final override def nodePatternsInternal: Seq[TreePattern] = 
Seq(EXISTS_SUBQUERY)
+  final override def nodePatternsInternal(): Seq[TreePattern] = 
Seq(EXISTS_SUBQUERY)
 }
diff --git 
a/sql/core/src/main/scala/org/apache/spark/sql/execution/ExternalAppendOnlyUnsafeRowArray.scala
 
b/sql/core/src/main/scala/org/apache/spark/sql/execution/ExternalAppendOnlyUnsafeRowArray.scala
index 4147d75186d..56289d73c07 100644
--- 
a/sql/core/src/main/scala/org/apache/spark/sql/execution/ExternalAppendOnlyUnsafeRowArray.scala
+++ 
b/sql/core/src/main/scala/org/apache/spark/sql/execution/ExternalAppendOnlyUnsafeRowArray.scala
@@ -209,7 +209,7 @@ private[sql] class ExternalAppendOnlyUnsafeRowArray(
 
     private var currentIndex = startIndex
 
-    override def hasNext(): Boolean = !isModified() && currentIndex < numRows
+    override def hasNext: Boolean = !isModified() && currentIndex < numRows
 
     override def next(): UnsafeRow = {
       throwExceptionIfModified()
@@ -226,7 +226,7 @@ private[sql] class ExternalAppendOnlyUnsafeRowArray(
 
     private val currentRow = new UnsafeRow(numFieldPerRow)
 
-    override def hasNext(): Boolean = !isModified() && iterator.hasNext
+    override def hasNext: Boolean = !isModified() && iterator.hasNext
 
     override def next(): UnsafeRow = {
       throwExceptionIfModified()
diff --git 
a/sql/core/src/main/scala/org/apache/spark/sql/execution/WholeStageCodegenEvaluatorFactory.scala
 
b/sql/core/src/main/scala/org/apache/spark/sql/execution/WholeStageCodegenEvaluatorFactory.scala
index edf5774cf02..f0f9c7cbc95 100644
--- 
a/sql/core/src/main/scala/org/apache/spark/sql/execution/WholeStageCodegenEvaluatorFactory.scala
+++ 
b/sql/core/src/main/scala/org/apache/spark/sql/execution/WholeStageCodegenEvaluatorFactory.scala
@@ -44,7 +44,7 @@ class WholeStageCodegenEvaluatorFactory(
           if (!v) durationMs += buffer.durationMs()
           v
         }
-        override def next: InternalRow = buffer.next()
+        override def next(): InternalRow = buffer.next()
       }
     }
   }
diff --git 
a/sql/core/src/main/scala/org/apache/spark/sql/execution/aggregate/ObjectAggregationIterator.scala
 
b/sql/core/src/main/scala/org/apache/spark/sql/execution/aggregate/ObjectAggregationIterator.scala
index fac3f7d6d8a..6d05be72b36 100644
--- 
a/sql/core/src/main/scala/org/apache/spark/sql/execution/aggregate/ObjectAggregationIterator.scala
+++ 
b/sql/core/src/main/scala/org/apache/spark/sql/execution/aggregate/ObjectAggregationIterator.scala
@@ -254,7 +254,7 @@ class SortBasedAggregator(
       private var result: AggregationBufferEntry = _
       private var groupingKey: UnsafeRow = _
 
-      override def hasNext(): Boolean = {
+      override def hasNext: Boolean = {
         result != null || findNextSortedGroup()
       }
 
diff --git 
a/sql/core/src/main/scala/org/apache/spark/sql/execution/joins/BroadcastNestedLoopJoinExec.scala
 
b/sql/core/src/main/scala/org/apache/spark/sql/execution/joins/BroadcastNestedLoopJoinExec.scala
index c56cc294a9f..32d23136225 100644
--- 
a/sql/core/src/main/scala/org/apache/spark/sql/execution/joins/BroadcastNestedLoopJoinExec.scala
+++ 
b/sql/core/src/main/scala/org/apache/spark/sql/execution/joins/BroadcastNestedLoopJoinExec.scala
@@ -181,7 +181,7 @@ case class BroadcastNestedLoopJoinExec(
           }
         }
 
-        override def hasNext(): Boolean = {
+        override def hasNext: Boolean = {
           resultRow != null || findNextMatch()
         }
         override def next(): InternalRow = {
diff --git 
a/sql/core/src/main/scala/org/apache/spark/sql/execution/metric/SQLMetrics.scala
 
b/sql/core/src/main/scala/org/apache/spark/sql/execution/metric/SQLMetrics.scala
index 3326c5d4cb9..69d288ae75c 100644
--- 
a/sql/core/src/main/scala/org/apache/spark/sql/execution/metric/SQLMetrics.scala
+++ 
b/sql/core/src/main/scala/org/apache/spark/sql/execution/metric/SQLMetrics.scala
@@ -63,7 +63,7 @@ class SQLMetric(val metricType: String, initValue: Long = 0L) 
extends Accumulato
       this.getClass.getName, other.getClass.getName)
   }
 
-  override def isZero(): Boolean = _value == _zeroValue
+  override def isZero: Boolean = _value == _zeroValue
 
   override def add(v: Long): Unit = {
     if (_value < 0) _value = 0
diff --git 
a/sql/core/src/main/scala/org/apache/spark/sql/execution/streaming/state/HDFSBackedStateStoreMap.scala
 
b/sql/core/src/main/scala/org/apache/spark/sql/execution/streaming/state/HDFSBackedStateStoreMap.scala
index a79a856b0ec..59b01b5b454 100644
--- 
a/sql/core/src/main/scala/org/apache/spark/sql/execution/streaming/state/HDFSBackedStateStoreMap.scala
+++ 
b/sql/core/src/main/scala/org/apache/spark/sql/execution/streaming/state/HDFSBackedStateStoreMap.scala
@@ -54,7 +54,7 @@ object HDFSBackedStateStoreMap {
 class NoPrefixHDFSBackedStateStoreMap extends HDFSBackedStateStoreMap {
   private val map = new HDFSBackedStateStoreMap.MapType()
 
-  override def size: Int = map.size()
+  override def size(): Int = map.size()
 
   override def get(key: UnsafeRow): UnsafeRow = map.get(key)
 
@@ -103,7 +103,7 @@ class PrefixScannableHDFSBackedStateStoreMap(
     UnsafeProjection.create(refs)
   }
 
-  override def size: Int = map.size()
+  override def size(): Int = map.size()
 
   override def get(key: UnsafeRow): UnsafeRow = map.get(key)
 
diff --git 
a/sql/core/src/main/scala/org/apache/spark/sql/execution/subquery.scala 
b/sql/core/src/main/scala/org/apache/spark/sql/execution/subquery.scala
index 41230c7792c..58f720154df 100644
--- a/sql/core/src/main/scala/org/apache/spark/sql/execution/subquery.scala
+++ b/sql/core/src/main/scala/org/apache/spark/sql/execution/subquery.scala
@@ -127,7 +127,7 @@ case class InSubqueryExec(
   override def nullable: Boolean = child.nullable
   override def toString: String = s"$child IN ${plan.name}"
   override def withNewPlan(plan: BaseSubqueryExec): InSubqueryExec = copy(plan 
= plan)
-  final override def nodePatternsInternal: Seq[TreePattern] = 
Seq(IN_SUBQUERY_EXEC)
+  final override def nodePatternsInternal(): Seq[TreePattern] = 
Seq(IN_SUBQUERY_EXEC)
 
   def updateResult(): Unit = {
     val rows = plan.executeCollect()
diff --git a/sql/core/src/test/scala/org/apache/spark/sql/SSBQuerySuite.scala 
b/sql/core/src/test/scala/org/apache/spark/sql/SSBQuerySuite.scala
index 099b559105f..8df91c6dbda 100644
--- a/sql/core/src/test/scala/org/apache/spark/sql/SSBQuerySuite.scala
+++ b/sql/core/src/test/scala/org/apache/spark/sql/SSBQuerySuite.scala
@@ -25,7 +25,7 @@ import org.apache.spark.sql.catalyst.util.resourceToString
  */
 class SSBQuerySuite extends BenchmarkQueryTest {
 
-  override def beforeAll: Unit = {
+  override def beforeAll(): Unit = {
     super.beforeAll
 
     sql(
diff --git 
a/sql/core/src/test/scala/org/apache/spark/sql/SparkSessionExtensionSuite.scala 
b/sql/core/src/test/scala/org/apache/spark/sql/SparkSessionExtensionSuite.scala
index 21518085ca4..c1b5d2761f7 100644
--- 
a/sql/core/src/test/scala/org/apache/spark/sql/SparkSessionExtensionSuite.scala
+++ 
b/sql/core/src/test/scala/org/apache/spark/sql/SparkSessionExtensionSuite.scala
@@ -824,7 +824,7 @@ class ColumnarWriteExec(
     staticPartitions: TablePartitionSpec) extends WriteFilesExec(
   child, fileFormat, partitionColumns, bucketSpec, options, staticPartitions) {
 
-  override def supportsColumnar(): Boolean = true
+  override def supportsColumnar: Boolean = true
 
   override def doExecuteWrite(writeFilesSpec: WriteFilesSpec): 
RDD[WriterCommitMessage] = {
     assert(child.supportsColumnar)
@@ -846,7 +846,7 @@ class BrokenColumnarAdd(
     failOnError: Boolean = false)
   extends Add(left, right, EvalMode.fromBoolean(failOnError)) with 
ColumnarExpression {
 
-  override def supportsColumnar(): Boolean = left.supportsColumnar && 
right.supportsColumnar
+  override def supportsColumnar: Boolean = left.supportsColumnar && 
right.supportsColumnar
 
   override def columnarEval(batch: ColumnarBatch): Any = {
     var lhs: Any = null
diff --git 
a/sql/core/src/test/scala/org/apache/spark/sql/execution/command/v1/DropNamespaceSuite.scala
 
b/sql/core/src/test/scala/org/apache/spark/sql/execution/command/v1/DropNamespaceSuite.scala
index ea03654f921..647247cc833 100644
--- 
a/sql/core/src/test/scala/org/apache/spark/sql/execution/command/v1/DropNamespaceSuite.scala
+++ 
b/sql/core/src/test/scala/org/apache/spark/sql/execution/command/v1/DropNamespaceSuite.scala
@@ -32,7 +32,7 @@ trait DropNamespaceSuiteBase extends 
command.DropNamespaceSuiteBase
   with command.TestsV1AndV2Commands {
   override protected def builtinTopNamespaces: Seq[String] = Seq("default")
 
-  override protected def namespaceAlias(): String = "database"
+  override protected def namespaceAlias: String = "database"
 
   test("drop default namespace") {
     checkError(
diff --git 
a/sql/core/src/test/scala/org/apache/spark/sql/execution/metric/SQLMetricsSuite.scala
 
b/sql/core/src/test/scala/org/apache/spark/sql/execution/metric/SQLMetricsSuite.scala
index 6347757e178..e71451f2f74 100644
--- 
a/sql/core/src/test/scala/org/apache/spark/sql/execution/metric/SQLMetricsSuite.scala
+++ 
b/sql/core/src/test/scala/org/apache/spark/sql/execution/metric/SQLMetricsSuite.scala
@@ -940,25 +940,25 @@ class SQLMetricsSuite extends SharedSparkSession with 
SQLMetricsTestUtils
     assert(SQLMetrics.createSizeMetric(sparkContext, name = "m", initValue = 
-1).value === 0)
     assert(SQLMetrics.createSizeMetric(sparkContext, name = "m", initValue = 
5).value === 5)
 
-    assert(SQLMetrics.createSizeMetric(sparkContext, name = "m").isZero())
-    assert(SQLMetrics.createSizeMetric(sparkContext, name = "m", initValue = 
-1).isZero())
-    assert(SQLMetrics.createSizeMetric(sparkContext, name = "m", initValue = 
5).isZero())
+    assert(SQLMetrics.createSizeMetric(sparkContext, name = "m").isZero)
+    assert(SQLMetrics.createSizeMetric(sparkContext, name = "m", initValue = 
-1).isZero)
+    assert(SQLMetrics.createSizeMetric(sparkContext, name = "m", initValue = 
5).isZero)
 
     assert(SQLMetrics.createTimingMetric(sparkContext, name = "m").value === 0)
     assert(SQLMetrics.createTimingMetric(sparkContext, name = "m", initValue = 
-1).value === 0)
     assert(SQLMetrics.createTimingMetric(sparkContext, name = "m", initValue = 
5).value === 5)
 
-    assert(SQLMetrics.createTimingMetric(sparkContext, name = "m").isZero())
-    assert(SQLMetrics.createTimingMetric(sparkContext, name = "m", initValue = 
-1).isZero())
-    assert(SQLMetrics.createTimingMetric(sparkContext, name = "m", initValue = 
5).isZero())
+    assert(SQLMetrics.createTimingMetric(sparkContext, name = "m").isZero)
+    assert(SQLMetrics.createTimingMetric(sparkContext, name = "m", initValue = 
-1).isZero)
+    assert(SQLMetrics.createTimingMetric(sparkContext, name = "m", initValue = 
5).isZero)
 
     assert(SQLMetrics.createNanoTimingMetric(sparkContext, name = "m").value 
=== 0)
     assert(SQLMetrics.createNanoTimingMetric(sparkContext, name = "m", 
initValue = -1).value === 0)
     assert(SQLMetrics.createNanoTimingMetric(sparkContext, name = "m", 
initValue = 5).value === 5)
 
-    assert(SQLMetrics.createNanoTimingMetric(sparkContext, name = 
"m").isZero())
-    assert(SQLMetrics.createNanoTimingMetric(sparkContext, name = "m", 
initValue = -1).isZero())
-    assert(SQLMetrics.createNanoTimingMetric(sparkContext, name = "m", 
initValue = 5).isZero())
+    assert(SQLMetrics.createNanoTimingMetric(sparkContext, name = "m").isZero)
+    assert(SQLMetrics.createNanoTimingMetric(sparkContext, name = "m", 
initValue = -1).isZero)
+    assert(SQLMetrics.createNanoTimingMetric(sparkContext, name = "m", 
initValue = 5).isZero)
   }
 }
 
diff --git 
a/sql/core/src/test/scala/org/apache/spark/sql/execution/ui/AllExecutionsPageSuite.scala
 
b/sql/core/src/test/scala/org/apache/spark/sql/execution/ui/AllExecutionsPageSuite.scala
index fe5e99d3f3f..17a5e0daaaf 100644
--- 
a/sql/core/src/test/scala/org/apache/spark/sql/execution/ui/AllExecutionsPageSuite.scala
+++ 
b/sql/core/src/test/scala/org/apache/spark/sql/execution/ui/AllExecutionsPageSuite.scala
@@ -275,7 +275,7 @@ class AllExecutionsPageWithInMemoryStoreSuite extends 
AllExecutionsPageSuite {
 @SlowSQLTest
 class AllExecutionsPageWithRocksDBBackendSuite extends AllExecutionsPageSuite {
   private val storePath = Utils.createTempDir()
-  override protected def createStatusStore(): SQLAppStatusStore = {
+  override protected def createStatusStore: SQLAppStatusStore = {
     val conf = sparkContext.conf
     conf.set(LIVE_UI_LOCAL_STORE_DIR, storePath.getCanonicalPath)
     val appStatusStore = AppStatusStore.createLiveStore(conf)
diff --git 
a/sql/hive/src/test/scala/org/apache/spark/sql/hive/execution/Hive_2_1_DDLSuite.scala
 
b/sql/hive/src/test/scala/org/apache/spark/sql/hive/execution/Hive_2_1_DDLSuite.scala
index 60a02461c93..f9dbae9b1aa 100644
--- 
a/sql/hive/src/test/scala/org/apache/spark/sql/hive/execution/Hive_2_1_DDLSuite.scala
+++ 
b/sql/hive/src/test/scala/org/apache/spark/sql/hive/execution/Hive_2_1_DDLSuite.scala
@@ -62,7 +62,7 @@ class Hive_2_1_DDLSuite extends SparkFunSuite with 
TestHiveSingleton {
     new HiveExternalCatalog(sparkConf, hadoopConf)
   }
 
-  override def afterEach: Unit = {
+  override def afterEach(): Unit = {
     catalog.listTables("default").foreach { t =>
       catalog.dropTable("default", t, true, false)
     }
diff --git 
a/sql/hive/src/test/scala/org/apache/spark/sql/hive/execution/PruneHiveTablePartitionsSuite.scala
 
b/sql/hive/src/test/scala/org/apache/spark/sql/hive/execution/PruneHiveTablePartitionsSuite.scala
index 42601be08e1..9a512871b54 100644
--- 
a/sql/hive/src/test/scala/org/apache/spark/sql/hive/execution/PruneHiveTablePartitionsSuite.scala
+++ 
b/sql/hive/src/test/scala/org/apache/spark/sql/hive/execution/PruneHiveTablePartitionsSuite.scala
@@ -34,7 +34,7 @@ import org.apache.spark.tags.SlowHiveTest
 @SlowHiveTest
 class PruneHiveTablePartitionsSuite extends PrunePartitionSuiteBase with 
TestHiveSingleton {
 
-  override def format(): String = "hive"
+  override def format: String = "hive"
 
   object Optimize extends RuleExecutor[LogicalPlan] {
     val batches =
diff --git 
a/streaming/src/test/scala/org/apache/spark/streaming/InputStreamsSuite.scala 
b/streaming/src/test/scala/org/apache/spark/streaming/InputStreamsSuite.scala
index e00651fe910..2a2243316a4 100644
--- 
a/streaming/src/test/scala/org/apache/spark/streaming/InputStreamsSuite.scala
+++ 
b/streaming/src/test/scala/org/apache/spark/streaming/InputStreamsSuite.scala
@@ -457,7 +457,7 @@ class InputStreamsSuite extends TestSuiteBase with 
BeforeAndAfter {
       }
 
       class TestReceiverInputDStream extends ReceiverInputDStream[String](ssc) 
{
-        def getReceiver: Receiver[String] = null
+        def getReceiver(): Receiver[String] = null
       }
 
       // Register input streams


---------------------------------------------------------------------
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org


Reply via email to