This is an automated email from the ASF dual-hosted git repository. gurwls223 pushed a commit to branch master in repository https://gitbox.apache.org/repos/asf/spark.git
The following commit(s) were added to refs/heads/master by this push: new 1fbb1d46feb [SPARK-39371][DOCS][CORE] Review and fix issues in Scala/Java API docs of Core module 1fbb1d46feb is described below commit 1fbb1d46feb992c3441f2a4f2c5d5179da465d4b Author: Yuanjian Li <yuanjian...@databricks.com> AuthorDate: Fri Jun 3 17:49:01 2022 +0900 [SPARK-39371][DOCS][CORE] Review and fix issues in Scala/Java API docs of Core module ### What changes were proposed in this pull request? Compare the 3.3.0 API doc with the latest release version 3.2.1. Fix the following issues: * Add missing Since annotation for new APIs * Remove the leaking class/object in API doc ### Why are the changes needed? Improve API docs ### Does this PR introduce _any_ user-facing change? No ### How was this patch tested? Existing UT Closes #36757 from xuanyuanking/doc. Authored-by: Yuanjian Li <yuanjian...@databricks.com> Signed-off-by: Hyukjin Kwon <gurwls...@apache.org> --- core/src/main/scala/org/apache/spark/errors/SparkCoreErrors.scala | 2 +- .../storage/BlockSavedOnDecommissionedBlockManagerException.scala | 2 +- .../src/main/java/org/apache/spark/launcher/AbstractLauncher.java | 2 +- .../src/main/java/org/apache/spark/launcher/InProcessLauncher.java | 2 +- .../src/main/java/org/apache/spark/launcher/JavaModuleOptions.java | 2 ++ .../scala/org/apache/spark/sql/diagnostic/DiagnosticListener.scala | 4 ++-- .../scala/org/apache/spark/sql/diagnostic/DiagnosticStore.scala | 6 +++--- 7 files changed, 11 insertions(+), 9 deletions(-) diff --git a/core/src/main/scala/org/apache/spark/errors/SparkCoreErrors.scala b/core/src/main/scala/org/apache/spark/errors/SparkCoreErrors.scala index aecef8ed2d6..1da02884462 100644 --- a/core/src/main/scala/org/apache/spark/errors/SparkCoreErrors.scala +++ b/core/src/main/scala/org/apache/spark/errors/SparkCoreErrors.scala @@ -30,7 +30,7 @@ import org.apache.spark.storage.{BlockId, BlockManagerId, BlockNotFoundException /** * Object for grouping error messages from (most) exceptions thrown during query execution. */ -object SparkCoreErrors { +private[spark] object SparkCoreErrors { def unexpectedPy4JServerError(other: Object): Throwable = { new RuntimeException(s"Unexpected Py4J server ${other.getClass}") } diff --git a/core/src/main/scala/org/apache/spark/storage/BlockSavedOnDecommissionedBlockManagerException.scala b/core/src/main/scala/org/apache/spark/storage/BlockSavedOnDecommissionedBlockManagerException.scala index 4684d9c6775..21a022864bb 100644 --- a/core/src/main/scala/org/apache/spark/storage/BlockSavedOnDecommissionedBlockManagerException.scala +++ b/core/src/main/scala/org/apache/spark/storage/BlockSavedOnDecommissionedBlockManagerException.scala @@ -17,5 +17,5 @@ package org.apache.spark.storage -class BlockSavedOnDecommissionedBlockManagerException(blockId: BlockId) +private[spark] class BlockSavedOnDecommissionedBlockManagerException(blockId: BlockId) extends Exception(s"Block $blockId cannot be saved on decommissioned executor") diff --git a/launcher/src/main/java/org/apache/spark/launcher/AbstractLauncher.java b/launcher/src/main/java/org/apache/spark/launcher/AbstractLauncher.java index eee15419209..a944950cf15 100644 --- a/launcher/src/main/java/org/apache/spark/launcher/AbstractLauncher.java +++ b/launcher/src/main/java/org/apache/spark/launcher/AbstractLauncher.java @@ -26,7 +26,7 @@ import static org.apache.spark.launcher.CommandBuilderUtils.*; /** * Base class for launcher implementations. * - * @since Spark 2.3.0 + * @since 2.3.0 */ public abstract class AbstractLauncher<T extends AbstractLauncher<T>> { diff --git a/launcher/src/main/java/org/apache/spark/launcher/InProcessLauncher.java b/launcher/src/main/java/org/apache/spark/launcher/InProcessLauncher.java index 688e1f763c2..6867518b321 100644 --- a/launcher/src/main/java/org/apache/spark/launcher/InProcessLauncher.java +++ b/launcher/src/main/java/org/apache/spark/launcher/InProcessLauncher.java @@ -37,7 +37,7 @@ import java.util.logging.Logger; * driver memory or configs which modify the driver's class path) do not take effect. Logging * configuration is also inherited from the parent application. * - * @since Spark 2.3.0 + * @since 2.3.0 */ public class InProcessLauncher extends AbstractLauncher<InProcessLauncher> { diff --git a/launcher/src/main/java/org/apache/spark/launcher/JavaModuleOptions.java b/launcher/src/main/java/org/apache/spark/launcher/JavaModuleOptions.java index c7d3df99c6e..978466cd77c 100644 --- a/launcher/src/main/java/org/apache/spark/launcher/JavaModuleOptions.java +++ b/launcher/src/main/java/org/apache/spark/launcher/JavaModuleOptions.java @@ -21,6 +21,8 @@ package org.apache.spark.launcher; * This helper class is used to place the all `--add-opens` options * required by Spark when using Java 17. `DEFAULT_MODULE_OPTIONS` has added * `-XX:+IgnoreUnrecognizedVMOptions` to be compatible with Java 8 and Java 11. + * + * @since 3.3.0 */ public class JavaModuleOptions { private static final String[] DEFAULT_MODULE_OPTIONS = { diff --git a/sql/core/src/main/scala/org/apache/spark/sql/diagnostic/DiagnosticListener.scala b/sql/core/src/main/scala/org/apache/spark/sql/diagnostic/DiagnosticListener.scala index 7ce1093e879..6899790603b 100644 --- a/sql/core/src/main/scala/org/apache/spark/sql/diagnostic/DiagnosticListener.scala +++ b/sql/core/src/main/scala/org/apache/spark/sql/diagnostic/DiagnosticListener.scala @@ -30,7 +30,7 @@ import org.apache.spark.status.{ElementTrackingStore, KVUtils} * * @param kvStore used to store the diagnostic information */ -class DiagnosticListener( +private[spark] class DiagnosticListener( conf: SparkConf, kvStore: ElementTrackingStore) extends SparkListener { @@ -107,6 +107,6 @@ class DiagnosticListener( } } -object DiagnosticListener { +private[spark] object DiagnosticListener { val QUEUE_NAME = "diagnostics" } diff --git a/sql/core/src/main/scala/org/apache/spark/sql/diagnostic/DiagnosticStore.scala b/sql/core/src/main/scala/org/apache/spark/sql/diagnostic/DiagnosticStore.scala index c13cc8a7f39..53ff787fe67 100644 --- a/sql/core/src/main/scala/org/apache/spark/sql/diagnostic/DiagnosticStore.scala +++ b/sql/core/src/main/scala/org/apache/spark/sql/diagnostic/DiagnosticStore.scala @@ -28,7 +28,7 @@ import org.apache.spark.util.kvstore.{KVIndex, KVStore} * information. There's no state kept in this class, so it's ok to have multiple instances * of it in an application. */ -class DiagnosticStore(store: KVStore) { +private[spark] class DiagnosticStore(store: KVStore) { def diagnosticsList(offset: Int, length: Int): Seq[ExecutionDiagnosticData] = { KVUtils.viewToSeq(store.view(classOf[ExecutionDiagnosticData]).skip(offset).max(length)) @@ -51,7 +51,7 @@ class DiagnosticStore(store: KVStore) { } /* Represents the diagnostic data of a SQL execution */ -class ExecutionDiagnosticData( +private[spark] class ExecutionDiagnosticData( @KVIndexParam val executionId: Long, val physicalPlan: String, val submissionTime: Long, @@ -59,7 +59,7 @@ class ExecutionDiagnosticData( val errorMessage: Option[String]) /* Represents the plan change of an adaptive execution */ -class AdaptiveExecutionUpdate( +private[spark] class AdaptiveExecutionUpdate( @KVIndexParam("id") val executionId: Long, @KVIndexParam(value = "updateTime", parent = "id") --------------------------------------------------------------------- To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org For additional commands, e-mail: commits-h...@spark.apache.org