This is an automated email from the ASF dual-hosted git repository. yao pushed a commit to branch branch-3.4 in repository https://gitbox.apache.org/repos/asf/spark.git
The following commit(s) were added to refs/heads/branch-3.4 by this push: new b74f7922577 [SPARK-42861][SQL] Use private[sql] instead of protected[sql] to avoid generating API doc b74f7922577 is described below commit b74f79225771a95e2bc045c806637d93e5dfaaa3 Author: Wenchen Fan <wenc...@databricks.com> AuthorDate: Fri Mar 24 14:45:43 2023 +0800 [SPARK-42861][SQL] Use private[sql] instead of protected[sql] to avoid generating API doc ### What changes were proposed in this pull request? This is the only issue I found during SQL module API auditing via https://github.com/apache/spark-website/pull/443/commits/615986022c573aedaff8d2b917a0d2d9dc2b67ef . Somehow `protected[sql]` also generates API doc which is unexpected. `private[sql]` solves the problem and I generated doc locally to verify it. Another API issue has been fixed by https://github.com/apache/spark/pull/40499 ### Why are the changes needed? fix api doc ### Does this PR introduce _any_ user-facing change? no ### How was this patch tested? N/A Closes #40541 from cloud-fan/auditing. Authored-by: Wenchen Fan <wenc...@databricks.com> Signed-off-by: Kent Yao <y...@apache.org> (cherry picked from commit f7421b498a15ea687eaf811a1b2c77091945ef90) Signed-off-by: Kent Yao <y...@apache.org> --- .../src/main/scala/org/apache/spark/sql/RuntimeConfig.scala | 10 +++++----- 1 file changed, 5 insertions(+), 5 deletions(-) diff --git a/sql/core/src/main/scala/org/apache/spark/sql/RuntimeConfig.scala b/sql/core/src/main/scala/org/apache/spark/sql/RuntimeConfig.scala index 532cc7e08e1..f879a13097b 100644 --- a/sql/core/src/main/scala/org/apache/spark/sql/RuntimeConfig.scala +++ b/sql/core/src/main/scala/org/apache/spark/sql/RuntimeConfig.scala @@ -64,7 +64,7 @@ class RuntimeConfig private[sql](sqlConf: SQLConf = new SQLConf) { /** * Sets the given Spark runtime configuration property. */ - protected[sql] def set[T](entry: ConfigEntry[T], value: T): Unit = { + private[sql] def set[T](entry: ConfigEntry[T], value: T): Unit = { requireNonStaticConf(entry.key) sqlConf.setConf(entry, value) } @@ -94,18 +94,18 @@ class RuntimeConfig private[sql](sqlConf: SQLConf = new SQLConf) { * Returns the value of Spark runtime configuration property for the given key. */ @throws[NoSuchElementException]("if the key is not set") - protected[sql] def get[T](entry: ConfigEntry[T]): T = { + private[sql] def get[T](entry: ConfigEntry[T]): T = { sqlConf.getConf(entry) } - protected[sql] def get[T](entry: OptionalConfigEntry[T]): Option[T] = { + private[sql] def get[T](entry: OptionalConfigEntry[T]): Option[T] = { sqlConf.getConf(entry) } /** * Returns the value of Spark runtime configuration property for the given key. */ - protected[sql] def get[T](entry: ConfigEntry[T], default: T): T = { + private[sql] def get[T](entry: ConfigEntry[T], default: T): T = { sqlConf.getConf(entry, default) } @@ -153,7 +153,7 @@ class RuntimeConfig private[sql](sqlConf: SQLConf = new SQLConf) { /** * Returns whether a particular key is set. */ - protected[sql] def contains(key: String): Boolean = { + private[sql] def contains(key: String): Boolean = { sqlConf.contains(key) } --------------------------------------------------------------------- To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org For additional commands, e-mail: commits-h...@spark.apache.org