This is an automated email from the ASF dual-hosted git repository. ruifengz pushed a commit to branch master in repository https://gitbox.apache.org/repos/asf/spark.git
The following commit(s) were added to refs/heads/master by this push: new e1865811cac9 [MINOR] Minor English fixes e1865811cac9 is described below commit e1865811cac9c809c89b4b1512843e4b716c3e46 Author: Nicholas Chammas <nicholas.cham...@gmail.com> AuthorDate: Tue Mar 12 15:52:43 2024 +0800 [MINOR] Minor English fixes ### What changes were proposed in this pull request? Minor English grammar and wording fixes. ### Why are the changes needed? They're not strictly needed, but give the project a tiny bit more polish. ### Does this PR introduce _any_ user-facing change? Yes, some user-facing errors have been tweaked. ### How was this patch tested? No testing beyond CI. ### Was this patch authored or co-authored using generative AI tooling? No. Closes #45461 from nchammas/minor-wording-tweaks. Authored-by: Nicholas Chammas <nicholas.cham...@gmail.com> Signed-off-by: Ruifeng Zheng <ruife...@apache.org> --- common/utils/src/main/resources/error/error-classes.json | 8 ++++---- .../org/apache/spark/storage/BlockManagerMasterEndpoint.scala | 6 +++--- docs/sql-error-conditions.md | 2 +- .../scala/org/apache/spark/sql/execution/ui/UISeleniumSuite.scala | 2 +- 4 files changed, 9 insertions(+), 9 deletions(-) diff --git a/common/utils/src/main/resources/error/error-classes.json b/common/utils/src/main/resources/error/error-classes.json index 99fbc585f981..93846e66df52 100644 --- a/common/utils/src/main/resources/error/error-classes.json +++ b/common/utils/src/main/resources/error/error-classes.json @@ -40,7 +40,7 @@ "AMBIGUOUS_COLUMN_REFERENCE" : { "message" : [ "Column <name> is ambiguous. It's because you joined several DataFrame together, and some of these DataFrames are the same.", - "This column points to one of the DataFrame but Spark is unable to figure out which one.", + "This column points to one of the DataFrames but Spark is unable to figure out which one.", "Please alias the DataFrames with different names via `DataFrame.alias` before joining them,", "and specify the column using qualified name, e.g. `df.alias(\"a\").join(df.alias(\"b\"), col(\"a.id\") > col(\"b.id\"))`." ], @@ -6184,17 +6184,17 @@ }, "_LEGACY_ERROR_TEMP_2109" : { "message" : [ - "Cannot build HashedRelation with more than 1/3 billions unique keys." + "Cannot build HashedRelation with more than 1/3 billion unique keys." ] }, "_LEGACY_ERROR_TEMP_2110" : { "message" : [ - "Can not build a HashedRelation that is larger than 8G." + "Cannot build a HashedRelation that is larger than 8G." ] }, "_LEGACY_ERROR_TEMP_2111" : { "message" : [ - "failed to push a row into <rowQueue>." + "Failed to push a row into <rowQueue>." ] }, "_LEGACY_ERROR_TEMP_2112" : { diff --git a/core/src/main/scala/org/apache/spark/storage/BlockManagerMasterEndpoint.scala b/core/src/main/scala/org/apache/spark/storage/BlockManagerMasterEndpoint.scala index b4920c7cb841..5dd536eeb304 100644 --- a/core/src/main/scala/org/apache/spark/storage/BlockManagerMasterEndpoint.scala +++ b/core/src/main/scala/org/apache/spark/storage/BlockManagerMasterEndpoint.scala @@ -323,11 +323,11 @@ class BlockManagerMasterEndpoint( val isAlive = try { driverEndpoint.askSync[Boolean](CoarseGrainedClusterMessages.IsExecutorAlive(executorId)) } catch { - // ignore the non-fatal error from driverEndpoint since the caller doesn't really - // care about the return result of removing blocks. And so we could avoid breaking + // Ignore the non-fatal error from driverEndpoint since the caller doesn't really + // care about the return result of removing blocks. That way we avoid breaking // down the whole application. case NonFatal(e) => - logError(s"Fail to know the executor $executorId is alive or not.", e) + logError(s"Cannot determine whether executor $executorId is alive or not.", e) false } if (!isAlive) { diff --git a/docs/sql-error-conditions.md b/docs/sql-error-conditions.md index b6b159f277c0..165f0a1d94af 100644 --- a/docs/sql-error-conditions.md +++ b/docs/sql-error-conditions.md @@ -71,7 +71,7 @@ Column or field `<name>` is ambiguous and has `<n>` matches. [SQLSTATE: 42702](sql-error-conditions-sqlstates.html#class-42-syntax-error-or-access-rule-violation) Column `<name>` is ambiguous. It's because you joined several DataFrame together, and some of these DataFrames are the same. -This column points to one of the DataFrame but Spark is unable to figure out which one. +This column points to one of the DataFrames but Spark is unable to figure out which one. Please alias the DataFrames with different names via `DataFrame.alias` before joining them, and specify the column using qualified name, e.g. `df.alias("a").join(df.alias("b"), col("a.id") > col("b.id"))`. diff --git a/sql/core/src/test/scala/org/apache/spark/sql/execution/ui/UISeleniumSuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/execution/ui/UISeleniumSuite.scala index dc617046c430..111e233c04e3 100644 --- a/sql/core/src/test/scala/org/apache/spark/sql/execution/ui/UISeleniumSuite.scala +++ b/sql/core/src/test/scala/org/apache/spark/sql/execution/ui/UISeleniumSuite.scala @@ -102,7 +102,7 @@ class UISeleniumSuite extends SparkFunSuite with WebBrowser { test("SPARK-44801: Analyzer failure shall show the query in failed table") { spark = creatSparkSessionWithUI - intercept[Exception](spark.sql("SELECT * FROM I_AM_A_INVISIBLE_TABLE").isEmpty) + intercept[Exception](spark.sql("SELECT * FROM I_AM_AN_INVISIBLE_TABLE").isEmpty) eventually(timeout(10.seconds), interval(100.milliseconds)) { val sd = findErrorMessageOnSQLUI() assert(sd.size === 1, "Analyze fail shall show the query in failed table") --------------------------------------------------------------------- To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org For additional commands, e-mail: commits-h...@spark.apache.org