[ https://issues.apache.org/jira/browse/SPARK-38587?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17511152#comment-17511152 ]
Kent Yao commented on SPARK-38587: ---------------------------------- resolved by https://github.com/apache/spark/pull/35895 > Validating new location for rename command should use formatted names > --------------------------------------------------------------------- > > Key: SPARK-38587 > URL: https://issues.apache.org/jira/browse/SPARK-38587 > Project: Spark > Issue Type: Bug > Components: SQL > Affects Versions: 3.0.3, 3.1.2, 3.2.1, 3.3.0 > Reporter: Kent Yao > Assignee: Kent Yao > Priority: Major > Fix For: 3.1.3, 3.0.4, 3.3.0, 3.2.2 > > > ``` > {code:java} > [info] - ALTER TABLE .. RENAME using V1 catalog V1 command: newName *** > FAILED *** (61 milliseconds) > [info] org.apache.spark.sql.catalyst.analysis.NoSuchDatabaseException: > Database 'CaseUpperCaseLower' not found > [info] at > org.apache.spark.sql.catalyst.catalog.ExternalCatalog.requireDbExists(ExternalCatalog.scala:42) > [info] at > org.apache.spark.sql.catalyst.catalog.ExternalCatalog.requireDbExists$(ExternalCatalog.scala:40) > [info] at > org.apache.spark.sql.catalyst.catalog.InMemoryCatalog.requireDbExists(InMemoryCatalog.scala:47) > [info] at > org.apache.spark.sql.catalyst.catalog.InMemoryCatalog.getDatabase(InMemoryCatalog.scala:171) > [info] at > org.apache.spark.sql.catalyst.catalog.ExternalCatalogWithListener.getDatabase(ExternalCatalogWithListener.scala:65) > [info] at > org.apache.spark.sql.catalyst.catalog.SessionCatalog.validateNewLocationOfRename(SessionCatalog.scala:1863) > [info] at > org.apache.spark.sql.catalyst.catalog.SessionCatalog.renameTable(SessionCatalog.scala:739) > [info] at > org.apache.spark.sql.execution.command.AlterTableRenameCommand.run(tables.scala:209 > [info] at > org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult$lzycompute(commands.scala:75) > [info] at > org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult(commands.scala:73) > [info] at > org.apache.spark.sql.execution.command.ExecutedCommandExec.executeCollect(commands.scala:84) > [info] at > org.apache.spark.sql.execution.QueryExecution$$anonfun$eagerlyExecuteCommands$1.$anonfun$applyOrElse$1(QueryExecution.scala:110) > [info] at > org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$6(SQLExecution.scala:109) > [info] at > org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:169) > [info] at > org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$1(SQLExecution.scala:95) > [info] at > org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:779) > [info] at > org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:64) > [info] at > org.apache.spark.sql.execution.QueryExecution$$anonfun$eagerlyExecuteCommands$1.applyOrElse(QueryExecution.scala:110) > [info] at > org.apache.spark.sql.execution.QueryExecution$$anonfun$eagerlyExecuteCommands$1.applyOrElse(QueryExecution.scala:106) > [info] at > org.apache.spark.sql.catalyst.trees.TreeNode.$anonfun$transformDownWithPruning$1(TreeNode.scala:491) > [info] at > org.apache.spark.sql.catalyst.trees.CurrentOrigin$.withOrigin(TreeNode.scala:83) > [info] at > org.apache.spark.sql.catalyst.trees.TreeNode.transformDownWithPruning(TreeNode.scala:491) > [info] at > org.apache.spark.sql.catalyst.plans.logical.LogicalPlan.org$apache$spark$sql$catalyst$plans$logical$AnalysisHelper$$super$transformDownWithPruning(LogicalPlan.scala:30) > [info] at > org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper.transformDownWithPruning(AnalysisHelper.scala:267) > [info] at > org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper.transformDownWithPruning$(AnalysisHelper.scala:263) > [info] at > org.apache.spark.sql.catalyst.plans.logical.LogicalPlan.transformDownWithPruning(LogicalPlan.scala:30) > [info] at > org.apache.spark.sql.catalyst.plans.logical.LogicalPlan.transformDownWithPruning(LogicalPlan.scala:30) > [info] at > org.apache.spark.sql.catalyst.trees.TreeNode.transformDown(TreeNode.scala:467) > [info] at > org.apache.spark.sql.execution.QueryExecution.eagerlyExecuteCommands(QueryExecution.scala:106) > [info] at > org.apache.spark.sql.execution.QueryExecution.commandExecuted$lzycompute(QueryExecution.scala:93) > [info] at > org.apache.spark.sql.execution.QueryExecution.commandExecuted(QueryExecution.scala:91) > [info] at org.apache.spark.sql.Dataset.<init>(Dataset.scala:220) > [info] at org.apache.spark.sql.Dataset$.$anonfun$ofRows$2(Dataset.scala:100) > [info] at > org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:779) > [info] at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:97) > [info] at > org.apache.spark.sql.SparkSession.$anonfun$sql$1(SparkSession.scala:622) > [info] at > org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:779) > [info] at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:617) > [info] at > org.apache.spark.sql.test.SQLTestUtilsBase.$anonfun$sql$1(SQLTestUtils.scala:232) > [info] at > org.apache.spark.sql.execution.command.AlterTableRenameSuiteBase.$anonfun$$init$$19(AlterTableRenameSuiteBase.scala:143) > [info] at > org.apache.spark.sql.execution.command.AlterTableRenameSuiteBase.$anonfun$$init$$19$adapted(AlterTableRenameSuiteBase.scala:141) > [info] at > org.apache.spark.sql.execution.command.DDLCommandTestUtils.$anonfun$withNamespaceAndTable$2(DDLCommandTestUtils.scala:67) > [info] at > scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23) > [info] at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:1478) > [info] at > org.apache.spark.sql.test.SQLTestUtilsBase.withTable(SQLTestUtils.scala:306) > [info] at > org.apache.spark.sql.test.SQLTestUtilsBase.withTable$(SQLTestUtils.scala:304) > [info] at > org.apache.spark.sql.execution.command.v1.AlterTableRenameSuite.withTable(AlterTableRenameSuite.scala:83) > [info] at > org.apache.spark.sql.execution.command.DDLCommandTestUtils.$anonfun$withNamespaceAndTable$1(DDLCommandTestUtils.scala:67) > [info] at > scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23) > [info] at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:1478) > [info] at > org.apache.spark.sql.test.SQLTestUtilsBase.withNamespace(SQLTestUtils.scala:389) > [info] at > org.apache.spark.sql.test.SQLTestUtilsBase.withNamespace$(SQLTestUtils.scala:387) > [info] at > org.apache.spark.sql.execution.command.v1.AlterTableRenameSuite.withNamespace(AlterTableRenameSuite.scala:83) > [info] at > org.apache.spark.sql.execution.command.DDLCommandTestUtils.withNamespaceAndTable(DDLCommandTestUtils.scala:63) > [info] at > org.apache.spark.sql.execution.command.DDLCommandTestUtils.withNamespaceAndTable$(DDLCommandTestUtils.scala:60) > [info] at > org.apache.spark.sql.execution.command.v1.AlterTableRenameSuite.withNamespaceAndTable(AlterTableRenameSuite.scala:83) > [info] at > org.apache.spark.sql.execution.command.AlterTableRenameSuiteBase.$anonfun$$init$$18(AlterTableRenameSuiteBase.scala:141) > [info] at > scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23) > [info] at org.scalatest.OutcomeOf.outcomeOf(OutcomeOf.scala:85) > [info] at org.scalatest.OutcomeOf.outcomeOf$(OutcomeOf.scala:83) > [info] at org.scalatest.OutcomeOf$.outcomeOf(OutcomeOf.scala:104) > [info] at org.scalatest.Transformer.apply(Transformer.scala:22) > [info] at org.scalatest.Transformer.apply(Transformer.scala:20) > [info] at > org.scalatest.funsuite.AnyFunSuiteLike$$anon$1.apply(AnyFunSuiteLike.scala:190) > [info] at > org.apache.spark.SparkFunSuite.withFixture(SparkFunSuite.scala:203) > [info] at > org.scalatest.funsuite.AnyFunSuiteLike.invokeWithFixture$1(AnyFunSuiteLike.scala:188) > [info] at > org.scalatest.funsuite.AnyFunSuiteLike.$anonfun$runTest$1(AnyFunSuiteLike.scala:200 > [info] at org.scalatest.SuperEngine.runTestImpl(Engine.scala:306) > [info] at > org.scalatest.funsuite.AnyFunSuiteLike.runTest(AnyFunSuiteLike.scala:200) > [info] at > org.scalatest.funsuite.AnyFunSuiteLike.runTest$(AnyFunSuiteLike.scala:182) > [info] at > org.apache.spark.SparkFunSuite.org$scalatest$BeforeAndAfterEach$$super$runTest(SparkFunSuite.scala:64) > [info] at > org.scalatest.BeforeAndAfterEach.runTest(BeforeAndAfterEach.scala:234) > [info] at > org.scalatest.BeforeAndAfterEach.runTest$(BeforeAndAfterEach.scala:227) > [info] at org.apache.spark.SparkFunSuite.runTest(SparkFunSuite.scala:64) > [info] at > org.scalatest.funsuite.AnyFunSuiteLike.$anonfun$runTests$1(AnyFunSuiteLike.scala:233) > [info] at > org.scalatest.SuperEngine.$anonfun$runTestsInBranch$1(Engine.scala:413) > [info] at scala.collection.immutable.List.foreach(List.scala:431) > [info] at org.scalatest.SuperEngine.traverseSubNodes$1(Engine.scala:401) > [info] at org.scalatest.SuperEngine.runTestsInBranch(Engine.scala:396) > [info] at org.scalatest.SuperEngine.runTestsImpl(Engine.scala:475) > [info] at > org.scalatest.funsuite.AnyFunSuiteLike.runTests(AnyFunSuiteLike.scala:233) > [info] at > org.scalatest.funsuite.AnyFunSuiteLike.runTests$(AnyFunSuiteLike.scala:232) > [info] at > org.scalatest.funsuite.AnyFunSuite.runTests(AnyFunSuite.scala:1563) > [info] at org.scalatest.Suite.run(Suite.scala:1112) > [info] at org.scalatest.Suite.run$(Suite.scala:1094) > [info] at > org.scalatest.funsuite.AnyFunSuite.org$scalatest$funsuite$AnyFunSuiteLike$$super$run(AnyFunSuite.scala:1563) > [info] at > org.scalatest.funsuite.AnyFunSuiteLike.$anonfun$run$1(AnyFunSuiteLike.scala:237) > [info] at org.scalatest.SuperEngine.runImpl(Engine.scala:535) > [info] at > org.scalatest.funsuite.AnyFunSuiteLike.run(AnyFunSuiteLike.scala:237) > [info] at > org.scalatest.funsuite.AnyFunSuiteLike.run$(AnyFunSuiteLike.scala:236) > [info] at > org.apache.spark.SparkFunSuite.org$scalatest$BeforeAndAfterAll$$super$run(SparkFunSuite.scala:64) > [info] at > org.scalatest.BeforeAndAfterAll.liftedTree1$1(BeforeAndAfterAll.scala:213) > [info] at org.scalatest.BeforeAndAfterAll.run(BeforeAndAfterAll.scala:210) > [info] at org.scalatest.BeforeAndAfterAll.run$(BeforeAndAfterAll.scala:208) > [info] at org.apache.spark.SparkFunSuite.run(SparkFunSuite.scala:64) > [info] at > org.scalatest.tools.Framework.org$scalatest$tools$Framework$$runSuite(Framework.scala:318) > [info] at > org.scalatest.tools.Framework$ScalaTestTask.execute(Framework.scala:513) > [info] at sbt.ForkMain$Run.lambda$runTest$1(ForkMain.java:413) > [info] at java.util.concurrent.FutureTask.run(FutureTask.java:266) > [info] at > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) > [info] at > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) > [info] at java.lang.Thread.run(Thread.java:748) > {code} -- This message was sent by Atlassian Jira (v8.20.1#820001) --------------------------------------------------------------------- To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org