This is an automated email from the ASF dual-hosted git repository.
singhpk234 pushed a commit to branch main
in repository https://gitbox.apache.org/repos/asf/iceberg.git
The following commit(s) were added to refs/heads/main by this push:
new cf27769762 Spark: Fix scala warnings in View code (#14703)
cf27769762 is described below
commit cf27769762d678a8a790481318b6346c2dd480ff
Author: Eduard Tudenhoefner <[email protected]>
AuthorDate: Fri Nov 28 02:41:34 2025 +0100
Spark: Fix scala warnings in View code (#14703)
---
.../scala/org/apache/spark/sql/catalyst/analysis/CheckViews.scala | 2 +-
.../scala/org/apache/spark/sql/catalyst/analysis/ResolveViews.scala | 4 ++--
.../apache/spark/sql/catalyst/analysis/RewriteViewCommands.scala | 2 +-
.../spark/sql/execution/datasources/v2/CreateV2ViewExec.scala | 6 +++---
.../spark/sql/execution/datasources/v2/DescribeV2ViewExec.scala | 2 +-
.../spark/sql/execution/datasources/v2/ShowCreateV2ViewExec.scala | 2 +-
.../sql/execution/datasources/v2/ShowV2ViewPropertiesExec.scala | 2 +-
.../scala/org/apache/spark/sql/catalyst/analysis/CheckViews.scala | 2 +-
.../scala/org/apache/spark/sql/catalyst/analysis/ResolveViews.scala | 4 ++--
.../apache/spark/sql/catalyst/analysis/RewriteViewCommands.scala | 2 +-
.../spark/sql/execution/datasources/v2/CreateV2ViewExec.scala | 6 +++---
.../spark/sql/execution/datasources/v2/DescribeV2ViewExec.scala | 2 +-
.../spark/sql/execution/datasources/v2/ShowCreateV2ViewExec.scala | 2 +-
.../sql/execution/datasources/v2/ShowV2ViewPropertiesExec.scala | 2 +-
.../scala/org/apache/spark/sql/catalyst/analysis/CheckViews.scala | 2 +-
.../scala/org/apache/spark/sql/catalyst/analysis/ResolveViews.scala | 4 ++--
.../apache/spark/sql/catalyst/analysis/RewriteViewCommands.scala | 2 +-
.../spark/sql/execution/datasources/v2/CreateV2ViewExec.scala | 6 +++---
.../spark/sql/execution/datasources/v2/DescribeV2ViewExec.scala | 4 ++--
.../spark/sql/execution/datasources/v2/ShowCreateV2ViewExec.scala | 2 +-
.../sql/execution/datasources/v2/ShowV2ViewPropertiesExec.scala | 2 +-
21 files changed, 31 insertions(+), 31 deletions(-)
diff --git
a/spark/v3.4/spark-extensions/src/main/scala/org/apache/spark/sql/catalyst/analysis/CheckViews.scala
b/spark/v3.4/spark-extensions/src/main/scala/org/apache/spark/sql/catalyst/analysis/CheckViews.scala
index 685b85a0d7..44b22e68d0 100644
---
a/spark/v3.4/spark-extensions/src/main/scala/org/apache/spark/sql/catalyst/analysis/CheckViews.scala
+++
b/spark/v3.4/spark-extensions/src/main/scala/org/apache/spark/sql/catalyst/analysis/CheckViews.scala
@@ -41,7 +41,7 @@ object CheckViews extends (LogicalPlan => Unit) {
case CreateIcebergView(resolvedIdent@ResolvedIdentifier(_: ViewCatalog,
_), _, query, columnAliases, _,
_, _, _, _, replace, _) =>
verifyColumnCount(resolvedIdent, columnAliases, query)
- SchemaUtils.checkColumnNameDuplication(query.schema.fieldNames,
SQLConf.get.resolver)
+
SchemaUtils.checkColumnNameDuplication(query.schema.fieldNames.toIndexedSeq,
SQLConf.get.resolver)
if (replace) {
val viewIdent: Seq[String] = resolvedIdent.catalog.name() +:
resolvedIdent.identifier.asMultipartIdentifier
checkCyclicViewReference(viewIdent, query, Seq(viewIdent))
diff --git
a/spark/v3.4/spark-extensions/src/main/scala/org/apache/spark/sql/catalyst/analysis/ResolveViews.scala
b/spark/v3.4/spark-extensions/src/main/scala/org/apache/spark/sql/catalyst/analysis/ResolveViews.scala
index 37791dbbec..b77cc03326 100644
---
a/spark/v3.4/spark-extensions/src/main/scala/org/apache/spark/sql/catalyst/analysis/ResolveViews.scala
+++
b/spark/v3.4/spark-extensions/src/main/scala/org/apache/spark/sql/catalyst/analysis/ResolveViews.scala
@@ -64,7 +64,7 @@ case class ResolveViews(spark: SparkSession) extends
Rule[LogicalPlan] with Look
case c@CreateIcebergView(ResolvedIdentifier(_, _), _, query,
columnAliases, columnComments, _, _, _, _, _, _)
if query.resolved && !c.rewritten =>
val aliased = aliasColumns(query, columnAliases, columnComments)
- c.copy(query = aliased, queryColumnNames = query.schema.fieldNames,
rewritten = true)
+ c.copy(query = aliased, queryColumnNames =
query.schema.fieldNames.toIndexedSeq, rewritten = true)
}
private def aliasColumns(
@@ -99,7 +99,7 @@ case class ResolveViews(spark: SparkSession) extends
Rule[LogicalPlan] with Look
val aliases = view.schema.fields.zipWithIndex.map { case (expected, pos) =>
val attr = GetColumnByOrdinal(pos, expected.dataType)
Alias(UpCast(attr, expected.dataType), expected.name)(explicitMetadata =
Some(expected.metadata))
- }
+ }.toIndexedSeq
SubqueryAlias(nameParts, Project(aliases, rewritten))
}
diff --git
a/spark/v3.4/spark-extensions/src/main/scala/org/apache/spark/sql/catalyst/analysis/RewriteViewCommands.scala
b/spark/v3.4/spark-extensions/src/main/scala/org/apache/spark/sql/catalyst/analysis/RewriteViewCommands.scala
index 3e854dfcb9..c7f09b2f5a 100644
---
a/spark/v3.4/spark-extensions/src/main/scala/org/apache/spark/sql/catalyst/analysis/RewriteViewCommands.scala
+++
b/spark/v3.4/spark-extensions/src/main/scala/org/apache/spark/sql/catalyst/analysis/RewriteViewCommands.scala
@@ -68,7 +68,7 @@ case class RewriteViewCommands(spark: SparkSession) extends
Rule[LogicalPlan] wi
case view @ ShowViews(UnresolvedNamespace(Seq()), pattern, output) =>
if (ViewUtil.isViewCatalog(catalogManager.currentCatalog)) {
- ShowIcebergViews(ResolvedNamespace(catalogManager.currentCatalog,
catalogManager.currentNamespace),
+ ShowIcebergViews(ResolvedNamespace(catalogManager.currentCatalog,
catalogManager.currentNamespace.toIndexedSeq),
pattern, output)
} else {
view
diff --git
a/spark/v3.4/spark-extensions/src/main/scala/org/apache/spark/sql/execution/datasources/v2/CreateV2ViewExec.scala
b/spark/v3.4/spark-extensions/src/main/scala/org/apache/spark/sql/execution/datasources/v2/CreateV2ViewExec.scala
index d2db0f4992..eb4cc7de48 100644
---
a/spark/v3.4/spark-extensions/src/main/scala/org/apache/spark/sql/execution/datasources/v2/CreateV2ViewExec.scala
+++
b/spark/v3.4/spark-extensions/src/main/scala/org/apache/spark/sql/execution/datasources/v2/CreateV2ViewExec.scala
@@ -27,7 +27,7 @@ import org.apache.spark.sql.catalyst.expressions.Attribute
import org.apache.spark.sql.connector.catalog.Identifier
import org.apache.spark.sql.connector.catalog.ViewCatalog
import org.apache.spark.sql.types.StructType
-import scala.collection.JavaConverters._
+import scala.jdk.CollectionConverters._
case class CreateV2ViewExec(
@@ -52,8 +52,8 @@ case class CreateV2ViewExec(
val engineVersion = "Spark " + org.apache.spark.SPARK_VERSION
val newProperties = properties ++
- comment.map(ViewCatalog.PROP_COMMENT -> _) +
- (ViewCatalog.PROP_CREATE_ENGINE_VERSION -> engineVersion,
+ comment.map(ViewCatalog.PROP_COMMENT -> _) ++
+ Map(ViewCatalog.PROP_CREATE_ENGINE_VERSION -> engineVersion,
ViewCatalog.PROP_ENGINE_VERSION -> engineVersion)
if (replace) {
diff --git
a/spark/v3.4/spark-extensions/src/main/scala/org/apache/spark/sql/execution/datasources/v2/DescribeV2ViewExec.scala
b/spark/v3.4/spark-extensions/src/main/scala/org/apache/spark/sql/execution/datasources/v2/DescribeV2ViewExec.scala
index 2a5566a15f..a9c7312d4a 100644
---
a/spark/v3.4/spark-extensions/src/main/scala/org/apache/spark/sql/execution/datasources/v2/DescribeV2ViewExec.scala
+++
b/spark/v3.4/spark-extensions/src/main/scala/org/apache/spark/sql/execution/datasources/v2/DescribeV2ViewExec.scala
@@ -25,7 +25,7 @@ import
org.apache.spark.sql.catalyst.util.escapeSingleQuotedString
import org.apache.spark.sql.connector.catalog.View
import org.apache.spark.sql.connector.catalog.ViewCatalog
import org.apache.spark.sql.execution.LeafExecNode
-import scala.collection.JavaConverters._
+import scala.jdk.CollectionConverters._
case class DescribeV2ViewExec(
output: Seq[Attribute],
diff --git
a/spark/v3.4/spark-extensions/src/main/scala/org/apache/spark/sql/execution/datasources/v2/ShowCreateV2ViewExec.scala
b/spark/v3.4/spark-extensions/src/main/scala/org/apache/spark/sql/execution/datasources/v2/ShowCreateV2ViewExec.scala
index 3be0f15031..ecd3e0d3c9 100644
---
a/spark/v3.4/spark-extensions/src/main/scala/org/apache/spark/sql/execution/datasources/v2/ShowCreateV2ViewExec.scala
+++
b/spark/v3.4/spark-extensions/src/main/scala/org/apache/spark/sql/execution/datasources/v2/ShowCreateV2ViewExec.scala
@@ -25,7 +25,7 @@ import
org.apache.spark.sql.catalyst.util.escapeSingleQuotedString
import org.apache.spark.sql.connector.catalog.View
import org.apache.spark.sql.connector.catalog.ViewCatalog
import org.apache.spark.sql.execution.LeafExecNode
-import scala.collection.JavaConverters._
+import scala.jdk.CollectionConverters._
case class ShowCreateV2ViewExec(output: Seq[Attribute], view: View)
extends V2CommandExec with LeafExecNode {
diff --git
a/spark/v3.4/spark-extensions/src/main/scala/org/apache/spark/sql/execution/datasources/v2/ShowV2ViewPropertiesExec.scala
b/spark/v3.4/spark-extensions/src/main/scala/org/apache/spark/sql/execution/datasources/v2/ShowV2ViewPropertiesExec.scala
index 89fafe99ef..553b808330 100644
---
a/spark/v3.4/spark-extensions/src/main/scala/org/apache/spark/sql/execution/datasources/v2/ShowV2ViewPropertiesExec.scala
+++
b/spark/v3.4/spark-extensions/src/main/scala/org/apache/spark/sql/execution/datasources/v2/ShowV2ViewPropertiesExec.scala
@@ -24,7 +24,7 @@ import org.apache.spark.sql.catalyst.expressions.Attribute
import org.apache.spark.sql.connector.catalog.View
import org.apache.spark.sql.connector.catalog.ViewCatalog
import org.apache.spark.sql.execution.LeafExecNode
-import scala.collection.JavaConverters._
+import scala.jdk.CollectionConverters._
case class ShowV2ViewPropertiesExec(
output: Seq[Attribute],
diff --git
a/spark/v3.5/spark-extensions/src/main/scala/org/apache/spark/sql/catalyst/analysis/CheckViews.scala
b/spark/v3.5/spark-extensions/src/main/scala/org/apache/spark/sql/catalyst/analysis/CheckViews.scala
index b8cd102029..2a0e573b18 100644
---
a/spark/v3.5/spark-extensions/src/main/scala/org/apache/spark/sql/catalyst/analysis/CheckViews.scala
+++
b/spark/v3.5/spark-extensions/src/main/scala/org/apache/spark/sql/catalyst/analysis/CheckViews.scala
@@ -41,7 +41,7 @@ object CheckViews extends (LogicalPlan => Unit) {
case CreateIcebergView(resolvedIdent@ResolvedIdentifier(_: ViewCatalog,
_), _, query, columnAliases, _,
_, _, _, _, replace, _) =>
verifyColumnCount(resolvedIdent, columnAliases, query)
- SchemaUtils.checkColumnNameDuplication(query.schema.fieldNames,
SQLConf.get.resolver)
+
SchemaUtils.checkColumnNameDuplication(query.schema.fieldNames.toIndexedSeq,
SQLConf.get.resolver)
if (replace) {
val viewIdent: Seq[String] = resolvedIdent.catalog.name() +:
resolvedIdent.identifier.asMultipartIdentifier
checkCyclicViewReference(viewIdent, query, Seq(viewIdent))
diff --git
a/spark/v3.5/spark-extensions/src/main/scala/org/apache/spark/sql/catalyst/analysis/ResolveViews.scala
b/spark/v3.5/spark-extensions/src/main/scala/org/apache/spark/sql/catalyst/analysis/ResolveViews.scala
index c80ce448e0..0fb5d30a91 100644
---
a/spark/v3.5/spark-extensions/src/main/scala/org/apache/spark/sql/catalyst/analysis/ResolveViews.scala
+++
b/spark/v3.5/spark-extensions/src/main/scala/org/apache/spark/sql/catalyst/analysis/ResolveViews.scala
@@ -64,7 +64,7 @@ case class ResolveViews(spark: SparkSession) extends
Rule[LogicalPlan] with Look
case c@CreateIcebergView(ResolvedIdentifier(_, _), _, query,
columnAliases, columnComments, _, _, _, _, _, _)
if query.resolved && !c.rewritten =>
val aliased = aliasColumns(query, columnAliases, columnComments)
- c.copy(query = aliased, queryColumnNames = query.schema.fieldNames,
rewritten = true)
+ c.copy(query = aliased, queryColumnNames =
query.schema.fieldNames.toIndexedSeq, rewritten = true)
}
private def aliasColumns(
@@ -100,7 +100,7 @@ case class ResolveViews(spark: SparkSession) extends
Rule[LogicalPlan] with Look
val aliases = view.schema.fields.zipWithIndex.map { case (expected, pos) =>
val attr = GetColumnByOrdinal(pos, expected.dataType)
Alias(UpCast(attr, expected.dataType), expected.name)(explicitMetadata =
Some(expected.metadata))
- }
+ }.toIndexedSeq
SubqueryAlias(nameParts, Project(aliases, rewritten))
}
diff --git
a/spark/v3.5/spark-extensions/src/main/scala/org/apache/spark/sql/catalyst/analysis/RewriteViewCommands.scala
b/spark/v3.5/spark-extensions/src/main/scala/org/apache/spark/sql/catalyst/analysis/RewriteViewCommands.scala
index c5bd648f2e..800ad1f66a 100644
---
a/spark/v3.5/spark-extensions/src/main/scala/org/apache/spark/sql/catalyst/analysis/RewriteViewCommands.scala
+++
b/spark/v3.5/spark-extensions/src/main/scala/org/apache/spark/sql/catalyst/analysis/RewriteViewCommands.scala
@@ -68,7 +68,7 @@ case class RewriteViewCommands(spark: SparkSession) extends
Rule[LogicalPlan] wi
case view @ ShowViews(UnresolvedNamespace(Seq()), pattern, output) =>
if (ViewUtil.isViewCatalog(catalogManager.currentCatalog)) {
- ShowIcebergViews(ResolvedNamespace(catalogManager.currentCatalog,
catalogManager.currentNamespace),
+ ShowIcebergViews(ResolvedNamespace(catalogManager.currentCatalog,
catalogManager.currentNamespace.toIndexedSeq),
pattern, output)
} else {
view
diff --git
a/spark/v3.5/spark-extensions/src/main/scala/org/apache/spark/sql/execution/datasources/v2/CreateV2ViewExec.scala
b/spark/v3.5/spark-extensions/src/main/scala/org/apache/spark/sql/execution/datasources/v2/CreateV2ViewExec.scala
index d2db0f4992..eb4cc7de48 100644
---
a/spark/v3.5/spark-extensions/src/main/scala/org/apache/spark/sql/execution/datasources/v2/CreateV2ViewExec.scala
+++
b/spark/v3.5/spark-extensions/src/main/scala/org/apache/spark/sql/execution/datasources/v2/CreateV2ViewExec.scala
@@ -27,7 +27,7 @@ import org.apache.spark.sql.catalyst.expressions.Attribute
import org.apache.spark.sql.connector.catalog.Identifier
import org.apache.spark.sql.connector.catalog.ViewCatalog
import org.apache.spark.sql.types.StructType
-import scala.collection.JavaConverters._
+import scala.jdk.CollectionConverters._
case class CreateV2ViewExec(
@@ -52,8 +52,8 @@ case class CreateV2ViewExec(
val engineVersion = "Spark " + org.apache.spark.SPARK_VERSION
val newProperties = properties ++
- comment.map(ViewCatalog.PROP_COMMENT -> _) +
- (ViewCatalog.PROP_CREATE_ENGINE_VERSION -> engineVersion,
+ comment.map(ViewCatalog.PROP_COMMENT -> _) ++
+ Map(ViewCatalog.PROP_CREATE_ENGINE_VERSION -> engineVersion,
ViewCatalog.PROP_ENGINE_VERSION -> engineVersion)
if (replace) {
diff --git
a/spark/v3.5/spark-extensions/src/main/scala/org/apache/spark/sql/execution/datasources/v2/DescribeV2ViewExec.scala
b/spark/v3.5/spark-extensions/src/main/scala/org/apache/spark/sql/execution/datasources/v2/DescribeV2ViewExec.scala
index 2a5566a15f..a9c7312d4a 100644
---
a/spark/v3.5/spark-extensions/src/main/scala/org/apache/spark/sql/execution/datasources/v2/DescribeV2ViewExec.scala
+++
b/spark/v3.5/spark-extensions/src/main/scala/org/apache/spark/sql/execution/datasources/v2/DescribeV2ViewExec.scala
@@ -25,7 +25,7 @@ import
org.apache.spark.sql.catalyst.util.escapeSingleQuotedString
import org.apache.spark.sql.connector.catalog.View
import org.apache.spark.sql.connector.catalog.ViewCatalog
import org.apache.spark.sql.execution.LeafExecNode
-import scala.collection.JavaConverters._
+import scala.jdk.CollectionConverters._
case class DescribeV2ViewExec(
output: Seq[Attribute],
diff --git
a/spark/v3.5/spark-extensions/src/main/scala/org/apache/spark/sql/execution/datasources/v2/ShowCreateV2ViewExec.scala
b/spark/v3.5/spark-extensions/src/main/scala/org/apache/spark/sql/execution/datasources/v2/ShowCreateV2ViewExec.scala
index 3be0f15031..ecd3e0d3c9 100644
---
a/spark/v3.5/spark-extensions/src/main/scala/org/apache/spark/sql/execution/datasources/v2/ShowCreateV2ViewExec.scala
+++
b/spark/v3.5/spark-extensions/src/main/scala/org/apache/spark/sql/execution/datasources/v2/ShowCreateV2ViewExec.scala
@@ -25,7 +25,7 @@ import
org.apache.spark.sql.catalyst.util.escapeSingleQuotedString
import org.apache.spark.sql.connector.catalog.View
import org.apache.spark.sql.connector.catalog.ViewCatalog
import org.apache.spark.sql.execution.LeafExecNode
-import scala.collection.JavaConverters._
+import scala.jdk.CollectionConverters._
case class ShowCreateV2ViewExec(output: Seq[Attribute], view: View)
extends V2CommandExec with LeafExecNode {
diff --git
a/spark/v3.5/spark-extensions/src/main/scala/org/apache/spark/sql/execution/datasources/v2/ShowV2ViewPropertiesExec.scala
b/spark/v3.5/spark-extensions/src/main/scala/org/apache/spark/sql/execution/datasources/v2/ShowV2ViewPropertiesExec.scala
index 89fafe99ef..553b808330 100644
---
a/spark/v3.5/spark-extensions/src/main/scala/org/apache/spark/sql/execution/datasources/v2/ShowV2ViewPropertiesExec.scala
+++
b/spark/v3.5/spark-extensions/src/main/scala/org/apache/spark/sql/execution/datasources/v2/ShowV2ViewPropertiesExec.scala
@@ -24,7 +24,7 @@ import org.apache.spark.sql.catalyst.expressions.Attribute
import org.apache.spark.sql.connector.catalog.View
import org.apache.spark.sql.connector.catalog.ViewCatalog
import org.apache.spark.sql.execution.LeafExecNode
-import scala.collection.JavaConverters._
+import scala.jdk.CollectionConverters._
case class ShowV2ViewPropertiesExec(
output: Seq[Attribute],
diff --git
a/spark/v4.0/spark-extensions/src/main/scala/org/apache/spark/sql/catalyst/analysis/CheckViews.scala
b/spark/v4.0/spark-extensions/src/main/scala/org/apache/spark/sql/catalyst/analysis/CheckViews.scala
index 109aa5360e..12c1fab36b 100644
---
a/spark/v4.0/spark-extensions/src/main/scala/org/apache/spark/sql/catalyst/analysis/CheckViews.scala
+++
b/spark/v4.0/spark-extensions/src/main/scala/org/apache/spark/sql/catalyst/analysis/CheckViews.scala
@@ -41,7 +41,7 @@ object CheckViews extends (LogicalPlan => Unit) {
case CreateIcebergView(resolvedIdent@ResolvedIdentifier(_: ViewCatalog,
_), _, query, columnAliases, _,
_, _, _, _, replace, _) =>
verifyColumnCount(resolvedIdent, columnAliases, query)
- SchemaUtils.checkColumnNameDuplication(query.schema.fieldNames,
SQLConf.get.resolver)
+
SchemaUtils.checkColumnNameDuplication(query.schema.fieldNames.toIndexedSeq,
SQLConf.get.resolver)
if (replace) {
val viewIdent: Seq[String] = resolvedIdent.catalog.name() +:
resolvedIdent.identifier.asMultipartIdentifier
checkCyclicViewReference(viewIdent, query, Seq(viewIdent))
diff --git
a/spark/v4.0/spark-extensions/src/main/scala/org/apache/spark/sql/catalyst/analysis/ResolveViews.scala
b/spark/v4.0/spark-extensions/src/main/scala/org/apache/spark/sql/catalyst/analysis/ResolveViews.scala
index e69db3fce5..fe94625234 100644
---
a/spark/v4.0/spark-extensions/src/main/scala/org/apache/spark/sql/catalyst/analysis/ResolveViews.scala
+++
b/spark/v4.0/spark-extensions/src/main/scala/org/apache/spark/sql/catalyst/analysis/ResolveViews.scala
@@ -64,7 +64,7 @@ case class ResolveViews(spark: SparkSession) extends
Rule[LogicalPlan] with Look
case c@CreateIcebergView(ResolvedIdentifier(_, _), _, query,
columnAliases, columnComments, _, _, _, _, _, _)
if query.resolved && !c.rewritten =>
val aliased = aliasColumns(query, columnAliases, columnComments)
- c.copy(query = aliased, queryColumnNames = query.schema.fieldNames,
rewritten = true)
+ c.copy(query = aliased, queryColumnNames =
query.schema.fieldNames.toIndexedSeq, rewritten = true)
}
private def aliasColumns(
@@ -100,7 +100,7 @@ case class ResolveViews(spark: SparkSession) extends
Rule[LogicalPlan] with Look
val aliases = view.schema.fields.zipWithIndex.map { case (expected, pos) =>
val attr = GetColumnByOrdinal(pos, expected.dataType)
Alias(UpCast(attr, expected.dataType), expected.name)(explicitMetadata =
Some(expected.metadata))
- }
+ }.toIndexedSeq
SubqueryAlias(nameParts, Project(aliases, rewritten))
}
diff --git
a/spark/v4.0/spark-extensions/src/main/scala/org/apache/spark/sql/catalyst/analysis/RewriteViewCommands.scala
b/spark/v4.0/spark-extensions/src/main/scala/org/apache/spark/sql/catalyst/analysis/RewriteViewCommands.scala
index db9ecfcd9e..1eeb50df8d 100644
---
a/spark/v4.0/spark-extensions/src/main/scala/org/apache/spark/sql/catalyst/analysis/RewriteViewCommands.scala
+++
b/spark/v4.0/spark-extensions/src/main/scala/org/apache/spark/sql/catalyst/analysis/RewriteViewCommands.scala
@@ -67,7 +67,7 @@ case class RewriteViewCommands(spark: SparkSession) extends
Rule[LogicalPlan] wi
case view @ ShowViews(CurrentNamespace, pattern, output) =>
if (ViewUtil.isViewCatalog(catalogManager.currentCatalog)) {
- ShowIcebergViews(ResolvedNamespace(catalogManager.currentCatalog,
catalogManager.currentNamespace),
+ ShowIcebergViews(ResolvedNamespace(catalogManager.currentCatalog,
catalogManager.currentNamespace.toIndexedSeq),
pattern, output)
} else {
view
diff --git
a/spark/v4.0/spark-extensions/src/main/scala/org/apache/spark/sql/execution/datasources/v2/CreateV2ViewExec.scala
b/spark/v4.0/spark-extensions/src/main/scala/org/apache/spark/sql/execution/datasources/v2/CreateV2ViewExec.scala
index 9015fb338e..97077a5ecf 100644
---
a/spark/v4.0/spark-extensions/src/main/scala/org/apache/spark/sql/execution/datasources/v2/CreateV2ViewExec.scala
+++
b/spark/v4.0/spark-extensions/src/main/scala/org/apache/spark/sql/execution/datasources/v2/CreateV2ViewExec.scala
@@ -28,7 +28,7 @@ import org.apache.spark.sql.connector.catalog.Identifier
import org.apache.spark.sql.connector.catalog.ViewCatalog
import org.apache.spark.sql.connector.catalog.ViewInfo
import org.apache.spark.sql.types.StructType
-import scala.collection.JavaConverters._
+import scala.jdk.CollectionConverters._
case class CreateV2ViewExec(
@@ -53,8 +53,8 @@ case class CreateV2ViewExec(
val engineVersion = "Spark " + org.apache.spark.SPARK_VERSION
val newProperties = properties ++
- comment.map(ViewCatalog.PROP_COMMENT -> _) +
- (ViewCatalog.PROP_CREATE_ENGINE_VERSION -> engineVersion,
+ comment.map(ViewCatalog.PROP_COMMENT -> _) ++
+ Map(ViewCatalog.PROP_CREATE_ENGINE_VERSION -> engineVersion,
ViewCatalog.PROP_ENGINE_VERSION -> engineVersion)
if (replace) {
diff --git
a/spark/v4.0/spark-extensions/src/main/scala/org/apache/spark/sql/execution/datasources/v2/DescribeV2ViewExec.scala
b/spark/v4.0/spark-extensions/src/main/scala/org/apache/spark/sql/execution/datasources/v2/DescribeV2ViewExec.scala
index 2a5566a15f..38385d2979 100644
---
a/spark/v4.0/spark-extensions/src/main/scala/org/apache/spark/sql/execution/datasources/v2/DescribeV2ViewExec.scala
+++
b/spark/v4.0/spark-extensions/src/main/scala/org/apache/spark/sql/execution/datasources/v2/DescribeV2ViewExec.scala
@@ -25,7 +25,7 @@ import
org.apache.spark.sql.catalyst.util.escapeSingleQuotedString
import org.apache.spark.sql.connector.catalog.View
import org.apache.spark.sql.connector.catalog.ViewCatalog
import org.apache.spark.sql.execution.LeafExecNode
-import scala.collection.JavaConverters._
+import scala.jdk.CollectionConverters._
case class DescribeV2ViewExec(
output: Seq[Attribute],
@@ -55,7 +55,7 @@ case class DescribeV2ViewExec(
private def describeExtended: Seq[InternalRow] = {
val outputColumns = view.queryColumnNames.mkString("[", ", ", "]")
val properties: Map[String, String] = view.properties.asScala.toMap --
ViewCatalog.RESERVED_PROPERTIES.asScala
- val viewCatalogAndNamespace: Seq[String] = view.name.split("\\.").take(2)
+ val viewCatalogAndNamespace: Seq[String] =
view.name.split("\\.").take(2).toIndexedSeq
val viewProperties = properties.toSeq.sortBy(_._1).map {
case (key, value) =>
s"'${escapeSingleQuotedString(key)}' =
'${escapeSingleQuotedString(value)}'"
diff --git
a/spark/v4.0/spark-extensions/src/main/scala/org/apache/spark/sql/execution/datasources/v2/ShowCreateV2ViewExec.scala
b/spark/v4.0/spark-extensions/src/main/scala/org/apache/spark/sql/execution/datasources/v2/ShowCreateV2ViewExec.scala
index 3be0f15031..ecd3e0d3c9 100644
---
a/spark/v4.0/spark-extensions/src/main/scala/org/apache/spark/sql/execution/datasources/v2/ShowCreateV2ViewExec.scala
+++
b/spark/v4.0/spark-extensions/src/main/scala/org/apache/spark/sql/execution/datasources/v2/ShowCreateV2ViewExec.scala
@@ -25,7 +25,7 @@ import
org.apache.spark.sql.catalyst.util.escapeSingleQuotedString
import org.apache.spark.sql.connector.catalog.View
import org.apache.spark.sql.connector.catalog.ViewCatalog
import org.apache.spark.sql.execution.LeafExecNode
-import scala.collection.JavaConverters._
+import scala.jdk.CollectionConverters._
case class ShowCreateV2ViewExec(output: Seq[Attribute], view: View)
extends V2CommandExec with LeafExecNode {
diff --git
a/spark/v4.0/spark-extensions/src/main/scala/org/apache/spark/sql/execution/datasources/v2/ShowV2ViewPropertiesExec.scala
b/spark/v4.0/spark-extensions/src/main/scala/org/apache/spark/sql/execution/datasources/v2/ShowV2ViewPropertiesExec.scala
index 89fafe99ef..553b808330 100644
---
a/spark/v4.0/spark-extensions/src/main/scala/org/apache/spark/sql/execution/datasources/v2/ShowV2ViewPropertiesExec.scala
+++
b/spark/v4.0/spark-extensions/src/main/scala/org/apache/spark/sql/execution/datasources/v2/ShowV2ViewPropertiesExec.scala
@@ -24,7 +24,7 @@ import org.apache.spark.sql.catalyst.expressions.Attribute
import org.apache.spark.sql.connector.catalog.View
import org.apache.spark.sql.connector.catalog.ViewCatalog
import org.apache.spark.sql.execution.LeafExecNode
-import scala.collection.JavaConverters._
+import scala.jdk.CollectionConverters._
case class ShowV2ViewPropertiesExec(
output: Seq[Attribute],