This is an automated email from the ASF dual-hosted git repository.

singhpk234 pushed a commit to branch main
in repository https://gitbox.apache.org/repos/asf/iceberg.git


The following commit(s) were added to refs/heads/main by this push:
     new cf27769762 Spark: Fix scala warnings in View code (#14703)
cf27769762 is described below

commit cf27769762d678a8a790481318b6346c2dd480ff
Author: Eduard Tudenhoefner <[email protected]>
AuthorDate: Fri Nov 28 02:41:34 2025 +0100

    Spark: Fix scala warnings in View code (#14703)
---
 .../scala/org/apache/spark/sql/catalyst/analysis/CheckViews.scala   | 2 +-
 .../scala/org/apache/spark/sql/catalyst/analysis/ResolveViews.scala | 4 ++--
 .../apache/spark/sql/catalyst/analysis/RewriteViewCommands.scala    | 2 +-
 .../spark/sql/execution/datasources/v2/CreateV2ViewExec.scala       | 6 +++---
 .../spark/sql/execution/datasources/v2/DescribeV2ViewExec.scala     | 2 +-
 .../spark/sql/execution/datasources/v2/ShowCreateV2ViewExec.scala   | 2 +-
 .../sql/execution/datasources/v2/ShowV2ViewPropertiesExec.scala     | 2 +-
 .../scala/org/apache/spark/sql/catalyst/analysis/CheckViews.scala   | 2 +-
 .../scala/org/apache/spark/sql/catalyst/analysis/ResolveViews.scala | 4 ++--
 .../apache/spark/sql/catalyst/analysis/RewriteViewCommands.scala    | 2 +-
 .../spark/sql/execution/datasources/v2/CreateV2ViewExec.scala       | 6 +++---
 .../spark/sql/execution/datasources/v2/DescribeV2ViewExec.scala     | 2 +-
 .../spark/sql/execution/datasources/v2/ShowCreateV2ViewExec.scala   | 2 +-
 .../sql/execution/datasources/v2/ShowV2ViewPropertiesExec.scala     | 2 +-
 .../scala/org/apache/spark/sql/catalyst/analysis/CheckViews.scala   | 2 +-
 .../scala/org/apache/spark/sql/catalyst/analysis/ResolveViews.scala | 4 ++--
 .../apache/spark/sql/catalyst/analysis/RewriteViewCommands.scala    | 2 +-
 .../spark/sql/execution/datasources/v2/CreateV2ViewExec.scala       | 6 +++---
 .../spark/sql/execution/datasources/v2/DescribeV2ViewExec.scala     | 4 ++--
 .../spark/sql/execution/datasources/v2/ShowCreateV2ViewExec.scala   | 2 +-
 .../sql/execution/datasources/v2/ShowV2ViewPropertiesExec.scala     | 2 +-
 21 files changed, 31 insertions(+), 31 deletions(-)

diff --git 
a/spark/v3.4/spark-extensions/src/main/scala/org/apache/spark/sql/catalyst/analysis/CheckViews.scala
 
b/spark/v3.4/spark-extensions/src/main/scala/org/apache/spark/sql/catalyst/analysis/CheckViews.scala
index 685b85a0d7..44b22e68d0 100644
--- 
a/spark/v3.4/spark-extensions/src/main/scala/org/apache/spark/sql/catalyst/analysis/CheckViews.scala
+++ 
b/spark/v3.4/spark-extensions/src/main/scala/org/apache/spark/sql/catalyst/analysis/CheckViews.scala
@@ -41,7 +41,7 @@ object CheckViews extends (LogicalPlan => Unit) {
       case CreateIcebergView(resolvedIdent@ResolvedIdentifier(_: ViewCatalog, 
_), _, query, columnAliases, _,
       _, _, _, _, replace, _) =>
         verifyColumnCount(resolvedIdent, columnAliases, query)
-        SchemaUtils.checkColumnNameDuplication(query.schema.fieldNames, 
SQLConf.get.resolver)
+        
SchemaUtils.checkColumnNameDuplication(query.schema.fieldNames.toIndexedSeq, 
SQLConf.get.resolver)
         if (replace) {
           val viewIdent: Seq[String] = resolvedIdent.catalog.name() +: 
resolvedIdent.identifier.asMultipartIdentifier
           checkCyclicViewReference(viewIdent, query, Seq(viewIdent))
diff --git 
a/spark/v3.4/spark-extensions/src/main/scala/org/apache/spark/sql/catalyst/analysis/ResolveViews.scala
 
b/spark/v3.4/spark-extensions/src/main/scala/org/apache/spark/sql/catalyst/analysis/ResolveViews.scala
index 37791dbbec..b77cc03326 100644
--- 
a/spark/v3.4/spark-extensions/src/main/scala/org/apache/spark/sql/catalyst/analysis/ResolveViews.scala
+++ 
b/spark/v3.4/spark-extensions/src/main/scala/org/apache/spark/sql/catalyst/analysis/ResolveViews.scala
@@ -64,7 +64,7 @@ case class ResolveViews(spark: SparkSession) extends 
Rule[LogicalPlan] with Look
     case c@CreateIcebergView(ResolvedIdentifier(_, _), _, query, 
columnAliases, columnComments, _, _, _, _, _, _)
       if query.resolved && !c.rewritten =>
       val aliased = aliasColumns(query, columnAliases, columnComments)
-      c.copy(query = aliased, queryColumnNames = query.schema.fieldNames, 
rewritten = true)
+      c.copy(query = aliased, queryColumnNames = 
query.schema.fieldNames.toIndexedSeq, rewritten = true)
   }
 
   private def aliasColumns(
@@ -99,7 +99,7 @@ case class ResolveViews(spark: SparkSession) extends 
Rule[LogicalPlan] with Look
     val aliases = view.schema.fields.zipWithIndex.map { case (expected, pos) =>
       val attr = GetColumnByOrdinal(pos, expected.dataType)
       Alias(UpCast(attr, expected.dataType), expected.name)(explicitMetadata = 
Some(expected.metadata))
-    }
+    }.toIndexedSeq
 
     SubqueryAlias(nameParts, Project(aliases, rewritten))
   }
diff --git 
a/spark/v3.4/spark-extensions/src/main/scala/org/apache/spark/sql/catalyst/analysis/RewriteViewCommands.scala
 
b/spark/v3.4/spark-extensions/src/main/scala/org/apache/spark/sql/catalyst/analysis/RewriteViewCommands.scala
index 3e854dfcb9..c7f09b2f5a 100644
--- 
a/spark/v3.4/spark-extensions/src/main/scala/org/apache/spark/sql/catalyst/analysis/RewriteViewCommands.scala
+++ 
b/spark/v3.4/spark-extensions/src/main/scala/org/apache/spark/sql/catalyst/analysis/RewriteViewCommands.scala
@@ -68,7 +68,7 @@ case class RewriteViewCommands(spark: SparkSession) extends 
Rule[LogicalPlan] wi
 
     case view @ ShowViews(UnresolvedNamespace(Seq()), pattern, output) =>
       if (ViewUtil.isViewCatalog(catalogManager.currentCatalog)) {
-        ShowIcebergViews(ResolvedNamespace(catalogManager.currentCatalog, 
catalogManager.currentNamespace),
+        ShowIcebergViews(ResolvedNamespace(catalogManager.currentCatalog, 
catalogManager.currentNamespace.toIndexedSeq),
           pattern, output)
       } else {
         view
diff --git 
a/spark/v3.4/spark-extensions/src/main/scala/org/apache/spark/sql/execution/datasources/v2/CreateV2ViewExec.scala
 
b/spark/v3.4/spark-extensions/src/main/scala/org/apache/spark/sql/execution/datasources/v2/CreateV2ViewExec.scala
index d2db0f4992..eb4cc7de48 100644
--- 
a/spark/v3.4/spark-extensions/src/main/scala/org/apache/spark/sql/execution/datasources/v2/CreateV2ViewExec.scala
+++ 
b/spark/v3.4/spark-extensions/src/main/scala/org/apache/spark/sql/execution/datasources/v2/CreateV2ViewExec.scala
@@ -27,7 +27,7 @@ import org.apache.spark.sql.catalyst.expressions.Attribute
 import org.apache.spark.sql.connector.catalog.Identifier
 import org.apache.spark.sql.connector.catalog.ViewCatalog
 import org.apache.spark.sql.types.StructType
-import scala.collection.JavaConverters._
+import scala.jdk.CollectionConverters._
 
 
 case class CreateV2ViewExec(
@@ -52,8 +52,8 @@ case class CreateV2ViewExec(
 
     val engineVersion = "Spark " + org.apache.spark.SPARK_VERSION
     val newProperties = properties ++
-      comment.map(ViewCatalog.PROP_COMMENT -> _) +
-      (ViewCatalog.PROP_CREATE_ENGINE_VERSION -> engineVersion,
+      comment.map(ViewCatalog.PROP_COMMENT -> _) ++
+      Map(ViewCatalog.PROP_CREATE_ENGINE_VERSION -> engineVersion,
         ViewCatalog.PROP_ENGINE_VERSION -> engineVersion)
 
     if (replace) {
diff --git 
a/spark/v3.4/spark-extensions/src/main/scala/org/apache/spark/sql/execution/datasources/v2/DescribeV2ViewExec.scala
 
b/spark/v3.4/spark-extensions/src/main/scala/org/apache/spark/sql/execution/datasources/v2/DescribeV2ViewExec.scala
index 2a5566a15f..a9c7312d4a 100644
--- 
a/spark/v3.4/spark-extensions/src/main/scala/org/apache/spark/sql/execution/datasources/v2/DescribeV2ViewExec.scala
+++ 
b/spark/v3.4/spark-extensions/src/main/scala/org/apache/spark/sql/execution/datasources/v2/DescribeV2ViewExec.scala
@@ -25,7 +25,7 @@ import 
org.apache.spark.sql.catalyst.util.escapeSingleQuotedString
 import org.apache.spark.sql.connector.catalog.View
 import org.apache.spark.sql.connector.catalog.ViewCatalog
 import org.apache.spark.sql.execution.LeafExecNode
-import scala.collection.JavaConverters._
+import scala.jdk.CollectionConverters._
 
 case class DescribeV2ViewExec(
   output: Seq[Attribute],
diff --git 
a/spark/v3.4/spark-extensions/src/main/scala/org/apache/spark/sql/execution/datasources/v2/ShowCreateV2ViewExec.scala
 
b/spark/v3.4/spark-extensions/src/main/scala/org/apache/spark/sql/execution/datasources/v2/ShowCreateV2ViewExec.scala
index 3be0f15031..ecd3e0d3c9 100644
--- 
a/spark/v3.4/spark-extensions/src/main/scala/org/apache/spark/sql/execution/datasources/v2/ShowCreateV2ViewExec.scala
+++ 
b/spark/v3.4/spark-extensions/src/main/scala/org/apache/spark/sql/execution/datasources/v2/ShowCreateV2ViewExec.scala
@@ -25,7 +25,7 @@ import 
org.apache.spark.sql.catalyst.util.escapeSingleQuotedString
 import org.apache.spark.sql.connector.catalog.View
 import org.apache.spark.sql.connector.catalog.ViewCatalog
 import org.apache.spark.sql.execution.LeafExecNode
-import scala.collection.JavaConverters._
+import scala.jdk.CollectionConverters._
 
 case class ShowCreateV2ViewExec(output: Seq[Attribute], view: View)
   extends V2CommandExec with LeafExecNode {
diff --git 
a/spark/v3.4/spark-extensions/src/main/scala/org/apache/spark/sql/execution/datasources/v2/ShowV2ViewPropertiesExec.scala
 
b/spark/v3.4/spark-extensions/src/main/scala/org/apache/spark/sql/execution/datasources/v2/ShowV2ViewPropertiesExec.scala
index 89fafe99ef..553b808330 100644
--- 
a/spark/v3.4/spark-extensions/src/main/scala/org/apache/spark/sql/execution/datasources/v2/ShowV2ViewPropertiesExec.scala
+++ 
b/spark/v3.4/spark-extensions/src/main/scala/org/apache/spark/sql/execution/datasources/v2/ShowV2ViewPropertiesExec.scala
@@ -24,7 +24,7 @@ import org.apache.spark.sql.catalyst.expressions.Attribute
 import org.apache.spark.sql.connector.catalog.View
 import org.apache.spark.sql.connector.catalog.ViewCatalog
 import org.apache.spark.sql.execution.LeafExecNode
-import scala.collection.JavaConverters._
+import scala.jdk.CollectionConverters._
 
 case class ShowV2ViewPropertiesExec(
   output: Seq[Attribute],
diff --git 
a/spark/v3.5/spark-extensions/src/main/scala/org/apache/spark/sql/catalyst/analysis/CheckViews.scala
 
b/spark/v3.5/spark-extensions/src/main/scala/org/apache/spark/sql/catalyst/analysis/CheckViews.scala
index b8cd102029..2a0e573b18 100644
--- 
a/spark/v3.5/spark-extensions/src/main/scala/org/apache/spark/sql/catalyst/analysis/CheckViews.scala
+++ 
b/spark/v3.5/spark-extensions/src/main/scala/org/apache/spark/sql/catalyst/analysis/CheckViews.scala
@@ -41,7 +41,7 @@ object CheckViews extends (LogicalPlan => Unit) {
       case CreateIcebergView(resolvedIdent@ResolvedIdentifier(_: ViewCatalog, 
_), _, query, columnAliases, _,
       _, _, _, _, replace, _) =>
         verifyColumnCount(resolvedIdent, columnAliases, query)
-        SchemaUtils.checkColumnNameDuplication(query.schema.fieldNames, 
SQLConf.get.resolver)
+        
SchemaUtils.checkColumnNameDuplication(query.schema.fieldNames.toIndexedSeq, 
SQLConf.get.resolver)
         if (replace) {
           val viewIdent: Seq[String] = resolvedIdent.catalog.name() +: 
resolvedIdent.identifier.asMultipartIdentifier
           checkCyclicViewReference(viewIdent, query, Seq(viewIdent))
diff --git 
a/spark/v3.5/spark-extensions/src/main/scala/org/apache/spark/sql/catalyst/analysis/ResolveViews.scala
 
b/spark/v3.5/spark-extensions/src/main/scala/org/apache/spark/sql/catalyst/analysis/ResolveViews.scala
index c80ce448e0..0fb5d30a91 100644
--- 
a/spark/v3.5/spark-extensions/src/main/scala/org/apache/spark/sql/catalyst/analysis/ResolveViews.scala
+++ 
b/spark/v3.5/spark-extensions/src/main/scala/org/apache/spark/sql/catalyst/analysis/ResolveViews.scala
@@ -64,7 +64,7 @@ case class ResolveViews(spark: SparkSession) extends 
Rule[LogicalPlan] with Look
     case c@CreateIcebergView(ResolvedIdentifier(_, _), _, query, 
columnAliases, columnComments, _, _, _, _, _, _)
       if query.resolved && !c.rewritten =>
       val aliased = aliasColumns(query, columnAliases, columnComments)
-      c.copy(query = aliased, queryColumnNames = query.schema.fieldNames, 
rewritten = true)
+      c.copy(query = aliased, queryColumnNames = 
query.schema.fieldNames.toIndexedSeq, rewritten = true)
   }
 
   private def aliasColumns(
@@ -100,7 +100,7 @@ case class ResolveViews(spark: SparkSession) extends 
Rule[LogicalPlan] with Look
     val aliases = view.schema.fields.zipWithIndex.map { case (expected, pos) =>
       val attr = GetColumnByOrdinal(pos, expected.dataType)
       Alias(UpCast(attr, expected.dataType), expected.name)(explicitMetadata = 
Some(expected.metadata))
-    }
+    }.toIndexedSeq
 
     SubqueryAlias(nameParts, Project(aliases, rewritten))
   }
diff --git 
a/spark/v3.5/spark-extensions/src/main/scala/org/apache/spark/sql/catalyst/analysis/RewriteViewCommands.scala
 
b/spark/v3.5/spark-extensions/src/main/scala/org/apache/spark/sql/catalyst/analysis/RewriteViewCommands.scala
index c5bd648f2e..800ad1f66a 100644
--- 
a/spark/v3.5/spark-extensions/src/main/scala/org/apache/spark/sql/catalyst/analysis/RewriteViewCommands.scala
+++ 
b/spark/v3.5/spark-extensions/src/main/scala/org/apache/spark/sql/catalyst/analysis/RewriteViewCommands.scala
@@ -68,7 +68,7 @@ case class RewriteViewCommands(spark: SparkSession) extends 
Rule[LogicalPlan] wi
 
     case view @ ShowViews(UnresolvedNamespace(Seq()), pattern, output) =>
       if (ViewUtil.isViewCatalog(catalogManager.currentCatalog)) {
-        ShowIcebergViews(ResolvedNamespace(catalogManager.currentCatalog, 
catalogManager.currentNamespace),
+        ShowIcebergViews(ResolvedNamespace(catalogManager.currentCatalog, 
catalogManager.currentNamespace.toIndexedSeq),
           pattern, output)
       } else {
         view
diff --git 
a/spark/v3.5/spark-extensions/src/main/scala/org/apache/spark/sql/execution/datasources/v2/CreateV2ViewExec.scala
 
b/spark/v3.5/spark-extensions/src/main/scala/org/apache/spark/sql/execution/datasources/v2/CreateV2ViewExec.scala
index d2db0f4992..eb4cc7de48 100644
--- 
a/spark/v3.5/spark-extensions/src/main/scala/org/apache/spark/sql/execution/datasources/v2/CreateV2ViewExec.scala
+++ 
b/spark/v3.5/spark-extensions/src/main/scala/org/apache/spark/sql/execution/datasources/v2/CreateV2ViewExec.scala
@@ -27,7 +27,7 @@ import org.apache.spark.sql.catalyst.expressions.Attribute
 import org.apache.spark.sql.connector.catalog.Identifier
 import org.apache.spark.sql.connector.catalog.ViewCatalog
 import org.apache.spark.sql.types.StructType
-import scala.collection.JavaConverters._
+import scala.jdk.CollectionConverters._
 
 
 case class CreateV2ViewExec(
@@ -52,8 +52,8 @@ case class CreateV2ViewExec(
 
     val engineVersion = "Spark " + org.apache.spark.SPARK_VERSION
     val newProperties = properties ++
-      comment.map(ViewCatalog.PROP_COMMENT -> _) +
-      (ViewCatalog.PROP_CREATE_ENGINE_VERSION -> engineVersion,
+      comment.map(ViewCatalog.PROP_COMMENT -> _) ++
+      Map(ViewCatalog.PROP_CREATE_ENGINE_VERSION -> engineVersion,
         ViewCatalog.PROP_ENGINE_VERSION -> engineVersion)
 
     if (replace) {
diff --git 
a/spark/v3.5/spark-extensions/src/main/scala/org/apache/spark/sql/execution/datasources/v2/DescribeV2ViewExec.scala
 
b/spark/v3.5/spark-extensions/src/main/scala/org/apache/spark/sql/execution/datasources/v2/DescribeV2ViewExec.scala
index 2a5566a15f..a9c7312d4a 100644
--- 
a/spark/v3.5/spark-extensions/src/main/scala/org/apache/spark/sql/execution/datasources/v2/DescribeV2ViewExec.scala
+++ 
b/spark/v3.5/spark-extensions/src/main/scala/org/apache/spark/sql/execution/datasources/v2/DescribeV2ViewExec.scala
@@ -25,7 +25,7 @@ import 
org.apache.spark.sql.catalyst.util.escapeSingleQuotedString
 import org.apache.spark.sql.connector.catalog.View
 import org.apache.spark.sql.connector.catalog.ViewCatalog
 import org.apache.spark.sql.execution.LeafExecNode
-import scala.collection.JavaConverters._
+import scala.jdk.CollectionConverters._
 
 case class DescribeV2ViewExec(
   output: Seq[Attribute],
diff --git 
a/spark/v3.5/spark-extensions/src/main/scala/org/apache/spark/sql/execution/datasources/v2/ShowCreateV2ViewExec.scala
 
b/spark/v3.5/spark-extensions/src/main/scala/org/apache/spark/sql/execution/datasources/v2/ShowCreateV2ViewExec.scala
index 3be0f15031..ecd3e0d3c9 100644
--- 
a/spark/v3.5/spark-extensions/src/main/scala/org/apache/spark/sql/execution/datasources/v2/ShowCreateV2ViewExec.scala
+++ 
b/spark/v3.5/spark-extensions/src/main/scala/org/apache/spark/sql/execution/datasources/v2/ShowCreateV2ViewExec.scala
@@ -25,7 +25,7 @@ import 
org.apache.spark.sql.catalyst.util.escapeSingleQuotedString
 import org.apache.spark.sql.connector.catalog.View
 import org.apache.spark.sql.connector.catalog.ViewCatalog
 import org.apache.spark.sql.execution.LeafExecNode
-import scala.collection.JavaConverters._
+import scala.jdk.CollectionConverters._
 
 case class ShowCreateV2ViewExec(output: Seq[Attribute], view: View)
   extends V2CommandExec with LeafExecNode {
diff --git 
a/spark/v3.5/spark-extensions/src/main/scala/org/apache/spark/sql/execution/datasources/v2/ShowV2ViewPropertiesExec.scala
 
b/spark/v3.5/spark-extensions/src/main/scala/org/apache/spark/sql/execution/datasources/v2/ShowV2ViewPropertiesExec.scala
index 89fafe99ef..553b808330 100644
--- 
a/spark/v3.5/spark-extensions/src/main/scala/org/apache/spark/sql/execution/datasources/v2/ShowV2ViewPropertiesExec.scala
+++ 
b/spark/v3.5/spark-extensions/src/main/scala/org/apache/spark/sql/execution/datasources/v2/ShowV2ViewPropertiesExec.scala
@@ -24,7 +24,7 @@ import org.apache.spark.sql.catalyst.expressions.Attribute
 import org.apache.spark.sql.connector.catalog.View
 import org.apache.spark.sql.connector.catalog.ViewCatalog
 import org.apache.spark.sql.execution.LeafExecNode
-import scala.collection.JavaConverters._
+import scala.jdk.CollectionConverters._
 
 case class ShowV2ViewPropertiesExec(
   output: Seq[Attribute],
diff --git 
a/spark/v4.0/spark-extensions/src/main/scala/org/apache/spark/sql/catalyst/analysis/CheckViews.scala
 
b/spark/v4.0/spark-extensions/src/main/scala/org/apache/spark/sql/catalyst/analysis/CheckViews.scala
index 109aa5360e..12c1fab36b 100644
--- 
a/spark/v4.0/spark-extensions/src/main/scala/org/apache/spark/sql/catalyst/analysis/CheckViews.scala
+++ 
b/spark/v4.0/spark-extensions/src/main/scala/org/apache/spark/sql/catalyst/analysis/CheckViews.scala
@@ -41,7 +41,7 @@ object CheckViews extends (LogicalPlan => Unit) {
       case CreateIcebergView(resolvedIdent@ResolvedIdentifier(_: ViewCatalog, 
_), _, query, columnAliases, _,
       _, _, _, _, replace, _) =>
         verifyColumnCount(resolvedIdent, columnAliases, query)
-        SchemaUtils.checkColumnNameDuplication(query.schema.fieldNames, 
SQLConf.get.resolver)
+        
SchemaUtils.checkColumnNameDuplication(query.schema.fieldNames.toIndexedSeq, 
SQLConf.get.resolver)
         if (replace) {
           val viewIdent: Seq[String] = resolvedIdent.catalog.name() +: 
resolvedIdent.identifier.asMultipartIdentifier
           checkCyclicViewReference(viewIdent, query, Seq(viewIdent))
diff --git 
a/spark/v4.0/spark-extensions/src/main/scala/org/apache/spark/sql/catalyst/analysis/ResolveViews.scala
 
b/spark/v4.0/spark-extensions/src/main/scala/org/apache/spark/sql/catalyst/analysis/ResolveViews.scala
index e69db3fce5..fe94625234 100644
--- 
a/spark/v4.0/spark-extensions/src/main/scala/org/apache/spark/sql/catalyst/analysis/ResolveViews.scala
+++ 
b/spark/v4.0/spark-extensions/src/main/scala/org/apache/spark/sql/catalyst/analysis/ResolveViews.scala
@@ -64,7 +64,7 @@ case class ResolveViews(spark: SparkSession) extends 
Rule[LogicalPlan] with Look
     case c@CreateIcebergView(ResolvedIdentifier(_, _), _, query, 
columnAliases, columnComments, _, _, _, _, _, _)
       if query.resolved && !c.rewritten =>
       val aliased = aliasColumns(query, columnAliases, columnComments)
-      c.copy(query = aliased, queryColumnNames = query.schema.fieldNames, 
rewritten = true)
+      c.copy(query = aliased, queryColumnNames = 
query.schema.fieldNames.toIndexedSeq, rewritten = true)
   }
 
   private def aliasColumns(
@@ -100,7 +100,7 @@ case class ResolveViews(spark: SparkSession) extends 
Rule[LogicalPlan] with Look
     val aliases = view.schema.fields.zipWithIndex.map { case (expected, pos) =>
       val attr = GetColumnByOrdinal(pos, expected.dataType)
       Alias(UpCast(attr, expected.dataType), expected.name)(explicitMetadata = 
Some(expected.metadata))
-    }
+    }.toIndexedSeq
 
     SubqueryAlias(nameParts, Project(aliases, rewritten))
   }
diff --git 
a/spark/v4.0/spark-extensions/src/main/scala/org/apache/spark/sql/catalyst/analysis/RewriteViewCommands.scala
 
b/spark/v4.0/spark-extensions/src/main/scala/org/apache/spark/sql/catalyst/analysis/RewriteViewCommands.scala
index db9ecfcd9e..1eeb50df8d 100644
--- 
a/spark/v4.0/spark-extensions/src/main/scala/org/apache/spark/sql/catalyst/analysis/RewriteViewCommands.scala
+++ 
b/spark/v4.0/spark-extensions/src/main/scala/org/apache/spark/sql/catalyst/analysis/RewriteViewCommands.scala
@@ -67,7 +67,7 @@ case class RewriteViewCommands(spark: SparkSession) extends 
Rule[LogicalPlan] wi
 
     case view @ ShowViews(CurrentNamespace, pattern, output) =>
       if (ViewUtil.isViewCatalog(catalogManager.currentCatalog)) {
-        ShowIcebergViews(ResolvedNamespace(catalogManager.currentCatalog, 
catalogManager.currentNamespace),
+        ShowIcebergViews(ResolvedNamespace(catalogManager.currentCatalog, 
catalogManager.currentNamespace.toIndexedSeq),
           pattern, output)
       } else {
         view
diff --git 
a/spark/v4.0/spark-extensions/src/main/scala/org/apache/spark/sql/execution/datasources/v2/CreateV2ViewExec.scala
 
b/spark/v4.0/spark-extensions/src/main/scala/org/apache/spark/sql/execution/datasources/v2/CreateV2ViewExec.scala
index 9015fb338e..97077a5ecf 100644
--- 
a/spark/v4.0/spark-extensions/src/main/scala/org/apache/spark/sql/execution/datasources/v2/CreateV2ViewExec.scala
+++ 
b/spark/v4.0/spark-extensions/src/main/scala/org/apache/spark/sql/execution/datasources/v2/CreateV2ViewExec.scala
@@ -28,7 +28,7 @@ import org.apache.spark.sql.connector.catalog.Identifier
 import org.apache.spark.sql.connector.catalog.ViewCatalog
 import org.apache.spark.sql.connector.catalog.ViewInfo
 import org.apache.spark.sql.types.StructType
-import scala.collection.JavaConverters._
+import scala.jdk.CollectionConverters._
 
 
 case class CreateV2ViewExec(
@@ -53,8 +53,8 @@ case class CreateV2ViewExec(
 
     val engineVersion = "Spark " + org.apache.spark.SPARK_VERSION
     val newProperties = properties ++
-      comment.map(ViewCatalog.PROP_COMMENT -> _) +
-      (ViewCatalog.PROP_CREATE_ENGINE_VERSION -> engineVersion,
+      comment.map(ViewCatalog.PROP_COMMENT -> _) ++
+      Map(ViewCatalog.PROP_CREATE_ENGINE_VERSION -> engineVersion,
         ViewCatalog.PROP_ENGINE_VERSION -> engineVersion)
 
     if (replace) {
diff --git 
a/spark/v4.0/spark-extensions/src/main/scala/org/apache/spark/sql/execution/datasources/v2/DescribeV2ViewExec.scala
 
b/spark/v4.0/spark-extensions/src/main/scala/org/apache/spark/sql/execution/datasources/v2/DescribeV2ViewExec.scala
index 2a5566a15f..38385d2979 100644
--- 
a/spark/v4.0/spark-extensions/src/main/scala/org/apache/spark/sql/execution/datasources/v2/DescribeV2ViewExec.scala
+++ 
b/spark/v4.0/spark-extensions/src/main/scala/org/apache/spark/sql/execution/datasources/v2/DescribeV2ViewExec.scala
@@ -25,7 +25,7 @@ import 
org.apache.spark.sql.catalyst.util.escapeSingleQuotedString
 import org.apache.spark.sql.connector.catalog.View
 import org.apache.spark.sql.connector.catalog.ViewCatalog
 import org.apache.spark.sql.execution.LeafExecNode
-import scala.collection.JavaConverters._
+import scala.jdk.CollectionConverters._
 
 case class DescribeV2ViewExec(
   output: Seq[Attribute],
@@ -55,7 +55,7 @@ case class DescribeV2ViewExec(
   private def describeExtended: Seq[InternalRow] = {
     val outputColumns = view.queryColumnNames.mkString("[", ", ", "]")
     val properties: Map[String, String] = view.properties.asScala.toMap -- 
ViewCatalog.RESERVED_PROPERTIES.asScala
-    val viewCatalogAndNamespace: Seq[String] = view.name.split("\\.").take(2)
+    val viewCatalogAndNamespace: Seq[String] = 
view.name.split("\\.").take(2).toIndexedSeq
     val viewProperties = properties.toSeq.sortBy(_._1).map {
       case (key, value) =>
         s"'${escapeSingleQuotedString(key)}' = 
'${escapeSingleQuotedString(value)}'"
diff --git 
a/spark/v4.0/spark-extensions/src/main/scala/org/apache/spark/sql/execution/datasources/v2/ShowCreateV2ViewExec.scala
 
b/spark/v4.0/spark-extensions/src/main/scala/org/apache/spark/sql/execution/datasources/v2/ShowCreateV2ViewExec.scala
index 3be0f15031..ecd3e0d3c9 100644
--- 
a/spark/v4.0/spark-extensions/src/main/scala/org/apache/spark/sql/execution/datasources/v2/ShowCreateV2ViewExec.scala
+++ 
b/spark/v4.0/spark-extensions/src/main/scala/org/apache/spark/sql/execution/datasources/v2/ShowCreateV2ViewExec.scala
@@ -25,7 +25,7 @@ import 
org.apache.spark.sql.catalyst.util.escapeSingleQuotedString
 import org.apache.spark.sql.connector.catalog.View
 import org.apache.spark.sql.connector.catalog.ViewCatalog
 import org.apache.spark.sql.execution.LeafExecNode
-import scala.collection.JavaConverters._
+import scala.jdk.CollectionConverters._
 
 case class ShowCreateV2ViewExec(output: Seq[Attribute], view: View)
   extends V2CommandExec with LeafExecNode {
diff --git 
a/spark/v4.0/spark-extensions/src/main/scala/org/apache/spark/sql/execution/datasources/v2/ShowV2ViewPropertiesExec.scala
 
b/spark/v4.0/spark-extensions/src/main/scala/org/apache/spark/sql/execution/datasources/v2/ShowV2ViewPropertiesExec.scala
index 89fafe99ef..553b808330 100644
--- 
a/spark/v4.0/spark-extensions/src/main/scala/org/apache/spark/sql/execution/datasources/v2/ShowV2ViewPropertiesExec.scala
+++ 
b/spark/v4.0/spark-extensions/src/main/scala/org/apache/spark/sql/execution/datasources/v2/ShowV2ViewPropertiesExec.scala
@@ -24,7 +24,7 @@ import org.apache.spark.sql.catalyst.expressions.Attribute
 import org.apache.spark.sql.connector.catalog.View
 import org.apache.spark.sql.connector.catalog.ViewCatalog
 import org.apache.spark.sql.execution.LeafExecNode
-import scala.collection.JavaConverters._
+import scala.jdk.CollectionConverters._
 
 case class ShowV2ViewPropertiesExec(
   output: Seq[Attribute],

Reply via email to