This is an automated email from the ASF dual-hosted git repository.

chengchengjin pushed a commit to branch main
in repository https://gitbox.apache.org/repos/asf/incubator-gluten.git


The following commit(s) were added to refs/heads/main by this push:
     new f1fc5697d0 [CORE] Shim: Remove unused Spark33Scan / Spark34Scan / 
Spark35Scan (#10051)
f1fc5697d0 is described below

commit f1fc5697d086a5b4a518edba318f44b7c7bf0860
Author: Hongze Zhang <[email protected]>
AuthorDate: Fri Jun 27 12:04:00 2025 +0800

    [CORE] Shim: Remove unused Spark33Scan / Spark34Scan / Spark35Scan (#10051)
---
 .../sql/execution/datasources/v2/Spark33Scan.scala | 47 --------------------
 .../sql/execution/datasources/v2/Spark34Scan.scala | 50 ----------------------
 .../sql/execution/datasources/v2/Spark35Scan.scala | 50 ----------------------
 3 files changed, 147 deletions(-)

diff --git 
a/shims/spark33/src/main/scala/org/apache/spark/sql/execution/datasources/v2/Spark33Scan.scala
 
b/shims/spark33/src/main/scala/org/apache/spark/sql/execution/datasources/v2/Spark33Scan.scala
deleted file mode 100644
index 169d5c3cb8..0000000000
--- 
a/shims/spark33/src/main/scala/org/apache/spark/sql/execution/datasources/v2/Spark33Scan.scala
+++ /dev/null
@@ -1,47 +0,0 @@
-/*
- * Licensed to the Apache Software Foundation (ASF) under one or more
- * contributor license agreements.  See the NOTICE file distributed with
- * this work for additional information regarding copyright ownership.
- * The ASF licenses this file to You under the Apache License, Version 2.0
- * (the "License"); you may not use this file except in compliance with
- * the License.  You may obtain a copy of the License at
- *
- *    http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-package org.apache.spark.sql.execution.datasources.v2
-
-import org.apache.spark.rdd.RDD
-import org.apache.spark.sql.catalyst.InternalRow
-import org.apache.spark.sql.catalyst.expressions.{Attribute, Expression}
-import org.apache.spark.sql.connector.read.{InputPartition, 
PartitionReaderFactory, Scan}
-
-class Spark33Scan extends DataSourceV2ScanExecBase {
-
-  override def scan: Scan = throw new 
UnsupportedOperationException("Spark33Scan")
-
-  override def readerFactory: PartitionReaderFactory =
-    throw new UnsupportedOperationException("Spark33Scan")
-
-  override def keyGroupedPartitioning: Option[Seq[Expression]] =
-    throw new UnsupportedOperationException("Spark33Scan")
-
-  override protected def inputPartitions: Seq[InputPartition] =
-    throw new UnsupportedOperationException("Spark33Scan")
-
-  override def inputRDD: RDD[InternalRow] = throw new 
UnsupportedOperationException("Spark33Scan")
-
-  override def output: Seq[Attribute] = throw new 
UnsupportedOperationException("Spark33Scan")
-
-  override def productElement(n: Int): Any = throw new 
UnsupportedOperationException("Spark33Scan")
-
-  override def productArity: Int = throw new 
UnsupportedOperationException("Spark33Scan")
-
-  override def canEqual(that: Any): Boolean = throw new 
UnsupportedOperationException("Spark33Scan")
-
-}
diff --git 
a/shims/spark34/src/main/scala/org/apache/spark/sql/execution/datasources/v2/Spark34Scan.scala
 
b/shims/spark34/src/main/scala/org/apache/spark/sql/execution/datasources/v2/Spark34Scan.scala
deleted file mode 100644
index 393db5b87c..0000000000
--- 
a/shims/spark34/src/main/scala/org/apache/spark/sql/execution/datasources/v2/Spark34Scan.scala
+++ /dev/null
@@ -1,50 +0,0 @@
-/*
- * Licensed to the Apache Software Foundation (ASF) under one or more
- * contributor license agreements.  See the NOTICE file distributed with
- * this work for additional information regarding copyright ownership.
- * The ASF licenses this file to You under the Apache License, Version 2.0
- * (the "License"); you may not use this file except in compliance with
- * the License.  You may obtain a copy of the License at
- *
- *    http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-package org.apache.spark.sql.execution.datasources.v2
-
-import org.apache.spark.rdd.RDD
-import org.apache.spark.sql.catalyst.InternalRow
-import org.apache.spark.sql.catalyst.expressions.{Attribute, Expression, 
SortOrder}
-import org.apache.spark.sql.connector.read.{InputPartition, 
PartitionReaderFactory, Scan}
-
-class Spark34Scan extends DataSourceV2ScanExecBase {
-
-  override def scan: Scan = throw new 
UnsupportedOperationException("Spark34Scan")
-
-  override def ordering: Option[Seq[SortOrder]] = throw new 
UnsupportedOperationException(
-    "Spark34Scan")
-
-  override def readerFactory: PartitionReaderFactory =
-    throw new UnsupportedOperationException("Spark34Scan")
-
-  override def keyGroupedPartitioning: Option[Seq[Expression]] =
-    throw new UnsupportedOperationException("Spark34Scan")
-
-  override protected def inputPartitions: Seq[InputPartition] =
-    throw new UnsupportedOperationException("Spark34Scan")
-
-  override def inputRDD: RDD[InternalRow] = throw new 
UnsupportedOperationException("Spark34Scan")
-
-  override def output: Seq[Attribute] = throw new 
UnsupportedOperationException("Spark34Scan")
-
-  override def productElement(n: Int): Any = throw new 
UnsupportedOperationException("Spark34Scan")
-
-  override def productArity: Int = throw new 
UnsupportedOperationException("Spark34Scan")
-
-  override def canEqual(that: Any): Boolean = throw new 
UnsupportedOperationException("Spark34Scan")
-
-}
diff --git 
a/shims/spark35/src/main/scala/org/apache/spark/sql/execution/datasources/v2/Spark35Scan.scala
 
b/shims/spark35/src/main/scala/org/apache/spark/sql/execution/datasources/v2/Spark35Scan.scala
deleted file mode 100644
index 98fcfa5483..0000000000
--- 
a/shims/spark35/src/main/scala/org/apache/spark/sql/execution/datasources/v2/Spark35Scan.scala
+++ /dev/null
@@ -1,50 +0,0 @@
-/*
- * Licensed to the Apache Software Foundation (ASF) under one or more
- * contributor license agreements.  See the NOTICE file distributed with
- * this work for additional information regarding copyright ownership.
- * The ASF licenses this file to You under the Apache License, Version 2.0
- * (the "License"); you may not use this file except in compliance with
- * the License.  You may obtain a copy of the License at
- *
- *    http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-package org.apache.spark.sql.execution.datasources.v2
-
-import org.apache.spark.rdd.RDD
-import org.apache.spark.sql.catalyst.InternalRow
-import org.apache.spark.sql.catalyst.expressions.{Attribute, Expression, 
SortOrder}
-import org.apache.spark.sql.connector.read.{InputPartition, 
PartitionReaderFactory, Scan}
-
-class Spark35Scan extends DataSourceV2ScanExecBase {
-
-  override def scan: Scan = throw new 
UnsupportedOperationException("Spark35Scan")
-
-  override def ordering: Option[Seq[SortOrder]] = throw new 
UnsupportedOperationException(
-    "Spark35Scan")
-
-  override def readerFactory: PartitionReaderFactory =
-    throw new UnsupportedOperationException("Spark35Scan")
-
-  override def keyGroupedPartitioning: Option[Seq[Expression]] =
-    throw new UnsupportedOperationException("Spark35Scan")
-
-  override protected def inputPartitions: Seq[InputPartition] =
-    throw new UnsupportedOperationException("Spark35Scan")
-
-  override def inputRDD: RDD[InternalRow] = throw new 
UnsupportedOperationException("Spark35Scan")
-
-  override def output: Seq[Attribute] = throw new 
UnsupportedOperationException("Spark35Scan")
-
-  override def productElement(n: Int): Any = throw new 
UnsupportedOperationException("Spark35Scan")
-
-  override def productArity: Int = throw new 
UnsupportedOperationException("Spark35Scan")
-
-  override def canEqual(that: Any): Boolean = throw new 
UnsupportedOperationException("Spark35Scan")
-
-}


---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to