This is an automated email from the ASF dual-hosted git repository.
chengchengjin pushed a commit to branch main
in repository https://gitbox.apache.org/repos/asf/incubator-gluten.git
The following commit(s) were added to refs/heads/main by this push:
new f1fc5697d0 [CORE] Shim: Remove unused Spark33Scan / Spark34Scan /
Spark35Scan (#10051)
f1fc5697d0 is described below
commit f1fc5697d086a5b4a518edba318f44b7c7bf0860
Author: Hongze Zhang <[email protected]>
AuthorDate: Fri Jun 27 12:04:00 2025 +0800
[CORE] Shim: Remove unused Spark33Scan / Spark34Scan / Spark35Scan (#10051)
---
.../sql/execution/datasources/v2/Spark33Scan.scala | 47 --------------------
.../sql/execution/datasources/v2/Spark34Scan.scala | 50 ----------------------
.../sql/execution/datasources/v2/Spark35Scan.scala | 50 ----------------------
3 files changed, 147 deletions(-)
diff --git
a/shims/spark33/src/main/scala/org/apache/spark/sql/execution/datasources/v2/Spark33Scan.scala
b/shims/spark33/src/main/scala/org/apache/spark/sql/execution/datasources/v2/Spark33Scan.scala
deleted file mode 100644
index 169d5c3cb8..0000000000
---
a/shims/spark33/src/main/scala/org/apache/spark/sql/execution/datasources/v2/Spark33Scan.scala
+++ /dev/null
@@ -1,47 +0,0 @@
-/*
- * Licensed to the Apache Software Foundation (ASF) under one or more
- * contributor license agreements. See the NOTICE file distributed with
- * this work for additional information regarding copyright ownership.
- * The ASF licenses this file to You under the Apache License, Version 2.0
- * (the "License"); you may not use this file except in compliance with
- * the License. You may obtain a copy of the License at
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-package org.apache.spark.sql.execution.datasources.v2
-
-import org.apache.spark.rdd.RDD
-import org.apache.spark.sql.catalyst.InternalRow
-import org.apache.spark.sql.catalyst.expressions.{Attribute, Expression}
-import org.apache.spark.sql.connector.read.{InputPartition,
PartitionReaderFactory, Scan}
-
-class Spark33Scan extends DataSourceV2ScanExecBase {
-
- override def scan: Scan = throw new
UnsupportedOperationException("Spark33Scan")
-
- override def readerFactory: PartitionReaderFactory =
- throw new UnsupportedOperationException("Spark33Scan")
-
- override def keyGroupedPartitioning: Option[Seq[Expression]] =
- throw new UnsupportedOperationException("Spark33Scan")
-
- override protected def inputPartitions: Seq[InputPartition] =
- throw new UnsupportedOperationException("Spark33Scan")
-
- override def inputRDD: RDD[InternalRow] = throw new
UnsupportedOperationException("Spark33Scan")
-
- override def output: Seq[Attribute] = throw new
UnsupportedOperationException("Spark33Scan")
-
- override def productElement(n: Int): Any = throw new
UnsupportedOperationException("Spark33Scan")
-
- override def productArity: Int = throw new
UnsupportedOperationException("Spark33Scan")
-
- override def canEqual(that: Any): Boolean = throw new
UnsupportedOperationException("Spark33Scan")
-
-}
diff --git
a/shims/spark34/src/main/scala/org/apache/spark/sql/execution/datasources/v2/Spark34Scan.scala
b/shims/spark34/src/main/scala/org/apache/spark/sql/execution/datasources/v2/Spark34Scan.scala
deleted file mode 100644
index 393db5b87c..0000000000
---
a/shims/spark34/src/main/scala/org/apache/spark/sql/execution/datasources/v2/Spark34Scan.scala
+++ /dev/null
@@ -1,50 +0,0 @@
-/*
- * Licensed to the Apache Software Foundation (ASF) under one or more
- * contributor license agreements. See the NOTICE file distributed with
- * this work for additional information regarding copyright ownership.
- * The ASF licenses this file to You under the Apache License, Version 2.0
- * (the "License"); you may not use this file except in compliance with
- * the License. You may obtain a copy of the License at
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-package org.apache.spark.sql.execution.datasources.v2
-
-import org.apache.spark.rdd.RDD
-import org.apache.spark.sql.catalyst.InternalRow
-import org.apache.spark.sql.catalyst.expressions.{Attribute, Expression,
SortOrder}
-import org.apache.spark.sql.connector.read.{InputPartition,
PartitionReaderFactory, Scan}
-
-class Spark34Scan extends DataSourceV2ScanExecBase {
-
- override def scan: Scan = throw new
UnsupportedOperationException("Spark34Scan")
-
- override def ordering: Option[Seq[SortOrder]] = throw new
UnsupportedOperationException(
- "Spark34Scan")
-
- override def readerFactory: PartitionReaderFactory =
- throw new UnsupportedOperationException("Spark34Scan")
-
- override def keyGroupedPartitioning: Option[Seq[Expression]] =
- throw new UnsupportedOperationException("Spark34Scan")
-
- override protected def inputPartitions: Seq[InputPartition] =
- throw new UnsupportedOperationException("Spark34Scan")
-
- override def inputRDD: RDD[InternalRow] = throw new
UnsupportedOperationException("Spark34Scan")
-
- override def output: Seq[Attribute] = throw new
UnsupportedOperationException("Spark34Scan")
-
- override def productElement(n: Int): Any = throw new
UnsupportedOperationException("Spark34Scan")
-
- override def productArity: Int = throw new
UnsupportedOperationException("Spark34Scan")
-
- override def canEqual(that: Any): Boolean = throw new
UnsupportedOperationException("Spark34Scan")
-
-}
diff --git
a/shims/spark35/src/main/scala/org/apache/spark/sql/execution/datasources/v2/Spark35Scan.scala
b/shims/spark35/src/main/scala/org/apache/spark/sql/execution/datasources/v2/Spark35Scan.scala
deleted file mode 100644
index 98fcfa5483..0000000000
---
a/shims/spark35/src/main/scala/org/apache/spark/sql/execution/datasources/v2/Spark35Scan.scala
+++ /dev/null
@@ -1,50 +0,0 @@
-/*
- * Licensed to the Apache Software Foundation (ASF) under one or more
- * contributor license agreements. See the NOTICE file distributed with
- * this work for additional information regarding copyright ownership.
- * The ASF licenses this file to You under the Apache License, Version 2.0
- * (the "License"); you may not use this file except in compliance with
- * the License. You may obtain a copy of the License at
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-package org.apache.spark.sql.execution.datasources.v2
-
-import org.apache.spark.rdd.RDD
-import org.apache.spark.sql.catalyst.InternalRow
-import org.apache.spark.sql.catalyst.expressions.{Attribute, Expression,
SortOrder}
-import org.apache.spark.sql.connector.read.{InputPartition,
PartitionReaderFactory, Scan}
-
-class Spark35Scan extends DataSourceV2ScanExecBase {
-
- override def scan: Scan = throw new
UnsupportedOperationException("Spark35Scan")
-
- override def ordering: Option[Seq[SortOrder]] = throw new
UnsupportedOperationException(
- "Spark35Scan")
-
- override def readerFactory: PartitionReaderFactory =
- throw new UnsupportedOperationException("Spark35Scan")
-
- override def keyGroupedPartitioning: Option[Seq[Expression]] =
- throw new UnsupportedOperationException("Spark35Scan")
-
- override protected def inputPartitions: Seq[InputPartition] =
- throw new UnsupportedOperationException("Spark35Scan")
-
- override def inputRDD: RDD[InternalRow] = throw new
UnsupportedOperationException("Spark35Scan")
-
- override def output: Seq[Attribute] = throw new
UnsupportedOperationException("Spark35Scan")
-
- override def productElement(n: Int): Any = throw new
UnsupportedOperationException("Spark35Scan")
-
- override def productArity: Int = throw new
UnsupportedOperationException("Spark35Scan")
-
- override def canEqual(that: Any): Boolean = throw new
UnsupportedOperationException("Spark35Scan")
-
-}
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]