This is an automated email from the ASF dual-hosted git repository.

beliefer pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/spark.git


The following commit(s) were added to refs/heads/master by this push:
     new f5cf3101991 [SPARK-45412][PYTHON][CONNECT][FOLLOW-UP] Remove 
unnecessary check
f5cf3101991 is described below

commit f5cf310199132f62b779e0244d15f7680e2ba856
Author: Ruifeng Zheng <ruife...@apache.org>
AuthorDate: Mon Oct 9 18:07:59 2023 +0800

    [SPARK-45412][PYTHON][CONNECT][FOLLOW-UP] Remove unnecessary check
    
    ### What changes were proposed in this pull request?
    Remove unnecessary check
    
    ### Why are the changes needed?
    https://github.com/apache/spark/pull/43215 already validates the plan in 
`__init__`
    
    ### Does this PR introduce _any_ user-facing change?
    no
    
    ### How was this patch tested?
    ci
    
    ### Was this patch authored or co-authored using generative AI tooling?
    no
    
    Closes #43287 from zhengruifeng/SPARK-45412-followup.
    
    Authored-by: Ruifeng Zheng <ruife...@apache.org>
    Signed-off-by: Jiaan Geng <belie...@163.com>
---
 python/pyspark/sql/connect/dataframe.py | 16 +---------------
 1 file changed, 1 insertion(+), 15 deletions(-)

diff --git a/python/pyspark/sql/connect/dataframe.py 
b/python/pyspark/sql/connect/dataframe.py
index 2c0a75fad46..4044fab3bb3 100644
--- a/python/pyspark/sql/connect/dataframe.py
+++ b/python/pyspark/sql/connect/dataframe.py
@@ -169,7 +169,6 @@ class DataFrame:
 
     @property
     def write(self) -> "DataFrameWriter":
-        assert self._plan is not None
         return DataFrameWriter(self._plan, self._session)
 
     write.__doc__ = PySparkDataFrame.write.__doc__
@@ -1096,11 +1095,6 @@ class DataFrame:
     union.__doc__ = PySparkDataFrame.union.__doc__
 
     def unionAll(self, other: "DataFrame") -> "DataFrame":
-        if other._plan is None:
-            raise PySparkValueError(
-                error_class="MISSING_VALID_PLAN",
-                message_parameters={"operator": "Union"},
-            )
         self._check_same_session(other)
         return DataFrame.withPlan(
             plan.SetOperation(self._plan, other._plan, "union", is_all=True), 
session=self._session
@@ -2030,8 +2024,6 @@ class DataFrame:
     mapInArrow.__doc__ = PySparkDataFrame.mapInArrow.__doc__
 
     def foreach(self, f: Callable[[Row], None]) -> None:
-        assert self._plan is not None
-
         def foreach_func(row: Any) -> None:
             f(row)
 
@@ -2042,8 +2034,6 @@ class DataFrame:
     foreach.__doc__ = PySparkDataFrame.foreach.__doc__
 
     def foreachPartition(self, f: Callable[[Iterator[Row]], None]) -> None:
-        assert self._plan is not None
-
         schema = self.schema
         field_converters = [
             ArrowTableToRowsConversion._create_converter(f.dataType) for f in 
schema.fields
@@ -2069,14 +2059,12 @@ class DataFrame:
 
     @property
     def writeStream(self) -> DataStreamWriter:
-        assert self._plan is not None
         return DataStreamWriter(plan=self._plan, session=self._session)
 
     writeStream.__doc__ = PySparkDataFrame.writeStream.__doc__
 
     def sameSemantics(self, other: "DataFrame") -> bool:
-        assert self._plan is not None
-        assert other._plan is not None
+        self._check_same_session(other)
         return self._session.client.same_semantics(
             plan=self._plan.to_proto(self._session.client),
             other=other._plan.to_proto(other._session.client),
@@ -2085,7 +2073,6 @@ class DataFrame:
     sameSemantics.__doc__ = PySparkDataFrame.sameSemantics.__doc__
 
     def semanticHash(self) -> int:
-        assert self._plan is not None
         return self._session.client.semantic_hash(
             plan=self._plan.to_proto(self._session.client),
         )
@@ -2093,7 +2080,6 @@ class DataFrame:
     semanticHash.__doc__ = PySparkDataFrame.semanticHash.__doc__
 
     def writeTo(self, table: str) -> "DataFrameWriterV2":
-        assert self._plan is not None
         return DataFrameWriterV2(self._plan, self._session, table)
 
     writeTo.__doc__ = PySparkDataFrame.writeTo.__doc__


---------------------------------------------------------------------
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org

Reply via email to