This is an automated email from the ASF dual-hosted git repository. gurwls223 pushed a commit to branch master in repository https://gitbox.apache.org/repos/asf/spark.git
The following commit(s) were added to refs/heads/master by this push: new dada63c24ec [SPARK-43603][PS][CONNECT][TESTS][FOLLOW-UP] Delete unused `test_parity_dataframe.py` dada63c24ec is described below commit dada63c24ecbe1b65dc4dac6632b34db05fbef98 Author: Ruifeng Zheng <ruife...@apache.org> AuthorDate: Tue May 30 12:05:16 2023 +0900 [SPARK-43603][PS][CONNECT][TESTS][FOLLOW-UP] Delete unused `test_parity_dataframe.py` ### What changes were proposed in this pull request? `pyspark.pandas.tests.connect.test_parity_dataframe` had been split in https://github.com/apache/spark/commit/cb5bd57aac40c06921321929df00d2086e37ba34, but I forgot to remove the test file. this PR removes unused `test_parity_dataframe.py` ### Why are the changes needed? `test_parity_dataframe.py` is no longer needed ### Does this PR introduce _any_ user-facing change? No, test-only ### How was this patch tested? CI Closes #41372 from zhengruifeng/reorg_ps_df_tests_followup. Authored-by: Ruifeng Zheng <ruife...@apache.org> Signed-off-by: Hyukjin Kwon <gurwls...@apache.org> --- .../pandas/tests/connect/test_parity_dataframe.py | 154 --------------------- 1 file changed, 154 deletions(-) diff --git a/python/pyspark/pandas/tests/connect/test_parity_dataframe.py b/python/pyspark/pandas/tests/connect/test_parity_dataframe.py deleted file mode 100644 index c1b9ae2ee11..00000000000 --- a/python/pyspark/pandas/tests/connect/test_parity_dataframe.py +++ /dev/null @@ -1,154 +0,0 @@ -# -# Licensed to the Apache Software Foundation (ASF) under one or more -# contributor license agreements. See the NOTICE file distributed with -# this work for additional information regarding copyright ownership. -# The ASF licenses this file to You under the Apache License, Version 2.0 -# (the "License"); you may not use this file except in compliance with -# the License. You may obtain a copy of the License at -# -# http://www.apache.org/licenses/LICENSE-2.0 -# -# Unless required by applicable law or agreed to in writing, software -# distributed under the License is distributed on an "AS IS" BASIS, -# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -# See the License for the specific language governing permissions and -# limitations under the License. -# -import unittest - -from pyspark import pandas as ps -from pyspark.pandas.tests.test_dataframe import DataFrameTestsMixin -from pyspark.testing.connectutils import ReusedConnectTestCase -from pyspark.testing.pandasutils import PandasOnSparkTestUtils - - -class DataFrameParityTests(DataFrameTestsMixin, PandasOnSparkTestUtils, ReusedConnectTestCase): - @property - def psdf(self): - return ps.from_pandas(self.pdf) - - @unittest.skip( - "TODO(SPARK-43610): Enable `InternalFrame.attach_distributed_column` in Spark Connect." - ) - def test_aggregate(self): - super().test_aggregate() - - @unittest.skip("TODO(SPARK-41876): Implement DataFrame `toLocalIterator`") - def test_iterrows(self): - super().test_iterrows() - - @unittest.skip("TODO(SPARK-41876): Implement DataFrame `toLocalIterator`") - def test_itertuples(self): - super().test_itertuples() - - @unittest.skip( - "TODO(SPARK-43611): Fix unexpected `AnalysisException` from Spark Connect client." - ) - def test_cummax(self): - super().test_cummax() - - @unittest.skip( - "TODO(SPARK-43611): Fix unexpected `AnalysisException` from Spark Connect client." - ) - def test_cummax_multiindex_columns(self): - super().test_cummax_multiindex_columns() - - @unittest.skip( - "TODO(SPARK-43611): Fix unexpected `AnalysisException` from Spark Connect client." - ) - def test_cummin(self): - super().test_cummin() - - @unittest.skip( - "TODO(SPARK-43611): Fix unexpected `AnalysisException` from Spark Connect client." - ) - def test_cummin_multiindex_columns(self): - super().test_cummin_multiindex_columns() - - @unittest.skip( - "TODO(SPARK-43611): Fix unexpected `AnalysisException` from Spark Connect client." - ) - def test_cumprod(self): - super().test_cumprod() - - @unittest.skip( - "TODO(SPARK-43611): Fix unexpected `AnalysisException` from Spark Connect client." - ) - def test_cumprod_multiindex_columns(self): - super().test_cumprod_multiindex_columns() - - @unittest.skip( - "TODO(SPARK-43611): Fix unexpected `AnalysisException` from Spark Connect client." - ) - def test_cumsum(self): - super().test_cumsum() - - @unittest.skip( - "TODO(SPARK-43611): Fix unexpected `AnalysisException` from Spark Connect client." - ) - def test_cumsum_multiindex_columns(self): - super().test_cumsum_multiindex_columns() - - @unittest.skip( - "TODO(SPARK-43616): Enable pyspark.pandas.spark.functions.repeat in Spark Connect." - ) - def test_binary_operator_multiply(self): - super().test_binary_operator_multiply() - - @unittest.skip("TODO(SPARK-43622): Enable pyspark.pandas.spark.functions.var in Spark Connect.") - def test_dataframe(self): - super().test_dataframe() - - @unittest.skip( - "TODO(SPARK-43611): Fix unexpected `AnalysisException` from Spark Connect client." - ) - def test_fillna(self): - return super().test_fillna() - - @unittest.skip( - "TODO(SPARK-43611): Fix unexpected `AnalysisException` from Spark Connect client." - ) - def test_pivot_table(self): - super().test_pivot_table() - - @unittest.skip( - "TODO(SPARK-43611): Fix unexpected `AnalysisException` from Spark Connect client." - ) - def test_pivot_table_dtypes(self): - super().test_pivot_table_dtypes() - - @unittest.skip( - "TODO(SPARK-43611): Fix unexpected `AnalysisException` from Spark Connect client." - ) - def test_reset_index_with_default_index_types(self): - super().test_reset_index_with_default_index_types() - - @unittest.skip( - "TODO(SPARK-43611): Fix unexpected `AnalysisException` from Spark Connect client." - ) - def test_transpose(self): - super().test_transpose() - - @unittest.skip( - "TODO(SPARK-43610): Enable `InternalFrame.attach_distributed_column` in Spark Connect." - ) - def test_unstack(self): - super().test_unstack() - - @unittest.skip( - "TODO(SPARK-43611): Fix unexpected `AnalysisException` from Spark Connect client." - ) - def test_append(self): - super().test_append() - - -if __name__ == "__main__": - from pyspark.pandas.tests.connect.test_parity_dataframe import * # noqa: F401 - - try: - import xmlrunner # type: ignore[import] - - testRunner = xmlrunner.XMLTestRunner(output="target/test-reports", verbosity=2) - except ImportError: - testRunner = None - unittest.main(testRunner=testRunner, verbosity=2) --------------------------------------------------------------------- To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org For additional commands, e-mail: commits-h...@spark.apache.org