Hyukjin Kwon created SPARK-46181:
------------------------------------

             Summary: Split scheduled Python build
                 Key: SPARK-46181
                 URL: https://issues.apache.org/jira/browse/SPARK-46181
             Project: Spark
          Issue Type: Sub-task
          Components: Project Infra, PySpark
    Affects Versions: 4.0.0
            Reporter: Hyukjin Kwon


Python 3.12 build fails as below:

{code}
/__w/spark/spark/python/pyspark/pandas/utils.py:1015: 
PandasAPIOnSparkAdviceWarning: `to_pandas` loads all data into the driver's 
memory. It should only be used if the resulting pandas Series is expected to be 
small.
  warnings.warn(message, PandasAPIOnSparkAdviceWarning)
/__w/spark/spark/python/pyspark/testing/pandasutils.py:401: FutureWarning: 
`assertPandasOnSparkEqual` will be removed in Spark 4.0.0. Use 
`ps.testing.assert_frame_equal`, `ps.testing.assert_series_equal` and 
`ps.testing.assert_index_equal` instead.
  warnings.warn(
/__w/spark/spark/python/pyspark/pandas/utils.py:1015: 
PandasAPIOnSparkAdviceWarning: `to_pandas` loads all data into the driver's 
memory. It should only be used if the resulting pandas DataFrame is expected to 
be small.
  warnings.warn(message, PandasAPIOnSparkAdviceWarning)
ok (15.809s)
  test_groupby_rolling_sum 
(pyspark.pandas.tests.connect.test_parity_ops_on_diff_frames_groupby_rolling.OpsOnDiffFramesGroupByRollingParityTests.test_groupby_rolling_sum)
 ... ERROR StatusCo
Had test failures in 
pyspark.pandas.tests.connect.test_parity_ops_on_diff_frames_groupby_rolling 
with python3.12; see logs.
Error:  running /__w/spark/spark/python/run-tests 
--modules=pyspark-pandas-connect-part2 --parallelism=1 
--python-executables=pypy3,python3.10,python3.11,python3.12 ; received return 
code 255
Error: Process completed with exit code 19.
{code}

https://github.com/apache/spark/actions/runs/7034467411/job/19154767856

I suspect this by OOM. We could split the build.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org

Reply via email to