This is an automated email from the ASF dual-hosted git repository. ruifengz pushed a commit to branch master in repository https://gitbox.apache.org/repos/asf/spark.git
The following commit(s) were added to refs/heads/master by this push: new f8ab74e6a6c3 [SPARK-52703][INFRA][PS] Upgrade minimum python version of pandas api to 3.10 f8ab74e6a6c3 is described below commit f8ab74e6a6c309f42e2823c3674e8c1475804853 Author: Ruifeng Zheng <ruife...@apache.org> AuthorDate: Tue Jul 8 17:36:56 2025 +0800 [SPARK-52703][INFRA][PS] Upgrade minimum python version of pandas api to 3.10 ### What changes were proposed in this pull request? Upgrade minimum python version of pandas api to 3.10 ### Why are the changes needed? python 3.9 is reaching the EOL, we should upgrade the minimum python version ### Does this PR introduce _any_ user-facing change? No, infra-only ### How was this patch tested? PR builder with env ``` default: '{"PYSPARK_IMAGE_TO_TEST": "python-ps-minimum", "PYTHON_TO_TEST": "python3.10"}' ``` https://github.com/zhengruifeng/spark/actions/runs/16133332146/job/45534172036 ### Was this patch authored or co-authored using generative AI tooling? no Closes #51397 from zhengruifeng/ps_py_310. Authored-by: Ruifeng Zheng <ruife...@apache.org> Signed-off-by: Ruifeng Zheng <ruife...@apache.org> --- .github/workflows/build_python_ps_minimum.yml | 2 +- dev/spark-test-image/python-ps-minimum/Dockerfile | 25 ++++++++--------------- 2 files changed, 9 insertions(+), 18 deletions(-) diff --git a/.github/workflows/build_python_ps_minimum.yml b/.github/workflows/build_python_ps_minimum.yml index 3aa83ff06a99..ed80a904ebd7 100644 --- a/.github/workflows/build_python_ps_minimum.yml +++ b/.github/workflows/build_python_ps_minimum.yml @@ -38,7 +38,7 @@ jobs: envs: >- { "PYSPARK_IMAGE_TO_TEST": "python-ps-minimum", - "PYTHON_TO_TEST": "python3.9" + "PYTHON_TO_TEST": "python3.10" } jobs: >- { diff --git a/dev/spark-test-image/python-ps-minimum/Dockerfile b/dev/spark-test-image/python-ps-minimum/Dockerfile index 0cdf1fa6aa1f..2b579c5e10b9 100644 --- a/dev/spark-test-image/python-ps-minimum/Dockerfile +++ b/dev/spark-test-image/python-ps-minimum/Dockerfile @@ -24,11 +24,12 @@ LABEL org.opencontainers.image.ref.name="Apache Spark Infra Image For Pandas API # Overwrite this label to avoid exposing the underlying Ubuntu OS version label LABEL org.opencontainers.image.version="" -ENV FULL_REFRESH_DATE=20250327 +ENV FULL_REFRESH_DATE=20250708 ENV DEBIAN_FRONTEND=noninteractive ENV DEBCONF_NONINTERACTIVE_SEEN=true +# Should keep the installation consistent with https://apache.github.io/spark/api/python/getting_started/install.html RUN apt-get update && apt-get install -y \ build-essential \ ca-certificates \ @@ -52,6 +53,8 @@ RUN apt-get update && apt-get install -y \ libxml2-dev \ openjdk-17-jdk-headless \ pkg-config \ + python3.10 \ + python3-psutil \ qpdf \ tzdata \ software-properties-common \ @@ -59,23 +62,11 @@ RUN apt-get update && apt-get install -y \ zlib1g-dev -# Should keep the installation consistent with https://apache.github.io/spark/api/python/getting_started/install.html - -# Install Python 3.9 -RUN add-apt-repository ppa:deadsnakes/ppa -RUN apt-get update && apt-get install -y \ - python3.9 \ - python3.9-distutils \ - && apt-get autoremove --purge -y \ - && apt-get clean \ - && rm -rf /var/lib/apt/lists/* - - ARG BASIC_PIP_PKGS="pyarrow==11.0.0 pandas==2.2.0 six==1.16.0 numpy scipy coverage unittest-xml-reporting" # Python deps for Spark Connect ARG CONNECT_PIP_PKGS="grpcio==1.67.0 grpcio-status==1.67.0 googleapis-common-protos==1.65.0 graphviz==0.20 protobuf" -# Install Python 3.9 packages -RUN curl -sS https://bootstrap.pypa.io/get-pip.py | python3.9 -RUN python3.9 -m pip install --force $BASIC_PIP_PKGS $CONNECT_PIP_PKGS && \ - python3.9 -m pip cache purge +# Install Python 3.10 packages +RUN curl -sS https://bootstrap.pypa.io/get-pip.py | python3.10 +RUN python3.10 -m pip install --force $BASIC_PIP_PKGS $CONNECT_PIP_PKGS && \ + python3.10 -m pip cache purge --------------------------------------------------------------------- To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org For additional commands, e-mail: commits-h...@spark.apache.org