kou commented on a change in pull request #9210:
URL: https://github.com/apache/arrow/pull/9210#discussion_r569110478
##########
File path: docker-compose.yml
##########
@@ -1308,3 +1308,34 @@ services:
/arrow/ci/scripts/python_build.sh /arrow /build &&
/arrow/ci/scripts/java_build.sh /arrow /build &&
/arrow/ci/scripts/integration_spark.sh /arrow /spark"]
+
+ conda-python-spark-pyarrow-only:
+ # Usage:
+ # docker-compose build conda-cpp
+ # docker-compose build conda-python
+ # docker-compose build conda-python-spark-pyarrow-only
+ # docker-compose run conda-python-spark-pyarrow-only
+ image: ${REPO}:${ARCH}-conda-python-${PYTHON}-spark-${SPARK}-pyarrow-only
+ build:
+ context: .
+ dockerfile: ci/docker/conda-python-spark.dockerfile
+ cache_from:
+ - ${REPO}:${ARCH}-conda-python-${PYTHON}-spark-${SPARK}-pyarrow-only
+ args:
+ repo: ${REPO}
+ arch: ${ARCH}
+ python: ${PYTHON}
+ jdk: ${JDK}
+ # conda-forge doesn't have 3.5.4 so pinning explicitly, but this should
+ # be set to ${MAVEN}
+ maven: 3.5
+ spark: ${SPARK}
+ shm_size: *shm-size
+ environment:
+ <<: *ccache
+ volumes: *conda-maven-volumes
+ command:
+ ["/arrow/ci/scripts/cpp_build.sh /arrow /build &&
+ /arrow/ci/scripts/python_build.sh /arrow /build &&
+ /arrow/ci/scripts/java_build.sh /arrow /build &&
+ /arrow/ci/scripts/integration_spark.sh /arrow /spark true"]
Review comment:
How about the following?
```diff
diff --git a/dev/tasks/tasks.yml b/dev/tasks/tasks.yml
index 63f38ab8f..af112c036 100644
--- a/dev/tasks/tasks.yml
+++ b/dev/tasks/tasks.yml
@@ -1945,8 +1945,9 @@ tasks:
env:
PYTHON: 3.7
SPARK: "branch-3.0"
+ TEST_PYARROW_ONLY: "true"
# use the branch-3.0 of spark, so prevent reusing any layers
- run: --no-leaf-cache conda-python-spark-pyarrow-only
+ run: --no-leaf-cache conda-python-spark
test-conda-python-3.8-spark-master:
ci: github
diff --git a/docker-compose.yml b/docker-compose.yml
index aa688de01..ca8989796 100644
--- a/docker-compose.yml
+++ b/docker-compose.yml
@@ -91,7 +91,6 @@ x-hierarchy:
- conda-python-turbodbc
- conda-python-kartothek
- conda-python-spark
- - conda-python-spark-pyarrow-only
- debian-cpp:
- debian-c-glib:
- debian-ruby
@@ -1340,35 +1339,4 @@ services:
["/arrow/ci/scripts/cpp_build.sh /arrow /build &&
/arrow/ci/scripts/python_build.sh /arrow /build &&
/arrow/ci/scripts/java_build.sh /arrow /build &&
- /arrow/ci/scripts/integration_spark.sh /arrow /spark"]
-
- conda-python-spark-pyarrow-only:
- # Usage:
- # docker-compose build conda-cpp
- # docker-compose build conda-python
- # docker-compose build conda-python-spark-pyarrow-only
- # docker-compose run conda-python-spark-pyarrow-only
- image:
${REPO}:${ARCH}-conda-python-${PYTHON}-spark-${SPARK}-pyarrow-only
- build:
- context: .
- dockerfile: ci/docker/conda-python-spark.dockerfile
- cache_from:
- - ${REPO}:${ARCH}-conda-python-${PYTHON}-spark-${SPARK}-pyarrow-only
- args:
- repo: ${REPO}
- arch: ${ARCH}
- python: ${PYTHON}
- jdk: ${JDK}
- # conda-forge doesn't have 3.5.4 so pinning explicitly, but this
should
- # be set to ${MAVEN}
- maven: 3.5
- spark: ${SPARK}
- shm_size: *shm-size
- environment:
- <<: *ccache
- volumes: *conda-maven-volumes
- command:
- ["/arrow/ci/scripts/cpp_build.sh /arrow /build &&
- /arrow/ci/scripts/python_build.sh /arrow /build &&
- /arrow/ci/scripts/java_build.sh /arrow /build &&
- /arrow/ci/scripts/integration_spark.sh /arrow /spark true"]
+ /arrow/ci/scripts/integration_spark.sh /arrow /spark
${TEST_PYARROW_ONLY:-false}"]
```
----------------------------------------------------------------
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
For queries about this service, please contact Infrastructure at:
[email protected]