[spark] branch master updated (b890fdc -> 8e36a8f)
This is an automated email from the ASF dual-hosted git repository. gurwls223 pushed a change to branch master in repository https://gitbox.apache.org/repos/asf/spark.git. from b890fdc [SPARK-32387][SS] Extract UninterruptibleThread runner logic from KafkaOffsetReader add 8e36a8f [SPARK-32419][PYTHON][BUILD] Avoid using subshell for Conda env (de)activation in pip packaging test No new revisions were added by this update. Summary of changes: dev/run-pip-tests | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) - To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org For additional commands, e-mail: commits-h...@spark.apache.org
[spark] branch master updated (b890fdc -> 8e36a8f)
This is an automated email from the ASF dual-hosted git repository. gurwls223 pushed a change to branch master in repository https://gitbox.apache.org/repos/asf/spark.git. from b890fdc [SPARK-32387][SS] Extract UninterruptibleThread runner logic from KafkaOffsetReader add 8e36a8f [SPARK-32419][PYTHON][BUILD] Avoid using subshell for Conda env (de)activation in pip packaging test No new revisions were added by this update. Summary of changes: dev/run-pip-tests | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) - To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org For additional commands, e-mail: commits-h...@spark.apache.org
[spark] branch master updated (b890fdc -> 8e36a8f)
This is an automated email from the ASF dual-hosted git repository. gurwls223 pushed a change to branch master in repository https://gitbox.apache.org/repos/asf/spark.git. from b890fdc [SPARK-32387][SS] Extract UninterruptibleThread runner logic from KafkaOffsetReader add 8e36a8f [SPARK-32419][PYTHON][BUILD] Avoid using subshell for Conda env (de)activation in pip packaging test No new revisions were added by this update. Summary of changes: dev/run-pip-tests | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) - To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org For additional commands, e-mail: commits-h...@spark.apache.org
[spark] branch master updated (b890fdc -> 8e36a8f)
This is an automated email from the ASF dual-hosted git repository. gurwls223 pushed a change to branch master in repository https://gitbox.apache.org/repos/asf/spark.git. from b890fdc [SPARK-32387][SS] Extract UninterruptibleThread runner logic from KafkaOffsetReader add 8e36a8f [SPARK-32419][PYTHON][BUILD] Avoid using subshell for Conda env (de)activation in pip packaging test No new revisions were added by this update. Summary of changes: dev/run-pip-tests | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) - To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org For additional commands, e-mail: commits-h...@spark.apache.org
[spark] branch master updated: [SPARK-32419][PYTHON][BUILD] Avoid using subshell for Conda env (de)activation in pip packaging test
This is an automated email from the ASF dual-hosted git repository. gurwls223 pushed a commit to branch master in repository https://gitbox.apache.org/repos/asf/spark.git The following commit(s) were added to refs/heads/master by this push: new 8e36a8f [SPARK-32419][PYTHON][BUILD] Avoid using subshell for Conda env (de)activation in pip packaging test 8e36a8f is described below commit 8e36a8f33fb9d20c481e9eeb0ea8155aa1569439 Author: HyukjinKwon AuthorDate: Sat Jul 25 13:09:23 2020 +0900 [SPARK-32419][PYTHON][BUILD] Avoid using subshell for Conda env (de)activation in pip packaging test ### What changes were proposed in this pull request? This PR proposes to avoid using subshell when it activates Conda environment. Looks like it ends up with activating the env within the subshell even if you use `conda` command. ### Why are the changes needed? If you take a close look for GitHub Actions log: ``` Installing dist into virtual env Processing ./python/dist/pyspark-3.1.0.dev0.tar.gz Collecting py4j==0.10.9 Downloading py4j-0.10.9-py2.py3-none-any.whl (198 kB) Using legacy setup.py install for pyspark, since package 'wheel' is not installed. Installing collected packages: py4j, pyspark Running setup.py install for pyspark: started Running setup.py install for pyspark: finished with status 'done' Successfully installed py4j-0.10.9 pyspark-3.1.0.dev0 ... Installing dist into virtual env Obtaining file:///home/runner/work/spark/spark/python Collecting py4j==0.10.9 Downloading py4j-0.10.9-py2.py3-none-any.whl (198 kB) Installing collected packages: py4j, pyspark Attempting uninstall: py4j Found existing installation: py4j 0.10.9 Uninstalling py4j-0.10.9: Successfully uninstalled py4j-0.10.9 Attempting uninstall: pyspark Found existing installation: pyspark 3.1.0.dev0 Uninstalling pyspark-3.1.0.dev0: Successfully uninstalled pyspark-3.1.0.dev0 Running setup.py develop for pyspark Successfully installed py4j-0.10.9 pyspark ``` It looks not properly using Conda as it removes the previously installed one when it reinstalls again. We should ideally test it with Conda environment as it's intended. ### Does this PR introduce _any_ user-facing change? No, dev-only. ### How was this patch tested? GitHub Actions will test. I also manually tested in my local. Closes #29212 from HyukjinKwon/SPARK-32419. Authored-by: HyukjinKwon Signed-off-by: HyukjinKwon --- dev/run-pip-tests | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/dev/run-pip-tests b/dev/run-pip-tests index be96ed9..b322d3f 100755 --- a/dev/run-pip-tests +++ b/dev/run-pip-tests @@ -85,7 +85,7 @@ for python in "${PYTHON_EXECS[@]}"; do source "$CONDA_PREFIX/etc/profile.d/conda.sh" fi conda create -y -p "$VIRTUALENV_PATH" python=$python numpy pandas pip setuptools - source activate "$VIRTUALENV_PATH" || (echo "Falling back to 'conda activate'" && conda activate "$VIRTUALENV_PATH") + source activate "$VIRTUALENV_PATH" || conda activate "$VIRTUALENV_PATH" else mkdir -p "$VIRTUALENV_PATH" virtualenv --python=$python "$VIRTUALENV_PATH" @@ -128,7 +128,7 @@ for python in "${PYTHON_EXECS[@]}"; do # conda / virtualenv environments need to be deactivated differently if [ -n "$USE_CONDA" ]; then - source deactivate || (echo "Falling back to 'conda deactivate'" && conda deactivate) + source deactivate || conda deactivate else deactivate fi - To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org For additional commands, e-mail: commits-h...@spark.apache.org
[spark] branch master updated (e6ef27b -> b890fdc)
This is an automated email from the ASF dual-hosted git repository. dongjoon pushed a change to branch master in repository https://gitbox.apache.org/repos/asf/spark.git. from e6ef27b [SPARK-32287][TESTS] Flaky Test: ExecutorAllocationManagerSuite.add executors default profile add b890fdc [SPARK-32387][SS] Extract UninterruptibleThread runner logic from KafkaOffsetReader No new revisions were added by this update. Summary of changes: .../spark/util/UninterruptibleThreadRunner.scala | 55 +++ .../util/UninterruptibleThreadRunnerSuite.scala| 64 ++ .../spark/sql/kafka010/KafkaOffsetReader.scala | 46 3 files changed, 129 insertions(+), 36 deletions(-) create mode 100644 core/src/main/scala/org/apache/spark/util/UninterruptibleThreadRunner.scala create mode 100644 core/src/test/scala/org/apache/spark/util/UninterruptibleThreadRunnerSuite.scala - To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org For additional commands, e-mail: commits-h...@spark.apache.org
[spark] branch master updated (e6ef27b -> b890fdc)
This is an automated email from the ASF dual-hosted git repository. dongjoon pushed a change to branch master in repository https://gitbox.apache.org/repos/asf/spark.git. from e6ef27b [SPARK-32287][TESTS] Flaky Test: ExecutorAllocationManagerSuite.add executors default profile add b890fdc [SPARK-32387][SS] Extract UninterruptibleThread runner logic from KafkaOffsetReader No new revisions were added by this update. Summary of changes: .../spark/util/UninterruptibleThreadRunner.scala | 55 +++ .../util/UninterruptibleThreadRunnerSuite.scala| 64 ++ .../spark/sql/kafka010/KafkaOffsetReader.scala | 46 3 files changed, 129 insertions(+), 36 deletions(-) create mode 100644 core/src/main/scala/org/apache/spark/util/UninterruptibleThreadRunner.scala create mode 100644 core/src/test/scala/org/apache/spark/util/UninterruptibleThreadRunnerSuite.scala - To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org For additional commands, e-mail: commits-h...@spark.apache.org
[spark] branch master updated (e6ef27b -> b890fdc)
This is an automated email from the ASF dual-hosted git repository. dongjoon pushed a change to branch master in repository https://gitbox.apache.org/repos/asf/spark.git. from e6ef27b [SPARK-32287][TESTS] Flaky Test: ExecutorAllocationManagerSuite.add executors default profile add b890fdc [SPARK-32387][SS] Extract UninterruptibleThread runner logic from KafkaOffsetReader No new revisions were added by this update. Summary of changes: .../spark/util/UninterruptibleThreadRunner.scala | 55 +++ .../util/UninterruptibleThreadRunnerSuite.scala| 64 ++ .../spark/sql/kafka010/KafkaOffsetReader.scala | 46 3 files changed, 129 insertions(+), 36 deletions(-) create mode 100644 core/src/main/scala/org/apache/spark/util/UninterruptibleThreadRunner.scala create mode 100644 core/src/test/scala/org/apache/spark/util/UninterruptibleThreadRunnerSuite.scala - To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org For additional commands, e-mail: commits-h...@spark.apache.org
[spark] branch master updated (e6ef27b -> b890fdc)
This is an automated email from the ASF dual-hosted git repository. dongjoon pushed a change to branch master in repository https://gitbox.apache.org/repos/asf/spark.git. from e6ef27b [SPARK-32287][TESTS] Flaky Test: ExecutorAllocationManagerSuite.add executors default profile add b890fdc [SPARK-32387][SS] Extract UninterruptibleThread runner logic from KafkaOffsetReader No new revisions were added by this update. Summary of changes: .../spark/util/UninterruptibleThreadRunner.scala | 55 +++ .../util/UninterruptibleThreadRunnerSuite.scala| 64 ++ .../spark/sql/kafka010/KafkaOffsetReader.scala | 46 3 files changed, 129 insertions(+), 36 deletions(-) create mode 100644 core/src/main/scala/org/apache/spark/util/UninterruptibleThreadRunner.scala create mode 100644 core/src/test/scala/org/apache/spark/util/UninterruptibleThreadRunnerSuite.scala - To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org For additional commands, e-mail: commits-h...@spark.apache.org
[spark] branch master updated (e6ef27b -> b890fdc)
This is an automated email from the ASF dual-hosted git repository. dongjoon pushed a change to branch master in repository https://gitbox.apache.org/repos/asf/spark.git. from e6ef27b [SPARK-32287][TESTS] Flaky Test: ExecutorAllocationManagerSuite.add executors default profile add b890fdc [SPARK-32387][SS] Extract UninterruptibleThread runner logic from KafkaOffsetReader No new revisions were added by this update. Summary of changes: .../spark/util/UninterruptibleThreadRunner.scala | 55 +++ .../util/UninterruptibleThreadRunnerSuite.scala| 64 ++ .../spark/sql/kafka010/KafkaOffsetReader.scala | 46 3 files changed, 129 insertions(+), 36 deletions(-) create mode 100644 core/src/main/scala/org/apache/spark/util/UninterruptibleThreadRunner.scala create mode 100644 core/src/test/scala/org/apache/spark/util/UninterruptibleThreadRunnerSuite.scala - To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org For additional commands, e-mail: commits-h...@spark.apache.org
[spark] branch branch-3.0 updated: [SPARK-32287][TESTS] Flaky Test: ExecutorAllocationManagerSuite.add executors default profile
This is an automated email from the ASF dual-hosted git repository. dongjoon pushed a commit to branch branch-3.0 in repository https://gitbox.apache.org/repos/asf/spark.git The following commit(s) were added to refs/heads/branch-3.0 by this push: new ab14d05 [SPARK-32287][TESTS] Flaky Test: ExecutorAllocationManagerSuite.add executors default profile ab14d05 is described below commit ab14d05624f6cc88b02e29dd5ae8d302ffcfd09d Author: Thomas Graves AuthorDate: Fri Jul 24 11:12:28 2020 -0700 [SPARK-32287][TESTS] Flaky Test: ExecutorAllocationManagerSuite.add executors default profile I wasn't able to reproduce the failure but the best I can tell is that the allocation manager timer triggers and call doRequest. The timeout is 10s so try to increase that to 30seconds. test failure no unit test Closes #29225 from tgravescs/SPARK-32287. Authored-by: Thomas Graves Signed-off-by: Dongjoon Hyun (cherry picked from commit e6ef27be52dcd14dc94384c2ada85861be44d843) Signed-off-by: Dongjoon Hyun --- .../test/scala/org/apache/spark/ExecutorAllocationManagerSuite.scala| 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/core/src/test/scala/org/apache/spark/ExecutorAllocationManagerSuite.scala b/core/src/test/scala/org/apache/spark/ExecutorAllocationManagerSuite.scala index 8d95849..0b19146 100644 --- a/core/src/test/scala/org/apache/spark/ExecutorAllocationManagerSuite.scala +++ b/core/src/test/scala/org/apache/spark/ExecutorAllocationManagerSuite.scala @@ -1142,7 +1142,7 @@ class ExecutorAllocationManagerSuite extends SparkFunSuite { .set(config.DYN_ALLOCATION_TESTING, true) // SPARK-22864: effectively disable the allocation schedule by setting the period to a // really long value. - .set(TEST_SCHEDULE_INTERVAL, 1L) + .set(TEST_SCHEDULE_INTERVAL, 3L) } private def createManager( - To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org For additional commands, e-mail: commits-h...@spark.apache.org
[spark] branch master updated (64a01c0 -> e6ef27b)
This is an automated email from the ASF dual-hosted git repository. dongjoon pushed a change to branch master in repository https://gitbox.apache.org/repos/asf/spark.git. from 64a01c0 [SPARK-32430][SQL] Extend SparkSessionExtensions to inject rules into AQE query stage preparation add e6ef27b [SPARK-32287][TESTS] Flaky Test: ExecutorAllocationManagerSuite.add executors default profile No new revisions were added by this update. Summary of changes: .../test/scala/org/apache/spark/ExecutorAllocationManagerSuite.scala| 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) - To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org For additional commands, e-mail: commits-h...@spark.apache.org
[spark] branch branch-3.0 updated: [SPARK-32287][TESTS] Flaky Test: ExecutorAllocationManagerSuite.add executors default profile
This is an automated email from the ASF dual-hosted git repository. dongjoon pushed a commit to branch branch-3.0 in repository https://gitbox.apache.org/repos/asf/spark.git The following commit(s) were added to refs/heads/branch-3.0 by this push: new ab14d05 [SPARK-32287][TESTS] Flaky Test: ExecutorAllocationManagerSuite.add executors default profile ab14d05 is described below commit ab14d05624f6cc88b02e29dd5ae8d302ffcfd09d Author: Thomas Graves AuthorDate: Fri Jul 24 11:12:28 2020 -0700 [SPARK-32287][TESTS] Flaky Test: ExecutorAllocationManagerSuite.add executors default profile I wasn't able to reproduce the failure but the best I can tell is that the allocation manager timer triggers and call doRequest. The timeout is 10s so try to increase that to 30seconds. test failure no unit test Closes #29225 from tgravescs/SPARK-32287. Authored-by: Thomas Graves Signed-off-by: Dongjoon Hyun (cherry picked from commit e6ef27be52dcd14dc94384c2ada85861be44d843) Signed-off-by: Dongjoon Hyun --- .../test/scala/org/apache/spark/ExecutorAllocationManagerSuite.scala| 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/core/src/test/scala/org/apache/spark/ExecutorAllocationManagerSuite.scala b/core/src/test/scala/org/apache/spark/ExecutorAllocationManagerSuite.scala index 8d95849..0b19146 100644 --- a/core/src/test/scala/org/apache/spark/ExecutorAllocationManagerSuite.scala +++ b/core/src/test/scala/org/apache/spark/ExecutorAllocationManagerSuite.scala @@ -1142,7 +1142,7 @@ class ExecutorAllocationManagerSuite extends SparkFunSuite { .set(config.DYN_ALLOCATION_TESTING, true) // SPARK-22864: effectively disable the allocation schedule by setting the period to a // really long value. - .set(TEST_SCHEDULE_INTERVAL, 1L) + .set(TEST_SCHEDULE_INTERVAL, 3L) } private def createManager( - To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org For additional commands, e-mail: commits-h...@spark.apache.org
[spark] branch master updated (64a01c0 -> e6ef27b)
This is an automated email from the ASF dual-hosted git repository. dongjoon pushed a change to branch master in repository https://gitbox.apache.org/repos/asf/spark.git. from 64a01c0 [SPARK-32430][SQL] Extend SparkSessionExtensions to inject rules into AQE query stage preparation add e6ef27b [SPARK-32287][TESTS] Flaky Test: ExecutorAllocationManagerSuite.add executors default profile No new revisions were added by this update. Summary of changes: .../test/scala/org/apache/spark/ExecutorAllocationManagerSuite.scala| 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) - To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org For additional commands, e-mail: commits-h...@spark.apache.org
[spark] branch branch-3.0 updated: [SPARK-32287][TESTS] Flaky Test: ExecutorAllocationManagerSuite.add executors default profile
This is an automated email from the ASF dual-hosted git repository. dongjoon pushed a commit to branch branch-3.0 in repository https://gitbox.apache.org/repos/asf/spark.git The following commit(s) were added to refs/heads/branch-3.0 by this push: new ab14d05 [SPARK-32287][TESTS] Flaky Test: ExecutorAllocationManagerSuite.add executors default profile ab14d05 is described below commit ab14d05624f6cc88b02e29dd5ae8d302ffcfd09d Author: Thomas Graves AuthorDate: Fri Jul 24 11:12:28 2020 -0700 [SPARK-32287][TESTS] Flaky Test: ExecutorAllocationManagerSuite.add executors default profile I wasn't able to reproduce the failure but the best I can tell is that the allocation manager timer triggers and call doRequest. The timeout is 10s so try to increase that to 30seconds. test failure no unit test Closes #29225 from tgravescs/SPARK-32287. Authored-by: Thomas Graves Signed-off-by: Dongjoon Hyun (cherry picked from commit e6ef27be52dcd14dc94384c2ada85861be44d843) Signed-off-by: Dongjoon Hyun --- .../test/scala/org/apache/spark/ExecutorAllocationManagerSuite.scala| 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/core/src/test/scala/org/apache/spark/ExecutorAllocationManagerSuite.scala b/core/src/test/scala/org/apache/spark/ExecutorAllocationManagerSuite.scala index 8d95849..0b19146 100644 --- a/core/src/test/scala/org/apache/spark/ExecutorAllocationManagerSuite.scala +++ b/core/src/test/scala/org/apache/spark/ExecutorAllocationManagerSuite.scala @@ -1142,7 +1142,7 @@ class ExecutorAllocationManagerSuite extends SparkFunSuite { .set(config.DYN_ALLOCATION_TESTING, true) // SPARK-22864: effectively disable the allocation schedule by setting the period to a // really long value. - .set(TEST_SCHEDULE_INTERVAL, 1L) + .set(TEST_SCHEDULE_INTERVAL, 3L) } private def createManager( - To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org For additional commands, e-mail: commits-h...@spark.apache.org
[spark] branch master updated (64a01c0 -> e6ef27b)
This is an automated email from the ASF dual-hosted git repository. dongjoon pushed a change to branch master in repository https://gitbox.apache.org/repos/asf/spark.git. from 64a01c0 [SPARK-32430][SQL] Extend SparkSessionExtensions to inject rules into AQE query stage preparation add e6ef27b [SPARK-32287][TESTS] Flaky Test: ExecutorAllocationManagerSuite.add executors default profile No new revisions were added by this update. Summary of changes: .../test/scala/org/apache/spark/ExecutorAllocationManagerSuite.scala| 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) - To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org For additional commands, e-mail: commits-h...@spark.apache.org
[spark] branch branch-3.0 updated: [SPARK-32287][TESTS] Flaky Test: ExecutorAllocationManagerSuite.add executors default profile
This is an automated email from the ASF dual-hosted git repository. dongjoon pushed a commit to branch branch-3.0 in repository https://gitbox.apache.org/repos/asf/spark.git The following commit(s) were added to refs/heads/branch-3.0 by this push: new ab14d05 [SPARK-32287][TESTS] Flaky Test: ExecutorAllocationManagerSuite.add executors default profile ab14d05 is described below commit ab14d05624f6cc88b02e29dd5ae8d302ffcfd09d Author: Thomas Graves AuthorDate: Fri Jul 24 11:12:28 2020 -0700 [SPARK-32287][TESTS] Flaky Test: ExecutorAllocationManagerSuite.add executors default profile I wasn't able to reproduce the failure but the best I can tell is that the allocation manager timer triggers and call doRequest. The timeout is 10s so try to increase that to 30seconds. test failure no unit test Closes #29225 from tgravescs/SPARK-32287. Authored-by: Thomas Graves Signed-off-by: Dongjoon Hyun (cherry picked from commit e6ef27be52dcd14dc94384c2ada85861be44d843) Signed-off-by: Dongjoon Hyun --- .../test/scala/org/apache/spark/ExecutorAllocationManagerSuite.scala| 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/core/src/test/scala/org/apache/spark/ExecutorAllocationManagerSuite.scala b/core/src/test/scala/org/apache/spark/ExecutorAllocationManagerSuite.scala index 8d95849..0b19146 100644 --- a/core/src/test/scala/org/apache/spark/ExecutorAllocationManagerSuite.scala +++ b/core/src/test/scala/org/apache/spark/ExecutorAllocationManagerSuite.scala @@ -1142,7 +1142,7 @@ class ExecutorAllocationManagerSuite extends SparkFunSuite { .set(config.DYN_ALLOCATION_TESTING, true) // SPARK-22864: effectively disable the allocation schedule by setting the period to a // really long value. - .set(TEST_SCHEDULE_INTERVAL, 1L) + .set(TEST_SCHEDULE_INTERVAL, 3L) } private def createManager( - To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org For additional commands, e-mail: commits-h...@spark.apache.org
[spark] branch master updated (64a01c0 -> e6ef27b)
This is an automated email from the ASF dual-hosted git repository. dongjoon pushed a change to branch master in repository https://gitbox.apache.org/repos/asf/spark.git. from 64a01c0 [SPARK-32430][SQL] Extend SparkSessionExtensions to inject rules into AQE query stage preparation add e6ef27b [SPARK-32287][TESTS] Flaky Test: ExecutorAllocationManagerSuite.add executors default profile No new revisions were added by this update. Summary of changes: .../test/scala/org/apache/spark/ExecutorAllocationManagerSuite.scala| 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) - To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org For additional commands, e-mail: commits-h...@spark.apache.org
[spark] branch branch-3.0 updated: [SPARK-32287][TESTS] Flaky Test: ExecutorAllocationManagerSuite.add executors default profile
This is an automated email from the ASF dual-hosted git repository. dongjoon pushed a commit to branch branch-3.0 in repository https://gitbox.apache.org/repos/asf/spark.git The following commit(s) were added to refs/heads/branch-3.0 by this push: new ab14d05 [SPARK-32287][TESTS] Flaky Test: ExecutorAllocationManagerSuite.add executors default profile ab14d05 is described below commit ab14d05624f6cc88b02e29dd5ae8d302ffcfd09d Author: Thomas Graves AuthorDate: Fri Jul 24 11:12:28 2020 -0700 [SPARK-32287][TESTS] Flaky Test: ExecutorAllocationManagerSuite.add executors default profile I wasn't able to reproduce the failure but the best I can tell is that the allocation manager timer triggers and call doRequest. The timeout is 10s so try to increase that to 30seconds. test failure no unit test Closes #29225 from tgravescs/SPARK-32287. Authored-by: Thomas Graves Signed-off-by: Dongjoon Hyun (cherry picked from commit e6ef27be52dcd14dc94384c2ada85861be44d843) Signed-off-by: Dongjoon Hyun --- .../test/scala/org/apache/spark/ExecutorAllocationManagerSuite.scala| 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/core/src/test/scala/org/apache/spark/ExecutorAllocationManagerSuite.scala b/core/src/test/scala/org/apache/spark/ExecutorAllocationManagerSuite.scala index 8d95849..0b19146 100644 --- a/core/src/test/scala/org/apache/spark/ExecutorAllocationManagerSuite.scala +++ b/core/src/test/scala/org/apache/spark/ExecutorAllocationManagerSuite.scala @@ -1142,7 +1142,7 @@ class ExecutorAllocationManagerSuite extends SparkFunSuite { .set(config.DYN_ALLOCATION_TESTING, true) // SPARK-22864: effectively disable the allocation schedule by setting the period to a // really long value. - .set(TEST_SCHEDULE_INTERVAL, 1L) + .set(TEST_SCHEDULE_INTERVAL, 3L) } private def createManager( - To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org For additional commands, e-mail: commits-h...@spark.apache.org
[spark] branch master updated (64a01c0 -> e6ef27b)
This is an automated email from the ASF dual-hosted git repository. dongjoon pushed a change to branch master in repository https://gitbox.apache.org/repos/asf/spark.git. from 64a01c0 [SPARK-32430][SQL] Extend SparkSessionExtensions to inject rules into AQE query stage preparation add e6ef27b [SPARK-32287][TESTS] Flaky Test: ExecutorAllocationManagerSuite.add executors default profile No new revisions were added by this update. Summary of changes: .../test/scala/org/apache/spark/ExecutorAllocationManagerSuite.scala| 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) - To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org For additional commands, e-mail: commits-h...@spark.apache.org
[spark] branch branch-3.0 updated: [SPARK-32430][SQL] Extend SparkSessionExtensions to inject rules into AQE query stage preparation
This is an automated email from the ASF dual-hosted git repository. dongjoon pushed a commit to branch branch-3.0 in repository https://gitbox.apache.org/repos/asf/spark.git The following commit(s) were added to refs/heads/branch-3.0 by this push: new 7004c98 [SPARK-32430][SQL] Extend SparkSessionExtensions to inject rules into AQE query stage preparation 7004c98 is described below commit 7004c989048b08891fb5f62ce2fcf0c89ce1496a Author: Andy Grove AuthorDate: Fri Jul 24 11:03:57 2020 -0700 [SPARK-32430][SQL] Extend SparkSessionExtensions to inject rules into AQE query stage preparation ### What changes were proposed in this pull request? Provide a generic mechanism for plugins to inject rules into the AQE "query prep" stage that happens before query stage creation. This goes along with https://issues.apache.org/jira/browse/SPARK-32332 where the current AQE implementation doesn't allow for users to properly extend it for columnar processing. ### Why are the changes needed? The issue here is that we create new query stages but we do not have access to the parent plan of the new query stage so certain things can not be determined because you have to know what the parent did. With this change it would allow you to add TAGs to be able to figure out what is going on. ### Does this PR introduce _any_ user-facing change? No. ### How was this patch tested? A new unit test is included in the PR. Closes #29224 from andygrove/insert-aqe-rule. Authored-by: Andy Grove Signed-off-by: Dongjoon Hyun (cherry picked from commit 64a01c0a559396fccd615dc00576a80bc8cc5648) Signed-off-by: Dongjoon Hyun --- .../apache/spark/sql/SparkSessionExtensions.scala | 20 - .../execution/adaptive/AdaptiveSparkPlanExec.scala | 2 +- .../sql/internal/BaseSessionStateBuilder.scala | 9 +++- .../apache/spark/sql/internal/SessionState.scala | 4 +- .../spark/sql/SparkSessionExtensionSuite.scala | 49 ++ 5 files changed, 79 insertions(+), 5 deletions(-) diff --git a/sql/core/src/main/scala/org/apache/spark/sql/SparkSessionExtensions.scala b/sql/core/src/main/scala/org/apache/spark/sql/SparkSessionExtensions.scala index 1c2bf9e..bd870fb 100644 --- a/sql/core/src/main/scala/org/apache/spark/sql/SparkSessionExtensions.scala +++ b/sql/core/src/main/scala/org/apache/spark/sql/SparkSessionExtensions.scala @@ -27,7 +27,7 @@ import org.apache.spark.sql.catalyst.expressions.ExpressionInfo import org.apache.spark.sql.catalyst.parser.ParserInterface import org.apache.spark.sql.catalyst.plans.logical.LogicalPlan import org.apache.spark.sql.catalyst.rules.Rule -import org.apache.spark.sql.execution.ColumnarRule +import org.apache.spark.sql.execution.{ColumnarRule, SparkPlan} /** * :: Experimental :: @@ -44,6 +44,7 @@ import org.apache.spark.sql.execution.ColumnarRule * Customized Parser. * (External) Catalog listeners. * Columnar Rules. + * Adaptive Query Stage Preparation Rules. * * * The extensions can be used by calling `withExtensions` on the [[SparkSession.Builder]], for @@ -96,8 +97,10 @@ class SparkSessionExtensions { type ParserBuilder = (SparkSession, ParserInterface) => ParserInterface type FunctionDescription = (FunctionIdentifier, ExpressionInfo, FunctionBuilder) type ColumnarRuleBuilder = SparkSession => ColumnarRule + type QueryStagePrepRuleBuilder = SparkSession => Rule[SparkPlan] private[this] val columnarRuleBuilders = mutable.Buffer.empty[ColumnarRuleBuilder] + private[this] val queryStagePrepRuleBuilders = mutable.Buffer.empty[QueryStagePrepRuleBuilder] /** * Build the override rules for columnar execution. @@ -107,12 +110,27 @@ class SparkSessionExtensions { } /** + * Build the override rules for the query stage preparation phase of adaptive query execution. + */ + private[sql] def buildQueryStagePrepRules(session: SparkSession): Seq[Rule[SparkPlan]] = { +queryStagePrepRuleBuilders.map(_.apply(session)).toSeq + } + + /** * Inject a rule that can override the columnar execution of an executor. */ def injectColumnar(builder: ColumnarRuleBuilder): Unit = { columnarRuleBuilders += builder } + /** + * Inject a rule that can override the the query stage preparation phase of adaptive query + * execution. + */ + def injectQueryStagePrepRule(builder: QueryStagePrepRuleBuilder): Unit = { +queryStagePrepRuleBuilders += builder + } + private[this] val resolutionRuleBuilders = mutable.Buffer.empty[RuleBuilder] /** diff --git a/sql/core/src/main/scala/org/apache/spark/sql/execution/adaptive/AdaptiveSparkPlanExec.scala b/sql/core/src/main/scala/org/apache/spark/sql/execution/adaptive/AdaptiveSparkPlanExec.scala index f6a..5714c33 100644 --- a/sql/core/src/main/scala/org/apache/spark/sql/execution/adaptive/AdaptiveSparkPlanExec.scala +++
[spark] branch branch-3.0 updated: [SPARK-32430][SQL] Extend SparkSessionExtensions to inject rules into AQE query stage preparation
This is an automated email from the ASF dual-hosted git repository. dongjoon pushed a commit to branch branch-3.0 in repository https://gitbox.apache.org/repos/asf/spark.git The following commit(s) were added to refs/heads/branch-3.0 by this push: new 7004c98 [SPARK-32430][SQL] Extend SparkSessionExtensions to inject rules into AQE query stage preparation 7004c98 is described below commit 7004c989048b08891fb5f62ce2fcf0c89ce1496a Author: Andy Grove AuthorDate: Fri Jul 24 11:03:57 2020 -0700 [SPARK-32430][SQL] Extend SparkSessionExtensions to inject rules into AQE query stage preparation ### What changes were proposed in this pull request? Provide a generic mechanism for plugins to inject rules into the AQE "query prep" stage that happens before query stage creation. This goes along with https://issues.apache.org/jira/browse/SPARK-32332 where the current AQE implementation doesn't allow for users to properly extend it for columnar processing. ### Why are the changes needed? The issue here is that we create new query stages but we do not have access to the parent plan of the new query stage so certain things can not be determined because you have to know what the parent did. With this change it would allow you to add TAGs to be able to figure out what is going on. ### Does this PR introduce _any_ user-facing change? No. ### How was this patch tested? A new unit test is included in the PR. Closes #29224 from andygrove/insert-aqe-rule. Authored-by: Andy Grove Signed-off-by: Dongjoon Hyun (cherry picked from commit 64a01c0a559396fccd615dc00576a80bc8cc5648) Signed-off-by: Dongjoon Hyun --- .../apache/spark/sql/SparkSessionExtensions.scala | 20 - .../execution/adaptive/AdaptiveSparkPlanExec.scala | 2 +- .../sql/internal/BaseSessionStateBuilder.scala | 9 +++- .../apache/spark/sql/internal/SessionState.scala | 4 +- .../spark/sql/SparkSessionExtensionSuite.scala | 49 ++ 5 files changed, 79 insertions(+), 5 deletions(-) diff --git a/sql/core/src/main/scala/org/apache/spark/sql/SparkSessionExtensions.scala b/sql/core/src/main/scala/org/apache/spark/sql/SparkSessionExtensions.scala index 1c2bf9e..bd870fb 100644 --- a/sql/core/src/main/scala/org/apache/spark/sql/SparkSessionExtensions.scala +++ b/sql/core/src/main/scala/org/apache/spark/sql/SparkSessionExtensions.scala @@ -27,7 +27,7 @@ import org.apache.spark.sql.catalyst.expressions.ExpressionInfo import org.apache.spark.sql.catalyst.parser.ParserInterface import org.apache.spark.sql.catalyst.plans.logical.LogicalPlan import org.apache.spark.sql.catalyst.rules.Rule -import org.apache.spark.sql.execution.ColumnarRule +import org.apache.spark.sql.execution.{ColumnarRule, SparkPlan} /** * :: Experimental :: @@ -44,6 +44,7 @@ import org.apache.spark.sql.execution.ColumnarRule * Customized Parser. * (External) Catalog listeners. * Columnar Rules. + * Adaptive Query Stage Preparation Rules. * * * The extensions can be used by calling `withExtensions` on the [[SparkSession.Builder]], for @@ -96,8 +97,10 @@ class SparkSessionExtensions { type ParserBuilder = (SparkSession, ParserInterface) => ParserInterface type FunctionDescription = (FunctionIdentifier, ExpressionInfo, FunctionBuilder) type ColumnarRuleBuilder = SparkSession => ColumnarRule + type QueryStagePrepRuleBuilder = SparkSession => Rule[SparkPlan] private[this] val columnarRuleBuilders = mutable.Buffer.empty[ColumnarRuleBuilder] + private[this] val queryStagePrepRuleBuilders = mutable.Buffer.empty[QueryStagePrepRuleBuilder] /** * Build the override rules for columnar execution. @@ -107,12 +110,27 @@ class SparkSessionExtensions { } /** + * Build the override rules for the query stage preparation phase of adaptive query execution. + */ + private[sql] def buildQueryStagePrepRules(session: SparkSession): Seq[Rule[SparkPlan]] = { +queryStagePrepRuleBuilders.map(_.apply(session)).toSeq + } + + /** * Inject a rule that can override the columnar execution of an executor. */ def injectColumnar(builder: ColumnarRuleBuilder): Unit = { columnarRuleBuilders += builder } + /** + * Inject a rule that can override the the query stage preparation phase of adaptive query + * execution. + */ + def injectQueryStagePrepRule(builder: QueryStagePrepRuleBuilder): Unit = { +queryStagePrepRuleBuilders += builder + } + private[this] val resolutionRuleBuilders = mutable.Buffer.empty[RuleBuilder] /** diff --git a/sql/core/src/main/scala/org/apache/spark/sql/execution/adaptive/AdaptiveSparkPlanExec.scala b/sql/core/src/main/scala/org/apache/spark/sql/execution/adaptive/AdaptiveSparkPlanExec.scala index f6a..5714c33 100644 --- a/sql/core/src/main/scala/org/apache/spark/sql/execution/adaptive/AdaptiveSparkPlanExec.scala +++
[spark] branch master updated (d3596c0 -> 64a01c0)
This is an automated email from the ASF dual-hosted git repository. dongjoon pushed a change to branch master in repository https://gitbox.apache.org/repos/asf/spark.git. from d3596c0 [SPARK-32406][SQL] Make RESET syntax support single configuration reset add 64a01c0 [SPARK-32430][SQL] Extend SparkSessionExtensions to inject rules into AQE query stage preparation No new revisions were added by this update. Summary of changes: .../apache/spark/sql/SparkSessionExtensions.scala | 20 - .../execution/adaptive/AdaptiveSparkPlanExec.scala | 2 +- .../sql/internal/BaseSessionStateBuilder.scala | 9 +++- .../apache/spark/sql/internal/SessionState.scala | 4 +- .../spark/sql/SparkSessionExtensionSuite.scala | 49 ++ 5 files changed, 79 insertions(+), 5 deletions(-) - To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org For additional commands, e-mail: commits-h...@spark.apache.org
[spark] branch branch-3.0 updated: [SPARK-32430][SQL] Extend SparkSessionExtensions to inject rules into AQE query stage preparation
This is an automated email from the ASF dual-hosted git repository. dongjoon pushed a commit to branch branch-3.0 in repository https://gitbox.apache.org/repos/asf/spark.git The following commit(s) were added to refs/heads/branch-3.0 by this push: new 7004c98 [SPARK-32430][SQL] Extend SparkSessionExtensions to inject rules into AQE query stage preparation 7004c98 is described below commit 7004c989048b08891fb5f62ce2fcf0c89ce1496a Author: Andy Grove AuthorDate: Fri Jul 24 11:03:57 2020 -0700 [SPARK-32430][SQL] Extend SparkSessionExtensions to inject rules into AQE query stage preparation ### What changes were proposed in this pull request? Provide a generic mechanism for plugins to inject rules into the AQE "query prep" stage that happens before query stage creation. This goes along with https://issues.apache.org/jira/browse/SPARK-32332 where the current AQE implementation doesn't allow for users to properly extend it for columnar processing. ### Why are the changes needed? The issue here is that we create new query stages but we do not have access to the parent plan of the new query stage so certain things can not be determined because you have to know what the parent did. With this change it would allow you to add TAGs to be able to figure out what is going on. ### Does this PR introduce _any_ user-facing change? No. ### How was this patch tested? A new unit test is included in the PR. Closes #29224 from andygrove/insert-aqe-rule. Authored-by: Andy Grove Signed-off-by: Dongjoon Hyun (cherry picked from commit 64a01c0a559396fccd615dc00576a80bc8cc5648) Signed-off-by: Dongjoon Hyun --- .../apache/spark/sql/SparkSessionExtensions.scala | 20 - .../execution/adaptive/AdaptiveSparkPlanExec.scala | 2 +- .../sql/internal/BaseSessionStateBuilder.scala | 9 +++- .../apache/spark/sql/internal/SessionState.scala | 4 +- .../spark/sql/SparkSessionExtensionSuite.scala | 49 ++ 5 files changed, 79 insertions(+), 5 deletions(-) diff --git a/sql/core/src/main/scala/org/apache/spark/sql/SparkSessionExtensions.scala b/sql/core/src/main/scala/org/apache/spark/sql/SparkSessionExtensions.scala index 1c2bf9e..bd870fb 100644 --- a/sql/core/src/main/scala/org/apache/spark/sql/SparkSessionExtensions.scala +++ b/sql/core/src/main/scala/org/apache/spark/sql/SparkSessionExtensions.scala @@ -27,7 +27,7 @@ import org.apache.spark.sql.catalyst.expressions.ExpressionInfo import org.apache.spark.sql.catalyst.parser.ParserInterface import org.apache.spark.sql.catalyst.plans.logical.LogicalPlan import org.apache.spark.sql.catalyst.rules.Rule -import org.apache.spark.sql.execution.ColumnarRule +import org.apache.spark.sql.execution.{ColumnarRule, SparkPlan} /** * :: Experimental :: @@ -44,6 +44,7 @@ import org.apache.spark.sql.execution.ColumnarRule * Customized Parser. * (External) Catalog listeners. * Columnar Rules. + * Adaptive Query Stage Preparation Rules. * * * The extensions can be used by calling `withExtensions` on the [[SparkSession.Builder]], for @@ -96,8 +97,10 @@ class SparkSessionExtensions { type ParserBuilder = (SparkSession, ParserInterface) => ParserInterface type FunctionDescription = (FunctionIdentifier, ExpressionInfo, FunctionBuilder) type ColumnarRuleBuilder = SparkSession => ColumnarRule + type QueryStagePrepRuleBuilder = SparkSession => Rule[SparkPlan] private[this] val columnarRuleBuilders = mutable.Buffer.empty[ColumnarRuleBuilder] + private[this] val queryStagePrepRuleBuilders = mutable.Buffer.empty[QueryStagePrepRuleBuilder] /** * Build the override rules for columnar execution. @@ -107,12 +110,27 @@ class SparkSessionExtensions { } /** + * Build the override rules for the query stage preparation phase of adaptive query execution. + */ + private[sql] def buildQueryStagePrepRules(session: SparkSession): Seq[Rule[SparkPlan]] = { +queryStagePrepRuleBuilders.map(_.apply(session)).toSeq + } + + /** * Inject a rule that can override the columnar execution of an executor. */ def injectColumnar(builder: ColumnarRuleBuilder): Unit = { columnarRuleBuilders += builder } + /** + * Inject a rule that can override the the query stage preparation phase of adaptive query + * execution. + */ + def injectQueryStagePrepRule(builder: QueryStagePrepRuleBuilder): Unit = { +queryStagePrepRuleBuilders += builder + } + private[this] val resolutionRuleBuilders = mutable.Buffer.empty[RuleBuilder] /** diff --git a/sql/core/src/main/scala/org/apache/spark/sql/execution/adaptive/AdaptiveSparkPlanExec.scala b/sql/core/src/main/scala/org/apache/spark/sql/execution/adaptive/AdaptiveSparkPlanExec.scala index f6a..5714c33 100644 --- a/sql/core/src/main/scala/org/apache/spark/sql/execution/adaptive/AdaptiveSparkPlanExec.scala +++
[spark] branch master updated (d3596c0 -> 64a01c0)
This is an automated email from the ASF dual-hosted git repository. dongjoon pushed a change to branch master in repository https://gitbox.apache.org/repos/asf/spark.git. from d3596c0 [SPARK-32406][SQL] Make RESET syntax support single configuration reset add 64a01c0 [SPARK-32430][SQL] Extend SparkSessionExtensions to inject rules into AQE query stage preparation No new revisions were added by this update. Summary of changes: .../apache/spark/sql/SparkSessionExtensions.scala | 20 - .../execution/adaptive/AdaptiveSparkPlanExec.scala | 2 +- .../sql/internal/BaseSessionStateBuilder.scala | 9 +++- .../apache/spark/sql/internal/SessionState.scala | 4 +- .../spark/sql/SparkSessionExtensionSuite.scala | 49 ++ 5 files changed, 79 insertions(+), 5 deletions(-) - To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org For additional commands, e-mail: commits-h...@spark.apache.org
[spark] branch branch-3.0 updated: [SPARK-32430][SQL] Extend SparkSessionExtensions to inject rules into AQE query stage preparation
This is an automated email from the ASF dual-hosted git repository. dongjoon pushed a commit to branch branch-3.0 in repository https://gitbox.apache.org/repos/asf/spark.git The following commit(s) were added to refs/heads/branch-3.0 by this push: new 7004c98 [SPARK-32430][SQL] Extend SparkSessionExtensions to inject rules into AQE query stage preparation 7004c98 is described below commit 7004c989048b08891fb5f62ce2fcf0c89ce1496a Author: Andy Grove AuthorDate: Fri Jul 24 11:03:57 2020 -0700 [SPARK-32430][SQL] Extend SparkSessionExtensions to inject rules into AQE query stage preparation ### What changes were proposed in this pull request? Provide a generic mechanism for plugins to inject rules into the AQE "query prep" stage that happens before query stage creation. This goes along with https://issues.apache.org/jira/browse/SPARK-32332 where the current AQE implementation doesn't allow for users to properly extend it for columnar processing. ### Why are the changes needed? The issue here is that we create new query stages but we do not have access to the parent plan of the new query stage so certain things can not be determined because you have to know what the parent did. With this change it would allow you to add TAGs to be able to figure out what is going on. ### Does this PR introduce _any_ user-facing change? No. ### How was this patch tested? A new unit test is included in the PR. Closes #29224 from andygrove/insert-aqe-rule. Authored-by: Andy Grove Signed-off-by: Dongjoon Hyun (cherry picked from commit 64a01c0a559396fccd615dc00576a80bc8cc5648) Signed-off-by: Dongjoon Hyun --- .../apache/spark/sql/SparkSessionExtensions.scala | 20 - .../execution/adaptive/AdaptiveSparkPlanExec.scala | 2 +- .../sql/internal/BaseSessionStateBuilder.scala | 9 +++- .../apache/spark/sql/internal/SessionState.scala | 4 +- .../spark/sql/SparkSessionExtensionSuite.scala | 49 ++ 5 files changed, 79 insertions(+), 5 deletions(-) diff --git a/sql/core/src/main/scala/org/apache/spark/sql/SparkSessionExtensions.scala b/sql/core/src/main/scala/org/apache/spark/sql/SparkSessionExtensions.scala index 1c2bf9e..bd870fb 100644 --- a/sql/core/src/main/scala/org/apache/spark/sql/SparkSessionExtensions.scala +++ b/sql/core/src/main/scala/org/apache/spark/sql/SparkSessionExtensions.scala @@ -27,7 +27,7 @@ import org.apache.spark.sql.catalyst.expressions.ExpressionInfo import org.apache.spark.sql.catalyst.parser.ParserInterface import org.apache.spark.sql.catalyst.plans.logical.LogicalPlan import org.apache.spark.sql.catalyst.rules.Rule -import org.apache.spark.sql.execution.ColumnarRule +import org.apache.spark.sql.execution.{ColumnarRule, SparkPlan} /** * :: Experimental :: @@ -44,6 +44,7 @@ import org.apache.spark.sql.execution.ColumnarRule * Customized Parser. * (External) Catalog listeners. * Columnar Rules. + * Adaptive Query Stage Preparation Rules. * * * The extensions can be used by calling `withExtensions` on the [[SparkSession.Builder]], for @@ -96,8 +97,10 @@ class SparkSessionExtensions { type ParserBuilder = (SparkSession, ParserInterface) => ParserInterface type FunctionDescription = (FunctionIdentifier, ExpressionInfo, FunctionBuilder) type ColumnarRuleBuilder = SparkSession => ColumnarRule + type QueryStagePrepRuleBuilder = SparkSession => Rule[SparkPlan] private[this] val columnarRuleBuilders = mutable.Buffer.empty[ColumnarRuleBuilder] + private[this] val queryStagePrepRuleBuilders = mutable.Buffer.empty[QueryStagePrepRuleBuilder] /** * Build the override rules for columnar execution. @@ -107,12 +110,27 @@ class SparkSessionExtensions { } /** + * Build the override rules for the query stage preparation phase of adaptive query execution. + */ + private[sql] def buildQueryStagePrepRules(session: SparkSession): Seq[Rule[SparkPlan]] = { +queryStagePrepRuleBuilders.map(_.apply(session)).toSeq + } + + /** * Inject a rule that can override the columnar execution of an executor. */ def injectColumnar(builder: ColumnarRuleBuilder): Unit = { columnarRuleBuilders += builder } + /** + * Inject a rule that can override the the query stage preparation phase of adaptive query + * execution. + */ + def injectQueryStagePrepRule(builder: QueryStagePrepRuleBuilder): Unit = { +queryStagePrepRuleBuilders += builder + } + private[this] val resolutionRuleBuilders = mutable.Buffer.empty[RuleBuilder] /** diff --git a/sql/core/src/main/scala/org/apache/spark/sql/execution/adaptive/AdaptiveSparkPlanExec.scala b/sql/core/src/main/scala/org/apache/spark/sql/execution/adaptive/AdaptiveSparkPlanExec.scala index f6a..5714c33 100644 --- a/sql/core/src/main/scala/org/apache/spark/sql/execution/adaptive/AdaptiveSparkPlanExec.scala +++
[spark] branch master updated (d3596c0 -> 64a01c0)
This is an automated email from the ASF dual-hosted git repository. dongjoon pushed a change to branch master in repository https://gitbox.apache.org/repos/asf/spark.git. from d3596c0 [SPARK-32406][SQL] Make RESET syntax support single configuration reset add 64a01c0 [SPARK-32430][SQL] Extend SparkSessionExtensions to inject rules into AQE query stage preparation No new revisions were added by this update. Summary of changes: .../apache/spark/sql/SparkSessionExtensions.scala | 20 - .../execution/adaptive/AdaptiveSparkPlanExec.scala | 2 +- .../sql/internal/BaseSessionStateBuilder.scala | 9 +++- .../apache/spark/sql/internal/SessionState.scala | 4 +- .../spark/sql/SparkSessionExtensionSuite.scala | 49 ++ 5 files changed, 79 insertions(+), 5 deletions(-) - To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org For additional commands, e-mail: commits-h...@spark.apache.org
[spark] branch branch-3.0 updated: [SPARK-32430][SQL] Extend SparkSessionExtensions to inject rules into AQE query stage preparation
This is an automated email from the ASF dual-hosted git repository. dongjoon pushed a commit to branch branch-3.0 in repository https://gitbox.apache.org/repos/asf/spark.git The following commit(s) were added to refs/heads/branch-3.0 by this push: new 7004c98 [SPARK-32430][SQL] Extend SparkSessionExtensions to inject rules into AQE query stage preparation 7004c98 is described below commit 7004c989048b08891fb5f62ce2fcf0c89ce1496a Author: Andy Grove AuthorDate: Fri Jul 24 11:03:57 2020 -0700 [SPARK-32430][SQL] Extend SparkSessionExtensions to inject rules into AQE query stage preparation ### What changes were proposed in this pull request? Provide a generic mechanism for plugins to inject rules into the AQE "query prep" stage that happens before query stage creation. This goes along with https://issues.apache.org/jira/browse/SPARK-32332 where the current AQE implementation doesn't allow for users to properly extend it for columnar processing. ### Why are the changes needed? The issue here is that we create new query stages but we do not have access to the parent plan of the new query stage so certain things can not be determined because you have to know what the parent did. With this change it would allow you to add TAGs to be able to figure out what is going on. ### Does this PR introduce _any_ user-facing change? No. ### How was this patch tested? A new unit test is included in the PR. Closes #29224 from andygrove/insert-aqe-rule. Authored-by: Andy Grove Signed-off-by: Dongjoon Hyun (cherry picked from commit 64a01c0a559396fccd615dc00576a80bc8cc5648) Signed-off-by: Dongjoon Hyun --- .../apache/spark/sql/SparkSessionExtensions.scala | 20 - .../execution/adaptive/AdaptiveSparkPlanExec.scala | 2 +- .../sql/internal/BaseSessionStateBuilder.scala | 9 +++- .../apache/spark/sql/internal/SessionState.scala | 4 +- .../spark/sql/SparkSessionExtensionSuite.scala | 49 ++ 5 files changed, 79 insertions(+), 5 deletions(-) diff --git a/sql/core/src/main/scala/org/apache/spark/sql/SparkSessionExtensions.scala b/sql/core/src/main/scala/org/apache/spark/sql/SparkSessionExtensions.scala index 1c2bf9e..bd870fb 100644 --- a/sql/core/src/main/scala/org/apache/spark/sql/SparkSessionExtensions.scala +++ b/sql/core/src/main/scala/org/apache/spark/sql/SparkSessionExtensions.scala @@ -27,7 +27,7 @@ import org.apache.spark.sql.catalyst.expressions.ExpressionInfo import org.apache.spark.sql.catalyst.parser.ParserInterface import org.apache.spark.sql.catalyst.plans.logical.LogicalPlan import org.apache.spark.sql.catalyst.rules.Rule -import org.apache.spark.sql.execution.ColumnarRule +import org.apache.spark.sql.execution.{ColumnarRule, SparkPlan} /** * :: Experimental :: @@ -44,6 +44,7 @@ import org.apache.spark.sql.execution.ColumnarRule * Customized Parser. * (External) Catalog listeners. * Columnar Rules. + * Adaptive Query Stage Preparation Rules. * * * The extensions can be used by calling `withExtensions` on the [[SparkSession.Builder]], for @@ -96,8 +97,10 @@ class SparkSessionExtensions { type ParserBuilder = (SparkSession, ParserInterface) => ParserInterface type FunctionDescription = (FunctionIdentifier, ExpressionInfo, FunctionBuilder) type ColumnarRuleBuilder = SparkSession => ColumnarRule + type QueryStagePrepRuleBuilder = SparkSession => Rule[SparkPlan] private[this] val columnarRuleBuilders = mutable.Buffer.empty[ColumnarRuleBuilder] + private[this] val queryStagePrepRuleBuilders = mutable.Buffer.empty[QueryStagePrepRuleBuilder] /** * Build the override rules for columnar execution. @@ -107,12 +110,27 @@ class SparkSessionExtensions { } /** + * Build the override rules for the query stage preparation phase of adaptive query execution. + */ + private[sql] def buildQueryStagePrepRules(session: SparkSession): Seq[Rule[SparkPlan]] = { +queryStagePrepRuleBuilders.map(_.apply(session)).toSeq + } + + /** * Inject a rule that can override the columnar execution of an executor. */ def injectColumnar(builder: ColumnarRuleBuilder): Unit = { columnarRuleBuilders += builder } + /** + * Inject a rule that can override the the query stage preparation phase of adaptive query + * execution. + */ + def injectQueryStagePrepRule(builder: QueryStagePrepRuleBuilder): Unit = { +queryStagePrepRuleBuilders += builder + } + private[this] val resolutionRuleBuilders = mutable.Buffer.empty[RuleBuilder] /** diff --git a/sql/core/src/main/scala/org/apache/spark/sql/execution/adaptive/AdaptiveSparkPlanExec.scala b/sql/core/src/main/scala/org/apache/spark/sql/execution/adaptive/AdaptiveSparkPlanExec.scala index f6a..5714c33 100644 --- a/sql/core/src/main/scala/org/apache/spark/sql/execution/adaptive/AdaptiveSparkPlanExec.scala +++
[spark] branch master updated (d3596c0 -> 64a01c0)
This is an automated email from the ASF dual-hosted git repository. dongjoon pushed a change to branch master in repository https://gitbox.apache.org/repos/asf/spark.git. from d3596c0 [SPARK-32406][SQL] Make RESET syntax support single configuration reset add 64a01c0 [SPARK-32430][SQL] Extend SparkSessionExtensions to inject rules into AQE query stage preparation No new revisions were added by this update. Summary of changes: .../apache/spark/sql/SparkSessionExtensions.scala | 20 - .../execution/adaptive/AdaptiveSparkPlanExec.scala | 2 +- .../sql/internal/BaseSessionStateBuilder.scala | 9 +++- .../apache/spark/sql/internal/SessionState.scala | 4 +- .../spark/sql/SparkSessionExtensionSuite.scala | 49 ++ 5 files changed, 79 insertions(+), 5 deletions(-) - To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org For additional commands, e-mail: commits-h...@spark.apache.org
[spark] branch master updated (d3596c0 -> 64a01c0)
This is an automated email from the ASF dual-hosted git repository. dongjoon pushed a change to branch master in repository https://gitbox.apache.org/repos/asf/spark.git. from d3596c0 [SPARK-32406][SQL] Make RESET syntax support single configuration reset add 64a01c0 [SPARK-32430][SQL] Extend SparkSessionExtensions to inject rules into AQE query stage preparation No new revisions were added by this update. Summary of changes: .../apache/spark/sql/SparkSessionExtensions.scala | 20 - .../execution/adaptive/AdaptiveSparkPlanExec.scala | 2 +- .../sql/internal/BaseSessionStateBuilder.scala | 9 +++- .../apache/spark/sql/internal/SessionState.scala | 4 +- .../spark/sql/SparkSessionExtensionSuite.scala | 49 ++ 5 files changed, 79 insertions(+), 5 deletions(-) - To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org For additional commands, e-mail: commits-h...@spark.apache.org
[spark] branch branch-3.0 updated (f50432f -> 8a52bda)
This is an automated email from the ASF dual-hosted git repository. huaxingao pushed a change to branch branch-3.0 in repository https://gitbox.apache.org/repos/asf/spark.git. from f50432f [SPARK-32363][PYTHON][BUILD][3.0] Fix flakiness in pip package testing in Jenkins add 8a52bda [SPARK-32310][ML][PYSPARK][3.0] ML params default value parity No new revisions were added by this update. Summary of changes: .../spark/ml/classification/FMClassifier.scala | 10 -- .../apache/spark/ml/classification/LinearSVC.scala | 11 +-- .../ml/classification/LogisticRegression.scala | 13 +-- .../spark/ml/classification/NaiveBayes.scala | 4 +- .../spark/ml/clustering/BisectingKMeans.scala | 7 +- .../spark/ml/clustering/GaussianMixture.scala | 7 +- .../org/apache/spark/ml/clustering/KMeans.scala| 11 +-- .../scala/org/apache/spark/ml/clustering/LDA.scala | 11 +-- .../ml/clustering/PowerIterationClustering.scala | 7 +- .../evaluation/BinaryClassificationEvaluator.scala | 4 +- .../MulticlassClassificationEvaluator.scala| 8 +- .../MultilabelClassificationEvaluator.scala| 6 +- .../spark/ml/evaluation/RankingEvaluator.scala | 6 +- .../spark/ml/evaluation/RegressionEvaluator.scala | 4 +- .../apache/spark/ml/feature/ChiSqSelector.scala| 9 +- .../org/apache/spark/ml/feature/Imputer.scala | 4 +- .../org/apache/spark/ml/feature/MinMaxScaler.scala | 4 +- .../apache/spark/ml/feature/OneHotEncoder.scala| 5 +- .../spark/ml/feature/QuantileDiscretizer.scala | 4 +- .../org/apache/spark/ml/feature/RFormula.scala | 6 +- .../org/apache/spark/ml/feature/RobustScaler.scala | 8 +- .../apache/spark/ml/feature/StringIndexer.scala| 6 +- .../apache/spark/ml/feature/VectorIndexer.scala| 6 +- .../org/apache/spark/ml/feature/VectorSlicer.scala | 6 +- .../org/apache/spark/ml/feature/Word2Vec.scala | 9 +- .../scala/org/apache/spark/ml/fpm/FPGrowth.scala | 5 +- .../ml/regression/AFTSurvivalRegression.scala | 10 +- .../spark/ml/regression/LinearRegression.scala | 14 +-- .../org/apache/spark/ml/tree/treeParams.scala | 16 +-- .../spark/ml/util/DefaultReadWriteTest.scala | 3 + python/pyspark/ml/classification.py| 86 +++- python/pyspark/ml/clustering.py| 43 ++-- python/pyspark/ml/feature.py | 110 ++--- python/pyspark/ml/fpm.py | 12 ++- python/pyspark/ml/recommendation.py| 20 ++-- python/pyspark/ml/regression.py| 88 - python/pyspark/ml/tests/test_param.py | 7 +- python/pyspark/ml/tuning.py| 16 ++- 38 files changed, 368 insertions(+), 238 deletions(-) - To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org For additional commands, e-mail: commits-h...@spark.apache.org
[spark] branch branch-3.0 updated (f50432f -> 8a52bda)
This is an automated email from the ASF dual-hosted git repository. huaxingao pushed a change to branch branch-3.0 in repository https://gitbox.apache.org/repos/asf/spark.git. from f50432f [SPARK-32363][PYTHON][BUILD][3.0] Fix flakiness in pip package testing in Jenkins add 8a52bda [SPARK-32310][ML][PYSPARK][3.0] ML params default value parity No new revisions were added by this update. Summary of changes: .../spark/ml/classification/FMClassifier.scala | 10 -- .../apache/spark/ml/classification/LinearSVC.scala | 11 +-- .../ml/classification/LogisticRegression.scala | 13 +-- .../spark/ml/classification/NaiveBayes.scala | 4 +- .../spark/ml/clustering/BisectingKMeans.scala | 7 +- .../spark/ml/clustering/GaussianMixture.scala | 7 +- .../org/apache/spark/ml/clustering/KMeans.scala| 11 +-- .../scala/org/apache/spark/ml/clustering/LDA.scala | 11 +-- .../ml/clustering/PowerIterationClustering.scala | 7 +- .../evaluation/BinaryClassificationEvaluator.scala | 4 +- .../MulticlassClassificationEvaluator.scala| 8 +- .../MultilabelClassificationEvaluator.scala| 6 +- .../spark/ml/evaluation/RankingEvaluator.scala | 6 +- .../spark/ml/evaluation/RegressionEvaluator.scala | 4 +- .../apache/spark/ml/feature/ChiSqSelector.scala| 9 +- .../org/apache/spark/ml/feature/Imputer.scala | 4 +- .../org/apache/spark/ml/feature/MinMaxScaler.scala | 4 +- .../apache/spark/ml/feature/OneHotEncoder.scala| 5 +- .../spark/ml/feature/QuantileDiscretizer.scala | 4 +- .../org/apache/spark/ml/feature/RFormula.scala | 6 +- .../org/apache/spark/ml/feature/RobustScaler.scala | 8 +- .../apache/spark/ml/feature/StringIndexer.scala| 6 +- .../apache/spark/ml/feature/VectorIndexer.scala| 6 +- .../org/apache/spark/ml/feature/VectorSlicer.scala | 6 +- .../org/apache/spark/ml/feature/Word2Vec.scala | 9 +- .../scala/org/apache/spark/ml/fpm/FPGrowth.scala | 5 +- .../ml/regression/AFTSurvivalRegression.scala | 10 +- .../spark/ml/regression/LinearRegression.scala | 14 +-- .../org/apache/spark/ml/tree/treeParams.scala | 16 +-- .../spark/ml/util/DefaultReadWriteTest.scala | 3 + python/pyspark/ml/classification.py| 86 +++- python/pyspark/ml/clustering.py| 43 ++-- python/pyspark/ml/feature.py | 110 ++--- python/pyspark/ml/fpm.py | 12 ++- python/pyspark/ml/recommendation.py| 20 ++-- python/pyspark/ml/regression.py| 88 - python/pyspark/ml/tests/test_param.py | 7 +- python/pyspark/ml/tuning.py| 16 ++- 38 files changed, 368 insertions(+), 238 deletions(-) - To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org For additional commands, e-mail: commits-h...@spark.apache.org
[spark] branch branch-3.0 updated (f50432f -> 8a52bda)
This is an automated email from the ASF dual-hosted git repository. huaxingao pushed a change to branch branch-3.0 in repository https://gitbox.apache.org/repos/asf/spark.git. from f50432f [SPARK-32363][PYTHON][BUILD][3.0] Fix flakiness in pip package testing in Jenkins add 8a52bda [SPARK-32310][ML][PYSPARK][3.0] ML params default value parity No new revisions were added by this update. Summary of changes: .../spark/ml/classification/FMClassifier.scala | 10 -- .../apache/spark/ml/classification/LinearSVC.scala | 11 +-- .../ml/classification/LogisticRegression.scala | 13 +-- .../spark/ml/classification/NaiveBayes.scala | 4 +- .../spark/ml/clustering/BisectingKMeans.scala | 7 +- .../spark/ml/clustering/GaussianMixture.scala | 7 +- .../org/apache/spark/ml/clustering/KMeans.scala| 11 +-- .../scala/org/apache/spark/ml/clustering/LDA.scala | 11 +-- .../ml/clustering/PowerIterationClustering.scala | 7 +- .../evaluation/BinaryClassificationEvaluator.scala | 4 +- .../MulticlassClassificationEvaluator.scala| 8 +- .../MultilabelClassificationEvaluator.scala| 6 +- .../spark/ml/evaluation/RankingEvaluator.scala | 6 +- .../spark/ml/evaluation/RegressionEvaluator.scala | 4 +- .../apache/spark/ml/feature/ChiSqSelector.scala| 9 +- .../org/apache/spark/ml/feature/Imputer.scala | 4 +- .../org/apache/spark/ml/feature/MinMaxScaler.scala | 4 +- .../apache/spark/ml/feature/OneHotEncoder.scala| 5 +- .../spark/ml/feature/QuantileDiscretizer.scala | 4 +- .../org/apache/spark/ml/feature/RFormula.scala | 6 +- .../org/apache/spark/ml/feature/RobustScaler.scala | 8 +- .../apache/spark/ml/feature/StringIndexer.scala| 6 +- .../apache/spark/ml/feature/VectorIndexer.scala| 6 +- .../org/apache/spark/ml/feature/VectorSlicer.scala | 6 +- .../org/apache/spark/ml/feature/Word2Vec.scala | 9 +- .../scala/org/apache/spark/ml/fpm/FPGrowth.scala | 5 +- .../ml/regression/AFTSurvivalRegression.scala | 10 +- .../spark/ml/regression/LinearRegression.scala | 14 +-- .../org/apache/spark/ml/tree/treeParams.scala | 16 +-- .../spark/ml/util/DefaultReadWriteTest.scala | 3 + python/pyspark/ml/classification.py| 86 +++- python/pyspark/ml/clustering.py| 43 ++-- python/pyspark/ml/feature.py | 110 ++--- python/pyspark/ml/fpm.py | 12 ++- python/pyspark/ml/recommendation.py| 20 ++-- python/pyspark/ml/regression.py| 88 - python/pyspark/ml/tests/test_param.py | 7 +- python/pyspark/ml/tuning.py| 16 ++- 38 files changed, 368 insertions(+), 238 deletions(-) - To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org For additional commands, e-mail: commits-h...@spark.apache.org
[spark] branch branch-3.0 updated (f50432f -> 8a52bda)
This is an automated email from the ASF dual-hosted git repository. huaxingao pushed a change to branch branch-3.0 in repository https://gitbox.apache.org/repos/asf/spark.git. from f50432f [SPARK-32363][PYTHON][BUILD][3.0] Fix flakiness in pip package testing in Jenkins add 8a52bda [SPARK-32310][ML][PYSPARK][3.0] ML params default value parity No new revisions were added by this update. Summary of changes: .../spark/ml/classification/FMClassifier.scala | 10 -- .../apache/spark/ml/classification/LinearSVC.scala | 11 +-- .../ml/classification/LogisticRegression.scala | 13 +-- .../spark/ml/classification/NaiveBayes.scala | 4 +- .../spark/ml/clustering/BisectingKMeans.scala | 7 +- .../spark/ml/clustering/GaussianMixture.scala | 7 +- .../org/apache/spark/ml/clustering/KMeans.scala| 11 +-- .../scala/org/apache/spark/ml/clustering/LDA.scala | 11 +-- .../ml/clustering/PowerIterationClustering.scala | 7 +- .../evaluation/BinaryClassificationEvaluator.scala | 4 +- .../MulticlassClassificationEvaluator.scala| 8 +- .../MultilabelClassificationEvaluator.scala| 6 +- .../spark/ml/evaluation/RankingEvaluator.scala | 6 +- .../spark/ml/evaluation/RegressionEvaluator.scala | 4 +- .../apache/spark/ml/feature/ChiSqSelector.scala| 9 +- .../org/apache/spark/ml/feature/Imputer.scala | 4 +- .../org/apache/spark/ml/feature/MinMaxScaler.scala | 4 +- .../apache/spark/ml/feature/OneHotEncoder.scala| 5 +- .../spark/ml/feature/QuantileDiscretizer.scala | 4 +- .../org/apache/spark/ml/feature/RFormula.scala | 6 +- .../org/apache/spark/ml/feature/RobustScaler.scala | 8 +- .../apache/spark/ml/feature/StringIndexer.scala| 6 +- .../apache/spark/ml/feature/VectorIndexer.scala| 6 +- .../org/apache/spark/ml/feature/VectorSlicer.scala | 6 +- .../org/apache/spark/ml/feature/Word2Vec.scala | 9 +- .../scala/org/apache/spark/ml/fpm/FPGrowth.scala | 5 +- .../ml/regression/AFTSurvivalRegression.scala | 10 +- .../spark/ml/regression/LinearRegression.scala | 14 +-- .../org/apache/spark/ml/tree/treeParams.scala | 16 +-- .../spark/ml/util/DefaultReadWriteTest.scala | 3 + python/pyspark/ml/classification.py| 86 +++- python/pyspark/ml/clustering.py| 43 ++-- python/pyspark/ml/feature.py | 110 ++--- python/pyspark/ml/fpm.py | 12 ++- python/pyspark/ml/recommendation.py| 20 ++-- python/pyspark/ml/regression.py| 88 - python/pyspark/ml/tests/test_param.py | 7 +- python/pyspark/ml/tuning.py| 16 ++- 38 files changed, 368 insertions(+), 238 deletions(-) - To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org For additional commands, e-mail: commits-h...@spark.apache.org
[spark] branch branch-3.0 updated (f50432f -> 8a52bda)
This is an automated email from the ASF dual-hosted git repository. huaxingao pushed a change to branch branch-3.0 in repository https://gitbox.apache.org/repos/asf/spark.git. from f50432f [SPARK-32363][PYTHON][BUILD][3.0] Fix flakiness in pip package testing in Jenkins add 8a52bda [SPARK-32310][ML][PYSPARK][3.0] ML params default value parity No new revisions were added by this update. Summary of changes: .../spark/ml/classification/FMClassifier.scala | 10 -- .../apache/spark/ml/classification/LinearSVC.scala | 11 +-- .../ml/classification/LogisticRegression.scala | 13 +-- .../spark/ml/classification/NaiveBayes.scala | 4 +- .../spark/ml/clustering/BisectingKMeans.scala | 7 +- .../spark/ml/clustering/GaussianMixture.scala | 7 +- .../org/apache/spark/ml/clustering/KMeans.scala| 11 +-- .../scala/org/apache/spark/ml/clustering/LDA.scala | 11 +-- .../ml/clustering/PowerIterationClustering.scala | 7 +- .../evaluation/BinaryClassificationEvaluator.scala | 4 +- .../MulticlassClassificationEvaluator.scala| 8 +- .../MultilabelClassificationEvaluator.scala| 6 +- .../spark/ml/evaluation/RankingEvaluator.scala | 6 +- .../spark/ml/evaluation/RegressionEvaluator.scala | 4 +- .../apache/spark/ml/feature/ChiSqSelector.scala| 9 +- .../org/apache/spark/ml/feature/Imputer.scala | 4 +- .../org/apache/spark/ml/feature/MinMaxScaler.scala | 4 +- .../apache/spark/ml/feature/OneHotEncoder.scala| 5 +- .../spark/ml/feature/QuantileDiscretizer.scala | 4 +- .../org/apache/spark/ml/feature/RFormula.scala | 6 +- .../org/apache/spark/ml/feature/RobustScaler.scala | 8 +- .../apache/spark/ml/feature/StringIndexer.scala| 6 +- .../apache/spark/ml/feature/VectorIndexer.scala| 6 +- .../org/apache/spark/ml/feature/VectorSlicer.scala | 6 +- .../org/apache/spark/ml/feature/Word2Vec.scala | 9 +- .../scala/org/apache/spark/ml/fpm/FPGrowth.scala | 5 +- .../ml/regression/AFTSurvivalRegression.scala | 10 +- .../spark/ml/regression/LinearRegression.scala | 14 +-- .../org/apache/spark/ml/tree/treeParams.scala | 16 +-- .../spark/ml/util/DefaultReadWriteTest.scala | 3 + python/pyspark/ml/classification.py| 86 +++- python/pyspark/ml/clustering.py| 43 ++-- python/pyspark/ml/feature.py | 110 ++--- python/pyspark/ml/fpm.py | 12 ++- python/pyspark/ml/recommendation.py| 20 ++-- python/pyspark/ml/regression.py| 88 - python/pyspark/ml/tests/test_param.py | 7 +- python/pyspark/ml/tuning.py| 16 ++- 38 files changed, 368 insertions(+), 238 deletions(-) - To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org For additional commands, e-mail: commits-h...@spark.apache.org
[spark] branch master updated (fa184c3 -> d3596c0)
This is an automated email from the ASF dual-hosted git repository. dongjoon pushed a change to branch master in repository https://gitbox.apache.org/repos/asf/spark.git. from fa184c3 [SPARK-32408][BUILD] Enable crossPaths back to prevent side effects add d3596c0 [SPARK-32406][SQL] Make RESET syntax support single configuration reset No new revisions were added by this update. Summary of changes: docs/sql-ref-syntax-aux-conf-mgmt-reset.md| 19 +-- .../org/apache/spark/sql/catalyst/parser/SqlBase.g4 | 2 +- .../apache/spark/sql/execution/SparkSqlParser.scala | 3 ++- .../spark/sql/execution/command/SetCommand.scala | 14 ++ .../org/apache/spark/sql/internal/SQLConfSuite.scala | 15 +++ 5 files changed, 45 insertions(+), 8 deletions(-) - To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org For additional commands, e-mail: commits-h...@spark.apache.org
[spark] branch master updated (fa184c3 -> d3596c0)
This is an automated email from the ASF dual-hosted git repository. dongjoon pushed a change to branch master in repository https://gitbox.apache.org/repos/asf/spark.git. from fa184c3 [SPARK-32408][BUILD] Enable crossPaths back to prevent side effects add d3596c0 [SPARK-32406][SQL] Make RESET syntax support single configuration reset No new revisions were added by this update. Summary of changes: docs/sql-ref-syntax-aux-conf-mgmt-reset.md| 19 +-- .../org/apache/spark/sql/catalyst/parser/SqlBase.g4 | 2 +- .../apache/spark/sql/execution/SparkSqlParser.scala | 3 ++- .../spark/sql/execution/command/SetCommand.scala | 14 ++ .../org/apache/spark/sql/internal/SQLConfSuite.scala | 15 +++ 5 files changed, 45 insertions(+), 8 deletions(-) - To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org For additional commands, e-mail: commits-h...@spark.apache.org
[spark] branch master updated (fa184c3 -> d3596c0)
This is an automated email from the ASF dual-hosted git repository. dongjoon pushed a change to branch master in repository https://gitbox.apache.org/repos/asf/spark.git. from fa184c3 [SPARK-32408][BUILD] Enable crossPaths back to prevent side effects add d3596c0 [SPARK-32406][SQL] Make RESET syntax support single configuration reset No new revisions were added by this update. Summary of changes: docs/sql-ref-syntax-aux-conf-mgmt-reset.md| 19 +-- .../org/apache/spark/sql/catalyst/parser/SqlBase.g4 | 2 +- .../apache/spark/sql/execution/SparkSqlParser.scala | 3 ++- .../spark/sql/execution/command/SetCommand.scala | 14 ++ .../org/apache/spark/sql/internal/SQLConfSuite.scala | 15 +++ 5 files changed, 45 insertions(+), 8 deletions(-) - To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org For additional commands, e-mail: commits-h...@spark.apache.org
[spark] branch master updated (fa184c3 -> d3596c0)
This is an automated email from the ASF dual-hosted git repository. dongjoon pushed a change to branch master in repository https://gitbox.apache.org/repos/asf/spark.git. from fa184c3 [SPARK-32408][BUILD] Enable crossPaths back to prevent side effects add d3596c0 [SPARK-32406][SQL] Make RESET syntax support single configuration reset No new revisions were added by this update. Summary of changes: docs/sql-ref-syntax-aux-conf-mgmt-reset.md| 19 +-- .../org/apache/spark/sql/catalyst/parser/SqlBase.g4 | 2 +- .../apache/spark/sql/execution/SparkSqlParser.scala | 3 ++- .../spark/sql/execution/command/SetCommand.scala | 14 ++ .../org/apache/spark/sql/internal/SQLConfSuite.scala | 15 +++ 5 files changed, 45 insertions(+), 8 deletions(-) - To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org For additional commands, e-mail: commits-h...@spark.apache.org
[spark] branch master updated (fa184c3 -> d3596c0)
This is an automated email from the ASF dual-hosted git repository. dongjoon pushed a change to branch master in repository https://gitbox.apache.org/repos/asf/spark.git. from fa184c3 [SPARK-32408][BUILD] Enable crossPaths back to prevent side effects add d3596c0 [SPARK-32406][SQL] Make RESET syntax support single configuration reset No new revisions were added by this update. Summary of changes: docs/sql-ref-syntax-aux-conf-mgmt-reset.md| 19 +-- .../org/apache/spark/sql/catalyst/parser/SqlBase.g4 | 2 +- .../apache/spark/sql/execution/SparkSqlParser.scala | 3 ++- .../spark/sql/execution/command/SetCommand.scala | 14 ++ .../org/apache/spark/sql/internal/SQLConfSuite.scala | 15 +++ 5 files changed, 45 insertions(+), 8 deletions(-) - To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org For additional commands, e-mail: commits-h...@spark.apache.org
[spark] branch master updated (8bc799f9 -> fa184c3)
This is an automated email from the ASF dual-hosted git repository. dongjoon pushed a change to branch master in repository https://gitbox.apache.org/repos/asf/spark.git. from 8bc799f9 [SPARK-32375][SQL] Basic functionality of table catalog v2 for JDBC add fa184c3 [SPARK-32408][BUILD] Enable crossPaths back to prevent side effects No new revisions were added by this update. Summary of changes: project/SparkBuild.scala | 2 -- 1 file changed, 2 deletions(-) - To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org For additional commands, e-mail: commits-h...@spark.apache.org
[spark] branch master updated (8bc799f9 -> fa184c3)
This is an automated email from the ASF dual-hosted git repository. dongjoon pushed a change to branch master in repository https://gitbox.apache.org/repos/asf/spark.git. from 8bc799f9 [SPARK-32375][SQL] Basic functionality of table catalog v2 for JDBC add fa184c3 [SPARK-32408][BUILD] Enable crossPaths back to prevent side effects No new revisions were added by this update. Summary of changes: project/SparkBuild.scala | 2 -- 1 file changed, 2 deletions(-) - To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org For additional commands, e-mail: commits-h...@spark.apache.org
[spark] branch master updated (8bc799f9 -> fa184c3)
This is an automated email from the ASF dual-hosted git repository. dongjoon pushed a change to branch master in repository https://gitbox.apache.org/repos/asf/spark.git. from 8bc799f9 [SPARK-32375][SQL] Basic functionality of table catalog v2 for JDBC add fa184c3 [SPARK-32408][BUILD] Enable crossPaths back to prevent side effects No new revisions were added by this update. Summary of changes: project/SparkBuild.scala | 2 -- 1 file changed, 2 deletions(-) - To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org For additional commands, e-mail: commits-h...@spark.apache.org
[spark] branch master updated (8bc799f9 -> fa184c3)
This is an automated email from the ASF dual-hosted git repository. dongjoon pushed a change to branch master in repository https://gitbox.apache.org/repos/asf/spark.git. from 8bc799f9 [SPARK-32375][SQL] Basic functionality of table catalog v2 for JDBC add fa184c3 [SPARK-32408][BUILD] Enable crossPaths back to prevent side effects No new revisions were added by this update. Summary of changes: project/SparkBuild.scala | 2 -- 1 file changed, 2 deletions(-) - To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org For additional commands, e-mail: commits-h...@spark.apache.org
[spark] branch master updated (8bc799f9 -> fa184c3)
This is an automated email from the ASF dual-hosted git repository. dongjoon pushed a change to branch master in repository https://gitbox.apache.org/repos/asf/spark.git. from 8bc799f9 [SPARK-32375][SQL] Basic functionality of table catalog v2 for JDBC add fa184c3 [SPARK-32408][BUILD] Enable crossPaths back to prevent side effects No new revisions were added by this update. Summary of changes: project/SparkBuild.scala | 2 -- 1 file changed, 2 deletions(-) - To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org For additional commands, e-mail: commits-h...@spark.apache.org
[spark] branch branch-3.0 updated: [SPARK-32363][PYTHON][BUILD][3.0] Fix flakiness in pip package testing in Jenkins
This is an automated email from the ASF dual-hosted git repository. dongjoon pushed a commit to branch branch-3.0 in repository https://gitbox.apache.org/repos/asf/spark.git The following commit(s) were added to refs/heads/branch-3.0 by this push: new f50432f [SPARK-32363][PYTHON][BUILD][3.0] Fix flakiness in pip package testing in Jenkins f50432f is described below commit f50432fd3d835f8b5de3283edce4c146cfafe827 Author: HyukjinKwon AuthorDate: Fri Jul 24 07:18:15 2020 -0700 [SPARK-32363][PYTHON][BUILD][3.0] Fix flakiness in pip package testing in Jenkins ### What changes were proposed in this pull request? This PR backports https://github.com/apache/spark/pull/29117 to branch-3.0 as the flakiness was found in branch-3.0 too: https://github.com/apache/spark/pull/29201#issuecomment-663114741 and https://github.com/apache/spark/pull/29201#issuecomment-663114741 This PR proposes: - ~~Don't use `--user` in pip packaging test~~ - ~~Pull `source` out of the subshell, and place it first.~~ - Exclude user sitepackages in Python path during pip installation test to address the flakiness of the pip packaging test in Jenkins. ~~(I think) #29116 caused this flakiness given my observation in the Jenkins log. I had to work around by specifying `--user` but it turned out that it does not properly work in old Conda on Jenkins for some reasons. Therefore, reverting this change back.~~ (I think) the installation at user site-packages affects other environments created by Conda in the old Conda version that Jenkins has. Seems it fails to isolate the environments for some reasons. So, it excludes user sitepackages in the Python path during the test. ~~In addition, #29116 also added some fallback logics of `conda (de)activate` and `source (de)activate` because Conda prefers to use `conda (de)activate` now per the official documentation and `source (de)activate` doesn't work for some reasons in certain environments (see also https://github.com/conda/conda/issues/7980). The problem was that `source` loads things to the current shell so does not affect the current shell. Therefore, this PR pulls `source` out of the subshell.~~ Disclaimer: I made the analysis purely based on Jenkins machine's log in this PR. It may have a different reason I missed during my observation. ### Why are the changes needed? To make the build and tests pass in Jenkins. ### Does this PR introduce _any_ user-facing change? No, dev-only. ### How was this patch tested? Jenkins tests should test it out. Closes #29215 from HyukjinKwon/SPARK-32363-3.0. Authored-by: HyukjinKwon Signed-off-by: Dongjoon Hyun --- dev/run-pip-tests | 4 1 file changed, 4 insertions(+) diff --git a/dev/run-pip-tests b/dev/run-pip-tests index 470f21e..81e33a6 100755 --- a/dev/run-pip-tests +++ b/dev/run-pip-tests @@ -68,6 +68,10 @@ PIP_OPTIONS="--upgrade --no-cache-dir --force-reinstall " PIP_COMMANDS=("pip install $PIP_OPTIONS $PYSPARK_DIST" "pip install $PIP_OPTIONS -e python/") +# Jenkins has PySpark installed under user sitepackages shared for some reasons. +# In this test, explicitly exclude user sitepackages to prevent side effects +export PYTHONNOUSERSITE=1 + for python in "${PYTHON_EXECS[@]}"; do for install_command in "${PIP_COMMANDS[@]}"; do echo "Testing pip installation with python $python" - To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org For additional commands, e-mail: commits-h...@spark.apache.org
[spark] branch branch-3.0 updated: [SPARK-32363][PYTHON][BUILD][3.0] Fix flakiness in pip package testing in Jenkins
This is an automated email from the ASF dual-hosted git repository. dongjoon pushed a commit to branch branch-3.0 in repository https://gitbox.apache.org/repos/asf/spark.git The following commit(s) were added to refs/heads/branch-3.0 by this push: new f50432f [SPARK-32363][PYTHON][BUILD][3.0] Fix flakiness in pip package testing in Jenkins f50432f is described below commit f50432fd3d835f8b5de3283edce4c146cfafe827 Author: HyukjinKwon AuthorDate: Fri Jul 24 07:18:15 2020 -0700 [SPARK-32363][PYTHON][BUILD][3.0] Fix flakiness in pip package testing in Jenkins ### What changes were proposed in this pull request? This PR backports https://github.com/apache/spark/pull/29117 to branch-3.0 as the flakiness was found in branch-3.0 too: https://github.com/apache/spark/pull/29201#issuecomment-663114741 and https://github.com/apache/spark/pull/29201#issuecomment-663114741 This PR proposes: - ~~Don't use `--user` in pip packaging test~~ - ~~Pull `source` out of the subshell, and place it first.~~ - Exclude user sitepackages in Python path during pip installation test to address the flakiness of the pip packaging test in Jenkins. ~~(I think) #29116 caused this flakiness given my observation in the Jenkins log. I had to work around by specifying `--user` but it turned out that it does not properly work in old Conda on Jenkins for some reasons. Therefore, reverting this change back.~~ (I think) the installation at user site-packages affects other environments created by Conda in the old Conda version that Jenkins has. Seems it fails to isolate the environments for some reasons. So, it excludes user sitepackages in the Python path during the test. ~~In addition, #29116 also added some fallback logics of `conda (de)activate` and `source (de)activate` because Conda prefers to use `conda (de)activate` now per the official documentation and `source (de)activate` doesn't work for some reasons in certain environments (see also https://github.com/conda/conda/issues/7980). The problem was that `source` loads things to the current shell so does not affect the current shell. Therefore, this PR pulls `source` out of the subshell.~~ Disclaimer: I made the analysis purely based on Jenkins machine's log in this PR. It may have a different reason I missed during my observation. ### Why are the changes needed? To make the build and tests pass in Jenkins. ### Does this PR introduce _any_ user-facing change? No, dev-only. ### How was this patch tested? Jenkins tests should test it out. Closes #29215 from HyukjinKwon/SPARK-32363-3.0. Authored-by: HyukjinKwon Signed-off-by: Dongjoon Hyun --- dev/run-pip-tests | 4 1 file changed, 4 insertions(+) diff --git a/dev/run-pip-tests b/dev/run-pip-tests index 470f21e..81e33a6 100755 --- a/dev/run-pip-tests +++ b/dev/run-pip-tests @@ -68,6 +68,10 @@ PIP_OPTIONS="--upgrade --no-cache-dir --force-reinstall " PIP_COMMANDS=("pip install $PIP_OPTIONS $PYSPARK_DIST" "pip install $PIP_OPTIONS -e python/") +# Jenkins has PySpark installed under user sitepackages shared for some reasons. +# In this test, explicitly exclude user sitepackages to prevent side effects +export PYTHONNOUSERSITE=1 + for python in "${PYTHON_EXECS[@]}"; do for install_command in "${PIP_COMMANDS[@]}"; do echo "Testing pip installation with python $python" - To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org For additional commands, e-mail: commits-h...@spark.apache.org
[spark] branch branch-3.0 updated: [SPARK-32363][PYTHON][BUILD][3.0] Fix flakiness in pip package testing in Jenkins
This is an automated email from the ASF dual-hosted git repository. dongjoon pushed a commit to branch branch-3.0 in repository https://gitbox.apache.org/repos/asf/spark.git The following commit(s) were added to refs/heads/branch-3.0 by this push: new f50432f [SPARK-32363][PYTHON][BUILD][3.0] Fix flakiness in pip package testing in Jenkins f50432f is described below commit f50432fd3d835f8b5de3283edce4c146cfafe827 Author: HyukjinKwon AuthorDate: Fri Jul 24 07:18:15 2020 -0700 [SPARK-32363][PYTHON][BUILD][3.0] Fix flakiness in pip package testing in Jenkins ### What changes were proposed in this pull request? This PR backports https://github.com/apache/spark/pull/29117 to branch-3.0 as the flakiness was found in branch-3.0 too: https://github.com/apache/spark/pull/29201#issuecomment-663114741 and https://github.com/apache/spark/pull/29201#issuecomment-663114741 This PR proposes: - ~~Don't use `--user` in pip packaging test~~ - ~~Pull `source` out of the subshell, and place it first.~~ - Exclude user sitepackages in Python path during pip installation test to address the flakiness of the pip packaging test in Jenkins. ~~(I think) #29116 caused this flakiness given my observation in the Jenkins log. I had to work around by specifying `--user` but it turned out that it does not properly work in old Conda on Jenkins for some reasons. Therefore, reverting this change back.~~ (I think) the installation at user site-packages affects other environments created by Conda in the old Conda version that Jenkins has. Seems it fails to isolate the environments for some reasons. So, it excludes user sitepackages in the Python path during the test. ~~In addition, #29116 also added some fallback logics of `conda (de)activate` and `source (de)activate` because Conda prefers to use `conda (de)activate` now per the official documentation and `source (de)activate` doesn't work for some reasons in certain environments (see also https://github.com/conda/conda/issues/7980). The problem was that `source` loads things to the current shell so does not affect the current shell. Therefore, this PR pulls `source` out of the subshell.~~ Disclaimer: I made the analysis purely based on Jenkins machine's log in this PR. It may have a different reason I missed during my observation. ### Why are the changes needed? To make the build and tests pass in Jenkins. ### Does this PR introduce _any_ user-facing change? No, dev-only. ### How was this patch tested? Jenkins tests should test it out. Closes #29215 from HyukjinKwon/SPARK-32363-3.0. Authored-by: HyukjinKwon Signed-off-by: Dongjoon Hyun --- dev/run-pip-tests | 4 1 file changed, 4 insertions(+) diff --git a/dev/run-pip-tests b/dev/run-pip-tests index 470f21e..81e33a6 100755 --- a/dev/run-pip-tests +++ b/dev/run-pip-tests @@ -68,6 +68,10 @@ PIP_OPTIONS="--upgrade --no-cache-dir --force-reinstall " PIP_COMMANDS=("pip install $PIP_OPTIONS $PYSPARK_DIST" "pip install $PIP_OPTIONS -e python/") +# Jenkins has PySpark installed under user sitepackages shared for some reasons. +# In this test, explicitly exclude user sitepackages to prevent side effects +export PYTHONNOUSERSITE=1 + for python in "${PYTHON_EXECS[@]}"; do for install_command in "${PIP_COMMANDS[@]}"; do echo "Testing pip installation with python $python" - To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org For additional commands, e-mail: commits-h...@spark.apache.org
[spark] branch master updated (8896f4a -> 8bc799f9)
This is an automated email from the ASF dual-hosted git repository. wenchen pushed a change to branch master in repository https://gitbox.apache.org/repos/asf/spark.git. from 8896f4a Revert "[SPARK-32253][INFRA] Show errors only for the sbt tests of github actions" add 8bc799f9 [SPARK-32375][SQL] Basic functionality of table catalog v2 for JDBC No new revisions were added by this update. Summary of changes: .../datasources/jdbc/JdbcRelationProvider.scala| 4 +- .../sql/execution/datasources/jdbc/JdbcUtils.scala | 50 +-- .../jdbc/JDBCTable.scala} | 28 ++-- .../datasources/v2/jdbc/JDBCTableCatalog.scala | 158 + .../org/apache/spark/sql/jdbc/JdbcDialects.scala | 13 ++ .../v2/jdbc/JDBCTableCatalogSuite.scala| 109 ++ .../org/apache/spark/sql/jdbc/JDBCSuite.scala | 5 +- .../org/apache/spark/sql/jdbc/JDBCWriteSuite.scala | 6 +- 8 files changed, 343 insertions(+), 30 deletions(-) copy sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/{jdbc/connection/BasicConnectionProvider.scala => v2/jdbc/JDBCTable.scala} (53%) create mode 100644 sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/v2/jdbc/JDBCTableCatalog.scala create mode 100644 sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/v2/jdbc/JDBCTableCatalogSuite.scala - To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org For additional commands, e-mail: commits-h...@spark.apache.org
[spark] branch master updated (8896f4a -> 8bc799f9)
This is an automated email from the ASF dual-hosted git repository. wenchen pushed a change to branch master in repository https://gitbox.apache.org/repos/asf/spark.git. from 8896f4a Revert "[SPARK-32253][INFRA] Show errors only for the sbt tests of github actions" add 8bc799f9 [SPARK-32375][SQL] Basic functionality of table catalog v2 for JDBC No new revisions were added by this update. Summary of changes: .../datasources/jdbc/JdbcRelationProvider.scala| 4 +- .../sql/execution/datasources/jdbc/JdbcUtils.scala | 50 +-- .../jdbc/JDBCTable.scala} | 28 ++-- .../datasources/v2/jdbc/JDBCTableCatalog.scala | 158 + .../org/apache/spark/sql/jdbc/JdbcDialects.scala | 13 ++ .../v2/jdbc/JDBCTableCatalogSuite.scala| 109 ++ .../org/apache/spark/sql/jdbc/JDBCSuite.scala | 5 +- .../org/apache/spark/sql/jdbc/JDBCWriteSuite.scala | 6 +- 8 files changed, 343 insertions(+), 30 deletions(-) copy sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/{jdbc/connection/BasicConnectionProvider.scala => v2/jdbc/JDBCTable.scala} (53%) create mode 100644 sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/v2/jdbc/JDBCTableCatalog.scala create mode 100644 sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/v2/jdbc/JDBCTableCatalogSuite.scala - To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org For additional commands, e-mail: commits-h...@spark.apache.org
[spark] branch master updated (8896f4a -> 8bc799f9)
This is an automated email from the ASF dual-hosted git repository. wenchen pushed a change to branch master in repository https://gitbox.apache.org/repos/asf/spark.git. from 8896f4a Revert "[SPARK-32253][INFRA] Show errors only for the sbt tests of github actions" add 8bc799f9 [SPARK-32375][SQL] Basic functionality of table catalog v2 for JDBC No new revisions were added by this update. Summary of changes: .../datasources/jdbc/JdbcRelationProvider.scala| 4 +- .../sql/execution/datasources/jdbc/JdbcUtils.scala | 50 +-- .../jdbc/JDBCTable.scala} | 28 ++-- .../datasources/v2/jdbc/JDBCTableCatalog.scala | 158 + .../org/apache/spark/sql/jdbc/JdbcDialects.scala | 13 ++ .../v2/jdbc/JDBCTableCatalogSuite.scala| 109 ++ .../org/apache/spark/sql/jdbc/JDBCSuite.scala | 5 +- .../org/apache/spark/sql/jdbc/JDBCWriteSuite.scala | 6 +- 8 files changed, 343 insertions(+), 30 deletions(-) copy sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/{jdbc/connection/BasicConnectionProvider.scala => v2/jdbc/JDBCTable.scala} (53%) create mode 100644 sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/v2/jdbc/JDBCTableCatalog.scala create mode 100644 sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/v2/jdbc/JDBCTableCatalogSuite.scala - To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org For additional commands, e-mail: commits-h...@spark.apache.org
[spark] branch master updated (8896f4a -> 8bc799f9)
This is an automated email from the ASF dual-hosted git repository. wenchen pushed a change to branch master in repository https://gitbox.apache.org/repos/asf/spark.git. from 8896f4a Revert "[SPARK-32253][INFRA] Show errors only for the sbt tests of github actions" add 8bc799f9 [SPARK-32375][SQL] Basic functionality of table catalog v2 for JDBC No new revisions were added by this update. Summary of changes: .../datasources/jdbc/JdbcRelationProvider.scala| 4 +- .../sql/execution/datasources/jdbc/JdbcUtils.scala | 50 +-- .../jdbc/JDBCTable.scala} | 28 ++-- .../datasources/v2/jdbc/JDBCTableCatalog.scala | 158 + .../org/apache/spark/sql/jdbc/JdbcDialects.scala | 13 ++ .../v2/jdbc/JDBCTableCatalogSuite.scala| 109 ++ .../org/apache/spark/sql/jdbc/JDBCSuite.scala | 5 +- .../org/apache/spark/sql/jdbc/JDBCWriteSuite.scala | 6 +- 8 files changed, 343 insertions(+), 30 deletions(-) copy sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/{jdbc/connection/BasicConnectionProvider.scala => v2/jdbc/JDBCTable.scala} (53%) create mode 100644 sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/v2/jdbc/JDBCTableCatalog.scala create mode 100644 sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/v2/jdbc/JDBCTableCatalogSuite.scala - To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org For additional commands, e-mail: commits-h...@spark.apache.org
[spark] branch master updated (8896f4a -> 8bc799f9)
This is an automated email from the ASF dual-hosted git repository. wenchen pushed a change to branch master in repository https://gitbox.apache.org/repos/asf/spark.git. from 8896f4a Revert "[SPARK-32253][INFRA] Show errors only for the sbt tests of github actions" add 8bc799f9 [SPARK-32375][SQL] Basic functionality of table catalog v2 for JDBC No new revisions were added by this update. Summary of changes: .../datasources/jdbc/JdbcRelationProvider.scala| 4 +- .../sql/execution/datasources/jdbc/JdbcUtils.scala | 50 +-- .../jdbc/JDBCTable.scala} | 28 ++-- .../datasources/v2/jdbc/JDBCTableCatalog.scala | 158 + .../org/apache/spark/sql/jdbc/JdbcDialects.scala | 13 ++ .../v2/jdbc/JDBCTableCatalogSuite.scala| 109 ++ .../org/apache/spark/sql/jdbc/JDBCSuite.scala | 5 +- .../org/apache/spark/sql/jdbc/JDBCWriteSuite.scala | 6 +- 8 files changed, 343 insertions(+), 30 deletions(-) copy sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/{jdbc/connection/BasicConnectionProvider.scala => v2/jdbc/JDBCTable.scala} (53%) create mode 100644 sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/v2/jdbc/JDBCTableCatalog.scala create mode 100644 sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/v2/jdbc/JDBCTableCatalogSuite.scala - To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org For additional commands, e-mail: commits-h...@spark.apache.org
[spark] branch master updated (84efa04 -> 8896f4a)
This is an automated email from the ASF dual-hosted git repository. gengliang pushed a change to branch master in repository https://gitbox.apache.org/repos/asf/spark.git. from 84efa04 [SPARK-32308][SQL] Move by-name resolution logic of unionByName from API code to analysis phase add 8896f4a Revert "[SPARK-32253][INFRA] Show errors only for the sbt tests of github actions" No new revisions were added by this update. Summary of changes: dev/run-tests.py | 3 --- project/SparkBuild.scala | 7 --- 2 files changed, 10 deletions(-) - To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org For additional commands, e-mail: commits-h...@spark.apache.org
[spark] branch master updated (84efa04 -> 8896f4a)
This is an automated email from the ASF dual-hosted git repository. gengliang pushed a change to branch master in repository https://gitbox.apache.org/repos/asf/spark.git. from 84efa04 [SPARK-32308][SQL] Move by-name resolution logic of unionByName from API code to analysis phase add 8896f4a Revert "[SPARK-32253][INFRA] Show errors only for the sbt tests of github actions" No new revisions were added by this update. Summary of changes: dev/run-tests.py | 3 --- project/SparkBuild.scala | 7 --- 2 files changed, 10 deletions(-) - To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org For additional commands, e-mail: commits-h...@spark.apache.org
[spark] branch master updated (84efa04 -> 8896f4a)
This is an automated email from the ASF dual-hosted git repository. gengliang pushed a change to branch master in repository https://gitbox.apache.org/repos/asf/spark.git. from 84efa04 [SPARK-32308][SQL] Move by-name resolution logic of unionByName from API code to analysis phase add 8896f4a Revert "[SPARK-32253][INFRA] Show errors only for the sbt tests of github actions" No new revisions were added by this update. Summary of changes: dev/run-tests.py | 3 --- project/SparkBuild.scala | 7 --- 2 files changed, 10 deletions(-) - To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org For additional commands, e-mail: commits-h...@spark.apache.org
[spark] branch master updated (84efa04 -> 8896f4a)
This is an automated email from the ASF dual-hosted git repository. gengliang pushed a change to branch master in repository https://gitbox.apache.org/repos/asf/spark.git. from 84efa04 [SPARK-32308][SQL] Move by-name resolution logic of unionByName from API code to analysis phase add 8896f4a Revert "[SPARK-32253][INFRA] Show errors only for the sbt tests of github actions" No new revisions were added by this update. Summary of changes: dev/run-tests.py | 3 --- project/SparkBuild.scala | 7 --- 2 files changed, 10 deletions(-) - To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org For additional commands, e-mail: commits-h...@spark.apache.org
[spark] branch master updated (84efa04 -> 8896f4a)
This is an automated email from the ASF dual-hosted git repository. gengliang pushed a change to branch master in repository https://gitbox.apache.org/repos/asf/spark.git. from 84efa04 [SPARK-32308][SQL] Move by-name resolution logic of unionByName from API code to analysis phase add 8896f4a Revert "[SPARK-32253][INFRA] Show errors only for the sbt tests of github actions" No new revisions were added by this update. Summary of changes: dev/run-tests.py | 3 --- project/SparkBuild.scala | 7 --- 2 files changed, 10 deletions(-) - To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org For additional commands, e-mail: commits-h...@spark.apache.org