See <https://ci-beam.apache.org/job/beam_PostCommit_Python39/2805/display/redirect>
Changes: ------------------------------------------ [...truncated 12.37 MB...] File "apache_beam/runners/worker/operations.py", line 953, in apache_beam.runners.worker.operations.DoOperation.process File "apache_beam/runners/worker/operations.py", line 954, in apache_beam.runners.worker.operations.DoOperation.process File "apache_beam/runners/common.py", line 1437, in apache_beam.runners.common.DoFnRunner.process File "apache_beam/runners/common.py", line 1547, in apache_beam.runners.common.DoFnRunner._reraise_augmented File "apache_beam/runners/common.py", line 1435, in apache_beam.runners.common.DoFnRunner.process File "apache_beam/runners/common.py", line 851, in apache_beam.runners.common.PerWindowInvoker.invoke_process File "apache_beam/runners/common.py", line 995, in apache_beam.runners.common.PerWindowInvoker._invoke_process_per_window File "apache_beam/runners/common.py", line 1611, in apache_beam.runners.common._OutputHandler.handle_process_outputs File "/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.9/site-packages/apache_beam/io/gcp/bigquery_file_loads.py", line 750, in process job_reference = self.bq_wrapper.perform_load_job( File "/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.9/site-packages/apache_beam/io/gcp/bigquery_tools.py", line 1022, in perform_load_job return self._insert_load_job( File "/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.9/site-packages/apache_beam/utils/retry.py", line 298, in wrapper return fun(*args, **kwargs) File "/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.9/site-packages/apache_beam/io/gcp/bigquery_tools.py", line 559, in _insert_load_job return self._start_job(request, stream=source_stream).jobReference File "/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.9/site-packages/apache_beam/io/gcp/bigquery_tools.py", line 578, in _start_job response = self.client.jobs.Insert(request, upload=upload) File "/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.9/site-packages/apache_beam/io/gcp/internal/clients/bigquery/bigquery_v2_client.py", line 356, in Insert return self._RunMethod( File "/usr/local/lib/python3.9/site-packages/apitools/base/py/base_api.py", line 731, in _RunMethod return self.ProcessHttpResponse(method_config, http_response, request) File "/usr/local/lib/python3.9/site-packages/apitools/base/py/base_api.py", line 737, in ProcessHttpResponse self.__ProcessHttpResponse(method_config, http_response, request)) File "/usr/local/lib/python3.9/site-packages/apitools/base/py/base_api.py", line 603, in __ProcessHttpResponse raise exceptions.HttpError.FromResponse( RuntimeError: apitools.base.py.exceptions.HttpBadRequestError: HttpError accessing <https://bigquery.googleapis.com/bigquery/v2/projects/apache-beam-testing/jobs?alt=json>: response: <{'vary': 'Origin, X-Origin, Referer', 'content-type': 'application/json; charset=UTF-8', 'date': 'Thu, 25 Jan 2024 00:29:10 GMT', 'server': 'ESF', 'cache-control': 'private', 'x-xss-protection': '0', 'x-frame-options': 'SAMEORIGIN', 'x-content-type-options': 'nosniff', 'transfer-encoding': 'chunked', 'status': '400', 'content-length': '318', '-content-encoding': 'gzip'}>, content <{ "error": { "code": 400, "message": "Load configuration must specify at least one source URI", "errors": [ { "message": "Load configuration must specify at least one source URI", "domain": "global", "reason": "invalid" } ], "status": "INVALID_ARGUMENT" } } > [while running > 'WriteWithMultipleDestsFreely/BigQueryBatchFileLoads/TriggerLoadJobsWithTempTables/ParDo(TriggerLoadJobs)-ptransform-71'] Worker ID: beamapp-jenkins-012500254-01241625-uy51-harness-w2wg [33mWARNING [0m apache_beam.runners.dataflow.dataflow_runner:dataflow_runner.py:203 2024-01-25T00:29:11.982Z: JOB_MESSAGE_WARNING: Unable to delete temp files: "gs://temp-storage-for-end-to-end-tests/temp-it/beamapp-jenkins-0125002540-066487-zmwbf9ko.1706142340.066646/dax-tmp-2024-01-24_16_25_40-3576056548624936316-S56-2-ba1d8146837ee981/tmp-ba1d8146837eed58@DAX.recordio." [33mWARNING [0m apache_beam.runners.dataflow.dataflow_runner:dataflow_runner.py:203 2024-01-25T00:29:11.983Z: JOB_MESSAGE_WARNING: Unable to delete temp files: "gs://temp-storage-for-end-to-end-tests/temp-it/beamapp-jenkins-0125002540-066487-zmwbf9ko.1706142340.066646/dax-tmp-2024-01-24_16_25_40-3576056548624936316-S56-1-a88f64f614a2a437/tmp-a88f64f614a2a2e2@DAX.recordio." [33mWARNING [0m apache_beam.runners.dataflow.dataflow_runner:dataflow_runner.py:203 2024-01-25T00:29:11.987Z: JOB_MESSAGE_WARNING: Unable to delete temp files: "gs://temp-storage-for-end-to-end-tests/temp-it/beamapp-jenkins-0125002540-066487-zmwbf9ko.1706142340.066646/dax-tmp-2024-01-24_16_25_40-3576056548624936316-S56-0-aa9e464cd130d213/tmp-aa9e464cd130d6be@DAX.recordio." [33mWARNING [0m apache_beam.runners.dataflow.dataflow_runner:dataflow_runner.py:203 2024-01-25T00:29:12.006Z: JOB_MESSAGE_WARNING: S56:WriteWithMultipleDests/BigQueryBatchFileLoads/GroupFilesByTableDestinations/Read+WriteWithMultipleDests/BigQueryBatchFileLoads/ParDo(PartitionFiles)/ParDo(PartitionFiles)+WriteWithMultipleDests/BigQueryBatchFileLoads/FlattenPartitions+WriteWithMultipleDests/BigQueryBatchFileLoads/TriggerLoadJobsWithTempTables/ParDo(TriggerLoadJobs)+WriteWithMultipleDests/BigQueryBatchFileLoads/MapTuple(<lambda at bigquery_file_loads.py:1135>)+WriteWithMultipleDests/BigQueryBatchFileLoads/ParDo(UpdateDestinationSchema)+WriteWithMultipleDests/BigQueryBatchFileLoads/GroupByKey/Write failed. [32mINFO [0m apache_beam.runners.dataflow.dataflow_runner:dataflow_runner.py:201 2024-01-25T00:29:12.031Z: JOB_MESSAGE_BASIC: Finished operation WriteWithMultipleDests/BigQueryBatchFileLoads/GroupFilesByTableDestinations/Read+WriteWithMultipleDests/BigQueryBatchFileLoads/ParDo(PartitionFiles)/ParDo(PartitionFiles)+WriteWithMultipleDests/BigQueryBatchFileLoads/FlattenPartitions+WriteWithMultipleDests/BigQueryBatchFileLoads/TriggerLoadJobsWithTempTables/ParDo(TriggerLoadJobs)+WriteWithMultipleDests/BigQueryBatchFileLoads/MapTuple(<lambda at bigquery_file_loads.py:1135>)+WriteWithMultipleDests/BigQueryBatchFileLoads/ParDo(UpdateDestinationSchema)+WriteWithMultipleDests/BigQueryBatchFileLoads/GroupByKey/Write [32mINFO [0m apache_beam.runners.dataflow.dataflow_runner:dataflow_runner.py:201 2024-01-25T00:29:12.427Z: JOB_MESSAGE_BASIC: Stopping worker pool... [32mINFO [0m apache_beam.runners.dataflow.dataflow_runner:dataflow_runner.py:201 2024-01-25T00:31:37.231Z: JOB_MESSAGE_BASIC: Worker pool stopped. [32mINFO [0m apache_beam.runners.dataflow.dataflow_runner:dataflow_runner.py:151 Job 2024-01-24_16_25_40-3576056548624936316 is in state JOB_STATE_FAILED [1m[31mERROR [0m apache_beam.runners.dataflow.dataflow_runner:dataflow_runner.py:777 Console URL: https://console.cloud.google.com/dataflow/jobs/<RegionId>/2024-01-24_16_25_40-3576056548624936316?project=<ProjectId> [32mINFO [0m apache_beam.io.gcp.bigquery_file_loads_test:bigquery_file_loads_test.py:1242 Deleting dataset python_bq_file_loads_170614233681bd22 in project apache-beam-testing [33m=============================== warnings summary ===============================[0m apache_beam/io/gcp/bigquery.py:2634 <https://ci-beam.apache.org/job/beam_PostCommit_Python39/ws/src/sdks/python/apache_beam/io/gcp/bigquery.py>:2634: DeprecationWarning: invalid escape sequence \# """Read data from BigQuery. apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_native_source <https://ci-beam.apache.org/job/beam_PostCommit_Python39/ws/src/sdks/python/apache_beam/io/gcp/bigquery_read_it_test.py>:170: BeamDeprecationWarning: BigQuerySource is deprecated since 2.25.0. Use ReadFromBigQuery instead. beam.io.BigQuerySource(query=self.query, use_standard_sql=True))) apache_beam/io/gcp/bigquery_test.py::PubSubBigQueryIT::test_file_loads apache_beam/io/gcp/bigquery_test.py::PubSubBigQueryIT::test_streaming_inserts apache_beam/examples/complete/game/leader_board_it_test.py::LeaderBoardIT::test_leader_board_it apache_beam/examples/complete/game/game_stats_it_test.py::GameStatsIT::test_game_stats_it apache_beam/examples/complete/game/hourly_team_score_it_test.py::HourlyTeamScoreIT::test_hourly_team_score_it <https://ci-beam.apache.org/job/beam_PostCommit_Python39/ws/src/sdks/python/apache_beam/io/gcp/tests/utils.py>:63: PendingDeprecationWarning: Client.dataset is deprecated and will be removed in a future version. Use a string like 'my_project.my_dataset' or a cloud.google.bigquery.DatasetReference object, instead. dataset_ref = client.dataset(unique_dataset_name, project=project) apache_beam/examples/cookbook/bigquery_tornadoes_it_test.py::BigqueryTornadoesIT::test_bigquery_tornadoes_it <https://ci-beam.apache.org/job/beam_PostCommit_Python39/ws/src/sdks/python/apache_beam/io/gcp/tests/utils.py>:100: PendingDeprecationWarning: Client.dataset is deprecated and will be removed in a future version. Use a string like 'my_project.my_dataset' or a cloud.google.bigquery.DatasetReference object, instead. table_ref = client.dataset(dataset_id).table(table_id) apache_beam/io/gcp/bigquery_read_it_test.py::ReadNewTypesTests::test_native_source <https://ci-beam.apache.org/job/beam_PostCommit_Python39/ws/src/sdks/python/apache_beam/io/gcp/bigquery_read_it_test.py>:706: BeamDeprecationWarning: BigQuerySource is deprecated since 2.25.0. Use ReadFromBigQuery instead. beam.io.BigQuerySource(query=self.query, use_standard_sql=True))) -- Docs: https://docs.pytest.org/en/stable/how-to/capture-warnings.html - generated xml file: <https://ci-beam.apache.org/job/beam_PostCommit_Python39/ws/src/sdks/python/pytest_postCommitIT-df-py39.xml> - [36m[1m=========================== short test summary info ============================[0m [31mFAILED[0m apache_beam/io/gcp/bigquery_file_loads_test.py::[1mBigQueryFileLoadsIT::test_batch_copy_jobs_with_no_input_schema[0m - AssertionError: Expected: (<[<BigqueryFullResultMatcher(Expected data is [(0,), (1,), (2,), (3,), (4,)])>, <BigqueryFullResultMatcher(Expected data is [(5,), (6,), (7,), (8,), (9,)])>]>) but: <[<BigqueryFullResultMatcher(Expected data is [(0,), (1,), (2,), (3,), (4,)])>, <BigqueryFullResultMatcher(Expected data is [(5,), (6,), (7,), (8,), (9,)])>]> was <DataflowPipelineResult 2024-01-24_16_18_12-12488496925338720336 DONE> [31mFAILED[0m apache_beam/io/gcp/bigquery_file_loads_test.py::[1mBigQueryFileLoadsIT::test_multiple_destinations_transform[0m - apache_beam.runners.dataflow.dataflow_runner.DataflowRuntimeException: Dataflow pipeline failed. State: FAILED, Error: Traceback (most recent call last): File "apache_beam/runners/common.py", line 1435, in apache_beam.runners.common.DoFnRunner.process File "apache_beam/runners/common.py", line 851, in apache_beam.runners.common.PerWindowInvoker.invoke_process File "apache_beam/runners/common.py", line 995, in apache_beam.runners.common.PerWindowInvoker._invoke_process_per_window File "apache_beam/runners/common.py", line 1611, in apache_beam.runners.common._OutputHandler.handle_process_outputs File "/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.9/site-packages/apache_beam/io/gcp/bigquery_file_loads.py", line 750, in process job_reference = self.bq_wrapper.perform_load_job( File "/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.9/site-packages/apache_beam/io/gcp/bigquery_tools.py", line 1022, in perform_load_job return self._insert_load_job( File "/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.9/site-packages/apache_beam/utils/retry.py", line 298, in wrapper return fun(*args, **kwargs) File "/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.9/site-packages/apache_beam/io/gcp/bigquery_tools.py", line 559, in _insert_load_job return self._start_job(request, stream=source_stream).jobReference File "/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.9/site-packages/apache_beam/io/gcp/bigquery_tools.py", line 578, in _start_job response = self.client.jobs.Insert(request, upload=upload) File "/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.9/site-packages/apache_beam/io/gcp/internal/clients/bigquery/bigquery_v2_client.py", line 356, in Insert return self._RunMethod( File "/usr/local/lib/python3.9/site-packages/apitools/base/py/base_api.py", line 731, in _RunMethod return self.ProcessHttpResponse(method_config, http_response, request) File "/usr/local/lib/python3.9/site-packages/apitools/base/py/base_api.py", line 737, in ProcessHttpResponse self.__ProcessHttpResponse(method_config, http_response, request)) File "/usr/local/lib/python3.9/site-packages/apitools/base/py/base_api.py", line 603, in __ProcessHttpResponse raise exceptions.HttpError.FromResponse( apitools.base.py.exceptions.HttpBadRequestError: HttpError accessing <https://bigquery.googleapis.com/bigquery/v2/projects/apache-beam-testing/jobs?alt=json>: response: <{'vary': 'Origin, X-Origin, Referer', 'content-type': 'application/json; charset=UTF-8', 'date': 'Thu, 25 Jan 2024 00:29:10 GMT', 'server': 'ESF', 'cache-control': 'private', 'x-xss-protection': '0', 'x-frame-options': 'SAMEORIGIN', 'x-content-type-options': 'nosniff', 'transfer-encoding': 'chunked', 'status': '400', 'content-length': '318', '-content-encoding': 'gzip'}>, content <{ "error": { "code": 400, "message": "Load configuration must specify at least one source URI", "errors": [ { "message": "Load configuration must specify at least one source URI", "domain": "global", "reason": "invalid" } ], "status": "INVALID_ARGUMENT" } } > During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.9/site-packages/apache_beam/runners/worker/sdk_worker.py", line 311, in _execute response = task() File "/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.9/site-packages/apache_beam/runners/worker/sdk_worker.py", line 386, in <lambda> lambda: self.create_worker().do_instruction(request), request) File "/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.9/site-packages/apache_beam/runners/worker/sdk_worker.py", line 650, in do_instruction return getattr(self, request_type)( File "/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.9/site-packages/apache_beam/runners/worker/sdk_worker.py", line 688, in process_bundle bundle_processor.process_bundle(instruction_id)) File "/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.9/site-packages/apache_beam/runners/worker/bundle_processor.py", line 1113, in process_bundle input_op_by_transform_id[element.transform_id].process_encoded( File "/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.9/site-packages/apache_beam/runners/worker/bundle_processor.py", line 237, in process_encoded self.output(decoded_value) File "apache_beam/runners/worker/operations.py", line 570, in apache_beam.runners.worker.operations.Operation.output File "apache_beam/runners/worker/operations.py", line 572, in apache_beam.runners.worker.operations.Operation.output File "apache_beam/runners/worker/operations.py", line 263, in apache_beam.runners.worker.operations.SingletonElementConsumerSet.receive File "apache_beam/runners/worker/operations.py", line 266, in apache_beam.runners.worker.operations.SingletonElementConsumerSet.receive File "apache_beam/runners/worker/operations.py", line 953, in apache_beam.runners.worker.operations.DoOperation.process File "apache_beam/runners/worker/operations.py", line 954, in apache_beam.runners.worker.operations.DoOperation.process File "apache_beam/runners/common.py", line 1437, in apache_beam.runners.common.DoFnRunner.process File "apache_beam/runners/common.py", line 1526, in apache_beam.runners.common.DoFnRunner._reraise_augmented File "apache_beam/runners/common.py", line 1435, in apache_beam.runners.common.DoFnRunner.process File "apache_beam/runners/common.py", line 636, in apache_beam.runners.common.SimpleInvoker.invoke_process File "apache_beam/runners/common.py", line 1621, in apache_beam.runners.common._OutputHandler.handle_process_outputs File "apache_beam/runners/common.py", line 1736, in apache_beam.runners.common._OutputHandler._write_value_to_tag File "apache_beam/runners/worker/operations.py", line 263, in apache_beam.runners.worker.operations.SingletonElementConsumerSet.receive File "apache_beam/runners/worker/operations.py", line 266, in apache_beam.runners.worker.operations.SingletonElementConsumerSet.receive File "apache_beam/runners/worker/operations.py", line 1351, in apache_beam.runners.worker.operations.FlattenOperation.process File "apache_beam/runners/worker/operations.py", line 1354, in apache_beam.runners.worker.operations.FlattenOperation.process File "apache_beam/runners/worker/operations.py", line 572, in apache_beam.runners.worker.operations.Operation.output File "apache_beam/runners/worker/operations.py", line 263, in apache_beam.runners.worker.operations.SingletonElementConsumerSet.receive File "apache_beam/runners/worker/operations.py", line 266, in apache_beam.runners.worker.operations.SingletonElementConsumerSet.receive File "apache_beam/runners/worker/operations.py", line 953, in apache_beam.runners.worker.operations.DoOperation.process File "apache_beam/runners/worker/operations.py", line 954, in apache_beam.runners.worker.operations.DoOperation.process File "apache_beam/runners/common.py", line 1437, in apache_beam.runners.common.DoFnRunner.process File "apache_beam/runners/common.py", line 1547, in apache_beam.runners.common.DoFnRunner._reraise_augmented File "apache_beam/runners/common.py", line 1435, in apache_beam.runners.common.DoFnRunner.process File "apache_beam/runners/common.py", line 851, in apache_beam.runners.common.PerWindowInvoker.invoke_process File "apache_beam/runners/common.py", line 995, in apache_beam.runners.common.PerWindowInvoker._invoke_process_per_window File "apache_beam/runners/common.py", line 1611, in apache_beam.runners.common._OutputHandler.handle_process_outputs File "/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.9/site-packages/apache_beam/io/gcp/bigquery_file_loads.py", line 750, in process job_reference = self.bq_wrapper.perform_load_job( File "/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.9/site-packages/apache_beam/io/gcp/bigquery_tools.py", line 1022, in perform_load_job return self._insert_load_job( File "/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.9/site-packages/apache_beam/utils/retry.py", line 298, in wrapper return fun(*args, **kwargs) File "/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.9/site-packages/apache_beam/io/gcp/bigquery_tools.py", line 559, in _insert_load_job return self._start_job(request, stream=source_stream).jobReference File "/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.9/site-packages/apache_beam/io/gcp/bigquery_tools.py", line 578, in _start_job response = self.client.jobs.Insert(request, upload=upload) File "/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.9/site-packages/apache_beam/io/gcp/internal/clients/bigquery/bigquery_v2_client.py", line 356, in Insert return self._RunMethod( File "/usr/local/lib/python3.9/site-packages/apitools/base/py/base_api.py", line 731, in _RunMethod return self.ProcessHttpResponse(method_config, http_response, request) File "/usr/local/lib/python3.9/site-packages/apitools/base/py/base_api.py", line 737, in ProcessHttpResponse self.__ProcessHttpResponse(method_config, http_response, request)) File "/usr/local/lib/python3.9/site-packages/apitools/base/py/base_api.py", line 603, in __ProcessHttpResponse raise exceptions.HttpError.FromResponse( RuntimeError: apitools.base.py.exceptions.HttpBadRequestError: HttpError accessing <https://bigquery.googleapis.com/bigquery/v2/projects/apache-beam-testing/jobs?alt=json>: response: <{'vary': 'Origin, X-Origin, Referer', 'content-type': 'application/json; charset=UTF-8', 'date': 'Thu, 25 Jan 2024 00:29:10 GMT', 'server': 'ESF', 'cache-control': 'private', 'x-xss-protection': '0', 'x-frame-options': 'SAMEORIGIN', 'x-content-type-options': 'nosniff', 'transfer-encoding': 'chunked', 'status': '400', 'content-length': '318', '-content-encoding': 'gzip'}>, content <{ "error": { "code": 400, "message": "Load configuration must specify at least one source URI", "errors": [ { "message": "Load configuration must specify at least one source URI", "domain": "global", "reason": "invalid" } ], "status": "INVALID_ARGUMENT" } } > [while running > 'WriteWithMultipleDestsFreely/BigQueryBatchFileLoads/TriggerLoadJobsWithTempTables/ParDo(TriggerLoadJobs)-ptransform-71'] [31m====== [31m[1m2 failed[0m, [32m90 passed[0m, [33m51 skipped[0m, [33m9 warnings[0m[31m in 5741.10s (1:35:41)[0m[31m =======[0m > Task :sdks:python:test-suites:dataflow:py39:postCommitIT FAILED FAILURE: Build completed with 2 failures. 1: Task failed with an exception. ----------- * Where: Script '<https://ci-beam.apache.org/job/beam_PostCommit_Python39/ws/src/sdks/python/test-suites/direct/common.gradle'> line: 52 * What went wrong: Execution failed for task ':sdks:python:test-suites:direct:py39:postCommitIT'. > Process 'command 'sh'' finished with non-zero exit value 1 * Try: > Run with --stacktrace option to get the stack trace. > Run with --info or --debug option to get more log output. > Get more help at https://help.gradle.org. ============================================================================== 2: Task failed with an exception. ----------- * Where: Script '<https://ci-beam.apache.org/job/beam_PostCommit_Python39/ws/src/sdks/python/test-suites/dataflow/common.gradle'> line: 139 * What went wrong: Execution failed for task ':sdks:python:test-suites:dataflow:py39:postCommitIT'. > Process 'command 'sh'' finished with non-zero exit value 1 * Try: > Run with --stacktrace option to get the stack trace. > Run with --info or --debug option to get more log output. > Get more help at https://help.gradle.org. ============================================================================== Deprecated Gradle features were used in this build, making it incompatible with Gradle 9.0. You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins. For more on this, please refer to https://docs.gradle.org/8.4/userguide/command_line_interface.html#sec:command_line_warnings in the Gradle documentation. BUILD FAILED in 1h 42m 5s 222 actionable tasks: 158 executed, 60 from cache, 4 up-to-date Publishing build scan... https://ge.apache.org/s/3drzg7ginkvdq Build step 'Invoke Gradle script' changed build result to FAILURE Build step 'Invoke Gradle script' marked build as failure --------------------------------------------------------------------- To unsubscribe, e-mail: builds-unsubscr...@beam.apache.org For additional commands, e-mail: builds-h...@beam.apache.org