This is an automated email from the ASF dual-hosted git repository. chesnay pushed a commit to branch experiment_gha_docs in repository https://gitbox.apache.org/repos/asf/flink.git
commit 89cd9dcb583ae80f3ba89e972dfabbe061004592 Author: Chesnay Schepler <ches...@apache.org> AuthorDate: Mon Mar 14 12:21:59 2022 +0100 disable azure --- azure-pipelines.yml | 98 ----------------------------------------------------- 1 file changed, 98 deletions(-) diff --git a/azure-pipelines.yml b/azure-pipelines.yml deleted file mode 100644 index 649643f..0000000 --- a/azure-pipelines.yml +++ /dev/null @@ -1,98 +0,0 @@ -# Licensed to the Apache Software Foundation (ASF) under one or more -# contributor license agreements. See the NOTICE file distributed with -# this work for additional information regarding copyright ownership. -# The ASF licenses this file to You under the Apache License, Version 2.0 -# (the "License"); you may not use this file except in compliance with -# the License. You may obtain a copy of the License at -# -# http://www.apache.org/licenses/LICENSE-2.0 -# -# Unless required by applicable law or agreed to in writing, software -# distributed under the License is distributed on an "AS IS" BASIS, -# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -# See the License for the specific language governing permissions and -# limitations under the License. - -# -# This file defines an Azure Pipeline build for testing Flink. It is intended to be used -# with a free Azure Pipelines account. -# It has the following features: -# - default builds for pushes / pull requests -# - end-to-end tests -# -# -# For the "apache/flink" repository, we are using the pipeline definition located in -# tools/azure-pipelines/build-apache-repo.yml -# That file points to custom, self-hosted build agents for faster pull request build processing and -# integration with Flinkbot. -# The custom pipeline definition file is configured in the "Pipeline settings" screen -# of the Azure Pipelines web ui. -# - -trigger: - branches: - include: - - '*' # must quote since "*" is a YAML reserved character; we want a string - -resources: - containers: - # Container with Maven 3.2.5, SSL to have the same environment everywhere. - # see https://github.com/flink-ci/flink-ci-docker - - container: flink-build-container - image: chesnay/flink-ci:java_8_11_17 - # On AZP provided machines, set this flag to allow writing coredumps in docker - options: --privileged - -# Define variables: -# - See tools/azure-pipelines/jobs-template.yml for a short summary of the caching -# - See https://stackoverflow.com/questions/60742105/how-can-i-access-a-secret-value-from-an-azure-pipelines-expression -# to understand why the secrets are handled like this -variables: - MAVEN_CACHE_FOLDER: $(Pipeline.Workspace)/.m2/repository - E2E_CACHE_FOLDER: $(Pipeline.Workspace)/e2e_cache - E2E_TARBALL_CACHE: $(Pipeline.Workspace)/e2e_artifact_cache - MAVEN_ARGS: '-Dmaven.repo.local=$(MAVEN_CACHE_FOLDER)' - CACHE_KEY: maven | $(Agent.OS) | **/pom.xml, !**/target/** - CACHE_FALLBACK_KEY: maven | $(Agent.OS) - DOCKER_IMAGES_CACHE_KEY: docker-images-cache | $(Agent.OS) | **/cache_docker_images.sh | flink-test-utils-parent/**/DockerImageVersions.java - DOCKER_IMAGES_CACHE_FOLDER: $(Pipeline.Workspace)/.docker-cache - FLINK_ARTIFACT_DIR: $(Pipeline.Workspace)/flink_artifact - SECRET_S3_BUCKET: $[variables.IT_CASE_S3_BUCKET] - SECRET_S3_ACCESS_KEY: $[variables.IT_CASE_S3_ACCESS_KEY] - SECRET_S3_SECRET_KEY: $[variables.IT_CASE_S3_SECRET_KEY] - SECRET_GLUE_SCHEMA_ACCESS_KEY: $[variables.IT_CASE_GLUE_SCHEMA_ACCESS_KEY] - SECRET_GLUE_SCHEMA_SECRET_KEY: $[variables.IT_CASE_GLUE_SCHEMA_SECRET_KEY] - - -stages: - # CI / PR triggered stage: - - stage: ci - displayName: "CI build (custom builders)" - condition: not(eq(variables['MODE'], 'release')) - jobs: - - template: tools/azure-pipelines/jobs-template.yml - parameters: # see template file for a definition of the parameters. - stage_name: ci_build - test_pool_definition: - vmImage: 'ubuntu-20.04' - e2e_pool_definition: - vmImage: 'ubuntu-20.04' - environment: PROFILE="-Dhadoop.version=2.8.5 -Dinclude_hadoop_aws -Dscala-2.12" - run_end_to_end: false - container: flink-build-container - jdk: 8 - - job: docs_404_check # run on a MSFT provided machine - pool: - vmImage: 'ubuntu-20.04' - steps: - - script: ./tools/ci/docs.sh - # CI / Special stage for release, e.g. building PyFlink wheel packages, etc: - - stage: ci_release - displayName: "CI build (release)" - condition: and(eq(variables['Build.Reason'], 'Manual'), eq(variables['MODE'], 'release')) - jobs: - - template: tools/azure-pipelines/build-python-wheels.yml - parameters: - stage_name: cron_python_wheels - environment: PROFILE="-Dhadoop.version=2.8.5 -Dinclude_hadoop_aws -Dscala-2.12" - container: flink-build-container