See <https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/1121/display/redirect?page=changes>
Changes: [lukasz.gajowy] [BEAM-3798] Remove error check on dataflow when getting batch job state ------------------------------------------ [...truncated 124.24 KB...] File was already downloaded /tmp/dataflow-requirements-cache/six-1.11.0.tar.gz test_as_dict_twice (apache_beam.transforms.sideinputs_test.SideInputsTest) ... ok Collecting funcsigs>=1 (from mock->-r postcommit_requirements.txt (line 2)) File was already downloaded /tmp/dataflow-requirements-cache/funcsigs-1.0.2.tar.gz Collecting pbr>=0.11 (from mock->-r postcommit_requirements.txt (line 2)) File was already downloaded /tmp/dataflow-requirements-cache/pbr-3.1.1.tar.gz Successfully downloaded pyhamcrest mock setuptools six funcsigs pbr Collecting pyhamcrest (from -r postcommit_requirements.txt (line 1)) File was already downloaded /tmp/dataflow-requirements-cache/PyHamcrest-1.9.0.tar.gz Collecting mock (from -r postcommit_requirements.txt (line 2)) File was already downloaded /tmp/dataflow-requirements-cache/mock-2.0.0.tar.gz test_as_list_twice (apache_beam.transforms.sideinputs_test.SideInputsTest) ... ok Collecting setuptools (from pyhamcrest->-r postcommit_requirements.txt (line 1)) File was already downloaded /tmp/dataflow-requirements-cache/setuptools-38.6.0.zip Collecting six (from pyhamcrest->-r postcommit_requirements.txt (line 1)) File was already downloaded /tmp/dataflow-requirements-cache/six-1.11.0.tar.gz Collecting funcsigs>=1 (from mock->-r postcommit_requirements.txt (line 2)) File was already downloaded /tmp/dataflow-requirements-cache/funcsigs-1.0.2.tar.gz Collecting pbr>=0.11 (from mock->-r postcommit_requirements.txt (line 2)) File was already downloaded /tmp/dataflow-requirements-cache/pbr-3.1.1.tar.gz Collecting pyhamcrest (from -r postcommit_requirements.txt (line 1)) File was already downloaded /tmp/dataflow-requirements-cache/PyHamcrest-1.9.0.tar.gz Successfully downloaded pyhamcrest mock setuptools six funcsigs pbr Collecting mock (from -r postcommit_requirements.txt (line 2)) File was already downloaded /tmp/dataflow-requirements-cache/mock-2.0.0.tar.gz Collecting setuptools (from pyhamcrest->-r postcommit_requirements.txt (line 1)) File was already downloaded /tmp/dataflow-requirements-cache/setuptools-38.6.0.zip Collecting six (from pyhamcrest->-r postcommit_requirements.txt (line 1)) File was already downloaded /tmp/dataflow-requirements-cache/six-1.11.0.tar.gz Collecting funcsigs>=1 (from mock->-r postcommit_requirements.txt (line 2)) File was already downloaded /tmp/dataflow-requirements-cache/funcsigs-1.0.2.tar.gz Collecting pbr>=0.11 (from mock->-r postcommit_requirements.txt (line 2)) File was already downloaded /tmp/dataflow-requirements-cache/pbr-3.1.1.tar.gz Successfully downloaded pyhamcrest mock setuptools six funcsigs pbr test_as_singleton_with_different_defaults (apache_beam.transforms.sideinputs_test.SideInputsTest) ... ok Collecting pyhamcrest (from -r postcommit_requirements.txt (line 1)) File was already downloaded /tmp/dataflow-requirements-cache/PyHamcrest-1.9.0.tar.gz Collecting mock (from -r postcommit_requirements.txt (line 2)) File was already downloaded /tmp/dataflow-requirements-cache/mock-2.0.0.tar.gz Collecting setuptools (from pyhamcrest->-r postcommit_requirements.txt (line 1)) File was already downloaded /tmp/dataflow-requirements-cache/setuptools-38.6.0.zip Collecting six (from pyhamcrest->-r postcommit_requirements.txt (line 1)) File was already downloaded /tmp/dataflow-requirements-cache/six-1.11.0.tar.gz Collecting funcsigs>=1 (from mock->-r postcommit_requirements.txt (line 2)) File was already downloaded /tmp/dataflow-requirements-cache/funcsigs-1.0.2.tar.gz Collecting pbr>=0.11 (from mock->-r postcommit_requirements.txt (line 2)) File was already downloaded /tmp/dataflow-requirements-cache/pbr-3.1.1.tar.gz Successfully downloaded pyhamcrest mock setuptools six funcsigs pbr test_as_singleton_without_unique_labels (apache_beam.transforms.sideinputs_test.SideInputsTest) ... ok Collecting pyhamcrest (from -r postcommit_requirements.txt (line 1)) <https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/src/sdks/python/local/lib/python2.7/site-packages/pip/_vendor/requests/packages/urllib3/util/ssl_.py>:318: SNIMissingWarning: An HTTPS request has been made, but the SNI (Subject Name Indication) extension to TLS is not available on this platform. This may cause the server to present an incorrect TLS certificate, which can cause validation failures. You can upgrade to a newer version of Python to solve this. For more information, see https://urllib3.readthedocs.io/en/latest/security.html#snimissingwarning. SNIMissingWarning <https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/src/sdks/python/local/lib/python2.7/site-packages/pip/_vendor/requests/packages/urllib3/util/ssl_.py>:122: InsecurePlatformWarning: A true SSLContext object is not available. This prevents urllib3 from configuring SSL appropriately and may cause certain SSL connections to fail. You can upgrade to a newer version of Python to solve this. For more information, see https://urllib3.readthedocs.io/en/latest/security.html#insecureplatformwarning. InsecurePlatformWarning File was already downloaded /tmp/dataflow-requirements-cache/PyHamcrest-1.9.0.tar.gz Collecting mock (from -r postcommit_requirements.txt (line 2)) File was already downloaded /tmp/dataflow-requirements-cache/mock-2.0.0.tar.gz Collecting setuptools (from pyhamcrest->-r postcommit_requirements.txt (line 1)) File was already downloaded /tmp/dataflow-requirements-cache/setuptools-38.6.0.zip Collecting six (from pyhamcrest->-r postcommit_requirements.txt (line 1)) File was already downloaded /tmp/dataflow-requirements-cache/six-1.11.0.tar.gz Collecting funcsigs>=1 (from mock->-r postcommit_requirements.txt (line 2)) File was already downloaded /tmp/dataflow-requirements-cache/funcsigs-1.0.2.tar.gz Collecting pbr>=0.11 (from mock->-r postcommit_requirements.txt (line 2)) File was already downloaded /tmp/dataflow-requirements-cache/pbr-3.1.1.tar.gz Successfully downloaded pyhamcrest mock setuptools six funcsigs pbr test_default_value_singleton_side_input (apache_beam.transforms.sideinputs_test.SideInputsTest) ... ok Collecting pyhamcrest (from -r postcommit_requirements.txt (line 1)) File was already downloaded /tmp/dataflow-requirements-cache/PyHamcrest-1.9.0.tar.gz Collecting mock (from -r postcommit_requirements.txt (line 2)) File was already downloaded /tmp/dataflow-requirements-cache/mock-2.0.0.tar.gz Collecting setuptools (from pyhamcrest->-r postcommit_requirements.txt (line 1)) File was already downloaded /tmp/dataflow-requirements-cache/setuptools-38.6.0.zip Collecting six (from pyhamcrest->-r postcommit_requirements.txt (line 1)) File was already downloaded /tmp/dataflow-requirements-cache/six-1.11.0.tar.gz Collecting funcsigs>=1 (from mock->-r postcommit_requirements.txt (line 2)) File was already downloaded /tmp/dataflow-requirements-cache/funcsigs-1.0.2.tar.gz Collecting pbr>=0.11 (from mock->-r postcommit_requirements.txt (line 2)) File was already downloaded /tmp/dataflow-requirements-cache/pbr-3.1.1.tar.gz Successfully downloaded pyhamcrest mock setuptools six funcsigs pbr test_flattened_side_input (apache_beam.transforms.sideinputs_test.SideInputsTest) ... ok test_empty_singleton_side_input (apache_beam.transforms.sideinputs_test.SideInputsTest) ... ok test_iterable_side_input (apache_beam.transforms.sideinputs_test.SideInputsTest) ... ok test_multi_valued_singleton_side_input (apache_beam.transforms.sideinputs_test.SideInputsTest) ... ok ====================================================================== ERROR: test_multiple_empty_outputs (apache_beam.transforms.ptransform_test.PTransformTest) ---------------------------------------------------------------------- Traceback (most recent call last): File "<https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/src/sdks/python/apache_beam/transforms/ptransform_test.py",> line 271, in test_multiple_empty_outputs pipeline.run() File "<https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/src/sdks/python/apache_beam/testing/test_pipeline.py",> line 102, in run result = super(TestPipeline, self).run() File "<https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/src/sdks/python/apache_beam/pipeline.py",> line 369, in run self.to_runner_api(), self.runner, self._options).run(False) File "<https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/src/sdks/python/apache_beam/pipeline.py",> line 382, in run return self.runner.run_pipeline(self) File "<https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/src/sdks/python/apache_beam/runners/dataflow/test_dataflow_runner.py",> line 39, in run_pipeline self.result = super(TestDataflowRunner, self).run_pipeline(pipeline) File "<https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/src/sdks/python/apache_beam/runners/dataflow/dataflow_runner.py",> line 324, in run_pipeline self.dataflow_client.create_job(self.job), self) File "<https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/src/sdks/python/apache_beam/utils/retry.py",> line 180, in wrapper return fun(*args, **kwargs) File "<https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/src/sdks/python/apache_beam/runners/dataflow/internal/apiclient.py",> line 462, in create_job self.create_job_description(job) File "<https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/src/sdks/python/apache_beam/runners/dataflow/internal/apiclient.py",> line 492, in create_job_description job.options, file_copy=self._gcs_file_copy) File "<https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/src/sdks/python/apache_beam/runners/dataflow/internal/dependency.py",> line 328, in stage_job_resources setup_options.requirements_file, requirements_cache_path) File "<https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/src/sdks/python/apache_beam/runners/dataflow/internal/dependency.py",> line 262, in _populate_requirements_cache processes.check_call(cmd_args) File "<https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/src/sdks/python/apache_beam/utils/processes.py",> line 44, in check_call return subprocess.check_call(*args, **kwargs) File "/usr/lib/python2.7/subprocess.py", line 540, in check_call raise CalledProcessError(retcode, cmd) CalledProcessError: Command '['<https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/src/sdks/python/bin/python',> '-m', 'pip', 'download', '--dest', '/tmp/dataflow-requirements-cache', '-r', 'postcommit_requirements.txt', '--no-binary', ':all:']' returned non-zero exit status 2 -------------------- >> begin captured logging << -------------------- root: DEBUG: PValue computed by Some Numbers/Read (tag None): refcount: 1 => 0 root: DEBUG: PValue computed by ClassifyNumbers/FlatMap(<lambda at ptransform_test.py:265>) (tag None): refcount: 1 => 0 root: DEBUG: PValue computed by assert_that/WindowInto(WindowIntoFn) (tag None): refcount: 1 => 0 root: DEBUG: PValue computed by assert_that/Create/Read (tag None): refcount: 1 => 0 root: DEBUG: PValue computed by assert_that/ToVoidKey (tag None): refcount: 1 => 0 root: DEBUG: PValue computed by assert_that/Group/pair_with_0 (tag None): refcount: 1 => 0 root: DEBUG: PValue computed by assert_that/Group/pair_with_1 (tag None): refcount: 1 => 0 root: DEBUG: PValue computed by assert_that/Group/Flatten (tag None): refcount: 1 => 0 root: DEBUG: PValue computed by assert_that/Group/GroupByKey (tag None): refcount: 1 => 0 root: DEBUG: PValue computed by assert_that/Group/Map(_merge_tagged_vals_under_key) (tag None): refcount: 1 => 0 root: DEBUG: PValue computed by assert_that/Unkey (tag None): refcount: 1 => 0 root: DEBUG: PValue computed by ClassifyNumbers/FlatMap(<lambda at ptransform_test.py:265>) (tag odd): refcount: 1 => 0 root: DEBUG: PValue computed by assert:odd/WindowInto(WindowIntoFn) (tag None): refcount: 1 => 0 root: DEBUG: PValue computed by assert:odd/Create/Read (tag None): refcount: 1 => 0 root: DEBUG: PValue computed by assert:odd/ToVoidKey (tag None): refcount: 1 => 0 root: DEBUG: PValue computed by assert:odd/Group/pair_with_0 (tag None): refcount: 1 => 0 root: DEBUG: PValue computed by assert:odd/Group/pair_with_1 (tag None): refcount: 1 => 0 root: DEBUG: PValue computed by assert:odd/Group/Flatten (tag None): refcount: 1 => 0 root: DEBUG: PValue computed by assert:odd/Group/GroupByKey (tag None): refcount: 1 => 0 root: DEBUG: PValue computed by assert:odd/Group/Map(_merge_tagged_vals_under_key) (tag None): refcount: 1 => 0 root: DEBUG: PValue computed by assert:odd/Unkey (tag None): refcount: 1 => 0 root: DEBUG: PValue computed by ClassifyNumbers/FlatMap(<lambda at ptransform_test.py:265>) (tag even): refcount: 1 => 0 root: DEBUG: PValue computed by assert:even/WindowInto(WindowIntoFn) (tag None): refcount: 1 => 0 root: DEBUG: PValue computed by assert:even/Create/Read (tag None): refcount: 1 => 0 root: DEBUG: PValue computed by assert:even/ToVoidKey (tag None): refcount: 1 => 0 root: DEBUG: PValue computed by assert:even/Group/pair_with_0 (tag None): refcount: 1 => 0 root: DEBUG: PValue computed by assert:even/Group/pair_with_1 (tag None): refcount: 1 => 0 root: DEBUG: PValue computed by assert:even/Group/Flatten (tag None): refcount: 1 => 0 root: DEBUG: PValue computed by assert:even/Group/GroupByKey (tag None): refcount: 1 => 0 root: DEBUG: PValue computed by assert:even/Group/Map(_merge_tagged_vals_under_key) (tag None): refcount: 1 => 0 root: DEBUG: PValue computed by assert:even/Unkey (tag None): refcount: 1 => 0 root: DEBUG: Connecting using Google Application Default Credentials. root: INFO: Starting GCS upload to gs://temp-storage-for-end-to-end-tests/staging-validatesrunner-test/beamapp-jenkins-0316000509-356838.1521158709.357126/pipeline.pb... oauth2client.transport: INFO: Attempting refresh to obtain initial access_token root: INFO: Completed GCS upload to gs://temp-storage-for-end-to-end-tests/staging-validatesrunner-test/beamapp-jenkins-0316000509-356838.1521158709.357126/pipeline.pb root: INFO: Starting GCS upload to gs://temp-storage-for-end-to-end-tests/staging-validatesrunner-test/beamapp-jenkins-0316000509-356838.1521158709.357126/requirements.txt... root: INFO: Completed GCS upload to gs://temp-storage-for-end-to-end-tests/staging-validatesrunner-test/beamapp-jenkins-0316000509-356838.1521158709.357126/requirements.txt root: INFO: Executing command: ['<https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/src/sdks/python/bin/python',> '-m', 'pip', 'download', '--dest', '/tmp/dataflow-requirements-cache', '-r', 'postcommit_requirements.txt', '--no-binary', ':all:'] --------------------- >> end captured logging << --------------------- ====================================================================== ERROR: test_par_do_with_multiple_outputs_and_using_return (apache_beam.transforms.ptransform_test.PTransformTest) ---------------------------------------------------------------------- Traceback (most recent call last): File "<https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/src/sdks/python/apache_beam/transforms/ptransform_test.py",> line 244, in test_par_do_with_multiple_outputs_and_using_return pipeline.run() File "<https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/src/sdks/python/apache_beam/testing/test_pipeline.py",> line 102, in run result = super(TestPipeline, self).run() File "<https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/src/sdks/python/apache_beam/pipeline.py",> line 369, in run self.to_runner_api(), self.runner, self._options).run(False) File "<https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/src/sdks/python/apache_beam/pipeline.py",> line 382, in run return self.runner.run_pipeline(self) File "<https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/src/sdks/python/apache_beam/runners/dataflow/test_dataflow_runner.py",> line 39, in run_pipeline self.result = super(TestDataflowRunner, self).run_pipeline(pipeline) File "<https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/src/sdks/python/apache_beam/runners/dataflow/dataflow_runner.py",> line 324, in run_pipeline self.dataflow_client.create_job(self.job), self) File "<https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/src/sdks/python/apache_beam/utils/retry.py",> line 180, in wrapper return fun(*args, **kwargs) File "<https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/src/sdks/python/apache_beam/runners/dataflow/internal/apiclient.py",> line 462, in create_job self.create_job_description(job) File "<https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/src/sdks/python/apache_beam/runners/dataflow/internal/apiclient.py",> line 492, in create_job_description job.options, file_copy=self._gcs_file_copy) File "<https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/src/sdks/python/apache_beam/runners/dataflow/internal/dependency.py",> line 328, in stage_job_resources setup_options.requirements_file, requirements_cache_path) File "<https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/src/sdks/python/apache_beam/runners/dataflow/internal/dependency.py",> line 262, in _populate_requirements_cache processes.check_call(cmd_args) File "<https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/src/sdks/python/apache_beam/utils/processes.py",> line 44, in check_call return subprocess.check_call(*args, **kwargs) File "/usr/lib/python2.7/subprocess.py", line 540, in check_call raise CalledProcessError(retcode, cmd) CalledProcessError: Command '['<https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/src/sdks/python/bin/python',> '-m', 'pip', 'download', '--dest', '/tmp/dataflow-requirements-cache', '-r', 'postcommit_requirements.txt', '--no-binary', ':all:']' returned non-zero exit status 2 -------------------- >> begin captured logging << -------------------- root: DEBUG: PValue computed by Some Numbers/Read (tag None): refcount: 1 => 0 root: DEBUG: PValue computed by ClassifyNumbers/FlatMap(some_fn) (tag None): refcount: 1 => 0 root: DEBUG: PValue computed by assert_that/WindowInto(WindowIntoFn) (tag None): refcount: 1 => 0 root: DEBUG: PValue computed by assert_that/Create/Read (tag None): refcount: 1 => 0 root: DEBUG: PValue computed by assert_that/ToVoidKey (tag None): refcount: 1 => 0 root: DEBUG: PValue computed by assert_that/Group/pair_with_0 (tag None): refcount: 1 => 0 root: DEBUG: PValue computed by assert_that/Group/pair_with_1 (tag None): refcount: 1 => 0 root: DEBUG: PValue computed by assert_that/Group/Flatten (tag None): refcount: 1 => 0 root: DEBUG: PValue computed by assert_that/Group/GroupByKey (tag None): refcount: 1 => 0 root: DEBUG: PValue computed by assert_that/Group/Map(_merge_tagged_vals_under_key) (tag None): refcount: 1 => 0 root: DEBUG: PValue computed by assert_that/Unkey (tag None): refcount: 1 => 0 root: DEBUG: PValue computed by ClassifyNumbers/FlatMap(some_fn) (tag odd): refcount: 1 => 0 root: DEBUG: PValue computed by assert:odd/WindowInto(WindowIntoFn) (tag None): refcount: 1 => 0 root: DEBUG: PValue computed by assert:odd/Create/Read (tag None): refcount: 1 => 0 root: DEBUG: PValue computed by assert:odd/ToVoidKey (tag None): refcount: 1 => 0 root: DEBUG: PValue computed by assert:odd/Group/pair_with_0 (tag None): refcount: 1 => 0 root: DEBUG: PValue computed by assert:odd/Group/pair_with_1 (tag None): refcount: 1 => 0 root: DEBUG: PValue computed by assert:odd/Group/Flatten (tag None): refcount: 1 => 0 root: DEBUG: PValue computed by assert:odd/Group/GroupByKey (tag None): refcount: 1 => 0 root: DEBUG: PValue computed by assert:odd/Group/Map(_merge_tagged_vals_under_key) (tag None): refcount: 1 => 0 root: DEBUG: PValue computed by assert:odd/Unkey (tag None): refcount: 1 => 0 root: DEBUG: PValue computed by ClassifyNumbers/FlatMap(some_fn) (tag even): refcount: 1 => 0 root: DEBUG: PValue computed by assert:even/WindowInto(WindowIntoFn) (tag None): refcount: 1 => 0 root: DEBUG: PValue computed by assert:even/Create/Read (tag None): refcount: 1 => 0 root: DEBUG: PValue computed by assert:even/ToVoidKey (tag None): refcount: 1 => 0 root: DEBUG: PValue computed by assert:even/Group/pair_with_0 (tag None): refcount: 1 => 0 root: DEBUG: PValue computed by assert:even/Group/pair_with_1 (tag None): refcount: 1 => 0 root: DEBUG: PValue computed by assert:even/Group/Flatten (tag None): refcount: 1 => 0 root: DEBUG: PValue computed by assert:even/Group/GroupByKey (tag None): refcount: 1 => 0 root: DEBUG: PValue computed by assert:even/Group/Map(_merge_tagged_vals_under_key) (tag None): refcount: 1 => 0 root: DEBUG: PValue computed by assert:even/Unkey (tag None): refcount: 1 => 0 root: DEBUG: Connecting using Google Application Default Credentials. root: INFO: Starting GCS upload to gs://temp-storage-for-end-to-end-tests/staging-validatesrunner-test/beamapp-jenkins-0316000509-311140.1521158709.311431/pipeline.pb... oauth2client.transport: INFO: Attempting refresh to obtain initial access_token root: INFO: Completed GCS upload to gs://temp-storage-for-end-to-end-tests/staging-validatesrunner-test/beamapp-jenkins-0316000509-311140.1521158709.311431/pipeline.pb root: INFO: Starting GCS upload to gs://temp-storage-for-end-to-end-tests/staging-validatesrunner-test/beamapp-jenkins-0316000509-311140.1521158709.311431/requirements.txt... root: INFO: Completed GCS upload to gs://temp-storage-for-end-to-end-tests/staging-validatesrunner-test/beamapp-jenkins-0316000509-311140.1521158709.311431/requirements.txt root: INFO: Executing command: ['<https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/src/sdks/python/bin/python',> '-m', 'pip', 'download', '--dest', '/tmp/dataflow-requirements-cache', '-r', 'postcommit_requirements.txt', '--no-binary', ':all:'] --------------------- >> end captured logging << --------------------- ---------------------------------------------------------------------- Ran 16 tests in 1716.055s FAILED (errors=2) Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-03-15_17_05_22-6111824375320159776?project=apache-beam-testing Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-03-15_17_12_22-15130686219676834271?project=apache-beam-testing Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-03-15_17_19_57-6079551500235601098?project=apache-beam-testing Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-03-15_17_27_12-16664985073794012740?project=apache-beam-testing Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-03-15_17_05_20-4884419995845335871?project=apache-beam-testing Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-03-15_17_12_40-2887441947640232848?project=apache-beam-testing Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-03-15_17_19_59-14792912523926695189?project=apache-beam-testing Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-03-15_17_26_38-7810449162696663465?project=apache-beam-testing Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-03-15_17_05_20-6440177208692820242?project=apache-beam-testing Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-03-15_17_12_59-497946836065052167?project=apache-beam-testing Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-03-15_17_19_59-3383330772375850991?project=apache-beam-testing Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-03-15_17_05_24-6993913432603044962?project=apache-beam-testing Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-03-15_17_13_34-5442479741021901859?project=apache-beam-testing Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-03-15_17_20_18-5583325922203574871?project=apache-beam-testing Build step 'Execute shell' marked build as failure Not sending mail to unregistered user lukasz.gaj...@polidea.com