See <https://builds.apache.org/job/beam_PostCommit_Python_Verify/2102/display/redirect?page=changes>
Changes: [lcwik] [BEAM-1871] Move GCP specific serialization CloudObject and supporting ------------------------------------------ [...truncated 585.04 KB...] { "kind": "ParallelDo", "name": "s13", "properties": { "display_data": [ { "key": "fn", "label": "Transform Function", "namespace": "apache_beam.transforms.core.CallableWrapperDoFn", "type": "STRING", "value": "_merge_tagged_vals_under_key" }, { "key": "fn", "label": "Transform Function", "namespace": "apache_beam.transforms.core.ParDo", "shortValue": "CallableWrapperDoFn", "type": "STRING", "value": "apache_beam.transforms.core.CallableWrapperDoFn" } ], "non_parallel_inputs": {}, "output_info": [ { "encoding": { "@type": "kind:windowed_value", "component_encodings": [ { "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/", "component_encodings": [ { "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/", "component_encodings": [] }, { "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/", "component_encodings": [] } ], "is_pair_like": true }, { "@type": "kind:global_window" } ], "is_wrapper": true }, "output_name": "out", "user_name": "assert_that/Group/Map(_merge_tagged_vals_under_key).out" } ], "parallel_input": { "@type": "OutputReference", "output_name": "out", "step_name": "s12" }, "serialized_fn": "<string of 1332 bytes>", "user_name": "assert_that/Group/Map(_merge_tagged_vals_under_key)" } }, { "kind": "ParallelDo", "name": "s14", "properties": { "display_data": [ { "key": "fn", "label": "Transform Function", "namespace": "apache_beam.transforms.core.CallableWrapperDoFn", "type": "STRING", "value": "<lambda>" }, { "key": "fn", "label": "Transform Function", "namespace": "apache_beam.transforms.core.ParDo", "shortValue": "CallableWrapperDoFn", "type": "STRING", "value": "apache_beam.transforms.core.CallableWrapperDoFn" } ], "non_parallel_inputs": {}, "output_info": [ { "encoding": { "@type": "kind:windowed_value", "component_encodings": [ { "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/", "component_encodings": [ { "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/", "component_encodings": [] }, { "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/", "component_encodings": [] } ], "is_pair_like": true }, { "@type": "kind:global_window" } ], "is_wrapper": true }, "output_name": "out", "user_name": "assert_that/Unkey.out" } ], "parallel_input": { "@type": "OutputReference", "output_name": "out", "step_name": "s13" }, "serialized_fn": "<string of 956 bytes>", "user_name": "assert_that/Unkey" } }, { "kind": "ParallelDo", "name": "s15", "properties": { "display_data": [ { "key": "fn", "label": "Transform Function", "namespace": "apache_beam.transforms.core.CallableWrapperDoFn", "type": "STRING", "value": "match" }, { "key": "fn", "label": "Transform Function", "namespace": "apache_beam.transforms.core.ParDo", "shortValue": "CallableWrapperDoFn", "type": "STRING", "value": "apache_beam.transforms.core.CallableWrapperDoFn" } ], "non_parallel_inputs": {}, "output_info": [ { "encoding": { "@type": "kind:windowed_value", "component_encodings": [ { "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/", "component_encodings": [ { "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/", "component_encodings": [] }, { "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/", "component_encodings": [] } ], "is_pair_like": true }, { "@type": "kind:global_window" } ], "is_wrapper": true }, "output_name": "out", "user_name": "assert_that/Match.out" } ], "parallel_input": { "@type": "OutputReference", "output_name": "out", "step_name": "s14" }, "serialized_fn": "<string of 1104 bytes>", "user_name": "assert_that/Match" } } ], "type": "JOB_TYPE_BATCH" } root: DEBUG: Response returned status 429, retrying root: DEBUG: Retrying request to url https://dataflow.googleapis.com/v1b3/projects/apache-beam-testing/locations/us-central1/jobs?alt=json after exception HttpError accessing <https://dataflow.googleapis.com/v1b3/projects/apache-beam-testing/locations/us-central1/jobs?alt=json>: response: <{'status': '429', 'content-length': '441', 'x-xss-protection': '1; mode=block', 'x-content-type-options': 'nosniff', 'transfer-encoding': 'chunked', 'vary': 'Origin, X-Origin, Referer', 'server': 'ESF', '-content-encoding': 'gzip', 'cache-control': 'private', 'date': 'Wed, 03 May 2017 21:06:08 GMT', 'x-frame-options': 'SAMEORIGIN', 'content-type': 'application/json; charset=UTF-8'}>, content <{ "error": { "code": 429, "message": "(a0317cd51e7799a8): The workflow could not be created. Causes: (896bdf8c7556dd88): Too many running jobs. Project apache-beam-testing is running 25 jobs and project limit for active jobs is 25. To fix this, cancel an existing workflow via the UI, wait for a workflow to finish or contact dataflow-feedb...@google.com to request an increase in quota.", "status": "RESOURCE_EXHAUSTED" } } > root: DEBUG: Response returned status 429, retrying root: DEBUG: Retrying request to url https://dataflow.googleapis.com/v1b3/projects/apache-beam-testing/locations/us-central1/jobs?alt=json after exception HttpError accessing <https://dataflow.googleapis.com/v1b3/projects/apache-beam-testing/locations/us-central1/jobs?alt=json>: response: <{'status': '429', 'content-length': '441', 'x-xss-protection': '1; mode=block', 'x-content-type-options': 'nosniff', 'transfer-encoding': 'chunked', 'vary': 'Origin, X-Origin, Referer', 'server': 'ESF', '-content-encoding': 'gzip', 'cache-control': 'private', 'date': 'Wed, 03 May 2017 21:06:10 GMT', 'x-frame-options': 'SAMEORIGIN', 'content-type': 'application/json; charset=UTF-8'}>, content <{ "error": { "code": 429, "message": "(bde39265c3ade645): The workflow could not be created. Causes: (58a37b1aae7522da): Too many running jobs. Project apache-beam-testing is running 25 jobs and project limit for active jobs is 25. To fix this, cancel an existing workflow via the UI, wait for a workflow to finish or contact dataflow-feedb...@google.com to request an increase in quota.", "status": "RESOURCE_EXHAUSTED" } } > root: DEBUG: Response returned status 429, retrying root: DEBUG: Retrying request to url https://dataflow.googleapis.com/v1b3/projects/apache-beam-testing/locations/us-central1/jobs?alt=json after exception HttpError accessing <https://dataflow.googleapis.com/v1b3/projects/apache-beam-testing/locations/us-central1/jobs?alt=json>: response: <{'status': '429', 'content-length': '441', 'x-xss-protection': '1; mode=block', 'x-content-type-options': 'nosniff', 'transfer-encoding': 'chunked', 'vary': 'Origin, X-Origin, Referer', 'server': 'ESF', '-content-encoding': 'gzip', 'cache-control': 'private', 'date': 'Wed, 03 May 2017 21:06:14 GMT', 'x-frame-options': 'SAMEORIGIN', 'content-type': 'application/json; charset=UTF-8'}>, content <{ "error": { "code": 429, "message": "(5f6f73707d90ce57): The workflow could not be created. Causes: (5eb5e05ad5b71fb8): Too many running jobs. Project apache-beam-testing is running 25 jobs and project limit for active jobs is 25. To fix this, cancel an existing workflow via the UI, wait for a workflow to finish or contact dataflow-feedb...@google.com to request an increase in quota.", "status": "RESOURCE_EXHAUSTED" } } > root: DEBUG: Response returned status 429, retrying root: DEBUG: Retrying request to url https://dataflow.googleapis.com/v1b3/projects/apache-beam-testing/locations/us-central1/jobs?alt=json after exception HttpError accessing <https://dataflow.googleapis.com/v1b3/projects/apache-beam-testing/locations/us-central1/jobs?alt=json>: response: <{'status': '429', 'content-length': '441', 'x-xss-protection': '1; mode=block', 'x-content-type-options': 'nosniff', 'transfer-encoding': 'chunked', 'vary': 'Origin, X-Origin, Referer', 'server': 'ESF', '-content-encoding': 'gzip', 'cache-control': 'private', 'date': 'Wed, 03 May 2017 21:06:24 GMT', 'x-frame-options': 'SAMEORIGIN', 'content-type': 'application/json; charset=UTF-8'}>, content <{ "error": { "code": 429, "message": "(250485361cede6a4): The workflow could not be created. Causes: (3d51f9ce06f8f449): Too many running jobs. Project apache-beam-testing is running 25 jobs and project limit for active jobs is 25. To fix this, cancel an existing workflow via the UI, wait for a workflow to finish or contact dataflow-feedb...@google.com to request an increase in quota.", "status": "RESOURCE_EXHAUSTED" } } > root: ERROR: HTTP status 429 trying to create job at dataflow service endpoint https://dataflow.googleapis.com root: CRITICAL: details of server error: HttpError accessing <https://dataflow.googleapis.com/v1b3/projects/apache-beam-testing/locations/us-central1/jobs?alt=json>: response: <{'status': '429', 'content-length': '441', 'x-xss-protection': '1; mode=block', 'x-content-type-options': 'nosniff', 'transfer-encoding': 'chunked', 'vary': 'Origin, X-Origin, Referer', 'server': 'ESF', '-content-encoding': 'gzip', 'cache-control': 'private', 'date': 'Wed, 03 May 2017 21:06:40 GMT', 'x-frame-options': 'SAMEORIGIN', 'content-type': 'application/json; charset=UTF-8'}>, content <{ "error": { "code": 429, "message": "(cafa8d725cdb5c05): The workflow could not be created. Causes: (122d3cce05f43e61): Too many running jobs. Project apache-beam-testing is running 25 jobs and project limit for active jobs is 25. To fix this, cancel an existing workflow via the UI, wait for a workflow to finish or contact dataflow-feedb...@google.com to request an increase in quota.", "status": "RESOURCE_EXHAUSTED" } } > --------------------- >> end captured logging << --------------------- ---------------------------------------------------------------------- Ran 15 tests in 1459.553s FAILED (errors=1) Found: https://console.cloud.google.com/dataflow/job/2017-05-03_13_53_59-6274220879832354093?project=apache-beam-testing Found: https://console.cloud.google.com/dataflow/job/2017-05-03_14_00_09-2115043220571521373?project=apache-beam-testing Found: https://console.cloud.google.com/dataflow/job/2017-05-03_14_06_46-16905698005168122708?project=apache-beam-testing Found: https://console.cloud.google.com/dataflow/job/2017-05-03_13_54_00-8390210282497876890?project=apache-beam-testing Found: https://console.cloud.google.com/dataflow/job/2017-05-03_13_59_45-17788673529223745000?project=apache-beam-testing Found: https://console.cloud.google.com/dataflow/job/2017-05-03_14_05_46-14034928885123604810?project=apache-beam-testing Found: https://console.cloud.google.com/dataflow/job/2017-05-03_14_11_21-7072714885907982823?project=apache-beam-testing Found: https://console.cloud.google.com/dataflow/job/2017-05-03_13_54_00-6119317239514633671?project=apache-beam-testing Found: https://console.cloud.google.com/dataflow/job/2017-05-03_14_01_06-11170993981454170247?project=apache-beam-testing Found: https://console.cloud.google.com/dataflow/job/2017-05-03_14_07_25-12220563990235853812?project=apache-beam-testing Found: https://console.cloud.google.com/dataflow/job/2017-05-03_13_53_59-9526256855014008237?project=apache-beam-testing Found: https://console.cloud.google.com/dataflow/job/2017-05-03_13_59_34-7356116023296377244?project=apache-beam-testing Found: https://console.cloud.google.com/dataflow/job/2017-05-03_14_06_55-2914167711325786900?project=apache-beam-testing Found: https://console.cloud.google.com/dataflow/job/2017-05-03_14_12_14-4746701238801138816?project=apache-beam-testing Build step 'Execute shell' marked build as failure