See <https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/432/display/redirect?page=changes>
Changes: [altay] Adding debug server to sdk worker to get threaddumps ------------------------------------------ [...truncated 1.07 MB...] "non_parallel_inputs": {}, "output_info": [ { "encoding": { "@type": "kind:windowed_value", "component_encodings": [ { "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/", "component_encodings": [ { "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/", "component_encodings": [] }, { "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/", "component_encodings": [] } ], "is_pair_like": true }, { "@type": "kind:global_window" } ], "is_wrapper": true }, "output_name": "out", "user_name": "assert_that/Group/Map(_merge_tagged_vals_under_key).out" } ], "parallel_input": { "@type": "OutputReference", "output_name": "out", "step_name": "s13" }, "serialized_fn": "<string of 1380 bytes>", "user_name": "assert_that/Group/Map(_merge_tagged_vals_under_key)" } }, { "kind": "ParallelDo", "name": "s15", "properties": { "display_data": [ { "key": "fn", "label": "Transform Function", "namespace": "apache_beam.transforms.core.CallableWrapperDoFn", "type": "STRING", "value": "<lambda>" }, { "key": "fn", "label": "Transform Function", "namespace": "apache_beam.transforms.core.ParDo", "shortValue": "CallableWrapperDoFn", "type": "STRING", "value": "apache_beam.transforms.core.CallableWrapperDoFn" } ], "non_parallel_inputs": {}, "output_info": [ { "encoding": { "@type": "kind:windowed_value", "component_encodings": [ { "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/", "component_encodings": [ { "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/", "component_encodings": [] }, { "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/", "component_encodings": [] } ], "is_pair_like": true }, { "@type": "kind:global_window" } ], "is_wrapper": true }, "output_name": "out", "user_name": "assert_that/Unkey.out" } ], "parallel_input": { "@type": "OutputReference", "output_name": "out", "step_name": "s14" }, "serialized_fn": "<string of 980 bytes>", "user_name": "assert_that/Unkey" } }, { "kind": "ParallelDo", "name": "s16", "properties": { "display_data": [ { "key": "fn", "label": "Transform Function", "namespace": "apache_beam.transforms.core.CallableWrapperDoFn", "type": "STRING", "value": "_equal" }, { "key": "fn", "label": "Transform Function", "namespace": "apache_beam.transforms.core.ParDo", "shortValue": "CallableWrapperDoFn", "type": "STRING", "value": "apache_beam.transforms.core.CallableWrapperDoFn" } ], "non_parallel_inputs": {}, "output_info": [ { "encoding": { "@type": "kind:windowed_value", "component_encodings": [ { "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/", "component_encodings": [ { "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/", "component_encodings": [] }, { "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/", "component_encodings": [] } ], "is_pair_like": true }, { "@type": "kind:global_window" } ], "is_wrapper": true }, "output_name": "out", "user_name": "assert_that/Match.out" } ], "parallel_input": { "@type": "OutputReference", "output_name": "out", "step_name": "s15" }, "serialized_fn": "<string of 1160 bytes>", "user_name": "assert_that/Match" } } ], "type": "JOB_TYPE_BATCH" } root: INFO: Create job: <Job createTime: u'2017-12-11T18:18:20.790421Z' currentStateTime: u'1970-01-01T00:00:00Z' id: u'2017-12-11_10_18_20-331615697938870470' location: u'us-central1' name: u'beamapp-jenkins-1211181759-019285' projectId: u'apache-beam-testing' stageStates: [] steps: [] tempFiles: [] type: TypeValueValuesEnum(JOB_TYPE_BATCH, 1)> root: INFO: Created job with id: [2017-12-11_10_18_20-331615697938870470] root: INFO: To access the Dataflow monitoring console, please navigate to https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2017-12-11_10_18_20-331615697938870470?project=apache-beam-testing root: INFO: Job 2017-12-11_10_18_20-331615697938870470 is in state JOB_STATE_PENDING root: INFO: 2017-12-11T18:18:20.085Z: JOB_MESSAGE_WARNING: (49a22b86596fed2): Setting the number of workers (1) disables autoscaling for this job. If you are trying to cap autoscaling, consider only setting max_num_workers. If you want to disable autoscaling altogether, the documented way is to explicitly use autoscalingAlgorithm=NONE. root: INFO: 2017-12-11T18:18:22.765Z: JOB_MESSAGE_DETAILED: (a72caaa2580bb22a): Checking required Cloud APIs are enabled. root: INFO: 2017-12-11T18:18:25.547Z: JOB_MESSAGE_DETAILED: (a72caaa2580bbc42): Expanding CoGroupByKey operations into optimizable parts. root: INFO: 2017-12-11T18:18:25.570Z: JOB_MESSAGE_DEBUG: (a72caaa2580bb603): Combiner lifting skipped for step assert_that/Group/GroupByKey: GroupByKey not followed by a combiner. root: INFO: 2017-12-11T18:18:25.586Z: JOB_MESSAGE_DETAILED: (a72caaa2580bb1d9): Expanding GroupByKey operations into optimizable parts. root: INFO: 2017-12-11T18:18:25.606Z: JOB_MESSAGE_DETAILED: (a72caaa2580bbdaf): Lifting ValueCombiningMappingFns into MergeBucketsMappingFns root: INFO: 2017-12-11T18:18:25.632Z: JOB_MESSAGE_DEBUG: (a72caaa2580bb55b): Annotating graph with Autotuner information. root: INFO: 2017-12-11T18:18:25.693Z: JOB_MESSAGE_DETAILED: (a72caaa2580bbd07): Fusing adjacent ParDo, Read, Write, and Flatten operations root: INFO: 2017-12-11T18:18:25.722Z: JOB_MESSAGE_DETAILED: (a72caaa2580bb8dd): Unzipping flatten s12 for input s10.out root: INFO: 2017-12-11T18:18:25.756Z: JOB_MESSAGE_DETAILED: (a72caaa2580bb4b3): Fusing unzipped copy of assert_that/Group/GroupByKey/Reify, through flatten assert_that/Group/Flatten, into producer assert_that/Group/pair_with_0 root: INFO: 2017-12-11T18:18:25.783Z: JOB_MESSAGE_DETAILED: (a72caaa2580bb089): Fusing consumer assert_that/Group/Map(_merge_tagged_vals_under_key) into assert_that/Group/GroupByKey/GroupByWindow root: INFO: 2017-12-11T18:18:25.803Z: JOB_MESSAGE_DETAILED: (a72caaa2580bbc5f): Fusing consumer assert_that/Unkey into assert_that/Group/Map(_merge_tagged_vals_under_key) root: INFO: 2017-12-11T18:18:25.822Z: JOB_MESSAGE_DETAILED: (a72caaa2580bb835): Fusing consumer assert_that/Group/GroupByKey/GroupByWindow into assert_that/Group/GroupByKey/Read root: INFO: 2017-12-11T18:18:25.849Z: JOB_MESSAGE_DETAILED: (a72caaa2580bb40b): Fusing consumer assert_that/Match into assert_that/Unkey root: INFO: 2017-12-11T18:18:25.871Z: JOB_MESSAGE_DETAILED: (a72caaa2580bbfe1): Unzipping flatten s12-u13 for input s13-reify-value0-c11 root: INFO: 2017-12-11T18:18:25.894Z: JOB_MESSAGE_DETAILED: (a72caaa2580bbbb7): Fusing unzipped copy of assert_that/Group/GroupByKey/Write, through flatten s12-u13, into producer assert_that/Group/GroupByKey/Reify root: INFO: 2017-12-11T18:18:25.911Z: JOB_MESSAGE_DETAILED: (a72caaa2580bb78d): Fusing consumer assert_that/Group/pair_with_0 into assert_that/Create/Read root: INFO: 2017-12-11T18:18:25.932Z: JOB_MESSAGE_DETAILED: (a72caaa2580bb363): Fusing consumer assert_that/Group/GroupByKey/Write into assert_that/Group/GroupByKey/Reify root: INFO: 2017-12-11T18:18:25.952Z: JOB_MESSAGE_DETAILED: (a72caaa2580bbf39): Fusing consumer assert_that/Group/GroupByKey/Reify into assert_that/Group/pair_with_1 root: INFO: 2017-12-11T18:18:25.972Z: JOB_MESSAGE_DETAILED: (a72caaa2580bbb0f): Fusing consumer FlatMap(<lambda at sideinputs_test.py:302>)/FlatMap(<lambda at sideinputs_test.py:302>) into main input/Read root: INFO: 2017-12-11T18:18:25.991Z: JOB_MESSAGE_DETAILED: (a72caaa2580bb6e5): Fusing consumer assert_that/Group/pair_with_1 into assert_that/ToVoidKey root: INFO: 2017-12-11T18:18:26.015Z: JOB_MESSAGE_DETAILED: (a72caaa2580bb2bb): Fusing consumer assert_that/WindowInto(WindowIntoFn) into FlatMap(<lambda at sideinputs_test.py:302>)/FlatMap(<lambda at sideinputs_test.py:302>) root: INFO: 2017-12-11T18:18:26.033Z: JOB_MESSAGE_DETAILED: (a72caaa2580bbe91): Fusing consumer assert_that/ToVoidKey into assert_that/WindowInto(WindowIntoFn) root: INFO: 2017-12-11T18:18:26.054Z: JOB_MESSAGE_DEBUG: (a72caaa2580bba67): Workflow config is missing a default resource spec. root: INFO: 2017-12-11T18:18:26.074Z: JOB_MESSAGE_DEBUG: (a72caaa2580bb63d): Adding StepResource setup and teardown to workflow graph. root: INFO: 2017-12-11T18:18:26.099Z: JOB_MESSAGE_DEBUG: (a72caaa2580bb213): Adding workflow start and stop steps. root: INFO: 2017-12-11T18:18:26.122Z: JOB_MESSAGE_DEBUG: (a72caaa2580bbde9): Assigning stage ids. root: INFO: Job 2017-12-11_10_18_20-331615697938870470 is in state JOB_STATE_RUNNING root: INFO: 2017-12-11T18:18:26.316Z: JOB_MESSAGE_DEBUG: (62475f0455e388be): Executing wait step start22 root: INFO: 2017-12-11T18:18:26.374Z: JOB_MESSAGE_BASIC: (8dd961acd8e88041): Executing operation side1/Read root: INFO: 2017-12-11T18:18:26.409Z: JOB_MESSAGE_BASIC: (62475f0455e38fe0): Executing operation side2/Read root: INFO: 2017-12-11T18:18:26.441Z: JOB_MESSAGE_DEBUG: (8dd961acd8e88167): Value "side1/Read.out" materialized. root: INFO: 2017-12-11T18:18:26.472Z: JOB_MESSAGE_DEBUG: (62475f0455e38702): Value "side2/Read.out" materialized. root: INFO: 2017-12-11T18:18:26.537Z: JOB_MESSAGE_BASIC: (62475f0455e38249): Executing operation Flatten root: INFO: 2017-12-11T18:18:26.624Z: JOB_MESSAGE_DEBUG: (62475f0455e388d7): Value "Flatten.out" materialized. root: INFO: 2017-12-11T18:18:26.684Z: JOB_MESSAGE_BASIC: (8dd961acd8e8825c): Executing operation FlatMap(<lambda at sideinputs_test.py:302>)/_UnpickledSideInput(Flatten.out.0) root: INFO: 2017-12-11T18:18:26.756Z: JOB_MESSAGE_DEBUG: (8dd961acd8e8850a): Value "FlatMap(<lambda at sideinputs_test.py:302>)/_UnpickledSideInput(Flatten.out.0).output" materialized. root: INFO: 2017-12-11T18:18:26.819Z: JOB_MESSAGE_BASIC: (62475f0455e38262): Executing operation assert_that/Group/GroupByKey/Create root: INFO: 2017-12-11T18:18:26.861Z: JOB_MESSAGE_DEBUG: (cd1119aa2b5c67b6): Starting worker pool setup. root: INFO: 2017-12-11T18:18:26.890Z: JOB_MESSAGE_BASIC: (cd1119aa2b5c68d0): Starting 1 workers in us-central1-f... root: INFO: 2017-12-11T18:18:26.953Z: JOB_MESSAGE_DEBUG: (9f404f4a3c7dd91b): Value "assert_that/Group/GroupByKey/Session" materialized. root: INFO: 2017-12-11T18:18:27.018Z: JOB_MESSAGE_BASIC: (9f404f4a3c7dda0c): Executing operation main input/Read+FlatMap(<lambda at sideinputs_test.py:302>)/FlatMap(<lambda at sideinputs_test.py:302>)+assert_that/WindowInto(WindowIntoFn)+assert_that/ToVoidKey+assert_that/Group/pair_with_1+assert_that/Group/GroupByKey/Reify+assert_that/Group/GroupByKey/Write root: INFO: 2017-12-11T18:18:27.044Z: JOB_MESSAGE_BASIC: (62475f0455e388f0): Executing operation assert_that/Create/Read+assert_that/Group/pair_with_0+assert_that/Group/GroupByKey/Reify+assert_that/Group/GroupByKey/Write root: INFO: 2017-12-11T18:18:33.851Z: JOB_MESSAGE_DETAILED: (4bd4c805d4422ebb): Autoscaling: Raised the number of workers to 0 based on the rate of progress in the currently running step(s). root: INFO: 2017-12-11T18:18:57.734Z: JOB_MESSAGE_ERROR: (4bd4c805d4422b8c): Startup of the worker pool in zone us-central1-f failed to bring up any of the desired 1 workers. ZONE_RESOURCE_POOL_EXHAUSTED: The zone 'projects/apache-beam-testing/zones/us-central1-f' does not have enough resources available to fulfill the request. Try a different zone, or try again later. root: INFO: 2017-12-11T18:18:57.783Z: JOB_MESSAGE_ERROR: (4bd4c805d4422e16): Workflow failed. root: INFO: 2017-12-11T18:18:58.229Z: JOB_MESSAGE_DETAILED: (a72caaa2580bbb2c): Cleaning up. root: INFO: 2017-12-11T18:18:58.315Z: JOB_MESSAGE_DEBUG: (a72caaa2580bb2d8): Starting worker pool teardown. root: INFO: 2017-12-11T18:18:58.348Z: JOB_MESSAGE_BASIC: (a72caaa2580bbeae): Stopping worker pool... root: INFO: 2017-12-11T18:20:06.426Z: JOB_MESSAGE_BASIC: (a72caaa2580bbe06): Worker pool stopped. root: INFO: 2017-12-11T18:20:06.461Z: JOB_MESSAGE_DEBUG: (a72caaa2580bb5b2): Tearing down pending resources... root: INFO: Job 2017-12-11_10_18_20-331615697938870470 is in state JOB_STATE_FAILED --------------------- >> end captured logging << --------------------- ---------------------------------------------------------------------- Ran 15 tests in 924.728s FAILED (errors=10) Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2017-12-11_10_11_17-2050635262786868742?project=apache-beam-testing Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2017-12-11_10_14_12-5347517398133682125?project=apache-beam-testing Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2017-12-11_10_19_08-960612389348443737?project=apache-beam-testing Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2017-12-11_10_11_17-15962750727635144905?project=apache-beam-testing Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2017-12-11_10_13_39-4091447656855636797?project=apache-beam-testing Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2017-12-11_10_16_30-14639992260540082984?project=apache-beam-testing Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2017-12-11_10_11_15-9665064188856117708?project=apache-beam-testing Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2017-12-11_10_12_49-11904819970812392070?project=apache-beam-testing Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2017-12-11_10_15_37-13733329581878086187?project=apache-beam-testing Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2017-12-11_10_18_20-331615697938870470?project=apache-beam-testing Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2017-12-11_10_20_38-88395723950014942?project=apache-beam-testing Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2017-12-11_10_11_18-2561225770366590499?project=apache-beam-testing Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2017-12-11_10_14_01-1920037737539727064?project=apache-beam-testing Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2017-12-11_10_16_49-8350530028329332064?project=apache-beam-testing Build step 'Execute shell' marked build as failure