See <https://builds.apache.org/job/beam_PostCommit_Python_Verify/855/>
------------------------------------------ [...truncated 2731 lines...] { "@type": "PickleCoder$eJxrYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlwBmcnZOanOIA5XIYNmYyFjbSFTkh4A9DYR/g==", "component_encodings": [] }, { "@type": "PickleCoder$eJxrYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlwBmcnZOanOIA5XIYNmYyFjbSFTkh4A9DYR/g==", "component_encodings": [] } ], "is_pair_like": true }, { "@type": "TimestampCoder$eJxrYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlwhmbmpxSWJuQXOID5XIYNmYyFjbSFTkh4ANWETWg==", "component_encodings": [] }, { "@type": "PickleCoder$eJxrYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlwBmcnZOanOIA5XIYNmYyFjbSFTkh4A9DYR/g==", "component_encodings": [ { "@type": "PickleCoder$eJxrYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlwBmcnZOanOIA5XIYNmYyFjbSFTkh4A9DYR/g==", "component_encodings": [] }, { "@type": "PickleCoder$eJxrYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlwBmcnZOanOIA5XIYNmYyFjbSFTkh4A9DYR/g==", "component_encodings": [] } ], "is_pair_like": true } ], "is_wrapper": true } ] }, "output_name": "out", "user_name": "ViewAsSingleton(side.None)/CreatePCollectionView.out" } ], "parallel_input": { "@type": "OutputReference", "output_name": "out", "step_name": "s2" }, "user_name": "ViewAsSingleton(side.None)/CreatePCollectionView" } }, { "kind": "ParallelDo", "name": "s4", "properties": { "display_data": [ { "key": "fn", "label": "Transform Function", "namespace": "apache_beam.transforms.core.CallableWrapperDoFn", "type": "STRING", "value": "<lambda>" }, { "key": "fn", "label": "Transform Function", "namespace": "apache_beam.transforms.core.ParDo", "shortValue": "CallableWrapperDoFn", "type": "STRING", "value": "apache_beam.transforms.core.CallableWrapperDoFn" } ], "non_parallel_inputs": { "s3": { "@type": "OutputReference", "output_name": "out", "step_name": "s3" } }, "output_info": [ { "encoding": { "@type": "WindowedValueCoder$<string of 408 bytes>", "component_encodings": [ { "@type": "FastPrimitivesCoder$eJxrYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/", "component_encodings": [ { "@type": "FastPrimitivesCoder$eJxrYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/", "component_encodings": [] }, { "@type": "FastPrimitivesCoder$eJxrYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/", "component_encodings": [] } ], "is_pair_like": true }, { "@type": "TimestampCoder$eJxrYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlwhmbmpxSWJuQXOID5XIYNmYyFjbSFTkh4ANWETWg==", "component_encodings": [] }, { "@type": "SingletonCoder$<string of 256 bytes>", "component_encodings": [] } ], "is_wrapper": true }, "output_name": "out", "user_name": "compute.out" } ], "parallel_input": { "@type": "OutputReference", "output_name": "out", "step_name": "s1" }, "serialized_fn": "<string of 1116 bytes>", "user_name": "compute" } } ], "type": "JOB_TYPE_BATCH" } root: INFO: Create job: <Job id: u'2016-12-10_01_49_34-8027451739738855844' projectId: u'apache-beam-testing' steps: [] tempFiles: [] type: TypeValueValuesEnum(JOB_TYPE_BATCH, 1)> root: INFO: Created job with id: [2016-12-10_01_49_34-8027451739738855844] root: INFO: To access the Dataflow monitoring console, please navigate to https://console.developers.google.com/project/apache-beam-testing/dataflow/job/2016-12-10_01_49_34-8027451739738855844 root: INFO: Job 2016-12-10_01_49_34-8027451739738855844 is in state JOB_STATE_RUNNING root: INFO: 2016-12-10_01_49_34-8027451739738855844_00000158e82344f3: 2016-12-10T09:49:35.347Z: JOB_MESSAGE_DETAILED: (2076fc0f8032d25c): Checking required Cloud APIs are enabled. root: INFO: 2016-12-10_01_49_34-8027451739738855844_00000158e823465a: 2016-12-10T09:49:35.706Z: JOB_MESSAGE_DETAILED: (2076fc0f8032db3b): Expanding GroupByKey operations into optimizable parts. root: INFO: 2016-12-10_01_49_34-8027451739738855844_00000158e823465c: 2016-12-10T09:49:35.708Z: JOB_MESSAGE_DETAILED: (2076fc0f8032d289): Lifting ValueCombiningMappingFns into MergeBucketsMappingFns root: INFO: 2016-12-10_01_49_34-8027451739738855844_00000158e823465f: 2016-12-10T09:49:35.711Z: JOB_MESSAGE_DETAILED: (2076fc0f8032d9d7): Annotating graph with Autotuner information. root: INFO: 2016-12-10_01_49_34-8027451739738855844_00000158e8234667: 2016-12-10T09:49:35.719Z: JOB_MESSAGE_DETAILED: (2076fc0f8032dfc1): Fusing adjacent ParDo, Read, Write, and Flatten operations root: INFO: 2016-12-10_01_49_34-8027451739738855844_00000158e8234674: 2016-12-10T09:49:35.732Z: JOB_MESSAGE_DEBUG: (2076fc0f8032d2e3): Workflow config is missing a default resource spec. root: INFO: 2016-12-10_01_49_34-8027451739738855844_00000158e8234676: 2016-12-10T09:49:35.734Z: JOB_MESSAGE_DETAILED: (2076fc0f8032da31): Adding StepResource setup and teardown to workflow graph. root: INFO: 2016-12-10_01_49_34-8027451739738855844_00000158e8234691: 2016-12-10T09:49:35.761Z: JOB_MESSAGE_DEBUG: (aff11933d595349f): Adding workflow start and stop steps. root: INFO: 2016-12-10_01_49_34-8027451739738855844_00000158e823469a: 2016-12-10T09:49:35.770Z: JOB_MESSAGE_DEBUG: (aff11933d595341d): Assigning stage ids. root: INFO: 2016-12-10_01_49_34-8027451739738855844_00000158e82346ff: 2016-12-10T09:49:35.871Z: JOB_MESSAGE_DEBUG: (44cf5c35fec5af66): Executing wait step start2 root: INFO: 2016-12-10_01_49_34-8027451739738855844_00000158e8234708: 2016-12-10T09:49:35.880Z: JOB_MESSAGE_DEBUG: (8a6e7a5719b93808): Executing operation side root: INFO: 2016-12-10_01_49_34-8027451739738855844_00000158e823470a: 2016-12-10T09:49:35.882Z: JOB_MESSAGE_DEBUG: (a853400cf892c9b7): Executing operation start root: INFO: 2016-12-10_01_49_34-8027451739738855844_00000158e8234715: 2016-12-10T09:49:35.893Z: JOB_MESSAGE_DEBUG: (46058dddabad1016): Value "start.out" materialized. root: INFO: 2016-12-10_01_49_34-8027451739738855844_00000158e8234727: 2016-12-10T09:49:35.911Z: JOB_MESSAGE_DEBUG: (134825dfbf05c885): Value "side.out" materialized. root: INFO: 2016-12-10_01_49_34-8027451739738855844_00000158e8234731: 2016-12-10T09:49:35.921Z: JOB_MESSAGE_BASIC: S01: (a853400cf892cb89): Executing operation ViewAsSingleton(side.None)/CreatePCollectionView root: INFO: 2016-12-10_01_49_34-8027451739738855844_00000158e823473c: 2016-12-10T09:49:35.932Z: JOB_MESSAGE_DEBUG: (da1f77c5a0e3f0b7): Value "ViewAsSingleton(side.None)/CreatePCollectionView.out" materialized. root: INFO: 2016-12-10_01_49_34-8027451739738855844_00000158e8234745: 2016-12-10T09:49:35.941Z: JOB_MESSAGE_BASIC: S02: (a853400cf892c8a1): Executing operation compute root: INFO: 2016-12-10_01_49_34-8027451739738855844_00000158e8234811: 2016-12-10T09:49:36.145Z: JOB_MESSAGE_DEBUG: (3439fbde293e209a): Starting worker pool setup. root: INFO: 2016-12-10_01_49_34-8027451739738855844_00000158e8234813: 2016-12-10T09:49:36.147Z: JOB_MESSAGE_BASIC: (3439fbde293e2a04): Starting 1 workers... root: INFO: 2016-12-10_01_49_34-8027451739738855844_00000158e824d152: 2016-12-10T09:51:16.818Z: JOB_MESSAGE_DETAILED: (54dd093694616bc0): Workers have started successfully. root: INFO: 2016-12-10_01_49_34-8027451739738855844_00000158e8258c2b: 2016-12-10T09:52:04.651Z: JOB_MESSAGE_ERROR: (d5fbdcb92ef7bb9): Traceback (most recent call last): File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", line 514, in do_work work_executor.execute() File "dataflow_worker/executor.py", line 892, in dataflow_worker.executor.MapTaskExecutor.execute (dataflow_worker/executor.c:24008) op.start() File "dataflow_worker/executor.py", line 456, in dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:13870) def start(self): File "dataflow_worker/executor.py", line 483, in dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:13685) self.dofn_runner = common.DoFnRunner( File "apache_beam/runners/common.py", line 89, in apache_beam.runners.common.DoFnRunner.__init__ (apache_beam/runners/common.c:3469) args, kwargs, [side_input[global_window] File "/usr/local/lib/python2.7/dist-packages/apache_beam/transforms/sideinputs.py", line 192, in __getitem__ _FilteringIterable(self._iterable, target_window), self._view_options) File "/usr/local/lib/python2.7/dist-packages/apache_beam/pvalue.py", line 279, in _from_runtime_iterable 'PCollection with more than one element accessed as ' ValueError: PCollection with more than one element accessed as a singleton view. root: INFO: 2016-12-10_01_49_34-8027451739738855844_00000158e8258cb8: 2016-12-10T09:52:04.792Z: JOB_MESSAGE_ERROR: (d5fbdcb92ef7949): Traceback (most recent call last): File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", line 514, in do_work work_executor.execute() File "dataflow_worker/executor.py", line 892, in dataflow_worker.executor.MapTaskExecutor.execute (dataflow_worker/executor.c:24008) op.start() File "dataflow_worker/executor.py", line 456, in dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:13870) def start(self): File "dataflow_worker/executor.py", line 483, in dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:13685) self.dofn_runner = common.DoFnRunner( File "apache_beam/runners/common.py", line 89, in apache_beam.runners.common.DoFnRunner.__init__ (apache_beam/runners/common.c:3469) args, kwargs, [side_input[global_window] File "/usr/local/lib/python2.7/dist-packages/apache_beam/transforms/sideinputs.py", line 192, in __getitem__ _FilteringIterable(self._iterable, target_window), self._view_options) File "/usr/local/lib/python2.7/dist-packages/apache_beam/pvalue.py", line 279, in _from_runtime_iterable 'PCollection with more than one element accessed as ' ValueError: PCollection with more than one element accessed as a singleton view. root: INFO: 2016-12-10_01_49_34-8027451739738855844_00000158e8258d27: 2016-12-10T09:52:04.903Z: JOB_MESSAGE_ERROR: (d5fbdcb92ef76d9): Traceback (most recent call last): File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", line 514, in do_work work_executor.execute() File "dataflow_worker/executor.py", line 892, in dataflow_worker.executor.MapTaskExecutor.execute (dataflow_worker/executor.c:24008) op.start() File "dataflow_worker/executor.py", line 456, in dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:13870) def start(self): File "dataflow_worker/executor.py", line 483, in dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:13685) self.dofn_runner = common.DoFnRunner( File "apache_beam/runners/common.py", line 89, in apache_beam.runners.common.DoFnRunner.__init__ (apache_beam/runners/common.c:3469) args, kwargs, [side_input[global_window] File "/usr/local/lib/python2.7/dist-packages/apache_beam/transforms/sideinputs.py", line 192, in __getitem__ _FilteringIterable(self._iterable, target_window), self._view_options) File "/usr/local/lib/python2.7/dist-packages/apache_beam/pvalue.py", line 279, in _from_runtime_iterable 'PCollection with more than one element accessed as ' ValueError: PCollection with more than one element accessed as a singleton view. root: INFO: 2016-12-10_01_49_34-8027451739738855844_00000158e8258d8d: 2016-12-10T09:52:05.005Z: JOB_MESSAGE_ERROR: (d5fbdcb92ef7469): Traceback (most recent call last): File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", line 514, in do_work work_executor.execute() File "dataflow_worker/executor.py", line 892, in dataflow_worker.executor.MapTaskExecutor.execute (dataflow_worker/executor.c:24008) op.start() File "dataflow_worker/executor.py", line 456, in dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:13870) def start(self): File "dataflow_worker/executor.py", line 483, in dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:13685) self.dofn_runner = common.DoFnRunner( File "apache_beam/runners/common.py", line 89, in apache_beam.runners.common.DoFnRunner.__init__ (apache_beam/runners/common.c:3469) args, kwargs, [side_input[global_window] File "/usr/local/lib/python2.7/dist-packages/apache_beam/transforms/sideinputs.py", line 192, in __getitem__ _FilteringIterable(self._iterable, target_window), self._view_options) File "/usr/local/lib/python2.7/dist-packages/apache_beam/pvalue.py", line 279, in _from_runtime_iterable 'PCollection with more than one element accessed as ' ValueError: PCollection with more than one element accessed as a singleton view. root: INFO: 2016-12-10_01_49_34-8027451739738855844_00000158e8258e00: 2016-12-10T09:52:05.120Z: JOB_MESSAGE_ERROR: (d5fbdcb92ef71f9): Traceback (most recent call last): File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", line 514, in do_work work_executor.execute() File "dataflow_worker/executor.py", line 892, in dataflow_worker.executor.MapTaskExecutor.execute (dataflow_worker/executor.c:24008) op.start() File "dataflow_worker/executor.py", line 456, in dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:13870) def start(self): File "dataflow_worker/executor.py", line 483, in dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:13685) self.dofn_runner = common.DoFnRunner( File "apache_beam/runners/common.py", line 89, in apache_beam.runners.common.DoFnRunner.__init__ (apache_beam/runners/common.c:3469) args, kwargs, [side_input[global_window] File "/usr/local/lib/python2.7/dist-packages/apache_beam/transforms/sideinputs.py", line 192, in __getitem__ _FilteringIterable(self._iterable, target_window), self._view_options) File "/usr/local/lib/python2.7/dist-packages/apache_beam/pvalue.py", line 279, in _from_runtime_iterable 'PCollection with more than one element accessed as ' ValueError: PCollection with more than one element accessed as a singleton view. root: INFO: 2016-12-10_01_49_34-8027451739738855844_00000158e8258e22: 2016-12-10T09:52:05.154Z: JOB_MESSAGE_DEBUG: (a853400cf892ce17): Executing failure step failure1 root: INFO: 2016-12-10_01_49_34-8027451739738855844_00000158e8258e24: 2016-12-10T09:52:05.156Z: JOB_MESSAGE_ERROR: (a853400cf892c95d): Workflow failed. Causes: (a853400cf892c9e8): S02:compute failed. root: INFO: 2016-12-10_01_49_34-8027451739738855844_00000158e8258f52: 2016-12-10T09:52:05.458Z: JOB_MESSAGE_DETAILED: (750e62c57747b351): Cleaning up. root: INFO: 2016-12-10_01_49_34-8027451739738855844_00000158e8258f59: 2016-12-10T09:52:05.465Z: JOB_MESSAGE_DEBUG: (750e62c57747bc68): Starting worker pool teardown. root: INFO: 2016-12-10_01_49_34-8027451739738855844_00000158e8258f5b: 2016-12-10T09:52:05.467Z: JOB_MESSAGE_BASIC: (750e62c57747bd22): Stopping worker pool... root: INFO: 2016-12-10_01_49_34-8027451739738855844_00000158e826c7e2: 2016-12-10T09:53:25.474Z: JOB_MESSAGE_BASIC: (750e62c57747b639): Worker pool stopped. root: INFO: 2016-12-10_01_49_34-8027451739738855844_00000158e826c7f3: 2016-12-10T09:53:25.491Z: JOB_MESSAGE_DEBUG: (750e62c57747b00a): Tearing down pending resources... root: INFO: Job 2016-12-10_01_49_34-8027451739738855844 is in state JOB_STATE_FAILED --------------------- >> end captured logging << --------------------- ---------------------------------------------------------------------- Ran 21 tests in 5095.881s FAILED (errors=1) Build step 'Execute shell' marked build as failure