See <https://builds.apache.org/job/beam_PostCommit_Python_Verify/864/>

------------------------------------------
[...truncated 2714 lines...]
    }, 
    {
      "kind": "CollectionToSingleton", 
      "name": "s3", 
      "properties": {
        "display_data": [], 
        "output_info": [
          {
            "encoding": {
              "@type": "WindowedValueCoder$<string of 192 bytes>", 
              "component_encodings": [
                {
                  "@type": "WindowedValueCoder$<string of 192 bytes>", 
                  "component_encodings": [
                    {
                      "@type": 
"PickleCoder$eJxrYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlwBmcnZOanOIA5XIYNmYyFjbSFTkh4A9DYR/g==",
 
                      "component_encodings": [
                        {
                          "@type": 
"PickleCoder$eJxrYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlwBmcnZOanOIA5XIYNmYyFjbSFTkh4A9DYR/g==",
 
                          "component_encodings": []
                        }, 
                        {
                          "@type": 
"PickleCoder$eJxrYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlwBmcnZOanOIA5XIYNmYyFjbSFTkh4A9DYR/g==",
 
                          "component_encodings": []
                        }
                      ], 
                      "is_pair_like": true
                    }, 
                    {
                      "@type": 
"TimestampCoder$eJxrYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlwhmbmpxSWJuQXOID5XIYNmYyFjbSFTkh4ANWETWg==",
 
                      "component_encodings": []
                    }, 
                    {
                      "@type": 
"PickleCoder$eJxrYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlwBmcnZOanOIA5XIYNmYyFjbSFTkh4A9DYR/g==",
 
                      "component_encodings": [
                        {
                          "@type": 
"PickleCoder$eJxrYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlwBmcnZOanOIA5XIYNmYyFjbSFTkh4A9DYR/g==",
 
                          "component_encodings": []
                        }, 
                        {
                          "@type": 
"PickleCoder$eJxrYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlwBmcnZOanOIA5XIYNmYyFjbSFTkh4A9DYR/g==",
 
                          "component_encodings": []
                        }
                      ], 
                      "is_pair_like": true
                    }
                  ], 
                  "is_wrapper": true
                }
              ]
            }, 
            "output_name": "out", 
            "user_name": "ViewAsSingleton(side.None)/CreatePCollectionView.out"
          }
        ], 
        "parallel_input": {
          "@type": "OutputReference", 
          "output_name": "out", 
          "step_name": "s2"
        }, 
        "user_name": "ViewAsSingleton(side.None)/CreatePCollectionView"
      }
    }, 
    {
      "kind": "ParallelDo", 
      "name": "s4", 
      "properties": {
        "display_data": [
          {
            "key": "fn", 
            "label": "Transform Function", 
            "namespace": "apache_beam.transforms.core.CallableWrapperDoFn", 
            "type": "STRING", 
            "value": "<lambda>"
          }, 
          {
            "key": "fn", 
            "label": "Transform Function", 
            "namespace": "apache_beam.transforms.core.ParDo", 
            "shortValue": "CallableWrapperDoFn", 
            "type": "STRING", 
            "value": "apache_beam.transforms.core.CallableWrapperDoFn"
          }
        ], 
        "non_parallel_inputs": {
          "s3": {
            "@type": "OutputReference", 
            "output_name": "out", 
            "step_name": "s3"
          }
        }, 
        "output_info": [
          {
            "encoding": {
              "@type": "WindowedValueCoder$<string of 408 bytes>", 
              "component_encodings": [
                {
                  "@type": 
"FastPrimitivesCoder$eJxrYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/",
 
                  "component_encodings": [
                    {
                      "@type": 
"FastPrimitivesCoder$eJxrYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/",
 
                      "component_encodings": []
                    }, 
                    {
                      "@type": 
"FastPrimitivesCoder$eJxrYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/",
 
                      "component_encodings": []
                    }
                  ], 
                  "is_pair_like": true
                }, 
                {
                  "@type": 
"TimestampCoder$eJxrYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlwhmbmpxSWJuQXOID5XIYNmYyFjbSFTkh4ANWETWg==",
 
                  "component_encodings": []
                }, 
                {
                  "@type": "SingletonCoder$<string of 256 bytes>", 
                  "component_encodings": []
                }
              ], 
              "is_wrapper": true
            }, 
            "output_name": "out", 
            "user_name": "compute.out"
          }
        ], 
        "parallel_input": {
          "@type": "OutputReference", 
          "output_name": "out", 
          "step_name": "s1"
        }, 
        "serialized_fn": "<string of 1116 bytes>", 
        "user_name": "compute"
      }
    }
  ], 
  "type": "JOB_TYPE_BATCH"
}
root: INFO: Create job: <Job
 id: u'2016-12-12_07_50_24-6463664622346895841'
 projectId: u'apache-beam-testing'
 steps: []
 tempFiles: []
 type: TypeValueValuesEnum(JOB_TYPE_BATCH, 1)>
root: INFO: Created job with id: [2016-12-12_07_50_24-6463664622346895841]
root: INFO: To access the Dataflow monitoring console, please navigate to 
https://console.developers.google.com/project/apache-beam-testing/dataflow/job/2016-12-12_07_50_24-6463664622346895841
root: INFO: Job 2016-12-12_07_50_24-6463664622346895841 is in state 
JOB_STATE_RUNNING
root: INFO: 2016-12-12_07_50_24-6463664622346895841_00000158f3ba5ac0: 
2016-12-12T15:50:26.240Z: JOB_MESSAGE_DETAILED: (a4af248cc1e28e44): Checking 
required Cloud APIs are enabled.
root: INFO: 2016-12-12_07_50_24-6463664622346895841_00000158f3ba5c3d: 
2016-12-12T15:50:26.621Z: JOB_MESSAGE_DETAILED: (a4af248cc1e28d09): Expanding 
GroupByKey operations into optimizable parts.
root: INFO: 2016-12-12_07_50_24-6463664622346895841_00000158f3ba5c3f: 
2016-12-12T15:50:26.623Z: JOB_MESSAGE_DETAILED: (a4af248cc1e28cc3): Lifting 
ValueCombiningMappingFns into MergeBucketsMappingFns
root: INFO: 2016-12-12_07_50_24-6463664622346895841_00000158f3ba5c42: 
2016-12-12T15:50:26.626Z: JOB_MESSAGE_DETAILED: (a4af248cc1e28c7d): Annotating 
graph with Autotuner information.
root: INFO: 2016-12-12_07_50_24-6463664622346895841_00000158f3ba5c4b: 
2016-12-12T15:50:26.635Z: JOB_MESSAGE_DETAILED: (a4af248cc1e28bab): Fusing 
adjacent ParDo, Read, Write, and Flatten operations
root: INFO: 2016-12-12_07_50_24-6463664622346895841_00000158f3ba5c5b: 
2016-12-12T15:50:26.651Z: JOB_MESSAGE_DEBUG: (a4af248cc1e289c1): Workflow 
config is missing a default resource spec.
root: INFO: 2016-12-12_07_50_24-6463664622346895841_00000158f3ba5c5d: 
2016-12-12T15:50:26.653Z: JOB_MESSAGE_DETAILED: (a4af248cc1e2897b): Adding 
StepResource setup and teardown to workflow graph.
root: INFO: 2016-12-12_07_50_24-6463664622346895841_00000158f3ba5c98: 
2016-12-12T15:50:26.712Z: JOB_MESSAGE_DEBUG: (7e90f8ebe8480cde): Adding 
workflow start and stop steps.
root: INFO: 2016-12-12_07_50_24-6463664622346895841_00000158f3ba5cca: 
2016-12-12T15:50:26.762Z: JOB_MESSAGE_DEBUG: (d8a1c8535c37ea18): Assigning 
stage ids.
root: INFO: 2016-12-12_07_50_24-6463664622346895841_00000158f3ba5d36: 
2016-12-12T15:50:26.870Z: JOB_MESSAGE_DEBUG: (877e4ceff4313187): Executing wait 
step start2
root: INFO: 2016-12-12_07_50_24-6463664622346895841_00000158f3ba5d3f: 
2016-12-12T15:50:26.879Z: JOB_MESSAGE_DEBUG: (31cd13e4fae54049): Executing 
operation side
root: INFO: 2016-12-12_07_50_24-6463664622346895841_00000158f3ba5d4b: 
2016-12-12T15:50:26.891Z: JOB_MESSAGE_DEBUG: (263b1beb2e4eb699): Value 
"side.out" materialized.
root: INFO: 2016-12-12_07_50_24-6463664622346895841_00000158f3ba5d5f: 
2016-12-12T15:50:26.911Z: JOB_MESSAGE_BASIC: S01: (81ad196c466d6570): Executing 
operation ViewAsSingleton(side.None)/CreatePCollectionView
root: INFO: 2016-12-12_07_50_24-6463664622346895841_00000158f3ba5d65: 
2016-12-12T15:50:26.917Z: JOB_MESSAGE_DEBUG: (d8a1c8535c37e4ba): Executing 
operation start
root: INFO: 2016-12-12_07_50_24-6463664622346895841_00000158f3ba5d6c: 
2016-12-12T15:50:26.924Z: JOB_MESSAGE_DEBUG: (7e90f8ebe8480d21): Value 
"ViewAsSingleton(side.None)/CreatePCollectionView.out" materialized.
root: INFO: 2016-12-12_07_50_24-6463664622346895841_00000158f3ba5d70: 
2016-12-12T15:50:26.928Z: JOB_MESSAGE_DEBUG: (667056a065d57d76): Value 
"start.out" materialized.
root: INFO: 2016-12-12_07_50_24-6463664622346895841_00000158f3ba5d7c: 
2016-12-12T15:50:26.940Z: JOB_MESSAGE_BASIC: S02: (263b1beb2e4eb07b): Executing 
operation compute
root: INFO: 2016-12-12_07_50_24-6463664622346895841_00000158f3ba5e47: 
2016-12-12T15:50:27.143Z: JOB_MESSAGE_DEBUG: (98ee7be1df6055ce): Starting 
worker pool setup.
root: INFO: 2016-12-12_07_50_24-6463664622346895841_00000158f3ba5e49: 
2016-12-12T15:50:27.145Z: JOB_MESSAGE_BASIC: (98ee7be1df605c54): Starting 1 
workers...
root: INFO: 2016-12-12_07_50_24-6463664622346895841_00000158f3bc3dc7: 
2016-12-12T15:52:29.895Z: JOB_MESSAGE_DETAILED: (6607e90d3f03422d): Workers 
have started successfully.
root: INFO: 2016-12-12_07_50_24-6463664622346895841_00000158f3bd1134: 
2016-12-12T15:53:24.020Z: JOB_MESSAGE_ERROR: (cc11e033844b1905): Traceback 
(most recent call last):
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", 
line 514, in do_work
    work_executor.execute()
  File "dataflow_worker/executor.py", line 892, in 
dataflow_worker.executor.MapTaskExecutor.execute 
(dataflow_worker/executor.c:24008)
    op.start()
  File "dataflow_worker/executor.py", line 456, in 
dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:13870)
    def start(self):
  File "dataflow_worker/executor.py", line 483, in 
dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:13685)
    self.dofn_runner = common.DoFnRunner(
  File "apache_beam/runners/common.py", line 89, in 
apache_beam.runners.common.DoFnRunner.__init__ 
(apache_beam/runners/common.c:3469)
    args, kwargs, [side_input[global_window]
  File 
"/usr/local/lib/python2.7/dist-packages/apache_beam/transforms/sideinputs.py", 
line 192, in __getitem__
    _FilteringIterable(self._iterable, target_window), self._view_options)
  File "/usr/local/lib/python2.7/dist-packages/apache_beam/pvalue.py", line 
279, in _from_runtime_iterable
    'PCollection with more than one element accessed as '
ValueError: PCollection with more than one element accessed as a singleton view.

root: INFO: 2016-12-12_07_50_24-6463664622346895841_00000158f3bd11a4: 
2016-12-12T15:53:24.132Z: JOB_MESSAGE_ERROR: (cc11e033844b1835): Traceback 
(most recent call last):
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", 
line 514, in do_work
    work_executor.execute()
  File "dataflow_worker/executor.py", line 892, in 
dataflow_worker.executor.MapTaskExecutor.execute 
(dataflow_worker/executor.c:24008)
    op.start()
  File "dataflow_worker/executor.py", line 456, in 
dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:13870)
    def start(self):
  File "dataflow_worker/executor.py", line 483, in 
dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:13685)
    self.dofn_runner = common.DoFnRunner(
  File "apache_beam/runners/common.py", line 89, in 
apache_beam.runners.common.DoFnRunner.__init__ 
(apache_beam/runners/common.c:3469)
    args, kwargs, [side_input[global_window]
  File 
"/usr/local/lib/python2.7/dist-packages/apache_beam/transforms/sideinputs.py", 
line 192, in __getitem__
    _FilteringIterable(self._iterable, target_window), self._view_options)
  File "/usr/local/lib/python2.7/dist-packages/apache_beam/pvalue.py", line 
279, in _from_runtime_iterable
    'PCollection with more than one element accessed as '
ValueError: PCollection with more than one element accessed as a singleton view.

root: INFO: 2016-12-12_07_50_24-6463664622346895841_00000158f3bd121c: 
2016-12-12T15:53:24.252Z: JOB_MESSAGE_ERROR: (cc11e033844b1765): Traceback 
(most recent call last):
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", 
line 514, in do_work
    work_executor.execute()
  File "dataflow_worker/executor.py", line 892, in 
dataflow_worker.executor.MapTaskExecutor.execute 
(dataflow_worker/executor.c:24008)
    op.start()
  File "dataflow_worker/executor.py", line 456, in 
dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:13870)
    def start(self):
  File "dataflow_worker/executor.py", line 483, in 
dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:13685)
    self.dofn_runner = common.DoFnRunner(
  File "apache_beam/runners/common.py", line 89, in 
apache_beam.runners.common.DoFnRunner.__init__ 
(apache_beam/runners/common.c:3469)
    args, kwargs, [side_input[global_window]
  File 
"/usr/local/lib/python2.7/dist-packages/apache_beam/transforms/sideinputs.py", 
line 192, in __getitem__
    _FilteringIterable(self._iterable, target_window), self._view_options)
  File "/usr/local/lib/python2.7/dist-packages/apache_beam/pvalue.py", line 
279, in _from_runtime_iterable
    'PCollection with more than one element accessed as '
ValueError: PCollection with more than one element accessed as a singleton view.

root: INFO: 2016-12-12_07_50_24-6463664622346895841_00000158f3bd127f: 
2016-12-12T15:53:24.351Z: JOB_MESSAGE_ERROR: (cc11e033844b1695): Traceback 
(most recent call last):
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", 
line 514, in do_work
    work_executor.execute()
  File "dataflow_worker/executor.py", line 892, in 
dataflow_worker.executor.MapTaskExecutor.execute 
(dataflow_worker/executor.c:24008)
    op.start()
  File "dataflow_worker/executor.py", line 456, in 
dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:13870)
    def start(self):
  File "dataflow_worker/executor.py", line 483, in 
dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:13685)
    self.dofn_runner = common.DoFnRunner(
  File "apache_beam/runners/common.py", line 89, in 
apache_beam.runners.common.DoFnRunner.__init__ 
(apache_beam/runners/common.c:3469)
    args, kwargs, [side_input[global_window]
  File 
"/usr/local/lib/python2.7/dist-packages/apache_beam/transforms/sideinputs.py", 
line 192, in __getitem__
    _FilteringIterable(self._iterable, target_window), self._view_options)
  File "/usr/local/lib/python2.7/dist-packages/apache_beam/pvalue.py", line 
279, in _from_runtime_iterable
    'PCollection with more than one element accessed as '
ValueError: PCollection with more than one element accessed as a singleton view.

root: INFO: 2016-12-12_07_50_24-6463664622346895841_00000158f3bd12a4: 
2016-12-12T15:53:24.388Z: JOB_MESSAGE_DEBUG: (263b1beb2e4ebe21): Executing 
failure step failure1
root: INFO: 2016-12-12_07_50_24-6463664622346895841_00000158f3bd12a6: 
2016-12-12T15:53:24.390Z: JOB_MESSAGE_ERROR: (263b1beb2e4ebc17): Workflow 
failed. Causes: (263b1beb2e4eb33a): S02:compute failed.
root: INFO: 2016-12-12_07_50_24-6463664622346895841_00000158f3bd12da: 
2016-12-12T15:53:24.442Z: JOB_MESSAGE_DETAILED: (d8bd0f04c4d5c1ec): Cleaning up.
root: INFO: 2016-12-12_07_50_24-6463664622346895841_00000158f3bd1362: 
2016-12-12T15:53:24.578Z: JOB_MESSAGE_DEBUG: (d8bd0f04c4d5c39b): Starting 
worker pool teardown.
root: INFO: 2016-12-12_07_50_24-6463664622346895841_00000158f3bd1364: 
2016-12-12T15:53:24.580Z: JOB_MESSAGE_BASIC: (d8bd0f04c4d5cf65): Stopping 
worker pool...
root: INFO: 2016-12-12_07_50_24-6463664622346895841_00000158f3be4bea: 
2016-12-12T15:54:44.586Z: JOB_MESSAGE_BASIC: (d8bd0f04c4d5c114): Worker pool 
stopped.
root: INFO: 2016-12-12_07_50_24-6463664622346895841_00000158f3be4bfc: 
2016-12-12T15:54:44.604Z: JOB_MESSAGE_DEBUG: (d8bd0f04c4d5ce8d): Tearing down 
pending resources...
root: INFO: Job 2016-12-12_07_50_24-6463664622346895841 is in state 
JOB_STATE_FAILED
--------------------- >> end captured logging << ---------------------

----------------------------------------------------------------------
Ran 21 tests in 5235.902s

FAILED (errors=1)
Build step 'Execute shell' marked build as failure

Reply via email to