See <https://builds.apache.org/job/beam_PostCommit_Py_VR_Dataflow/596/display/redirect?page=changes>
Changes: [github] [BEAM-4752] Add support for newer dill dependency (#5931) ------------------------------------------ [...truncated 1.31 MB...] } ], "parallel_input": { "@type": "OutputReference", "output_name": "out", "step_name": "s14" }, "serialized_fn": "<string of 1124 bytes>", "user_name": "assert_that/Match" } } ], "type": "JOB_TYPE_BATCH" } root: INFO: Create job: <Job createTime: u'2018-07-13T18:06:48.527308Z' currentStateTime: u'1970-01-01T00:00:00Z' id: u'2018-07-13_11_06_47-6503774580458991719' location: u'us-central1' name: u'beamapp-jenkins-0713180641-930106' projectId: u'apache-beam-testing' stageStates: [] steps: [] tempFiles: [] type: TypeValueValuesEnum(JOB_TYPE_BATCH, 1)> root: INFO: Created job with id: [2018-07-13_11_06_47-6503774580458991719] root: INFO: To access the Dataflow monitoring console, please navigate to https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-07-13_11_06_47-6503774580458991719?project=apache-beam-testing root: INFO: Job 2018-07-13_11_06_47-6503774580458991719 is in state JOB_STATE_RUNNING root: INFO: 2018-07-13T18:06:47.811Z: JOB_MESSAGE_DETAILED: Autoscaling is enabled for job 2018-07-13_11_06_47-6503774580458991719. The number of workers will be between 1 and 1000. root: INFO: 2018-07-13T18:06:47.854Z: JOB_MESSAGE_DETAILED: Autoscaling was automatically enabled for job 2018-07-13_11_06_47-6503774580458991719. root: INFO: 2018-07-13T18:06:50.209Z: JOB_MESSAGE_DETAILED: Checking required Cloud APIs are enabled. root: INFO: 2018-07-13T18:06:50.333Z: JOB_MESSAGE_DETAILED: Checking permissions granted to controller Service Account. root: INFO: 2018-07-13T18:06:51.176Z: JOB_MESSAGE_BASIC: Worker configuration: n1-standard-1 in us-central1-f. root: INFO: 2018-07-13T18:06:51.738Z: JOB_MESSAGE_DETAILED: Expanding CoGroupByKey operations into optimizable parts. root: INFO: 2018-07-13T18:06:51.787Z: JOB_MESSAGE_DEBUG: Combiner lifting skipped for step assert_that/Group/GroupByKey: GroupByKey not followed by a combiner. root: INFO: 2018-07-13T18:06:51.832Z: JOB_MESSAGE_DETAILED: Expanding GroupByKey operations into optimizable parts. root: INFO: 2018-07-13T18:06:51.869Z: JOB_MESSAGE_DETAILED: Lifting ValueCombiningMappingFns into MergeBucketsMappingFns root: INFO: 2018-07-13T18:06:51.952Z: JOB_MESSAGE_DEBUG: Annotating graph with Autotuner information. root: INFO: 2018-07-13T18:06:51.998Z: JOB_MESSAGE_DETAILED: Fusing adjacent ParDo, Read, Write, and Flatten operations root: INFO: 2018-07-13T18:06:52.036Z: JOB_MESSAGE_DETAILED: Unzipping flatten s11 for input s10.out root: INFO: 2018-07-13T18:06:52.069Z: JOB_MESSAGE_DETAILED: Fusing unzipped copy of assert_that/Group/GroupByKey/Reify, through flatten assert_that/Group/Flatten, into producer assert_that/Group/pair_with_1 root: INFO: 2018-07-13T18:06:52.104Z: JOB_MESSAGE_DETAILED: Fusing consumer assert_that/Group/Map(_merge_tagged_vals_under_key) into assert_that/Group/GroupByKey/GroupByWindow root: INFO: 2018-07-13T18:06:52.144Z: JOB_MESSAGE_DETAILED: Fusing consumer assert_that/Match into assert_that/Unkey root: INFO: 2018-07-13T18:06:52.185Z: JOB_MESSAGE_DETAILED: Fusing consumer assert_that/Unkey into assert_that/Group/Map(_merge_tagged_vals_under_key) root: INFO: 2018-07-13T18:06:52.229Z: JOB_MESSAGE_DETAILED: Fusing consumer assert_that/Group/GroupByKey/GroupByWindow into assert_that/Group/GroupByKey/Read root: INFO: 2018-07-13T18:06:52.271Z: JOB_MESSAGE_DETAILED: Unzipping flatten s11-u13 for input s12-reify-value0-c11 root: INFO: 2018-07-13T18:06:52.307Z: JOB_MESSAGE_DETAILED: Fusing unzipped copy of assert_that/Group/GroupByKey/Write, through flatten s11-u13, into producer assert_that/Group/GroupByKey/Reify root: INFO: 2018-07-13T18:06:52.346Z: JOB_MESSAGE_DETAILED: Fusing consumer assert_that/Group/GroupByKey/Reify into assert_that/Group/pair_with_0 root: INFO: 2018-07-13T18:06:52.387Z: JOB_MESSAGE_DETAILED: Fusing consumer assert_that/Group/GroupByKey/Write into assert_that/Group/GroupByKey/Reify root: INFO: 2018-07-13T18:06:52.430Z: JOB_MESSAGE_DETAILED: Fusing consumer assert_that/Group/pair_with_0 into assert_that/Create/Read root: INFO: 2018-07-13T18:06:52.472Z: JOB_MESSAGE_DETAILED: Fusing consumer Map(<lambda at sideinputs_test.py:213>)/Map(<lambda at sideinputs_test.py:213>) into main input/Read root: INFO: 2018-07-13T18:06:52.512Z: JOB_MESSAGE_DETAILED: Fusing consumer assert_that/WindowInto(WindowIntoFn) into Map(<lambda at sideinputs_test.py:213>)/Map(<lambda at sideinputs_test.py:213>) root: INFO: 2018-07-13T18:06:52.549Z: JOB_MESSAGE_DETAILED: Fusing consumer assert_that/Group/pair_with_1 into assert_that/ToVoidKey root: INFO: 2018-07-13T18:06:52.592Z: JOB_MESSAGE_DETAILED: Fusing consumer assert_that/ToVoidKey into assert_that/WindowInto(WindowIntoFn) root: INFO: 2018-07-13T18:06:52.631Z: JOB_MESSAGE_DEBUG: Workflow config is missing a default resource spec. root: INFO: 2018-07-13T18:06:52.676Z: JOB_MESSAGE_DEBUG: Adding StepResource setup and teardown to workflow graph. root: INFO: 2018-07-13T18:06:52.712Z: JOB_MESSAGE_DEBUG: Adding workflow start and stop steps. root: INFO: 2018-07-13T18:06:52.757Z: JOB_MESSAGE_DEBUG: Assigning stage ids. root: INFO: 2018-07-13T18:06:52.931Z: JOB_MESSAGE_DEBUG: Executing wait step start21 root: INFO: 2018-07-13T18:06:53.012Z: JOB_MESSAGE_BASIC: Executing operation assert_that/Group/GroupByKey/Create root: INFO: 2018-07-13T18:06:53.048Z: JOB_MESSAGE_BASIC: Executing operation side list/Read root: INFO: 2018-07-13T18:06:53.071Z: JOB_MESSAGE_DEBUG: Starting worker pool setup. root: INFO: 2018-07-13T18:06:53.117Z: JOB_MESSAGE_BASIC: Starting 1 workers in us-central1-f... root: INFO: 2018-07-13T18:06:53.141Z: JOB_MESSAGE_DEBUG: Value "side list/Read.out" materialized. root: INFO: 2018-07-13T18:06:53.203Z: JOB_MESSAGE_DEBUG: Value "assert_that/Group/GroupByKey/Session" materialized. root: INFO: 2018-07-13T18:06:53.237Z: JOB_MESSAGE_BASIC: Executing operation Map(<lambda at sideinputs_test.py:213>)/_UnpickledSideInput(Read.out.0) root: INFO: 2018-07-13T18:06:53.271Z: JOB_MESSAGE_BASIC: Executing operation Map(<lambda at sideinputs_test.py:213>)/_UnpickledSideInput(Read.out.1) root: INFO: 2018-07-13T18:06:53.316Z: JOB_MESSAGE_BASIC: Executing operation assert_that/Create/Read+assert_that/Group/pair_with_0+assert_that/Group/GroupByKey/Reify+assert_that/Group/GroupByKey/Write root: INFO: 2018-07-13T18:06:53.361Z: JOB_MESSAGE_DEBUG: Value "Map(<lambda at sideinputs_test.py:213>)/_UnpickledSideInput(Read.out.0).output" materialized. root: INFO: 2018-07-13T18:06:53.397Z: JOB_MESSAGE_DEBUG: Value "Map(<lambda at sideinputs_test.py:213>)/_UnpickledSideInput(Read.out.1).output" materialized. root: INFO: 2018-07-13T18:06:53.478Z: JOB_MESSAGE_BASIC: Executing operation main input/Read+Map(<lambda at sideinputs_test.py:213>)/Map(<lambda at sideinputs_test.py:213>)+assert_that/WindowInto(WindowIntoFn)+assert_that/ToVoidKey+assert_that/Group/pair_with_1+assert_that/Group/GroupByKey/Reify+assert_that/Group/GroupByKey/Write root: INFO: 2018-07-13T18:07:03.133Z: JOB_MESSAGE_DETAILED: Autoscaling: Raised the number of workers to 0 based on the rate of progress in the currently running step(s). root: INFO: 2018-07-13T18:07:34.533Z: JOB_MESSAGE_DETAILED: Autoscaling: Raised the number of workers to 1 based on the rate of progress in the currently running step(s). root: INFO: 2018-07-13T18:07:34.576Z: JOB_MESSAGE_DETAILED: Autoscaling: Would further reduce the number of workers but reached the minimum number allowed for the job. root: INFO: 2018-07-13T18:08:46.323Z: JOB_MESSAGE_DETAILED: Workers have started successfully. root: INFO: 2018-07-13T18:11:26.770Z: JOB_MESSAGE_ERROR: Traceback (most recent call last): File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", line 642, in do_work work_executor.execute() File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/executor.py", line 156, in execute op.start() File "apache_beam/runners/worker/operations.py", line 344, in apache_beam.runners.worker.operations.DoOperation.start def start(self): File "apache_beam/runners/worker/operations.py", line 345, in apache_beam.runners.worker.operations.DoOperation.start with self.scoped_start_state: File "apache_beam/runners/worker/operations.py", line 350, in apache_beam.runners.worker.operations.DoOperation.start pickler.loads(self.spec.serialized_fn)) File "/usr/local/lib/python2.7/dist-packages/apache_beam/internal/pickler.py", line 238, in loads return dill.loads(s) File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 277, in loads return load(file) File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 266, in load obj = pik.load() File "/usr/lib/python2.7/pickle.py", line 864, in load dispatch[key](self) File "/usr/lib/python2.7/pickle.py", line 1096, in load_global klass = self.find_class(module, name) File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 423, in find_class return StockUnpickler.find_class(self, module, name) File "/usr/lib/python2.7/pickle.py", line 1130, in find_class __import__(module) ImportError: No module named _dill root: INFO: 2018-07-13T18:11:29.950Z: JOB_MESSAGE_ERROR: Traceback (most recent call last): File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", line 642, in do_work work_executor.execute() File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/executor.py", line 156, in execute op.start() File "apache_beam/runners/worker/operations.py", line 344, in apache_beam.runners.worker.operations.DoOperation.start def start(self): File "apache_beam/runners/worker/operations.py", line 345, in apache_beam.runners.worker.operations.DoOperation.start with self.scoped_start_state: File "apache_beam/runners/worker/operations.py", line 350, in apache_beam.runners.worker.operations.DoOperation.start pickler.loads(self.spec.serialized_fn)) File "/usr/local/lib/python2.7/dist-packages/apache_beam/internal/pickler.py", line 238, in loads return dill.loads(s) File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 277, in loads return load(file) File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 266, in load obj = pik.load() File "/usr/lib/python2.7/pickle.py", line 864, in load dispatch[key](self) File "/usr/lib/python2.7/pickle.py", line 1096, in load_global klass = self.find_class(module, name) File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 423, in find_class return StockUnpickler.find_class(self, module, name) File "/usr/lib/python2.7/pickle.py", line 1130, in find_class __import__(module) ImportError: No module named _dill root: INFO: 2018-07-13T18:11:33.161Z: JOB_MESSAGE_ERROR: Traceback (most recent call last): File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", line 642, in do_work work_executor.execute() File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/executor.py", line 156, in execute op.start() File "apache_beam/runners/worker/operations.py", line 344, in apache_beam.runners.worker.operations.DoOperation.start def start(self): File "apache_beam/runners/worker/operations.py", line 345, in apache_beam.runners.worker.operations.DoOperation.start with self.scoped_start_state: File "apache_beam/runners/worker/operations.py", line 350, in apache_beam.runners.worker.operations.DoOperation.start pickler.loads(self.spec.serialized_fn)) File "/usr/local/lib/python2.7/dist-packages/apache_beam/internal/pickler.py", line 238, in loads return dill.loads(s) File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 277, in loads return load(file) File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 266, in load obj = pik.load() File "/usr/lib/python2.7/pickle.py", line 864, in load dispatch[key](self) File "/usr/lib/python2.7/pickle.py", line 1096, in load_global klass = self.find_class(module, name) File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 423, in find_class return StockUnpickler.find_class(self, module, name) File "/usr/lib/python2.7/pickle.py", line 1130, in find_class __import__(module) ImportError: No module named _dill root: INFO: 2018-07-13T18:11:36.347Z: JOB_MESSAGE_ERROR: Traceback (most recent call last): File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", line 642, in do_work work_executor.execute() File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/executor.py", line 156, in execute op.start() File "apache_beam/runners/worker/operations.py", line 344, in apache_beam.runners.worker.operations.DoOperation.start def start(self): File "apache_beam/runners/worker/operations.py", line 345, in apache_beam.runners.worker.operations.DoOperation.start with self.scoped_start_state: File "apache_beam/runners/worker/operations.py", line 350, in apache_beam.runners.worker.operations.DoOperation.start pickler.loads(self.spec.serialized_fn)) File "/usr/local/lib/python2.7/dist-packages/apache_beam/internal/pickler.py", line 238, in loads return dill.loads(s) File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 277, in loads return load(file) File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 266, in load obj = pik.load() File "/usr/lib/python2.7/pickle.py", line 864, in load dispatch[key](self) File "/usr/lib/python2.7/pickle.py", line 1096, in load_global klass = self.find_class(module, name) File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 423, in find_class return StockUnpickler.find_class(self, module, name) File "/usr/lib/python2.7/pickle.py", line 1130, in find_class __import__(module) ImportError: No module named _dill root: INFO: 2018-07-13T18:11:36.397Z: JOB_MESSAGE_DEBUG: Executing failure step failure20 root: INFO: 2018-07-13T18:11:36.444Z: JOB_MESSAGE_ERROR: Workflow failed. Causes: S06:main input/Read+Map(<lambda at sideinputs_test.py:213>)/Map(<lambda at sideinputs_test.py:213>)+assert_that/WindowInto(WindowIntoFn)+assert_that/ToVoidKey+assert_that/Group/pair_with_1+assert_that/Group/GroupByKey/Reify+assert_that/Group/GroupByKey/Write failed., A work item was attempted 4 times without success. Each time the worker eventually lost contact with the service. The work item was attempted on: beamapp-jenkins-071318064-07131106-eubu-harness-030m, beamapp-jenkins-071318064-07131106-eubu-harness-030m, beamapp-jenkins-071318064-07131106-eubu-harness-030m, beamapp-jenkins-071318064-07131106-eubu-harness-030m root: INFO: 2018-07-13T18:11:36.621Z: JOB_MESSAGE_DETAILED: Cleaning up. root: INFO: 2018-07-13T18:11:36.678Z: JOB_MESSAGE_DEBUG: Starting worker pool teardown. root: INFO: 2018-07-13T18:11:36.722Z: JOB_MESSAGE_BASIC: Stopping worker pool... root: INFO: 2018-07-13T18:13:14.536Z: JOB_MESSAGE_DETAILED: Autoscaling: Reduced the number of workers to 0 based on the rate of progress in the currently running step(s). root: INFO: 2018-07-13T18:13:14.587Z: JOB_MESSAGE_BASIC: Worker pool stopped. root: INFO: 2018-07-13T18:13:14.619Z: JOB_MESSAGE_DEBUG: Tearing down pending resources... root: INFO: Job 2018-07-13_11_06_47-6503774580458991719 is in state JOB_STATE_FAILED --------------------- >> end captured logging << --------------------- ---------------------------------------------------------------------- XML: <https://builds.apache.org/job/beam_PostCommit_Py_VR_Dataflow/ws/src/sdks/python/nosetests.xml> ---------------------------------------------------------------------- Ran 16 tests in 763.905s FAILED (errors=15) Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-07-13_11_01_04-8861744825699680641?project=apache-beam-testing. Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-07-13_11_07_57-16361049430650504259?project=apache-beam-testing. Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-07-13_11_01_05-9924871110631356788?project=apache-beam-testing. Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-07-13_11_07_08-3767373755959007140?project=apache-beam-testing. Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-07-13_11_01_05-3715755387236352179?project=apache-beam-testing. Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-07-13_11_07_08-12883390187929332919?project=apache-beam-testing. Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-07-13_11_01_04-15291842371424654171?project=apache-beam-testing. Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-07-13_11_06_58-6497891430942434023?project=apache-beam-testing. Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-07-13_11_01_04-11024953510760491991?project=apache-beam-testing. Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-07-13_11_06_37-6721000336552321767?project=apache-beam-testing. Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-07-13_11_01_04-8375333924129771047?project=apache-beam-testing. Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-07-13_11_07_22-4157069115947863857?project=apache-beam-testing. Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-07-13_11_01_04-8345425708502502648?project=apache-beam-testing. Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-07-13_11_06_47-6503774580458991719?project=apache-beam-testing. Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-07-13_11_01_05-7292488924303189325?project=apache-beam-testing. Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-07-13_11_06_43-5067674007401283983?project=apache-beam-testing. > Task :beam-sdks-python:validatesRunnerTests FAILED :beam-sdks-python:validatesRunnerTests (Thread[Task worker for ':',5,main]) completed. Took 12 mins 45.718 secs. FAILURE: Build failed with an exception. * Where: Build file '<https://builds.apache.org/job/beam_PostCommit_Py_VR_Dataflow/ws/src/sdks/python/build.gradle'> line: 237 * What went wrong: Execution failed for task ':beam-sdks-python:validatesRunnerTests'. > Process 'command 'sh'' finished with non-zero exit value 1 * Try: Run with --stacktrace option to get the stack trace. Run with --debug option to get more log output. Run with --scan to get full insights. * Get more help at https://help.gradle.org Deprecated Gradle features were used in this build, making it incompatible with Gradle 5.0. See https://docs.gradle.org/4.8/userguide/command_line_interface.html#sec:command_line_warnings BUILD FAILED in 13m 14s 3 actionable tasks: 3 executed Publishing build scan... https://gradle.com/s/dodzjwzhubkvm Build step 'Invoke Gradle script' changed build result to FAILURE Build step 'Invoke Gradle script' marked build as failure