See 
<https://builds.apache.org/job/beam_PostRelease_NightlySnapshot/63/display/redirect?page=changes>

Changes:

[rmannibucau] extracting the scheduled executor service in a factory variable 
in SDF

[sidhom] Run NeedsRunner tests from direct runner gradle build

[ccy] Fix issue from incomplete removal of has_cache

[sidhom] Address review comments

[sidhom] Remove old sourceSets.test.output references

[robertwb] Avoid warning in our default runner.

[github] [BEAM-3719] Adds support for reading side-inputs from SDFs

[github] print() is a function in Python 3

[robertwb] [maven-release-plugin] prepare branch release-2.4.0

[robertwb] [maven-release-plugin] prepare for next development iteration

[robertwb] Bump Python dev version.

------------------------------------------
[...truncated 3.63 MB...]
        at 
org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)
        at 
org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)
        at 
org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)
        at 
org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)
        at 
org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)
        at 
org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:96)
        at 
org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:53)
        at org.apache.spark.scheduler.Task.run(Task.scala:108)
        at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:338)
        at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
        at java.lang.Thread.run(Thread.java:748)

Mar 02, 2018 11:04:33 AM org.apache.spark.internal.Logging$class logError
SEVERE: Task 3 in stage 0.0 failed 1 times; aborting job
Mar 02, 2018 11:04:33 AM org.apache.spark.internal.Logging$class logInfo
INFO: Removed TaskSet 0.0, whose tasks have all completed, from pool 
Mar 02, 2018 11:04:33 AM org.apache.spark.internal.Logging$class logInfo
INFO: Lost task 0.0 in stage 0.0 (TID 0) on localhost, executor driver: 
java.lang.NoSuchMethodError 
(org.apache.beam.runners.core.DoFnRunners.simpleRunner(Lorg/apache/beam/sdk/options/PipelineOptions;Lorg/apache/beam/sdk/transforms/DoFn;Lorg/apache/beam/runners/core/SideInputReader;Lorg/apache/beam/runners/core/DoFnRunners$OutputManager;Lorg/apache/beam/sdk/values/TupleTag;Ljava/util/List;Lorg/apache/beam/runners/core/StepContext;Lorg/apache/beam/sdk/values/WindowingStrategy;)Lorg/apache/beam/runners/core/DoFnRunner;)
 [duplicate 1]
Mar 02, 2018 11:04:33 AM org.apache.spark.internal.Logging$class logInfo
INFO: Removed TaskSet 0.0, whose tasks have all completed, from pool 
Mar 02, 2018 11:04:33 AM org.apache.spark.internal.Logging$class logInfo
INFO: Lost task 1.0 in stage 0.0 (TID 1) on localhost, executor driver: 
java.lang.NoSuchMethodError 
(org.apache.beam.runners.core.DoFnRunners.simpleRunner(Lorg/apache/beam/sdk/options/PipelineOptions;Lorg/apache/beam/sdk/transforms/DoFn;Lorg/apache/beam/runners/core/SideInputReader;Lorg/apache/beam/runners/core/DoFnRunners$OutputManager;Lorg/apache/beam/sdk/values/TupleTag;Ljava/util/List;Lorg/apache/beam/runners/core/StepContext;Lorg/apache/beam/sdk/values/WindowingStrategy;)Lorg/apache/beam/runners/core/DoFnRunner;)
 [duplicate 2]
Mar 02, 2018 11:04:33 AM org.apache.spark.internal.Logging$class logInfo
INFO: Removed TaskSet 0.0, whose tasks have all completed, from pool 
Mar 02, 2018 11:04:33 AM org.apache.spark.internal.Logging$class logInfo
INFO: Lost task 2.0 in stage 0.0 (TID 2) on localhost, executor driver: 
java.lang.NoSuchMethodError 
(org.apache.beam.runners.core.DoFnRunners.simpleRunner(Lorg/apache/beam/sdk/options/PipelineOptions;Lorg/apache/beam/sdk/transforms/DoFn;Lorg/apache/beam/runners/core/SideInputReader;Lorg/apache/beam/runners/core/DoFnRunners$OutputManager;Lorg/apache/beam/sdk/values/TupleTag;Ljava/util/List;Lorg/apache/beam/runners/core/StepContext;Lorg/apache/beam/sdk/values/WindowingStrategy;)Lorg/apache/beam/runners/core/DoFnRunner;)
 [duplicate 3]
Mar 02, 2018 11:04:33 AM org.apache.spark.internal.Logging$class logInfo
INFO: Removed TaskSet 0.0, whose tasks have all completed, from pool 
Mar 02, 2018 11:04:33 AM org.apache.spark.internal.Logging$class logInfo
INFO: Cancelling stage 0
Mar 02, 2018 11:04:33 AM org.apache.spark.internal.Logging$class logInfo
INFO: ShuffleMapStage 0 (mapToPair at GroupCombineFunctions.java:184) failed in 
2.081 s due to Job aborted due to stage failure: Task 3 in stage 0.0 failed 1 
times, most recent failure: Lost task 3.0 in stage 0.0 (TID 3, localhost, 
executor driver): java.lang.NoSuchMethodError: 
org.apache.beam.runners.core.DoFnRunners.simpleRunner(Lorg/apache/beam/sdk/options/PipelineOptions;Lorg/apache/beam/sdk/transforms/DoFn;Lorg/apache/beam/runners/core/SideInputReader;Lorg/apache/beam/runners/core/DoFnRunners$OutputManager;Lorg/apache/beam/sdk/values/TupleTag;Ljava/util/List;Lorg/apache/beam/runners/core/StepContext;Lorg/apache/beam/sdk/values/WindowingStrategy;)Lorg/apache/beam/runners/core/DoFnRunner;
        at 
org.apache.beam.runners.spark.translation.MultiDoFnFunction.call(MultiDoFnFunction.java:137)
        at 
org.apache.beam.runners.spark.translation.MultiDoFnFunction.call(MultiDoFnFunction.java:58)
        at 
org.apache.spark.api.java.JavaRDDLike$$anonfun$fn$7$1.apply(JavaRDDLike.scala:186)
        at 
org.apache.spark.api.java.JavaRDDLike$$anonfun$fn$7$1.apply(JavaRDDLike.scala:186)
        at 
org.apache.spark.rdd.RDD$$anonfun$mapPartitions$1$$anonfun$apply$23.apply(RDD.scala:797)
        at 
org.apache.spark.rdd.RDD$$anonfun$mapPartitions$1$$anonfun$apply$23.apply(RDD.scala:797)
        at 
org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)
        at 
org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)
        at 
org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)
        at 
org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)
        at 
org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)
        at 
org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)
        at 
org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)
        at 
org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)
        at 
org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:96)
        at 
org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:53)
        at org.apache.spark.scheduler.Task.run(Task.scala:108)
        at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:338)
        at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
        at java.lang.Thread.run(Thread.java:748)

Driver stacktrace:
Mar 02, 2018 11:04:33 AM org.apache.spark.internal.Logging$class logInfo
INFO: Job 0 failed: collect at BoundedDataset.java:87, took 2.409393 s
Mar 02, 2018 11:04:33 AM org.spark_project.jetty.server.AbstractConnector doStop
INFO: Stopped Spark@54524a6a{HTTP/1.1,[http/1.1]}{127.0.0.1:4040}
Mar 02, 2018 11:04:33 AM org.apache.spark.internal.Logging$class logInfo
INFO: Stopped Spark web UI at http://127.0.0.1:4040
Mar 02, 2018 11:04:33 AM org.apache.spark.internal.Logging$class logInfo
INFO: MapOutputTrackerMasterEndpoint stopped!
Mar 02, 2018 11:04:33 AM org.apache.spark.internal.Logging$class logInfo
INFO: MemoryStore cleared
Mar 02, 2018 11:04:33 AM org.apache.spark.internal.Logging$class logInfo
INFO: BlockManager stopped
Mar 02, 2018 11:04:33 AM org.apache.spark.internal.Logging$class logInfo
INFO: BlockManagerMaster stopped
Mar 02, 2018 11:04:33 AM org.apache.spark.internal.Logging$class logInfo
INFO: OutputCommitCoordinator stopped!
Mar 02, 2018 11:04:33 AM org.apache.spark.internal.Logging$class logInfo
INFO: Successfully stopped SparkContext
[WARNING] 
org.apache.beam.sdk.Pipeline$PipelineExecutionException: 
java.lang.NoSuchMethodError: 
org.apache.beam.runners.core.DoFnRunners.simpleRunner(Lorg/apache/beam/sdk/options/PipelineOptions;Lorg/apache/beam/sdk/transforms/DoFn;Lorg/apache/beam/runners/core/SideInputReader;Lorg/apache/beam/runners/core/DoFnRunners$OutputManager;Lorg/apache/beam/sdk/values/TupleTag;Ljava/util/List;Lorg/apache/beam/runners/core/StepContext;Lorg/apache/beam/sdk/values/WindowingStrategy;)Lorg/apache/beam/runners/core/DoFnRunner;
    at org.apache.beam.runners.spark.SparkPipelineResult.beamExceptionFrom 
(SparkPipelineResult.java:68)
    at org.apache.beam.runners.spark.SparkPipelineResult.waitUntilFinish 
(SparkPipelineResult.java:99)
    at org.apache.beam.runners.spark.SparkPipelineResult.waitUntilFinish 
(SparkPipelineResult.java:87)
    at org.apache.beam.examples.WordCount.main (WordCount.java:187)
    at sun.reflect.NativeMethodAccessorImpl.invoke0 (Native Method)
    at sun.reflect.NativeMethodAccessorImpl.invoke 
(NativeMethodAccessorImpl.java:62)
    at sun.reflect.DelegatingMethodAccessorImpl.invoke 
(DelegatingMethodAccessorImpl.java:43)
    at java.lang.reflect.Method.invoke (Method.java:498)
    at org.codehaus.mojo.exec.ExecJavaMojo$1.run (ExecJavaMojo.java:282)
    at java.lang.Thread.run (Thread.java:748)
Caused by: java.lang.NoSuchMethodError: 
org.apache.beam.runners.core.DoFnRunners.simpleRunner(Lorg/apache/beam/sdk/options/PipelineOptions;Lorg/apache/beam/sdk/transforms/DoFn;Lorg/apache/beam/runners/core/SideInputReader;Lorg/apache/beam/runners/core/DoFnRunners$OutputManager;Lorg/apache/beam/sdk/values/TupleTag;Ljava/util/List;Lorg/apache/beam/runners/core/StepContext;Lorg/apache/beam/sdk/values/WindowingStrategy;)Lorg/apache/beam/runners/core/DoFnRunner;
    at org.apache.beam.runners.spark.translation.MultiDoFnFunction.call 
(MultiDoFnFunction.java:137)
    at org.apache.beam.runners.spark.translation.MultiDoFnFunction.call 
(MultiDoFnFunction.java:58)
    at org.apache.spark.api.java.JavaRDDLike$$anonfun$fn$7$1.apply 
(JavaRDDLike.scala:186)
    at org.apache.spark.api.java.JavaRDDLike$$anonfun$fn$7$1.apply 
(JavaRDDLike.scala:186)
    at 
org.apache.spark.rdd.RDD$$anonfun$mapPartitions$1$$anonfun$apply$23.apply 
(RDD.scala:797)
    at 
org.apache.spark.rdd.RDD$$anonfun$mapPartitions$1$$anonfun$apply$23.apply 
(RDD.scala:797)
    at org.apache.spark.rdd.MapPartitionsRDD.compute (MapPartitionsRDD.scala:38)
    at org.apache.spark.rdd.RDD.computeOrReadCheckpoint (RDD.scala:323)
    at org.apache.spark.rdd.RDD.iterator (RDD.scala:287)
    at org.apache.spark.rdd.MapPartitionsRDD.compute (MapPartitionsRDD.scala:38)
    at org.apache.spark.rdd.RDD.computeOrReadCheckpoint (RDD.scala:323)
    at org.apache.spark.rdd.RDD.iterator (RDD.scala:287)
    at org.apache.spark.rdd.MapPartitionsRDD.compute (MapPartitionsRDD.scala:38)
    at org.apache.spark.rdd.RDD.computeOrReadCheckpoint (RDD.scala:323)
    at org.apache.spark.rdd.RDD.iterator (RDD.scala:287)
    at org.apache.spark.rdd.MapPartitionsRDD.compute (MapPartitionsRDD.scala:38)
    at org.apache.spark.rdd.RDD.computeOrReadCheckpoint (RDD.scala:323)
    at org.apache.spark.rdd.RDD.iterator (RDD.scala:287)
    at org.apache.spark.rdd.MapPartitionsRDD.compute (MapPartitionsRDD.scala:38)
    at org.apache.spark.rdd.RDD.computeOrReadCheckpoint (RDD.scala:323)
    at org.apache.spark.rdd.RDD.iterator (RDD.scala:287)
    at org.apache.spark.rdd.MapPartitionsRDD.compute (MapPartitionsRDD.scala:38)
    at org.apache.spark.rdd.RDD.computeOrReadCheckpoint (RDD.scala:323)
    at org.apache.spark.rdd.RDD.iterator (RDD.scala:287)
    at org.apache.spark.rdd.MapPartitionsRDD.compute (MapPartitionsRDD.scala:38)
    at org.apache.spark.rdd.RDD.computeOrReadCheckpoint (RDD.scala:323)
    at org.apache.spark.rdd.RDD.iterator (RDD.scala:287)
    at org.apache.spark.rdd.MapPartitionsRDD.compute (MapPartitionsRDD.scala:38)
    at org.apache.spark.rdd.RDD.computeOrReadCheckpoint (RDD.scala:323)
    at org.apache.spark.rdd.RDD.iterator (RDD.scala:287)
    at org.apache.spark.scheduler.ShuffleMapTask.runTask 
(ShuffleMapTask.scala:96)
    at org.apache.spark.scheduler.ShuffleMapTask.runTask 
(ShuffleMapTask.scala:53)
    at org.apache.spark.scheduler.Task.run (Task.scala:108)
    at org.apache.spark.executor.Executor$TaskRunner.run (Executor.scala:338)
    at java.util.concurrent.ThreadPoolExecutor.runWorker 
(ThreadPoolExecutor.java:1149)
    at java.util.concurrent.ThreadPoolExecutor$Worker.run 
(ThreadPoolExecutor.java:624)
    at java.lang.Thread.run (Thread.java:748)
[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 01:44 min
[INFO] Finished at: 2018-03-02T11:04:33Z
[INFO] Final Memory: 87M/836M
[INFO] ------------------------------------------------------------------------
[ERROR] Failed to execute goal org.codehaus.mojo:exec-maven-plugin:1.6.0:java 
(default-cli) on project word-count-beam: An exception occured while executing 
the Java class. java.lang.NoSuchMethodError: 
org.apache.beam.runners.core.DoFnRunners.simpleRunner(Lorg/apache/beam/sdk/options/PipelineOptions;Lorg/apache/beam/sdk/transforms/DoFn;Lorg/apache/beam/runners/core/SideInputReader;Lorg/apache/beam/runners/core/DoFnRunners$OutputManager;Lorg/apache/beam/sdk/values/TupleTag;Ljava/util/List;Lorg/apache/beam/runners/core/StepContext;Lorg/apache/beam/sdk/values/WindowingStrategy;)Lorg/apache/beam/runners/core/DoFnRunner;
 -> [Help 1]
[ERROR] 
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e 
switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR] 
[ERROR] For more information about the errors and possible solutions, please 
read the following articles:
[ERROR] [Help 1] 
http://cwiki.apache.org/confluence/display/MAVEN/MojoExecutionException
[ERROR] Failed command
:runners:spark:runQuickstartJavaSpark FAILED
Mar 02, 2018 11:04:36 AM 
org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2018-03-02T11:04:35.954Z: (1329671cafed2f34): Executing operation 
WordCount.CountWords/Count.PerElement/Combine.perKey(Count)/GroupByKey/Close
Mar 02, 2018 11:04:36 AM 
org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2018-03-02T11:04:36.011Z: (1329671cafed2acc): Executing operation 
WriteCounts/WriteFiles/WriteUnshardedBundlesToTempFiles/GroupUnwritten/Create
Mar 02, 2018 11:04:36 AM 
org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2018-03-02T11:04:36.141Z: (1329671cafed2abd): Executing operation 
WordCount.CountWords/Count.PerElement/Combine.perKey(Count)/GroupByKey/Read+WordCount.CountWords/Count.PerElement/Combine.perKey(Count)/Combine.GroupedValues+WordCount.CountWords/Count.PerElement/Combine.perKey(Count)/Combine.GroupedValues/Extract+MapElements/Map+WriteCounts/WriteFiles/RewindowIntoGlobal/Window.Assign+WriteCounts/WriteFiles/WriteUnshardedBundlesToTempFiles/WriteUnshardedBundles+WriteCounts/WriteFiles/GatherTempFileResults/View.AsList/ParDo(ToIsmRecordForGlobalWindow)+WriteCounts/WriteFiles/WriteUnshardedBundlesToTempFiles/GroupUnwritten/Reify+WriteCounts/WriteFiles/WriteUnshardedBundlesToTempFiles/GroupUnwritten/Write
Mar 02, 2018 11:04:46 AM 
org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2018-03-02T11:04:46.669Z: (1329671cafed21ed): Executing operation 
WriteCounts/WriteFiles/WriteUnshardedBundlesToTempFiles/GroupUnwritten/Close
Mar 02, 2018 11:04:46 AM 
org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2018-03-02T11:04:46.733Z: (1329671cafed2c12): Executing operation 
WriteCounts/WriteFiles/WriteUnshardedBundlesToTempFiles/GroupUnwritten/Read+WriteCounts/WriteFiles/WriteUnshardedBundlesToTempFiles/GroupUnwritten/GroupByWindow+WriteCounts/WriteFiles/WriteUnshardedBundlesToTempFiles/WriteUnwritten+WriteCounts/WriteFiles/WriteUnshardedBundlesToTempFiles/DropShardNum+WriteCounts/WriteFiles/GatherTempFileResults/View.AsList/ParDo(ToIsmRecordForGlobalWindow)
Mar 02, 2018 11:04:49 AM 
org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2018-03-02T11:04:48.816Z: (556da1fb7f20c8bc): Executing operation s12-u31
Mar 02, 2018 11:04:49 AM 
org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2018-03-02T11:04:49.029Z: (1329671cafed2ecb): Executing operation 
WriteCounts/WriteFiles/GatherTempFileResults/View.AsList/CreateDataflowView
Mar 02, 2018 11:04:49 AM 
org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2018-03-02T11:04:49.222Z: (556da1fb7f20c5d9): Executing operation 
WriteCounts/WriteFiles/GatherTempFileResults/Reify.ReifyViewInGlobalWindow/Create.Values/Read(CreateSource)+WriteCounts/WriteFiles/GatherTempFileResults/Reify.ReifyViewInGlobalWindow/Reify.ReifyView/ParDo(Anonymous)+WriteCounts/WriteFiles/GatherTempFileResults/Reify.ReifyViewInGlobalWindow/Values/Values/Map+WriteCounts/WriteFiles/FinalizeTempFileBundles/Finalize
Mar 02, 2018 11:04:51 AM 
org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2018-03-02T11:04:50.985Z: (25f20689db9d9e37): Cleaning up.
Mar 02, 2018 11:04:51 AM 
org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2018-03-02T11:04:51.067Z: (25f20689db9d9f01): Stopping worker pool...
Mar 02, 2018 11:07:06 AM 
org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2018-03-02T11:07:04.059Z: (2991a79c2f0cd870): Autoscaling: Resized worker 
pool from 1 to 0.
Mar 02, 2018 11:07:06 AM 
org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2018-03-02T11:07:04.085Z: (2991a79c2f0cdaba): Autoscaling: Would further 
reduce the number of workers but reached the minimum number allowed for the job.
Mar 02, 2018 11:07:14 AM org.apache.beam.runners.dataflow.DataflowPipelineJob 
waitUntilFinish
INFO: Job 2018-03-02_03_02_32-15408355914220323339 finished with status DONE.
[INFO] ------------------------------------------------------------------------
[INFO] BUILD SUCCESS
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 05:58 min
[INFO] Finished at: 2018-03-02T11:07:14Z
[INFO] Final Memory: 58M/694M
[INFO] ------------------------------------------------------------------------
gsutil cat gs://temp-storage-for-release-validation-tests/quickstart/count* | 
grep Montague:
Montague: 47
Verified Montague: 47
gsutil rm gs://temp-storage-for-release-validation-tests/quickstart/count*
Removing 
gs://temp-storage-for-release-validation-tests/quickstart/counts-00000-of-00003...
Removing 
gs://temp-storage-for-release-validation-tests/quickstart/counts-00001-of-00003...
Removing 
gs://temp-storage-for-release-validation-tests/quickstart/counts-00002-of-00003...
[SUCCESS]

FAILURE: Build completed with 3 failures.

1: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':runners:flink:runQuickstartJavaFlinkLocal'.
> Process 'command '/usr/local/asfpackages/java/jdk1.8.0_152/bin/java'' 
> finished with non-zero exit value 1

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug 
option to get more log output. Run with --scan to get full insights.
==============================================================================

2: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':runners:apex:runQuickstartJavaApex'.
> Process 'command '/usr/local/asfpackages/java/jdk1.8.0_152/bin/java'' 
> finished with non-zero exit value 1

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug 
option to get more log output. Run with --scan to get full insights.
==============================================================================

3: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':runners:spark:runQuickstartJavaSpark'.
> Process 'command '/usr/local/asfpackages/java/jdk1.8.0_152/bin/java'' 
> finished with non-zero exit value 1

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug 
option to get more log output. Run with --scan to get full insights.
==============================================================================

* Get more help at https://help.gradle.org

BUILD FAILED in 6m 51s
6 actionable tasks: 6 executed
Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure
Not sending mail to unregistered user sid...@google.com
Not sending mail to unregistered user c...@google.com
Not sending mail to unregistered user git...@alasdairhodge.co.uk

Reply via email to