See <https://builds.apache.org/job/beam_PostCommit_Java_GradleBuild/240/display/redirect?page=changes>
Changes: [herohde] [BEAM-4175] Fix direct ouput pardo issue ------------------------------------------ [...truncated 19.14 MB...] Apr 30, 2018 11:35:05 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process INFO: 2018-04-30T23:35:03.318Z: Workers have started successfully. org.apache.beam.sdk.io.gcp.datastore.V1ReadIT > testE2EV1ReadWithGQLQueryWithNoLimit STANDARD_ERROR Apr 30, 2018 11:35:11 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process INFO: 2018-04-30T23:35:10.833Z: Autoscaling: Raised the number of workers to 5 based on the rate of progress in the currently running step(s). Apr 30, 2018 11:35:28 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process INFO: 2018-04-30T23:35:27.505Z: Workers have started successfully. org.apache.beam.sdk.io.gcp.datastore.V1WriteIT > testE2EV1Write STANDARD_ERROR Apr 30, 2018 11:35:36 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process INFO: 2018-04-30T23:35:35.302Z: Cleaning up. Apr 30, 2018 11:35:36 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process INFO: 2018-04-30T23:35:35.411Z: Stopping worker pool... org.apache.beam.sdk.io.gcp.datastore.V1ReadIT > testE2EV1ReadWithGQLQueryWithNoLimit STANDARD_ERROR Apr 30, 2018 11:35:54 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process INFO: 2018-04-30T23:35:53.923Z: Executing operation DatastoreV1.Read/Reshuffle/Reshuffle/GroupByKey/Close Apr 30, 2018 11:35:54 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process INFO: 2018-04-30T23:35:53.999Z: Executing operation DatastoreV1.Read/Reshuffle/Reshuffle/GroupByKey/Read+DatastoreV1.Read/Reshuffle/Reshuffle/GroupByKey/GroupByWindow+DatastoreV1.Read/Reshuffle/Reshuffle/ExpandIterable+DatastoreV1.Read/Reshuffle/Values/Values/Map+DatastoreV1.Read/Read+Combine.globally(Count)/WithKeys/AddKeys/Map+Combine.globally(Count)/Combine.perKey(Count)/GroupByKey+Combine.globally(Count)/Combine.perKey(Count)/Combine.GroupedValues/Partial+Combine.globally(Count)/Combine.perKey(Count)/GroupByKey/Reify+Combine.globally(Count)/Combine.perKey(Count)/GroupByKey/Write Apr 30, 2018 11:36:04 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process INFO: 2018-04-30T23:36:03.420Z: Executing operation Combine.globally(Count)/Combine.perKey(Count)/GroupByKey/Close Apr 30, 2018 11:36:04 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process INFO: 2018-04-30T23:36:03.506Z: Executing operation Combine.globally(Count)/Combine.perKey(Count)/GroupByKey/Read+Combine.globally(Count)/Combine.perKey(Count)/Combine.GroupedValues+Combine.globally(Count)/Combine.perKey(Count)/Combine.GroupedValues/Extract+Combine.globally(Count)/Values/Values/Map+PAssert$3/CreateActual/FilterActuals/Window.Assign+PAssert$3/CreateActual/GatherPanes/Reify.Window/ParDo(Anonymous)+PAssert$3/CreateActual/GatherPanes/WithKeys/AddKeys/Map+PAssert$3/CreateActual/GatherPanes/Window.Into()/Window.Assign+PAssert$3/CreateActual/GatherPanes/GroupByKey/Reify+PAssert$3/CreateActual/GatherPanes/GroupByKey/Write+Combine.globally(Count)/View.AsIterable/ParDo(ToIsmRecordForGlobalWindow) Apr 30, 2018 11:36:15 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process INFO: 2018-04-30T23:36:14.393Z: Executing operation Combine.globally(Count)/View.AsIterable/CreateDataflowView Apr 30, 2018 11:36:15 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process INFO: 2018-04-30T23:36:14.607Z: Executing operation Combine.globally(Count)/CreateVoid/Read(CreateSource)+Combine.globally(Count)/ProduceDefault+PAssert$3/CreateActual/FilterActuals/Window.Assign+PAssert$3/CreateActual/GatherPanes/Reify.Window/ParDo(Anonymous)+PAssert$3/CreateActual/GatherPanes/WithKeys/AddKeys/Map+PAssert$3/CreateActual/GatherPanes/Window.Into()/Window.Assign+PAssert$3/CreateActual/GatherPanes/GroupByKey/Reify+PAssert$3/CreateActual/GatherPanes/GroupByKey/Write Apr 30, 2018 11:36:19 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process INFO: 2018-04-30T23:36:17.112Z: Executing operation PAssert$3/CreateActual/GatherPanes/GroupByKey/Close Apr 30, 2018 11:36:19 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process INFO: 2018-04-30T23:36:17.206Z: Executing operation PAssert$3/CreateActual/GatherPanes/GroupByKey/Read+PAssert$3/CreateActual/GatherPanes/GroupByKey/GroupByWindow+PAssert$3/CreateActual/GatherPanes/Values/Values/Map+PAssert$3/CreateActual/ExtractPane/Map+PAssert$3/CreateActual/Flatten.Iterables/FlattenIterables/FlatMap+PAssert$3/CreateActual/RewindowActuals/Window.Assign+PAssert$3/CreateActual/ParDo(Anonymous)+PAssert$3/CreateActual/View.AsSingleton/Combine.GloballyAsSingletonView/Combine.globally(Singleton)/WithKeys/AddKeys/Map+PAssert$3/CreateActual/View.AsSingleton/Combine.GloballyAsSingletonView/Combine.globally(Singleton)/Combine.perKey(Singleton)/GroupByKey+PAssert$3/CreateActual/View.AsSingleton/Combine.GloballyAsSingletonView/Combine.globally(Singleton)/Combine.perKey(Singleton)/Combine.GroupedValues/Partial+PAssert$3/CreateActual/View.AsSingleton/Combine.GloballyAsSingletonView/Combine.globally(Singleton)/Combine.perKey(Singleton)/GroupByKey/Reify+PAssert$3/CreateActual/View.AsSingleton/Combine.GloballyAsSingletonView/Combine.globally(Singleton)/Combine.perKey(Singleton)/GroupByKey/Write Apr 30, 2018 11:36:19 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process INFO: 2018-04-30T23:36:19.192Z: Executing operation PAssert$3/CreateActual/View.AsSingleton/Combine.GloballyAsSingletonView/Combine.globally(Singleton)/Combine.perKey(Singleton)/GroupByKey/Close Apr 30, 2018 11:36:19 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process INFO: 2018-04-30T23:36:19.289Z: Executing operation PAssert$3/CreateActual/View.AsSingleton/Combine.GloballyAsSingletonView/Combine.globally(Singleton)/Combine.perKey(Singleton)/GroupByKey/Read+PAssert$3/CreateActual/View.AsSingleton/Combine.GloballyAsSingletonView/Combine.globally(Singleton)/Combine.perKey(Singleton)/Combine.GroupedValues+PAssert$3/CreateActual/View.AsSingleton/Combine.GloballyAsSingletonView/Combine.globally(Singleton)/Combine.perKey(Singleton)/Combine.GroupedValues/Extract+PAssert$3/CreateActual/View.AsSingleton/Combine.GloballyAsSingletonView/Combine.globally(Singleton)/Values/Values/Map+PAssert$3/CreateActual/View.AsSingleton/Combine.GloballyAsSingletonView/BatchViewOverrides.GroupByWindowHashAsKeyAndWindowAsSortKey/ParDo(UseWindowHashAsKeyAndWindowAsSortKey)+PAssert$3/CreateActual/View.AsSingleton/Combine.GloballyAsSingletonView/BatchViewOverrides.GroupByWindowHashAsKeyAndWindowAsSortKey/BatchViewOverrides.GroupByKeyAndSortValuesOnly/Write Apr 30, 2018 11:36:22 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process INFO: 2018-04-30T23:36:20.921Z: Executing operation PAssert$3/CreateActual/View.AsSingleton/Combine.GloballyAsSingletonView/BatchViewOverrides.GroupByWindowHashAsKeyAndWindowAsSortKey/BatchViewOverrides.GroupByKeyAndSortValuesOnly/Close Apr 30, 2018 11:36:22 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process INFO: 2018-04-30T23:36:21.014Z: Executing operation PAssert$3/CreateActual/View.AsSingleton/Combine.GloballyAsSingletonView/BatchViewOverrides.GroupByWindowHashAsKeyAndWindowAsSortKey/BatchViewOverrides.GroupByKeyAndSortValuesOnly/Read+PAssert$3/CreateActual/View.AsSingleton/Combine.GloballyAsSingletonView/ParDo(IsmRecordForSingularValuePerWindow) Apr 30, 2018 11:36:27 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process INFO: 2018-04-30T23:36:24.610Z: Executing operation PAssert$3/CreateActual/View.AsSingleton/Combine.GloballyAsSingletonView/CreateDataflowView Apr 30, 2018 11:36:27 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process INFO: 2018-04-30T23:36:24.812Z: Executing operation PAssert$3/Create.Values/Read(CreateSource)+PAssert$3/WindowToken/Window.Assign+PAssert$3/RunChecks+PAssert$3/VerifyAssertions/ParDo(DefaultConclude) Apr 30, 2018 11:36:27 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process INFO: 2018-04-30T23:36:26.940Z: Cleaning up. Apr 30, 2018 11:36:27 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process INFO: 2018-04-30T23:36:27.134Z: Stopping worker pool... org.apache.beam.sdk.io.gcp.datastore.V1WriteIT > testE2EV1Write STANDARD_ERROR Apr 30, 2018 11:37:32 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process INFO: 2018-04-30T23:37:32.611Z: Autoscaling: Resized worker pool from 1 to 0. Apr 30, 2018 11:37:32 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process INFO: 2018-04-30T23:37:32.643Z: Autoscaling: Would further reduce the number of workers but reached the minimum number allowed for the job. Apr 30, 2018 11:37:32 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process INFO: 2018-04-30T23:37:32.689Z: Worker pool stopped. Apr 30, 2018 11:37:40 PM org.apache.beam.runners.dataflow.DataflowPipelineJob waitUntilFinish INFO: Job 2018-04-30_16_34_07-2128685051185044055 finished with status DONE. Apr 30, 2018 11:37:40 PM org.apache.beam.runners.dataflow.TestDataflowRunner checkForPAssertSuccess INFO: Success result for Dataflow job 2018-04-30_16_34_07-2128685051185044055. Found 0 success, 0 failures out of 0 expected assertions. Apr 30, 2018 11:37:41 PM org.apache.beam.sdk.io.gcp.datastore.V1TestUtil$V1TestWriter flushBatch INFO: Writing batch of 500 entities Apr 30, 2018 11:37:42 PM org.apache.beam.sdk.io.gcp.datastore.V1TestUtil$V1TestWriter flushBatch INFO: Successfully wrote 500 entities Apr 30, 2018 11:37:42 PM org.apache.beam.sdk.io.gcp.datastore.V1TestUtil$V1TestWriter flushBatch INFO: Writing batch of 500 entities Apr 30, 2018 11:37:42 PM org.apache.beam.sdk.io.gcp.datastore.V1TestUtil$V1TestWriter flushBatch INFO: Successfully wrote 500 entities Apr 30, 2018 11:37:42 PM org.apache.beam.sdk.io.gcp.datastore.V1TestUtil deleteAllEntities INFO: Successfully deleted 1000 entities Gradle Test Executor 130 finished executing tests. > Task > :beam-runners-google-cloud-dataflow-java:googleCloudPlatformIntegrationTest org.apache.beam.sdk.io.gcp.spanner.SpannerReadIT > testQuery FAILED java.lang.NoClassDefFoundError: com/google/api/gax/retrying/ExceptionRetryAlgorithm at java.lang.ClassLoader.defineClass1(Native Method) at java.lang.ClassLoader.defineClass(ClassLoader.java:763) at java.security.SecureClassLoader.defineClass(SecureClassLoader.java:142) at java.net.URLClassLoader.defineClass(URLClassLoader.java:467) at java.net.URLClassLoader.access$100(URLClassLoader.java:73) at java.net.URLClassLoader$1.run(URLClassLoader.java:368) at java.net.URLClassLoader$1.run(URLClassLoader.java:362) at java.security.AccessController.doPrivileged(Native Method) at java.net.URLClassLoader.findClass(URLClassLoader.java:361) at java.lang.ClassLoader.loadClass(ClassLoader.java:424) at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:338) at java.lang.ClassLoader.loadClass(ClassLoader.java:357) at com.google.cloud.BaseService.<clinit>(BaseService.java:48) at com.google.cloud.spanner.SpannerOptions$DefaultSpannerFactory.create(SpannerOptions.java:60) at com.google.cloud.spanner.SpannerOptions$DefaultSpannerFactory.create(SpannerOptions.java:55) at com.google.cloud.ServiceOptions.getService(ServiceOptions.java:426) at org.apache.beam.sdk.io.gcp.spanner.SpannerReadIT.setUp(SpannerReadIT.java:93) Caused by: java.lang.ClassNotFoundException: com.google.api.gax.retrying.ExceptionRetryAlgorithm at java.net.URLClassLoader.findClass(URLClassLoader.java:381) at java.lang.ClassLoader.loadClass(ClassLoader.java:424) at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:338) at java.lang.ClassLoader.loadClass(ClassLoader.java:357) ... 17 more java.lang.NullPointerException at org.apache.beam.sdk.io.gcp.spanner.SpannerReadIT.tearDown(SpannerReadIT.java:217) org.apache.beam.sdk.io.gcp.spanner.SpannerReadIT > testRead FAILED java.lang.NoClassDefFoundError: Could not initialize class com.google.cloud.spanner.SpannerImpl at com.google.cloud.spanner.SpannerOptions$DefaultSpannerFactory.create(SpannerOptions.java:60) at com.google.cloud.spanner.SpannerOptions$DefaultSpannerFactory.create(SpannerOptions.java:55) at com.google.cloud.ServiceOptions.getService(ServiceOptions.java:426) at org.apache.beam.sdk.io.gcp.spanner.SpannerReadIT.setUp(SpannerReadIT.java:93) java.lang.NullPointerException at org.apache.beam.sdk.io.gcp.spanner.SpannerReadIT.tearDown(SpannerReadIT.java:217) org.apache.beam.sdk.io.gcp.spanner.SpannerReadIT > testReadAllRecordsInDb FAILED java.lang.NoClassDefFoundError: Could not initialize class com.google.cloud.spanner.SpannerImpl at com.google.cloud.spanner.SpannerOptions$DefaultSpannerFactory.create(SpannerOptions.java:60) at com.google.cloud.spanner.SpannerOptions$DefaultSpannerFactory.create(SpannerOptions.java:55) at com.google.cloud.ServiceOptions.getService(ServiceOptions.java:426) at org.apache.beam.sdk.io.gcp.spanner.SpannerReadIT.setUp(SpannerReadIT.java:93) java.lang.NullPointerException at org.apache.beam.sdk.io.gcp.spanner.SpannerReadIT.tearDown(SpannerReadIT.java:217) org.apache.beam.sdk.io.gcp.datastore.V1ReadIT > testE2EV1ReadWithGQLQueryWithNoLimit STANDARD_ERROR Apr 30, 2018 11:38:15 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process INFO: 2018-04-30T23:38:14.329Z: Autoscaling: Resized worker pool from 5 to 0. Apr 30, 2018 11:38:15 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process INFO: 2018-04-30T23:38:14.363Z: Autoscaling: Would further reduce the number of workers but reached the minimum number allowed for the job. Apr 30, 2018 11:38:15 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process INFO: 2018-04-30T23:38:14.412Z: Worker pool stopped. Apr 30, 2018 11:38:23 PM org.apache.beam.runners.dataflow.DataflowPipelineJob waitUntilFinish INFO: Job 2018-04-30_16_34_02-14624331325458396126 finished with status DONE. Apr 30, 2018 11:38:23 PM org.apache.beam.runners.dataflow.TestDataflowRunner checkForPAssertSuccess INFO: Success result for Dataflow job 2018-04-30_16_34_02-14624331325458396126. Found 1 success, 0 failures out of 1 expected assertions. Apr 30, 2018 11:38:24 PM org.apache.beam.sdk.io.gcp.datastore.V1TestUtil$V1TestWriter flushBatch INFO: Writing batch of 500 entities Apr 30, 2018 11:38:24 PM org.apache.beam.sdk.io.gcp.datastore.V1TestUtil$V1TestWriter flushBatch INFO: Successfully wrote 500 entities Apr 30, 2018 11:38:25 PM org.apache.beam.sdk.io.gcp.datastore.V1TestUtil$V1TestWriter flushBatch INFO: Writing batch of 500 entities Apr 30, 2018 11:38:25 PM org.apache.beam.sdk.io.gcp.datastore.V1TestUtil$V1TestWriter flushBatch INFO: Successfully wrote 500 entities Apr 30, 2018 11:38:25 PM org.apache.beam.sdk.io.gcp.datastore.V1TestUtil deleteAllEntities INFO: Successfully deleted 1000 entities org.apache.beam.sdk.io.gcp.datastore.SplitQueryFnIT > testSplitQueryFnWithLargeDataset STANDARD_ERROR Apr 30, 2018 11:38:25 PM org.apache.beam.sdk.transforms.DoFnTester of WARNING: Your tests use DoFnTester, which may not exercise DoFns correctly. Please use TestPipeline instead. Apr 30, 2018 11:38:25 PM org.apache.beam.sdk.io.gcp.datastore.DatastoreV1$Read getEstimatedSizeBytes INFO: Latest stats timestamp for kind sort_1G is 1524901390000000 Apr 30, 2018 11:38:25 PM org.apache.beam.sdk.io.gcp.datastore.DatastoreV1$Read getEstimatedNumSplits INFO: Estimated size bytes for the query is: 2130000000 Apr 30, 2018 11:38:25 PM org.apache.beam.sdk.io.gcp.datastore.DatastoreV1$Read$SplitQueryFn processElement INFO: Splitting the query into 32 splits org.apache.beam.sdk.io.gcp.datastore.SplitQueryFnIT > testSplitQueryFnWithSmallDataset STANDARD_ERROR Apr 30, 2018 11:38:26 PM org.apache.beam.sdk.transforms.DoFnTester of WARNING: Your tests use DoFnTester, which may not exercise DoFns correctly. Please use TestPipeline instead. Apr 30, 2018 11:38:26 PM org.apache.beam.sdk.io.gcp.datastore.DatastoreV1$Read getEstimatedSizeBytes INFO: Latest stats timestamp for kind shakespeare is 1524901390000000 Apr 30, 2018 11:38:26 PM org.apache.beam.sdk.io.gcp.datastore.DatastoreV1$Read getEstimatedNumSplits INFO: Estimated size bytes for the query is: 26383451 Apr 30, 2018 11:38:26 PM org.apache.beam.sdk.io.gcp.datastore.DatastoreV1$Read$SplitQueryFn processElement INFO: Splitting the query into 12 splits Gradle Test Executor 122 finished executing tests. > Task > :beam-runners-google-cloud-dataflow-java:googleCloudPlatformIntegrationTest > FAILED org.apache.beam.sdk.io.gcp.spanner.SpannerWriteIT > testWrite FAILED java.lang.NoClassDefFoundError: com/google/api/gax/retrying/ExceptionRetryAlgorithm at java.lang.ClassLoader.defineClass1(Native Method) at java.lang.ClassLoader.defineClass(ClassLoader.java:763) at java.security.SecureClassLoader.defineClass(SecureClassLoader.java:142) at java.net.URLClassLoader.defineClass(URLClassLoader.java:467) at java.net.URLClassLoader.access$100(URLClassLoader.java:73) at java.net.URLClassLoader$1.run(URLClassLoader.java:368) at java.net.URLClassLoader$1.run(URLClassLoader.java:362) at java.security.AccessController.doPrivileged(Native Method) at java.net.URLClassLoader.findClass(URLClassLoader.java:361) at java.lang.ClassLoader.loadClass(ClassLoader.java:424) at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:338) at java.lang.ClassLoader.loadClass(ClassLoader.java:357) at com.google.cloud.BaseService.<clinit>(BaseService.java:48) at com.google.cloud.spanner.SpannerOptions$DefaultSpannerFactory.create(SpannerOptions.java:60) at com.google.cloud.spanner.SpannerOptions$DefaultSpannerFactory.create(SpannerOptions.java:55) at com.google.cloud.ServiceOptions.getService(ServiceOptions.java:426) at org.apache.beam.sdk.io.gcp.spanner.SpannerWriteIT.setUp(SpannerWriteIT.java:91) Caused by: java.lang.ClassNotFoundException: com.google.api.gax.retrying.ExceptionRetryAlgorithm at java.net.URLClassLoader.findClass(URLClassLoader.java:381) at java.lang.ClassLoader.loadClass(ClassLoader.java:424) at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:338) at java.lang.ClassLoader.loadClass(ClassLoader.java:357) ... 17 more java.lang.NullPointerException at org.apache.beam.sdk.io.gcp.spanner.SpannerWriteIT.tearDown(SpannerWriteIT.java:148) 13 tests completed, 4 failed Finished generating test XML results (0.007 secs) into: <https://builds.apache.org/job/beam_PostCommit_Java_GradleBuild/ws/src/runners/google-cloud-dataflow-java/build/test-results/googleCloudPlatformIntegrationTest> Generating HTML test report... Finished generating test html results (0.011 secs) into: <https://builds.apache.org/job/beam_PostCommit_Java_GradleBuild/ws/src/runners/google-cloud-dataflow-java/build/reports/tests/googleCloudPlatformIntegrationTest> :beam-runners-google-cloud-dataflow-java:googleCloudPlatformIntegrationTest (Thread[Task worker for ':',5,main]) completed. Took 17 mins 23.27 secs. FAILURE: Build failed with an exception. * What went wrong: Execution failed for task ':beam-runners-google-cloud-dataflow-java:googleCloudPlatformIntegrationTest'. > There were failing tests. See the report at: > file://<https://builds.apache.org/job/beam_PostCommit_Java_GradleBuild/ws/src/runners/google-cloud-dataflow-java/build/reports/tests/googleCloudPlatformIntegrationTest/index.html> * Try: Run with --stacktrace option to get the stack trace. Run with --debug option to get more log output. Run with --scan to get full insights. * Get more help at https://help.gradle.org Deprecated Gradle features were used in this build, making it incompatible with Gradle 5.0. See https://docs.gradle.org/4.7/userguide/command_line_interface.html#sec:command_line_warnings BUILD FAILED in 55m 5s 545 actionable tasks: 545 executed Publishing build scan... https://gradle.com/s/w6vm2vnvzw6fo Build step 'Invoke Gradle script' changed build result to FAILURE Build step 'Invoke Gradle script' marked build as failure Recording test results Not sending mail to unregistered user sweg...@google.com Not sending mail to unregistered user pger...@us.ibm.com Not sending mail to unregistered user aal...@gmail.com Not sending mail to unregistered user sid...@google.com Not sending mail to unregistered user katarzyna.kucharc...@polidea.com Not sending mail to unregistered user ankurgoe...@gmail.com Not sending mail to unregistered user hero...@google.com Not sending mail to unregistered user ro...@frantil.com Not sending mail to unregistered user w...@google.com Not sending mail to unregistered user szewi...@gmail.com Not sending mail to unregistered user git...@alasdairhodge.co.uk Not sending mail to unregistered user ke...@google.com Not sending mail to unregistered user ekirpic...@gmail.com Not sending mail to unregistered user aljoscha.kret...@gmail.com Not sending mail to unregistered user apill...@google.com Not sending mail to unregistered user re...@relax-macbookpro2.roam.corp.google.com Not sending mail to unregistered user kirpic...@google.com