See 
<https://builds.apache.org/job/beam_PostCommit_Java_GradleBuild/610/display/redirect>

------------------------------------------
[...truncated 17.88 MB...]
    }
        at 
com.google.api.client.googleapis.json.GoogleJsonResponseException.from(GoogleJsonResponseException.java:146)
        at 
com.google.api.client.googleapis.services.json.AbstractGoogleJsonClientRequest.newExceptionOnError(AbstractGoogleJsonClientRequest.java:113)
        at 
com.google.api.client.googleapis.services.json.AbstractGoogleJsonClientRequest.newExceptionOnError(AbstractGoogleJsonClientRequest.java:40)
        at 
com.google.api.client.googleapis.services.AbstractGoogleClientRequest.executeUnparsed(AbstractGoogleClientRequest.java:432)
        at 
com.google.api.client.googleapis.services.AbstractGoogleClientRequest.executeUnparsed(AbstractGoogleClientRequest.java:352)
        at 
com.google.api.client.googleapis.services.AbstractGoogleClientRequest.execute(AbstractGoogleClientRequest.java:469)
        at 
com.google.cloud.hadoop.util.AbstractGoogleAsyncWriteChannel$UploadOperation.call(AbstractGoogleAsyncWriteChannel.java:357)
        ... 4 more

    May 31, 2018 12:45:49 PM 
org.apache.beam.runners.dataflow.TestDataflowRunner$ErrorMonitorMessagesHandler 
process
    INFO: Dataflow job 2018-05-31_05_44_02-17161200251859747600 threw 
exception. Failure message was: java.io.IOException: 
com.google.api.client.googleapis.json.GoogleJsonResponseException: 410 Gone
    {
      "code" : 500,
      "errors" : [ {
        "domain" : "global",
        "message" : "Backend Error",
        "reason" : "backendError"
      } ],
      "message" : "Backend Error"
    }
        at 
com.google.cloud.hadoop.util.AbstractGoogleAsyncWriteChannel.waitForCompletionAndThrowIfUploadFailed(AbstractGoogleAsyncWriteChannel.java:431)
        at 
com.google.cloud.hadoop.util.AbstractGoogleAsyncWriteChannel.close(AbstractGoogleAsyncWriteChannel.java:289)
        at java.nio.channels.Channels$1.close(Channels.java:178)
        at 
com.google.cloud.dataflow.worker.repackaged.com.google.common.io.CountingOutputStream.close(CountingOutputStream.java:68)
        at 
com.google.cloud.dataflow.worker.IsmSink$IsmSinkWriter.close(IsmSink.java:296)
        at 
com.google.cloud.dataflow.worker.SizeReportingSinkWrapper$SizeLimitingWriterWrapper.close(SizeReportingSinkWrapper.java:53)
        at 
com.google.cloud.dataflow.worker.util.common.worker.WriteOperation.finish(WriteOperation.java:95)
        at 
com.google.cloud.dataflow.worker.util.common.worker.MapTaskExecutor.execute(MapTaskExecutor.java:83)
        at 
com.google.cloud.dataflow.worker.BatchDataflowWorker.executeWork(BatchDataflowWorker.java:383)
        at 
com.google.cloud.dataflow.worker.BatchDataflowWorker.doWork(BatchDataflowWorker.java:355)
        at 
com.google.cloud.dataflow.worker.BatchDataflowWorker.getAndPerformWork(BatchDataflowWorker.java:286)
        at 
com.google.cloud.dataflow.worker.DataflowBatchWorkerHarness$WorkerThread.doWork(DataflowBatchWorkerHarness.java:134)
        at 
com.google.cloud.dataflow.worker.DataflowBatchWorkerHarness$WorkerThread.call(DataflowBatchWorkerHarness.java:114)
        at 
com.google.cloud.dataflow.worker.DataflowBatchWorkerHarness$WorkerThread.call(DataflowBatchWorkerHarness.java:101)
        at java.util.concurrent.FutureTask.run(FutureTask.java:266)
        at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
        at java.lang.Thread.run(Thread.java:745)
    Caused by: 
com.google.api.client.googleapis.json.GoogleJsonResponseException: 410 Gone
    {
      "code" : 500,
      "errors" : [ {
        "domain" : "global",
        "message" : "Backend Error",
        "reason" : "backendError"
      } ],
      "message" : "Backend Error"
    }
        at 
com.google.api.client.googleapis.json.GoogleJsonResponseException.from(GoogleJsonResponseException.java:146)
        at 
com.google.api.client.googleapis.services.json.AbstractGoogleJsonClientRequest.newExceptionOnError(AbstractGoogleJsonClientRequest.java:113)
        at 
com.google.api.client.googleapis.services.json.AbstractGoogleJsonClientRequest.newExceptionOnError(AbstractGoogleJsonClientRequest.java:40)
        at 
com.google.api.client.googleapis.services.AbstractGoogleClientRequest.executeUnparsed(AbstractGoogleClientRequest.java:432)
        at 
com.google.api.client.googleapis.services.AbstractGoogleClientRequest.executeUnparsed(AbstractGoogleClientRequest.java:352)
        at 
com.google.api.client.googleapis.services.AbstractGoogleClientRequest.execute(AbstractGoogleClientRequest.java:469)
        at 
com.google.cloud.hadoop.util.AbstractGoogleAsyncWriteChannel$UploadOperation.call(AbstractGoogleAsyncWriteChannel.java:357)
        ... 4 more

    May 31, 2018 12:46:16 PM 
org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    SEVERE: 2018-05-31T12:46:15.199Z: java.io.IOException: 
com.google.api.client.googleapis.json.GoogleJsonResponseException: 410 Gone
    {
      "code" : 500,
      "errors" : [ {
        "domain" : "global",
        "message" : "Backend Error",
        "reason" : "backendError"
      } ],
      "message" : "Backend Error"
    }
        at 
com.google.cloud.hadoop.util.AbstractGoogleAsyncWriteChannel.waitForCompletionAndThrowIfUploadFailed(AbstractGoogleAsyncWriteChannel.java:431)
        at 
com.google.cloud.hadoop.util.AbstractGoogleAsyncWriteChannel.close(AbstractGoogleAsyncWriteChannel.java:289)
        at java.nio.channels.Channels$1.close(Channels.java:178)
        at 
com.google.cloud.dataflow.worker.repackaged.com.google.common.io.CountingOutputStream.close(CountingOutputStream.java:68)
        at 
com.google.cloud.dataflow.worker.IsmSink$IsmSinkWriter.close(IsmSink.java:296)
        at 
com.google.cloud.dataflow.worker.SizeReportingSinkWrapper$SizeLimitingWriterWrapper.close(SizeReportingSinkWrapper.java:53)
        at 
com.google.cloud.dataflow.worker.util.common.worker.WriteOperation.finish(WriteOperation.java:95)
        at 
com.google.cloud.dataflow.worker.util.common.worker.MapTaskExecutor.execute(MapTaskExecutor.java:83)
        at 
com.google.cloud.dataflow.worker.BatchDataflowWorker.executeWork(BatchDataflowWorker.java:383)
        at 
com.google.cloud.dataflow.worker.BatchDataflowWorker.doWork(BatchDataflowWorker.java:355)
        at 
com.google.cloud.dataflow.worker.BatchDataflowWorker.getAndPerformWork(BatchDataflowWorker.java:286)
        at 
com.google.cloud.dataflow.worker.DataflowBatchWorkerHarness$WorkerThread.doWork(DataflowBatchWorkerHarness.java:134)
        at 
com.google.cloud.dataflow.worker.DataflowBatchWorkerHarness$WorkerThread.call(DataflowBatchWorkerHarness.java:114)
        at 
com.google.cloud.dataflow.worker.DataflowBatchWorkerHarness$WorkerThread.call(DataflowBatchWorkerHarness.java:101)
        at java.util.concurrent.FutureTask.run(FutureTask.java:266)
        at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
        at java.lang.Thread.run(Thread.java:745)
    Caused by: 
com.google.api.client.googleapis.json.GoogleJsonResponseException: 410 Gone
    {
      "code" : 500,
      "errors" : [ {
        "domain" : "global",
        "message" : "Backend Error",
        "reason" : "backendError"
      } ],
      "message" : "Backend Error"
    }
        at 
com.google.api.client.googleapis.json.GoogleJsonResponseException.from(GoogleJsonResponseException.java:146)
        at 
com.google.api.client.googleapis.services.json.AbstractGoogleJsonClientRequest.newExceptionOnError(AbstractGoogleJsonClientRequest.java:113)
        at 
com.google.api.client.googleapis.services.json.AbstractGoogleJsonClientRequest.newExceptionOnError(AbstractGoogleJsonClientRequest.java:40)
        at 
com.google.api.client.googleapis.services.AbstractGoogleClientRequest.executeUnparsed(AbstractGoogleClientRequest.java:432)
        at 
com.google.api.client.googleapis.services.AbstractGoogleClientRequest.executeUnparsed(AbstractGoogleClientRequest.java:352)
        at 
com.google.api.client.googleapis.services.AbstractGoogleClientRequest.execute(AbstractGoogleClientRequest.java:469)
        at 
com.google.cloud.hadoop.util.AbstractGoogleAsyncWriteChannel$UploadOperation.call(AbstractGoogleAsyncWriteChannel.java:357)
        ... 4 more

    May 31, 2018 12:46:16 PM 
org.apache.beam.runners.dataflow.TestDataflowRunner$ErrorMonitorMessagesHandler 
process
    INFO: Dataflow job 2018-05-31_05_44_02-17161200251859747600 threw 
exception. Failure message was: java.io.IOException: 
com.google.api.client.googleapis.json.GoogleJsonResponseException: 410 Gone
    {
      "code" : 500,
      "errors" : [ {
        "domain" : "global",
        "message" : "Backend Error",
        "reason" : "backendError"
      } ],
      "message" : "Backend Error"
    }
        at 
com.google.cloud.hadoop.util.AbstractGoogleAsyncWriteChannel.waitForCompletionAndThrowIfUploadFailed(AbstractGoogleAsyncWriteChannel.java:431)
        at 
com.google.cloud.hadoop.util.AbstractGoogleAsyncWriteChannel.close(AbstractGoogleAsyncWriteChannel.java:289)
        at java.nio.channels.Channels$1.close(Channels.java:178)
        at 
com.google.cloud.dataflow.worker.repackaged.com.google.common.io.CountingOutputStream.close(CountingOutputStream.java:68)
        at 
com.google.cloud.dataflow.worker.IsmSink$IsmSinkWriter.close(IsmSink.java:296)
        at 
com.google.cloud.dataflow.worker.SizeReportingSinkWrapper$SizeLimitingWriterWrapper.close(SizeReportingSinkWrapper.java:53)
        at 
com.google.cloud.dataflow.worker.util.common.worker.WriteOperation.finish(WriteOperation.java:95)
        at 
com.google.cloud.dataflow.worker.util.common.worker.MapTaskExecutor.execute(MapTaskExecutor.java:83)
        at 
com.google.cloud.dataflow.worker.BatchDataflowWorker.executeWork(BatchDataflowWorker.java:383)
        at 
com.google.cloud.dataflow.worker.BatchDataflowWorker.doWork(BatchDataflowWorker.java:355)
        at 
com.google.cloud.dataflow.worker.BatchDataflowWorker.getAndPerformWork(BatchDataflowWorker.java:286)
        at 
com.google.cloud.dataflow.worker.DataflowBatchWorkerHarness$WorkerThread.doWork(DataflowBatchWorkerHarness.java:134)
        at 
com.google.cloud.dataflow.worker.DataflowBatchWorkerHarness$WorkerThread.call(DataflowBatchWorkerHarness.java:114)
        at 
com.google.cloud.dataflow.worker.DataflowBatchWorkerHarness$WorkerThread.call(DataflowBatchWorkerHarness.java:101)
        at java.util.concurrent.FutureTask.run(FutureTask.java:266)
        at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
        at java.lang.Thread.run(Thread.java:745)
    Caused by: 
com.google.api.client.googleapis.json.GoogleJsonResponseException: 410 Gone
    {
      "code" : 500,
      "errors" : [ {
        "domain" : "global",
        "message" : "Backend Error",
        "reason" : "backendError"
      } ],
      "message" : "Backend Error"
    }
        at 
com.google.api.client.googleapis.json.GoogleJsonResponseException.from(GoogleJsonResponseException.java:146)
        at 
com.google.api.client.googleapis.services.json.AbstractGoogleJsonClientRequest.newExceptionOnError(AbstractGoogleJsonClientRequest.java:113)
        at 
com.google.api.client.googleapis.services.json.AbstractGoogleJsonClientRequest.newExceptionOnError(AbstractGoogleJsonClientRequest.java:40)
        at 
com.google.api.client.googleapis.services.AbstractGoogleClientRequest.executeUnparsed(AbstractGoogleClientRequest.java:432)
        at 
com.google.api.client.googleapis.services.AbstractGoogleClientRequest.executeUnparsed(AbstractGoogleClientRequest.java:352)
        at 
com.google.api.client.googleapis.services.AbstractGoogleClientRequest.execute(AbstractGoogleClientRequest.java:469)
        at 
com.google.cloud.hadoop.util.AbstractGoogleAsyncWriteChannel$UploadOperation.call(AbstractGoogleAsyncWriteChannel.java:357)
        ... 4 more

    May 31, 2018 12:46:53 PM 
org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-05-31T12:46:53.346Z: Executing operation SpannerIO.Write/Write 
mutations to Cloud Spanner/Wait.OnSignal/To wait view 
0/View.AsList/CreateDataflowView
    May 31, 2018 12:46:53 PM 
org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-05-31T12:46:53.633Z: Executing operation SpannerIO.Write/Write 
mutations to Cloud Spanner/Create seed/Read(CreateSource)+SpannerIO.Write/Write 
mutations to Cloud Spanner/Wait.OnSignal/Wait/Map+SpannerIO.Write/Write 
mutations to Cloud Spanner/Read information schema+SpannerIO.Write/Write 
mutations to Cloud Spanner/Schema 
View/Combine.GloballyAsSingletonView/Combine.globally(Singleton)/WithKeys/AddKeys/Map+SpannerIO.Write/Write
 mutations to Cloud Spanner/Schema 
View/Combine.GloballyAsSingletonView/Combine.globally(Singleton)/Combine.perKey(Singleton)/GroupByKey+SpannerIO.Write/Write
 mutations to Cloud Spanner/Schema 
View/Combine.GloballyAsSingletonView/Combine.globally(Singleton)/Combine.perKey(Singleton)/Combine.GroupedValues/Partial+SpannerIO.Write/Write
 mutations to Cloud Spanner/Schema 
View/Combine.GloballyAsSingletonView/Combine.globally(Singleton)/Combine.perKey(Singleton)/GroupByKey/Reify+SpannerIO.Write/Write
 mutations to Cloud Spanner/Schema 
View/Combine.GloballyAsSingletonView/Combine.globally(Singleton)/Combine.perKey(Singleton)/GroupByKey/Write
    May 31, 2018 12:47:02 PM 
org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-05-31T12:47:00.414Z: Executing operation SpannerIO.Write/Write 
mutations to Cloud Spanner/Schema 
View/Combine.GloballyAsSingletonView/Combine.globally(Singleton)/Combine.perKey(Singleton)/GroupByKey/Close
    May 31, 2018 12:47:02 PM 
org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-05-31T12:47:00.526Z: Executing operation SpannerIO.Write/Write 
mutations to Cloud Spanner/Schema 
View/Combine.GloballyAsSingletonView/Combine.globally(Singleton)/Combine.perKey(Singleton)/GroupByKey/Read+SpannerIO.Write/Write
 mutations to Cloud Spanner/Schema 
View/Combine.GloballyAsSingletonView/Combine.globally(Singleton)/Combine.perKey(Singleton)/Combine.GroupedValues+SpannerIO.Write/Write
 mutations to Cloud Spanner/Schema 
View/Combine.GloballyAsSingletonView/Combine.globally(Singleton)/Combine.perKey(Singleton)/Combine.GroupedValues/Extract+SpannerIO.Write/Write
 mutations to Cloud Spanner/Schema 
View/Combine.GloballyAsSingletonView/Combine.globally(Singleton)/Values/Values/Map+SpannerIO.Write/Write
 mutations to Cloud Spanner/Schema 
View/Combine.GloballyAsSingletonView/BatchViewOverrides.GroupByWindowHashAsKeyAndWindowAsSortKey/ParDo(UseWindowHashAsKeyAndWindowAsSortKey)+SpannerIO.Write/Write
 mutations to Cloud Spanner/Schema 
View/Combine.GloballyAsSingletonView/BatchViewOverrides.GroupByWindowHashAsKeyAndWindowAsSortKey/BatchViewOverrides.GroupByKeyAndSortValuesOnly/Write
    May 31, 2018 12:47:05 PM 
org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-05-31T12:47:05.195Z: Executing operation SpannerIO.Write/Write 
mutations to Cloud Spanner/Schema 
View/Combine.GloballyAsSingletonView/BatchViewOverrides.GroupByWindowHashAsKeyAndWindowAsSortKey/BatchViewOverrides.GroupByKeyAndSortValuesOnly/Close
    May 31, 2018 12:47:05 PM 
org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-05-31T12:47:05.283Z: Executing operation SpannerIO.Write/Write 
mutations to Cloud Spanner/Schema 
View/Combine.GloballyAsSingletonView/BatchViewOverrides.GroupByWindowHashAsKeyAndWindowAsSortKey/BatchViewOverrides.GroupByKeyAndSortValuesOnly/Read+SpannerIO.Write/Write
 mutations to Cloud Spanner/Schema 
View/Combine.GloballyAsSingletonView/ParDo(IsmRecordForSingularValuePerWindow)
    May 31, 2018 12:47:13 PM 
org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-05-31T12:47:12.722Z: Executing operation SpannerIO.Write/Write 
mutations to Cloud Spanner/Schema 
View/Combine.GloballyAsSingletonView/CreateDataflowView
    May 31, 2018 12:47:13 PM 
org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-05-31T12:47:12.912Z: Executing operation SpannerIO.Write/Write 
mutations to Cloud Spanner/Serialize mutations+SpannerIO.Write/Write mutations 
to Cloud Spanner/Extract keys+SpannerIO.Write/Write mutations to Cloud 
Spanner/Sample keys/GroupByKey+SpannerIO.Write/Write mutations to Cloud 
Spanner/Sample keys/Combine.GroupedValues/Partial+SpannerIO.Write/Write 
mutations to Cloud Spanner/Sample keys/GroupByKey/Reify+SpannerIO.Write/Write 
mutations to Cloud Spanner/Sample keys/GroupByKey/Write
    May 31, 2018 12:47:23 PM 
org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-05-31T12:47:22.355Z: Executing operation SpannerIO.Write/Write 
mutations to Cloud Spanner/Sample keys/GroupByKey/Close
    May 31, 2018 12:47:23 PM 
org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-05-31T12:47:22.437Z: Executing operation SpannerIO.Write/Write 
mutations to Cloud Spanner/Sample keys/GroupByKey/Read+SpannerIO.Write/Write 
mutations to Cloud Spanner/Sample 
keys/Combine.GroupedValues+SpannerIO.Write/Write mutations to Cloud 
Spanner/Sample keys/Combine.GroupedValues/Extract+SpannerIO.Write/Write 
mutations to Cloud Spanner/Keys sample as 
view/GBKaSVForData/ParDo(GroupByKeyHashAndSortByKeyAndWindow)+SpannerIO.Write/Write
 mutations to Cloud Spanner/Keys sample as 
view/GBKaSVForData/BatchViewOverrides.GroupByKeyAndSortValuesOnly/Write
    May 31, 2018 12:47:25 PM 
org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-05-31T12:47:23.935Z: Executing operation SpannerIO.Write/Write 
mutations to Cloud Spanner/Keys sample as 
view/GBKaSVForData/BatchViewOverrides.GroupByKeyAndSortValuesOnly/Close
    May 31, 2018 12:47:25 PM 
org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-05-31T12:47:24.028Z: Executing operation SpannerIO.Write/Write 
mutations to Cloud Spanner/Keys sample as 
view/GBKaSVForData/BatchViewOverrides.GroupByKeyAndSortValuesOnly/Read+SpannerIO.Write/Write
 mutations to Cloud Spanner/Keys sample as 
view/ParMultiDo(ToIsmRecordForMapLike)+SpannerIO.Write/Write mutations to Cloud 
Spanner/Keys sample as view/GBKaSVForSize/Write+SpannerIO.Write/Write mutations 
to Cloud Spanner/Keys sample as view/GBKaSVForKeys/Write
    May 31, 2018 12:47:30 PM 
org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-05-31T12:47:30.178Z: Executing operation SpannerIO.Write/Write 
mutations to Cloud Spanner/Keys sample as view/GBKaSVForSize/Close
    May 31, 2018 12:47:30 PM 
org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-05-31T12:47:30.220Z: Executing operation SpannerIO.Write/Write 
mutations to Cloud Spanner/Keys sample as view/GBKaSVForKeys/Close
    May 31, 2018 12:47:30 PM 
org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-05-31T12:47:30.256Z: Executing operation SpannerIO.Write/Write 
mutations to Cloud Spanner/Keys sample as 
view/GBKaSVForSize/Read+SpannerIO.Write/Write mutations to Cloud Spanner/Keys 
sample as view/ParDo(ToIsmMetadataRecordForSize)
    May 31, 2018 12:47:30 PM 
org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-05-31T12:47:30.291Z: Executing operation SpannerIO.Write/Write 
mutations to Cloud Spanner/Keys sample as 
view/GBKaSVForKeys/Read+SpannerIO.Write/Write mutations to Cloud Spanner/Keys 
sample as view/ParDo(ToIsmMetadataRecordForKey)
    May 31, 2018 12:47:39 PM 
org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-05-31T12:47:39.083Z: Executing operation SpannerIO.Write/Write 
mutations to Cloud Spanner/Keys sample as view/Flatten.PCollections
    May 31, 2018 12:47:39 PM 
org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-05-31T12:47:39.339Z: Executing operation SpannerIO.Write/Write 
mutations to Cloud Spanner/Keys sample as view/CreateDataflowView
    May 31, 2018 12:47:42 PM 
org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-05-31T12:47:39.519Z: Executing operation SpannerIO.Write/Write 
mutations to Cloud Spanner/Partition input+SpannerIO.Write/Write mutations to 
Cloud Spanner/Group by partition/Reify+SpannerIO.Write/Write mutations to Cloud 
Spanner/Group by partition/Write
    May 31, 2018 12:47:45 PM 
org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-05-31T12:47:44.455Z: Executing operation SpannerIO.Write/Write 
mutations to Cloud Spanner/Group by partition/Close
    May 31, 2018 12:47:45 PM 
org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-05-31T12:47:44.530Z: Executing operation SpannerIO.Write/Write 
mutations to Cloud Spanner/Group by partition/Read+SpannerIO.Write/Write 
mutations to Cloud Spanner/Group by 
partition/GroupByWindow+SpannerIO.Write/Write mutations to Cloud Spanner/Batch 
mutations together+SpannerIO.Write/Write mutations to Cloud Spanner/Write 
mutations to Spanner
    May 31, 2018 12:47:51 PM 
org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-05-31T12:47:51.460Z: Cleaning up.
    May 31, 2018 12:47:52 PM 
org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-05-31T12:47:51.830Z: Stopping worker pool...
    May 31, 2018 12:49:26 PM 
org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-05-31T12:49:26.195Z: Autoscaling: Resized worker pool from 1 to 
0.
    May 31, 2018 12:49:26 PM 
org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-05-31T12:49:26.240Z: Autoscaling: Would further reduce the 
number of workers but reached the minimum number allowed for the job.
    May 31, 2018 12:49:35 PM 
org.apache.beam.runners.dataflow.DataflowPipelineJob waitUntilFinish
    INFO: Job 2018-05-31_05_44_02-17161200251859747600 finished with status 
DONE.
    May 31, 2018 12:49:35 PM 
org.apache.beam.runners.dataflow.TestDataflowRunner checkForPAssertSuccess
    INFO: Success result for Dataflow job 
2018-05-31_05_44_02-17161200251859747600. Found 0 success, 0 failures out of 0 
expected assertions.
    May 31, 2018 12:49:35 PM 
org.apache.beam.runners.dataflow.DataflowPipelineJob waitUntilFinish
    INFO: Job 2018-05-31_05_44_02-17161200251859747600 finished with status 
DONE.

Gradle Test Executor 121 finished executing tests.

> Task 
> :beam-runners-google-cloud-dataflow-java:googleCloudPlatformIntegrationTest
Finished generating test XML results (0.009 secs) into: 
<https://builds.apache.org/job/beam_PostCommit_Java_GradleBuild/ws/src/runners/google-cloud-dataflow-java/build/test-results/googleCloudPlatformIntegrationTest>
Generating HTML test report...
Finished generating test html results (0.012 secs) into: 
<https://builds.apache.org/job/beam_PostCommit_Java_GradleBuild/ws/src/runners/google-cloud-dataflow-java/build/reports/tests/googleCloudPlatformIntegrationTest>
Packing task 
':beam-runners-google-cloud-dataflow-java:googleCloudPlatformIntegrationTest'
:beam-runners-google-cloud-dataflow-java:googleCloudPlatformIntegrationTest 
(Thread[Task worker for ':' Thread 3,5,main]) completed. Took 33 mins 35.287 
secs.
:beam-runners-google-cloud-dataflow-java:postCommit (Thread[Task worker for ':' 
Thread 3,5,main]) started.

> Task :beam-runners-google-cloud-dataflow-java:postCommit
Skipping task ':beam-runners-google-cloud-dataflow-java:postCommit' as it has 
no actions.
:beam-runners-google-cloud-dataflow-java:postCommit (Thread[Task worker for ':' 
Thread 3,5,main]) completed. Took 0.0 secs.

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':beam-sdks-java-extensions-sql:integrationTest'.
> There were failing tests. See the report at: 
> file://<https://builds.apache.org/job/beam_PostCommit_Java_GradleBuild/ws/src/sdks/java/extensions/sql/build/reports/tests/integrationTest/index.html>

* Try:
Run with --stacktrace option to get the stack trace. Run with --debug option to 
get more log output. Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with 
Gradle 5.0.
See 
https://docs.gradle.org/4.7/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 48m 56s
535 actionable tasks: 530 executed, 5 from cache

Publishing build scan...
https://gradle.com/s/5eraffa7jdcby

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure
Recording test results
Not sending mail to unregistered user [email protected]

Reply via email to