See
<https://builds.apache.org/job/beam_PostCommit_Java_GradleBuild/610/display/redirect>
------------------------------------------
[...truncated 17.88 MB...]
}
at
com.google.api.client.googleapis.json.GoogleJsonResponseException.from(GoogleJsonResponseException.java:146)
at
com.google.api.client.googleapis.services.json.AbstractGoogleJsonClientRequest.newExceptionOnError(AbstractGoogleJsonClientRequest.java:113)
at
com.google.api.client.googleapis.services.json.AbstractGoogleJsonClientRequest.newExceptionOnError(AbstractGoogleJsonClientRequest.java:40)
at
com.google.api.client.googleapis.services.AbstractGoogleClientRequest.executeUnparsed(AbstractGoogleClientRequest.java:432)
at
com.google.api.client.googleapis.services.AbstractGoogleClientRequest.executeUnparsed(AbstractGoogleClientRequest.java:352)
at
com.google.api.client.googleapis.services.AbstractGoogleClientRequest.execute(AbstractGoogleClientRequest.java:469)
at
com.google.cloud.hadoop.util.AbstractGoogleAsyncWriteChannel$UploadOperation.call(AbstractGoogleAsyncWriteChannel.java:357)
... 4 more
May 31, 2018 12:45:49 PM
org.apache.beam.runners.dataflow.TestDataflowRunner$ErrorMonitorMessagesHandler
process
INFO: Dataflow job 2018-05-31_05_44_02-17161200251859747600 threw
exception. Failure message was: java.io.IOException:
com.google.api.client.googleapis.json.GoogleJsonResponseException: 410 Gone
{
"code" : 500,
"errors" : [ {
"domain" : "global",
"message" : "Backend Error",
"reason" : "backendError"
} ],
"message" : "Backend Error"
}
at
com.google.cloud.hadoop.util.AbstractGoogleAsyncWriteChannel.waitForCompletionAndThrowIfUploadFailed(AbstractGoogleAsyncWriteChannel.java:431)
at
com.google.cloud.hadoop.util.AbstractGoogleAsyncWriteChannel.close(AbstractGoogleAsyncWriteChannel.java:289)
at java.nio.channels.Channels$1.close(Channels.java:178)
at
com.google.cloud.dataflow.worker.repackaged.com.google.common.io.CountingOutputStream.close(CountingOutputStream.java:68)
at
com.google.cloud.dataflow.worker.IsmSink$IsmSinkWriter.close(IsmSink.java:296)
at
com.google.cloud.dataflow.worker.SizeReportingSinkWrapper$SizeLimitingWriterWrapper.close(SizeReportingSinkWrapper.java:53)
at
com.google.cloud.dataflow.worker.util.common.worker.WriteOperation.finish(WriteOperation.java:95)
at
com.google.cloud.dataflow.worker.util.common.worker.MapTaskExecutor.execute(MapTaskExecutor.java:83)
at
com.google.cloud.dataflow.worker.BatchDataflowWorker.executeWork(BatchDataflowWorker.java:383)
at
com.google.cloud.dataflow.worker.BatchDataflowWorker.doWork(BatchDataflowWorker.java:355)
at
com.google.cloud.dataflow.worker.BatchDataflowWorker.getAndPerformWork(BatchDataflowWorker.java:286)
at
com.google.cloud.dataflow.worker.DataflowBatchWorkerHarness$WorkerThread.doWork(DataflowBatchWorkerHarness.java:134)
at
com.google.cloud.dataflow.worker.DataflowBatchWorkerHarness$WorkerThread.call(DataflowBatchWorkerHarness.java:114)
at
com.google.cloud.dataflow.worker.DataflowBatchWorkerHarness$WorkerThread.call(DataflowBatchWorkerHarness.java:101)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
at java.lang.Thread.run(Thread.java:745)
Caused by:
com.google.api.client.googleapis.json.GoogleJsonResponseException: 410 Gone
{
"code" : 500,
"errors" : [ {
"domain" : "global",
"message" : "Backend Error",
"reason" : "backendError"
} ],
"message" : "Backend Error"
}
at
com.google.api.client.googleapis.json.GoogleJsonResponseException.from(GoogleJsonResponseException.java:146)
at
com.google.api.client.googleapis.services.json.AbstractGoogleJsonClientRequest.newExceptionOnError(AbstractGoogleJsonClientRequest.java:113)
at
com.google.api.client.googleapis.services.json.AbstractGoogleJsonClientRequest.newExceptionOnError(AbstractGoogleJsonClientRequest.java:40)
at
com.google.api.client.googleapis.services.AbstractGoogleClientRequest.executeUnparsed(AbstractGoogleClientRequest.java:432)
at
com.google.api.client.googleapis.services.AbstractGoogleClientRequest.executeUnparsed(AbstractGoogleClientRequest.java:352)
at
com.google.api.client.googleapis.services.AbstractGoogleClientRequest.execute(AbstractGoogleClientRequest.java:469)
at
com.google.cloud.hadoop.util.AbstractGoogleAsyncWriteChannel$UploadOperation.call(AbstractGoogleAsyncWriteChannel.java:357)
... 4 more
May 31, 2018 12:46:16 PM
org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
SEVERE: 2018-05-31T12:46:15.199Z: java.io.IOException:
com.google.api.client.googleapis.json.GoogleJsonResponseException: 410 Gone
{
"code" : 500,
"errors" : [ {
"domain" : "global",
"message" : "Backend Error",
"reason" : "backendError"
} ],
"message" : "Backend Error"
}
at
com.google.cloud.hadoop.util.AbstractGoogleAsyncWriteChannel.waitForCompletionAndThrowIfUploadFailed(AbstractGoogleAsyncWriteChannel.java:431)
at
com.google.cloud.hadoop.util.AbstractGoogleAsyncWriteChannel.close(AbstractGoogleAsyncWriteChannel.java:289)
at java.nio.channels.Channels$1.close(Channels.java:178)
at
com.google.cloud.dataflow.worker.repackaged.com.google.common.io.CountingOutputStream.close(CountingOutputStream.java:68)
at
com.google.cloud.dataflow.worker.IsmSink$IsmSinkWriter.close(IsmSink.java:296)
at
com.google.cloud.dataflow.worker.SizeReportingSinkWrapper$SizeLimitingWriterWrapper.close(SizeReportingSinkWrapper.java:53)
at
com.google.cloud.dataflow.worker.util.common.worker.WriteOperation.finish(WriteOperation.java:95)
at
com.google.cloud.dataflow.worker.util.common.worker.MapTaskExecutor.execute(MapTaskExecutor.java:83)
at
com.google.cloud.dataflow.worker.BatchDataflowWorker.executeWork(BatchDataflowWorker.java:383)
at
com.google.cloud.dataflow.worker.BatchDataflowWorker.doWork(BatchDataflowWorker.java:355)
at
com.google.cloud.dataflow.worker.BatchDataflowWorker.getAndPerformWork(BatchDataflowWorker.java:286)
at
com.google.cloud.dataflow.worker.DataflowBatchWorkerHarness$WorkerThread.doWork(DataflowBatchWorkerHarness.java:134)
at
com.google.cloud.dataflow.worker.DataflowBatchWorkerHarness$WorkerThread.call(DataflowBatchWorkerHarness.java:114)
at
com.google.cloud.dataflow.worker.DataflowBatchWorkerHarness$WorkerThread.call(DataflowBatchWorkerHarness.java:101)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
at java.lang.Thread.run(Thread.java:745)
Caused by:
com.google.api.client.googleapis.json.GoogleJsonResponseException: 410 Gone
{
"code" : 500,
"errors" : [ {
"domain" : "global",
"message" : "Backend Error",
"reason" : "backendError"
} ],
"message" : "Backend Error"
}
at
com.google.api.client.googleapis.json.GoogleJsonResponseException.from(GoogleJsonResponseException.java:146)
at
com.google.api.client.googleapis.services.json.AbstractGoogleJsonClientRequest.newExceptionOnError(AbstractGoogleJsonClientRequest.java:113)
at
com.google.api.client.googleapis.services.json.AbstractGoogleJsonClientRequest.newExceptionOnError(AbstractGoogleJsonClientRequest.java:40)
at
com.google.api.client.googleapis.services.AbstractGoogleClientRequest.executeUnparsed(AbstractGoogleClientRequest.java:432)
at
com.google.api.client.googleapis.services.AbstractGoogleClientRequest.executeUnparsed(AbstractGoogleClientRequest.java:352)
at
com.google.api.client.googleapis.services.AbstractGoogleClientRequest.execute(AbstractGoogleClientRequest.java:469)
at
com.google.cloud.hadoop.util.AbstractGoogleAsyncWriteChannel$UploadOperation.call(AbstractGoogleAsyncWriteChannel.java:357)
... 4 more
May 31, 2018 12:46:16 PM
org.apache.beam.runners.dataflow.TestDataflowRunner$ErrorMonitorMessagesHandler
process
INFO: Dataflow job 2018-05-31_05_44_02-17161200251859747600 threw
exception. Failure message was: java.io.IOException:
com.google.api.client.googleapis.json.GoogleJsonResponseException: 410 Gone
{
"code" : 500,
"errors" : [ {
"domain" : "global",
"message" : "Backend Error",
"reason" : "backendError"
} ],
"message" : "Backend Error"
}
at
com.google.cloud.hadoop.util.AbstractGoogleAsyncWriteChannel.waitForCompletionAndThrowIfUploadFailed(AbstractGoogleAsyncWriteChannel.java:431)
at
com.google.cloud.hadoop.util.AbstractGoogleAsyncWriteChannel.close(AbstractGoogleAsyncWriteChannel.java:289)
at java.nio.channels.Channels$1.close(Channels.java:178)
at
com.google.cloud.dataflow.worker.repackaged.com.google.common.io.CountingOutputStream.close(CountingOutputStream.java:68)
at
com.google.cloud.dataflow.worker.IsmSink$IsmSinkWriter.close(IsmSink.java:296)
at
com.google.cloud.dataflow.worker.SizeReportingSinkWrapper$SizeLimitingWriterWrapper.close(SizeReportingSinkWrapper.java:53)
at
com.google.cloud.dataflow.worker.util.common.worker.WriteOperation.finish(WriteOperation.java:95)
at
com.google.cloud.dataflow.worker.util.common.worker.MapTaskExecutor.execute(MapTaskExecutor.java:83)
at
com.google.cloud.dataflow.worker.BatchDataflowWorker.executeWork(BatchDataflowWorker.java:383)
at
com.google.cloud.dataflow.worker.BatchDataflowWorker.doWork(BatchDataflowWorker.java:355)
at
com.google.cloud.dataflow.worker.BatchDataflowWorker.getAndPerformWork(BatchDataflowWorker.java:286)
at
com.google.cloud.dataflow.worker.DataflowBatchWorkerHarness$WorkerThread.doWork(DataflowBatchWorkerHarness.java:134)
at
com.google.cloud.dataflow.worker.DataflowBatchWorkerHarness$WorkerThread.call(DataflowBatchWorkerHarness.java:114)
at
com.google.cloud.dataflow.worker.DataflowBatchWorkerHarness$WorkerThread.call(DataflowBatchWorkerHarness.java:101)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
at java.lang.Thread.run(Thread.java:745)
Caused by:
com.google.api.client.googleapis.json.GoogleJsonResponseException: 410 Gone
{
"code" : 500,
"errors" : [ {
"domain" : "global",
"message" : "Backend Error",
"reason" : "backendError"
} ],
"message" : "Backend Error"
}
at
com.google.api.client.googleapis.json.GoogleJsonResponseException.from(GoogleJsonResponseException.java:146)
at
com.google.api.client.googleapis.services.json.AbstractGoogleJsonClientRequest.newExceptionOnError(AbstractGoogleJsonClientRequest.java:113)
at
com.google.api.client.googleapis.services.json.AbstractGoogleJsonClientRequest.newExceptionOnError(AbstractGoogleJsonClientRequest.java:40)
at
com.google.api.client.googleapis.services.AbstractGoogleClientRequest.executeUnparsed(AbstractGoogleClientRequest.java:432)
at
com.google.api.client.googleapis.services.AbstractGoogleClientRequest.executeUnparsed(AbstractGoogleClientRequest.java:352)
at
com.google.api.client.googleapis.services.AbstractGoogleClientRequest.execute(AbstractGoogleClientRequest.java:469)
at
com.google.cloud.hadoop.util.AbstractGoogleAsyncWriteChannel$UploadOperation.call(AbstractGoogleAsyncWriteChannel.java:357)
... 4 more
May 31, 2018 12:46:53 PM
org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2018-05-31T12:46:53.346Z: Executing operation SpannerIO.Write/Write
mutations to Cloud Spanner/Wait.OnSignal/To wait view
0/View.AsList/CreateDataflowView
May 31, 2018 12:46:53 PM
org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2018-05-31T12:46:53.633Z: Executing operation SpannerIO.Write/Write
mutations to Cloud Spanner/Create seed/Read(CreateSource)+SpannerIO.Write/Write
mutations to Cloud Spanner/Wait.OnSignal/Wait/Map+SpannerIO.Write/Write
mutations to Cloud Spanner/Read information schema+SpannerIO.Write/Write
mutations to Cloud Spanner/Schema
View/Combine.GloballyAsSingletonView/Combine.globally(Singleton)/WithKeys/AddKeys/Map+SpannerIO.Write/Write
mutations to Cloud Spanner/Schema
View/Combine.GloballyAsSingletonView/Combine.globally(Singleton)/Combine.perKey(Singleton)/GroupByKey+SpannerIO.Write/Write
mutations to Cloud Spanner/Schema
View/Combine.GloballyAsSingletonView/Combine.globally(Singleton)/Combine.perKey(Singleton)/Combine.GroupedValues/Partial+SpannerIO.Write/Write
mutations to Cloud Spanner/Schema
View/Combine.GloballyAsSingletonView/Combine.globally(Singleton)/Combine.perKey(Singleton)/GroupByKey/Reify+SpannerIO.Write/Write
mutations to Cloud Spanner/Schema
View/Combine.GloballyAsSingletonView/Combine.globally(Singleton)/Combine.perKey(Singleton)/GroupByKey/Write
May 31, 2018 12:47:02 PM
org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2018-05-31T12:47:00.414Z: Executing operation SpannerIO.Write/Write
mutations to Cloud Spanner/Schema
View/Combine.GloballyAsSingletonView/Combine.globally(Singleton)/Combine.perKey(Singleton)/GroupByKey/Close
May 31, 2018 12:47:02 PM
org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2018-05-31T12:47:00.526Z: Executing operation SpannerIO.Write/Write
mutations to Cloud Spanner/Schema
View/Combine.GloballyAsSingletonView/Combine.globally(Singleton)/Combine.perKey(Singleton)/GroupByKey/Read+SpannerIO.Write/Write
mutations to Cloud Spanner/Schema
View/Combine.GloballyAsSingletonView/Combine.globally(Singleton)/Combine.perKey(Singleton)/Combine.GroupedValues+SpannerIO.Write/Write
mutations to Cloud Spanner/Schema
View/Combine.GloballyAsSingletonView/Combine.globally(Singleton)/Combine.perKey(Singleton)/Combine.GroupedValues/Extract+SpannerIO.Write/Write
mutations to Cloud Spanner/Schema
View/Combine.GloballyAsSingletonView/Combine.globally(Singleton)/Values/Values/Map+SpannerIO.Write/Write
mutations to Cloud Spanner/Schema
View/Combine.GloballyAsSingletonView/BatchViewOverrides.GroupByWindowHashAsKeyAndWindowAsSortKey/ParDo(UseWindowHashAsKeyAndWindowAsSortKey)+SpannerIO.Write/Write
mutations to Cloud Spanner/Schema
View/Combine.GloballyAsSingletonView/BatchViewOverrides.GroupByWindowHashAsKeyAndWindowAsSortKey/BatchViewOverrides.GroupByKeyAndSortValuesOnly/Write
May 31, 2018 12:47:05 PM
org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2018-05-31T12:47:05.195Z: Executing operation SpannerIO.Write/Write
mutations to Cloud Spanner/Schema
View/Combine.GloballyAsSingletonView/BatchViewOverrides.GroupByWindowHashAsKeyAndWindowAsSortKey/BatchViewOverrides.GroupByKeyAndSortValuesOnly/Close
May 31, 2018 12:47:05 PM
org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2018-05-31T12:47:05.283Z: Executing operation SpannerIO.Write/Write
mutations to Cloud Spanner/Schema
View/Combine.GloballyAsSingletonView/BatchViewOverrides.GroupByWindowHashAsKeyAndWindowAsSortKey/BatchViewOverrides.GroupByKeyAndSortValuesOnly/Read+SpannerIO.Write/Write
mutations to Cloud Spanner/Schema
View/Combine.GloballyAsSingletonView/ParDo(IsmRecordForSingularValuePerWindow)
May 31, 2018 12:47:13 PM
org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2018-05-31T12:47:12.722Z: Executing operation SpannerIO.Write/Write
mutations to Cloud Spanner/Schema
View/Combine.GloballyAsSingletonView/CreateDataflowView
May 31, 2018 12:47:13 PM
org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2018-05-31T12:47:12.912Z: Executing operation SpannerIO.Write/Write
mutations to Cloud Spanner/Serialize mutations+SpannerIO.Write/Write mutations
to Cloud Spanner/Extract keys+SpannerIO.Write/Write mutations to Cloud
Spanner/Sample keys/GroupByKey+SpannerIO.Write/Write mutations to Cloud
Spanner/Sample keys/Combine.GroupedValues/Partial+SpannerIO.Write/Write
mutations to Cloud Spanner/Sample keys/GroupByKey/Reify+SpannerIO.Write/Write
mutations to Cloud Spanner/Sample keys/GroupByKey/Write
May 31, 2018 12:47:23 PM
org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2018-05-31T12:47:22.355Z: Executing operation SpannerIO.Write/Write
mutations to Cloud Spanner/Sample keys/GroupByKey/Close
May 31, 2018 12:47:23 PM
org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2018-05-31T12:47:22.437Z: Executing operation SpannerIO.Write/Write
mutations to Cloud Spanner/Sample keys/GroupByKey/Read+SpannerIO.Write/Write
mutations to Cloud Spanner/Sample
keys/Combine.GroupedValues+SpannerIO.Write/Write mutations to Cloud
Spanner/Sample keys/Combine.GroupedValues/Extract+SpannerIO.Write/Write
mutations to Cloud Spanner/Keys sample as
view/GBKaSVForData/ParDo(GroupByKeyHashAndSortByKeyAndWindow)+SpannerIO.Write/Write
mutations to Cloud Spanner/Keys sample as
view/GBKaSVForData/BatchViewOverrides.GroupByKeyAndSortValuesOnly/Write
May 31, 2018 12:47:25 PM
org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2018-05-31T12:47:23.935Z: Executing operation SpannerIO.Write/Write
mutations to Cloud Spanner/Keys sample as
view/GBKaSVForData/BatchViewOverrides.GroupByKeyAndSortValuesOnly/Close
May 31, 2018 12:47:25 PM
org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2018-05-31T12:47:24.028Z: Executing operation SpannerIO.Write/Write
mutations to Cloud Spanner/Keys sample as
view/GBKaSVForData/BatchViewOverrides.GroupByKeyAndSortValuesOnly/Read+SpannerIO.Write/Write
mutations to Cloud Spanner/Keys sample as
view/ParMultiDo(ToIsmRecordForMapLike)+SpannerIO.Write/Write mutations to Cloud
Spanner/Keys sample as view/GBKaSVForSize/Write+SpannerIO.Write/Write mutations
to Cloud Spanner/Keys sample as view/GBKaSVForKeys/Write
May 31, 2018 12:47:30 PM
org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2018-05-31T12:47:30.178Z: Executing operation SpannerIO.Write/Write
mutations to Cloud Spanner/Keys sample as view/GBKaSVForSize/Close
May 31, 2018 12:47:30 PM
org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2018-05-31T12:47:30.220Z: Executing operation SpannerIO.Write/Write
mutations to Cloud Spanner/Keys sample as view/GBKaSVForKeys/Close
May 31, 2018 12:47:30 PM
org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2018-05-31T12:47:30.256Z: Executing operation SpannerIO.Write/Write
mutations to Cloud Spanner/Keys sample as
view/GBKaSVForSize/Read+SpannerIO.Write/Write mutations to Cloud Spanner/Keys
sample as view/ParDo(ToIsmMetadataRecordForSize)
May 31, 2018 12:47:30 PM
org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2018-05-31T12:47:30.291Z: Executing operation SpannerIO.Write/Write
mutations to Cloud Spanner/Keys sample as
view/GBKaSVForKeys/Read+SpannerIO.Write/Write mutations to Cloud Spanner/Keys
sample as view/ParDo(ToIsmMetadataRecordForKey)
May 31, 2018 12:47:39 PM
org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2018-05-31T12:47:39.083Z: Executing operation SpannerIO.Write/Write
mutations to Cloud Spanner/Keys sample as view/Flatten.PCollections
May 31, 2018 12:47:39 PM
org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2018-05-31T12:47:39.339Z: Executing operation SpannerIO.Write/Write
mutations to Cloud Spanner/Keys sample as view/CreateDataflowView
May 31, 2018 12:47:42 PM
org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2018-05-31T12:47:39.519Z: Executing operation SpannerIO.Write/Write
mutations to Cloud Spanner/Partition input+SpannerIO.Write/Write mutations to
Cloud Spanner/Group by partition/Reify+SpannerIO.Write/Write mutations to Cloud
Spanner/Group by partition/Write
May 31, 2018 12:47:45 PM
org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2018-05-31T12:47:44.455Z: Executing operation SpannerIO.Write/Write
mutations to Cloud Spanner/Group by partition/Close
May 31, 2018 12:47:45 PM
org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2018-05-31T12:47:44.530Z: Executing operation SpannerIO.Write/Write
mutations to Cloud Spanner/Group by partition/Read+SpannerIO.Write/Write
mutations to Cloud Spanner/Group by
partition/GroupByWindow+SpannerIO.Write/Write mutations to Cloud Spanner/Batch
mutations together+SpannerIO.Write/Write mutations to Cloud Spanner/Write
mutations to Spanner
May 31, 2018 12:47:51 PM
org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2018-05-31T12:47:51.460Z: Cleaning up.
May 31, 2018 12:47:52 PM
org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2018-05-31T12:47:51.830Z: Stopping worker pool...
May 31, 2018 12:49:26 PM
org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2018-05-31T12:49:26.195Z: Autoscaling: Resized worker pool from 1 to
0.
May 31, 2018 12:49:26 PM
org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2018-05-31T12:49:26.240Z: Autoscaling: Would further reduce the
number of workers but reached the minimum number allowed for the job.
May 31, 2018 12:49:35 PM
org.apache.beam.runners.dataflow.DataflowPipelineJob waitUntilFinish
INFO: Job 2018-05-31_05_44_02-17161200251859747600 finished with status
DONE.
May 31, 2018 12:49:35 PM
org.apache.beam.runners.dataflow.TestDataflowRunner checkForPAssertSuccess
INFO: Success result for Dataflow job
2018-05-31_05_44_02-17161200251859747600. Found 0 success, 0 failures out of 0
expected assertions.
May 31, 2018 12:49:35 PM
org.apache.beam.runners.dataflow.DataflowPipelineJob waitUntilFinish
INFO: Job 2018-05-31_05_44_02-17161200251859747600 finished with status
DONE.
Gradle Test Executor 121 finished executing tests.
> Task
> :beam-runners-google-cloud-dataflow-java:googleCloudPlatformIntegrationTest
Finished generating test XML results (0.009 secs) into:
<https://builds.apache.org/job/beam_PostCommit_Java_GradleBuild/ws/src/runners/google-cloud-dataflow-java/build/test-results/googleCloudPlatformIntegrationTest>
Generating HTML test report...
Finished generating test html results (0.012 secs) into:
<https://builds.apache.org/job/beam_PostCommit_Java_GradleBuild/ws/src/runners/google-cloud-dataflow-java/build/reports/tests/googleCloudPlatformIntegrationTest>
Packing task
':beam-runners-google-cloud-dataflow-java:googleCloudPlatformIntegrationTest'
:beam-runners-google-cloud-dataflow-java:googleCloudPlatformIntegrationTest
(Thread[Task worker for ':' Thread 3,5,main]) completed. Took 33 mins 35.287
secs.
:beam-runners-google-cloud-dataflow-java:postCommit (Thread[Task worker for ':'
Thread 3,5,main]) started.
> Task :beam-runners-google-cloud-dataflow-java:postCommit
Skipping task ':beam-runners-google-cloud-dataflow-java:postCommit' as it has
no actions.
:beam-runners-google-cloud-dataflow-java:postCommit (Thread[Task worker for ':'
Thread 3,5,main]) completed. Took 0.0 secs.
FAILURE: Build failed with an exception.
* What went wrong:
Execution failed for task ':beam-sdks-java-extensions-sql:integrationTest'.
> There were failing tests. See the report at:
> file://<https://builds.apache.org/job/beam_PostCommit_Java_GradleBuild/ws/src/sdks/java/extensions/sql/build/reports/tests/integrationTest/index.html>
* Try:
Run with --stacktrace option to get the stack trace. Run with --debug option to
get more log output. Run with --scan to get full insights.
* Get more help at https://help.gradle.org
Deprecated Gradle features were used in this build, making it incompatible with
Gradle 5.0.
See
https://docs.gradle.org/4.7/userguide/command_line_interface.html#sec:command_line_warnings
BUILD FAILED in 48m 56s
535 actionable tasks: 530 executed, 5 from cache
Publishing build scan...
https://gradle.com/s/5eraffa7jdcby
Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure
Recording test results
Not sending mail to unregistered user [email protected]