Hmm, perhaps it only happens sometimes. The other half of the time I "Run Java Precommit" on this PR I hit this different failure:
The connection is not obvious to me, if its related to my PR. https://github.com/apache/beam/pull/14804 I only added some Precondition checks. But I don't see those failing anywhere. (Unless something indirect is causing it and stacktrace for that is not printed, i.e. like in a subprocess). Any ideas? Are these tests known to be failing right now? ---- https://ci-beam.apache.org/job/beam_PreCommit_Java_Phrase/3742/#showFailuresLink Test Result (32 failures / +32) org.apache.beam.sdk.io.elasticsearch.ElasticsearchIOTest.testWriteScriptedUpsert org.apache.beam.sdk.io.elasticsearch.ElasticsearchIOTest.testReadWithMetadata org.apache.beam.sdk.io.elasticsearch.ElasticsearchIOTest.testWriteWithIndexFn org.apache.beam.sdk.io.elasticsearch.ElasticsearchIOTest.testMaxParallelRequestsPerWindow org.apache.beam.sdk.io.elasticsearch.ElasticsearchIOTest.testWriteRetryValidRequest org.apache.beam.sdk.io.elasticsearch.ElasticsearchIOTest.testWriteWithMaxBatchSize org.apache.beam.sdk.io.elasticsearch.ElasticsearchIOTest.testWriteRetry org.apache.beam.sdk.io.elasticsearch.ElasticsearchIOTest.testReadWithQueryString org.apache.beam.sdk.io.elasticsearch.ElasticsearchIOTest.testSizes org.apache.beam.sdk.io.elasticsearch.ElasticsearchIOTest.testWriteWithMaxBatchSizeBytes org.apache.beam.sdk.io.elasticsearch.ElasticsearchIOTest.testWriteWithDocVersion org.apache.beam.sdk.io.elasticsearch.ElasticsearchIOTest.testWriteWithAllowableErrors org.apache.beam.sdk.io.elasticsearch.ElasticsearchIOTest.testWriteWithTypeFn org.apache.beam.sdk.io.elasticsearch.ElasticsearchIOTest.testWriteScriptedUpsert org.apache.beam.sdk.io.elasticsearch.ElasticsearchIOTest.testReadWithQueryValueProvider org.apache.beam.sdk.io.elasticsearch.ElasticsearchIOTest.testSplit org.apache.beam.sdk.io.elasticsearch.ElasticsearchIOTest.testWriteRetryValidRequest org.apache.beam.sdk.io.elasticsearch.ElasticsearchIOTest.testWriteWithDocVersion org.apache.beam.sdk.io.elasticsearch.ElasticsearchIOTest.testSizes org.apache.beam.sdk.io.elasticsearch.ElasticsearchIOTest.testMaxParallelRequestsPerWindow org.apache.beam.sdk.io.elasticsearch.ElasticsearchIOTest.testReadWithQueryString org.apache.beam.sdk.io.elasticsearch.ElasticsearchIOTest.testWritePartialUpdate org.apache.beam.sdk.io.elasticsearch.ElasticsearchIOTest.testWriteWithMaxBatchSizeBytes org.apache.beam.sdk.io.elasticsearch.ElasticsearchIOTest.testDefaultRetryPredicate org.apache.beam.sdk.io.elasticsearch.ElasticsearchIOTest.testWriteWithIndexFn org.apache.beam.sdk.io.elasticsearch.ElasticsearchIOTest.testWriteWithRouting org.apache.beam.sdk.io.elasticsearch.ElasticsearchIOTest.testWriteRetry org.apache.beam.sdk.io.elasticsearch.ElasticsearchIOTest.testReadWithMetadata org.apache.beam.sdk.io.elasticsearch.ElasticsearchIOTest.testWriteFullAddressing org.apache.beam.sdk.io.elasticsearch.ElasticsearchIOTest.testWriteWithMaxBatchSize org.apache.beam.sdk.io.elasticsearch.ElasticsearchIOTest.testWriteWithIsDeleteFn org.apache.beam.sdk.io.elasticsearch.ElasticsearchIOTest.testWrite On Wed, Jun 16, 2021 at 5:24 PM Robert Burke <[email protected]> wrote: > Very odd as those paths do resolve now, redirecting to their pkg.go.dev > paths. Very odd. This feels transient, but it's not clear why that would > return a 404 vs some other error. > > On Wed, 16 Jun 2021 at 15:39, Kyle Weaver <[email protected]> wrote: > >> For tasks without structured JUnit output, we have to scroll up / ctrl-f / >> grep for more logs. In this case it looks like it was probably a server-side >> issue. These links work for me, so I'm assuming the problem has been >> resolved. >> >> >> *11:31:04* >* Task :release:go-licenses:java:dockerRun**11:31:04* package >> google.golang.org/protobuf/reflect/protoreflect: unrecognized import path >> "google.golang.org/protobuf/reflect/protoreflect": reading >> https://google.golang.org/protobuf/reflect/protoreflect?go-get=1: 404 Not >> Found*11:31:04* package google.golang.org/protobuf/runtime/protoimpl: >> unrecognized import path "google.golang.org/protobuf/runtime/protoimpl": >> reading https://google.golang.org/protobuf/runtime/protoimpl?go-get=1: 404 >> Not Found*11:31:04* package google.golang.org/protobuf/types/descriptorpb: >> unrecognized import path "google.golang.org/protobuf/types/descriptorpb": >> reading https://google.golang.org/protobuf/types/descriptorpb?go-get=1: 404 >> Not Found*11:31:04* package >> google.golang.org/protobuf/types/known/durationpb: unrecognized import path >> "google.golang.org/protobuf/types/known/durationpb": reading >> https://google.golang.org/protobuf/types/known/durationpb?go-get=1: 404 Not >> Found >> >> >> >> On Wed, Jun 16, 2021 at 2:35 PM Alex Amato <[email protected]> wrote: >> >>> For PR: https://github.com/apache/beam/pull/14804 >>> >>> Is something wrong on this machine? preventing it from running docker? >>> Seems to happen a few times after a run again as well. >>> >>> Anything I can do here to move my PR forward and get it merged? >>> >>> >>> https://ci-beam.apache.org/job/beam_PreCommit_Java_Phrase/3735/consoleFull >>> >>> *11:36:42* >* Task :sdks:java:core:buildDependents**11:36:42* *11:36:42* >>> FAILURE: Build failed with an exception.*11:36:42* *11:36:42* * What went >>> wrong:*11:36:42* Execution failed for task >>> ':release:go-licenses:java:dockerRun'.*11:36:42* > Process 'command >>> 'docker'' finished with non-zero exit value 1*11:36:42* *11:36:42* * >>> Try:*11:36:42* Run with --stacktrace option to get the stack trace. Run >>> with --info or --debug option to get more log output. Run with --scan to >>> get full insights.*11:36:42* *11:36:42* * Get more help at >>> https://help.gradle.org*11:36:42* *11:36:42* Deprecated Gradle features >>> were used in this build, making it incompatible with Gradle 7.0.*11:36:42* >>> Use '--warning-mode all' to show the individual deprecation >>> warnings.*11:36:42* See >>> https://docs.gradle.org/6.8.3/userguide/command_line_interface.html#sec:command_line_warnings*11:36:42* >>> *11:36:42* BUILD FAILED in 7m 20s*11:36:42* 1134 actionable tasks: 420 >>> executed, 712 from cache, 2 up-to-date*11:36:43* *11:36:43* Publishing >>> build scan...*11:36:43* https://gradle.com/s/3yfusnsnfll62 >>> >>>
