See <https://builds.apache.org/job/kafka-2.5-jdk8/115/display/redirect?page=changes>
Changes: [ismael] KAFKA-9996: Upgrade zookeeper to 3.5.8 (#8674) ------------------------------------------ [...truncated 2.05 MB...] org.apache.kafka.streams.integration.GlobalThreadShutDownOrderTest > shouldFinishGlobalStoreOperationOnShutDown PASSED org.apache.kafka.streams.integration.StoreUpgradeIntegrationTest > shouldMigrateInMemoryKeyValueStoreToTimestampedKeyValueStoreUsingPapi STARTED org.apache.kafka.streams.integration.StoreUpgradeIntegrationTest > shouldMigrateInMemoryKeyValueStoreToTimestampedKeyValueStoreUsingPapi PASSED org.apache.kafka.streams.integration.StoreUpgradeIntegrationTest > shouldProxyWindowStoreToTimestampedWindowStoreUsingPapi STARTED org.apache.kafka.streams.integration.StoreUpgradeIntegrationTest > shouldProxyWindowStoreToTimestampedWindowStoreUsingPapi PASSED org.apache.kafka.streams.integration.StoreUpgradeIntegrationTest > shouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi STARTED org.apache.kafka.streams.integration.StoreUpgradeIntegrationTest > shouldMigratePersistentKeyValueStoreToTimestampedKeyValueStoreUsingPapi PASSED org.apache.kafka.streams.integration.StoreUpgradeIntegrationTest > shouldMigratePersistentWindowStoreToTimestampedWindowStoreUsingPapi STARTED org.apache.kafka.streams.integration.StoreUpgradeIntegrationTest > shouldMigratePersistentWindowStoreToTimestampedWindowStoreUsingPapi PASSED org.apache.kafka.streams.integration.StoreUpgradeIntegrationTest > shouldProxyKeyValueStoreToTimestampedKeyValueStoreUsingPapi STARTED org.apache.kafka.streams.integration.StoreUpgradeIntegrationTest > shouldProxyKeyValueStoreToTimestampedKeyValueStoreUsingPapi PASSED org.apache.kafka.streams.integration.StoreUpgradeIntegrationTest > shouldMigrateInMemoryWindowStoreToTimestampedWindowStoreUsingPapi STARTED org.apache.kafka.streams.integration.StoreUpgradeIntegrationTest > shouldMigrateInMemoryWindowStoreToTimestampedWindowStoreUsingPapi PASSED org.apache.kafka.streams.integration.PurgeRepartitionTopicIntegrationTest > shouldRestoreState STARTED org.apache.kafka.streams.integration.PurgeRepartitionTopicIntegrationTest > shouldRestoreState PASSED org.apache.kafka.streams.integration.FineGrainedAutoResetIntegrationTest > shouldOnlyReadRecordsWhereEarliestSpecifiedWithNoCommittedOffsetsWithGlobalAutoOffsetResetLatest STARTED org.apache.kafka.streams.integration.FineGrainedAutoResetIntegrationTest > shouldOnlyReadRecordsWhereEarliestSpecifiedWithNoCommittedOffsetsWithGlobalAutoOffsetResetLatest PASSED org.apache.kafka.streams.integration.FineGrainedAutoResetIntegrationTest > shouldThrowExceptionOverlappingPattern STARTED org.apache.kafka.streams.integration.FineGrainedAutoResetIntegrationTest > shouldThrowExceptionOverlappingPattern PASSED org.apache.kafka.streams.integration.FineGrainedAutoResetIntegrationTest > shouldThrowExceptionOverlappingTopic STARTED org.apache.kafka.streams.integration.FineGrainedAutoResetIntegrationTest > shouldThrowExceptionOverlappingTopic PASSED org.apache.kafka.streams.integration.FineGrainedAutoResetIntegrationTest > shouldOnlyReadRecordsWhereEarliestSpecifiedWithInvalidCommittedOffsets STARTED org.apache.kafka.streams.integration.FineGrainedAutoResetIntegrationTest > shouldOnlyReadRecordsWhereEarliestSpecifiedWithInvalidCommittedOffsets PASSED org.apache.kafka.streams.integration.FineGrainedAutoResetIntegrationTest > shouldOnlyReadRecordsWhereEarliestSpecifiedWithNoCommittedOffsetsWithDefaultGlobalAutoOffsetResetEarliest STARTED org.apache.kafka.streams.integration.FineGrainedAutoResetIntegrationTest > shouldOnlyReadRecordsWhereEarliestSpecifiedWithNoCommittedOffsetsWithDefaultGlobalAutoOffsetResetEarliest PASSED org.apache.kafka.streams.integration.FineGrainedAutoResetIntegrationTest > shouldThrowStreamsExceptionNoResetSpecified STARTED org.apache.kafka.streams.integration.FineGrainedAutoResetIntegrationTest > shouldThrowStreamsExceptionNoResetSpecified PASSED org.apache.kafka.streams.integration.StreamTableJoinIntegrationTest > testInner[caching enabled = true] STARTED org.apache.kafka.streams.integration.StreamTableJoinIntegrationTest > testInner[caching enabled = true] PASSED org.apache.kafka.streams.integration.StreamTableJoinIntegrationTest > testLeft[caching enabled = true] STARTED org.apache.kafka.streams.integration.StreamTableJoinIntegrationTest > testLeft[caching enabled = true] PASSED org.apache.kafka.streams.integration.StreamTableJoinIntegrationTest > testShouldAutoShutdownOnIncompleteMetadata[caching enabled = true] STARTED org.apache.kafka.streams.integration.StreamTableJoinIntegrationTest > testShouldAutoShutdownOnIncompleteMetadata[caching enabled = true] PASSED org.apache.kafka.streams.integration.StreamTableJoinIntegrationTest > testInner[caching enabled = false] STARTED org.apache.kafka.streams.integration.StreamTableJoinIntegrationTest > testInner[caching enabled = false] PASSED org.apache.kafka.streams.integration.StreamTableJoinIntegrationTest > testLeft[caching enabled = false] STARTED org.apache.kafka.streams.integration.StreamTableJoinIntegrationTest > testLeft[caching enabled = false] PASSED org.apache.kafka.streams.integration.StreamTableJoinIntegrationTest > testShouldAutoShutdownOnIncompleteMetadata[caching enabled = false] STARTED org.apache.kafka.streams.integration.StreamTableJoinIntegrationTest > testShouldAutoShutdownOnIncompleteMetadata[caching enabled = false] PASSED org.apache.kafka.streams.integration.KTableKTableForeignKeyJoinDefaultSerdeTest > shouldUseExpectedTopicsWithSerde STARTED org.apache.kafka.streams.integration.KTableKTableForeignKeyJoinDefaultSerdeTest > shouldUseExpectedTopicsWithSerde PASSED org.apache.kafka.streams.integration.KTableKTableForeignKeyJoinDefaultSerdeTest > shouldWorkWithDefaultAndJoinResultSerdes STARTED org.apache.kafka.streams.integration.KTableKTableForeignKeyJoinDefaultSerdeTest > shouldWorkWithDefaultAndJoinResultSerdes PASSED org.apache.kafka.streams.integration.KTableKTableForeignKeyJoinDefaultSerdeTest > shouldWorkWithDefaultSerdes STARTED org.apache.kafka.streams.integration.KTableKTableForeignKeyJoinDefaultSerdeTest > shouldWorkWithDefaultSerdes PASSED org.apache.kafka.streams.integration.KTableKTableForeignKeyJoinDefaultSerdeTest > shouldWorkWithDefaultAndProducedSerdes STARTED org.apache.kafka.streams.integration.KTableKTableForeignKeyJoinDefaultSerdeTest > shouldWorkWithDefaultAndProducedSerdes PASSED org.apache.kafka.streams.integration.KTableKTableForeignKeyJoinDefaultSerdeTest > shouldWorkWithDefaultAndConsumedSerdes STARTED org.apache.kafka.streams.integration.KTableKTableForeignKeyJoinDefaultSerdeTest > shouldWorkWithDefaultAndConsumedSerdes PASSED org.apache.kafka.streams.integration.KTableKTableForeignKeyJoinDefaultSerdeTest > shouldWorkWithDefaultAndEquiJoinResultSerdes STARTED org.apache.kafka.streams.integration.KTableKTableForeignKeyJoinDefaultSerdeTest > shouldWorkWithDefaultAndEquiJoinResultSerdes PASSED org.apache.kafka.streams.integration.QueryableStateIntegrationTest > shouldBeAbleToQueryMapValuesState STARTED org.apache.kafka.streams.integration.QueryableStateIntegrationTest > shouldBeAbleToQueryMapValuesState PASSED org.apache.kafka.streams.integration.QueryableStateIntegrationTest > shouldNotMakeStoreAvailableUntilAllStoresAvailable STARTED org.apache.kafka.streams.integration.QueryableStateIntegrationTest > shouldNotMakeStoreAvailableUntilAllStoresAvailable PASSED org.apache.kafka.streams.integration.QueryableStateIntegrationTest > concurrentAccesses STARTED org.apache.kafka.streams.integration.QueryableStateIntegrationTest > concurrentAccesses PASSED org.apache.kafka.streams.integration.QueryableStateIntegrationTest > shouldAllowToQueryAfterThreadDied STARTED org.apache.kafka.streams.integration.QueryableStateIntegrationTest > shouldAllowToQueryAfterThreadDied PASSED org.apache.kafka.streams.integration.QueryableStateIntegrationTest > shouldBeAbleToQueryStateWithZeroSizedCache STARTED org.apache.kafka.streams.integration.QueryableStateIntegrationTest > shouldBeAbleToQueryStateWithZeroSizedCache PASSED org.apache.kafka.streams.integration.QueryableStateIntegrationTest > shouldBeAbleToQueryDuringRebalance STARTED org.apache.kafka.streams.integration.QueryableStateIntegrationTest > shouldBeAbleToQueryDuringRebalance PASSED org.apache.kafka.streams.integration.QueryableStateIntegrationTest > shouldBeAbleToQueryMapValuesAfterFilterState STARTED org.apache.kafka.streams.integration.QueryableStateIntegrationTest > shouldBeAbleToQueryMapValuesAfterFilterState PASSED org.apache.kafka.streams.integration.QueryableStateIntegrationTest > shouldBeAbleToQueryFilterState STARTED org.apache.kafka.streams.integration.QueryableStateIntegrationTest > shouldBeAbleToQueryFilterState PASSED org.apache.kafka.streams.integration.QueryableStateIntegrationTest > shouldBeAbleQueryStandbyStateDuringRebalance STARTED org.apache.kafka.streams.integration.QueryableStateIntegrationTest > shouldBeAbleQueryStandbyStateDuringRebalance PASSED org.apache.kafka.streams.integration.QueryableStateIntegrationTest > shouldBeAbleToQueryStateWithNonZeroSizedCache STARTED org.apache.kafka.streams.integration.QueryableStateIntegrationTest > shouldBeAbleToQueryStateWithNonZeroSizedCache PASSED org.apache.kafka.streams.integration.StreamsUpgradeTestIntegrationTest > testVersionProbingUpgrade STARTED org.apache.kafka.streams.integration.StreamsUpgradeTestIntegrationTest > testVersionProbingUpgrade PASSED org.apache.kafka.streams.integration.MetricsIntegrationTest > shouldAddMetricsForSessionStoreWithBuiltInMetricsLatestVersion STARTED org.apache.kafka.streams.integration.MetricsIntegrationTest > shouldAddMetricsForSessionStoreWithBuiltInMetricsLatestVersion PASSED org.apache.kafka.streams.integration.MetricsIntegrationTest > shouldAddMetricsOnAllLevelsWithBuiltInMetricsLatestVersion STARTED org.apache.kafka.streams.integration.MetricsIntegrationTest > shouldAddMetricsOnAllLevelsWithBuiltInMetricsLatestVersion PASSED org.apache.kafka.streams.integration.MetricsIntegrationTest > shouldAddMetricsOnAllLevelsWithBuiltInMetricsVersion0100To24 STARTED org.apache.kafka.streams.integration.MetricsIntegrationTest > shouldAddMetricsOnAllLevelsWithBuiltInMetricsVersion0100To24 PASSED org.apache.kafka.streams.integration.MetricsIntegrationTest > shouldAddMetricsForWindowStoreAndSuppressionBufferWithBuiltInMetricsLatestVersion STARTED org.apache.kafka.streams.integration.MetricsIntegrationTest > shouldAddMetricsForWindowStoreAndSuppressionBufferWithBuiltInMetricsLatestVersion PASSED org.apache.kafka.streams.integration.MetricsIntegrationTest > shouldAddMetricsForSessionStoreWithBuiltInMetricsVersion0100To24 STARTED org.apache.kafka.streams.integration.MetricsIntegrationTest > shouldAddMetricsForSessionStoreWithBuiltInMetricsVersion0100To24 PASSED org.apache.kafka.streams.integration.MetricsIntegrationTest > shouldAddMetricsForWindowStoreAndSuppressionBufferWithBuiltInMetricsVersion0100To24 STARTED org.apache.kafka.streams.integration.MetricsIntegrationTest > shouldAddMetricsForWindowStoreAndSuppressionBufferWithBuiltInMetricsVersion0100To24 PASSED org.apache.kafka.streams.integration.ResetPartitionTimeIntegrationTest > shouldPreservePartitionTimeOnKafkaStreamRestart[0: eosEnabled=false] STARTED org.apache.kafka.streams.integration.ResetPartitionTimeIntegrationTest > shouldPreservePartitionTimeOnKafkaStreamRestart[0: eosEnabled=false] PASSED org.apache.kafka.streams.integration.ResetPartitionTimeIntegrationTest > shouldPreservePartitionTimeOnKafkaStreamRestart[1: eosEnabled=true] STARTED org.apache.kafka.streams.integration.ResetPartitionTimeIntegrationTest > shouldPreservePartitionTimeOnKafkaStreamRestart[1: eosEnabled=true] PASSED org.apache.kafka.streams.integration.KTableSourceTopicRestartIntegrationTest > shouldRestoreAndProgressWhenTopicNotWrittenToDuringRestoration STARTED org.apache.kafka.streams.integration.KTableSourceTopicRestartIntegrationTest > shouldRestoreAndProgressWhenTopicNotWrittenToDuringRestoration PASSED org.apache.kafka.streams.integration.KTableSourceTopicRestartIntegrationTest > shouldRestoreAndProgressWhenTopicWrittenToDuringRestorationWithEosEnabled STARTED FATAL: command execution failed java.io.EOFException at java.io.ObjectInputStream$PeekInputStream.readFully(ObjectInputStream.java:2681) at java.io.ObjectInputStream$BlockDataInputStream.readShort(ObjectInputStream.java:3156) at java.io.ObjectInputStream.readStreamHeader(ObjectInputStream.java:862) at java.io.ObjectInputStream.<init>(ObjectInputStream.java:358) at hudson.remoting.ObjectInputStreamEx.<init>(ObjectInputStreamEx.java:49) at hudson.remoting.Command.readFrom(Command.java:140) at hudson.remoting.Command.readFrom(Command.java:126) at hudson.remoting.AbstractSynchronousByteArrayCommandTransport.read(AbstractSynchronousByteArrayCommandTransport.java:36) at hudson.remoting.SynchronousCommandTransport$ReaderThread.run(SynchronousCommandTransport.java:63) Caused: java.io.IOException: Unexpected termination of the channel at hudson.remoting.SynchronousCommandTransport$ReaderThread.run(SynchronousCommandTransport.java:77) Caused: hudson.remoting.ChannelClosedException: Channel "unknown": Remote call on H39 failed. The channel is closing down or has closed down at hudson.remoting.Channel.call(Channel.java:950) at hudson.remoting.RemoteInvocationHandler.invoke(RemoteInvocationHandler.java:283) at com.sun.proxy.$Proxy141.isAlive(Unknown Source) at hudson.Launcher$RemoteLauncher$ProcImpl.isAlive(Launcher.java:1150) at hudson.Launcher$RemoteLauncher$ProcImpl.join(Launcher.java:1142) at hudson.tasks.CommandInterpreter.join(CommandInterpreter.java:155) at hudson.tasks.CommandInterpreter.perform(CommandInterpreter.java:109) at hudson.tasks.CommandInterpreter.perform(CommandInterpreter.java:66) at hudson.tasks.BuildStepMonitor$1.perform(BuildStepMonitor.java:20) at hudson.model.AbstractBuild$AbstractBuildExecution.perform(AbstractBuild.java:741) at hudson.model.Build$BuildExecution.build(Build.java:206) at hudson.model.Build$BuildExecution.doRun(Build.java:163) at hudson.model.AbstractBuild$AbstractBuildExecution.run(AbstractBuild.java:504) at hudson.model.Run.execute(Run.java:1815) at hudson.model.FreeStyleBuild.run(FreeStyleBuild.java:43) at hudson.model.ResourceController.execute(ResourceController.java:97) at hudson.model.Executor.run(Executor.java:429) FATAL: Unable to delete script file /tmp/jenkins4115813822234750880.sh java.io.EOFException at java.io.ObjectInputStream$PeekInputStream.readFully(ObjectInputStream.java:2681) at java.io.ObjectInputStream$BlockDataInputStream.readShort(ObjectInputStream.java:3156) at java.io.ObjectInputStream.readStreamHeader(ObjectInputStream.java:862) at java.io.ObjectInputStream.<init>(ObjectInputStream.java:358) at hudson.remoting.ObjectInputStreamEx.<init>(ObjectInputStreamEx.java:49) at hudson.remoting.Command.readFrom(Command.java:140) at hudson.remoting.Command.readFrom(Command.java:126) at hudson.remoting.AbstractSynchronousByteArrayCommandTransport.read(AbstractSynchronousByteArrayCommandTransport.java:36) at hudson.remoting.SynchronousCommandTransport$ReaderThread.run(SynchronousCommandTransport.java:63) Caused: java.io.IOException: Unexpected termination of the channel at hudson.remoting.SynchronousCommandTransport$ReaderThread.run(SynchronousCommandTransport.java:77) Caused: hudson.remoting.ChannelClosedException: Channel "unknown": Remote call on H39 failed. The channel is closing down or has closed down at hudson.remoting.Channel.call(Channel.java:950) at hudson.FilePath.act(FilePath.java:1072) at hudson.FilePath.act(FilePath.java:1061) at hudson.FilePath.delete(FilePath.java:1542) at hudson.tasks.CommandInterpreter.perform(CommandInterpreter.java:123) at hudson.tasks.CommandInterpreter.perform(CommandInterpreter.java:66) at hudson.tasks.BuildStepMonitor$1.perform(BuildStepMonitor.java:20) at hudson.model.AbstractBuild$AbstractBuildExecution.perform(AbstractBuild.java:741) at hudson.model.Build$BuildExecution.build(Build.java:206) at hudson.model.Build$BuildExecution.doRun(Build.java:163) at hudson.model.AbstractBuild$AbstractBuildExecution.run(AbstractBuild.java:504) at hudson.model.Run.execute(Run.java:1815) at hudson.model.FreeStyleBuild.run(FreeStyleBuild.java:43) at hudson.model.ResourceController.execute(ResourceController.java:97) at hudson.model.Executor.run(Executor.java:429) Build step 'Execute shell' marked build as failure ERROR: Step ?[Deprecated] Publish FindBugs analysis results? failed: no workspace for kafka-2.5-jdk8 #115 ERROR: Step ?Publish JUnit test result report? failed: no workspace for kafka-2.5-jdk8 #115 ERROR: H39 is offline; cannot locate JDK 1.8 (latest) Not sending mail to unregistered user ism...@juma.me.uk