[ https://issues.apache.org/jira/browse/KAFKA-435?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Manikumar resolved KAFKA-435. ----------------------------- Resolution: Cannot Reproduce Closing inactive issue. > Keep track of the transient test failure for Kafka-343 on Apache Jenkins > ------------------------------------------------------------------------ > > Key: KAFKA-435 > URL: https://issues.apache.org/jira/browse/KAFKA-435 > Project: Kafka > Issue Type: Sub-task > Affects Versions: 0.8.0 > Reporter: Yang Ye > Assignee: Yang Ye > Priority: Minor > Labels: transient-unit-test-failure > > See: > http://mail-archives.apache.org/mod_mbox/incubator-kafka-commits/201208.mbox/browser > Error message: > ------------------------------------------ > [...truncated 3415 lines...] > [2012-08-01 17:27:08,432] ERROR KafkaApi on Broker 0, error when processing > request (test_topic,0,-1,1048576) > (kafka.server.KafkaApis:99) > kafka.common.OffsetOutOfRangeException: offset -1 is out of range > at kafka.log.Log$.findRange(Log.scala:46) > at kafka.log.Log.read(Log.scala:265) > at > kafka.server.KafkaApis.kafka$server$KafkaApis$$readMessageSet(KafkaApis.scala:377) > at > kafka.server.KafkaApis$$anonfun$kafka$server$KafkaApis$$readMessageSets$1$$anonfun$apply$21.apply(KafkaApis.scala:333) > at > kafka.server.KafkaApis$$anonfun$kafka$server$KafkaApis$$readMessageSets$1$$anonfun$apply$21.apply(KafkaApis.scala:332) > at > scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:57) > at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:43) > at > kafka.server.KafkaApis$$anonfun$kafka$server$KafkaApis$$readMessageSets$1.apply(KafkaApis.scala:332) > at > kafka.server.KafkaApis$$anonfun$kafka$server$KafkaApis$$readMessageSets$1.apply(KafkaApis.scala:328) > at > scala.collection.IndexedSeqOptimized$class.foreach(IndexedSeqOptimized.scala:34) > at scala.collection.mutable.WrappedArray.foreach(WrappedArray.scala:32) > at > kafka.server.KafkaApis.kafka$server$KafkaApis$$readMessageSets(KafkaApis.scala:328) > at kafka.server.KafkaApis.handleFetchRequest(KafkaApis.scala:272) > at kafka.server.KafkaApis.handle(KafkaApis.scala:59) > at kafka.server.KafkaRequestHandler.run(KafkaRequestHandler.scala:38) > at java.lang.Thread.run(Thread.java:662) > [2012-08-01 17:27:08,446] ERROR Closing socket for /67.195.138.9 because of > error (kafka.network.Processor:99) > java.io.IOException: Connection reset by peer > at sun.nio.ch.FileDispatcher.read0(Native Method) > at sun.nio.ch.SocketDispatcher.read(SocketDispatcher.java:21) > at sun.nio.ch.IOUtil.readIntoNativeBuffer(IOUtil.java:198) > at sun.nio.ch.IOUtil.read(IOUtil.java:171) > at sun.nio.ch.SocketChannelImpl.read(SocketChannelImpl.java:243) > at kafka.utils.Utils$.read(Utils.scala:630) > at > kafka.network.BoundedByteBufferReceive.readFrom(BoundedByteBufferReceive.scala:54) > at kafka.network.Processor.read(SocketServer.scala:296) > at kafka.network.Processor.run(SocketServer.scala:212) > at java.lang.Thread.run(Thread.java:662) > [0m[[0minfo[0m] [0mTest Passed: > testResetToEarliestWhenOffsetTooLow(kafka.integration.AutoOffsetResetTest)[0m > [0m[[0minfo[0m] [0mTest Starting: > testResetToLatestWhenOffsetTooHigh(kafka.integration.AutoOffsetResetTest)[0m > [2012-08-01 17:27:09,203] ERROR KafkaApi on Broker 0, error when processing > request (test_topic,0,10000,1048576) > (kafka.server.KafkaApis:99) > kafka.common.OffsetOutOfRangeException: offset 10000 is out of range > at kafka.log.Log$.findRange(Log.scala:46) > at kafka.log.Log.read(Log.scala:265) > at > kafka.server.KafkaApis.kafka$server$KafkaApis$$readMessageSet(KafkaApis.scala:377) > at > kafka.server.KafkaApis$$anonfun$kafka$server$KafkaApis$$readMessageSets$1$$anonfun$apply$21.apply(KafkaApis.scala:333) > at > kafka.server.KafkaApis$$anonfun$kafka$server$KafkaApis$$readMessageSets$1$$anonfun$apply$21.apply(KafkaApis.scala:332) > at > scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:57) > at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:43) > at > kafka.server.KafkaApis$$anonfun$kafka$server$KafkaApis$$readMessageSets$1.apply(KafkaApis.scala:332) > at > kafka.server.KafkaApis$$anonfun$kafka$server$KafkaApis$$readMessageSets$1.apply(KafkaApis.scala:328) > at > scala.collection.IndexedSeqOptimized$class.foreach(IndexedSeqOptimized.scala:34) > at scala.collection.mutable.WrappedArray.foreach(WrappedArray.scala:32) > at > kafka.server.KafkaApis.kafka$server$KafkaApis$$readMessageSets(KafkaApis.scala:328) > at kafka.server.KafkaApis.handleFetchRequest(KafkaApis.scala:272) > at kafka.server.KafkaApis.handle(KafkaApis.scala:59) > at kafka.server.KafkaRequestHandler.run(KafkaRequestHandler.scala:38) > at java.lang.Thread.run(Thread.java:662) > [2012-08-01 17:27:11,197] ERROR Closing socket for /67.195.138.9 because of > error (kafka.network.Processor:99) > java.io.IOException: Connection reset by peer > at sun.nio.ch.FileDispatcher.read0(Native Method) > at sun.nio.ch.SocketDispatcher.read(SocketDispatcher.java:21) > at sun.nio.ch.IOUtil.readIntoNativeBuffer(IOUtil.java:198) > at sun.nio.ch.IOUtil.read(IOUtil.java:171) > at sun.nio.ch.SocketChannelImpl.read(SocketChannelImpl.java:243) > at kafka.utils.Utils$.read(Utils.scala:630) > at > kafka.network.BoundedByteBufferReceive.readFrom(BoundedByteBufferReceive.scala:54) > at kafka.network.Processor.read(SocketServer.scala:296) > at kafka.network.Processor.run(SocketServer.scala:212) > at java.lang.Thread.run(Thread.java:662) > [0m[[0minfo[0m] [0mTest Passed: > testResetToLatestWhenOffsetTooHigh(kafka.integration.AutoOffsetResetTest)[0m > [0m[[0minfo[0m] [0mTest Starting: > testResetToLatestWhenOffsetTooLow(kafka.integration.AutoOffsetResetTest)[0m > [2012-08-01 17:27:12,365] ERROR KafkaApi on Broker 0, error when processing > request (test_topic,0,-1,1048576) > (kafka.server.KafkaApis:99) > kafka.common.OffsetOutOfRangeException: offset -1 is out of range > at kafka.log.Log$.findRange(Log.scala:46) > at kafka.log.Log.read(Log.scala:265) > at > kafka.server.KafkaApis.kafka$server$KafkaApis$$readMessageSet(KafkaApis.scala:377) > at > kafka.server.KafkaApis$$anonfun$kafka$server$KafkaApis$$readMessageSets$1$$anonfun$apply$21.apply(KafkaApis.scala:333) > at > kafka.server.KafkaApis$$anonfun$kafka$server$KafkaApis$$readMessageSets$1$$anonfun$apply$21.apply(KafkaApis.scala:332) > at > scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:57) > at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:43) > at > kafka.server.KafkaApis$$anonfun$kafka$server$KafkaApis$$readMessageSets$1.apply(KafkaApis.scala:332) > at > kafka.server.KafkaApis$$anonfun$kafka$server$KafkaApis$$readMessageSets$1.apply(KafkaApis.scala:328) > at > scala.collection.IndexedSeqOptimized$class.foreach(IndexedSeqOptimized.scala:34) > at scala.collection.mutable.WrappedArray.foreach(WrappedArray.scala:32) > at > kafka.server.KafkaApis.kafka$server$KafkaApis$$readMessageSets(KafkaApis.scala:328) > at kafka.server.KafkaApis.handleFetchRequest(KafkaApis.scala:272) > at kafka.server.KafkaApis.handle(KafkaApis.scala:59) > at kafka.server.KafkaRequestHandler.run(KafkaRequestHandler.scala:38) > at java.lang.Thread.run(Thread.java:662) > [2012-08-01 17:27:13,044] WARN EndOfStreamException: Unable to read > additional data from client > sessionid 0x138e33a0beb0012, likely client has closed socket > (org.apache.zookeeper.server.NIOServerCnxn:634) > [2012-08-01 17:27:13,246] WARN EndOfStreamException: Unable to read > additional data from client > sessionid 0x138e33a0beb0016, likely client has closed socket > (org.apache.zookeeper.server.NIOServerCnxn:634) > [2012-08-01 17:27:14,333] ERROR Closing socket for /67.195.138.9 because of > error (kafka.network.Processor:99) > java.io.IOException: Connection reset by peer > at sun.nio.ch.FileDispatcher.read0(Native Method) > at sun.nio.ch.SocketDispatcher.read(SocketDispatcher.java:21) > at sun.nio.ch.IOUtil.readIntoNativeBuffer(IOUtil.java:198) > at sun.nio.ch.IOUtil.read(IOUtil.java:171) > at sun.nio.ch.SocketChannelImpl.read(SocketChannelImpl.java:243) > at kafka.utils.Utils$.read(Utils.scala:630) > at > kafka.network.BoundedByteBufferReceive.readFrom(BoundedByteBufferReceive.scala:54) > at kafka.network.Processor.read(SocketServer.scala:296) > at kafka.network.Processor.run(SocketServer.scala:212) > at java.lang.Thread.run(Thread.java:662) > [2012-08-01 17:27:14,347] WARN EndOfStreamException: Unable to read > additional data from client > sessionid 0x138e33a167e0004, likely client has closed socket > (org.apache.zookeeper.server.NIOServerCnxn:634) > [0m[[0minfo[0m] [0mTest Passed: > testResetToLatestWhenOffsetTooLow(kafka.integration.AutoOffsetResetTest)[0m > [0m[[0minfo[0m] [34m== core-kafka / kafka.integration.AutoOffsetResetTest > ==[0m > [0m[[0minfo[0m] [34m[0m > [0m[[0minfo[0m] [34m== core-kafka / kafka.integration.TopicMetadataTest > ==[0m > [0m[[0minfo[0m] [0mTest Starting: > testTopicMetadataRequest(kafka.integration.TopicMetadataTest)[0m > [0m[[0minfo[0m] [0mTest Passed: > testTopicMetadataRequest(kafka.integration.TopicMetadataTest)[0m > [0m[[0minfo[0m] [0mTest Starting: > testBasicTopicMetadata(kafka.integration.TopicMetadataTest)[0m > [0m[[0minfo[0m] [0mTest Passed: > testBasicTopicMetadata(kafka.integration.TopicMetadataTest)[0m > [0m[[0minfo[0m] [0mTest Starting: > testAutoCreateTopic(kafka.integration.TopicMetadataTest)[0m > [0m[[0minfo[0m] [0mTest Passed: > testAutoCreateTopic(kafka.integration.TopicMetadataTest)[0m > [0m[[0minfo[0m] [34m== core-kafka / kafka.integration.TopicMetadataTest > ==[0m > [0m[[0minfo[0m] [34m[0m > [0m[[0minfo[0m] [34m== core-kafka / kafka.server.LeaderElectionTest ==[0m > [0m[[0minfo[0m] [0mTest Starting: > testLeaderElectionAndEpoch(kafka.server.LeaderElectionTest)[0m > [2012-08-01 17:27:15,189] ERROR Kafka Log on Broker 1, Cannot truncate log to > 0 since the > log start offset is 0 and end offset is 0 (kafka.log.Log:93) > [2012-08-01 17:27:15,694] ERROR Closing socket for /67.195.138.9 because of > error (kafka.network.Processor:99) > java.io.IOException: Connection reset by peer > at sun.nio.ch.FileDispatcher.write0(Native Method) > at sun.nio.ch.SocketDispatcher.write(SocketDispatcher.java:29) > at sun.nio.ch.IOUtil.writeFromNativeBuffer(IOUtil.java:69) > at sun.nio.ch.IOUtil.write(IOUtil.java:40) > at sun.nio.ch.SocketChannelImpl.write(SocketChannelImpl.java:334) > at kafka.api.PartitionDataSend.writeTo(FetchResponse.scala:66) > at kafka.network.MultiSend.writeTo(Transmission.scala:94) > at kafka.network.Send$class.writeCompletely(Transmission.scala:75) > at kafka.network.MultiSend.writeCompletely(Transmission.scala:87) > at kafka.api.TopicDataSend.writeTo(FetchResponse.scala:142) > at kafka.network.MultiSend.writeTo(Transmission.scala:94) > at kafka.network.Send$class.writeCompletely(Transmission.scala:75) > at kafka.network.MultiSend.writeCompletely(Transmission.scala:87) > at kafka.api.FetchResponseSend.writeTo(FetchResponse.scala:219) > at kafka.network.Processor.write(SocketServer.scala:321) > at kafka.network.Processor.run(SocketServer.scala:214) > at java.lang.Thread.run(Thread.java:662) > [2012-08-01 17:27:15,834] WARN EndOfStreamException: Unable to read > additional data from client > sessionid 0x138e33a167e0007, likely client has closed socket > (org.apache.zookeeper.server.NIOServerCnxn:634) > [2012-08-01 17:27:15,835] WARN EndOfStreamException: Unable to read > additional data from client > sessionid 0x138e33a167e0012, likely client has closed socket > (org.apache.zookeeper.server.NIOServerCnxn:634) > [2012-08-01 17:27:17,261] ERROR Kafka Log on Broker 1, Cannot truncate log to > 0 since the > log start offset is 0 and end offset is 0 (kafka.log.Log:93) > [2012-08-01 17:27:19,635] WARN EndOfStreamException: Unable to read > additional data from client > sessionid 0x138e33a252b0014, likely client has closed socket > (org.apache.zookeeper.server.NIOServerCnxn:634) > [2012-08-01 17:27:19,636] WARN EndOfStreamException: Unable to read > additional data from client > sessionid 0x138e33a252b0015, likely client has closed socket > (org.apache.zookeeper.server.NIOServerCnxn:634) > [2012-08-01 17:27:19,645] WARN EndOfStreamException: Unable to read > additional data from client > sessionid 0x138e33a252b0006, likely client has closed socket > (org.apache.zookeeper.server.NIOServerCnxn:634) > [2012-08-01 17:27:19,665] WARN EndOfStreamException: Unable to read > additional data from client > sessionid 0x138e33a252b0018, likely client has closed socket > (org.apache.zookeeper.server.NIOServerCnxn:634) > [2012-08-01 17:27:19,728] ERROR Unexpected Exception: > (org.apache.zookeeper.server.NIOServerCnxn:445) > java.nio.channels.CancelledKeyException > at sun.nio.ch.SelectionKeyImpl.ensureValid(SelectionKeyImpl.java:55) > at sun.nio.ch.SelectionKeyImpl.interestOps(SelectionKeyImpl.java:59) > at > org.apache.zookeeper.server.NIOServerCnxn.sendBuffer(NIOServerCnxn.java:418) > at > org.apache.zookeeper.server.NIOServerCnxn.sendResponse(NIOServerCnxn.java:1509) > at > org.apache.zookeeper.server.FinalRequestProcessor.processRequest(FinalRequestProcessor.java:171) > at > org.apache.zookeeper.server.SyncRequestProcessor.run(SyncRequestProcessor.java:135) > [2012-08-01 17:27:19,729] ERROR Unexpected Exception: > (org.apache.zookeeper.server.NIOServerCnxn:445) > java.nio.channels.CancelledKeyException > at sun.nio.ch.SelectionKeyImpl.ensureValid(SelectionKeyImpl.java:55) > at sun.nio.ch.SelectionKeyImpl.interestOps(SelectionKeyImpl.java:59) > at > org.apache.zookeeper.server.NIOServerCnxn.sendBuffer(NIOServerCnxn.java:418) > at > org.apache.zookeeper.server.NIOServerCnxn.sendResponse(NIOServerCnxn.java:1509) > at > org.apache.zookeeper.server.FinalRequestProcessor.processRequest(FinalRequestProcessor.java:171) > at > org.apache.zookeeper.server.SyncRequestProcessor.run(SyncRequestProcessor.java:135) > [2012-08-01 17:27:19,729] ERROR Unexpected Exception: > (org.apache.zookeeper.server.NIOServerCnxn:445) > java.nio.channels.CancelledKeyException > at sun.nio.ch.SelectionKeyImpl.ensureValid(SelectionKeyImpl.java:55) > at sun.nio.ch.SelectionKeyImpl.interestOps(SelectionKeyImpl.java:59) > at > org.apache.zookeeper.server.NIOServerCnxn.sendBuffer(NIOServerCnxn.java:418) > at > org.apache.zookeeper.server.NIOServerCnxn.sendResponse(NIOServerCnxn.java:1509) > at > org.apache.zookeeper.server.FinalRequestProcessor.processRequest(FinalRequestProcessor.java:171) > at > org.apache.zookeeper.server.SyncRequestProcessor.run(SyncRequestProcessor.java:135) > [2012-08-01 17:27:19,729] ERROR Unexpected Exception: > (org.apache.zookeeper.server.NIOServerCnxn:445) > java.nio.channels.CancelledKeyException > at sun.nio.ch.SelectionKeyImpl.ensureValid(SelectionKeyImpl.java:55) > at sun.nio.ch.SelectionKeyImpl.interestOps(SelectionKeyImpl.java:59) > at > org.apache.zookeeper.server.NIOServerCnxn.sendBuffer(NIOServerCnxn.java:418) > at > org.apache.zookeeper.server.NIOServerCnxn.sendResponse(NIOServerCnxn.java:1509) > at > org.apache.zookeeper.server.FinalRequestProcessor.processRequest(FinalRequestProcessor.java:171) > at > org.apache.zookeeper.server.SyncRequestProcessor.run(SyncRequestProcessor.java:135) > [0m[[0minfo[0m] [0mTest Passed: > testLeaderElectionAndEpoch(kafka.server.LeaderElectionTest)[0m > [0m[[0minfo[0m] [34m== core-kafka / kafka.server.LeaderElectionTest ==[0m > [0m[[0minfo[0m] [34m[0m > [0m[[0minfo[0m] [34m== core-kafka / kafka.log4j.KafkaLog4jAppenderTest > ==[0m > [0m[[0minfo[0m] [0mTest Starting: > testKafkaLog4jConfigs(kafka.log4j.KafkaLog4jAppenderTest)[0m > log4j:WARN No appenders could be found for logger > (org.I0Itec.zkclient.ZkEventThread). > log4j:WARN Please initialize the log4j system properly. > [0m[[0minfo[0m] [0mTest Passed: > testKafkaLog4jConfigs(kafka.log4j.KafkaLog4jAppenderTest)[0m > [0m[[0minfo[0m] [0mTest Starting: > testZkConnectLog4jAppends(kafka.log4j.KafkaLog4jAppenderTest)[0m > [0m[[0minfo[0m] [0mTest Passed: > testZkConnectLog4jAppends(kafka.log4j.KafkaLog4jAppenderTest)[0m > [0m[[0minfo[0m] [34m== core-kafka / kafka.log4j.KafkaLog4jAppenderTest > ==[0m > [0m[[0minfo[0m] [34m[0m > [0m[[0minfo[0m] [34m== core-kafka / > kafka.javaapi.consumer.ZookeeperConsumerConnectorTest > ==[0m > [0m[[0minfo[0m] [0mTest Starting: > testBasic(kafka.javaapi.consumer.ZookeeperConsumerConnectorTest)[0m > [0m[[0minfo[0m] [0mTest Passed: > testBasic(kafka.javaapi.consumer.ZookeeperConsumerConnectorTest)[0m > [0m[[0minfo[0m] [34m== core-kafka / > kafka.javaapi.consumer.ZookeeperConsumerConnectorTest > ==[0m > [0m[[0minfo[0m] [34m[0m > [0m[[0minfo[0m] [34m== core-kafka / Test cleanup 1 ==[0m > [0m[[0minfo[0m] [0mDeleting directory /tmp/sbt_501f0f08[0m > [0m[[0minfo[0m] [34m== core-kafka / Test cleanup 1 ==[0m > [0m[[0minfo[0m] [34m[0m > [0m[[0minfo[0m] [34m== core-kafka / test-finish ==[0m > [0m[[31merror[0m] [0mFailed: : Total 136, Failed 3, Errors 0, Passed 133, > Skipped 0[0m > [0m[[0minfo[0m] [34m== core-kafka / test-finish ==[0m > [0m[[0minfo[0m] [34m[0m > [0m[[0minfo[0m] [34m== core-kafka / test-cleanup ==[0m > [0m[[0minfo[0m] [34m== core-kafka / test-cleanup ==[0m > [0m[[0minfo[0m] [34m[0m > [0m[[0minfo[0m] [34m== java-examples / test-compile ==[0m > [0m[[0minfo[0m] [0m Source analysis: 0 new/modified, 0 indirectly > invalidated, 0 removed.[0m > [0m[[0minfo[0m] [0mCompiling test sources...[0m > [0m[[0minfo[0m] [0mNothing to compile.[0m > [0m[[0minfo[0m] [0m Post-analysis: 0 classes.[0m > [0m[[0minfo[0m] [34m== java-examples / test-compile ==[0m > [0m[[0minfo[0m] [34m[0m > [0m[[0minfo[0m] [34m== hadoop consumer / copy-test-resources ==[0m > [0m[[0minfo[0m] [34m== hadoop consumer / copy-test-resources ==[0m > [0m[[0minfo[0m] [34m[0m > [0m[[0minfo[0m] [34m== hadoop consumer / copy-resources ==[0m > [0m[[0minfo[0m] [34m== hadoop consumer / copy-resources ==[0m > [0m[[0minfo[0m] [34m[0m > [0m[[0minfo[0m] [34m== perf / copy-resources ==[0m > [0m[[0minfo[0m] [34m== perf / copy-resources ==[0m > [0m[[0minfo[0m] [34m[0m > [0m[[0minfo[0m] [34m== java-examples / copy-test-resources ==[0m > [0m[[0minfo[0m] [34m== java-examples / copy-test-resources ==[0m > [0m[[0minfo[0m] [34m[0m > [0m[[0minfo[0m] [34m== perf / test-compile ==[0m > [0m[[0minfo[0m] [0m Source analysis: 0 new/modified, 0 indirectly > invalidated, 0 removed.[0m > [0m[[0minfo[0m] [0mCompiling test sources...[0m > [0m[[0minfo[0m] [0mNothing to compile.[0m > [0m[[0minfo[0m] [0m Post-analysis: 0 classes.[0m > [0m[[0minfo[0m] [34m== perf / test-compile ==[0m > [0m[[0minfo[0m] [34m[0m > [0m[[0minfo[0m] [34m== hadoop consumer / test-compile ==[0m > [0m[[0minfo[0m] [0m Source analysis: 0 new/modified, 0 indirectly > invalidated, 0 removed.[0m > [0m[[0minfo[0m] [0mCompiling test sources...[0m > [0m[[0minfo[0m] [0mNothing to compile.[0m > [0m[[0minfo[0m] [0m Post-analysis: 0 classes.[0m > [0m[[0minfo[0m] [34m== hadoop consumer / test-compile ==[0m > [0m[[0minfo[0m] [34m[0m > [0m[[0minfo[0m] [34m== perf / copy-test-resources ==[0m > [0m[[0minfo[0m] [34m== perf / copy-test-resources ==[0m > [0m[[0minfo[0m] [34m[0m > [0m[[0minfo[0m] [34m== hadoop producer / copy-resources ==[0m > [0m[[0minfo[0m] [34m== hadoop producer / copy-resources ==[0m > [0m[[0minfo[0m] [34m[0m > [0m[[0minfo[0m] [34m== java-examples / copy-resources ==[0m > [0m[[0minfo[0m] [34m== java-examples / copy-resources ==[0m > [0m[[31merror[0m] [0mError running kafka.producer.SyncProducerTest: Test > FAILED[0m > [0m[[31merror[0m] [0mError running kafka.server.LogRecoveryTest: Test > FAILED[0m > [0m[[31merror[0m] [0mError running kafka.server.ServerShutdownTest: Test > FAILED[0m > [0m[[31merror[0m] [0mError running test: One or more subtasks failed[0m > [0m[[0minfo[0m] [0m[0m > [0m[[0minfo[0m] [0mTotal time: 229 s, completed Aug 1, 2012 5:27:24 PM[0m > [0m[[0minfo[0m] [0m[0m > [0m[[0minfo[0m] [0mTotal session time: 229 s, completed Aug 1, 2012 > 5:27:24 PM[0m > [0m[[31merror[0m] [0mError during build.[0m > Build step 'Execute shell' marked build as failure -- This message was sent by Atlassian JIRA (v6.4.14#64029)