[ https://issues.apache.org/jira/browse/CASSANDRA-4733?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13922300#comment-13922300 ]
Serj Veras edited comment on CASSANDRA-4733 at 3/6/14 10:57 AM: ---------------------------------------------------------------- I have the same error using Cassandra 2.0.5.22 (DataStax package). I use 3 DC with 3 nodes in each of them. Error is occurred during massive insert workload in one of the DCs. Target CF has replication factor 2 in each of the DCs. {code} ERROR [CompactionExecutor:26] 2014-03-06 10:18:45,760 CassandraDaemon.java (line 196) Exception in thread Thread[CompactionExecutor:26,1,main] java.lang.RuntimeException: Last written key DecoratedKey(-3718191715883699976, 36633732653439302d303730632d343139352d386461342d333736383265393965316335) >= current key DecoratedKey(-7629226534008815744, 62306334323161342d663662362d346364632d383965382d306563343832376639316536) writing into /data/db/cassandra/data/Sync/sy/Sync-sy-tmp-jb-41-Data.db at org.apache.cassandra.io.sstable.SSTableWriter.beforeAppend(SSTableWriter.java:142) at org.apache.cassandra.io.sstable.SSTableWriter.append(SSTableWriter.java:165) at org.apache.cassandra.db.compaction.CompactionTask.runWith(CompactionTask.java:160) at org.apache.cassandra.io.util.DiskAwareRunnable.runMayThrow(DiskAwareRunnable.java:48) at org.apache.cassandra.utils.WrappedRunnable.run(WrappedRunnable.java:28) at org.apache.cassandra.db.compaction.CompactionTask.executeInternal(CompactionTask.java:60) at org.apache.cassandra.db.compaction.AbstractCompactionTask.execute(AbstractCompactionTask.java:59) at org.apache.cassandra.db.compaction.CompactionManager$BackgroundCompactionTask.run(CompactionManager.java:197) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471) at java.util.concurrent.FutureTask.run(FutureTask.java:262) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) at java.lang.Thread.run(Thread.java:744) {code} Here is the state of my cluster after error has occurred. DC3 is the destination of workload writes. {code} Datacenter: DC1 ========== Address Rack Status State Load Owns Token 3074457345618258602 10.0.0.163 RAC1 Up Normal 14.43 GB 33.33% -9223372036854775808 10.0.0.166 RAC0 Up Normal 14.41 GB 33.33% -3074457345618258603 10.0.0.167 RAC2 Up Normal 14.33 GB 33.33% 3074457345618258602 Datacenter: DC2 ========== Address Rack Status State Load Owns Token 3074457345618258603 10.0.1.145 RAC0 Up Normal 14.46 GB 0.00% -9223372036854775807 10.0.1.147 RAC1 Up Normal 14.39 GB 0.00% -3074457345618258602 10.0.1.149 RAC2 Up Normal 14.43 GB 0.00% 3074457345618258603 Datacenter: DC3 ========== Address Rack Status State Load Owns Token 3074457345618258604 10.0.2.47 RAC0 Down Normal 12.84 GB 0.00% -9223372036854775806 10.0.2.49 RAC1 Down Normal 13.69 GB 0.00% -3074457345618258601 10.0.2.51 RAC2 Down Normal 12.34 GB 0.00% 3074457345618258604 {code} was (Author: sivikt): I have the same error using Cassandra 2.0.5.22 (DataStax package). I use 3 DC with 3 nodes in each of them. Error is occurred during massive insert workload in one of the DCs. Target CF has replication factor 2 in each of the DCs. {code} ERROR [CompactionExecutor:26] 2014-03-06 10:18:45,760 CassandraDaemon.java (line 196) Exception in thread Thread[CompactionExecutor:26,1,main] java.lang.RuntimeException: Last written key DecoratedKey(-3718191715883699976, 36633732653439302d303730632d343139352d386461342d333736383265393965316335) >= current key DecoratedKey(-7629226534008815744, 62306334323161342d663662362d346364632d383965382d306563343832376639316536) writing into /data/db/cassandra/data/Sync/sy/Sync-sy-tmp-jb-41-Data.db at org.apache.cassandra.io.sstable.SSTableWriter.beforeAppend(SSTableWriter.java:142) at org.apache.cassandra.io.sstable.SSTableWriter.append(SSTableWriter.java:165) at org.apache.cassandra.db.compaction.CompactionTask.runWith(CompactionTask.java:160) at org.apache.cassandra.io.util.DiskAwareRunnable.runMayThrow(DiskAwareRunnable.java:48) at org.apache.cassandra.utils.WrappedRunnable.run(WrappedRunnable.java:28) at org.apache.cassandra.db.compaction.CompactionTask.executeInternal(CompactionTask.java:60) at org.apache.cassandra.db.compaction.AbstractCompactionTask.execute(AbstractCompactionTask.java:59) at org.apache.cassandra.db.compaction.CompactionManager$BackgroundCompactionTask.run(CompactionManager.java:197) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471) at java.util.concurrent.FutureTask.run(FutureTask.java:262) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) at java.lang.Thread.run(Thread.java:744) {code} Here is the state of my cluster after error has occurred. DC3 is the destination of workload writes. {code} Datacenter: DC1 ========== Address Rack Status State Load Owns Token 3074457345618258602 10.0.0.163 RAC1 Up Normal 14.43 GB 33.33% -9223372036854775808 10.0.0.166 RAC0 Up Normal 14.41 GB 33.33% -3074457345618258603 10.0.0.167 RAC2 Up Normal 14.33 GB 33.33% 3074457345618258602 Datacenter: DC2 ========== Address Rack Status State Load Owns Token 3074457345618258603 10.0.1.145 RAC0 Up Normal 14.46 GB 0.00% -9223372036854775807 10.0.1.147 RAC1 Up Normal 14.39 GB 0.00% -3074457345618258602 10.0.1.149 RAC2 Up Normal 14.43 GB 0.00% 3074457345618258603 Datacenter: DC3 ========== Address Rack Status State Load Owns Token 3074457345618258604 10.0.2.47 RAC0 Down Normal 12.84 GB 0.00% -9223372036854775806 10.0.2.49 RAC1 Down Normal 13.69 GB 0.00% -3074457345618258601 10.0.2.51 RAC2 Down Normal 12.34 GB 0.00% 3074457345618258604 {code} > Last written key >= current key exception when streaming > -------------------------------------------------------- > > Key: CASSANDRA-4733 > URL: https://issues.apache.org/jira/browse/CASSANDRA-4733 > Project: Cassandra > Issue Type: Bug > Components: Core > Affects Versions: 1.2.0 beta 1 > Reporter: Brandon Williams > Assignee: Yuki Morishita > Fix For: 1.2.0 beta 2 > > > {noformat} > ERROR 16:52:56,260 Exception in thread Thread[Streaming to > /10.179.111.137:1,5,main] > java.lang.RuntimeException: java.io.IOException: Connection reset by peer > at com.google.common.base.Throwables.propagate(Throwables.java:160) > at > org.apache.cassandra.utils.WrappedRunnable.run(WrappedRunnable.java:32) > at > java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) > at > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) > at java.lang.Thread.run(Thread.java:662) > Caused by: java.io.IOException: Connection reset by peer > at sun.nio.ch.FileDispatcher.write0(Native Method) > at sun.nio.ch.SocketDispatcher.write(SocketDispatcher.java:29) > at sun.nio.ch.IOUtil.writeFromNativeBuffer(IOUtil.java:72) > at sun.nio.ch.IOUtil.write(IOUtil.java:43) > at sun.nio.ch.SocketChannelImpl.write(SocketChannelImpl.java:334) > at java.nio.channels.Channels.writeFullyImpl(Channels.java:59) > at java.nio.channels.Channels.writeFully(Channels.java:81) > at java.nio.channels.Channels.access$000(Channels.java:47) > at java.nio.channels.Channels$1.write(Channels.java:155) > at > com.ning.compress.lzf.ChunkEncoder.encodeAndWriteChunk(ChunkEncoder.java:133) > at > com.ning.compress.lzf.LZFOutputStream.writeCompressedBlock(LZFOutputStream.java:203) > at > com.ning.compress.lzf.LZFOutputStream.write(LZFOutputStream.java:97) > at > org.apache.cassandra.streaming.FileStreamTask.write(FileStreamTask.java:218) > at > org.apache.cassandra.streaming.FileStreamTask.stream(FileStreamTask.java:164) > at > org.apache.cassandra.streaming.FileStreamTask.runMayThrow(FileStreamTask.java:91) > at > org.apache.cassandra.utils.WrappedRunnable.run(WrappedRunnable.java:28) > ... 3 more > ERROR 16:53:03,951 Exception in thread Thread[Thread-11,5,main] > java.lang.RuntimeException: Last written key > DecoratedKey(113424593524874987650593774422007331058, 3036303936343535) >= > current key DecoratedKey(59229538317742990547810678738983628664, > 3036313133373139) writing into > /var/lib/cassandra/data/Keyspace1-Standard1-tmp-ia-95-Data.db > at > org.apache.cassandra.io.sstable.SSTableWriter.beforeAppend(SSTableWriter.java:132) > at > org.apache.cassandra.io.sstable.SSTableWriter.appendFromStream(SSTableWriter.java:208) > at > org.apache.cassandra.streaming.IncomingStreamReader.streamIn(IncomingStreamReader.java:164) > at > org.apache.cassandra.streaming.IncomingStreamReader.read(IncomingStreamReader.java:107) > at > org.apache.cassandra.net.IncomingTcpConnection.stream(IncomingTcpConnection.java:220) > at > org.apache.cassandra.net.IncomingTcpConnection.handleStream(IncomingTcpConnection.java:165) > at > org.apache.cassandra.net.IncomingTcpConnection.run(IncomingTcpConnection.java:65) > {noformat} > I didn't do anything fancy here, just inserted about 6M keys at rf=2, then > ran repair and got this. -- This message was sent by Atlassian JIRA (v6.2#6252)