[ https://issues.apache.org/jira/browse/CASSANDRA-1717?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13079382#comment-13079382 ]
Pavel Yaskevich commented on CASSANDRA-1717: -------------------------------------------- Making a checksum optional and off by default sounds good to me. bq. Not sure that's bulletproof... That is why I mentioned that if we have checksum per column it will work as a protection from wrong decompression on the block level and spares us additional read and check, isn't it? > Cassandra cannot detect corrupt-but-readable column data > -------------------------------------------------------- > > Key: CASSANDRA-1717 > URL: https://issues.apache.org/jira/browse/CASSANDRA-1717 > Project: Cassandra > Issue Type: New Feature > Components: Core > Reporter: Jonathan Ellis > Assignee: Pavel Yaskevich > Fix For: 1.0 > > Attachments: checksums.txt > > > Most corruptions of on-disk data due to bitrot render the column (or row) > unreadable, so the data can be replaced by read repair or anti-entropy. But > if the corruption keeps column data readable we do not detect it, and if it > corrupts to a higher timestamp value can even resist being overwritten by > newer values. -- This message is automatically generated by JIRA. For more information on JIRA, see: http://www.atlassian.com/software/jira