read failure during flush ------------------------- Key: CASSANDRA-1040 URL: https://issues.apache.org/jira/browse/CASSANDRA-1040 Project: Cassandra Issue Type: Bug Components: Core Reporter: Jonathan Ellis Assignee: Jonathan Ellis Priority: Critical
Joost Ouwerkerk writes: On a single-node cassandra cluster with basic config (-Xmx:1G) loop { * insert 5,000 records in a single columnfamily with UUID keys and random string values (between 1 and 1000 chars) in 5 different columns spanning two different supercolumns * delete all the data by iterating over the rows with get_range_slices(ONE) and calling remove(QUORUM) on each row id returned (path containing only columnfamily) * count number of non-tombstone rows by iterating over the rows with get_range_slices(ONE) and testing data. Break if not zero. } while this is running, call "bin/nodetool -h localhost -p 8081 flush KeySpace" in the background every minute or so. When the data hits some critical size, the loop will break. -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.