[ https://issues.apache.org/jira/browse/CASSANDRA-6487?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13969880#comment-13969880 ]
Benedict commented on CASSANDRA-6487: ------------------------------------- I suggest using the ColumnFamily.dataSize() method as Aleksey suggested: in the BatchStatement.executeWithConditions() and executeWithoutConditions() methods we have access to the fully constructed ColumnFamily objects we will apply. In the former we construct a single CF _updates_, and in the latter we can iterate over each of the IMutations and call _getColumnFamilies()_. Warning on the prepared size is probably not meaningful, because it does not say anything about how big the data we're applying is. > Log WARN on large batch sizes > ----------------------------- > > Key: CASSANDRA-6487 > URL: https://issues.apache.org/jira/browse/CASSANDRA-6487 > Project: Cassandra > Issue Type: Improvement > Reporter: Patrick McFadin > Assignee: Lyuben Todorov > Priority: Minor > Fix For: 2.0.8 > > Attachments: 6487_trunk.patch, 6487_trunk_v2.patch, > cassandra-2.0-6487.diff > > > Large batches on a coordinator can cause a lot of node stress. I propose > adding a WARN log entry if batch sizes go beyond a configurable size. This > will give more visibility to operators on something that can happen on the > developer side. > New yaml setting with 5k default. > {{# Log WARN on any batch size exceeding this value. 5k by default.}} > {{# Caution should be taken on increasing the size of this threshold as it > can lead to node instability.}} > {{batch_size_warn_threshold: 5k}} -- This message was sent by Atlassian JIRA (v6.2#6252)