[ 
https://issues.apache.org/jira/browse/CASSANDRA-6487?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13969880#comment-13969880
 ] 

Benedict commented on CASSANDRA-6487:
-------------------------------------

I suggest using the ColumnFamily.dataSize() method as Aleksey suggested: in the 
BatchStatement.executeWithConditions() and executeWithoutConditions() methods 
we have access to the fully constructed ColumnFamily objects we will apply. In 
the former we construct a single CF _updates_, and in the latter we can iterate 
over each of the IMutations and call _getColumnFamilies()_.

Warning on the prepared size is probably not meaningful, because it does not 
say anything about how big the data we're applying is.

> Log WARN on large batch sizes
> -----------------------------
>
>                 Key: CASSANDRA-6487
>                 URL: https://issues.apache.org/jira/browse/CASSANDRA-6487
>             Project: Cassandra
>          Issue Type: Improvement
>            Reporter: Patrick McFadin
>            Assignee: Lyuben Todorov
>            Priority: Minor
>             Fix For: 2.0.8
>
>         Attachments: 6487_trunk.patch, 6487_trunk_v2.patch, 
> cassandra-2.0-6487.diff
>
>
> Large batches on a coordinator can cause a lot of node stress. I propose 
> adding a WARN log entry if batch sizes go beyond a configurable size. This 
> will give more visibility to operators on something that can happen on the 
> developer side. 
> New yaml setting with 5k default.
> {{# Log WARN on any batch size exceeding this value. 5k by default.}}
> {{# Caution should be taken on increasing the size of this threshold as it 
> can lead to node instability.}}
> {{batch_size_warn_threshold: 5k}}



--
This message was sent by Atlassian JIRA
(v6.2#6252)

Reply via email to