small reminder that unless you have autosnapshot to false in cassandra.yaml, you will need to clear snapshot (nodetool clearsnapshot system_distributed) to actually delete the sstables
On Thu, Oct 6, 2016 at 9:25 AM, Saladi Naidu <naidusp2...@yahoo.com> wrote: > Thanks for the response. It makes sense to periodically truncate as it is > only for debugging purposes > > Naidu Saladi > > > On Wednesday, October 5, 2016 8:03 PM, Chris Lohfink <clohfin...@gmail.com> > wrote: > > > The only current solution is to truncate it periodically. I opened > https://issues.apache.org/jira/browse/CASSANDRA-12701 about it if > interested in following > > On Wed, Oct 5, 2016 at 4:23 PM, Saladi Naidu <naidusp2...@yahoo.com> > wrote: > > We are seeing following warnings in system.log, As *compaction_large_ > partition_warning_threshold_mb* in cassandra.yaml file is as default > value 100, we are seeing these warnings > > 110:WARN [CompactionExecutor:91798] 2016-10-05 00:54:05,554 > BigTableWriter.java:184 - Writing large partition > system_distributed/repair_ history:gccatmer:mer_admin_job (115943239 bytes) > > 111:WARN [CompactionExecutor:91798] 2016-10-05 00:54:13,303 > BigTableWriter.java:184 - Writing large partition system_distributed/repair_ > history:gcconfigsrvcks:user_ activation (163926097 bytes) > > > When I looked at the table definition it is partitioned by keyspace and > cloumnfamily, under this partition, repair history is maintained. When I > looked at the count of rows in this partition, most of the paritions have > >200,000 rows and these will keep growing because of the partition strategy > right. There is no TTL on this so any idea what is the solution for reducing > partition size. > > > I also looked at size_estimates table for this column family and found that > the mean partition size for each range is 50,610,179 which is very large > compared to any other tables. > > > > >