> After upgrading to cass 0.8.4 from cass 0.6.11.  I ran scrub.  That worked
> fine.  Then I ran nodetool repair on one of the nodes.  The disk usage on
> data directory increased from 40GB to 480GB, and it's still growing.

If you check your data directory, does it contain a lot of
"*Compacted" files? It sounds like you're churning sstables from a
combination of compactions/flushes (including triggered by repair) and
the old ones aren't being deleted. I wonder if there is still some
issue causing sstable retention

Since you're on 0.8.4, I'm a bit suspicious. I'd have to re-check each
JIRA but I think the major known repair problems should be fixed
except for CASSANDRA-2280 which is not your problem since you're going
form a total load of 40  gig to hundreds of gigs (so even with all
cf:s streaming, that's unexpected).

Do you have any old left-over streams active on the nodes? "nodetool
netstats". If there are "stuck" streams, they might be causing sstable
retention beyond what you'd expect.

-- 
/ Peter Schuller (@scode on twitter)

Reply via email to