> Is there anything to do before that? like drain or flush? For a clean shutdown I do
nodetool -h localhost disablethrift nodetool -h localhost disablegossip && sleep 10 nodetool -h localhost drain then kill > Would you recommend that? If I do it, how often should I do a full snapshot, > and how often should I backup the backup directory? Sounds like you could use Priam and be happier... http://techblog.netflix.com/2012/02/announcing-priam.html > I just saw that there is an option global_snapshot, is it still supported? I cannot find it. Try Piram or the instructions here, which are pretty much what you have described http://www.datastax.com/docs/1.0/operations/backup_restore Cheers ----------------- Aaron Morton Freelance Developer @aaronmorton http://www.thelastpickle.com On 15/08/2012, at 4:57 PM, Tamar Fraenkel <ta...@tok-media.com> wrote: > Aaron, > Thank you very much. I will do as you suggested. > > One last question regarding restart: > I assume, I should do it node by node. > Is there anything to do before that? like drain or flush? > > I am also considering enabling incremental backups on my cluster. Currently I > take a daily full snapshot of the cluster, tar it and load it to S3 (size now > is 3.1GB). Would you recommend that? If I do it, how often should I do a full > snapshot, and how often should I backup the backup directory? > > Another snapshot related question, currently I snapshot on each node and use > parallel-slurp to copy the snapshot to one node where I tar them. I just saw > that there is an option global_snapshot, is it still supported? Does that > mean that if I run it on one node the snapshot will contain data from all > cluster? How does it work in restore? Is it better than my current backup > system? > > Tamar Fraenkel > Senior Software Engineer, TOK Media > > > > ta...@tok-media.com > Tel: +972 2 6409736 > Mob: +972 54 8356490 > Fax: +972 2 5612956 > > > > > > On Tue, Aug 14, 2012 at 11:51 PM, aaron morton <aa...@thelastpickle.com> > wrote: >> According to cfstats there are the some CF with high Comacted row maximum >> sizes (1131752, 4866323 and 25109160). Others max sizes are < 1000000. Are >> these considered to be problematic, what can I do to solve that? > They are only 1, 4 and 25 MB. Not too big. > >> What should be the values of in_memory_compaction_limit_in_mb and >> concurrent_compactors and how do I change them? > Sounds like you dont have very big CF's, so changing the > in_memory_compaction_limit_in_mb may not make too much difference. > > Try changing concurrent_compactors to 2 in the yaml file. This change will > let you know if GC and compaction are related. > >> change yaml file and restart, > yes > >> What do I do about the long rows? What value is considered too big. > They churn more memory during compaction. If you have a lot of rows +32 MB I > would think about it, does not look that way. > > Cheers > > > ----------------- > Aaron Morton > Freelance Developer > @aaronmorton > http://www.thelastpickle.com > > On 15/08/2012, at 3:15 AM, Tamar Fraenkel <ta...@tok-media.com> wrote: > >> Hi! >> It helps, but before I do more actions I want to give you some more info, >> and ask some questions: >> >> Related Info >> According to my yaml file (where do I see these parameters in the jmx? I >> couldn't find them): >> in_memory_compaction_limit_in_mb: 64 >> concurrent_compactors: 1, but it is commented out, so I guess it is the >> default value >> multithreaded_compaction: false >> compaction_throughput_mb_per_sec: 16 >> compaction_preheat_key_cache: true >> According to cfstats there are the some CF with high Comacted row maximum >> sizes (1131752, 4866323 and 25109160). Others max sizes are < 1000000. Are >> these considered to be problematic, what can I do to solve that? >> During compactions Cassandra is slower >> Running Cassandra Version 1.0.8 >> Questions >> What should be the values of in_memory_compaction_limit_in_mb and >> concurrent_compactors and how do I change them? change yaml file and >> restart, or can it be done using jmx without restarting Cassandra? >> What do I do about the long rows? What value is considered too big. >> >> I appreciate your help! >> Thanks, >> >> >> >> Tamar Fraenkel >> Senior Software Engineer, TOK Media >> >> <tokLogo.png> >> >> ta...@tok-media.com >> Tel: +972 2 6409736 >> Mob: +972 54 8356490 >> Fax: +972 2 5612956 >> >> >> >> >> >> On Tue, Aug 14, 2012 at 1:22 PM, aaron morton <aa...@thelastpickle.com> >> wrote: >> There are a couple of steps you can take if compaction is causing GC. >> >> - if you have a lot of wide rows consider reducing the >> in_memory_compaction_limit_in_mb yaml setting. This will slow down >> compaction but will reduce the memory usage. >> >> - reduce concurrent_compactors >> >> Both of these may slow down compaction. Once you have GC under control you >> may want to play with memory settings. >> >> Hope that helps. >> ----------------- >> Aaron Morton >> Freelance Developer >> @aaronmorton >> http://www.thelastpickle.com >> >> On 14/08/2012, at 4:45 PM, Tamar Fraenkel <ta...@tok-media.com> wrote: >> >>> Hi! >>> I have 3 nodes ring running on Amazon EC2. >>> About once a week I see in the logs compaction messages and around the same >>> time info messages about GC (see below) that I think means it is taking too >>> long and happening too often. >>> >>> Does it mean I have to reduce my cache size? >>> Thanks, >>> Tamar >>> >>> INFO [ScheduledTasks:1] 2012-08-13 12:50:57,593 GCInspector.java (line >>> 122) GC for ParNew: 242 ms for 1 collections, 1541590352 used; max is >>> 1937768448 >>> INFO [ScheduledTasks:1] 2012-08-13 12:51:27,740 GCInspector.java (line >>> 122) GC for ParNew: 291 ms for 1 collections, 1458227032 used; max is >>> 1937768448 >>> INFO [ScheduledTasks:1] 2012-08-13 12:51:29,741 GCInspector.java (line >>> 122) GC for ParNew: 261 ms for 1 collections, 1228861368 used; max is >>> 1937768448 >>> INFO [ScheduledTasks:1] 2012-08-13 12:51:30,833 GCInspector.java (line >>> 122) GC for ParNew: 319 ms for 1 collections, 1120131360 used; max is >>> 1937768448 >>> INFO [ScheduledTasks:1] 2012-08-13 12:51:32,863 GCInspector.java (line >>> 122) GC for ParNew: 241 ms for 1 collections, 983144216 used; max is >>> 1937768448 >>> INFO [ScheduledTasks:1] 2012-08-13 12:51:33,864 GCInspector.java (line >>> 122) GC for ParNew: 215 ms for 1 collections, 967702720 used; max is >>> 1937768448 >>> INFO [ScheduledTasks:1] 2012-08-13 12:51:34,964 GCInspector.java (line >>> 122) GC for ParNew: 248 ms for 1 collections, 973803344 used; max is >>> 1937768448 >>> INFO [ScheduledTasks:1] 2012-08-13 12:51:41,211 GCInspector.java (line >>> 122) GC for ParNew: 265 ms for 1 collections, 1071933560 used; max is >>> 1937768448 >>> INFO [ScheduledTasks:1] 2012-08-13 12:51:43,212 GCInspector.java (line >>> 122) GC for ParNew: 326 ms for 1 collections, 1217367792 used; max is >>> 1937768448 >>> INFO [ScheduledTasks:1] 2012-08-13 12:51:44,212 GCInspector.java (line >>> 122) GC for ParNew: 245 ms for 1 collections, 1203481536 used; max is >>> 1937768448 >>> INFO [ScheduledTasks:1] 2012-08-13 12:51:45,213 GCInspector.java (line >>> 122) GC for ParNew: 209 ms for 1 collections, 1208819416 used; max is >>> 1937768448 >>> INFO [ScheduledTasks:1] 2012-08-13 12:51:46,237 GCInspector.java (line >>> 122) GC for ParNew: 248 ms for 1 collections, 1338361648 used; max is >>> 1937768448 >>> >>> >>> Tamar Fraenkel >>> Senior Software Engineer, TOK Media >>> >>> <tokLogo.png> >>> >>> >>> ta...@tok-media.com >>> Tel: +972 2 6409736 >>> Mob: +972 54 8356490 >>> Fax: +972 2 5612956 >>> >>> >>> >> >> > >