Hello, A couple of items,
1) your reads are going to be very slow with that many SSTables being accessed in the worse case. Personally i have never seen 72 SStables being accessed before, if I'm reading that correctly. 2) 16GB of Heap is very large and may actually increase the duration of GC events, though they GC events should occur less frequently. Looks like you actually have several issues: 1) GC issue - please add GC logging information, using the GC log flags included int he cassandra-env.sh file. Please share the results of the gc log file once collected. 2) Compaction/Tombstone/Data Model issues - this one is tougher to solve via irc as data modeling requires a bit more in-depth review of the problem statement/goal. But, what does your data model look like? Also, how are you doing reads and writes and deletes? - I would expect your GC issues are the result of the Data Model or access patterns based on the results of CFHistograms 3) What does tpstats look like? Jonathan Jonathan Lacefield Solutions Architect, DataStax (404) 822 3487 <http://www.linkedin.com/in/jlacefield> <http://www.datastax.com/what-we-offer/products-services/training/virtual-training> On Mon, Jan 27, 2014 at 7:07 AM, Dimetrio <dimet...@flysoft.ru> wrote: > No one advice did't help to me for reduce GC load > > I tried these: > > MAX_HEAP_SIZE from default(8GB) to 16G with HEAP_NEWSIZE from 400M to 9600M > key cache on/off > compacting memory size and other limits > 15 c3.4xlarge nodes (adding 5 nodes to 10 nodes cluster did't help): > and many other > > Reads ~5000 ops/s > Writes ~ 5000 ops/s > max batch is 50 > heavy reads and heavy writes (and heavy deletes) > sometimes i have message: > Read 1001 live and 2691 > Read 12 live and 2796 > > > sudo jstat -gcutil -h15 `sudo cat /var/run/cassandra/cassandra.pid` 250ms 0 > S0 S1 E O P YGC YGCT FGC FGCT GCT > 18.93 0.00 4.52 75.36 59.77 225 30.119 18 28.361 58.480 > 0.00 13.12 3.78 81.09 59.77 226 30.193 18 28.617 58.810 > 0.00 13.12 39.50 81.09 59.78 226 30.193 18 28.617 58.810 > 0.00 13.12 80.70 81.09 59.78 226 30.193 18 28.617 58.810 > 17.21 9.13 0.66 87.38 59.78 228 30.235 18 28.617 58.852 > 0.00 10.96 29.43 87.89 59.78 228 30.328 18 28.617 58.945 > 0.00 10.96 62.67 87.89 59.78 228 30.328 18 28.617 58.945 > 0.00 10.96 96.62 87.89 59.78 228 30.328 18 28.617 58.945 > 0.00 10.69 10.29 94.56 59.78 230 30.462 18 28.617 59.078 > 0.00 10.69 38.08 94.56 59.78 230 30.462 18 28.617 59.078 > 0.00 10.69 71.70 94.56 59.78 230 30.462 18 28.617 59.078 > 15.91 6.24 0.03 99.96 59.78 232 30.506 18 28.617 59.123 > 15.91 8.02 0.03 99.96 59.78 232 30.506 18 28.617 59.123 > 15.91 8.02 0.03 99.96 59.78 232 30.506 18 28.617 59.123 > 15.91 8.02 0.03 99.96 59.78 232 30.506 18 28.617 59.123 > S0 S1 E O P YGC YGCT FGC FGCT GCT > 15.91 8.02 0.03 99.96 59.78 232 30.506 18 28.617 59.123 > 15.91 8.02 0.03 99.96 59.78 232 30.506 18 28.617 59.123 > 15.91 8.02 0.03 99.96 59.78 232 30.506 18 28.617 59.123 > 15.91 8.02 0.03 99.96 59.78 232 30.506 18 28.617 59.123 > 15.91 8.02 0.03 99.96 59.78 232 30.506 18 28.617 59.123 > 15.91 8.02 0.03 99.96 59.78 232 30.506 18 28.617 59.123 > 15.91 8.02 0.03 99.96 59.78 232 30.506 18 28.617 59.123 > 15.91 8.02 0.03 99.96 59.78 232 30.506 18 28.617 59.123 > 15.91 8.02 0.03 99.96 59.78 232 30.506 18 28.617 59.123 > 15.91 8.02 0.03 99.96 59.78 232 30.506 18 28.617 59.123 > 15.91 8.02 0.03 99.96 59.78 232 30.506 18 28.617 59.123 > 15.91 8.02 0.03 99.96 59.78 232 30.506 18 28.617 59.123 > 15.91 8.02 0.03 99.96 59.78 232 30.506 18 28.617 59.123 > 15.91 8.02 0.03 99.96 59.78 232 30.506 18 28.617 59.123 > > > $ nodetool cfhistograms Social home_timeline > Social/home_timeline histograms > Offset SSTables Write Latency Read Latency Partition Size > Cell Count > (micros) (micros) (bytes) > 1 10458 0 0 0 > 26330 > 2 72428 0 0 0 > 0 > 3 339490 11 0 0 > 42398 > 4 661819 156 0 0 > 0 > 5 67186 893 0 0 > 0 > 6 33284 3064 0 0 > 15907 > 7 41287 10542 0 0 > 0 > 8 49085 34689 0 0 > 0 > 10 82941 244013 0 0 > 19068 > 12 49438 523284 0 0 > 17726 > 14 45109 724026 0 0 > 0 > 17 69144 1181873 0 0 > 18888 > 20 52563 1039934 0 0 > 9859 > 24 63041 1007711 0 0 > 16223 > 29 53042 702118 34 0 > 6289 > 35 31725 395098 78 0 > 10434 > 42 25571 212531 81 0 > 13735 > 50 11984 120692 69 0 > 8020 > 60 3985 71142 51 0 > 12874 > 72 4681 44988 52 0 > 10466 > 86 0 31096 50 12336 > 9162 > 103 0 25435 41 0 > 11421 > 124 0 20666 35 0 > 11204 > 149 0 17810 42 4799 > 9953 > 179 0 18568 45 22462 > 12347 > 215 0 20129 89 1492 > 12628 > 258 0 26381 174 6605 > 13687 > 310 0 34731 412 19060 > 12978 > 372 0 58609 577 3848 > 17080 > 446 0 59318 1355 9504 > 13231 > 535 0 16462 2778 10685 > 13791 > 642 0 9985 6555 13931 > 12703 > 770 0 6403 14811 16859 > 11261 > 924 0 4149 34059 12032 > 12311 > 1109 0 3910 77621 8868 > 11829 > 1331 0 3617 154224 12096 > 12886 > 1597 0 3400 234929 10006 > 11738 > 1916 0 3156 264185 9443 > 11546 > 2299 0 2862 199240 10557 > 12011 > 2759 0 2548 145502 10370 > 12039 > 3311 0 1809 125259 10600 > 14861 > 3973 0 1401 83855 10081 > 13866 > 4768 0 1035 69407 10028 > 13341 > 5722 0 795 55982 10594 > 12707 > 6866 0 688 47374 10450 > 11429 > 8239 0 338 37722 10608 > 10324 > 9887 0 277 34395 12180 > 9418 > 11864 0 255 31126 12896 > 8542 > 14237 0 223 25607 12829 > 7796 > 17084 0 178 20661 15812 > 6803 > 20501 0 193 15916 15590 > 6128 > 24601 0 198 12606 14920 > 5290 > 29521 0 186 11137 14068 > 4598 > 35425 0 195 11111 14377 > 3891 > 42510 0 138 9597 14468 > 3439 > 51012 0 123 7032 14298 > 2701 > 61214 0 413 5288 15194 > 2333 > 73457 0 1911 3991 13549 > 1942 > 88148 0 940 3392 13178 > 1595 > 105778 0 427 3045 12835 > 1234 > 126934 0 431 2691 12676 > 1080 > 152321 0 987 2504 15068 > 787 > 182785 0 484 2337 14566 > 600 > 219342 0 595 2052 13664 > 516 > 263210 0 480 1736 13538 > 377 > 315852 0 428 1428 12007 > 316 > 379022 0 367 1093 10802 > 240 > 454826 0 330 816 9819 > 172 > 545791 0 285 555 8931 > 155 > 654949 0 240 339 8153 > 121 > 785939 0 168 201 7212 > 99 > 943127 0 154 103 6477 > 64 > 1131752 0 141 89 5693 > 65 > 1358102 0 135 87 4885 > 51 > 1629722 0 97 66 4210 > 42 > 1955666 0 88 18 3689 > 38 > 2346799 0 78 7 3027 > 18 > 2816159 0 73 21 2528 > 23 > 3379391 0 120 54 2104 > 11 > 4055269 0 70 107 1797 > 7 > 4866323 0 118 37 1429 > 3 > 5839588 0 56 73 1155 > 3 > 7007506 0 31 52 895 > 4 > 8409007 0 36 52 678 > 3 > 10090808 0 29 4 614 > 4 > 12108970 0 34 17 402 > 2 > 14530764 0 31 7 369 > 3 > 17436917 0 22 47 229 > 2 > 20924300 0 20 1 195 > 0 > 25109160 0 34 36 154 > 1 > > > I'm at an impasse. > > > > -- > View this message in context: > http://cassandra-user-incubator-apache-org.3065146.n2.nabble.com/GC-eden-filled-instantly-any-size-Dropping-messages-tp7592447.html > Sent from the cassandra-u...@incubator.apache.org mailing list archive at > Nabble.com. >