Why is this a problem? CPU load and log file size. The CPU load reported by top increases from it's normal 0-1% resting state to 25-50% continually, and higher when searching, indexing, etc.
With 6.1.0 we just got a huge GC log file. With 6.4.1 the GC log is now cycled before it grows too large, but it does mean that they are quickly cycled (every 3 hours). Operations seem much slower in this state but I haven't collected figures for this. When Solr gets into this state it stays like it until it's restarted. Thanks, Leon Stringer > > On 14 February 2017 at 05:45 Erick Erickson <erickerick...@gmail.com> > wrote: > > > Why is this a problem? Are you seeing unacceptable slowness? > It's fairly common for Java to frequently do GC, the problem happens > when it uses stop-the-world GC. So unless you're seeing visibly > slow performance I'd say ignore it. > > Curiously, increasing the Java heap a little bit sometimes helps > as I've seen situations where the GC recovers so little memory that > another GC cycle immediately occurs. That said I don't see evidence > of this in what you showed. > > GCViewer is a nifty tool for visualizing the GC activity BTW. > > Best, > Erick > > On Mon, Feb 13, 2017 at 8:36 AM, Leon STRINGER > <leon.strin...@ntlworld.com> wrote: > > Hi, > > > > I get an issue where, when I'm deleting and adding Solr cores, it > > appears to go > > into a loop increasing CPU load and continually (every 2 seconds) > > logging to the > > garbage collection log. > > > > I had this problem with 6.1.0 so we've just upgraded to 6.4.1 and the > > issue > > still occurs. The entries being logged every 2 seconds are below (hope > > it's not > > too verbose). Obviously this means the log gets big quickly. > > > > We can work around the issue by restarting Solr but presumably something > > has > > gone wrong. Can anyone suggest if we're doing something incorrectly to > > cause > > this, or if it's an issue we can troubleshoot. > > > > Any advice gratefully received. > > > > On CentOS 7 with OpenJDK 1.8.0_91-b14. > > > > solr_gc.log.0.current logs the following every 2 seconds: > > > > 2017-02-13T16:19:11.230+0000: 5092.640: [GC (CMS Initial Mark) [1 > > CMS-initial-mark: 225270K(393216K)] 225280K(502464K), 0.0030517 secs] > > [Times: > > user=0.01 sys=0.00, real=0.01 secs] > > 2017-02-13T16:19:11.234+0000: 5092.643: Total time for which application > > threads > > were stopped: 0.0033800 seconds, Stopping threads took: 0.0000473 > > seconds > > 2017-02-13T16:19:11.234+0000: 5092.643: [CMS-concurrent-mark-start] > > 2017-02-13T16:19:11.359+0000: 5092.769: [CMS-concurrent-mark: > > 0.125/0.125 secs] > > [Times: user=0.50 sys=0.00, real=0.12 secs] > > 2017-02-13T16:19:11.359+0000: 5092.769: [CMS-concurrent-preclean-start] > > 2017-02-13T16:19:11.361+0000: 5092.771: [CMS-concurrent-preclean: > > 0.002/0.002 > > secs] [Times: user=0.00 sys=0.00, real=0.00 secs] > > 2017-02-13T16:19:11.362+0000: 5092.771: [GC (CMS Final Remark) [YG > > occupancy: 10 > > K (109248 K)]{Heap before GC invocations=3236 (full 1150): > > par new generation total 109248K, used 10K [0x00000000e0000000, > > 0x00000000e8000000, 0x00000000e8000000) > > eden space 87424K, 0% used [0x00000000e0000000, 0x00000000e0001020, > > 0x00000000e5560000) > > from space 21824K, 0% used [0x00000000e6ab0000, 0x00000000e6ab1830, > > 0x00000000e8000000) > > to space 21824K, 0% used [0x00000000e5560000, 0x00000000e5560000, > > 0x00000000e6ab0000) > > concurrent mark-sweep generation total 393216K, used 225270K > > [0x00000000e8000000, 0x0000000100000000, 0x0000000100000000) > > Metaspace used 176850K, capacity 179580K, committed 181092K, reserved > > 1210368K > > class space used 18794K, capacity 19506K, committed 19836K, reserved > > 1048576K > > 2017-02-13T16:19:11.362+0000: 5092.771: [GC (CMS Final Remark) > > 2017-02-13T16:19:11.362+0000: 5092.771: [ParNew > > Desired survivor size 20112992 bytes, new threshold 8 (max 8) > > - age 2: 160 bytes, 160 total > > - age 4: 32 bytes, 192 total > > : 10K->6K(109248K), 0.0041872 secs] 225280K->225276K(502464K), 0.0042455 > > secs] > > [Times: user=0.01 sys=0.00, real=0.01 secs] > > Heap after GC invocations=3237 (full 1150): > > par new generation total 109248K, used 6K [0x00000000e0000000, > > 0x00000000e8000000, 0x00000000e8000000) > > eden space 87424K, 0% used [0x00000000e0000000, 0x00000000e0000000, > > 0x00000000e5560000) > > from space 21824K, 0% used [0x00000000e5560000, 0x00000000e5561830, > > 0x00000000e6ab0000) > > to space 21824K, 0% used [0x00000000e6ab0000, 0x00000000e6ab0000, > > 0x00000000e8000000) > > concurrent mark-sweep generation total 393216K, used 225270K > > [0x00000000e8000000, 0x0000000100000000, 0x0000000100000000) > > Metaspace used 176850K, capacity 179580K, committed 181092K, reserved > > 1210368K > > class space used 18794K, capacity 19506K, committed 19836K, reserved > > 1048576K > > } > > 2017-02-13T16:19:11.366+0000: 5092.775: [Rescan (parallel) , 0.0018980 > > secs]2017-02-13T16:19:11.368+0000: 5092.777: [weak refs processing, > > 0.0004940 > > secs]2017-02-13T16:19:11.368+0000: 5092.778: [class unloading, 0.0580950 > > secs]2017-02-13T16:19:11.426+0000: 5092.836: [scrub symbol table, > > 0.0110875 > > secs]2017-02-13T16:19:11.438+0000: 5092.847: [scrub string table, > > 0.0019072 > > secs][1 CMS-remark: 225270K(393216K)] 225276K(502464K), 0.0780250 secs] > > [Times: > > user=0.09 sys=0.00, real=0.08 secs] > > 2017-02-13T16:19:11.440+0000: 5092.849: Total time for which application > > threads > > were stopped: 0.0782677 seconds, Stopping threads took: 0.0000411 > > seconds > > 2017-02-13T16:19:11.440+0000: 5092.849: [CMS-concurrent-sweep-start] > > 2017-02-13T16:19:11.546+0000: 5092.955: [CMS-concurrent-sweep: > > 0.106/0.106 secs] > > [Times: user=0.11 sys=0.00, real=0.11 secs] > > 2017-02-13T16:19:11.546+0000: 5092.955: [CMS-concurrent-reset-start] > > 2017-02-13T16:19:11.546+0000: 5092.956: [CMS-concurrent-reset: > > 0.001/0.001 secs] > > [Times: user=0.00 sys=0.00, real=0.00 secs] > > > > Regards, > > > > Leon Stringer >