Is sombody using the UseG1GC garbage collector with Solr and Tomcat 7?
Any extra options needed?
Thanks...
On 03/25/2013 08:34 AM, Arkadi Colson wrote:
I changed my system memory to 12GB. Solr now gets -Xms2048m -Xmx8192m
as parameters. I also added -XX:+UseG1GC to the java process. But now
the whole machine crashes! Any idea why?
Mar 22 20:30:01 solr01-gs kernel: [716098.077809] java invoked
oom-killer: gfp_mask=0x201da, order=0, oom_adj=0
Mar 22 20:30:01 solr01-gs kernel: [716098.077962] java cpuset=/
mems_allowed=0
Mar 22 20:30:01 solr01-gs kernel: [716098.078019] Pid: 29339, comm:
java Not tainted 2.6.32-5-amd64 #1
Mar 22 20:30:01 solr01-gs kernel: [716098.078095] Call Trace:
Mar 22 20:30:01 solr01-gs kernel: [716098.078155] [<ffffffff810b6324>]
? oom_kill_process+0x7f/0x23f
Mar 22 20:30:01 solr01-gs kernel: [716098.078233] [<ffffffff810b6848>]
? __out_of_memory+0x12a/0x141
Mar 22 20:30:01 solr01-gs kernel: [716098.078309] [<ffffffff810b699f>]
? out_of_memory+0x140/0x172
Mar 22 20:30:01 solr01-gs kernel: [716098.078385] [<ffffffff810ba704>]
? __alloc_pages_nodemask+0x4ec/0x5fc
Mar 22 20:30:01 solr01-gs kernel: [716098.078469] [<ffffffff812fb47a>]
? io_schedule+0x93/0xb7
Mar 22 20:30:01 solr01-gs kernel: [716098.078541] [<ffffffff810bbc69>]
? __do_page_cache_readahead+0x9b/0x1b4
Mar 22 20:30:01 solr01-gs kernel: [716098.078626] [<ffffffff81064fc0>]
? wake_bit_function+0x0/0x23
Mar 22 20:30:01 solr01-gs kernel: [716098.078702] [<ffffffff810bbd9e>]
? ra_submit+0x1c/0x20
Mar 22 20:30:01 solr01-gs kernel: [716098.078773] [<ffffffff810b4a72>]
? filemap_fault+0x17d/0x2f6
Mar 22 20:30:01 solr01-gs kernel: [716098.078849] [<ffffffff810ca9e2>]
? __do_fault+0x54/0x3c3
Mar 22 20:30:01 solr01-gs kernel: [716098.078921] [<ffffffff810ccd36>]
? handle_mm_fault+0x3b8/0x80f
Mar 22 20:30:01 solr01-gs kernel: [716098.078999] [<ffffffff8101166e>]
? apic_timer_interrupt+0xe/0x20
Mar 22 20:30:01 solr01-gs kernel: [716098.079078] [<ffffffff812febf6>]
? do_page_fault+0x2e0/0x2fc
Mar 22 20:30:01 solr01-gs kernel: [716098.079153] [<ffffffff812fca95>]
? page_fault+0x25/0x30
Mar 22 20:30:01 solr01-gs kernel: [716098.079222] Mem-Info:
Mar 22 20:30:01 solr01-gs kernel: [716098.079261] Node 0 DMA per-cpu:
Mar 22 20:30:01 solr01-gs kernel: [716098.079310] CPU 0: hi: 0,
btch: 1 usd: 0
Mar 22 20:30:01 solr01-gs kernel: [716098.079374] CPU 1: hi: 0,
btch: 1 usd: 0
Mar 22 20:30:01 solr01-gs kernel: [716098.079439] CPU 2: hi: 0,
btch: 1 usd: 0
Mar 22 20:30:01 solr01-gs kernel: [716098.079527] CPU 3: hi: 0,
btch: 1 usd: 0
Mar 22 20:30:01 solr01-gs kernel: [716098.079591] Node 0 DMA32 per-cpu:
Mar 22 20:30:01 solr01-gs kernel: [716098.079642] CPU 0: hi: 186,
btch: 31 usd: 0
Mar 22 20:30:01 solr01-gs kernel: [716098.079706] CPU 1: hi: 186,
btch: 31 usd: 0
Mar 22 20:30:01 solr01-gs kernel: [716098.079770] CPU 2: hi: 186,
btch: 31 usd: 0
Mar 22 20:30:01 solr01-gs kernel: [716098.079834] CPU 3: hi: 186,
btch: 31 usd: 0
Mar 22 20:30:01 solr01-gs kernel: [716098.079899] Node 0 Normal per-cpu:
Mar 22 20:30:01 solr01-gs kernel: [716098.079951] CPU 0: hi: 186,
btch: 31 usd: 17
Mar 22 20:30:01 solr01-gs kernel: [716098.080015] CPU 1: hi: 186,
btch: 31 usd: 0
Mar 22 20:30:01 solr01-gs kernel: [716098.080079] CPU 2: hi: 186,
btch: 31 usd: 2
Mar 22 20:30:01 solr01-gs kernel: [716098.080142] CPU 3: hi: 186,
btch: 31 usd: 0
Mar 22 20:30:01 solr01-gs kernel: [716098.080209] active_anon:2638016
inactive_anon:388557 isolated_anon:0
Mar 22 20:30:01 solr01-gs kernel: [716098.080209] active_file:68
inactive_file:236 isolated_file:0
Mar 22 20:30:01 solr01-gs kernel: [716098.080210] unevictable:0
dirty:5 writeback:5 unstable:0
Mar 22 20:30:01 solr01-gs kernel: [716098.080211] free:16573
slab_reclaimable:2398 slab_unreclaimable:2335
Mar 22 20:30:01 solr01-gs kernel: [716098.080212] mapped:36 shmem:0
pagetables:24750 bounce:0
Mar 22 20:30:01 solr01-gs kernel: [716098.080575] Node 0 DMA
free:15796kB min:16kB low:20kB high:24kB active_anon:0kB
inactive_anon:0kB active_file:0kB inactive_file:0kB unevictable:0kB
isolated(anon):0kB isolated(file):0kB present:15244kB mlocked:0kB
dirty:0kB writeback:0kB mapped:0kB shmem:0kB slab_reclaimable:0kB
slab_unreclaimable:8kB kernel_stack:0kB pagetables:0kB unstable:0kB
bounce:0kB writeback_tmp:0kB pages_scanned:0 all_unreclaimable? yes
Mar 22 20:30:01 solr01-gs kernel: [716098.081041] lowmem_reserve[]: 0
3000 12090 12090
Mar 22 20:30:01 solr01-gs kernel: [716098.081110] Node 0 DMA32
free:39824kB min:3488kB low:4360kB high:5232kB active_anon:2285240kB
inactive_anon:520624kB active_file:0kB inactive_file:188kB
unevictable:0kB isolated(anon):0kB isolated(file):0kB
present:3072096kB mlocked:0kB dirty:0kB writeback:0kB mapped:0kB
shmem:0kB slab_reclaimable:4152kB slab_unreclaimable:1640kB
kernel_stack:1104kB pagetables:31100kB unstable:0kB bounce:0kB
writeback_tmp:0kB pages_scanned:89 all_unreclaimable? no
Mar 22 20:30:01 solr01-gs kernel: [716098.081600] lowmem_reserve[]: 0
0 9090 9090
Mar 22 20:30:01 solr01-gs kernel: [716098.081664] Node 0 Normal
free:10672kB min:10572kB low:13212kB high:15856kB
active_anon:8266824kB inactive_anon:1033604kB active_file:292kB
inactive_file:756kB unevictable:0kB isolated(anon):0kB
isolated(file):0kB present:9308160kB mlocked:0kB dirty:20kB
writeback:20kB mapped:156kB shmem:0kB slab_reclaimable:5440kB
slab_unreclaimable:7692kB kernel_stack:1280kB pagetables:67900kB
unstable:0kB bounce:0kB writeback_tmp:0kB pages_scanned:256
all_unreclaimable? no
Mar 22 20:30:01 solr01-gs kernel: [716098.082171] lowmem_reserve[]: 0
0 0 0
Mar 22 20:30:01 solr01-gs kernel: [716098.082240] Node 0 DMA: 1*4kB
2*8kB 2*16kB 0*32kB 2*64kB 0*128kB 1*256kB 0*512kB 1*1024kB 1*2048kB
3*4096kB = 15796kB
Mar 22 20:30:01 solr01-gs kernel: [716098.082394] Node 0 DMA32:
4578*4kB 2434*8kB 4*16kB 1*32kB 1*64kB 2*128kB 0*256kB 2*512kB
1*1024kB 0*2048kB 0*4096kB = 40248kB
Mar 22 20:30:01 solr01-gs kernel: [716098.082555] Node 0 Normal:
1020*4kB 332*8kB 7*16kB 3*32kB 6*64kB 6*128kB 2*256kB 2*512kB 1*1024kB
0*2048kB 0*4096kB = 10656kB
Mar 22 20:30:01 solr01-gs kernel: [716098.082715] 7069 total pagecache
pages
Mar 22 20:30:01 solr01-gs kernel: [716098.082769] 6768 pages in swap
cache
Mar 22 20:30:01 solr01-gs kernel: [716098.114216] 203 pages shared
Mar 22 20:30:01 solr01-gs kernel: [716098.114261] 3067047 pages
non-shared
Mar 22 20:30:01 solr01-gs kernel: [716098.114314] Out of memory: kill
process 29301 (java) score 37654 or a child
Mar 22 20:30:01 solr01-gs kernel: [716098.114401] Killed process 29301
(java)
Thanks!
On 03/14/2013 04:00 PM, Arkadi Colson wrote:
On 03/14/2013 03:11 PM, Toke Eskildsen wrote:
On Thu, 2013-03-14 at 13:10 +0100, Arkadi Colson wrote:
When I shutdown tomcat free -m and top keeps telling me the same
values.
Almost no free memory...
Any idea?
Are you reading top & free right? It is standard behaviour for most
modern operating systems to have very little free memory. As long as
the
sum of free memory and cache is high, everything is fine.
Looking at the stats you gave previously we have
*top*
PID USER PR NI VIRT RES SHR S %CPU %MEM TIME+
COMMAND &nbs
p;
13666 root 20 0 86.8g 4.7g 248m S 101 39.7 478:37.45
4.7GB physical memory used and ~80GB used for memory mapping the index.
*free -m**
* total used free shared buffers
cached
Mem: 12047 11942 105 0 180 6363
-/+ buffers/cache: 5399 6648
Swap: 956 75 881
So 6648MB used for either general disk cache or memory mapped index.
This really translates to 6648MB (plus the 105MB above) available
memory
as any application asking for memory will get it immediately from that
pool (sorry if this is basic stuff for you).
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908)
at java.lang.Thread.run(Thread.java:662)
Caused by: java.lang.OutOfMemoryError
at java.util.zip.ZipFile.open(Native Method)
at java.util.zip.ZipFile.<init>(ZipFile.java:127)
at java.util.zip.ZipFile.<init>(ZipFile.java:144)
at
org.apache.poi.openxml4j.opc.internal.ZipHelper.openZipFile(ZipHelper.java:157)
[...]
Java HotSpot(TM) 64-Bit Server VM warning: Attempt to allocate stack
guard pages failed.
mmap failed for CEN and END part of zip file
A quick search shows that other people have had problems with
ZipFile in
at least some sub-versions of Java 1.7. However, another very common
cause for OOM with memory mapping is that the limit for allocating
virtual memory is too low.
We do not index zip files so that could not cause the problem
Try doing a
ulimit -v
on the machine. If the number is somewhere around 100000000 (100GB),
Lucene's memory mapping of your index (the 80GB) plus the ZipFile's
memory mapping plus other processes might hit the ceiling. If that is
the case, simply raise the limit.
- Toke
ulimit -v shows me unlimited
I decreased the hard commit time to 10 seconds and set
ramBufferSizeMB to 250. Hope this helps...
Will keep you informed!
Thanks for the explanation!