And when I'm checking memory usage per process, it looks normal, java took
only 300Mb as it supposed to, but free -m still shows no memory:
# ps axo pid,vsz,comm=|sort -n -k 2
PID VSZ
465 26396 systemd-logind
444 26724 dbus-daemon
454 27984 avahi-daemon
443 28108 avahi-daemon
344 32720 systemd-journal
1 41212 systemd
364 43132 systemd-udevd
27138 52688 sftp-server
511 53056 wpa_supplicant
769 82548 sshd
30734 83972 sshd
1068 91128 master
28534 91232 pickup
1073 91300 qmgr
519 110032 agetty
27029 115380 bash
27145 115380 bash
30736 116440 sort
385 116720 auditd
489 126332 crond
30733 139624 sshd
27027 140840 sshd
27136 140840 sshd
27143 140840 sshd
30735 148904 ps
438 242360 rsyslogd
466 447932 NetworkManager
510 527448 polkitd
770 553060 tuned
30074 2922460 java
# free -m
total used free shared buff/cache
available
Mem: 489 424 5 0 58
41
Swap: 255 57 198
On Mon, Apr 25, 2016 at 2:52 PM, Eugene Strokin <[email protected]> wrote:
> thanks for your help, but I still struggling with the System OOMKiller
> issue.
> I was doing more digging. And still couldn't find the problem.
> All settings are normal overcommit_memory=0, overcommit_ratio=50.
> free -m before the process starts:
>
> # free -m
> total used free shared buff/cache
> available
> Mem: 489 25 399 1 63
> 440
> Swap: 255 57 198
>
> I start my process like this:
>
> *java* -server -Xmx300m -Xms300m -XX:+HeapDumpOnOutOfMemoryError
> -XX:+UseConcMarkSweepGC -XX:CMSInitiatingOccupancyFraction=55 -jar
> /opt/ccio-image.jar
>
> So, I should still have about 99Mb of free memory, but:
>
> # free -m
> total used free shared buff/cache
> available
> Mem: 489 409 6 1 73
> 55
> Swap: 255 54 201
>
> And I didn't even make a single call to the process yet. It joined the
> cluster, and loaded data from overflow files. And all my free memory is
> gone. Even though I've set 300Mb max for Java.
> As I mentioned before, I've set off-heap memory setting to false:
>
> Cache cache = new CacheFactory()
> .set("locators", LOCATORS.get())
> .set("start-locator", LOCATOR_IP.get()+"["+LOCATOR_PORT.get()+"]")
> .set("bind-address", LOCATOR_IP.get())
> .create();
>
> cache.createDiskStoreFactory()
> .setMaxOplogSize(500)
> .setDiskDirsAndSizes(new File[] { new File("/opt/ccio/geode/store") }, new
> int[] { 18000 })
> .setCompactionThreshold(95)
> .create("-ccio-store");
>
> RegionFactory<String, byte[]> regionFactory = cache.createRegionFactory();
>
> Region<String, byte[]> region = regionFactory
> .setDiskStoreName("-ccio-store")
> .setDataPolicy(DataPolicy.PERSISTENT_PARTITION)
> .setOffHeap(false)
> .setMulticastEnabled(false)
> .setCacheLoader(new AwsS3CacheLoader())
> .create("ccio-images");
>
> I don't understand how the memory is getting overcommitted.
>
> Eugene
>
> On Fri, Apr 22, 2016 at 8:03 PM, Barry Oglesby <[email protected]>
> wrote:
>
>> The OOM killer uses the overcommit_memory and overcommit_ratio parameters
>> to determine if / when to kill a process.
>>
>> What are the settings for these parameters in your environment?
>>
>> The defaults are 0 and 50.
>>
>> cat /proc/sys/vm/overcommit_memory
>> 0
>>
>> cat /proc/sys/vm/overcommit_ratio
>> 50
>>
>> How much free memory is available before you start the JVM?
>>
>> How much free memory is available when your process is killed?
>>
>> You can monitor free memory using either free or vmstat before and during
>> your test.
>>
>> Run free -m in a loop to monitor free memory like:
>>
>> free -ms2
>> total used free shared buffers cached
>> Mem: 290639 35021 255617 0 9215 21396
>> -/+ buffers/cache: 4408 286230
>> Swap: 20473 0 20473
>>
>> Run vmstat in a loop to monitor memory like:
>>
>> vmstat -SM 2
>> procs -----------memory---------- ---swap-- -----io---- --system--
>> -----cpu-----
>> r b swpd free buff cache si so bi bo in cs us sy
>> id wa st
>> 0 0 0 255619 9215 21396 0 0 0 23 0 0 2 0
>> 98 0 0
>> 0 0 0 255619 9215 21396 0 0 0 0 121 198 0 0
>> 100 0 0
>> 0 0 0 255619 9215 21396 0 0 0 0 102 189 0 0
>> 100 0 0
>> 0 0 0 255619 9215 21396 0 0 0 0 110 195 0 0
>> 100 0 0
>> 0 0 0 255619 9215 21396 0 0 0 0 117 205 0 0
>> 100 0 0
>>
>>
>> Thanks,
>> Barry Oglesby
>>
>>
>> On Fri, Apr 22, 2016 at 4:44 PM, Dan Smith <[email protected]> wrote:
>>
>>> The java metaspace will also take up memory. Maybe try setting
>>> -XX:MaxMetaspaceSize
>>>
>>> -Dan
>>>
>>>
>>> -------- Original message --------
>>> From: Eugene Strokin <[email protected]>
>>> Date: 4/22/2016 4:34 PM (GMT-08:00)
>>> To: [email protected]
>>> Subject: Re: System Out of Memory
>>>
>>> The machine is small, it has only 512mb RAM, plus 256mb swap.
>>> But java is set max heap size to 400mb. I've tried less, no help. And
>>> the most interesting part is that I don't see Java OOM Exceptions at all. I
>>> even included a code with memory leak, and I saw the Java OOM Exceptions
>>> before the java process got killed then.
>>> I've browsed internet, and some people are actually noticed the same
>>> problem with other frameworks, not Geode. So, I'm suspecting this could be
>>> not Geode, but Geode was the first suspect because it has off-heap storage
>>> feature. They say that there was a memory leak, but for some reason OS was
>>> killing the process even before Java was getting OOM,
>>> I'll connect with JProbe, and will be monitoring the system with the
>>> console. Will let you know if I'll find something interesting.
>>>
>>> Thanks,
>>> Eugene
>>>
>>>
>>> On Fri, Apr 22, 2016 at 5:55 PM, Dan Smith <[email protected]> wrote:
>>>
>>>> What's your -Xmx for your JVM set to, and how much memory does your
>>>> droplet have? Does it have any swap space? My guess is you need to
>>>> reduce the heap size of your JVM and the OS is killing your process
>>>> because there is not enough memory left.
>>>>
>>>> -Dan
>>>>
>>>> On Fri, Apr 22, 2016 at 1:55 PM, Darrel Schneider <
>>>> [email protected]> wrote:
>>>> > I don't know why your OS would be killing your process which seems
>>>> like your
>>>> > main problem.
>>>> >
>>>> > But I did want you to know that if you don't have any regions with
>>>> > off-heap=true then you have no reason to have off-heap-memory-size to
>>>> be set
>>>> > to anything other than 0.
>>>> >
>>>> > On Fri, Apr 22, 2016 at 12:48 PM, Eugene Strokin <[email protected]
>>>> >
>>>> > wrote:
>>>> >>
>>>> >> I'm running load tests on the Geode cluster I've built.
>>>> >> The OS is killing my process occasionally, complaining that the
>>>> process
>>>> >> takes too much memory:
>>>> >>
>>>> >> # dmesg
>>>> >> [ 2544.932226] Out of memory: Kill process 5382 (java) score 780 or
>>>> >> sacrifice child
>>>> >> [ 2544.933591] Killed process 5382 (java) total-vm:3102804kB,
>>>> >> anon-rss:335780kB, file-rss:0kB
>>>> >>
>>>> >> Java doesn't have any problems, I don't see OOM exception.
>>>> >> Looks like Geode is using off-heap memory. But I set offHeap to
>>>> false for
>>>> >> my region, and I do have only one region:
>>>> >>
>>>> >> RegionFactory<String, byte[]> regionFactory =
>>>> cache.createRegionFactory();
>>>> >> regionFactory
>>>> >> .setDiskStoreName("-ccio-store")
>>>> >> .setDataPolicy(DataPolicy.PERSISTENT_PARTITION)
>>>> >> .setOffHeap(false)
>>>> >> .setCacheLoader(new AwsS3CacheLoader());
>>>> >>
>>>> >> Also, I've played with off-heap-memory-size setting, setting it to
>>>> small
>>>> >> number like 20M to prevent Geode to take too much off-heap memory,
>>>> but
>>>> >> result is the same.
>>>> >>
>>>> >> Do you have any other ideas what could I do here? I'm stack at this
>>>> point.
>>>> >>
>>>> >> Thank you,
>>>> >> Eugene
>>>> >
>>>> >
>>>>
>>>
>>>
>>
>