Yeah, you’re right about 400G, I dropped a digit reading it out of your top 
display. 

So what are you seeing in the dashboard, I’m not sure I understand the 
disconnect between the top you shared and what you’re seeing there. It shows 
lots more than 110G in use, I gather? Or are you seeing this on the hosts page 
per host mem use?

> On Dec 12, 2018, at 12:34 AM, Tony Brian Albers <t...@kb.dk> wrote:
> 
> I'm not following you on the 42G available, the way I see it there's
> 400+G available:
> 
> [root@man-001 ~]# free -h
>               total        used        free      shared  buff/cache   a
> vailable
> Mem:           503G         96G         19G        205M        387G    
>     405G
> Swap:          4.0G        520K        4.0G
> 
> And here's top sorted by %mem usage:
> 
> top - 07:29:00 up 104 days, 20:56,  1 user,  load average: 0.59, 0.68,
> 0.67
> Tasks: 564 total,   1 running, 563 sleeping,   0 stopped,   0 zombie
> %Cpu(s):  1.5 us,  0.2 sy,  0.0 ni, 98.2 id,  0.0 wa,  0.0 hi,  0.0
> si,  0.0 st
> KiB Mem : 52807689+total, 20085144 free, 10132981+used,
> 40666195+buff/cache
> KiB Swap:  4194300 total,  4193780 free,      520 used. 42491062+avail
> Mem 
> 
>    PID USER      PR  NI    VIRT    RES    SHR S  %CPU %MEM     TIME+
> COMMAND                                     
>   5517 qemu      20   0 9128268   8.1g  14084 S   3.0  1.6   5892:07
> qemu-kvm                                    
>  14187 qemu      20   0 9210236   8.1g  14072 S   5.3  1.6   6586:00
> qemu-kvm                                    
>  12791 qemu      20   0 9272448   8.1g  14140 S  14.2  1.6  17452:10
> qemu-kvm                                    
> 135526 qemu      20   0 9117748   8.1g  13664 S   2.3  1.6   5874:48
> qemu-kvm                                    
>   7938 qemu      20   0 9129936   8.1g  13744 S   2.3  1.6  22109:28
> qemu-kvm                                    
>  11764 qemu      20   0 9275520   8.1g  13720 S   3.3  1.6  10679:25
> qemu-kvm                                    
>  12066 qemu      20   0 9360552   8.1g  13708 S   3.0  1.6  10724:34
> qemu-kvm                                    
>  11153 qemu      20   0 9113544   8.1g  13700 S  15.6  1.6  19050:12
> qemu-kvm                                    
>  12436 qemu      20   0 9161800   8.1g  13712 S  16.2  1.6  21268:00
> qemu-kvm                                    
>   6902 qemu      20   0 9110480   8.0g  13580 S   0.7  1.6   1804:16
> qemu-kvm                                    
>   7621 qemu      20   0 9203816   4.8g  14264 S   1.7  1.0   3143:35
> qemu-kvm                                    
>   6587 qemu      20   0 4880980   4.1g  13744 S   0.7  0.8   2354:56
> qemu-kvm                                    
>   7249 qemu      20   0 4913084   1.6g  13712 S   0.7  0.3   1380:38
> qemu-kvm                                    
> 111877 qemu      20   0 1911088 <tel:1911088>   1.1g  14076 S   0.3  0.2 
> 419:58.70
> qemu-kvm                                    
>   4602 vdsm       0 -20 4803160 <tel:20%204803160> 114184  13860 S   1.3  0.0 
>   2143:44
> vdsmd                                       
>   4058 root      15  -5 1154020 <tel:1154020>  38804   9588 S   0.0  0.0   
> 0:00.81
> supervdsmd                                  
>    818 root      20   0   84576  35356  34940 S   0.0  0.0   1:05.60
> systemd-journal                             
>   3602 root      20   0 1496796 <tel:1496796>  32536   9232 S   0.0  0.0 
> 123:53.70
> python                                      
>   2672 root      20   0  358328  30228   7984 S   0.0  0.0   0:14.76
> firewalld                                   
>   4801 vdsm      20   0 1640996 <tel:1640996>  28904   5484 S   0.0  0.0   
> 1265:14
> python
> 
> 
> Rebooting a host doesn't help, (I've tried that earlier) the only thing
> that works is to stop all vm's, reboot all hosts at the same time and
> start vm's again. Then memory usage shown in the dashboard slowly
> increases over time again.
> 
> /tony
> 
> 
> 
> 
> 
> 
> On Tue, 2018-12-11 at 14:09 -0600, Darrell Budic wrote:
>> That’s only reporting 42G available of your 512, ok but something
>> still using it. Try sorting the top by memory %, should be ‘>’ while
>> it’s running.
>> 
>>> On Dec 11, 2018, at 1:39 AM, Tony Brian Albers <t...@kb.dk> wrote:
>>> 
>>> Looks ok to me:
>>> 
>>> top - 08:38:07 up 103 days, 22:05,  1 user,  load average: 0.68,
>>> 0.62,
>>> 0.57
>>> Tasks: 565 total,   1 running, 564 sleeping,   0 stopped,   0
>>> zombie
>>> %Cpu(s):  1.0 us,  0.5 sy,  0.0 ni, 98.5 id,  0.0 wa,  0.0 hi,  0.0
>>> si,  0.0 st
>>> KiB Mem : 52807689+total, 22355988 free, 10132873+used,
>>> 40439219+buff/cache
>>> KiB Swap:  4194300 total,  4193780 free,      520 used.
>>> 42492028+avail
>>> Mem 
>>> 
>>>    PID USER      PR  NI    VIRT    RES    SHR S  %CPU
>>> %MEM     TIME+
>>> COMMAND                  
>>>  14187 qemu      20   0 9144668   8.1g  14072
>>> S  12.6  1.6   6506:46
>>> qemu-kvm                 
>>>  11153 qemu      20   0 9244680   8.1g  13700
>>> S   4.3  1.6  18881:11
>>> qemu-kvm                 
>>>  12436 qemu      20   0 9292936   8.1g  13712
>>> S   3.3  1.6  21071:56
>>> qemu-kvm                 
>>>   5517 qemu      20   0 9128268   8.1g  14084
>>> S   3.0  1.6   5801:03
>>> qemu-kvm                 
>>>  11764 qemu      20   0 9185364   8.1g  13720
>>> S   3.0  1.6  10585:14
>>> qemu-kvm                 
>>>   7938 qemu      20   0 9252876   8.1g  13744
>>> S   2.6  1.6  21912:46
>>> qemu-kvm                 
>>>  12791 qemu      20   0 9182292   8.1g  14140
>>> S   2.6  1.6  17299:36
>>> qemu-kvm                 
>>>   4602 vdsm       0 -20 4803160 <tel:20%204803160> <tel:20%204803160 
>>> <tel:20%204803160>> 114132  13860
>>> S   2.3  0.0   2123:45
>>> vdsmd                    
>>>   7621 qemu      20   0 9187424   4.8g  14264
>>> S   2.3  1.0   3114:25
>>> qemu-kvm                 
>>>  12066 qemu      20   0 9188436   8.1g  13708
>>> S   2.3  1.6  10629:53
>>> qemu-kvm                 
>>> 135526 qemu      20   0 9298060   8.1g  13664
>>> S   2.0  1.6   5792:05
>>> qemu-kvm                 
>>>   6587 qemu      20   0 4883036   4.1g  13744
>>> S   1.3  0.8   2334:54
>>> qemu-kvm                 
>>>   3814 root      20   0 1450200 <tel:1450200> <tel:1450200 <tel:1450200>>  
>>> 25096  14208
>>> S   1.0  0.0 368:03.80
>>> libvirtd                 
>>>   6902 qemu      20   0 9110480   8.0g  13580
>>> S   1.0  1.6   1787:57
>>> qemu-kvm                 
>>>   7249 qemu      20   0 4913084   1.6g  13712
>>> S   0.7  0.3   1367:32
>>> qemu-kvm                 
>>> 
>>> 
>>> It looks like it's only in oVirt-engine that there's an issue. The
>>> host
>>> seems happy enough.
>>> 
>>> /tony
>>> 
>>> 
>>> 
>>> On Mon, 2018-12-10 at 20:14 -0600, Darrell Budic wrote:
>>>> Grab a shell on your hosts and check top memory use quick. Could
>>>> be
>>>> VDSMD, in which case restarting the process will give you a temp
>>>> fix.
>>>> If you’re running hyperconvered, check your gluster version,
>>>> there
>>>> was a leak in versions 3.12.7 - 3.1.12 or so, updating
>>>> ovirt/gluster
>>>> is the best fix for that.
>>>> 
>>>>> On Dec 10, 2018, at 7:36 AM, Tony Brian Albers <t...@kb.dk>
>>>>> wrote:
>>>>> 
>>>>> Hi guys,
>>>>> 
>>>>> We have a small test installation here running around 30 vms on
>>>>> 2
>>>>> hosts.
>>>>> 
>>>>> oVirt 4.2.5.3
>>>>> 
>>>>> The hosts each have 512 GB memory, and the vms are sized with
>>>>> 4-8
>>>>> GB
>>>>> each.
>>>>> 
>>>>> I have noticed that over the last months, the memory usage in
>>>>> the
>>>>> dashboard has been increasing and is now showing 946.8 GB used
>>>>> of
>>>>> 1007.2 GB.
>>>>> 
>>>>> What can be causing this?
>>>>> 
>>>>> TIA,
>>>>> 
>>>>> -- 
>>>>> -- 
>>>>> Tony Albers
>>>>> Systems Architect
>>>>> Systems Director, National Cultural Heritage Cluster
>>>>> Royal Danish Library, Victor Albecks Vej 1, 8000 Aarhus C,
>>>>> Denmark.
>>>>> Tel: +45 2566 2383 / +45 8946 2316
>>>>> _______________________________________________
>>>>> Users mailing list -- users@ovirt.org
>>>>> To unsubscribe send an email to users-le...@ovirt.org
>>>>> Privacy Statement: https://www.ovirt.org/site/privacy-policy/
>>>>> oVirt Code of Conduct: https://www.ovirt.org/community/about/co
>>>>> mmun
>>>>> ity-guidelines/
>>>>> List Archives: https://lists.ovirt.org/archives/list/users@ovir
>>>>> t.or
>>>>> g/message/SDDH2OC5RBOVYYCLGPOUF6HO676HWI5U/
>>>> 
>>>> 
>>> 
>>> -- 
>>> Tony Albers
>>> Systems Architect
>>> Systems Director, National Cultural Heritage Cluster
>>> Royal Danish Library, Victor Albecks Vej 1, 8000 Aarhus C, Denmark.
>>> Tel: +45 2566 2383 <tel:+45%202566%202383> <tel:+45%202566%202383 
>>> <tel:+45%202566%202383>> / +45 8946 2316 <tel:+45%208946%202316>
>>> <tel:+45%208946%202316 <tel:+45%208946%202316>>
> 
> -- 
> -- 
> Tony Albers
> Systems Architect
> Systems Director, National Cultural Heritage Cluster
> Royal Danish Library, Victor Albecks Vej 1, 8000 Aarhus C, Denmark.
> Tel: +45 2566 2383 <tel:+45%202566%202383> / +45 8946 2316 
> <tel:+45%208946%202316>
_______________________________________________
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/ZJYANHN72PLA4MKWY45TK7QNU7HEB7VQ/

Reply via email to