Yup, doing the 3 > drop_cache is clearly freeing up the memory.

I've filed bug 1659342 in rh bugzilla

Thanks Darrell, have a nice weekend.

/tony


On Thu, 2018-12-13 at 10:54 -0600, Darrell Budic wrote:
> Agree, looks like disk caching, which is considered in use but can be
> freed as things ask for it. What kind of storage are you using here?
> 
> I checked my converged group, the GUI memory use looks like it
> includes the cache as well, just not as extreme as yours :) If you
> wanted to confirm, you could clear it with
> 
> sync; echo 3 > /proc/sys/vm/drop_caches
> 
> and you should get all the cache memory back without rebooting.
> Presumably, it’s all holding filesystem data for quick access, so not
> a terrible thing, just alarming to see 97% utilization. Unless it
> kicked off unneeded MOM runs if you’re using same page merging and
> ballooning, then its doing unneeded work and maybe slowing your VMs…
> You’d probably see it in top if this was the case, it’s fairly CPU
> intensive.
> 
> Ultimately, you may want to file a bug/feature request to have the
> GUI differentiate cache from “regular’ in use memory for clarity.
> 
> > On Dec 13, 2018, at 12:06 AM, Tony Brian Albers <t...@kb.dk> wrote:
> > 
> > Screenshots attached
> > 
> > We have 2 hosts and 24 running vms, all vms are pretty small.
> > 
> > As you can see, top and the dashboard does not agree.
> > 
> > An interesting thing is that if I view the host itself in the
> > engine,
> > it says under "General" that 
> > Max free Memory for scheduling new VMs: 413360 MB
> > 
> > So maybe it's some sort of caching that's using the memory.
> > 
> > 
> > /tony
> > 
> > On Wed, 2018-12-12 at 09:59 -0600, Darrell Budic wrote:
> > > Yeah, you’re right about 400G, I dropped a digit reading it out
> > > of
> > > your top display. 
> > > 
> > > So what are you seeing in the dashboard, I’m not sure I
> > > understand
> > > the disconnect between the top you shared and what you’re seeing
> > > there. It shows lots more than 110G in use, I gather? Or are you
> > > seeing this on the hosts page per host mem use?
> > > 
> > > > On Dec 12, 2018, at 12:34 AM, Tony Brian Albers <t...@kb.dk>
> > > > wrote:
> > > > 
> > > > I'm not following you on the 42G available, the way I see it
> > > > there's
> > > > 400+G available:
> > > > 
> > > > [root@man-001 ~]# free -h
> > > >               total        used        free      shared  buff/c
> > > > ache
> > > >    a
> > > > vailable
> > > > Mem:           503G         96G         19G        205M        
> > > > 387G
> > > >     
> > > >     405G
> > > > Swap:          4.0G        520K        4.0G
> > > > 
> > > > And here's top sorted by %mem usage:
> > > > 
> > > > top - 07:29:00 up 104 days, 20:56,  1 user,  load average:
> > > > 0.59,
> > > > 0.68,
> > > > 0.67
> > > > Tasks: 564 total,   1 running, 563 sleeping,   0 stopped,   0
> > > > zombie
> > > > %Cpu(s):  1.5 us,  0.2 sy,  0.0 ni, 98.2 id,  0.0 wa,  0.0
> > > > hi,  0.0
> > > > si,  0.0 st
> > > > KiB Mem : 52807689+total, 20085144 free, 10132981+used,
> > > > 40666195+buff/cache
> > > > KiB Swap:  4194300 total,  4193780 free,      520 used.
> > > > 42491062+avail
> > > > Mem 
> > > > 
> > > >    PID USER      PR  NI    VIRT    RES    SHR S  %CPU
> > > > %MEM     TIME+
> > > > COMMAND                                     
> > > >   5517 qemu      20   0 9128268   8.1g  14084
> > > > S   3.0  1.6   5892:07
> > > > qemu-kvm                                    
> > > >  14187 qemu      20   0 9210236   8.1g  14072
> > > > S   5.3  1.6   6586:00
> > > > qemu-kvm                                    
> > > >  12791 qemu      20   0 9272448   8.1g  14140
> > > > S  14.2  1.6  17452:10
> > > > qemu-kvm                                    
> > > > 135526 qemu      20   0 9117748   8.1g  13664
> > > > S   2.3  1.6   5874:48
> > > > qemu-kvm                                    
> > > >   7938 qemu      20   0 9129936   8.1g  13744
> > > > S   2.3  1.6  22109:28
> > > > qemu-kvm                                    
> > > >  11764 qemu      20   0 9275520   8.1g  13720
> > > > S   3.3  1.6  10679:25
> > > > qemu-kvm                                    
> > > >  12066 qemu      20   0 9360552   8.1g  13708
> > > > S   3.0  1.6  10724:34
> > > > qemu-kvm                                    
> > > >  11153 qemu      20   0 9113544   8.1g  13700
> > > > S  15.6  1.6  19050:12
> > > > qemu-kvm                                    
> > > >  12436 qemu      20   0 9161800   8.1g  13712
> > > > S  16.2  1.6  21268:00
> > > > qemu-kvm                                    
> > > >   6902 qemu      20   0 9110480   8.0g  13580
> > > > S   0.7  1.6   1804:16
> > > > qemu-kvm                                    
> > > >   7621 qemu      20   0 9203816   4.8g  14264
> > > > S   1.7  1.0   3143:35
> > > > qemu-kvm                                    
> > > >   6587 qemu      20   0 4880980   4.1g  13744
> > > > S   0.7  0.8   2354:56
> > > > qemu-kvm                                    
> > > >   7249 qemu      20   0 4913084   1.6g  13712
> > > > S   0.7  0.3   1380:38
> > > > qemu-kvm                                    
> > > > 111877 qemu      20   0 1911088   1.1g  14076 S   0.3  0.2
> > > > 419:58.70
> > > > qemu-kvm                                    
> > > >   4602 vdsm       0 -20 4803160 114184  13860
> > > > S   1.3  0.0   2143:44
> > > > vdsmd                                       
> > > >   4058 root      15  -5 1154020  38804   9588
> > > > S   0.0  0.0   0:00.81
> > > > supervdsmd                                  
> > > >    818 root      20   0   84576  35356  34940
> > > > S   0.0  0.0   1:05.60
> > > > systemd-journal                             
> > > >   3602 root      20   0 1496796  32536   9232 S   0.0  0.0
> > > > 123:53.70
> > > > python                                      
> > > >   2672 root      20   0  358328  30228   7984
> > > > S   0.0  0.0   0:14.76
> > > > firewalld                                   
> > > >   4801 vdsm      20   0 1640996  28904   5484
> > > > S   0.0  0.0   1265:14
> > > > python
> > > > 
> > > > 
> > > > Rebooting a host doesn't help, (I've tried that earlier) the
> > > > only
> > > > thing
> > > > that works is to stop all vm's, reboot all hosts at the same
> > > > time
> > > > and
> > > > start vm's again. Then memory usage shown in the dashboard
> > > > slowly
> > > > increases over time again.
> > > > 
> > > > /tony
> > > > 
> > > > 
> > > > 
> > > > 
> > > > 
> > > > 
> > > > On Tue, 2018-12-11 at 14:09 -0600, Darrell Budic wrote:
> > > > > That’s only reporting 42G available of your 512, ok but
> > > > > something
> > > > > still using it. Try sorting the top by memory %, should be
> > > > > ‘>’
> > > > > while
> > > > > it’s running.
> > > > > 
> > > > > > On Dec 11, 2018, at 1:39 AM, Tony Brian Albers <t...@kb.dk>
> > > > > > wrote:
> > > > > > 
> > > > > > Looks ok to me:
> > > > > > 
> > > > > > top - 08:38:07 up 103 days, 22:05,  1 user,  load average:
> > > > > > 0.68,
> > > > > > 0.62,
> > > > > > 0.57
> > > > > > Tasks: 565 total,   1 running, 564 sleeping,   0
> > > > > > stopped,   0
> > > > > > zombie
> > > > > > %Cpu(s):  1.0 us,  0.5 sy,  0.0 ni, 98.5 id,  0.0 wa,  0.0
> > > > > > hi,  0.0
> > > > > > si,  0.0 st
> > > > > > KiB Mem : 52807689+total, 22355988 free, 10132873+used,
> > > > > > 40439219+buff/cache
> > > > > > KiB Swap:  4194300 total,  4193780 free,      520 used.
> > > > > > 42492028+avail
> > > > > > Mem 
> > > > > > 
> > > > > >    PID USER      PR  NI    VIRT    RES    SHR S  %CPU
> > > > > > %MEM     TIME+
> > > > > > COMMAND                  
> > > > > >  14187 qemu      20   0 9144668   8.1g  14072
> > > > > > S  12.6  1.6   6506:46
> > > > > > qemu-kvm                 
> > > > > >  11153 qemu      20   0 9244680   8.1g  13700
> > > > > > S   4.3  1.6  18881:11
> > > > > > qemu-kvm                 
> > > > > >  12436 qemu      20   0 9292936   8.1g  13712
> > > > > > S   3.3  1.6  21071:56
> > > > > > qemu-kvm                 
> > > > > >   5517 qemu      20   0 9128268   8.1g  14084
> > > > > > S   3.0  1.6   5801:03
> > > > > > qemu-kvm                 
> > > > > >  11764 qemu      20   0 9185364   8.1g  13720
> > > > > > S   3.0  1.6  10585:14
> > > > > > qemu-kvm                 
> > > > > >   7938 qemu      20   0 9252876   8.1g  13744
> > > > > > S   2.6  1.6  21912:46
> > > > > > qemu-kvm                 
> > > > > >  12791 qemu      20   0 9182292   8.1g  14140
> > > > > > S   2.6  1.6  17299:36
> > > > > > qemu-kvm                 
> > > > > >   4602 vdsm       0 -20 4803160 <tel:20%204803160>
> > > > > > 114132  13860
> > > > > > S   2.3  0.0   2123:45
> > > > > > vdsmd                    
> > > > > >   7621 qemu      20   0 9187424   4.8g  14264
> > > > > > S   2.3  1.0   3114:25
> > > > > > qemu-kvm                 
> > > > > >  12066 qemu      20   0 9188436   8.1g  13708
> > > > > > S   2.3  1.6  10629:53
> > > > > > qemu-kvm                 
> > > > > > 135526 qemu      20   0 9298060   8.1g  13664
> > > > > > S   2.0  1.6   5792:05
> > > > > > qemu-kvm                 
> > > > > >   6587 qemu      20   0 4883036   4.1g  13744
> > > > > > S   1.3  0.8   2334:54
> > > > > > qemu-kvm                 
> > > > > >   3814 root      20   0 1450200 <tel:1450200>  25096  14208
> > > > > > S   1.0  0.0 368:03.80
> > > > > > libvirtd                 
> > > > > >   6902 qemu      20   0 9110480   8.0g  13580
> > > > > > S   1.0  1.6   1787:57
> > > > > > qemu-kvm                 
> > > > > >   7249 qemu      20   0 4913084   1.6g  13712
> > > > > > S   0.7  0.3   1367:32
> > > > > > qemu-kvm                 
> > > > > > 
> > > > > > 
> > > > > > It looks like it's only in oVirt-engine that there's an
> > > > > > issue.
> > > > > > The
> > > > > > host
> > > > > > seems happy enough.
> > > > > > 
> > > > > > /tony
> > > > > > 
> > > > > > 
> > > > > > 
> > > > > > On Mon, 2018-12-10 at 20:14 -0600, Darrell Budic wrote:
> > > > > > > Grab a shell on your hosts and check top memory use
> > > > > > > quick.
> > > > > > > Could
> > > > > > > be
> > > > > > > VDSMD, in which case restarting the process will give you
> > > > > > > a
> > > > > > > temp
> > > > > > > fix.
> > > > > > > If you’re running hyperconvered, check your gluster
> > > > > > > version,
> > > > > > > there
> > > > > > > was a leak in versions 3.12.7 - 3.1.12 or so, updating
> > > > > > > ovirt/gluster
> > > > > > > is the best fix for that.
> > > > > > > 
> > > > > > > > On Dec 10, 2018, at 7:36 AM, Tony Brian Albers <tba@kb.
> > > > > > > > dk>
> > > > > > > > wrote:
> > > > > > > > 
> > > > > > > > Hi guys,
> > > > > > > > 
> > > > > > > > We have a small test installation here running around
> > > > > > > > 30
> > > > > > > > vms on
> > > > > > > > 2
> > > > > > > > hosts.
> > > > > > > > 
> > > > > > > > oVirt 4.2.5.3
> > > > > > > > 
> > > > > > > > The hosts each have 512 GB memory, and the vms are
> > > > > > > > sized
> > > > > > > > with
> > > > > > > > 4-8
> > > > > > > > GB
> > > > > > > > each.
> > > > > > > > 
> > > > > > > > I have noticed that over the last months, the memory
> > > > > > > > usage
> > > > > > > > in
> > > > > > > > the
> > > > > > > > dashboard has been increasing and is now showing 946.8
> > > > > > > > GB
> > > > > > > > used
> > > > > > > > of
> > > > > > > > 1007.2 GB.
> > > > > > > > 
> > > > > > > > What can be causing this?
> > > > > > > > 
> > > > > > > > TIA,
> > > > > > > > 
> > > > > > > > -- 
> > > > > > > > -- 
> > > > > > > > Tony Albers
> > > > > > > > Systems Architect
> > > > > > > > Systems Director, National Cultural Heritage Cluster
> > > > > > > > Royal Danish Library, Victor Albecks Vej 1, 8000 Aarhus
> > > > > > > > C,
> > > > > > > > Denmark.
> > > > > > > > Tel: +45 2566 2383 / +45 8946 2316
> > > > > > > > _______________________________________________
> > > > > > > > Users mailing list -- users@ovirt.org
> > > > > > > > To unsubscribe send an email to users-le...@ovirt.org
> > > > > > > > Privacy Statement: https://www.ovirt.org/site/privacy-p
> > > > > > > > olic
> > > > > > > > y/
> > > > > > > > oVirt Code of Conduct: https://www.ovirt.org/community/
> > > > > > > > abou
> > > > > > > > t/co
> > > > > > > > mmun
> > > > > > > > ity-guidelines/
> > > > > > > > List Archives: https://lists.ovirt.org/archives/list/us
> > > > > > > > ers@
> > > > > > > > ovir
> > > > > > > > t.or
> > > > > > > > g/message/SDDH2OC5RBOVYYCLGPOUF6HO676HWI5U/
> > > > > > 
> > > > > > -- 
> > > > > > Tony Albers
> > > > > > Systems Architect
> > > > > > Systems Director, National Cultural Heritage Cluster
> > > > > > Royal Danish Library, Victor Albecks Vej 1, 8000 Aarhus C,
> > > > > > Denmark.
> > > > > > Tel: +45 2566 2383 <tel:+45%202566%202383> / +45 8946 2316
> > > > > > <tel:+45%208946%202316>
> > > > 
> > > > -- 
> > > > -- 
> > > > Tony Albers
> > > > Systems Architect
> > > > Systems Director, National Cultural Heritage Cluster
> > > > Royal Danish Library, Victor Albecks Vej 1, 8000 Aarhus C,
> > > > Denmark.
> > > > Tel: +45 2566 2383 / +45 8946 2316
> > > 
> > > 
> > 
> > -- 
> > Tony Albers
> > Systems Architect
> > Systems Director, National Cultural Heritage Cluster
> > Royal Danish Library, Victor Albecks Vej 1, 8000 Aarhus C, Denmark.
> > Tel: +45 2566 2383 <tel:+45%202566%202383> / +45 8946 2316
> > <tel:+45%208946%202316><2018-12-13-070032_349x526_scrot.png><2018-
> > 12-13-070259_504x81_scrot.png>
> 
> 

-- 
-- 
Tony Albers
Systems Architect
Systems Director, National Cultural Heritage Cluster
Royal Danish Library, Victor Albecks Vej 1, 8000 Aarhus C, Denmark.
Tel: +45 2566 2383 / +45 8946 2316
_______________________________________________
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/33K5S6PH4PRKWMLIDOJSHQKON2RL4ILI/

Reply via email to