[ 
https://issues.apache.org/jira/browse/IGNITE-9305?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16598940#comment-16598940
 ] 

Pavel Pereslegin commented on IGNITE-9305:
------------------------------------------

Hello [~dmagda].

I changed format as you suggested, here an example of output:
{noformat}
    ^-- H/N/C [hosts=1, nodes=2, CPUs=8]
    ^-- CPU [cur=7.8%, avg=11.65%, GC=0%]
    ^-- PageMemory [pages=7447]
    ^-- Heap [used=114MB, free=96.78%, comm=242MB]
    ^-- Off-heap [used=29MB, free=87.2%, comm=230MB]
    ^--   sysMemPlc region [used=0MB, free=99.98%, comm=100MB]
    ^--   default region [used=29MB, free=2.1%, comm=30MB]
    ^--   metastoreMemPlc region [used=0MB, free=99.96%, comm=100MB]
    ^-- Ignite persistence [used=35MB]
    ^--   sysMemPlc region [used=0MB]
    ^--   default region [used=35MB]
    ^--   metastoreMemPlc region [used=0MB]
    ^-- Outbound messages queue [size=0]
    ^-- Public thread pool [active=0, idle=6, qSize=0]
    ^-- System thread pool [active=0, idle=6, qSize=0]
    ^-- Custom executor 0 [active=0, idle=0, qSize=0]
    ^-- Custom executor 1 [active=0, idle=0, qSize=0]
{noformat}

But there is a problem with tracking allocated pages (disk usage) for 
"metastoreMemPlc" persistence region (metastore).
Total allocated pages metric is always zero, because this region is recreated 
after FilePageStore was created.

I see the following options:
# Output actual value of this metric (don't change anything here) and create 
separate ticket for the problem.
# Exclude "metastoreMemPlc" region from persistence regions in log output.
# Output only those persistence regions for which metrics are enabled.

Any thoughts?

> Wrong off-heap size is reported for a node
> ------------------------------------------
>
>                 Key: IGNITE-9305
>                 URL: https://issues.apache.org/jira/browse/IGNITE-9305
>             Project: Ignite
>          Issue Type: Task
>    Affects Versions: 2.6
>            Reporter: Denis Magda
>            Assignee: Pavel Pereslegin
>            Priority: Blocker
>             Fix For: 2.7
>
>
> Was troubleshooting an Ignite deployment today and couldn't find out from the 
> logs what was the actual off-heap space used. 
> Those were the given memory resoures (Ignite 2.6):
> {code}
> [2018-08-16 15:07:49,961][INFO ][main][GridDiscoveryManager] Topology 
> snapshot [ver=1, servers=1, clients=0, CPUs=64, offheap=30.0GB, heap=24.0GB]
> {code}
> And that weird stuff was reported by the node (pay attention to the last 
> line):
> {code}
> [2018-08-16 15:45:50,211][INFO 
> ][grid-timeout-worker-#135%cluster_31-Dec-2017%][IgniteKernal%cluster_31-Dec-2017]
>  
> Metrics for local node (to disable set 'metricsLogFrequency' to 0)
>     ^-- Node [id=c033026e, name=cluster_31-Dec-2017, uptime=00:38:00.257]
>     ^-- H/N/C [hosts=1, nodes=1, CPUs=64]
>     ^-- CPU [cur=0.03%, avg=5.54%, GC=0%]
>     ^-- PageMemory [pages=6997377]
>     ^-- Heap [used=9706MB, free=61.18%, comm=22384MB]
>     ^-- Non heap [used=144MB, free=-1%, comm=148MB] - this line is always the 
> same!
> {code}
> Had to change the code by using 
> {code}dataRegion.getPhysicalMemoryPages(){code} to find out that actual 
> off-heap usage size was 
> {code}
> >>> Physical Memory Size: 28651614208 => 27324 MB, 26 GB
> {code}
> The logs have to report the following instead:
> {code}
>  ^-- Off-heap {Data Region 1} [used={dataRegion1.getPhysicalMemorySize()}, 
> free=X%, comm=dataRegion1.maxSize()]
>  ^-- Off-heap {Data Region 2} [used={dataRegion2.getPhysicalMemorySize()}, 
> free=X%, comm=dataRegion2.maxSize()]
> {code}
> If Ignite persistence is enabled then the following extra lines have to be 
> added to see the disk used space:
> {code}
>  ^-- Ignite persistence {Data Region 1}: 
> used={dataRegion1.getTotalAllocatedSize() - 
> dataRegion1.getPhysicalMemorySize()}
>  ^-- Ignite persistence {Data Region 2} 
> [used={dataRegion2.getTotalAllocatedSize() - 
> dataRegion2.getPhysicalMemorySize()}]
> {code}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

Reply via email to