On Sat, Apr 27, 2019 at 03:09:15PM -0300, Mauro Carvalho Chehab wrote: > Em Fri, 26 Apr 2019 23:31:40 +0800 > Changbin Du <changbin...@gmail.com> escreveu: > > > This converts the plain text documentation to reStructuredText format and > > add it to Sphinx TOC tree. No essential content change. > > > > Signed-off-by: Changbin Du <changbin...@gmail.com> > > --- > > Documentation/x86/index.rst | 1 + > > .../x86/{resctrl_ui.txt => resctrl_ui.rst} | 913 ++++++++++-------- > > 2 files changed, 490 insertions(+), 424 deletions(-) > > rename Documentation/x86/{resctrl_ui.txt => resctrl_ui.rst} (68%) > > > > diff --git a/Documentation/x86/index.rst b/Documentation/x86/index.rst > > index 2fcd10f29b87..4e9fa2b046df 100644 > > --- a/Documentation/x86/index.rst > > +++ b/Documentation/x86/index.rst > > @@ -23,3 +23,4 @@ Linux x86 Support > > amd-memory-encryption > > pti > > microcode > > + resctrl_ui > > diff --git a/Documentation/x86/resctrl_ui.txt > > b/Documentation/x86/resctrl_ui.rst > > similarity index 68% > > rename from Documentation/x86/resctrl_ui.txt > > rename to Documentation/x86/resctrl_ui.rst > > index c1f95b59e14d..81aaa271d5ea 100644 > > --- a/Documentation/x86/resctrl_ui.txt > > +++ b/Documentation/x86/resctrl_ui.rst > > @@ -1,33 +1,39 @@ > > +.. SPDX-License-Identifier: GPL-2.0 > > +.. include:: <isonum.txt> > > + > > +=========================================== > > User Interface for Resource Control feature > > +=========================================== > > > > -Intel refers to this feature as Intel Resource Director > > Technology(Intel(R) RDT). > > -AMD refers to this feature as AMD Platform Quality of Service(AMD QoS). > > +:Copyright: |copy| 2016 Intel Corporation > > +:Authors: - Fenghua Yu <fenghua...@intel.com> > > + - Tony Luck <tony.l...@intel.com> > > + - Vikas Shivappa <vikas.shiva...@intel.com> > > > > -Copyright (C) 2016 Intel Corporation > > > > -Fenghua Yu <fenghua...@intel.com> > > -Tony Luck <tony.l...@intel.com> > > -Vikas Shivappa <vikas.shiva...@intel.com> > > +Intel refers to this feature as Intel Resource Director > > Technology(Intel(R) RDT). > > +AMD refers to this feature as AMD Platform Quality of Service(AMD QoS). > > > > This feature is enabled by the CONFIG_X86_CPU_RESCTRL and the x86 > > /proc/cpuinfo > > -flag bits: > > -RDT (Resource Director Technology) Allocation - "rdt_a" > > -CAT (Cache Allocation Technology) - "cat_l3", "cat_l2" > > -CDP (Code and Data Prioritization ) - "cdp_l3", "cdp_l2" > > -CQM (Cache QoS Monitoring) - "cqm_llc", "cqm_occup_llc" > > -MBM (Memory Bandwidth Monitoring) - "cqm_mbm_total", "cqm_mbm_local" > > -MBA (Memory Bandwidth Allocation) - "mba" > > +flag bits:: > > + > > + RDT (Resource Director Technology) Allocation - "rdt_a" > > + CAT (Cache Allocation Technology) - "cat_l3", "cat_l2" > > + CDP (Code and Data Prioritization ) - "cdp_l3", "cdp_l2" > > + CQM (Cache QoS Monitoring) - "cqm_llc", "cqm_occup_llc" > > + MBM (Memory Bandwidth Monitoring) - "cqm_mbm_total", "cqm_mbm_local" > > + MBA (Memory Bandwidth Allocation) - "mba" > > I don't see any reason to convert this into a literal block. I > would either convert it into a table or into something like: > > RDT > (Resource Director Technology) Allocation - "rdt_a" > CAT > (Cache Allocation Technology) - "cat_l3", "cat_l2" > CDP > (Code and Data Prioritization ) - "cdp_l3", "cdp_l2" > CQM > (Cache QoS Monitoring) - "cqm_llc", "cqm_occup_llc" > MBM > (Memory Bandwidth Monitoring) - "cqm_mbm_total", "cqm_mbm_local" > MBA > (Memory Bandwidth Allocation) - "mba" > > > A table seems to be the best approach, though: > > ============================================= > ================================ > RDT (Resource Director Technology) Allocation "rdt_a" > CAT (Cache Allocation Technology) "cat_l3", "cat_l2" > CDP (Code and Data Prioritization) "cdp_l3", "cdp_l2" > CQM (Cache QoS Monitoring) "cqm_llc", > "cqm_occup_llc" > MBM (Memory Bandwidth Monitoring) "cqm_mbm_total", > "cqm_mbm_local" > MBA (Memory Bandwidth Allocation) "mba" > ============================================= > ================================ > I prefer table, too. Thanks.
> > > > -To use the feature mount the file system: > > +To use the feature mount the file system:: > > > > # mount -t resctrl resctrl [-o cdp[,cdpl2][,mba_MBps]] /sys/fs/resctrl > > > > mount options are: > > > > -"cdp": Enable code/data prioritization in L3 cache allocations. > > -"cdpl2": Enable code/data prioritization in L2 cache allocations. > > -"mba_MBps": Enable the MBA Software Controller(mba_sc) to specify MBA > > - bandwidth in MBps > > +* "cdp": Enable code/data prioritization in L3 cache allocations. > > +* "cdpl2": Enable code/data prioritization in L2 cache allocations. > > +* "mba_MBps": Enable the MBA Software Controller(mba_sc) to specify MBA > > + bandwidth in MBps > > I would add a \n\t after each :, in order to show the options in > bold and the explanation on a separate indented line, just like what > you did with the next similar set of options. > Done, and removed the '*'. > > > > L2 and L3 CDP are controlled seperately. > > > > @@ -44,7 +50,7 @@ For more details on the behavior of the interface during > > monitoring > > and allocation, see the "Resource alloc and monitor groups" section. > > > > Info directory > > --------------- > > +============== > > > > The 'info' directory contains information about the enabled > > resources. Each resource has its own subdirectory. The subdirectory > > @@ -56,77 +62,93 @@ allocation: > > Cache resource(L3/L2) subdirectory contains the following files > > related to allocation: > > > > -"num_closids": The number of CLOSIDs which are valid for this > > - resource. The kernel uses the smallest number of > > - CLOSIDs of all enabled resources as limit. > > - > > -"cbm_mask": The bitmask which is valid for this resource. > > - This mask is equivalent to 100%. > > - > > -"min_cbm_bits": The minimum number of consecutive bits which > > - must be set when writing a mask. > > - > > -"shareable_bits": Bitmask of shareable resource with other executing > > - entities (e.g. I/O). User can use this when > > - setting up exclusive cache partitions. Note that > > - some platforms support devices that have their > > - own settings for cache use which can over-ride > > - these bits. > > -"bit_usage": Annotated capacity bitmasks showing how all > > - instances of the resource are used. The legend is: > > - "0" - Corresponding region is unused. When the system's > > +"num_closids": > > + The number of CLOSIDs which are valid for this > > + resource. The kernel uses the smallest number of > > + CLOSIDs of all enabled resources as limit. > > +"cbm_mask": > > + The bitmask which is valid for this resource. > > + This mask is equivalent to 100%. > > +"min_cbm_bits": > > + The minimum number of consecutive bits which > > + must be set when writing a mask. > > + > > +"shareable_bits": > > + Bitmask of shareable resource with other executing > > + entities (e.g. I/O). User can use this when > > + setting up exclusive cache partitions. Note that > > + some platforms support devices that have their > > + own settings for cache use which can over-ride > > + these bits. > > +"bit_usage": > > + Annotated capacity bitmasks showing how all > > + instances of the resource are used. The legend is: > > + > > + "0": > > + Corresponding region is unused. When the system's > > resources have been allocated and a "0" is found > > in "bit_usage" it is a sign that resources are > > wasted. > > - "H" - Corresponding region is used by hardware only > > + > > + "H": > > + Corresponding region is used by hardware only > > but available for software use. If a resource > > has bits set in "shareable_bits" but not all > > of these bits appear in the resource groups' > > schematas then the bits appearing in > > "shareable_bits" but no resource group will > > be marked as "H". > > - "X" - Corresponding region is available for sharing and > > + "X": > > + Corresponding region is available for sharing and > > used by hardware and software. These are the > > bits that appear in "shareable_bits" as > > well as a resource group's allocation. > > - "S" - Corresponding region is used by software > > + "S": > > + Corresponding region is used by software > > and available for sharing. > > - "E" - Corresponding region is used exclusively by > > + "E": > > + Corresponding region is used exclusively by > > one resource group. No sharing allowed. > > - "P" - Corresponding region is pseudo-locked. No > > + "P": > > + Corresponding region is pseudo-locked. No > > sharing allowed. > > > > Memory bandwitdh(MB) subdirectory contains the following files > > with respect to allocation: > > > > -"min_bandwidth": The minimum memory bandwidth percentage which > > - user can request. > > +"min_bandwidth": > > + The minimum memory bandwidth percentage which > > + user can request. > > > > -"bandwidth_gran": The granularity in which the memory bandwidth > > - percentage is allocated. The allocated > > - b/w percentage is rounded off to the next > > - control step available on the hardware. The > > - available bandwidth control steps are: > > - min_bandwidth + N * bandwidth_gran. > > +"bandwidth_gran": > > + The granularity in which the memory bandwidth > > + percentage is allocated. The allocated > > + b/w percentage is rounded off to the next > > + control step available on the hardware. The > > + available bandwidth control steps are: > > + min_bandwidth + N * bandwidth_gran. > > > > -"delay_linear": Indicates if the delay scale is linear or > > - non-linear. This field is purely informational > > - only. > > +"delay_linear": > > + Indicates if the delay scale is linear or > > + non-linear. This field is purely informational > > + only. > > > > If RDT monitoring is available there will be an "L3_MON" directory > > with the following files: > > > > -"num_rmids": The number of RMIDs available. This is the > > - upper bound for how many "CTRL_MON" + "MON" > > - groups can be created. > > +"num_rmids": > > + The number of RMIDs available. This is the > > + upper bound for how many "CTRL_MON" + "MON" > > + groups can be created. > > > > -"mon_features": Lists the monitoring events if > > - monitoring is enabled for the resource. > > +"mon_features": > > + Lists the monitoring events if > > + monitoring is enabled for the resource. > > > > "max_threshold_occupancy": > > - Read/write file provides the largest value (in > > - bytes) at which a previously used LLC_occupancy > > - counter can be considered for re-use. > > + Read/write file provides the largest value (in > > + bytes) at which a previously used LLC_occupancy > > + counter can be considered for re-use. > > > > Finally, in the top level of the "info" directory there is a file > > named "last_cmd_status". This is reset with every "command" issued > > @@ -134,6 +156,7 @@ via the file system (making new directories or writing > > to any of the > > control files). If the command was successful, it will read as "ok". > > If the command failed, it will provide more information that can be > > conveyed in the error returns from file operations. E.g. > > +:: > > > > # echo L3:0=f7 > schemata > > bash: echo: write error: Invalid argument > > @@ -141,7 +164,7 @@ conveyed in the error returns from file operations. E.g. > > mask f7 has non-consecutive 1-bits > > > > Resource alloc and monitor groups > > ---------------------------------- > > +================================= > > > > Resource groups are represented as directories in the resctrl file > > system. The default group is the root directory which, immediately > > @@ -226,6 +249,7 @@ When monitoring is enabled all MON groups will also > > contain: > > > > Resource allocation rules > > ------------------------- > > + > > When a task is running the following rules define which resources are > > available to it: > > > > @@ -252,7 +276,7 @@ Resource monitoring rules > > > > > > Notes on cache occupancy monitoring and control > > ------------------------------------------------ > > +=============================================== > > When moving a task from one group to another you should remember that > > this only affects *new* cache allocations by the task. E.g. you may have > > a task in a monitor group showing 3 MB of cache occupancy. If you move > > @@ -321,7 +345,7 @@ of the capacity of the cache. You could partition the > > cache into four > > equal parts with masks: 0x1f, 0x3e0, 0x7c00, 0xf8000. > > > > Memory bandwidth Allocation and monitoring > > ------------------------------------------- > > +========================================== > > > > For Memory bandwidth resource, by default the user controls the resource > > by indicating the percentage of total memory bandwidth. > > @@ -369,7 +393,7 @@ In order to mitigate this and make the interface more > > user friendly, > > resctrl added support for specifying the bandwidth in MBps as well. The > > kernel underneath would use a software feedback mechanism or a "Software > > Controller(mba_sc)" which reads the actual bandwidth using MBM counters > > -and adjust the memowy bandwidth percentages to ensure > > +and adjust the memowy bandwidth percentages to ensure:: > > > > "actual bandwidth < user specified bandwidth". > > > > @@ -380,14 +404,14 @@ sections. > > > > L3 schemata file details (code and data prioritization disabled) > > ---------------------------------------------------------------- > > -With CDP disabled the L3 schemata format is: > > +With CDP disabled the L3 schemata format is:: > > > > L3:<cache_id0>=<cbm>;<cache_id1>=<cbm>;... > > > > L3 schemata file details (CDP enabled via mount option to resctrl) > > ------------------------------------------------------------------ > > When CDP is enabled L3 control is split into two separate resources > > -so you can specify independent masks for code and data like this: > > +so you can specify independent masks for code and data like this:: > > > > L3data:<cache_id0>=<cbm>;<cache_id1>=<cbm>;... > > L3code:<cache_id0>=<cbm>;<cache_id1>=<cbm>;... > > @@ -395,7 +419,7 @@ so you can specify independent masks for code and data > > like this: > > L2 schemata file details > > ------------------------ > > L2 cache does not support code and data prioritization, so the > > -schemata format is always: > > +schemata format is always:: > > > > L2:<cache_id0>=<cbm>;<cache_id1>=<cbm>;... > > > > @@ -403,6 +427,7 @@ Memory bandwidth Allocation (default mode) > > ------------------------------------------ > > > > Memory b/w domain is L3 cache. > > +:: > > > > MB:<cache_id0>=bandwidth0;<cache_id1>=bandwidth1;... > > > > @@ -410,6 +435,7 @@ Memory bandwidth Allocation specified in MBps > > --------------------------------------------- > > > > Memory bandwidth domain is L3 cache. > > +:: > > > > MB:<cache_id0>=bw_MBps0;<cache_id1>=bw_MBps1;... > > > > @@ -418,17 +444,18 @@ Reading/writing the schemata file > > Reading the schemata file will show the state of all resources > > on all domains. When writing you only need to specify those values > > which you wish to change. E.g. > > +:: > > > > -# cat schemata > > -L3DATA:0=fffff;1=fffff;2=fffff;3=fffff > > -L3CODE:0=fffff;1=fffff;2=fffff;3=fffff > > -# echo "L3DATA:2=3c0;" > schemata > > -# cat schemata > > -L3DATA:0=fffff;1=fffff;2=3c0;3=fffff > > -L3CODE:0=fffff;1=fffff;2=fffff;3=fffff > > + # cat schemata > > + L3DATA:0=fffff;1=fffff;2=fffff;3=fffff > > + L3CODE:0=fffff;1=fffff;2=fffff;3=fffff > > + # echo "L3DATA:2=3c0;" > schemata > > + # cat schemata > > + L3DATA:0=fffff;1=fffff;2=3c0;3=fffff > > + L3CODE:0=fffff;1=fffff;2=fffff;3=fffff > > > > Cache Pseudo-Locking > > --------------------- > > +==================== > > CAT enables a user to specify the amount of cache space that an > > application can fill. Cache pseudo-locking builds on the fact that a > > CPU can still read and write data pre-allocated outside its current > > @@ -442,6 +469,7 @@ a region of memory with reduced average read latency. > > The creation of a cache pseudo-locked region is triggered by a request > > from the user to do so that is accompanied by a schemata of the region > > to be pseudo-locked. The cache pseudo-locked region is created as follows: > > + > > - Create a CAT allocation CLOSNEW with a CBM matching the schemata > > from the user of the cache region that will contain the pseudo-locked > > memory. This region must not overlap with any current CAT allocation/CLOS > > @@ -480,6 +508,7 @@ initial mmap() handling, there is no enforcement > > afterwards and the > > application self needs to ensure it remains affine to the correct cores. > > > > Pseudo-locking is accomplished in two stages: > > + > > 1) During the first stage the system administrator allocates a portion > > of cache that should be dedicated to pseudo-locking. At this time an > > equivalent portion of memory is allocated, loaded into allocated > > @@ -506,7 +535,7 @@ by user space in order to obtain access to the > > pseudo-locked memory region. > > An example of cache pseudo-locked region creation and usage can be found > > below. > > > > Cache Pseudo-Locking Debugging Interface > > ---------------------------------------- > > +---------------------------------------- > > The pseudo-locking debugging interface is enabled by default (if > > CONFIG_DEBUG_FS is enabled) and can be found in /sys/kernel/debug/resctrl. > > > > @@ -514,6 +543,7 @@ There is no explicit way for the kernel to test if a > > provided memory > > location is present in the cache. The pseudo-locking debugging interface > > uses > > the tracing infrastructure to provide two ways to measure cache residency > > of > > the pseudo-locked region: > > + > > 1) Memory access latency using the pseudo_lock_mem_latency tracepoint. Data > > from these measurements are best visualized using a hist trigger (see > > example below). In this test the pseudo-locked region is traversed at > > @@ -529,87 +559,97 @@ it in debugfs as /sys/kernel/debug/resctrl/<newdir>. > > A single > > write-only file, pseudo_lock_measure, is present in this directory. The > > measurement of the pseudo-locked region depends on the number written to > > this > > debugfs file: > > -1 - writing "1" to the pseudo_lock_measure file will trigger the latency > > + > > +1: > > + writing "1" to the pseudo_lock_measure file will trigger the latency > > measurement captured in the pseudo_lock_mem_latency tracepoint. See > > example below. > > -2 - writing "2" to the pseudo_lock_measure file will trigger the L2 cache > > +2: > > + writing "2" to the pseudo_lock_measure file will trigger the L2 cache > > residency (cache hits and misses) measurement captured in the > > pseudo_lock_l2 tracepoint. See example below. > > -3 - writing "3" to the pseudo_lock_measure file will trigger the L3 cache > > +3: > > + writing "3" to the pseudo_lock_measure file will trigger the L3 cache > > residency (cache hits and misses) measurement captured in the > > pseudo_lock_l3 tracepoint. > > > > All measurements are recorded with the tracing infrastructure. This > > requires > > the relevant tracepoints to be enabled before the measurement is triggered. > > > > -Example of latency debugging interface: > > +Example of latency debugging interface > > +~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ > > In this example a pseudo-locked region named "newlock" was created. Here is > > how we can measure the latency in cycles of reading from this region and > > visualize this data with a histogram that is available if > > CONFIG_HIST_TRIGGERS > > -is set: > > -# :> /sys/kernel/debug/tracing/trace > > -# echo 'hist:keys=latency' > > > /sys/kernel/debug/tracing/events/resctrl/pseudo_lock_mem_latency/trigger > > -# echo 1 > > > /sys/kernel/debug/tracing/events/resctrl/pseudo_lock_mem_latency/enable > > -# echo 1 > /sys/kernel/debug/resctrl/newlock/pseudo_lock_measure > > -# echo 0 > > > /sys/kernel/debug/tracing/events/resctrl/pseudo_lock_mem_latency/enable > > -# cat /sys/kernel/debug/tracing/events/resctrl/pseudo_lock_mem_latency/hist > > - > > -# event histogram > > -# > > -# trigger info: hist:keys=latency:vals=hitcount:sort=hitcount:size=2048 > > [active] > > -# > > - > > -{ latency: 456 } hitcount: 1 > > -{ latency: 50 } hitcount: 83 > > -{ latency: 36 } hitcount: 96 > > -{ latency: 44 } hitcount: 174 > > -{ latency: 48 } hitcount: 195 > > -{ latency: 46 } hitcount: 262 > > -{ latency: 42 } hitcount: 693 > > -{ latency: 40 } hitcount: 3204 > > -{ latency: 38 } hitcount: 3484 > > - > > -Totals: > > - Hits: 8192 > > - Entries: 9 > > - Dropped: 0 > > - > > -Example of cache hits/misses debugging: > > +is set:: > > + > > + # :> /sys/kernel/debug/tracing/trace > > + # echo 'hist:keys=latency' > > > /sys/kernel/debug/tracing/events/resctrl/pseudo_lock_mem_latency/trigger > > + # echo 1 > > > /sys/kernel/debug/tracing/events/resctrl/pseudo_lock_mem_latency/enable > > + # echo 1 > /sys/kernel/debug/resctrl/newlock/pseudo_lock_measure > > + # echo 0 > > > /sys/kernel/debug/tracing/events/resctrl/pseudo_lock_mem_latency/enable > > + # cat > > /sys/kernel/debug/tracing/events/resctrl/pseudo_lock_mem_latency/hist > > + > > + # event histogram > > + # > > + # trigger info: hist:keys=latency:vals=hitcount:sort=hitcount:size=2048 > > [active] > > + # > > + > > + { latency: 456 } hitcount: 1 > > + { latency: 50 } hitcount: 83 > > + { latency: 36 } hitcount: 96 > > + { latency: 44 } hitcount: 174 > > + { latency: 48 } hitcount: 195 > > + { latency: 46 } hitcount: 262 > > + { latency: 42 } hitcount: 693 > > + { latency: 40 } hitcount: 3204 > > + { latency: 38 } hitcount: 3484 > > + > > + Totals: > > + Hits: 8192 > > + Entries: 9 > > + Dropped: 0 > > + > > +Example of cache hits/misses debugging > > +~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ > > In this example a pseudo-locked region named "newlock" was created on the > > L2 > > cache of a platform. Here is how we can obtain details of the cache hits > > and misses using the platform's precision counters. > > +:: > > > > -# :> /sys/kernel/debug/tracing/trace > > -# echo 1 > /sys/kernel/debug/tracing/events/resctrl/pseudo_lock_l2/enable > > -# echo 2 > /sys/kernel/debug/resctrl/newlock/pseudo_lock_measure > > -# echo 0 > /sys/kernel/debug/tracing/events/resctrl/pseudo_lock_l2/enable > > -# cat /sys/kernel/debug/tracing/trace > > + # :> /sys/kernel/debug/tracing/trace > > + # echo 1 > /sys/kernel/debug/tracing/events/resctrl/pseudo_lock_l2/enable > > + # echo 2 > /sys/kernel/debug/resctrl/newlock/pseudo_lock_measure > > + # echo 0 > /sys/kernel/debug/tracing/events/resctrl/pseudo_lock_l2/enable > > + # cat /sys/kernel/debug/tracing/trace > > > > -# tracer: nop > > -# > > -# _-----=> irqs-off > > -# / _----=> need-resched > > -# | / _---=> hardirq/softirq > > -# || / _--=> preempt-depth > > -# ||| / delay > > -# TASK-PID CPU# |||| TIMESTAMP FUNCTION > > -# | | | |||| | | > > - pseudo_lock_mea-1672 [002] .... 3132.860500: pseudo_lock_l2: hits=4097 > > miss=0 > > + # tracer: nop > > + # > > + # _-----=> irqs-off > > + # / _----=> need-resched > > + # | / _---=> hardirq/softirq > > + # || / _--=> preempt-depth > > + # ||| / delay > > + # TASK-PID CPU# |||| TIMESTAMP FUNCTION > > + # | | | |||| | | > > + pseudo_lock_mea-1672 [002] .... 3132.860500: pseudo_lock_l2: hits=4097 > > miss=0 > > > > > > -Examples for RDT allocation usage: > > +Examples for RDT allocation usage > > +~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ > > + > > +1) Example 1 > > > > -Example 1 > > ---------- > > On a two socket machine (one L3 cache per socket) with just four bits > > for cache bit masks, minimum b/w of 10% with a memory bandwidth > > -granularity of 10% > > +granularity of 10%. > > +:: > > > > -# mount -t resctrl resctrl /sys/fs/resctrl > > -# cd /sys/fs/resctrl > > -# mkdir p0 p1 > > -# echo "L3:0=3;1=c\nMB:0=50;1=50" > /sys/fs/resctrl/p0/schemata > > -# echo "L3:0=3;1=3\nMB:0=50;1=50" > /sys/fs/resctrl/p1/schemata > > + # mount -t resctrl resctrl /sys/fs/resctrl > > + # cd /sys/fs/resctrl > > + # mkdir p0 p1 > > + # echo "L3:0=3;1=c\nMB:0=50;1=50" > /sys/fs/resctrl/p0/schemata > > + # echo "L3:0=3;1=3\nMB:0=50;1=50" > /sys/fs/resctrl/p1/schemata > > > > The default resource group is unmodified, so we have access to all parts > > of all caches (its schemata file reads "L3:0=f;1=f"). > > @@ -628,100 +668,106 @@ the b/w accordingly. > > > > If the MBA is specified in MB(megabytes) then user can enter the max b/w > > in MB > > rather than the percentage values. > > +:: > > > > -# echo "L3:0=3;1=c\nMB:0=1024;1=500" > /sys/fs/resctrl/p0/schemata > > -# echo "L3:0=3;1=3\nMB:0=1024;1=500" > /sys/fs/resctrl/p1/schemata > > + # echo "L3:0=3;1=c\nMB:0=1024;1=500" > /sys/fs/resctrl/p0/schemata > > + # echo "L3:0=3;1=3\nMB:0=1024;1=500" > /sys/fs/resctrl/p1/schemata > > > > In the above example the tasks in "p1" and "p0" on socket 0 would use a > > max b/w > > of 1024MB where as on socket 1 they would use 500MB. > > > > -Example 2 > > ---------- > > +2) Example 2 > > + > > Again two sockets, but this time with a more realistic 20-bit mask. > > > > Two real time tasks pid=1234 running on processor 0 and pid=5678 running on > > processor 1 on socket 0 on a 2-socket and dual core machine. To avoid noisy > > neighbors, each of the two real-time tasks exclusively occupies one quarter > > of L3 cache on socket 0. > > +:: > > > > -# mount -t resctrl resctrl /sys/fs/resctrl > > -# cd /sys/fs/resctrl > > + # mount -t resctrl resctrl /sys/fs/resctrl > > + # cd /sys/fs/resctrl > > > > First we reset the schemata for the default group so that the "upper" > > 50% of the L3 cache on socket 0 and 50% of memory b/w cannot be used by > > -ordinary tasks: > > +ordinary tasks:: > > > > -# echo "L3:0=3ff;1=fffff\nMB:0=50;1=100" > schemata > > + # echo "L3:0=3ff;1=fffff\nMB:0=50;1=100" > schemata > > > > Next we make a resource group for our first real time task and give > > it access to the "top" 25% of the cache on socket 0. > > +:: > > > > -# mkdir p0 > > -# echo "L3:0=f8000;1=fffff" > p0/schemata > > + # mkdir p0 > > + # echo "L3:0=f8000;1=fffff" > p0/schemata > > > > Finally we move our first real time task into this resource group. We > > also use taskset(1) to ensure the task always runs on a dedicated CPU > > on socket 0. Most uses of resource groups will also constrain which > > processors tasks run on. > > +:: > > > > -# echo 1234 > p0/tasks > > -# taskset -cp 1 1234 > > + # echo 1234 > p0/tasks > > + # taskset -cp 1 1234 > > > > -Ditto for the second real time task (with the remaining 25% of cache): > > +Ditto for the second real time task (with the remaining 25% of cache):: > > > > -# mkdir p1 > > -# echo "L3:0=7c00;1=fffff" > p1/schemata > > -# echo 5678 > p1/tasks > > -# taskset -cp 2 5678 > > + # mkdir p1 > > + # echo "L3:0=7c00;1=fffff" > p1/schemata > > + # echo 5678 > p1/tasks > > + # taskset -cp 2 5678 > > > > For the same 2 socket system with memory b/w resource and CAT L3 the > > schemata would look like(Assume min_bandwidth 10 and bandwidth_gran is > > 10): > > > > -For our first real time task this would request 20% memory b/w on socket > > -0. > > +For our first real time task this would request 20% memory b/w on socket 0. > > +:: > > > > -# echo -e "L3:0=f8000;1=fffff\nMB:0=20;1=100" > p0/schemata > > + # echo -e "L3:0=f8000;1=fffff\nMB:0=20;1=100" > p0/schemata > > > > For our second real time task this would request an other 20% memory b/w > > on socket 0. > > +:: > > > > -# echo -e "L3:0=f8000;1=fffff\nMB:0=20;1=100" > p0/schemata > > + # echo -e "L3:0=f8000;1=fffff\nMB:0=20;1=100" > p0/schemata > > > > -Example 3 > > ---------- > > +3) Example 3 > > > > A single socket system which has real-time tasks running on core 4-7 and > > non real-time workload assigned to core 0-3. The real-time tasks share text > > and data, so a per task association is not required and due to interaction > > with the kernel it's desired that the kernel on these cores shares L3 with > > the tasks. > > +:: > > > > -# mount -t resctrl resctrl /sys/fs/resctrl > > -# cd /sys/fs/resctrl > > + # mount -t resctrl resctrl /sys/fs/resctrl > > + # cd /sys/fs/resctrl > > > > First we reset the schemata for the default group so that the "upper" > > 50% of the L3 cache on socket 0, and 50% of memory bandwidth on socket 0 > > -cannot be used by ordinary tasks: > > +cannot be used by ordinary tasks:: > > > > -# echo "L3:0=3ff\nMB:0=50" > schemata > > + # echo "L3:0=3ff\nMB:0=50" > schemata > > > > Next we make a resource group for our real time cores and give it access > > to the "top" 50% of the cache on socket 0 and 50% of memory bandwidth on > > socket 0. > > +:: > > > > -# mkdir p0 > > -# echo "L3:0=ffc00\nMB:0=50" > p0/schemata > > + # mkdir p0 > > + # echo "L3:0=ffc00\nMB:0=50" > p0/schemata > > > > Finally we move core 4-7 over to the new group and make sure that the > > kernel and the tasks running there get 50% of the cache. They should > > also get 50% of memory bandwidth assuming that the cores 4-7 are SMT > > siblings and only the real time threads are scheduled on the cores 4-7. > > +:: > > > > -# echo F0 > p0/cpus > > + # echo F0 > p0/cpus > > > > -Example 4 > > ---------- > > +4) Example 4 > > > > The resource groups in previous examples were all in the default > > "shareable" > > mode allowing sharing of their cache allocations. If one resource group > > @@ -732,157 +778,168 @@ In this example a new exclusive resource group will > > be created on a L2 CAT > > system with two L2 cache instances that can be configured with an 8-bit > > capacity bitmask. The new exclusive resource group will be configured to > > use > > 25% of each cache instance. > > +:: > > > > -# mount -t resctrl resctrl /sys/fs/resctrl/ > > -# cd /sys/fs/resctrl > > + # mount -t resctrl resctrl /sys/fs/resctrl/ > > + # cd /sys/fs/resctrl > > > > First, we observe that the default group is configured to allocate to all > > L2 > > -cache: > > +cache:: > > > > -# cat schemata > > -L2:0=ff;1=ff > > + # cat schemata > > + L2:0=ff;1=ff > > > > We could attempt to create the new resource group at this point, but it > > will > > -fail because of the overlap with the schemata of the default group: > > -# mkdir p0 > > -# echo 'L2:0=0x3;1=0x3' > p0/schemata > > -# cat p0/mode > > -shareable > > -# echo exclusive > p0/mode > > --sh: echo: write error: Invalid argument > > -# cat info/last_cmd_status > > -schemata overlaps > > +fail because of the overlap with the schemata of the default group:: > > + > > + # mkdir p0 > > + # echo 'L2:0=0x3;1=0x3' > p0/schemata > > + # cat p0/mode > > + shareable > > + # echo exclusive > p0/mode > > + -sh: echo: write error: Invalid argument > > + # cat info/last_cmd_status > > + schemata overlaps > > > > To ensure that there is no overlap with another resource group the default > > resource group's schemata has to change, making it possible for the new > > resource group to become exclusive. > > -# echo 'L2:0=0xfc;1=0xfc' > schemata > > -# echo exclusive > p0/mode > > -# grep . p0/* > > -p0/cpus:0 > > -p0/mode:exclusive > > -p0/schemata:L2:0=03;1=03 > > -p0/size:L2:0=262144;1=262144 > > +:: > > + > > + # echo 'L2:0=0xfc;1=0xfc' > schemata > > + # echo exclusive > p0/mode > > + # grep . p0/* > > + p0/cpus:0 > > + p0/mode:exclusive > > + p0/schemata:L2:0=03;1=03 > > + p0/size:L2:0=262144;1=262144 > > > > A new resource group will on creation not overlap with an exclusive > > resource > > -group: > > -# mkdir p1 > > -# grep . p1/* > > -p1/cpus:0 > > -p1/mode:shareable > > -p1/schemata:L2:0=fc;1=fc > > -p1/size:L2:0=786432;1=786432 > > - > > -The bit_usage will reflect how the cache is used: > > -# cat info/L2/bit_usage > > -0=SSSSSSEE;1=SSSSSSEE > > - > > -A resource group cannot be forced to overlap with an exclusive resource > > group: > > -# echo 'L2:0=0x1;1=0x1' > p1/schemata > > --sh: echo: write error: Invalid argument > > -# cat info/last_cmd_status > > -overlaps with exclusive group > > +group:: > > + > > + # mkdir p1 > > + # grep . p1/* > > + p1/cpus:0 > > + p1/mode:shareable > > + p1/schemata:L2:0=fc;1=fc > > + p1/size:L2:0=786432;1=786432 > > + > > +The bit_usage will reflect how the cache is used:: > > + > > + # cat info/L2/bit_usage > > + 0=SSSSSSEE;1=SSSSSSEE > > + > > +A resource group cannot be forced to overlap with an exclusive resource > > group:: > > + > > + # echo 'L2:0=0x1;1=0x1' > p1/schemata > > + -sh: echo: write error: Invalid argument > > + # cat info/last_cmd_status > > + overlaps with exclusive group > > > > Example of Cache Pseudo-Locking > > -------------------------------- > > +~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ > > Lock portion of L2 cache from cache id 1 using CBM 0x3. Pseudo-locked > > region is exposed at /dev/pseudo_lock/newlock that can be provided to > > application for argument to mmap(). > > +:: > > > > -# mount -t resctrl resctrl /sys/fs/resctrl/ > > -# cd /sys/fs/resctrl > > + # mount -t resctrl resctrl /sys/fs/resctrl/ > > + # cd /sys/fs/resctrl > > > > Ensure that there are bits available that can be pseudo-locked, since only > > unused bits can be pseudo-locked the bits to be pseudo-locked needs to be > > -removed from the default resource group's schemata: > > -# cat info/L2/bit_usage > > -0=SSSSSSSS;1=SSSSSSSS > > -# echo 'L2:1=0xfc' > schemata > > -# cat info/L2/bit_usage > > -0=SSSSSSSS;1=SSSSSS00 > > +removed from the default resource group's schemata:: > > + > > + # cat info/L2/bit_usage > > + 0=SSSSSSSS;1=SSSSSSSS > > + # echo 'L2:1=0xfc' > schemata > > + # cat info/L2/bit_usage > > + 0=SSSSSSSS;1=SSSSSS00 > > > > Create a new resource group that will be associated with the pseudo-locked > > region, indicate that it will be used for a pseudo-locked region, and > > -configure the requested pseudo-locked region capacity bitmask: > > +configure the requested pseudo-locked region capacity bitmask:: > > > > -# mkdir newlock > > -# echo pseudo-locksetup > newlock/mode > > -# echo 'L2:1=0x3' > newlock/schemata > > + # mkdir newlock > > + # echo pseudo-locksetup > newlock/mode > > + # echo 'L2:1=0x3' > newlock/schemata > > > > On success the resource group's mode will change to pseudo-locked, the > > bit_usage will reflect the pseudo-locked region, and the character device > > -exposing the pseudo-locked region will exist: > > - > > -# cat newlock/mode > > -pseudo-locked > > -# cat info/L2/bit_usage > > -0=SSSSSSSS;1=SSSSSSPP > > -# ls -l /dev/pseudo_lock/newlock > > -crw------- 1 root root 243, 0 Apr 3 05:01 /dev/pseudo_lock/newlock > > - > > -/* > > - * Example code to access one page of pseudo-locked cache region > > - * from user space. > > - */ > > -#define _GNU_SOURCE > > -#include <fcntl.h> > > -#include <sched.h> > > -#include <stdio.h> > > -#include <stdlib.h> > > -#include <unistd.h> > > -#include <sys/mman.h> > > - > > -/* > > - * It is required that the application runs with affinity to only > > - * cores associated with the pseudo-locked region. Here the cpu > > - * is hardcoded for convenience of example. > > - */ > > -static int cpuid = 2; > > - > > -int main(int argc, char *argv[]) > > -{ > > - cpu_set_t cpuset; > > - long page_size; > > - void *mapping; > > - int dev_fd; > > - int ret; > > - > > - page_size = sysconf(_SC_PAGESIZE); > > - > > - CPU_ZERO(&cpuset); > > - CPU_SET(cpuid, &cpuset); > > - ret = sched_setaffinity(0, sizeof(cpuset), &cpuset); > > - if (ret < 0) { > > - perror("sched_setaffinity"); > > - exit(EXIT_FAILURE); > > - } > > - > > - dev_fd = open("/dev/pseudo_lock/newlock", O_RDWR); > > - if (dev_fd < 0) { > > - perror("open"); > > - exit(EXIT_FAILURE); > > - } > > - > > - mapping = mmap(0, page_size, PROT_READ | PROT_WRITE, MAP_SHARED, > > - dev_fd, 0); > > - if (mapping == MAP_FAILED) { > > - perror("mmap"); > > - close(dev_fd); > > - exit(EXIT_FAILURE); > > - } > > - > > - /* Application interacts with pseudo-locked memory @mapping */ > > - > > - ret = munmap(mapping, page_size); > > - if (ret < 0) { > > - perror("munmap"); > > - close(dev_fd); > > - exit(EXIT_FAILURE); > > - } > > - > > - close(dev_fd); > > - exit(EXIT_SUCCESS); > > -} > > +exposing the pseudo-locked region will exist:: > > + > > + # cat newlock/mode > > + pseudo-locked > > + # cat info/L2/bit_usage > > + 0=SSSSSSSS;1=SSSSSSPP > > + # ls -l /dev/pseudo_lock/newlock > > + crw------- 1 root root 243, 0 Apr 3 05:01 /dev/pseudo_lock/newlock > > + > > +:: > > + > > + /* > > + * Example code to access one page of pseudo-locked cache region > > + * from user space. > > + */ > > + #define _GNU_SOURCE > > + #include <fcntl.h> > > + #include <sched.h> > > + #include <stdio.h> > > + #include <stdlib.h> > > + #include <unistd.h> > > + #include <sys/mman.h> > > + > > + /* > > + * It is required that the application runs with affinity to only > > + * cores associated with the pseudo-locked region. Here the cpu > > + * is hardcoded for convenience of example. > > + */ > > + static int cpuid = 2; > > + > > + int main(int argc, char *argv[]) > > + { > > + cpu_set_t cpuset; > > + long page_size; > > + void *mapping; > > + int dev_fd; > > + int ret; > > + > > + page_size = sysconf(_SC_PAGESIZE); > > + > > + CPU_ZERO(&cpuset); > > + CPU_SET(cpuid, &cpuset); > > + ret = sched_setaffinity(0, sizeof(cpuset), &cpuset); > > + if (ret < 0) { > > + perror("sched_setaffinity"); > > + exit(EXIT_FAILURE); > > + } > > + > > + dev_fd = open("/dev/pseudo_lock/newlock", O_RDWR); > > + if (dev_fd < 0) { > > + perror("open"); > > + exit(EXIT_FAILURE); > > + } > > + > > + mapping = mmap(0, page_size, PROT_READ | PROT_WRITE, MAP_SHARED, > > + dev_fd, 0); > > + if (mapping == MAP_FAILED) { > > + perror("mmap"); > > + close(dev_fd); > > + exit(EXIT_FAILURE); > > + } > > + > > + /* Application interacts with pseudo-locked memory @mapping */ > > + > > + ret = munmap(mapping, page_size); > > + if (ret < 0) { > > + perror("munmap"); > > + close(dev_fd); > > + exit(EXIT_FAILURE); > > + } > > + > > + close(dev_fd); > > + exit(EXIT_SUCCESS); > > + } > > > > Locking between applications > > ---------------------------- > > @@ -921,86 +978,86 @@ Read lock: > > B) If success read the directory structure. > > C) funlock > > > > -Example with bash: > > - > > -# Atomically read directory structure > > -$ flock -s /sys/fs/resctrl/ find /sys/fs/resctrl > > - > > -# Read directory contents and create new subdirectory > > - > > -$ cat create-dir.sh > > -find /sys/fs/resctrl/ > output.txt > > -mask = function-of(output.txt) > > -mkdir /sys/fs/resctrl/newres/ > > -echo mask > /sys/fs/resctrl/newres/schemata > > - > > -$ flock /sys/fs/resctrl/ ./create-dir.sh > > - > > -Example with C: > > - > > -/* > > - * Example code do take advisory locks > > - * before accessing resctrl filesystem > > - */ > > -#include <sys/file.h> > > -#include <stdlib.h> > > - > > -void resctrl_take_shared_lock(int fd) > > -{ > > - int ret; > > - > > - /* take shared lock on resctrl filesystem */ > > - ret = flock(fd, LOCK_SH); > > - if (ret) { > > - perror("flock"); > > - exit(-1); > > - } > > -} > > - > > -void resctrl_take_exclusive_lock(int fd) > > -{ > > - int ret; > > - > > - /* release lock on resctrl filesystem */ > > - ret = flock(fd, LOCK_EX); > > - if (ret) { > > - perror("flock"); > > - exit(-1); > > - } > > -} > > - > > -void resctrl_release_lock(int fd) > > -{ > > - int ret; > > - > > - /* take shared lock on resctrl filesystem */ > > - ret = flock(fd, LOCK_UN); > > - if (ret) { > > - perror("flock"); > > - exit(-1); > > - } > > -} > > - > > -void main(void) > > -{ > > - int fd, ret; > > - > > - fd = open("/sys/fs/resctrl", O_DIRECTORY); > > - if (fd == -1) { > > - perror("open"); > > - exit(-1); > > - } > > - resctrl_take_shared_lock(fd); > > - /* code to read directory contents */ > > - resctrl_release_lock(fd); > > - > > - resctrl_take_exclusive_lock(fd); > > - /* code to read and write directory contents */ > > - resctrl_release_lock(fd); > > -} > > - > > -Examples for RDT Monitoring along with allocation usage: > > - > > +Example with bash:: > > + > > + # Atomically read directory structure > > + $ flock -s /sys/fs/resctrl/ find /sys/fs/resctrl > > + > > + # Read directory contents and create new subdirectory > > + > > + $ cat create-dir.sh > > + find /sys/fs/resctrl/ > output.txt > > + mask = function-of(output.txt) > > + mkdir /sys/fs/resctrl/newres/ > > + echo mask > /sys/fs/resctrl/newres/schemata > > + > > + $ flock /sys/fs/resctrl/ ./create-dir.sh > > + > > +Example with C:: > > + > > + /* > > + * Example code do take advisory locks > > + * before accessing resctrl filesystem > > + */ > > + #include <sys/file.h> > > + #include <stdlib.h> > > + > > + void resctrl_take_shared_lock(int fd) > > + { > > + int ret; > > + > > + /* take shared lock on resctrl filesystem */ > > + ret = flock(fd, LOCK_SH); > > + if (ret) { > > + perror("flock"); > > + exit(-1); > > + } > > + } > > + > > + void resctrl_take_exclusive_lock(int fd) > > + { > > + int ret; > > + > > + /* release lock on resctrl filesystem */ > > + ret = flock(fd, LOCK_EX); > > + if (ret) { > > + perror("flock"); > > + exit(-1); > > + } > > + } > > + > > + void resctrl_release_lock(int fd) > > + { > > + int ret; > > + > > + /* take shared lock on resctrl filesystem */ > > + ret = flock(fd, LOCK_UN); > > + if (ret) { > > + perror("flock"); > > + exit(-1); > > + } > > + } > > + > > + void main(void) > > + { > > + int fd, ret; > > + > > + fd = open("/sys/fs/resctrl", O_DIRECTORY); > > + if (fd == -1) { > > + perror("open"); > > + exit(-1); > > + } > > + resctrl_take_shared_lock(fd); > > + /* code to read directory contents */ > > + resctrl_release_lock(fd); > > + > > + resctrl_take_exclusive_lock(fd); > > + /* code to read and write directory contents */ > > + resctrl_release_lock(fd); > > + } > > + > > +Examples for RDT Monitoring along with allocation usage > > +======================================================= > > Reading monitored data > > ---------------------- > > Reading an event file (for ex: mon_data/mon_L3_00/llc_occupancy) would > > @@ -1009,17 +1066,17 @@ group or CTRL_MON group. > > > > > > Example 1 (Monitor CTRL_MON group and subset of tasks in CTRL_MON group) > > ---------- > > +------------------------------------------------------------------------ > > On a two socket machine (one L3 cache per socket) with just four bits > > -for cache bit masks > > +for cache bit masks:: > > > > -# mount -t resctrl resctrl /sys/fs/resctrl > > -# cd /sys/fs/resctrl > > -# mkdir p0 p1 > > -# echo "L3:0=3;1=c" > /sys/fs/resctrl/p0/schemata > > -# echo "L3:0=3;1=3" > /sys/fs/resctrl/p1/schemata > > -# echo 5678 > p1/tasks > > -# echo 5679 > p1/tasks > > + # mount -t resctrl resctrl /sys/fs/resctrl > > + # cd /sys/fs/resctrl > > + # mkdir p0 p1 > > + # echo "L3:0=3;1=c" > /sys/fs/resctrl/p0/schemata > > + # echo "L3:0=3;1=3" > /sys/fs/resctrl/p1/schemata > > + # echo 5678 > p1/tasks > > + # echo 5679 > p1/tasks > > > > The default resource group is unmodified, so we have access to all parts > > of all caches (its schemata file reads "L3:0=f;1=f"). > > @@ -1029,47 +1086,51 @@ Tasks that are under the control of group "p0" may > > only allocate from the > > Tasks in group "p1" use the "lower" 50% of cache on both sockets. > > > > Create monitor groups and assign a subset of tasks to each monitor group. > > +:: > > > > -# cd /sys/fs/resctrl/p1/mon_groups > > -# mkdir m11 m12 > > -# echo 5678 > m11/tasks > > -# echo 5679 > m12/tasks > > + # cd /sys/fs/resctrl/p1/mon_groups > > + # mkdir m11 m12 > > + # echo 5678 > m11/tasks > > + # echo 5679 > m12/tasks > > > > fetch data (data shown in bytes) > > +:: > > > > -# cat m11/mon_data/mon_L3_00/llc_occupancy > > -16234000 > > -# cat m11/mon_data/mon_L3_01/llc_occupancy > > -14789000 > > -# cat m12/mon_data/mon_L3_00/llc_occupancy > > -16789000 > > + # cat m11/mon_data/mon_L3_00/llc_occupancy > > + 16234000 > > + # cat m11/mon_data/mon_L3_01/llc_occupancy > > + 14789000 > > + # cat m12/mon_data/mon_L3_00/llc_occupancy > > + 16789000 > > > > The parent ctrl_mon group shows the aggregated data. > > +:: > > > > -# cat /sys/fs/resctrl/p1/mon_data/mon_l3_00/llc_occupancy > > -31234000 > > + # cat /sys/fs/resctrl/p1/mon_data/mon_l3_00/llc_occupancy > > + 31234000 > > > > Example 2 (Monitor a task from its creation) > > ---------- > > -On a two socket machine (one L3 cache per socket) > > +-------------------------------------------- > > +On a two socket machine (one L3 cache per socket):: > > > > -# mount -t resctrl resctrl /sys/fs/resctrl > > -# cd /sys/fs/resctrl > > -# mkdir p0 p1 > > + # mount -t resctrl resctrl /sys/fs/resctrl > > + # cd /sys/fs/resctrl > > + # mkdir p0 p1 > > > > An RMID is allocated to the group once its created and hence the <cmd> > > below is monitored from its creation. > > +:: > > > > -# echo $$ > /sys/fs/resctrl/p1/tasks > > -# <cmd> > > + # echo $$ > /sys/fs/resctrl/p1/tasks > > + # <cmd> > > > > -Fetch the data > > +Fetch the data:: > > > > -# cat /sys/fs/resctrl/p1/mon_data/mon_l3_00/llc_occupancy > > -31789000 > > + # cat /sys/fs/resctrl/p1/mon_data/mon_l3_00/llc_occupancy > > + 31789000 > > > > Example 3 (Monitor without CAT support or before creating CAT groups) > > ---------- > > +--------------------------------------------------------------------- > > > > Assume a system like HSW has only CQM and no CAT support. In this case > > the resctrl will still mount but cannot create CTRL_MON directories. > > @@ -1078,27 +1139,29 @@ able to monitor all tasks including kernel threads. > > > > This can also be used to profile jobs cache size footprint before being > > able to allocate them to different allocation groups. > > +:: > > > > -# mount -t resctrl resctrl /sys/fs/resctrl > > -# cd /sys/fs/resctrl > > -# mkdir mon_groups/m01 > > -# mkdir mon_groups/m02 > > + # mount -t resctrl resctrl /sys/fs/resctrl > > + # cd /sys/fs/resctrl > > + # mkdir mon_groups/m01 > > + # mkdir mon_groups/m02 > > > > -# echo 3478 > /sys/fs/resctrl/mon_groups/m01/tasks > > -# echo 2467 > /sys/fs/resctrl/mon_groups/m02/tasks > > + # echo 3478 > /sys/fs/resctrl/mon_groups/m01/tasks > > + # echo 2467 > /sys/fs/resctrl/mon_groups/m02/tasks > > > > Monitor the groups separately and also get per domain data. From the > > below its apparent that the tasks are mostly doing work on > > domain(socket) 0. > > +:: > > > > -# cat /sys/fs/resctrl/mon_groups/m01/mon_L3_00/llc_occupancy > > -31234000 > > -# cat /sys/fs/resctrl/mon_groups/m01/mon_L3_01/llc_occupancy > > -34555 > > -# cat /sys/fs/resctrl/mon_groups/m02/mon_L3_00/llc_occupancy > > -31234000 > > -# cat /sys/fs/resctrl/mon_groups/m02/mon_L3_01/llc_occupancy > > -32789 > > + # cat /sys/fs/resctrl/mon_groups/m01/mon_L3_00/llc_occupancy > > + 31234000 > > + # cat /sys/fs/resctrl/mon_groups/m01/mon_L3_01/llc_occupancy > > + 34555 > > + # cat /sys/fs/resctrl/mon_groups/m02/mon_L3_00/llc_occupancy > > + 31234000 > > + # cat /sys/fs/resctrl/mon_groups/m02/mon_L3_01/llc_occupancy > > + 32789 > > > > > > Example 4 (Monitor real time tasks) > > @@ -1107,15 +1170,17 @@ Example 4 (Monitor real time tasks) > > A single socket system which has real time tasks running on cores 4-7 > > and non real time tasks on other cpus. We want to monitor the cache > > occupancy of the real time threads on these cores. > > +:: > > > > -# mount -t resctrl resctrl /sys/fs/resctrl > > -# cd /sys/fs/resctrl > > -# mkdir p1 > > + # mount -t resctrl resctrl /sys/fs/resctrl > > + # cd /sys/fs/resctrl > > + # mkdir p1 > > > > -Move the cpus 4-7 over to p1 > > -# echo f0 > p1/cpus > > +Move the cpus 4-7 over to p1:: > > + > > There are extra whitespaces at the tail of the above line. > Fixed. > > After fixing the above: > > Reviewed-by: Mauro Carvalho Chehab <mchehab+sams...@kernel.org> > > > + # echo f0 > p1/cpus > > > > -View the llc occupancy snapshot > > +View the llc occupancy snapshot:: > > > > -# cat /sys/fs/resctrl/p1/mon_data/mon_L3_00/llc_occupancy > > -11234000 > > + # cat /sys/fs/resctrl/p1/mon_data/mon_L3_00/llc_occupancy > > + 11234000 > > > > Thanks, > Mauro -- Cheers, Changbin Du