Hi Ritesh,

I test, the regression still existed in v5.8-rc1. Do you have time to take a look at it? Thanks.

On 4/14/2020 1:49 PM, Xing Zhengjun wrote:
Thanks for your quick response, if you need any more test information about the regression, please let me known.

On 4/13/2020 6:56 PM, Ritesh Harjani wrote:


On 4/13/20 2:07 PM, Xing Zhengjun wrote:
Hi Harjani,

    Do you have time to take a look at this? Thanks.

Hello Xing,

I do want to look into this. But as of now I am stuck with another
mballoc failure issue. I will get back at this once I have some handle
over that one.

BTW, are you planning to take look at this?

-ritesh



On 4/7/2020 4:00 PM, kernel test robot wrote:
Greeting,

FYI, we noticed a -60.5% regression of stress-ng.fiemap.ops_per_sec due to commit:


commit: d3b6f23f71670007817a5d59f3fbafab2b794e8c ("ext4: move ext4_fiemap to use iomap framework")
https://git.kernel.org/cgit/linux/kernel/git/torvalds/linux.git master

in testcase: stress-ng
on test machine: 96 threads Intel(R) Xeon(R) Gold 6252 CPU @ 2.10GHz with 192G memory
with following parameters:

    nr_threads: 10%
    disk: 1HDD
    testtime: 1s
    class: os
    cpufreq_governor: performance
    ucode: 0x500002c
    fs: ext4






Details are as below:
-------------------------------------------------------------------------------------------------->


To reproduce:

         git clone https://github.com/intel/lkp-tests.git
         cd lkp-tests
         bin/lkp install job.yaml  # job file is attached in this email
         bin/lkp run     job.yaml

========================================================================================= class/compiler/cpufreq_governor/disk/fs/kconfig/nr_threads/rootfs/tbox_group/testcase/testtime/ucode: os/gcc-7/performance/1HDD/ext4/x86_64-rhel-7.6/10%/debian-x86_64-20191114.cgz/lkp-csl-2sp5/stress-ng/1s/0x500002c

commit:
   b2c5764262 ("ext4: make ext4_ind_map_blocks work with fiemap")
   d3b6f23f71 ("ext4: move ext4_fiemap to use iomap framework")

b2c5764262edded1 d3b6f23f71670007817a5d59f3f
---------------- ---------------------------
        fail:runs  %reproduction    fail:runs
            |             |             |
            :4           25%           1:4 dmesg.WARNING:at#for_ip_interrupt_entry/0x            2:4            5%           2:4 perf-profile.calltrace.cycles-pp.sync_regs.error_entry            2:4            6%           3:4 perf-profile.calltrace.cycles-pp.error_entry            3:4            9%           3:4 perf-profile.children.cycles-pp.error_entry            0:4            1%           0:4 perf-profile.self.cycles-pp.error_entry
          %stddev     %change         %stddev
              \          |                \
      28623           +28.2%      36703 ± 12%  stress-ng.daemon.ops
      28632           +28.2%      36704 ± 12% stress-ng.daemon.ops_per_sec
     566.00 ± 22%     -53.2%     265.00 ± 53%  stress-ng.dev.ops
     278.81 ± 22%     -53.0%     131.00 ± 54% stress-ng.dev.ops_per_sec
      73160           -60.6%      28849 ±  3%  stress-ng.fiemap.ops
      72471           -60.5%      28612 ±  3% stress-ng.fiemap.ops_per_sec
      23421 ± 12%     +21.2%      28388 ±  6%  stress-ng.filename.ops
      22638 ± 12%     +20.3%      27241 ±  6% stress-ng.filename.ops_per_sec
      21.25 ±  7%     -10.6%      19.00 ±  3%  stress-ng.iomix.ops
      38.75 ± 49%     -47.7%      20.25 ± 96%  stress-ng.memhotplug.ops
      34.45 ± 52%     -51.8%      16.62 ±106% stress-ng.memhotplug.ops_per_sec
       1734 ± 10%     +31.4%       2278 ± 10%  stress-ng.resources.ops
     807.56 ±  5%     +35.2%       1091 ±  8% stress-ng.resources.ops_per_sec
    1007356 ±  3%     -16.5%     840642 ±  9%  stress-ng.revio.ops
    1007692 ±  3%     -16.6%     840711 ±  9% stress-ng.revio.ops_per_sec
      21812 ±  3%     +16.0%      25294 ±  5%  stress-ng.sysbadaddr.ops
      21821 ±  3%     +15.9%      25294 ±  5% stress-ng.sysbadaddr.ops_per_sec
     440.75 ±  4%     +21.9%     537.25 ±  9%  stress-ng.sysfs.ops
     440.53 ±  4%     +21.9%     536.86 ±  9% stress-ng.sysfs.ops_per_sec    13286582           -11.1%   11805520 ±  6% stress-ng.time.file_system_outputs    68253896            +2.4%   69860122 stress-ng.time.minor_page_faults
     197.00 ±  4%     -15.9%     165.75 ± 12%  stress-ng.xattr.ops
     192.45 ±  5%     -16.1%     161.46 ± 11% stress-ng.xattr.ops_per_sec
      15310           +62.5%      24875 ± 22%  stress-ng.zombie.ops
      15310           +62.5%      24874 ± 22% stress-ng.zombie.ops_per_sec
     203.50 ± 12%     -47.3%     107.25 ± 49%  vmstat.io.bi
     861318 ± 18%     -29.7%     605884 ±  5%  meminfo.AnonHugePages
    1062742 ± 14%     -20.2%     847853 ±  3%  meminfo.AnonPages
      31093 ±  6%      +9.6%      34090 ±  3%  meminfo.KernelStack
       7151 ± 34%     +55.8%      11145 ±  9%  meminfo.Mlocked
  1.082e+08 ±  5%     -40.2%   64705429 ± 31% numa-numastat.node0.local_node   1.082e+08 ±  5%     -40.2%   64739883 ± 31% numa-numastat.node0.numa_hit    46032662 ± 21%    +104.3%   94042918 ± 20% numa-numastat.node1.local_node    46074205 ± 21%    +104.2%   94072810 ± 20% numa-numastat.node1.numa_hit        3942 ±  3%     +14.2%       4501 ±  4% slabinfo.pool_workqueue.active_objs        4098 ±  3%     +14.3%       4683 ±  4% slabinfo.pool_workqueue.num_objs        4817 ±  7%     +13.3%       5456 ±  8% slabinfo.proc_dir_entry.active_objs        5153 ±  6%     +12.5%       5797 ±  8% slabinfo.proc_dir_entry.num_objs       18598 ± 13%     -33.1%      12437 ± 20% sched_debug.cfs_rq:/.load.avg      452595 ± 56%     -71.4%     129637 ± 76% sched_debug.cfs_rq:/.load.max       67675 ± 35%     -55.1%      30377 ± 42% sched_debug.cfs_rq:/.load.stddev       18114 ± 12%     -33.7%      12011 ± 20% sched_debug.cfs_rq:/.runnable_weight.avg      448215 ± 58%     -72.8%     121789 ± 82% sched_debug.cfs_rq:/.runnable_weight.max       67083 ± 37%     -56.3%      29305 ± 43% sched_debug.cfs_rq:/.runnable_weight.stddev      -38032          +434.3%    -203212 sched_debug.cfs_rq:/.spread0.avg     -204466           +95.8%    -400301 sched_debug.cfs_rq:/.spread0.min       90.02 ± 25%     -58.1%      37.69 ± 52% sched_debug.cfs_rq:/.util_est_enqueued.avg      677.54 ±  6%     -39.3%     411.50 ± 22% sched_debug.cfs_rq:/.util_est_enqueued.max      196.57 ±  8%     -47.6%     103.05 ± 36% sched_debug.cfs_rq:/.util_est_enqueued.stddev        3.34 ± 23%     +34.1%       4.48 ±  4% sched_debug.cpu.clock.stddev        3.34 ± 23%     +34.1%       4.48 ±  4% sched_debug.cpu.clock_task.stddev      402872 ±  7%     -11.9%     354819 ±  2% proc-vmstat.nr_active_anon
    1730331            -9.5%    1566418 ±  5%  proc-vmstat.nr_dirtied
      31042 ±  6%      +9.3%      33915 ±  3% proc-vmstat.nr_kernel_stack
     229047            -2.4%     223615        proc-vmstat.nr_mapped
      74008 ±  7%     +20.5%      89163 ±  8%  proc-vmstat.nr_written
     402872 ±  7%     -11.9%     354819 ±  2% proc-vmstat.nr_zone_active_anon       50587 ± 11%     -25.2%      37829 ± 14% proc-vmstat.numa_pages_migrated      457500           -23.1%     351918 ± 31% proc-vmstat.numa_pte_updates
   81382485            +1.9%   82907822        proc-vmstat.pgfault
  2.885e+08 ±  5%     -13.3%  2.502e+08 ±  6%  proc-vmstat.pgfree
      42206 ± 12%     -46.9%      22399 ± 49%  proc-vmstat.pgpgin
     431233 ± 13%     -64.8%     151736 ±109%  proc-vmstat.pgrotated
     176754 ±  7%     -40.2%     105637 ± 31% proc-vmstat.thp_fault_alloc      314.50 ± 82%    +341.5%       1388 ± 44% proc-vmstat.unevictable_pgs_stranded     1075269 ± 14%     -41.3%     631388 ± 17% numa-meminfo.node0.Active      976056 ± 12%     -39.7%     588727 ± 19% numa-meminfo.node0.Active(anon)      426857 ± 22%     -36.4%     271375 ± 13% numa-meminfo.node0.AnonHugePages      558590 ± 19%     -36.4%     355402 ± 14% numa-meminfo.node0.AnonPages     1794824 ±  9%     -28.8%    1277157 ± 20% numa-meminfo.node0.FilePages        8517 ± 92%     -82.7%       1473 ± 89% numa-meminfo.node0.Inactive(file)      633118 ±  2%     -41.7%     368920 ± 36% numa-meminfo.node0.Mapped     2958038 ± 12%     -27.7%    2139271 ± 12% numa-meminfo.node0.MemUsed      181401 ±  5%     -13.7%     156561 ±  4% numa-meminfo.node0.SUnreclaim
     258124 ±  6%     -13.0%     224535 ±  5%  numa-meminfo.node0.Slab
     702083 ± 16%     +31.0%     919406 ± 11% numa-meminfo.node1.Active       38663 ±107%    +137.8%      91951 ± 31% numa-meminfo.node1.Active(file)     1154975 ±  7%     +41.6%    1635593 ± 12% numa-meminfo.node1.FilePages      395813 ± 25%     +62.8%     644533 ± 16% numa-meminfo.node1.Inactive      394313 ± 25%     +62.5%     640686 ± 16% numa-meminfo.node1.Inactive(anon)      273317           +88.8%     515976 ± 25% numa-meminfo.node1.Mapped     2279237 ±  6%     +25.7%    2865582 ±  7% numa-meminfo.node1.MemUsed       10830 ± 18%     +29.6%      14033 ±  9% numa-meminfo.node1.PageTables      149390 ±  3%     +23.2%     184085 ±  3% numa-meminfo.node1.SUnreclaim
     569542 ± 16%     +74.8%     995336 ± 21%  numa-meminfo.node1.Shmem
     220774 ±  5%     +20.3%     265656 ±  3%  numa-meminfo.node1.Slab
   35623587 ±  5%     -11.7%   31444514 ±  3%  perf-stat.i.cache-misses
  2.576e+08 ±  5%      -6.8%    2.4e+08 ±  2% perf-stat.i.cache-references        3585            -7.3%       3323 ±  5% perf-stat.i.cpu-migrations
     180139 ±  2%      +4.2%     187668        perf-stat.i.minor-faults
      69.13            +2.6       71.75 perf-stat.i.node-load-miss-rate%     4313695 ±  2%      -7.4%    3994957 ±  2% perf-stat.i.node-load-misses
    5466253 ± 11%     -17.3%    4521173 ±  6%  perf-stat.i.node-loads
    2818674 ±  6%     -15.8%    2372542 ±  5%  perf-stat.i.node-stores
     227810            +4.6%     238290        perf-stat.i.page-faults
      12.67 ±  4%      -7.2%      11.76 ±  2%  perf-stat.overall.MPKI
       1.01 ±  4%      -0.0        0.97 ±  3% perf-stat.overall.branch-miss-rate%        1044           +13.1%       1181 ±  4% perf-stat.overall.cycles-between-cache-misses       40.37 ±  4%      +3.6       44.00 ±  2% perf-stat.overall.node-store-miss-rate%    36139526 ±  5%     -12.5%   31625519 ±  3% perf-stat.ps.cache-misses   2.566e+08 ±  5%      -6.9%  2.389e+08 ±  2% perf-stat.ps.cache-references        3562            -7.2%       3306 ±  5% perf-stat.ps.cpu-migrations      179088            +4.2%     186579 perf-stat.ps.minor-faults
    4323383 ±  2%      -7.5%    3999214 perf-stat.ps.node-load-misses
    5607721 ± 10%     -18.5%    4568664 ±  6%  perf-stat.ps.node-loads
    2855134 ±  7%     -16.4%    2387345 ±  5%  perf-stat.ps.node-stores
     226270            +4.6%     236709        perf-stat.ps.page-faults
     242305 ± 10%     -42.4%     139551 ± 18% numa-vmstat.node0.nr_active_anon      135983 ± 17%     -37.4%      85189 ± 10% numa-vmstat.node0.nr_anon_pages      209.25 ± 16%     -38.1%     129.50 ± 10% numa-vmstat.node0.nr_anon_transparent_hugepages      449367 ±  9%     -29.7%     315804 ± 20% numa-vmstat.node0.nr_file_pages        2167 ± 90%     -80.6%     419.75 ± 98% numa-vmstat.node0.nr_inactive_file      157405 ±  3%     -41.4%      92206 ± 35% numa-vmstat.node0.nr_mapped        2022 ± 30%     -73.3%     539.25 ± 91% numa-vmstat.node0.nr_mlock        3336 ± 10%     -24.3%       2524 ± 25% numa-vmstat.node0.nr_page_table_pages      286158 ± 10%     -41.2%     168337 ± 37% numa-vmstat.node0.nr_shmem       45493 ±  5%     -14.1%      39094 ±  4% numa-vmstat.node0.nr_slab_unreclaimable      242294 ± 10%     -42.4%     139547 ± 18% numa-vmstat.node0.nr_zone_active_anon        2167 ± 90%     -80.6%     419.75 ± 98% numa-vmstat.node0.nr_zone_inactive_file    54053924 ±  8%     -39.3%   32786242 ± 34% numa-vmstat.node0.numa_hit    53929628 ±  8%     -39.5%   32619715 ± 34% numa-vmstat.node0.numa_local        9701 ±107%    +136.9%      22985 ± 31% numa-vmstat.node1.nr_active_file      202.50 ± 16%     -25.1%     151.75 ± 23% numa-vmstat.node1.nr_anon_transparent_hugepages      284922 ±  7%     +43.3%     408195 ± 13% numa-vmstat.node1.nr_file_pages       96002 ± 26%     +67.5%     160850 ± 17% numa-vmstat.node1.nr_inactive_anon       68077 ±  2%     +90.3%     129533 ± 25% numa-vmstat.node1.nr_mapped      138482 ± 15%     +79.2%     248100 ± 22% numa-vmstat.node1.nr_shmem       37396 ±  3%     +23.3%      46094 ±  3% numa-vmstat.node1.nr_slab_unreclaimable        9701 ±107%    +136.9%      22985 ± 31% numa-vmstat.node1.nr_zone_active_file       96005 ± 26%     +67.5%     160846 ± 17% numa-vmstat.node1.nr_zone_inactive_anon    23343661 ± 17%     +99.9%   46664267 ± 23% numa-vmstat.node1.numa_hit    23248487 ± 17%    +100.5%   46610447 ± 23% numa-vmstat.node1.numa_local
     105745 ± 23%    +112.6%     224805 ± 24%  softirqs.CPU0.NET_RX
     133310 ± 36%     -45.3%      72987 ± 52%  softirqs.CPU1.NET_RX
     170110 ± 55%     -66.8%      56407 ±147%  softirqs.CPU11.NET_RX
      91465 ± 36%     -65.2%      31858 ±112%  softirqs.CPU13.NET_RX
     164491 ± 57%     -77.7%      36641 ±121%  softirqs.CPU15.NET_RX
     121069 ± 55%     -99.3%     816.75 ± 96%  softirqs.CPU17.NET_RX
      81019 ±  4%      -8.7%      73967 ±  4%  softirqs.CPU20.RCU
      72143 ± 63%     -89.8%       7360 ±172%  softirqs.CPU22.NET_RX
     270663 ± 17%     -57.9%     113915 ± 45%  softirqs.CPU24.NET_RX
      20149 ± 76%    +474.1%     115680 ± 62%  softirqs.CPU26.NET_RX
      14033 ± 70%    +977.5%     151211 ± 75%  softirqs.CPU27.NET_RX
      27834 ± 94%    +476.1%     160357 ± 28%  softirqs.CPU28.NET_RX
      35346 ± 68%    +212.0%     110290 ± 30%  softirqs.CPU29.NET_RX
      34347 ±103%    +336.5%     149941 ± 32%  softirqs.CPU32.NET_RX
      70077 ±  3%     +10.8%      77624 ±  3%  softirqs.CPU34.RCU
      36453 ± 84%    +339.6%     160253 ± 42%  softirqs.CPU36.NET_RX
      72367 ±  2%     +10.6%      80043        softirqs.CPU37.RCU
      25239 ±118%    +267.7%      92799 ± 45%  softirqs.CPU38.NET_RX
       4995 ±170%   +1155.8%      62734 ± 62%  softirqs.CPU39.NET_RX
       4641 ±145%   +1611.3%      79432 ± 90%  softirqs.CPU42.NET_RX
       7192 ± 65%    +918.0%      73225 ± 66%  softirqs.CPU45.NET_RX
       1772 ±166%   +1837.4%      34344 ± 63%  softirqs.CPU46.NET_RX
      13149 ± 81%    +874.7%     128170 ± 58%  softirqs.CPU47.NET_RX
      86484 ± 94%     -92.6%       6357 ±172%  softirqs.CPU48.NET_RX
     129128 ± 27%     -95.8%       5434 ±172%  softirqs.CPU55.NET_RX
      82772 ± 59%     -91.7%       6891 ±164%  softirqs.CPU56.NET_RX
     145313 ± 57%     -87.8%      17796 ± 88%  softirqs.CPU57.NET_RX
     118160 ± 33%     -86.3%      16226 ±109%  softirqs.CPU58.NET_RX
      94576 ± 56%     -94.1%       5557 ±173%  softirqs.CPU6.NET_RX
      82900 ± 77%     -66.8%      27508 ±171%  softirqs.CPU62.NET_RX
     157291 ± 30%     -81.1%      29656 ±111%  softirqs.CPU64.NET_RX
     135101 ± 28%     -80.2%      26748 ± 90%  softirqs.CPU67.NET_RX
     146574 ± 56%    -100.0%      69.75 ± 98%  softirqs.CPU68.NET_RX
      81347 ±  2%      -9.0%      74024 ±  2%  softirqs.CPU68.RCU
     201729 ± 37%     -99.6%     887.50 ±107%  softirqs.CPU69.NET_RX
     108454 ± 78%     -97.9%       2254 ±169%  softirqs.CPU70.NET_RX
      55289 ±104%     -89.3%       5942 ±172%  softirqs.CPU71.NET_RX
      10112 ±172%    +964.6%     107651 ± 89%  softirqs.CPU72.NET_RX
       3136 ±171%   +1522.2%      50879 ± 66%  softirqs.CPU73.NET_RX
      13353 ± 79%    +809.2%     121407 ±101%  softirqs.CPU74.NET_RX
      75194 ±  3%     +10.3%      82957 ±  5%  softirqs.CPU75.RCU
      11002 ±173%   +1040.8%     125512 ± 61%  softirqs.CPU76.NET_RX
       2463 ±173%   +2567.3%      65708 ± 77%  softirqs.CPU78.NET_RX
      25956 ±  3%      -7.8%      23932 ±  3%  softirqs.CPU78.SCHED
      16366 ±150%    +340.7%      72125 ± 91%  softirqs.CPU82.NET_RX
      14553 ±130%   +1513.4%     234809 ± 27%  softirqs.CPU93.NET_RX
      26314            -9.2%      23884 ±  3%  softirqs.CPU93.SCHED
       4582 ± 88%   +4903.4%     229268 ± 23%  softirqs.CPU94.NET_RX
      11214 ±111%   +1762.5%     208867 ± 18%  softirqs.CPU95.NET_RX
       1.53 ± 27%      -0.5        0.99 ± 17% perf-profile.calltrace.cycles-pp.exit_to_usermode_loop.do_syscall_64.entry_SYSCALL_64_after_hwframe        1.52 ± 27%      -0.5        0.99 ± 17% perf-profile.calltrace.cycles-pp.do_signal.exit_to_usermode_loop.do_syscall_64.entry_SYSCALL_64_after_hwframe        1.39 ± 29%      -0.5        0.88 ± 21% perf-profile.calltrace.cycles-pp.do_group_exit.get_signal.do_signal.exit_to_usermode_loop.do_syscall_64        1.39 ± 29%      -0.5        0.88 ± 21% perf-profile.calltrace.cycles-pp.get_signal.do_signal.exit_to_usermode_loop.do_syscall_64.entry_SYSCALL_64_after_hwframe        0.50 ± 59%      +0.3        0.81 ± 13% perf-profile.calltrace.cycles-pp.filemap_map_pages.handle_pte_fault.__handle_mm_fault.handle_mm_fault.do_page_fault        5.70 ±  9%      +0.8        6.47 ±  7% perf-profile.calltrace.cycles-pp.do_exit.do_group_exit.get_signal.do_signal.exit_to_usermode_loop        5.48 ±  9%      +0.8        6.27 ±  7% perf-profile.calltrace.cycles-pp.exit_mmap.mmput.do_exit.do_group_exit.get_signal        5.49 ±  9%      +0.8        6.28 ±  7% perf-profile.calltrace.cycles-pp.mmput.do_exit.do_group_exit.get_signal.do_signal        4.30 ±  4%      +1.3        5.60 ±  7% perf-profile.calltrace.cycles-pp.do_group_exit.get_signal.do_signal.exit_to_usermode_loop.prepare_exit_to_usermode        4.40 ±  4%      +1.3        5.69 ±  7% perf-profile.calltrace.cycles-pp.prepare_exit_to_usermode.swapgs_restore_regs_and_return_to_usermode        4.37 ±  4%      +1.3        5.66 ±  7% perf-profile.calltrace.cycles-pp.exit_to_usermode_loop.prepare_exit_to_usermode.swapgs_restore_regs_and_return_to_usermode        4.36 ±  4%      +1.3        5.66 ±  7% perf-profile.calltrace.cycles-pp.do_signal.exit_to_usermode_loop.prepare_exit_to_usermode.swapgs_restore_regs_and_return_to_usermode        4.33 ±  4%      +1.3        5.62 ±  7% perf-profile.calltrace.cycles-pp.get_signal.do_signal.exit_to_usermode_loop.prepare_exit_to_usermode.swapgs_restore_regs_and_return_to_usermode        4.44 ±  4%      +1.3        5.74 ±  7% perf-profile.calltrace.cycles-pp.swapgs_restore_regs_and_return_to_usermode        3.20 ± 10%      -2.4        0.78 ±156% perf-profile.children.cycles-pp.copy_page        0.16 ±  9%      -0.1        0.08 ± 64% perf-profile.children.cycles-pp.irq_work_interrupt        0.16 ±  9%      -0.1        0.08 ± 64% perf-profile.children.cycles-pp.smp_irq_work_interrupt        0.24 ±  5%      -0.1        0.17 ± 18% perf-profile.children.cycles-pp.irq_work_run_list        0.16 ±  9%      -0.1        0.10 ± 24% perf-profile.children.cycles-pp.irq_work_run        0.16 ±  9%      -0.1        0.10 ± 24% perf-profile.children.cycles-pp.printk        0.23 ±  6%      -0.1        0.17 ±  9% perf-profile.children.cycles-pp.__do_execve_file        0.08 ± 14%      -0.1        0.03 ±100% perf-profile.children.cycles-pp.delay_tsc        0.16 ±  6%      -0.1        0.11 ±  9% perf-profile.children.cycles-pp.load_elf_binary        0.16 ±  7%      -0.0        0.12 ± 13% perf-profile.children.cycles-pp.search_binary_handler        0.20 ±  7%      -0.0        0.15 ± 10% perf-profile.children.cycles-pp.call_usermodehelper_exec_async        0.19 ±  6%      -0.0        0.15 ± 11% perf-profile.children.cycles-pp.do_execve        0.08 ± 10%      -0.0        0.04 ± 59% perf-profile.children.cycles-pp.__vunmap        0.15 ±  3%      -0.0        0.11 ±  7% perf-profile.children.cycles-pp.rcu_idle_exit        0.12 ± 10%      -0.0        0.09 ± 14% perf-profile.children.cycles-pp.__switch_to_asm        0.09 ± 13%      -0.0        0.07 ±  5% perf-profile.children.cycles-pp.des3_ede_encrypt        0.06 ± 11%      +0.0        0.09 ± 13% perf-profile.children.cycles-pp.mark_page_accessed        0.15 ±  5%      +0.0        0.19 ± 12% perf-profile.children.cycles-pp.apparmor_cred_prepare        0.22 ±  8%      +0.0        0.27 ± 11% perf-profile.children.cycles-pp.mem_cgroup_throttle_swaprate        0.17 ±  2%      +0.0        0.22 ± 12% perf-profile.children.cycles-pp.security_prepare_creds        0.95 ± 17%      +0.3        1.22 ± 14% perf-profile.children.cycles-pp.filemap_map_pages        5.92 ±  8%      +0.7        6.65 ±  7% perf-profile.children.cycles-pp.get_signal        5.66 ±  9%      +0.8        6.44 ±  7% perf-profile.children.cycles-pp.mmput        5.65 ±  9%      +0.8        6.43 ±  7% perf-profile.children.cycles-pp.exit_mmap        4.40 ±  4%      +1.3        5.70 ±  7% perf-profile.children.cycles-pp.prepare_exit_to_usermode        4.45 ±  4%      +1.3        5.75 ±  7% perf-profile.children.cycles-pp.swapgs_restore_regs_and_return_to_usermode        3.16 ± 10%      -2.4        0.77 ±155% perf-profile.self.cycles-pp.copy_page        0.08 ± 14%      -0.1        0.03 ±100% perf-profile.self.cycles-pp.delay_tsc        0.12 ± 10%      -0.0        0.09 ± 14% perf-profile.self.cycles-pp.__switch_to_asm        0.08 ± 12%      -0.0        0.06 ± 17% perf-profile.self.cycles-pp.enqueue_task_fair        0.09 ± 13%      -0.0        0.07 ±  5% perf-profile.self.cycles-pp.des3_ede_encrypt        0.07 ± 13%      +0.0        0.08 ± 19% perf-profile.self.cycles-pp.__lru_cache_add        0.19 ±  9%      +0.0        0.22 ± 10% perf-profile.self.cycles-pp.mem_cgroup_throttle_swaprate        0.15 ±  5%      +0.0        0.19 ± 11% perf-profile.self.cycles-pp.apparmor_cred_prepare        0.05 ± 58%      +0.0        0.09 ± 13% perf-profile.self.cycles-pp.mark_page_accessed        0.58 ± 10%      +0.2        0.80 ± 20% perf-profile.self.cycles-pp.release_pages        0.75 ±173%  +1.3e+05%       1005 ±100% interrupts.127:PCI-MSI.31981660-edge.i40e-eth0-TxRx-91      820.75 ±111%     -99.9%       0.50 ±173% interrupts.47:PCI-MSI.31981580-edge.i40e-eth0-TxRx-11      449.25 ± 86%    -100.0%       0.00 interrupts.53:PCI-MSI.31981586-edge.i40e-eth0-TxRx-17       33.25 ±157%    -100.0%       0.00 interrupts.57:PCI-MSI.31981590-edge.i40e-eth0-TxRx-21        0.75 ±110%  +63533.3%     477.25 ±162% interrupts.61:PCI-MSI.31981594-edge.i40e-eth0-TxRx-25      561.50 ±160%    -100.0%       0.00 interrupts.65:PCI-MSI.31981598-edge.i40e-eth0-TxRx-29       82921 ±  8%     -11.1%      73748 ±  6% interrupts.CPU11.CAL:Function_call_interrupts       66509 ± 30%     -32.6%      44828 ±  8% interrupts.CPU14.TLB:TLB_shootdowns       43105 ± 98%     -90.3%       4183 ± 21% interrupts.CPU17.RES:Rescheduling_interrupts      148719 ± 70%     -69.4%      45471 ± 16% interrupts.CPU17.TLB:TLB_shootdowns       85589 ± 42%     -52.2%      40884 ±  5% interrupts.CPU20.TLB:TLB_shootdowns      222472 ± 41%     -98.0%       4360 ± 45% interrupts.CPU22.RES:Rescheduling_interrupts        0.50 ±173%  +95350.0%     477.25 ±162% interrupts.CPU25.61:PCI-MSI.31981594-edge.i40e-eth0-TxRx-25       76029 ± 10%     +14.9%      87389 ±  5% interrupts.CPU25.CAL:Function_call_interrupts      399042 ±  6%     +13.4%     452479 ±  8% interrupts.CPU27.LOC:Local_timer_interrupts      561.00 ±161%    -100.0%       0.00 interrupts.CPU29.65:PCI-MSI.31981598-edge.i40e-eth0-TxRx-29        7034 ± 46%   +1083.8%      83279 ±138% interrupts.CPU29.RES:Rescheduling_interrupts       17829 ± 99%     -71.0%       5172 ± 16% interrupts.CPU30.RES:Rescheduling_interrupts        5569 ± 15%   +2414.7%     140059 ± 94% interrupts.CPU31.RES:Rescheduling_interrupts       37674 ± 16%     +36.6%      51473 ± 25% interrupts.CPU31.TLB:TLB_shootdowns       47905 ± 39%     +76.6%      84583 ± 38% interrupts.CPU34.TLB:TLB_shootdowns      568.75 ±140%    +224.8%       1847 ± 90% interrupts.CPU36.NMI:Non-maskable_interrupts      568.75 ±140%    +224.8%       1847 ± 90% interrupts.CPU36.PMI:Performance_monitoring_interrupts        4236 ± 25%   +2168.5%      96092 ± 90% interrupts.CPU36.RES:Rescheduling_interrupts       52717 ± 27%     +43.3%      75565 ± 28% interrupts.CPU37.TLB:TLB_shootdowns       41418 ±  9%    +136.6%      98010 ± 50% interrupts.CPU39.TLB:TLB_shootdowns        5551 ±  8%    +847.8%      52615 ± 66% interrupts.CPU40.RES:Rescheduling_interrupts        4746 ± 25%    +865.9%      45841 ± 91% interrupts.CPU42.RES:Rescheduling_interrupts       37556 ± 11%     +24.6%      46808 ±  6% interrupts.CPU42.TLB:TLB_shootdowns       21846 ±124%     -84.4%       3415 ± 46% interrupts.CPU48.RES:Rescheduling_interrupts      891.50 ± 22%     -35.2%     577.25 ± 40% interrupts.CPU49.NMI:Non-maskable_interrupts      891.50 ± 22%     -35.2%     577.25 ± 40% interrupts.CPU49.PMI:Performance_monitoring_interrupts       20459 ±120%     -79.2%       4263 ± 14% interrupts.CPU49.RES:Rescheduling_interrupts       59840 ± 21%     -23.1%      46042 ± 16% interrupts.CPU5.TLB:TLB_shootdowns       65200 ± 19%     -34.5%      42678 ±  9% interrupts.CPU51.TLB:TLB_shootdowns       70923 ±153%     -94.0%       4270 ± 29% interrupts.CPU53.RES:Rescheduling_interrupts       65312 ± 22%     -28.7%      46578 ± 14% interrupts.CPU56.TLB:TLB_shootdowns       65828 ± 24%   �� -33.4%      43846 ±  4% interrupts.CPU59.TLB:TLB_shootdowns       72558 ±156%     -93.2%       4906 ±  9% interrupts.CPU6.RES:Rescheduling_interrupts       68698 ± 34%     -32.6%      46327 ± 18% interrupts.CPU61.TLB:TLB_shootdowns      109745 ± 44%     -57.4%      46711 ± 16% interrupts.CPU62.TLB:TLB_shootdowns       89714 ± 44%     -48.5%      46198 ±  7% interrupts.CPU63.TLB:TLB_shootdowns       59380 ±136%     -91.5%       5066 ± 13% interrupts.CPU69.RES:Rescheduling_interrupts       40094 ± 18%    +133.9%      93798 ± 44% interrupts.CPU78.TLB:TLB_shootdowns      129884 ± 72%     -55.3%      58034 ±157% interrupts.CPU8.RES:Rescheduling_interrupts       69984 ± 11%     +51.4%     105957 ± 20% interrupts.CPU80.CAL:Function_call_interrupts       32857 ± 10%    +128.7%      75131 ± 36% interrupts.CPU80.TLB:TLB_shootdowns       35726 ± 16%     +34.6%      48081 ± 12% interrupts.CPU82.TLB:TLB_shootdowns       73820 ± 17%     +28.2%      94643 ±  8% interrupts.CPU84.CAL:Function_call_interrupts       38829 ± 28%    +190.3%     112736 ± 42% interrupts.CPU84.TLB:TLB_shootdowns       36129 ±  4%     +47.6%      53329 ± 13% interrupts.CPU85.TLB:TLB_shootdowns        4693 ±  7%   +1323.0%      66793 ±145% interrupts.CPU86.RES:Rescheduling_interrupts       38003 ± 11%     +94.8%      74031 ± 43% interrupts.CPU86.TLB:TLB_shootdowns       78022 ±  3%      +7.9%      84210 ±  3% interrupts.CPU87.CAL:Function_call_interrupts       36359 ±  6%     +54.9%      56304 ± 48% interrupts.CPU88.TLB:TLB_shootdowns       89031 ±105%     -95.0%       4475 ± 40% interrupts.CPU9.RES:Rescheduling_interrupts       40085 ± 11%     +60.6%      64368 ± 27% interrupts.CPU91.TLB:TLB_shootdowns       42244 ± 10%     +44.8%      61162 ± 35% interrupts.CPU94.TLB:TLB_shootdowns       40959 ± 15%    +109.4%      85780 ± 41% interrupts.CPU95.TLB:TLB_shootdowns


                                 stress-ng.fiemap.ops
   80000 +-------------------------------------------------------------------+    75000 |..+.             .+..            .+..+..  .+. .+..                |          |    +..+..+..+.+.     .+..+..  .+       +.   +. +.+..+..+..+.+..|    70000 |-+                   + +.                                    |    65000 |-+                                                                 |    60000 |-+                                                                 |    55000 |-+                                                                 |
|                                                                   |
   50000 |-+                                                                 |    45000 |-+                                                                 |    40000 |-+                                                                 |
   35000 |-+ O                                                |
         |  O       O                       O     O O                     |    30000 |-+  O  O     O O     O O     O  O    O     O    O  O O  O O  O O  |    25000 +-------------------------------------------------------------------+
                             stress-ng.fiemap.ops_per_sec
   80000 +-------------------------------------------------------------------+
   75000 |..               .+.. .+..                           |
         |  +.  .+..+..+.+.     .+..+..  .+.+. +..+.+..+..+.+..+..+..+.+..|    70000 |-+  +.               + +.                                    |    65000 |-+                                                                 |    60000 |-+                                                                 |    55000 |-+                                                                 |
|                                                                   |
   50000 |-+                                                                 |    45000 |-+                                                                 |    40000 |-+                                                                 |
   35000 |-+ O                                                |
         |  O       O                       O O                          |    30000 |-+  O  O     O O     O O     O       O     O O  O  O    O O  O O  |    25000 +-------------------------------------------------------------------+
[*] bisect-good sample
[O] bisect-bad  sample



Disclaimer:
Results have been estimated based on internal Intel analysis and are provided for informational purposes only. Any difference in system hardware or software
design or configuration may affect actual performance.


Thanks,
Rong Chen


_______________________________________________
LKP mailing list -- l...@lists.01.org
To unsubscribe send an email to lkp-le...@lists.01.org





--
Zhengjun Xing

Reply via email to