linux-mm.kvack.org archive mirror
 help / color / mirror / Atom feed
* [linux-next:master] [drivers/char/mem]  1b057bd800: stress-ng.splice.ops_per_sec -99.8% regression
@ 2023-10-17 15:06 kernel test robot
  2023-10-17 16:56 ` Greg Kroah-Hartman
       [not found] ` <CAKPOu+_T8xk4yd2P4KT4j3eMoFqwYmkxqDHaFtv4Hii5-XyPuA@mail.gmail.com>
  0 siblings, 2 replies; 9+ messages in thread
From: kernel test robot @ 2023-10-17 15:06 UTC (permalink / raw)
  To: Max Kellermann
  Cc: oe-lkp, lkp, Linux Memory Management List, Greg Kroah-Hartman,
	linux-kernel, ying.huang, feng.tang, fengwei.yin, oliver.sang



Hello,

kernel test robot noticed a -99.8% regression of stress-ng.splice.ops_per_sec on:


commit: 1b057bd800c3ea0c926191d7950cd2365eddc9bb ("drivers/char/mem: implement splice() for /dev/zero, /dev/full")
https://git.kernel.org/cgit/linux/kernel/git/next/linux-next.git master

testcase: stress-ng
test machine: 64 threads 2 sockets Intel(R) Xeon(R) Gold 6346 CPU @ 3.10GHz (Ice Lake) with 256G memory
parameters:

	nr_threads: 100%
	testtime: 60s
	class: pipe
	test: splice
	cpufreq_governor: performance


In addition to that, the commit also has significant impact on the following tests:

+------------------+-------------------------------------------------------------------------------------------------+
| testcase: change | stress-ng: stress-ng.splice.ops_per_sec 38.9% improvement                                       |
| test machine     | 36 threads 1 sockets Intel(R) Core(TM) i9-10980XE CPU @ 3.00GHz (Cascade Lake) with 128G memory |
| test parameters  | class=os                                                                                        |
|                  | cpufreq_governor=performance                                                                    |
|                  | disk=1HDD                                                                                       |
|                  | fs=ext4                                                                                         |
|                  | nr_threads=1                                                                                    |
|                  | test=splice                                                                                     |
|                  | testtime=60s                                                                                    |
+------------------+-------------------------------------------------------------------------------------------------+


If you fix the issue in a separate patch/commit (i.e. not just a new version of
the same patch/commit), kindly add following tags
| Reported-by: kernel test robot <oliver.sang@intel.com>
| Closes: https://lore.kernel.org/oe-lkp/202310172247.b9959bd4-oliver.sang@intel.com


Details are as below:
-------------------------------------------------------------------------------------------------->


The kernel config and materials to reproduce are available at:
https://download.01.org/0day-ci/archive/20231017/202310172247.b9959bd4-oliver.sang@intel.com

=========================================================================================
class/compiler/cpufreq_governor/kconfig/nr_threads/rootfs/tbox_group/test/testcase/testtime:
  pipe/gcc-12/performance/x86_64-rhel-8.3/100%/debian-11.1-x86_64-20220510.cgz/lkp-icl-2sp8/splice/stress-ng/60s

commit: 
  19e3e6cdfd ("accessibility: speakup: refactor deprecated strncpy")
  1b057bd800 ("drivers/char/mem: implement splice() for /dev/zero, /dev/full")

19e3e6cdfdc73400 1b057bd800c3ea0c926191d7950 
---------------- --------------------------- 
         %stddev     %change         %stddev
             \          |                \  
      2272          +166.0%       6045        uptime.idle
 2.724e+08 ±  6%   +1401.7%  4.091e+09        cpuidle..time
    301247 ±  3%   +1283.6%    4167916        cpuidle..usage
 3.774e+08 ±  5%     -99.6%    1510553 ±  8%  numa-numastat.node0.local_node
 3.774e+08 ±  5%     -99.6%    1545040 ±  6%  numa-numastat.node0.numa_hit
 3.696e+08 ±  5%     -99.6%    1536537 ±  8%  numa-numastat.node1.local_node
 3.698e+08 ±  5%     -99.6%    1568287 ±  7%  numa-numastat.node1.numa_hit
    136270           -91.3%      11853 ±  3%  meminfo.Active
    136158           -91.4%      11741 ±  3%  meminfo.Active(anon)
   1318175           -12.0%    1160436        meminfo.Committed_AS
     57581           -35.5%      37162        meminfo.Mapped
    161552           -88.2%      19074        meminfo.Shmem
      5.78 ±  9%     +93.1       98.86        mpstat.cpu.all.idle%
      0.72            -0.1        0.62        mpstat.cpu.all.irq%
      0.00 ± 17%      +0.0        0.02 ±  4%  mpstat.cpu.all.soft%
     78.79           -78.6        0.20 ±  4%  mpstat.cpu.all.sys%
     14.69           -14.4        0.27        mpstat.cpu.all.usr%
    402.17 ± 11%     -99.5%       2.17 ± 86%  perf-c2c.DRAM.local
      4747 ±  3%     -99.5%      22.83 ± 16%  perf-c2c.DRAM.remote
      4301 ±  6%     -98.8%      53.00 ± 18%  perf-c2c.HITM.local
      2593 ±  7%     -99.5%      14.00 ± 20%  perf-c2c.HITM.remote
      6894 ±  2%     -99.0%      67.00 ± 15%  perf-c2c.HITM.total
      8.60 ±  6%   +1046.3%      98.61        vmstat.cpu.id
     77.15           -98.5%       1.14        vmstat.cpu.sy
     14.23           -98.1%       0.28 ±  2%  vmstat.cpu.us
     58.37           -99.0%       0.60 ±  4%  vmstat.procs.r
    112757           -41.9%      65497        vmstat.system.in
     14891 ± 17%     -85.7%       2127 ± 55%  numa-meminfo.node0.Active
     14872 ± 17%     -86.1%       2071 ± 55%  numa-meminfo.node0.Active(anon)
     21319 ± 16%     -67.5%       6920 ± 18%  numa-meminfo.node0.Shmem
    122229           -92.0%       9734 ± 12%  numa-meminfo.node1.Active
    122135           -92.1%       9678 ± 12%  numa-meminfo.node1.Active(anon)
    140624           -91.4%      12163 ± 12%  numa-meminfo.node1.Shmem
    743.57          +334.3%       3229 ±  3%  stress-ng.splice.MB_per_sec_splice_rate
  7.46e+08           -99.8%    1373628 ±  3%  stress-ng.splice.ops
  12433266           -99.8%      22893 ±  3%  stress-ng.splice.ops_per_sec
     58608 ± 19%     -99.9%      41.50 ± 79%  stress-ng.time.involuntary_context_switches
      6121           -99.9%       5.67 ±  8%  stress-ng.time.percent_of_cpu_this_job_got
      3212           -99.9%       2.93 ±  6%  stress-ng.time.system_time
    586.44           -99.8%       0.99 ±  5%  stress-ng.time.user_time
      3721 ± 17%     -86.1%     517.86 ± 55%  numa-vmstat.node0.nr_active_anon
      5334 ± 16%     -67.6%       1727 ± 17%  numa-vmstat.node0.nr_shmem
      3721 ± 17%     -86.1%     517.86 ± 55%  numa-vmstat.node0.nr_zone_active_anon
 3.774e+08 ±  5%     -99.6%    1544858 ±  6%  numa-vmstat.node0.numa_hit
 3.774e+08 ±  5%     -99.6%    1510371 ±  8%  numa-vmstat.node0.numa_local
     30543           -92.1%       2409 ± 12%  numa-vmstat.node1.nr_active_anon
     35175           -91.4%       3033 ± 12%  numa-vmstat.node1.nr_shmem
     30543           -92.1%       2409 ± 12%  numa-vmstat.node1.nr_zone_active_anon
 3.698e+08 ±  5%     -99.6%    1567973 ±  7%  numa-vmstat.node1.numa_hit
 3.696e+08 ±  5%     -99.6%    1536223 ±  8%  numa-vmstat.node1.numa_local
      3375           -98.6%      47.67        turbostat.Avg_MHz
     94.04           -92.7        1.32        turbostat.Busy%
    260617 ±  9%   +1489.4%    4142197        turbostat.C1
      6.02 ±  9%     +93.8       99.83        turbostat.C1%
      5.96 ±  9%   +1556.1%      98.68        turbostat.CPU%c1
     63.83 ±  3%     -22.5%      49.50 ±  2%  turbostat.CoreTmp
   7374866           -41.9%    4288223        turbostat.IRQ
     23.49 ± 30%     -23.5        0.01 ±100%  turbostat.PKG_%
     63.00 ±  2%     -21.4%      49.50 ±  2%  turbostat.PkgTmp
    400.87           -40.6%     238.28        turbostat.PkgWatt
     70.18 ±  2%     -13.5%      60.74        turbostat.RAMWatt
     34160           -91.4%       2935 ±  3%  proc-vmstat.nr_active_anon
     87556            -1.5%      86204        proc-vmstat.nr_anon_pages
    726993            -4.9%     691342        proc-vmstat.nr_file_pages
     93734            -6.0%      88078        proc-vmstat.nr_inactive_anon
     14153           -34.3%       9292        proc-vmstat.nr_mapped
     40421           -88.2%       4770        proc-vmstat.nr_shmem
     34160           -91.4%       2935 ±  3%  proc-vmstat.nr_zone_active_anon
     93734            -6.0%      88078        proc-vmstat.nr_zone_inactive_anon
     13484 ±  5%     -99.7%      36.33 ± 58%  proc-vmstat.numa_hint_faults
     12534 ±  6%    -100.0%       5.50 ±223%  proc-vmstat.numa_hint_faults_local
 7.472e+08           -99.6%    3115004 ±  3%  proc-vmstat.numa_hit
  7.47e+08           -99.6%    3048767 ±  3%  proc-vmstat.numa_local
      1482 ± 28%     -97.9%      30.83 ± 45%  proc-vmstat.numa_pages_migrated
     55167           -99.8%     120.00 ± 46%  proc-vmstat.numa_pte_updates
     65858 ±  2%     -93.2%       4468 ±  4%  proc-vmstat.pgactivate
 7.465e+08           -99.6%    3156108 ±  3%  proc-vmstat.pgalloc_normal
    358101           -18.6%     291467        proc-vmstat.pgfault
 7.464e+08           -99.6%    3150970 ±  3%  proc-vmstat.pgfree
      1482 ± 28%     -97.9%      30.83 ± 45%  proc-vmstat.pgmigrate_success
   1919511           -99.7%       5014 ±  3%  sched_debug.cfs_rq:/.avg_vruntime.avg
   1945627           -99.2%      16366 ± 10%  sched_debug.cfs_rq:/.avg_vruntime.max
   1822945           -99.9%     977.33 ± 13%  sched_debug.cfs_rq:/.avg_vruntime.min
     20450 ±  4%     -84.1%       3245 ±  6%  sched_debug.cfs_rq:/.avg_vruntime.stddev
      0.63           -78.0%       0.14 ±  9%  sched_debug.cfs_rq:/.h_nr_running.avg
      1.83 ± 12%     -45.5%       1.00        sched_debug.cfs_rq:/.h_nr_running.max
      0.50          -100.0%       0.00        sched_debug.cfs_rq:/.h_nr_running.min
    545483           +51.1%     824178 ± 11%  sched_debug.cfs_rq:/.load.max
      8018          -100.0%       0.00        sched_debug.cfs_rq:/.load.min
     66622           +68.9%     112536 ± 12%  sched_debug.cfs_rq:/.load.stddev
    548.42         +3600.6%      20294 ± 91%  sched_debug.cfs_rq:/.load_avg.max
      7.50          -100.0%       0.00        sched_debug.cfs_rq:/.load_avg.min
    184.01 ±  5%   +1389.8%       2741 ± 89%  sched_debug.cfs_rq:/.load_avg.stddev
   1919511           -99.7%       5014 ±  3%  sched_debug.cfs_rq:/.min_vruntime.avg
   1945627           -99.2%      16366 ± 10%  sched_debug.cfs_rq:/.min_vruntime.max
   1822945           -99.9%     977.33 ± 13%  sched_debug.cfs_rq:/.min_vruntime.min
     20450 ±  4%     -84.1%       3245 ±  6%  sched_debug.cfs_rq:/.min_vruntime.stddev
      0.62           -77.5%       0.14 ±  9%  sched_debug.cfs_rq:/.nr_running.avg
      0.50          -100.0%       0.00        sched_debug.cfs_rq:/.nr_running.min
      0.21 ±  3%     +44.5%       0.31 ±  3%  sched_debug.cfs_rq:/.nr_running.stddev
    766.20           -65.5%     264.60 ±  5%  sched_debug.cfs_rq:/.runnable_avg.avg
      1333 ±  4%     -20.1%       1065 ± 12%  sched_debug.cfs_rq:/.runnable_avg.max
    512.00          -100.0%       0.00        sched_debug.cfs_rq:/.runnable_avg.min
    195.91 ±  4%     +26.6%     248.04 ±  6%  sched_debug.cfs_rq:/.runnable_avg.stddev
    762.56           -65.4%     264.06 ±  5%  sched_debug.cfs_rq:/.util_avg.avg
    505.42          -100.0%       0.00        sched_debug.cfs_rq:/.util_avg.min
    176.81 ±  3%     +39.9%     247.39 ±  6%  sched_debug.cfs_rq:/.util_avg.stddev
    248.65 ±  4%     -90.2%      24.31 ± 26%  sched_debug.cfs_rq:/.util_est_enqueued.avg
     84.08 ± 12%    -100.0%       0.00        sched_debug.cfs_rq:/.util_est_enqueued.min
    736503           +10.7%     815492        sched_debug.cpu.avg_idle.avg
    125501 ± 37%    +150.5%     314353 ±  7%  sched_debug.cpu.avg_idle.min
    209723 ±  5%     -17.4%     173320 ±  4%  sched_debug.cpu.avg_idle.stddev
      3.56 ± 11%     -59.9%       1.43 ±  5%  sched_debug.cpu.clock.stddev
      2481           -81.4%     462.76 ± 10%  sched_debug.cpu.curr->pid.avg
      2027 ±  6%    -100.0%       0.00        sched_debug.cpu.curr->pid.min
    790.94 ±  2%     +45.0%       1146 ±  4%  sched_debug.cpu.curr->pid.stddev
      0.63           -78.0%       0.14 ± 10%  sched_debug.cpu.nr_running.avg
      1.83 ± 12%     -45.5%       1.00        sched_debug.cpu.nr_running.max
      0.50          -100.0%       0.00        sched_debug.cpu.nr_running.min
      0.74 ±  2%     -42.9%       0.42 ±  5%  perf-stat.i.MPKI
 2.563e+10           -98.6%  3.557e+08        perf-stat.i.branch-instructions
      0.16 ±  5%      +1.1        1.27        perf-stat.i.branch-miss-rate%
  23826465 ±  4%     -70.4%    7064101        perf-stat.i.branch-misses
     36.05 ±  3%     -26.9        9.14 ±  4%  perf-stat.i.cache-miss-rate%
  97109331 ±  2%     -99.5%     445320 ±  3%  perf-stat.i.cache-misses
 2.636e+08           -98.2%    4830384 ±  3%  perf-stat.i.cache-references
      1.67           +58.9%       2.65        perf-stat.i.cpi
 2.185e+11           -99.0%  2.225e+09 ±  2%  perf-stat.i.cpu-cycles
    142.39           -38.7%      87.35        perf-stat.i.cpu-migrations
      2287 ±  2%    +214.5%       7193 ±  6%  perf-stat.i.cycles-between-cache-misses
      0.00 ±  5%      +0.0        0.03 ± 10%  perf-stat.i.dTLB-load-miss-rate%
    126811 ±  6%     -55.6%      56317 ±  8%  perf-stat.i.dTLB-load-misses
 3.748e+10           -98.9%  4.216e+08        perf-stat.i.dTLB-loads
      0.00 ±  7%      +0.0        0.02 ±  4%  perf-stat.i.dTLB-store-miss-rate%
     66664           -67.3%      21800 ±  4%  perf-stat.i.dTLB-store-misses
 2.342e+10           -99.2%  1.814e+08        perf-stat.i.dTLB-stores
 1.294e+11           -98.6%  1.763e+09        perf-stat.i.instructions
      0.61           -19.6%       0.49        perf-stat.i.ipc
      0.16 ± 54%     -81.7%       0.03 ± 48%  perf-stat.i.major-faults
      3.41           -99.0%       0.03 ±  2%  perf-stat.i.metric.GHz
    559.53 ±  3%     +18.7%     663.98 ±  3%  perf-stat.i.metric.K/sec
      1356           -98.9%      14.37        perf-stat.i.metric.M/sec
      4046           -27.6%       2928        perf-stat.i.minor-faults
     90.60            -5.5       85.14 ±  2%  perf-stat.i.node-load-miss-rate%
  15399896 ±  4%     -99.5%      73111 ±  5%  perf-stat.i.node-load-misses
   1456459 ±  5%     -98.4%      22922 ± 10%  perf-stat.i.node-loads
     96.48           -59.4       37.04 ± 31%  perf-stat.i.node-store-miss-rate%
  17686801 ±  3%     -99.8%      38212 ± 22%  perf-stat.i.node-store-misses
    211214 ±  5%     -73.3%      56368 ± 13%  perf-stat.i.node-stores
      4046           -27.6%       2928        perf-stat.i.page-faults
      0.75 ±  2%     -66.4%       0.25 ±  3%  perf-stat.overall.MPKI
      0.09 ±  5%      +1.9        1.99        perf-stat.overall.branch-miss-rate%
     36.84 ±  3%     -27.6        9.22 ±  3%  perf-stat.overall.cache-miss-rate%
      1.69           -25.3%       1.26 ±  2%  perf-stat.overall.cpi
      2252 ±  2%    +122.0%       4998 ±  2%  perf-stat.overall.cycles-between-cache-misses
      0.00 ±  8%      +0.0        0.01 ±  7%  perf-stat.overall.dTLB-load-miss-rate%
      0.00            +0.0        0.01 ±  4%  perf-stat.overall.dTLB-store-miss-rate%
      0.59           +34.0%       0.79 ±  2%  perf-stat.overall.ipc
     91.32           -15.2       76.12 ±  3%  perf-stat.overall.node-load-miss-rate%
     98.82           -58.5       40.36 ± 21%  perf-stat.overall.node-store-miss-rate%
 2.522e+10           -98.6%  3.509e+08        perf-stat.ps.branch-instructions
  23422336 ±  4%     -70.2%    6970716        perf-stat.ps.branch-misses
  95558027 ±  2%     -99.5%     438886 ±  3%  perf-stat.ps.cache-misses
 2.595e+08           -98.2%    4763002 ±  3%  perf-stat.ps.cache-references
  2.15e+11           -99.0%  2.192e+09 ±  2%  perf-stat.ps.cpu-cycles
    140.46           -38.8%      85.97        perf-stat.ps.cpu-migrations
    129105 ±  8%     -57.0%      55458 ±  8%  perf-stat.ps.dTLB-load-misses
 3.689e+10           -98.9%  4.157e+08        perf-stat.ps.dTLB-loads
     65855           -67.4%      21474 ±  4%  perf-stat.ps.dTLB-store-misses
 2.304e+10           -99.2%  1.789e+08        perf-stat.ps.dTLB-stores
 1.273e+11           -98.6%  1.739e+09        perf-stat.ps.instructions
      0.16 ± 54%     -81.6%       0.03 ± 48%  perf-stat.ps.major-faults
      3974           -27.5%       2882        perf-stat.ps.minor-faults
  15153146 ±  4%     -99.5%      72034 ±  5%  perf-stat.ps.node-load-misses
   1435599 ±  5%     -98.4%      22594 ± 10%  perf-stat.ps.node-loads
  17403697 ±  3%     -99.8%      37693 ± 22%  perf-stat.ps.node-store-misses
    207650 ±  5%     -73.2%      55558 ± 13%  perf-stat.ps.node-stores
      3974           -27.5%       2882        perf-stat.ps.page-faults
 8.067e+12           -98.6%  1.098e+11        perf-stat.total.instructions
      2.08 ±  5%    -100.0%       0.00 ±223%  perf-sched.sch_delay.avg.ms.__cond_resched.stop_one_cpu.sched_exec.bprm_execve.part
      0.20 ±149%     -98.4%       0.00 ± 11%  perf-sched.sch_delay.avg.ms.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
      0.09 ±135%     -97.5%       0.00 ± 17%  perf-sched.sch_delay.avg.ms.do_nanosleep.hrtimer_nanosleep.common_nsleep.__x64_sys_clock_nanosleep
      0.54 ± 30%     -99.5%       0.00 ± 13%  perf-sched.sch_delay.avg.ms.do_task_dead.do_exit.do_group_exit.__x64_sys_exit_group.do_syscall_64
      0.02 ± 82%     -81.8%       0.00 ± 11%  perf-sched.sch_delay.avg.ms.do_wait.kernel_wait4.__do_sys_wait4.do_syscall_64
      0.02 ± 37%     -65.0%       0.01 ± 16%  perf-sched.sch_delay.avg.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.syscall_exit_to_user_mode.do_syscall_64
      0.01 ± 19%     -78.8%       0.00        perf-sched.sch_delay.avg.ms.pipe_read.vfs_read.ksys_read.do_syscall_64
      0.01 ±  4%     -74.5%       0.00        perf-sched.sch_delay.avg.ms.rcu_gp_kthread.kthread.ret_from_fork.ret_from_fork_asm
      0.00 ±  7%     -58.6%       0.00        perf-sched.sch_delay.avg.ms.schedule_hrtimeout_range_clock.do_poll.constprop.0.do_sys_poll
      1.18 ± 56%     -99.6%       0.00 ± 10%  perf-sched.sch_delay.avg.ms.schedule_hrtimeout_range_clock.do_select.core_sys_select.kern_select
      0.01 ±  6%     -60.0%       0.00        perf-sched.sch_delay.avg.ms.schedule_hrtimeout_range_clock.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
      0.09 ± 33%     -96.6%       0.00        perf-sched.sch_delay.avg.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone
      0.01 ±  6%     -62.2%       0.00 ± 20%  perf-sched.sch_delay.avg.ms.schedule_timeout.kcompactd.kthread.ret_from_fork
      0.01 ± 33%     -61.5%       0.00 ± 14%  perf-sched.sch_delay.avg.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
      0.01 ±  9%     -27.6%       0.01        perf-sched.sch_delay.avg.ms.syslog_print.do_syslog.kmsg_read.vfs_read
      1.20 ± 30%     -99.8%       0.00        perf-sched.sch_delay.avg.ms.wait_for_partner.fifo_open.do_dentry_open.do_open
      1.23 ±116%     -99.6%       0.01 ± 11%  perf-sched.sch_delay.max.ms.__cond_resched.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr.__sched_setaffinity
      3.93          -100.0%       0.00 ±223%  perf-sched.sch_delay.max.ms.__cond_resched.stop_one_cpu.sched_exec.bprm_execve.part
      0.97 ±153%     -99.5%       0.01 ± 34%  perf-sched.sch_delay.max.ms.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
      0.50 ±142%     -99.4%       0.00 ± 11%  perf-sched.sch_delay.max.ms.do_nanosleep.hrtimer_nanosleep.common_nsleep.__x64_sys_clock_nanosleep
      3.91           -99.7%       0.01 ±111%  perf-sched.sch_delay.max.ms.do_task_dead.do_exit.do_group_exit.__x64_sys_exit_group.do_syscall_64
      1.03 ±130%     -99.3%       0.01 ± 24%  perf-sched.sch_delay.max.ms.do_wait.kernel_wait4.__do_sys_wait4.do_syscall_64
      3.82 ±  9%     -99.8%       0.01 ± 16%  perf-sched.sch_delay.max.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.syscall_exit_to_user_mode.do_syscall_64
      1.50 ± 65%     -99.3%       0.01 ± 15%  perf-sched.sch_delay.max.ms.pipe_read.vfs_read.ksys_read.do_syscall_64
      0.01 ± 20%     -50.7%       0.01 ±  9%  perf-sched.sch_delay.max.ms.rcu_gp_kthread.kthread.ret_from_fork.ret_from_fork_asm
      0.01 ± 18%     -47.3%       0.00 ± 14%  perf-sched.sch_delay.max.ms.schedule_hrtimeout_range_clock.do_poll.constprop.0.do_sys_poll
      2.75 ± 31%     -99.8%       0.01 ±  6%  perf-sched.sch_delay.max.ms.schedule_hrtimeout_range_clock.do_select.core_sys_select.kern_select
      0.01 ±  9%     -44.4%       0.01 ± 11%  perf-sched.sch_delay.max.ms.schedule_hrtimeout_range_clock.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
      2.45 ± 46%     -99.7%       0.01 ± 46%  perf-sched.sch_delay.max.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone
      0.01 ± 25%     -59.4%       0.00 ± 21%  perf-sched.sch_delay.max.ms.schedule_timeout.kcompactd.kthread.ret_from_fork
      2.04 ± 56%     -93.7%       0.13 ±121%  perf-sched.sch_delay.max.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
      0.01 ±  6%     -27.9%       0.01 ±  4%  perf-sched.sch_delay.max.ms.syslog_print.do_syslog.kmsg_read.vfs_read
      3.94           -99.9%       0.00 ± 48%  perf-sched.sch_delay.max.ms.wait_for_partner.fifo_open.do_dentry_open.do_open
      0.06 ±  4%     -91.9%       0.00 ± 24%  perf-sched.total_sch_delay.average.ms
    111.94 ±  6%    +149.0%     278.74 ±  7%  perf-sched.total_wait_and_delay.average.ms
      5730 ±  6%     -61.0%       2233 ±  6%  perf-sched.total_wait_and_delay.count.ms
    111.89 ±  6%    +149.1%     278.74 ±  7%  perf-sched.total_wait_time.average.ms
      0.07 ± 22%    -100.0%       0.00        perf-sched.wait_and_delay.avg.ms.__cond_resched.mutex_lock.pipe_double_lock.splice_pipe_to_pipe.do_splice
      0.17 ± 84%    -100.0%       0.00        perf-sched.wait_and_delay.avg.ms.__cond_resched.mutex_lock.pipe_write.vfs_write.ksys_write
      0.07 ± 32%    -100.0%       0.00        perf-sched.wait_and_delay.avg.ms.__cond_resched.mutex_lock.splice_from_pipe.do_splice.__do_splice
      2.08 ±  5%    -100.0%       0.00        perf-sched.wait_and_delay.avg.ms.__cond_resched.stop_one_cpu.sched_exec.bprm_execve.part
      2.87 ± 18%     -93.6%       0.18 ±  2%  perf-sched.wait_and_delay.avg.ms.do_wait.kernel_wait4.__do_sys_wait4.do_syscall_64
      0.08 ± 31%    -100.0%       0.00        perf-sched.wait_and_delay.avg.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt
      0.39 ±171%    -100.0%       0.00        perf-sched.wait_and_delay.avg.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.syscall_exit_to_user_mode.do_syscall_64
    148.55 ± 35%     +95.5%     290.39        perf-sched.wait_and_delay.avg.ms.pipe_read.vfs_read.ksys_read.do_syscall_64
      2.82 ±  3%     +60.0%       4.51 ± 11%  perf-sched.wait_and_delay.avg.ms.rcu_gp_kthread.kthread.ret_from_fork.ret_from_fork_asm
      1.67 ±  3%     -68.5%       0.53 ± 58%  perf-sched.wait_and_delay.avg.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone
      6.72 ±  3%    +376.7%      32.02 ± 10%  perf-sched.wait_and_delay.avg.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
    505.64 ±  7%     +20.3%     608.06 ±  7%  perf-sched.wait_and_delay.avg.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
      1.33 ± 24%     -99.8%       0.00        perf-sched.wait_and_delay.avg.ms.wait_for_partner.fifo_open.do_dentry_open.do_open
    570.88 ±  4%     -24.1%     433.13 ±  4%  perf-sched.wait_and_delay.avg.ms.worker_thread.kthread.ret_from_fork.ret_from_fork_asm
    110.50 ±  7%    -100.0%       0.00        perf-sched.wait_and_delay.count.__cond_resched.mutex_lock.pipe_double_lock.splice_pipe_to_pipe.do_splice
     46.17 ±  7%    -100.0%       0.00        perf-sched.wait_and_delay.count.__cond_resched.mutex_lock.pipe_write.vfs_write.ksys_write
     41.33 ± 14%    -100.0%       0.00        perf-sched.wait_and_delay.count.__cond_resched.mutex_lock.splice_from_pipe.do_splice.__do_splice
     26.67 ± 16%     -90.6%       2.50 ± 38%  perf-sched.wait_and_delay.count.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
     61.17 ± 23%    -100.0%       0.00        perf-sched.wait_and_delay.count.__cond_resched.stop_one_cpu.sched_exec.bprm_execve.part
    104.33 ±  3%     +18.5%     123.67        perf-sched.wait_and_delay.count.do_wait.kernel_wait4.__do_sys_wait4.do_syscall_64
    305.17 ±  6%    -100.0%       0.00        perf-sched.wait_and_delay.count.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt
      1977 ±  7%    -100.0%       0.00        perf-sched.wait_and_delay.count.exit_to_user_mode_loop.exit_to_user_mode_prepare.syscall_exit_to_user_mode.do_syscall_64
    503.33 ± 25%     -53.6%     233.33        perf-sched.wait_and_delay.count.pipe_read.vfs_read.ksys_read.do_syscall_64
     51.00 ±  6%     +64.7%      84.00 ±  9%  perf-sched.wait_and_delay.count.rcu_gp_kthread.kthread.ret_from_fork.ret_from_fork_asm
    721.67 ±  3%     -79.8%     145.67 ± 10%  perf-sched.wait_and_delay.count.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
    615.33 ±  6%     -12.2%     540.17 ±  7%  perf-sched.wait_and_delay.count.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
    291.17 ±  7%     +15.7%     337.00 ±  4%  perf-sched.wait_and_delay.count.worker_thread.kthread.ret_from_fork.ret_from_fork_asm
      1.28 ± 24%    -100.0%       0.00        perf-sched.wait_and_delay.max.ms.__cond_resched.mutex_lock.pipe_double_lock.splice_pipe_to_pipe.do_splice
      4.26 ±130%    -100.0%       0.00        perf-sched.wait_and_delay.max.ms.__cond_resched.mutex_lock.pipe_write.vfs_write.ksys_write
      1.47 ± 71%    -100.0%       0.00        perf-sched.wait_and_delay.max.ms.__cond_resched.mutex_lock.splice_from_pipe.do_splice.__do_splice
      2527 ± 38%     -47.1%       1336 ± 55%  perf-sched.wait_and_delay.max.ms.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
      3.93          -100.0%       0.00        perf-sched.wait_and_delay.max.ms.__cond_resched.stop_one_cpu.sched_exec.bprm_execve.part
     50.82 ± 99%     -98.9%       0.57 ±  2%  perf-sched.wait_and_delay.max.ms.do_wait.kernel_wait4.__do_sys_wait4.do_syscall_64
      5.04 ±101%    -100.0%       0.00        perf-sched.wait_and_delay.max.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt
    176.65 ±208%    -100.0%       0.00        perf-sched.wait_and_delay.max.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.syscall_exit_to_user_mode.do_syscall_64
    189.67 ± 18%     +44.2%     273.50 ± 17%  perf-sched.wait_and_delay.max.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
      3.94           -99.9%       0.00 ± 45%  perf-sched.wait_and_delay.max.ms.wait_for_partner.fifo_open.do_dentry_open.do_open
      0.05 ± 36%    -100.0%       0.00        perf-sched.wait_time.avg.ms.__cond_resched.__alloc_pages.pipe_write.vfs_write.ksys_write
      0.04 ± 94%    -100.0%       0.00        perf-sched.wait_time.avg.ms.__cond_resched.__splice_from_pipe.splice_from_pipe.do_splice.__do_splice
      0.12 ± 98%    -100.0%       0.00        perf-sched.wait_time.avg.ms.__cond_resched.dput.__fput.__x64_sys_close.do_syscall_64
      0.07 ± 22%    -100.0%       0.00        perf-sched.wait_time.avg.ms.__cond_resched.mutex_lock.pipe_double_lock.splice_pipe_to_pipe.do_splice
      0.17 ± 84%    -100.0%       0.00        perf-sched.wait_time.avg.ms.__cond_resched.mutex_lock.pipe_write.vfs_write.ksys_write
      0.21 ±164%    -100.0%       0.00        perf-sched.wait_time.avg.ms.__cond_resched.mutex_lock.splice_file_to_pipe.do_splice.__do_splice
      0.07 ± 32%    -100.0%       0.00        perf-sched.wait_time.avg.ms.__cond_resched.mutex_lock.splice_from_pipe.do_splice.__do_splice
      0.10 ± 39%    -100.0%       0.00        perf-sched.wait_time.avg.ms.__cond_resched.mutex_lock.splice_pipe_to_pipe.do_splice.__do_splice
      2.85 ± 18%     -93.7%       0.18 ±  2%  perf-sched.wait_time.avg.ms.do_wait.kernel_wait4.__do_sys_wait4.do_syscall_64
      0.07 ± 32%    -100.0%       0.00        perf-sched.wait_time.avg.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt
      0.13 ± 77%    -100.0%       0.00        perf-sched.wait_time.avg.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_call_function_single
      0.37 ±178%    -100.0%       0.00        perf-sched.wait_time.avg.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.syscall_exit_to_user_mode.do_syscall_64
    148.54 ± 35%     +95.5%     290.38        perf-sched.wait_time.avg.ms.pipe_read.vfs_read.ksys_read.do_syscall_64
      2.81 ±  3%     +60.4%       4.51 ± 11%  perf-sched.wait_time.avg.ms.rcu_gp_kthread.kthread.ret_from_fork.ret_from_fork_asm
      1.58 ±  5%     -66.9%       0.52 ± 58%  perf-sched.wait_time.avg.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone
      6.71 ±  3%    +376.9%      32.01 ± 10%  perf-sched.wait_time.avg.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
    505.63 ±  7%     +20.3%     608.06 ±  7%  perf-sched.wait_time.avg.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
      0.12 ± 74%    -100.0%       0.00        perf-sched.wait_time.avg.ms.wait_for_partner.fifo_open.do_dentry_open.do_open
    570.86 ±  4%     -24.1%     433.12 ±  4%  perf-sched.wait_time.avg.ms.worker_thread.kthread.ret_from_fork.ret_from_fork_asm
      0.49 ± 75%    -100.0%       0.00        perf-sched.wait_time.max.ms.__cond_resched.__alloc_pages.pipe_write.vfs_write.ksys_write
      0.20 ±133%    -100.0%       0.00        perf-sched.wait_time.max.ms.__cond_resched.__splice_from_pipe.splice_from_pipe.do_splice.__do_splice
      0.54 ± 98%    -100.0%       0.00        perf-sched.wait_time.max.ms.__cond_resched.dput.__fput.__x64_sys_close.do_syscall_64
      1.28 ± 24%    -100.0%       0.00        perf-sched.wait_time.max.ms.__cond_resched.mutex_lock.pipe_double_lock.splice_pipe_to_pipe.do_splice
      4.26 ±130%    -100.0%       0.00        perf-sched.wait_time.max.ms.__cond_resched.mutex_lock.pipe_write.vfs_write.ksys_write
      3.16 ±184%    -100.0%       0.00        perf-sched.wait_time.max.ms.__cond_resched.mutex_lock.splice_file_to_pipe.do_splice.__do_splice
      1.47 ± 71%    -100.0%       0.00        perf-sched.wait_time.max.ms.__cond_resched.mutex_lock.splice_from_pipe.do_splice.__do_splice
      1.46 ± 78%    -100.0%       0.00        perf-sched.wait_time.max.ms.__cond_resched.mutex_lock.splice_pipe_to_pipe.do_splice.__do_splice
      2527 ± 38%     -47.1%       1336 ± 55%  perf-sched.wait_time.max.ms.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
     50.74 ± 99%     -98.9%       0.56 ±  2%  perf-sched.wait_time.max.ms.do_wait.kernel_wait4.__do_sys_wait4.do_syscall_64
      5.04 ±101%    -100.0%       0.00        perf-sched.wait_time.max.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt
      1.02 ± 77%    -100.0%       0.00        perf-sched.wait_time.max.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_call_function_single
    176.65 ±208%    -100.0%       0.00        perf-sched.wait_time.max.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.syscall_exit_to_user_mode.do_syscall_64
    189.66 ± 18%     +44.2%     273.50 ± 17%  perf-sched.wait_time.max.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
      2.77 ± 33%    -100.0%       0.00 ±223%  perf-sched.wait_time.max.ms.wait_for_partner.fifo_open.do_dentry_open.do_open
     81.68           -81.7        0.00        perf-profile.calltrace.cycles-pp.splice
     69.84           -69.8        0.00        perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.splice
     67.70           -67.7        0.00        perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.splice
     64.18           -64.2        0.00        perf-profile.calltrace.cycles-pp.__x64_sys_splice.do_syscall_64.entry_SYSCALL_64_after_hwframe.splice
     59.37           -59.4        0.00        perf-profile.calltrace.cycles-pp.__do_splice.__x64_sys_splice.do_syscall_64.entry_SYSCALL_64_after_hwframe.splice
     57.18           -57.2        0.00        perf-profile.calltrace.cycles-pp.do_splice.__do_splice.__x64_sys_splice.do_syscall_64.entry_SYSCALL_64_after_hwframe
     42.45           -42.5        0.00        perf-profile.calltrace.cycles-pp.splice_from_pipe.do_splice.__do_splice.__x64_sys_splice.do_syscall_64
     41.51           -41.5        0.00        perf-profile.calltrace.cycles-pp.__splice_from_pipe.splice_from_pipe.do_splice.__do_splice.__x64_sys_splice
     39.36           -39.4        0.00        perf-profile.calltrace.cycles-pp.__folio_put.__splice_from_pipe.splice_from_pipe.do_splice.__do_splice
     39.26           -39.3        0.00        perf-profile.calltrace.cycles-pp.__mem_cgroup_uncharge.__folio_put.__splice_from_pipe.splice_from_pipe.do_splice
     38.84           -38.8        0.00        perf-profile.calltrace.cycles-pp.uncharge_batch.__mem_cgroup_uncharge.__folio_put.__splice_from_pipe.splice_from_pipe
     36.10           -36.1        0.00        perf-profile.calltrace.cycles-pp.page_counter_uncharge.uncharge_batch.__mem_cgroup_uncharge.__folio_put.__splice_from_pipe
     12.98           -13.0        0.00        perf-profile.calltrace.cycles-pp.write
     12.06           -12.1        0.00        perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.write
     11.90           -11.9        0.00        perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.write
     11.57           -11.6        0.00        perf-profile.calltrace.cycles-pp.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe.write
     10.51           -10.5        0.00        perf-profile.calltrace.cycles-pp.vfs_write.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe.write
      9.76            -9.8        0.00        perf-profile.calltrace.cycles-pp.splice_pipe_to_pipe.do_splice.__do_splice.__x64_sys_splice.do_syscall_64
      9.45            -9.5        0.00        perf-profile.calltrace.cycles-pp.pipe_write.vfs_write.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe
      9.30 ±  5%      -9.3        0.00        perf-profile.calltrace.cycles-pp.propagate_protected_usage.page_counter_uncharge.uncharge_batch.__mem_cgroup_uncharge.__folio_put
      8.53            -8.5        0.00        perf-profile.calltrace.cycles-pp.__entry_text_start.splice
      5.44            -5.4        0.00        perf-profile.calltrace.cycles-pp.__alloc_pages.pipe_write.vfs_write.ksys_write.do_syscall_64
      0.00            +0.7        0.66 ±  7%  perf-profile.calltrace.cycles-pp.rcu_sched_clock_irq.update_process_times.tick_sched_handle.tick_sched_timer.__hrtimer_run_queues
      0.00            +0.7        0.70 ±  7%  perf-profile.calltrace.cycles-pp.native_apic_msr_eoi.__sysvec_apic_timer_interrupt.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.acpi_safe_halt
      0.00            +0.7        0.73 ±  8%  perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe
      0.00            +0.7        0.73 ±  8%  perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe
      0.00            +0.7        0.74 ± 11%  perf-profile.calltrace.cycles-pp.update_sg_lb_stats.update_sd_lb_stats.find_busiest_group.load_balance.rebalance_domains
      0.00            +0.9        0.94 ± 13%  perf-profile.calltrace.cycles-pp.tick_nohz_next_event.tick_nohz_get_sleep_length.menu_select.cpuidle_idle_call.do_idle
      0.00            +0.9        0.94 ±  8%  perf-profile.calltrace.cycles-pp.update_sd_lb_stats.find_busiest_group.load_balance.rebalance_domains.__do_softirq
      0.00            +1.0        1.00 ±  7%  perf-profile.calltrace.cycles-pp.find_busiest_group.load_balance.rebalance_domains.__do_softirq.__irq_exit_rcu
      0.00            +1.0        1.05 ±  8%  perf-profile.calltrace.cycles-pp.__intel_pmu_enable_all.perf_rotate_context.perf_mux_hrtimer_handler.__hrtimer_run_queues.hrtimer_interrupt
      0.00            +1.1        1.07 ± 26%  perf-profile.calltrace.cycles-pp.tick_irq_enter.irq_enter_rcu.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.acpi_safe_halt
      0.00            +1.1        1.12 ±  6%  perf-profile.calltrace.cycles-pp.lapic_next_deadline.clockevents_program_event.hrtimer_interrupt.__sysvec_apic_timer_interrupt.sysvec_apic_timer_interrupt
      0.00            +1.4        1.40 ± 23%  perf-profile.calltrace.cycles-pp.irq_enter_rcu.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.acpi_safe_halt.acpi_idle_enter
      0.00            +1.4        1.44 ± 18%  perf-profile.calltrace.cycles-pp.cpuidle_enter.cpuidle_idle_call.do_idle.cpu_startup_entry.rest_init
      0.00            +1.5        1.48 ± 11%  perf-profile.calltrace.cycles-pp.perf_rotate_context.perf_mux_hrtimer_handler.__hrtimer_run_queues.hrtimer_interrupt.__sysvec_apic_timer_interrupt
      0.00            +1.5        1.51 ± 18%  perf-profile.calltrace.cycles-pp.cpuidle_idle_call.do_idle.cpu_startup_entry.rest_init.arch_call_rest_init
      0.00            +1.5        1.54 ± 17%  perf-profile.calltrace.cycles-pp.do_idle.cpu_startup_entry.rest_init.arch_call_rest_init.start_kernel
      0.00            +1.5        1.54 ± 17%  perf-profile.calltrace.cycles-pp.cpu_startup_entry.rest_init.arch_call_rest_init.start_kernel.x86_64_start_reservations
      0.00            +1.5        1.54 ± 17%  perf-profile.calltrace.cycles-pp.x86_64_start_kernel.secondary_startup_64_no_verify
      0.00            +1.5        1.54 ± 17%  perf-profile.calltrace.cycles-pp.x86_64_start_reservations.x86_64_start_kernel.secondary_startup_64_no_verify
      0.00            +1.5        1.54 ± 17%  perf-profile.calltrace.cycles-pp.start_kernel.x86_64_start_reservations.x86_64_start_kernel.secondary_startup_64_no_verify
      0.00            +1.5        1.54 ± 17%  perf-profile.calltrace.cycles-pp.arch_call_rest_init.start_kernel.x86_64_start_reservations.x86_64_start_kernel.secondary_startup_64_no_verify
      0.00            +1.5        1.54 ± 17%  perf-profile.calltrace.cycles-pp.rest_init.arch_call_rest_init.start_kernel.x86_64_start_reservations.x86_64_start_kernel
      0.00            +1.6        1.57 ±  3%  perf-profile.calltrace.cycles-pp.clockevents_program_event.hrtimer_interrupt.__sysvec_apic_timer_interrupt.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt
      0.00            +1.6        1.59 ± 13%  perf-profile.calltrace.cycles-pp.tick_nohz_get_sleep_length.menu_select.cpuidle_idle_call.do_idle.cpu_startup_entry
      0.00            +1.6        1.62 ±  6%  perf-profile.calltrace.cycles-pp.load_balance.rebalance_domains.__do_softirq.__irq_exit_rcu.sysvec_apic_timer_interrupt
      0.00            +1.7        1.66 ± 14%  perf-profile.calltrace.cycles-pp.ktime_get_update_offsets_now.hrtimer_interrupt.__sysvec_apic_timer_interrupt.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt
      0.00            +2.2        2.24 ± 10%  perf-profile.calltrace.cycles-pp.perf_mux_hrtimer_handler.__hrtimer_run_queues.hrtimer_interrupt.__sysvec_apic_timer_interrupt.sysvec_apic_timer_interrupt
      0.00            +2.3        2.30 ± 13%  perf-profile.calltrace.cycles-pp.arch_scale_freq_tick.scheduler_tick.update_process_times.tick_sched_handle.tick_sched_timer
      0.00            +2.4        2.38 ±  5%  perf-profile.calltrace.cycles-pp.rebalance_domains.__do_softirq.__irq_exit_rcu.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt
      0.00            +3.2        3.20 ±  5%  perf-profile.calltrace.cycles-pp.menu_select.cpuidle_idle_call.do_idle.cpu_startup_entry.start_secondary
      0.00            +3.9        3.88 ±  7%  perf-profile.calltrace.cycles-pp.__do_softirq.__irq_exit_rcu.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.acpi_safe_halt
      0.00            +4.4        4.41 ±  7%  perf-profile.calltrace.cycles-pp.__irq_exit_rcu.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.acpi_safe_halt.acpi_idle_enter
      0.00            +8.0        8.00 ±  8%  perf-profile.calltrace.cycles-pp.__intel_pmu_enable_all.perf_adjust_freq_unthr_context.perf_event_task_tick.scheduler_tick.update_process_times
      0.00           +12.1       12.15 ±  6%  perf-profile.calltrace.cycles-pp.perf_adjust_freq_unthr_context.perf_event_task_tick.scheduler_tick.update_process_times.tick_sched_handle
      0.00           +12.5       12.54 ±  6%  perf-profile.calltrace.cycles-pp.perf_event_task_tick.scheduler_tick.update_process_times.tick_sched_handle.tick_sched_timer
      0.00           +16.7       16.65 ±  6%  perf-profile.calltrace.cycles-pp.scheduler_tick.update_process_times.tick_sched_handle.tick_sched_timer.__hrtimer_run_queues
      0.00           +18.4       18.44 ±  5%  perf-profile.calltrace.cycles-pp.update_process_times.tick_sched_handle.tick_sched_timer.__hrtimer_run_queues.hrtimer_interrupt
      0.00           +18.6       18.64 ±  5%  perf-profile.calltrace.cycles-pp.tick_sched_handle.tick_sched_timer.__hrtimer_run_queues.hrtimer_interrupt.__sysvec_apic_timer_interrupt
      0.00           +20.5       20.49 ±  8%  perf-profile.calltrace.cycles-pp.tick_sched_timer.__hrtimer_run_queues.hrtimer_interrupt.__sysvec_apic_timer_interrupt.sysvec_apic_timer_interrupt
      0.00           +24.3       24.29 ±  7%  perf-profile.calltrace.cycles-pp.__hrtimer_run_queues.hrtimer_interrupt.__sysvec_apic_timer_interrupt.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt
      0.00           +28.3       28.30 ±  7%  perf-profile.calltrace.cycles-pp.hrtimer_interrupt.__sysvec_apic_timer_interrupt.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.acpi_safe_halt
      0.00           +29.3       29.31 ±  7%  perf-profile.calltrace.cycles-pp.__sysvec_apic_timer_interrupt.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.acpi_safe_halt.acpi_idle_enter
      0.00           +37.3       37.26 ±  4%  perf-profile.calltrace.cycles-pp.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.acpi_safe_halt.acpi_idle_enter.cpuidle_enter_state
      0.00           +46.9       46.88 ±  3%  perf-profile.calltrace.cycles-pp.acpi_safe_halt.acpi_idle_enter.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call
      0.00           +88.4       88.42        perf-profile.calltrace.cycles-pp.acpi_idle_enter.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call.do_idle
      0.00           +89.0       88.95        perf-profile.calltrace.cycles-pp.cpuidle_enter.cpuidle_idle_call.do_idle.cpu_startup_entry.start_secondary
      0.00           +89.9       89.85        perf-profile.calltrace.cycles-pp.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call.do_idle.cpu_startup_entry
      0.00           +93.5       93.46        perf-profile.calltrace.cycles-pp.cpuidle_idle_call.do_idle.cpu_startup_entry.start_secondary.secondary_startup_64_no_verify
      0.00           +94.3       94.31        perf-profile.calltrace.cycles-pp.do_idle.cpu_startup_entry.start_secondary.secondary_startup_64_no_verify
      0.00           +94.5       94.46        perf-profile.calltrace.cycles-pp.cpu_startup_entry.start_secondary.secondary_startup_64_no_verify
      0.00           +94.5       94.46        perf-profile.calltrace.cycles-pp.start_secondary.secondary_startup_64_no_verify
      0.00           +96.0       96.00        perf-profile.calltrace.cycles-pp.secondary_startup_64_no_verify
      0.00          +123.6      123.62        perf-profile.calltrace.cycles-pp.asm_sysvec_apic_timer_interrupt.acpi_safe_halt.acpi_idle_enter.cpuidle_enter_state.cpuidle_enter
     82.00           -82.0        0.00        perf-profile.children.cycles-pp.splice
     82.07           -79.8        2.28 ±  7%  perf-profile.children.cycles-pp.entry_SYSCALL_64_after_hwframe
     80.39           -78.1        2.27 ±  8%  perf-profile.children.cycles-pp.do_syscall_64
     64.69           -64.7        0.00        perf-profile.children.cycles-pp.__x64_sys_splice
     59.95           -60.0        0.00        perf-profile.children.cycles-pp.__do_splice
     57.55           -57.6        0.00        perf-profile.children.cycles-pp.do_splice
     42.50           -42.5        0.00        perf-profile.children.cycles-pp.splice_from_pipe
     41.57           -41.6        0.00        perf-profile.children.cycles-pp.__splice_from_pipe
     39.38           -39.4        0.00        perf-profile.children.cycles-pp.__folio_put
     39.28           -39.3        0.00        perf-profile.children.cycles-pp.__mem_cgroup_uncharge
     38.90           -38.9        0.00        perf-profile.children.cycles-pp.uncharge_batch
     36.14           -36.1        0.00        perf-profile.children.cycles-pp.page_counter_uncharge
     13.29           -13.2        0.09 ± 30%  perf-profile.children.cycles-pp.write
     11.64           -11.6        0.07 ± 28%  perf-profile.children.cycles-pp.ksys_write
     10.58           -10.5        0.07 ± 31%  perf-profile.children.cycles-pp.vfs_write
     10.12           -10.1        0.00        perf-profile.children.cycles-pp.splice_pipe_to_pipe
      9.56            -9.5        0.04 ± 73%  perf-profile.children.cycles-pp.pipe_write
      9.45 ±  5%      -9.4        0.00        perf-profile.children.cycles-pp.propagate_protected_usage
      5.51            -5.5        0.01 ±223%  perf-profile.children.cycles-pp.__entry_text_start
      5.49            -5.5        0.03 ±102%  perf-profile.children.cycles-pp.__alloc_pages
      1.18            -1.1        0.06 ± 78%  perf-profile.children.cycles-pp.__cond_resched
      0.00            +0.1        0.06 ± 23%  perf-profile.children.cycles-pp.tlb_batch_pages_flush
      0.00            +0.1        0.06 ± 21%  perf-profile.children.cycles-pp.schedule_idle
      0.00            +0.1        0.07 ± 23%  perf-profile.children.cycles-pp.filename_lookup
      0.00            +0.1        0.07 ± 23%  perf-profile.children.cycles-pp.path_lookupat
      0.00            +0.1        0.07 ± 32%  perf-profile.children.cycles-pp.exec_mmap
      0.00            +0.1        0.07 ± 20%  perf-profile.children.cycles-pp.sched_clock_noinstr
      0.07 ±  6%      +0.1        0.15 ± 23%  perf-profile.children.cycles-pp._raw_spin_unlock_irqrestore
      0.00            +0.1        0.08 ± 25%  perf-profile.children.cycles-pp.evsel__read_counter
      0.00            +0.1        0.08 ± 49%  perf-profile.children.cycles-pp.setlocale
      0.00            +0.1        0.08 ± 24%  perf-profile.children.cycles-pp.drm_gem_get_pages
      0.00            +0.1        0.08 ± 24%  perf-profile.children.cycles-pp.drm_gem_shmem_get_pages
      0.00            +0.1        0.08 ± 23%  perf-profile.children.cycles-pp.release_pages
      0.00            +0.1        0.08 ± 35%  perf-profile.children.cycles-pp.can_stop_idle_tick
      0.00            +0.1        0.08 ± 27%  perf-profile.children.cycles-pp.begin_new_exec
      0.00            +0.1        0.09 ± 41%  perf-profile.children.cycles-pp.elf_map
      0.00            +0.1        0.09 ± 23%  perf-profile.children.cycles-pp.copy_strings
      0.00            +0.1        0.09 ± 26%  perf-profile.children.cycles-pp.tlb_finish_mmu
      0.00            +0.1        0.09 ± 36%  perf-profile.children.cycles-pp.drm_gem_vmap_unlocked
      0.00            +0.1        0.09 ± 36%  perf-profile.children.cycles-pp.drm_gem_vmap
      0.00            +0.1        0.09 ± 36%  perf-profile.children.cycles-pp.drm_gem_shmem_vmap
      0.00            +0.1        0.09 ± 22%  perf-profile.children.cycles-pp.tick_nohz_stop_idle
      0.00            +0.1        0.10 ± 36%  perf-profile.children.cycles-pp.dup_mm
      0.00            +0.1        0.10 ± 49%  perf-profile.children.cycles-pp.tick_program_event
      0.00            +0.1        0.10 ± 40%  perf-profile.children.cycles-pp.mas_store_prealloc
      0.00            +0.1        0.10 ± 39%  perf-profile.children.cycles-pp.irq_work_needs_cpu
      0.00            +0.1        0.10 ± 27%  perf-profile.children.cycles-pp.rb_insert_color
      0.00            +0.1        0.10 ± 21%  perf-profile.children.cycles-pp.rcu_do_batch
      0.00            +0.1        0.10 ± 12%  perf-profile.children.cycles-pp.seq_read_iter
      0.00            +0.1        0.10 ± 26%  perf-profile.children.cycles-pp.ct_kernel_exit
      0.00            +0.1        0.11 ± 21%  perf-profile.children.cycles-pp.tick_nohz_idle_retain_tick
      0.00            +0.1        0.11 ± 25%  perf-profile.children.cycles-pp.link_path_walk
      0.00            +0.1        0.11 ± 38%  perf-profile.children.cycles-pp.rb_next
      0.00            +0.1        0.12 ±  9%  perf-profile.children.cycles-pp.__update_load_avg_cfs_rq
      0.00            +0.1        0.12 ± 25%  perf-profile.children.cycles-pp.do_vmi_align_munmap
      0.00            +0.1        0.12 ± 46%  perf-profile.children.cycles-pp.copy_mc_enhanced_fast_string
      0.00            +0.1        0.12 ± 44%  perf-profile.children.cycles-pp.update_rt_rq_load_avg
      0.00            +0.1        0.12 ± 22%  perf-profile.children.cycles-pp.do_vmi_munmap
      0.00            +0.1        0.12 ± 26%  perf-profile.children.cycles-pp.schedule
      0.00            +0.1        0.12 ± 41%  perf-profile.children.cycles-pp.smpboot_thread_fn
      0.00            +0.1        0.12 ± 29%  perf-profile.children.cycles-pp.evlist_cpu_iterator__next
      0.00            +0.1        0.12 ± 38%  perf-profile.children.cycles-pp.__mmap
      0.00            +0.1        0.13 ± 10%  perf-profile.children.cycles-pp.tsc_verify_tsc_adjust
      0.00            +0.1        0.13 ± 36%  perf-profile.children.cycles-pp._dl_addr
      0.00            +0.1        0.13 ± 15%  perf-profile.children.cycles-pp.run_posix_cpu_timers
      0.00            +0.1        0.14 ± 41%  perf-profile.children.cycles-pp.__collapse_huge_page_copy
      0.00            +0.1        0.14 ± 45%  perf-profile.children.cycles-pp.rcu_report_qs_rdp
      0.00            +0.1        0.14 ± 28%  perf-profile.children.cycles-pp.__do_sys_clone
      0.00            +0.1        0.14 ± 25%  perf-profile.children.cycles-pp.drm_fb_helper_damage_work
      0.00            +0.1        0.14 ± 25%  perf-profile.children.cycles-pp.drm_fbdev_generic_helper_fb_dirty
      0.00            +0.1        0.14 ± 18%  perf-profile.children.cycles-pp.hrtimer_forward
      0.00            +0.1        0.14 ± 12%  perf-profile.children.cycles-pp.arch_cpu_idle_enter
      0.00            +0.2        0.15 ± 35%  perf-profile.children.cycles-pp.intel_pmu_disable_all
      0.00            +0.2        0.15 ± 38%  perf-profile.children.cycles-pp.collapse_huge_page
      0.00            +0.2        0.15 ± 33%  perf-profile.children.cycles-pp.irqentry_exit
      0.00            +0.2        0.16 ± 37%  perf-profile.children.cycles-pp.khugepaged
      0.00            +0.2        0.16 ± 37%  perf-profile.children.cycles-pp.khugepaged_scan_mm_slot
      0.00            +0.2        0.16 ± 37%  perf-profile.children.cycles-pp.hpage_collapse_scan_pmd
      0.00            +0.2        0.16 ± 32%  perf-profile.children.cycles-pp.menu_reflect
      0.00            +0.2        0.16 ± 11%  perf-profile.children.cycles-pp.cpu_util
      0.00            +0.2        0.16 ± 36%  perf-profile.children.cycles-pp.update_wall_time
      0.00            +0.2        0.16 ± 36%  perf-profile.children.cycles-pp.timekeeping_advance
      0.00            +0.2        0.17 ± 33%  perf-profile.children.cycles-pp.__libc_fork
      0.00            +0.2        0.17 ± 42%  perf-profile.children.cycles-pp.sched_setaffinity
      0.00            +0.2        0.17 ± 18%  perf-profile.children.cycles-pp.copy_process
      0.00            +0.2        0.17 ± 33%  perf-profile.children.cycles-pp.arch_cpu_idle_exit
      0.00            +0.2        0.18 ± 15%  perf-profile.children.cycles-pp.__update_blocked_fair
      0.00            +0.2        0.18 ± 35%  perf-profile.children.cycles-pp.asm_sysvec_call_function_single
      0.00            +0.2        0.18 ± 39%  perf-profile.children.cycles-pp.path_openat
      0.00            +0.2        0.19 ± 36%  perf-profile.children.cycles-pp.do_filp_open
      0.00            +0.2        0.19 ± 15%  perf-profile.children.cycles-pp.call_cpuidle
      0.00            +0.2        0.20 ± 27%  perf-profile.children.cycles-pp.filemap_map_pages
      0.00            +0.2        0.20 ± 17%  perf-profile.children.cycles-pp.__schedule
      0.00            +0.2        0.20 ± 25%  perf-profile.children.cycles-pp.check_cpu_stall
      0.00            +0.2        0.20 ± 13%  perf-profile.children.cycles-pp.kernel_clone
      0.00            +0.2        0.21 ± 41%  perf-profile.children.cycles-pp.cpuidle_reflect
      0.00            +0.2        0.21 ± 20%  perf-profile.children.cycles-pp.idle_cpu
      0.00            +0.2        0.21 ± 25%  perf-profile.children.cycles-pp.note_gp_changes
      0.00            +0.2        0.21 ± 23%  perf-profile.children.cycles-pp.do_sys_openat2
      0.00            +0.2        0.21 ± 30%  perf-profile.children.cycles-pp.error_entry
      0.00            +0.2        0.21 ± 17%  perf-profile.children.cycles-pp.__memcpy
      0.00            +0.2        0.21 ± 22%  perf-profile.children.cycles-pp.__x64_sys_openat
      0.00            +0.2        0.21 ± 22%  perf-profile.children.cycles-pp.exit_mm
      0.00            +0.2        0.21 ± 26%  perf-profile.children.cycles-pp.hrtimer_get_next_event
      0.00            +0.2        0.22 ± 33%  perf-profile.children.cycles-pp.do_read_fault
      0.00            +0.2        0.22 ± 34%  perf-profile.children.cycles-pp._find_next_and_bit
      0.00            +0.2        0.23 ± 20%  perf-profile.children.cycles-pp.read
      0.00            +0.2        0.23 ± 31%  perf-profile.children.cycles-pp.process_one_work
      0.00            +0.2        0.24 ±  9%  perf-profile.children.cycles-pp.read_counters
      0.00            +0.2        0.24 ±  8%  perf-profile.children.cycles-pp.cmd_stat
      0.00            +0.2        0.24 ±  8%  perf-profile.children.cycles-pp.dispatch_events
      0.00            +0.2        0.24 ±  8%  perf-profile.children.cycles-pp.process_interval
      0.00            +0.2        0.24 ± 23%  perf-profile.children.cycles-pp.tick_check_broadcast_expired
      0.00            +0.2        0.25 ± 17%  perf-profile.children.cycles-pp.ksys_mmap_pgoff
      0.00            +0.3        0.25 ± 29%  perf-profile.children.cycles-pp.trigger_load_balance
      0.00            +0.3        0.26 ±  7%  perf-profile.children.cycles-pp.main
      0.00            +0.3        0.26 ±  7%  perf-profile.children.cycles-pp.run_builtin
      0.00            +0.3        0.26 ±  7%  perf-profile.children.cycles-pp.__libc_start_main
      0.00            +0.3        0.26 ± 34%  perf-profile.children.cycles-pp.worker_thread
      0.00            +0.3        0.26 ± 21%  perf-profile.children.cycles-pp.exit_mmap
      0.00            +0.3        0.27 ± 32%  perf-profile.children.cycles-pp.timerqueue_del
      0.00            +0.3        0.27 ± 22%  perf-profile.children.cycles-pp.__mmput
      0.00            +0.3        0.27 ± 18%  perf-profile.children.cycles-pp.vfs_read
      0.00            +0.3        0.28 ± 17%  perf-profile.children.cycles-pp.ksys_read
      0.00            +0.3        0.28 ± 28%  perf-profile.children.cycles-pp.do_fault
      0.00            +0.3        0.28 ± 18%  perf-profile.children.cycles-pp.__x64_sys_exit_group
      0.00            +0.3        0.28 ± 18%  perf-profile.children.cycles-pp.do_group_exit
      0.00            +0.3        0.28 ± 18%  perf-profile.children.cycles-pp.do_exit
      0.00            +0.3        0.28 ± 21%  perf-profile.children.cycles-pp.hrtimer_update_next_event
      0.00            +0.3        0.28 ± 23%  perf-profile.children.cycles-pp.ct_nmi_enter
      0.00            +0.3        0.28 ± 23%  perf-profile.children.cycles-pp.ct_kernel_enter
      0.00            +0.3        0.29 ± 20%  perf-profile.children.cycles-pp.mmap_region
      0.00            +0.3        0.32 ± 21%  perf-profile.children.cycles-pp.do_mmap
      0.00            +0.3        0.33 ± 19%  perf-profile.children.cycles-pp.vm_mmap_pgoff
      0.00            +0.3        0.34 ± 12%  perf-profile.children.cycles-pp.irq_work_tick
      0.00            +0.4        0.35 ± 19%  perf-profile.children.cycles-pp.local_clock_noinstr
      0.00            +0.4        0.35 ± 22%  perf-profile.children.cycles-pp.load_elf_binary
      0.00            +0.4        0.37 ± 21%  perf-profile.children.cycles-pp.exec_binprm
      0.00            +0.4        0.37 ± 21%  perf-profile.children.cycles-pp.search_binary_handler
      0.00            +0.4        0.37 ± 13%  perf-profile.children.cycles-pp.timerqueue_add
      0.00            +0.4        0.38 ± 23%  perf-profile.children.cycles-pp.tick_nohz_irq_exit
      0.00            +0.4        0.38 ± 21%  perf-profile.children.cycles-pp.ct_idle_exit
      0.00            +0.4        0.38 ± 18%  perf-profile.children.cycles-pp.__hrtimer_next_event_base
      0.00            +0.4        0.40 ± 44%  perf-profile.children.cycles-pp.calc_global_load_tick
      0.00            +0.4        0.40 ± 20%  perf-profile.children.cycles-pp.update_rq_clock
      0.00            +0.4        0.40 ± 21%  perf-profile.children.cycles-pp.bprm_execve
      0.00            +0.4        0.40 ± 16%  perf-profile.children.cycles-pp.irqentry_enter
      0.00            +0.4        0.40 ± 33%  perf-profile.children.cycles-pp.__handle_mm_fault
      0.00            +0.4        0.41 ± 24%  perf-profile.children.cycles-pp.tick_check_oneshot_broadcast_this_cpu
      0.00            +0.4        0.42 ± 13%  perf-profile.children.cycles-pp.get_cpu_device
      0.30 ±  2%      +0.4        0.72 ± 13%  perf-profile.children.cycles-pp._raw_spin_lock_irqsave
      0.00            +0.4        0.42 ± 20%  perf-profile.children.cycles-pp.update_irq_load_avg
      0.00            +0.4        0.42 ± 29%  perf-profile.children.cycles-pp.handle_mm_fault
      0.00            +0.4        0.42 ± 20%  perf-profile.children.cycles-pp.x86_pmu_disable
      0.00            +0.4        0.44 ± 23%  perf-profile.children.cycles-pp.perf_pmu_nop_void
      0.00            +0.4        0.44 ± 26%  perf-profile.children.cycles-pp.exc_page_fault
      0.00            +0.4        0.44 ± 26%  perf-profile.children.cycles-pp.do_user_addr_fault
      0.00            +0.4        0.45 ± 20%  perf-profile.children.cycles-pp.should_we_balance
      0.00            +0.4        0.45 ± 12%  perf-profile.children.cycles-pp.enqueue_hrtimer
      0.00            +0.5        0.48 ± 25%  perf-profile.children.cycles-pp.ct_kernel_exit_state
      0.00            +0.5        0.48 ± 28%  perf-profile.children.cycles-pp.hrtimer_next_event_without
      0.00            +0.5        0.48 ± 23%  perf-profile.children.cycles-pp.rcu_core
      0.00            +0.5        0.49 ± 14%  perf-profile.children.cycles-pp.get_next_timer_interrupt
      0.00            +0.5        0.49 ±  7%  perf-profile.children.cycles-pp.rcu_pending
      0.00            +0.6        0.56 ± 28%  perf-profile.children.cycles-pp.asm_exc_page_fault
      0.00            +0.6        0.57 ± 11%  perf-profile.children.cycles-pp.cpuidle_governor_latency_req
      0.00            +0.6        0.58 ± 21%  perf-profile.children.cycles-pp.do_execveat_common
      0.00            +0.6        0.58 ± 21%  perf-profile.children.cycles-pp.__x64_sys_execve
      0.00            +0.6        0.59 ± 20%  perf-profile.children.cycles-pp.execve
      0.00            +0.6        0.62 ± 20%  perf-profile.children.cycles-pp.update_rq_clock_task
      0.00            +0.6        0.62 ± 33%  perf-profile.children.cycles-pp.kthread
      0.00            +0.6        0.63 ± 32%  perf-profile.children.cycles-pp.ret_from_fork
      0.00            +0.6        0.63 ± 21%  perf-profile.children.cycles-pp.update_blocked_averages
      0.00            +0.6        0.64 ± 32%  perf-profile.children.cycles-pp.ret_from_fork_asm
      0.00            +0.7        0.67 ±  6%  perf-profile.children.cycles-pp.rcu_sched_clock_irq
      0.00            +0.7        0.68 ± 19%  perf-profile.children.cycles-pp.run_rebalance_domains
      0.00            +0.8        0.75 ± 10%  perf-profile.children.cycles-pp.irqtime_account_irq
      0.00            +0.8        0.84 ± 10%  perf-profile.children.cycles-pp.update_sg_lb_stats
      0.00            +0.9        0.87 ± 17%  perf-profile.children.cycles-pp.sched_clock
      0.00            +0.9        0.88 ±  6%  perf-profile.children.cycles-pp.native_apic_msr_eoi
      0.00            +1.0        0.97 ± 15%  perf-profile.children.cycles-pp._raw_spin_lock
      0.00            +1.0        1.00 ± 12%  perf-profile.children.cycles-pp.tick_nohz_next_event
      0.00            +1.0        1.02 ± 15%  perf-profile.children.cycles-pp.sched_clock_cpu
      0.00            +1.1        1.06 ±  8%  perf-profile.children.cycles-pp.update_sd_lb_stats
      0.00            +1.1        1.09 ± 14%  perf-profile.children.cycles-pp.native_sched_clock
      0.00            +1.1        1.11 ±  7%  perf-profile.children.cycles-pp.find_busiest_group
      0.00            +1.1        1.14 ± 26%  perf-profile.children.cycles-pp.tick_irq_enter
      0.00            +1.2        1.22 ±  3%  perf-profile.children.cycles-pp.lapic_next_deadline
      0.00            +1.3        1.26 ± 60%  perf-profile.children.cycles-pp.tick_sched_do_timer
      0.00            +1.3        1.27 ±  8%  perf-profile.children.cycles-pp.read_tsc
      0.00            +1.5        1.46 ± 23%  perf-profile.children.cycles-pp.irq_enter_rcu
      0.00            +1.5        1.54 ± 17%  perf-profile.children.cycles-pp.x86_64_start_kernel
      0.00            +1.5        1.54 ± 17%  perf-profile.children.cycles-pp.x86_64_start_reservations
      0.00            +1.5        1.54 ± 17%  perf-profile.children.cycles-pp.start_kernel
      0.00            +1.5        1.54 ± 17%  perf-profile.children.cycles-pp.arch_call_rest_init
      0.00            +1.5        1.54 ± 17%  perf-profile.children.cycles-pp.rest_init
      0.00            +1.6        1.56 ± 12%  perf-profile.children.cycles-pp.perf_rotate_context
      0.00            +1.6        1.62 ±  2%  perf-profile.children.cycles-pp.clockevents_program_event
      0.00            +1.6        1.64 ± 12%  perf-profile.children.cycles-pp.tick_nohz_get_sleep_length
      0.00            +1.7        1.68 ± 15%  perf-profile.children.cycles-pp.ktime_get_update_offsets_now
      0.00            +1.7        1.74 ±  5%  perf-profile.children.cycles-pp.load_balance
      0.00            +1.8        1.76 ± 11%  perf-profile.children.cycles-pp.native_irq_return_iret
      0.00            +2.2        2.17 ± 23%  perf-profile.children.cycles-pp.ktime_get
      0.00            +2.3        2.35 ±  9%  perf-profile.children.cycles-pp.perf_mux_hrtimer_handler
      0.00            +2.4        2.36 ± 12%  perf-profile.children.cycles-pp.arch_scale_freq_tick
      0.00            +2.4        2.45 ±  4%  perf-profile.children.cycles-pp.rebalance_domains
      0.00            +3.3        3.27 ±  5%  perf-profile.children.cycles-pp.menu_select
      0.00            +4.0        3.99 ±  8%  perf-profile.children.cycles-pp.__do_softirq
      0.00            +4.5        4.54 ±  7%  perf-profile.children.cycles-pp.__irq_exit_rcu
      0.05 ±  8%      +9.2        9.29 ±  7%  perf-profile.children.cycles-pp.__intel_pmu_enable_all
      0.08 ± 10%     +12.6       12.65 ±  5%  perf-profile.children.cycles-pp.perf_adjust_freq_unthr_context
      0.08 ±  8%     +12.7       12.75 ±  6%  perf-profile.children.cycles-pp.perf_event_task_tick
      0.16 ±  3%     +16.8       17.00 ±  5%  perf-profile.children.cycles-pp.scheduler_tick
      0.18 ±  2%     +18.6       18.80 ±  5%  perf-profile.children.cycles-pp.update_process_times
      0.18 ±  2%     +18.8       18.94 ±  5%  perf-profile.children.cycles-pp.tick_sched_handle
      0.19 ±  3%     +20.7       20.87 ±  8%  perf-profile.children.cycles-pp.tick_sched_timer
      0.22 ±  2%     +24.5       24.69 ±  6%  perf-profile.children.cycles-pp.__hrtimer_run_queues
      0.25 ±  3%     +28.5       28.72 ±  6%  perf-profile.children.cycles-pp.hrtimer_interrupt
      0.25 ±  2%     +29.5       29.71 ±  6%  perf-profile.children.cycles-pp.__sysvec_apic_timer_interrupt
      0.26           +37.1       37.37 ±  4%  perf-profile.children.cycles-pp.sysvec_apic_timer_interrupt
      0.29 ±  3%     +80.7       81.02        perf-profile.children.cycles-pp.asm_sysvec_apic_timer_interrupt
      0.00           +88.3       88.31        perf-profile.children.cycles-pp.acpi_safe_halt
      0.00           +88.5       88.50        perf-profile.children.cycles-pp.acpi_idle_enter
      0.00           +90.0       89.96        perf-profile.children.cycles-pp.cpuidle_enter_state
      0.00           +90.4       90.40        perf-profile.children.cycles-pp.cpuidle_enter
      0.00           +94.5       94.46        perf-profile.children.cycles-pp.start_secondary
      0.00           +95.1       95.07        perf-profile.children.cycles-pp.cpuidle_idle_call
      0.00           +96.0       96.00        perf-profile.children.cycles-pp.secondary_startup_64_no_verify
      0.00           +96.0       96.00        perf-profile.children.cycles-pp.cpu_startup_entry
      0.00           +96.0       96.00        perf-profile.children.cycles-pp.do_idle
     26.75           -26.8        0.00        perf-profile.self.cycles-pp.page_counter_uncharge
      9.40 ±  5%      -9.4        0.00        perf-profile.self.cycles-pp.propagate_protected_usage
      0.06 ±  6%      +0.1        0.12 ± 28%  perf-profile.self.cycles-pp._raw_spin_unlock_irqrestore
      0.00            +0.1        0.07 ± 25%  perf-profile.self.cycles-pp.__update_blocked_fair
      0.00            +0.1        0.08 ± 35%  perf-profile.self.cycles-pp.can_stop_idle_tick
      0.00            +0.1        0.08 ± 35%  perf-profile.self.cycles-pp.update_blocked_averages
      0.00            +0.1        0.09 ± 36%  perf-profile.self.cycles-pp.intel_pmu_disable_all
      0.00            +0.1        0.10 ± 43%  perf-profile.self.cycles-pp.rb_next
      0.00            +0.1        0.10 ± 25%  perf-profile.self.cycles-pp.tick_nohz_idle_retain_tick
      0.00            +0.1        0.10 ± 17%  perf-profile.self.cycles-pp.hrtimer_update_next_event
      0.00            +0.1        0.10 ± 16%  perf-profile.self.cycles-pp.tsc_verify_tsc_adjust
      0.00            +0.1        0.10 ± 38%  perf-profile.self.cycles-pp.menu_reflect
      0.00            +0.1        0.11 ± 46%  perf-profile.self.cycles-pp.clockevents_program_event
      0.00            +0.1        0.11 ± 10%  perf-profile.self.cycles-pp.__update_load_avg_cfs_rq
      0.00            +0.1        0.11 ± 33%  perf-profile.self.cycles-pp.tick_nohz_get_sleep_length
      0.00            +0.1        0.11 ± 21%  perf-profile.self.cycles-pp.update_rq_clock
      0.00            +0.1        0.11 ± 39%  perf-profile.self.cycles-pp.update_rt_rq_load_avg
      0.00            +0.1        0.11 ± 27%  perf-profile.self.cycles-pp.sched_clock_cpu
      0.00            +0.1        0.11 ± 23%  perf-profile.self.cycles-pp.irqentry_enter
      0.00            +0.1        0.12 ± 19%  perf-profile.self.cycles-pp.__irq_exit_rcu
      0.00            +0.1        0.12 ± 23%  perf-profile.self.cycles-pp.hrtimer_forward
      0.00            +0.1        0.12 ± 46%  perf-profile.self.cycles-pp.copy_mc_enhanced_fast_string
      0.00            +0.1        0.12 ± 17%  perf-profile.self.cycles-pp.cpuidle_governor_latency_req
      0.00            +0.1        0.12 ± 41%  perf-profile.self.cycles-pp._dl_addr
      0.00            +0.1        0.12 ± 14%  perf-profile.self.cycles-pp.run_posix_cpu_timers
      0.00            +0.1        0.12 ± 37%  perf-profile.self.cycles-pp.timerqueue_del
      0.00            +0.1        0.13 ± 16%  perf-profile.self.cycles-pp.__sysvec_apic_timer_interrupt
      0.00            +0.1        0.13 ± 40%  perf-profile.self.cycles-pp.perf_mux_hrtimer_handler
      0.00            +0.1        0.13 ± 32%  perf-profile.self.cycles-pp.tick_sched_timer
      0.00            +0.1        0.14 ± 26%  perf-profile.self.cycles-pp.rebalance_domains
      0.00            +0.1        0.14 ± 31%  perf-profile.self.cycles-pp.__do_softirq
      0.00            +0.1        0.14 ± 21%  perf-profile.self.cycles-pp.note_gp_changes
      0.00            +0.2        0.15 ± 27%  perf-profile.self.cycles-pp.ct_kernel_enter
      0.00            +0.2        0.16 ± 40%  perf-profile.self.cycles-pp.cpuidle_reflect
      0.00            +0.2        0.16 ± 14%  perf-profile.self.cycles-pp.cpu_util
      0.00            +0.2        0.16 ± 36%  perf-profile.self.cycles-pp.acpi_idle_enter
      0.00            +0.2        0.16 ± 21%  perf-profile.self.cycles-pp.load_balance
      0.00            +0.2        0.16 ± 18%  perf-profile.self.cycles-pp.update_sd_lb_stats
      0.00            +0.2        0.17 ± 26%  perf-profile.self.cycles-pp.rcu_sched_clock_irq
      0.00            +0.2        0.17 ± 16%  perf-profile.self.cycles-pp.get_next_timer_interrupt
      0.00            +0.2        0.18 ± 12%  perf-profile.self.cycles-pp.call_cpuidle
      0.00            +0.2        0.18 ± 23%  perf-profile.self.cycles-pp.ct_nmi_enter
      0.00            +0.2        0.19 ± 32%  perf-profile.self.cycles-pp.irqtime_account_irq
      0.00            +0.2        0.19 ± 25%  perf-profile.self.cycles-pp.idle_cpu
      0.00            +0.2        0.20 ± 27%  perf-profile.self.cycles-pp.check_cpu_stall
      0.00            +0.2        0.20 ± 30%  perf-profile.self.cycles-pp.trigger_load_balance
      0.00            +0.2        0.20 ± 27%  perf-profile.self.cycles-pp.error_entry
      0.00            +0.2        0.20 ± 31%  perf-profile.self.cycles-pp.perf_pmu_nop_void
      0.00            +0.2        0.21 ± 16%  perf-profile.self.cycles-pp.__memcpy
      0.00            +0.2        0.21 ± 27%  perf-profile.self.cycles-pp.hrtimer_interrupt
      0.00            +0.2        0.21 ± 34%  perf-profile.self.cycles-pp._find_next_and_bit
      0.00            +0.2        0.21 ± 43%  perf-profile.self.cycles-pp.update_rq_clock_task
      0.00            +0.2        0.24 ± 25%  perf-profile.self.cycles-pp.tick_check_broadcast_expired
      0.00            +0.3        0.26 ± 14%  perf-profile.self.cycles-pp.rcu_pending
      0.00            +0.3        0.28 ± 33%  perf-profile.self.cycles-pp.update_process_times
      0.00            +0.3        0.28 ± 42%  perf-profile.self.cycles-pp.perf_rotate_context
      0.00            +0.3        0.29 ± 17%  perf-profile.self.cycles-pp.timerqueue_add
      0.00            +0.3        0.30 ± 24%  perf-profile.self.cycles-pp.__hrtimer_next_event_base
      0.00            +0.3        0.32 ± 22%  perf-profile.self.cycles-pp.irq_enter_rcu
      0.00            +0.3        0.32 ± 28%  perf-profile.self.cycles-pp.scheduler_tick
      0.30 ±  2%      +0.3        0.62 ± 13%  perf-profile.self.cycles-pp._raw_spin_lock_irqsave
      0.00            +0.3        0.33 ± 12%  perf-profile.self.cycles-pp.irq_work_tick
      0.00            +0.4        0.35 ± 17%  perf-profile.self.cycles-pp.tick_nohz_next_event
      0.00            +0.4        0.36 ± 22%  perf-profile.self.cycles-pp.x86_pmu_disable
      0.00            +0.4        0.36 ± 18%  perf-profile.self.cycles-pp.do_idle
      0.00            +0.4        0.37 ± 26%  perf-profile.self.cycles-pp.__hrtimer_run_queues
      0.00            +0.4        0.39 ± 45%  perf-profile.self.cycles-pp.calc_global_load_tick
      0.00            +0.4        0.39 ± 23%  perf-profile.self.cycles-pp.tick_check_oneshot_broadcast_this_cpu
      0.00            +0.4        0.40 ± 15%  perf-profile.self.cycles-pp.get_cpu_device
      0.00            +0.4        0.41 ± 21%  perf-profile.self.cycles-pp.update_irq_load_avg
      0.00            +0.4        0.43 ± 14%  perf-profile.self.cycles-pp.cpuidle_enter
      0.00            +0.4        0.44 ± 22%  perf-profile.self.cycles-pp.ct_kernel_exit_state
      0.00            +0.5        0.49 ± 15%  perf-profile.self.cycles-pp.cpuidle_enter_state
      0.00            +0.5        0.51 ± 14%  perf-profile.self.cycles-pp.sysvec_apic_timer_interrupt
      0.00            +0.6        0.58 ±  9%  perf-profile.self.cycles-pp.update_sg_lb_stats
      0.00            +0.7        0.66 ± 10%  perf-profile.self.cycles-pp.asm_sysvec_apic_timer_interrupt
      0.00            +0.7        0.69 ±  8%  perf-profile.self.cycles-pp.cpuidle_idle_call
      0.00            +0.9        0.88 ±  6%  perf-profile.self.cycles-pp.native_apic_msr_eoi
      0.00            +0.9        0.90 ± 13%  perf-profile.self.cycles-pp._raw_spin_lock
      0.00            +1.0        1.00 ±  9%  perf-profile.self.cycles-pp.menu_select
      0.00            +1.0        1.02 ± 67%  perf-profile.self.cycles-pp.tick_sched_do_timer
      0.00            +1.0        1.04 ± 14%  perf-profile.self.cycles-pp.native_sched_clock
      0.00            +1.1        1.10 ± 41%  perf-profile.self.cycles-pp.ktime_get
      0.00            +1.2        1.22 ±  4%  perf-profile.self.cycles-pp.lapic_next_deadline
      0.00            +1.2        1.24 ±  9%  perf-profile.self.cycles-pp.read_tsc
      0.00            +1.5        1.48 ± 14%  perf-profile.self.cycles-pp.ktime_get_update_offsets_now
      0.00            +1.8        1.76 ± 11%  perf-profile.self.cycles-pp.native_irq_return_iret
      0.00            +2.4        2.35 ± 12%  perf-profile.self.cycles-pp.arch_scale_freq_tick
      0.00            +3.7        3.69 ±  5%  perf-profile.self.cycles-pp.perf_adjust_freq_unthr_context
      0.05 ±  8%      +9.2        9.29 ±  7%  perf-profile.self.cycles-pp.__intel_pmu_enable_all
      0.00           +48.2       48.17 ±  2%  perf-profile.self.cycles-pp.acpi_safe_halt


***************************************************************************************************
lkp-csl-d02: 36 threads 1 sockets Intel(R) Core(TM) i9-10980XE CPU @ 3.00GHz (Cascade Lake) with 128G memory
=========================================================================================
class/compiler/cpufreq_governor/disk/fs/kconfig/nr_threads/rootfs/tbox_group/test/testcase/testtime:
  os/gcc-12/performance/1HDD/ext4/x86_64-rhel-8.3/1/debian-11.1-x86_64-20220510.cgz/lkp-csl-d02/splice/stress-ng/60s

commit: 
  19e3e6cdfd ("accessibility: speakup: refactor deprecated strncpy")
  1b057bd800 ("drivers/char/mem: implement splice() for /dev/zero, /dev/full")

19e3e6cdfdc73400 1b057bd800c3ea0c926191d7950 
---------------- --------------------------- 
         %stddev     %change         %stddev
             \          |                \  
      2.98            -3.3%       2.88        iostat.cpu.system
      0.51            +0.1        0.60 ±  2%  mpstat.cpu.all.usr%
      0.40           -10.1%       0.36 ±  3%  turbostat.IPC
      0.01 ± 55%    -100.0%       0.00        perf-sched.wait_time.avg.ms.__cond_resched.__alloc_pages.pipe_write.vfs_write.ksys_write
      0.02 ± 56%    -100.0%       0.00        perf-sched.wait_time.max.ms.__cond_resched.__alloc_pages.pipe_write.vfs_write.ksys_write
     65093           +70.3%     110836        stress-ng.splice.MB_per_sec_splice_rate
  10307909           +38.9%   14322626 ±  4%  stress-ng.splice.ops
    171798           +38.9%     238710 ±  4%  stress-ng.splice.ops_per_sec
  1.71e+08           +39.3%  2.383e+08 ±  2%  proc-vmstat.numa_hit
 1.751e+08 ±  2%     +35.5%  2.372e+08 ±  2%  proc-vmstat.numa_local
 1.652e+08           +43.1%  2.364e+08 ±  2%  proc-vmstat.pgalloc_normal
 1.652e+08           +43.1%  2.363e+08 ±  2%  proc-vmstat.pgfree
      0.05 ±  4%     +10.8%       0.06 ±  2%  perf-stat.i.MPKI
 1.277e+09           -15.1%  1.085e+09 ±  2%  perf-stat.i.branch-instructions
      0.71            +0.3        0.96        perf-stat.i.branch-miss-rate%
  10496910           +14.8%   12053487 ±  2%  perf-stat.i.branch-misses
      0.75           +13.7%       0.85 ±  2%  perf-stat.i.cpi
      0.00 ±  7%      +0.0        0.00 ±  6%  perf-stat.i.dTLB-load-miss-rate%
 1.831e+09           -15.0%  1.557e+09 ±  2%  perf-stat.i.dTLB-loads
 1.124e+09           -11.5%   9.94e+08 ±  2%  perf-stat.i.dTLB-stores
     89.35            +1.7       91.10        perf-stat.i.iTLB-load-miss-rate%
   4833975 ±  5%     +26.6%    6118619 ±  9%  perf-stat.i.iTLB-load-misses
 6.665e+09           -11.6%  5.892e+09 ±  2%  perf-stat.i.instructions
      1534 ±  4%     -25.7%       1140 ±  5%  perf-stat.i.instructions-per-iTLB-miss
      1.34           -11.8%       1.18 ±  3%  perf-stat.i.ipc
    117.54           -14.1%     100.98 ±  2%  perf-stat.i.metric.M/sec
      0.06 ±  3%     +16.3%       0.07 ±  2%  perf-stat.overall.MPKI
      0.82            +0.3        1.11        perf-stat.overall.branch-miss-rate%
      0.75           +13.2%       0.84 ±  2%  perf-stat.overall.cpi
      0.00 ±  6%      +0.0        0.00 ±  6%  perf-stat.overall.dTLB-load-miss-rate%
      0.00 ±  5%      +0.0        0.00 ±  5%  perf-stat.overall.dTLB-store-miss-rate%
     90.21            +1.9       92.08        perf-stat.overall.iTLB-load-miss-rate%
      1382 ±  5%     -29.9%     968.76 ±  6%  perf-stat.overall.instructions-per-iTLB-miss
      1.34           -11.6%       1.19 ±  3%  perf-stat.overall.ipc
 1.257e+09           -15.1%  1.068e+09 ±  2%  perf-stat.ps.branch-instructions
  10335583           +14.8%   11864548 ±  2%  perf-stat.ps.branch-misses
 1.802e+09           -15.0%  1.532e+09 ±  2%  perf-stat.ps.dTLB-loads
 1.106e+09           -11.5%  9.783e+08 ±  2%  perf-stat.ps.dTLB-stores
   4757545 ±  5%     +26.6%    6022044 ±  9%  perf-stat.ps.iTLB-load-misses
 6.559e+09           -11.6%  5.799e+09 ±  2%  perf-stat.ps.instructions
 4.149e+11           -11.6%  3.668e+11 ±  2%  perf-stat.total.instructions
     29.20 ± 11%     -29.2        0.00        perf-profile.calltrace.cycles-pp.write
     28.74 ± 11%     -28.7        0.00        perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.write
     28.67 ± 11%     -28.7        0.00        perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.write
     28.57 ± 11%     -28.6        0.00        perf-profile.calltrace.cycles-pp.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe.write
     28.41 ± 11%     -28.4        0.00        perf-profile.calltrace.cycles-pp.vfs_write.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe.write
     27.88 ± 11%     -27.9        0.00        perf-profile.calltrace.cycles-pp.pipe_write.vfs_write.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe
     11.96 ± 11%     -12.0        0.00        perf-profile.calltrace.cycles-pp.copy_page_from_iter.pipe_write.vfs_write.ksys_write.do_syscall_64
     11.62 ± 12%     -11.6        0.00        perf-profile.calltrace.cycles-pp._copy_from_iter.copy_page_from_iter.pipe_write.vfs_write.ksys_write
     11.58 ± 11%     -11.6        0.00        perf-profile.calltrace.cycles-pp.__alloc_pages.pipe_write.vfs_write.ksys_write.do_syscall_64
      9.48 ± 12%      -9.5        0.00        perf-profile.calltrace.cycles-pp.copyin._copy_from_iter.copy_page_from_iter.pipe_write.vfs_write
      9.47 ±  9%      -9.5        0.00        perf-profile.calltrace.cycles-pp.__mem_cgroup_uncharge.__folio_put.__splice_from_pipe.splice_from_pipe.do_splice
      9.90 ±  9%      -8.9        0.97 ± 17%  perf-profile.calltrace.cycles-pp.__folio_put.__splice_from_pipe.splice_from_pipe.do_splice.__do_splice
      8.42 ± 11%      -8.4        0.00        perf-profile.calltrace.cycles-pp.rep_movs_alternative.copyin._copy_from_iter.copy_page_from_iter.pipe_write
      7.82 ±  9%      -7.8        0.00        perf-profile.calltrace.cycles-pp.uncharge_batch.__mem_cgroup_uncharge.__folio_put.__splice_from_pipe.splice_from_pipe
     16.11 ±  9%      -6.1        9.97 ±  8%  perf-profile.calltrace.cycles-pp.__splice_from_pipe.splice_from_pipe.do_splice.__do_splice.__x64_sys_splice
     16.43 ±  9%      -6.0       10.44 ±  7%  perf-profile.calltrace.cycles-pp.splice_from_pipe.do_splice.__do_splice.__x64_sys_splice.do_syscall_64
      5.34 ± 12%      -5.3        0.00        perf-profile.calltrace.cycles-pp.get_page_from_freelist.__alloc_pages.pipe_write.vfs_write.ksys_write
      0.41 ± 71%      +0.4        0.81 ± 16%  perf-profile.calltrace.cycles-pp.pipe_double_lock.splice_pipe_to_pipe.do_splice.__do_splice.__x64_sys_splice
      1.01 ±  8%      +0.4        1.45 ±  4%  perf-profile.calltrace.cycles-pp.free_unref_page_prepare.free_unref_page.__splice_from_pipe.splice_from_pipe.do_splice
      1.02 ± 14%      +0.5        1.48 ± 11%  perf-profile.calltrace.cycles-pp._raw_spin_trylock.free_unref_page.__splice_from_pipe.splice_from_pipe.do_splice
      0.18 ±141%      +0.5        0.65 ± 11%  perf-profile.calltrace.cycles-pp.get_pfnblock_flags_mask.free_unref_page_prepare.free_unref_page.__splice_from_pipe.splice_from_pipe
      0.28 ±100%      +0.5        0.75 ± 16%  perf-profile.calltrace.cycles-pp.mutex_lock.pipe_double_lock.splice_pipe_to_pipe.do_splice.__do_splice
      0.86 ± 19%      +0.5        1.33 ± 11%  perf-profile.calltrace.cycles-pp.free_unref_page_commit.free_unref_page.__splice_from_pipe.splice_from_pipe.do_splice
      0.17 ±141%      +0.5        0.69 ± 17%  perf-profile.calltrace.cycles-pp.syscall_return_via_sysret.splice
      0.08 ±223%      +0.5        0.62 ± 10%  perf-profile.calltrace.cycles-pp.syscall_exit_to_user_mode.do_syscall_64.entry_SYSCALL_64_after_hwframe.splice
      1.36 ± 18%      +0.7        2.03 ±  9%  perf-profile.calltrace.cycles-pp.__fget_light.__x64_sys_splice.do_syscall_64.entry_SYSCALL_64_after_hwframe.splice
      0.00            +0.8        0.80 ±  9%  perf-profile.calltrace.cycles-pp.__kmem_cache_alloc_node.__kmalloc.copy_splice_read.splice_file_to_pipe.do_splice
      0.00            +1.1        1.06 ±  7%  perf-profile.calltrace.cycles-pp.__kmalloc.copy_splice_read.splice_file_to_pipe.do_splice.__do_splice
      0.09 ±223%      +1.1        1.20 ±  8%  perf-profile.calltrace.cycles-pp.vfs_splice_read.splice_file_to_pipe.do_splice.__do_splice.__x64_sys_splice
      3.43 ± 10%      +1.4        4.79 ±  7%  perf-profile.calltrace.cycles-pp.splice_pipe_to_pipe.do_splice.__do_splice.__x64_sys_splice.do_syscall_64
      3.59 ± 12%      +1.8        5.37 ±  6%  perf-profile.calltrace.cycles-pp.free_unref_page.__splice_from_pipe.splice_from_pipe.do_splice.__do_splice
      4.06 ± 13%      +2.1        6.12 ±  5%  perf-profile.calltrace.cycles-pp.__entry_text_start.splice
      0.00            +2.2        2.20 ± 10%  perf-profile.calltrace.cycles-pp.generic_pipe_buf_release.__splice_from_pipe.splice_from_pipe.do_splice.__do_splice
      0.00            +2.3        2.33 ±  4%  perf-profile.calltrace.cycles-pp.__alloc_pages_bulk.copy_splice_read.splice_file_to_pipe.do_splice.__do_splice
      0.00           +18.4       18.44 ± 11%  perf-profile.calltrace.cycles-pp.iov_iter_zero.read_iter_zero.copy_splice_read.splice_file_to_pipe.do_splice
      0.00           +19.0       19.02 ± 10%  perf-profile.calltrace.cycles-pp.read_iter_zero.copy_splice_read.splice_file_to_pipe.do_splice.__do_splice
     21.46 ±  9%     +22.8       44.25 ±  6%  perf-profile.calltrace.cycles-pp.do_splice.__do_splice.__x64_sys_splice.do_syscall_64.entry_SYSCALL_64_after_hwframe
     22.01 ±  9%     +23.1       45.09 ±  6%  perf-profile.calltrace.cycles-pp.__do_splice.__x64_sys_splice.do_syscall_64.entry_SYSCALL_64_after_hwframe.splice
     24.22 ±  9%     +24.3       48.48 ±  6%  perf-profile.calltrace.cycles-pp.__x64_sys_splice.do_syscall_64.entry_SYSCALL_64_after_hwframe.splice
     25.33 ±  9%     +24.8       50.10 ±  6%  perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.splice
     26.16 ±  9%     +25.2       51.36 ±  6%  perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.splice
      0.00           +25.2       25.21 ±  8%  perf-profile.calltrace.cycles-pp.copy_splice_read.splice_file_to_pipe.do_splice.__do_splice.__x64_sys_splice
      0.54 ± 46%     +26.7       27.26 ±  7%  perf-profile.calltrace.cycles-pp.splice_file_to_pipe.do_splice.__do_splice.__x64_sys_splice.do_syscall_64
     32.15 ± 10%     +27.9       60.02 ±  5%  perf-profile.calltrace.cycles-pp.splice
     29.33 ± 11%     -29.3        0.00        perf-profile.children.cycles-pp.write
     28.59 ± 11%     -28.6        0.00        perf-profile.children.cycles-pp.ksys_write
     28.44 ± 11%     -28.4        0.00        perf-profile.children.cycles-pp.vfs_write
     28.01 ± 11%     -28.0        0.00        perf-profile.children.cycles-pp.pipe_write
     12.00 ± 11%     -12.0        0.00        perf-profile.children.cycles-pp.copy_page_from_iter
     11.30 ± 11%     -11.3        0.00        perf-profile.children.cycles-pp._copy_from_iter
     11.73 ± 11%     -11.1        0.59 ±  9%  perf-profile.children.cycles-pp.__alloc_pages
     10.36 ± 11%     -10.4        0.00        perf-profile.children.cycles-pp.copyin
      9.51 ±  9%      -9.1        0.37 ± 16%  perf-profile.children.cycles-pp.__mem_cgroup_uncharge
     10.02 ±  9%      -8.9        1.12 ± 16%  perf-profile.children.cycles-pp.__folio_put
      8.62 ± 11%      -8.6        0.00        perf-profile.children.cycles-pp.rep_movs_alternative
      8.03 ±  9%      -8.0        0.00        perf-profile.children.cycles-pp.uncharge_batch
     16.19 ±  9%      -6.1       10.09 ±  8%  perf-profile.children.cycles-pp.__splice_from_pipe
     16.44 ±  9%      -6.0       10.45 ±  7%  perf-profile.children.cycles-pp.splice_from_pipe
      5.42 ± 12%      -5.0        0.38 ±  8%  perf-profile.children.cycles-pp.get_page_from_freelist
      4.21 ± 11%      -3.9        0.27 ± 18%  perf-profile.children.cycles-pp.rmqueue
      2.85 ±  9%      -1.1        1.74 ± 10%  perf-profile.children.cycles-pp._raw_spin_trylock
      1.04 ± 12%      -1.0        0.06 ± 59%  perf-profile.children.cycles-pp._raw_spin_lock_irq
      0.23 ± 14%      -0.1        0.15 ± 33%  perf-profile.children.cycles-pp._raw_spin_lock_irqsave
      0.06 ± 50%      +0.1        0.11 ± 18%  perf-profile.children.cycles-pp.pipe_clear_nowait
      0.23 ±  7%      +0.1        0.32 ± 11%  perf-profile.children.cycles-pp.__list_del_entry_valid_or_report
      0.10 ± 17%      +0.1        0.20 ± 15%  perf-profile.children.cycles-pp.clock_gettime
      0.28 ± 15%      +0.1        0.39 ± 13%  perf-profile.children.cycles-pp.syscall_enter_from_user_mode
      0.40 ± 14%      +0.2        0.55 ± 14%  perf-profile.children.cycles-pp.entry_SYSCALL_64_safe_stack
      0.35 ± 15%      +0.2        0.53 ± 13%  perf-profile.children.cycles-pp.stress_splice
      0.28 ± 18%      +0.2        0.46 ± 10%  perf-profile.children.cycles-pp.get_pipe_info
      0.00            +0.2        0.18 ± 14%  perf-profile.children.cycles-pp.kmalloc_slab
      0.62 ± 11%      +0.2        0.82 ±  7%  perf-profile.children.cycles-pp.syscall_exit_to_user_mode
      1.02 ±  8%      +0.2        1.23 ± 10%  perf-profile.children.cycles-pp.mutex_unlock
      0.00            +0.2        0.24 ± 23%  perf-profile.children.cycles-pp.kfree
      0.53 ± 19%      +0.2        0.77 ± 14%  perf-profile.children.cycles-pp.syscall_return_via_sysret
      0.59 ± 13%      +0.2        0.84 ± 15%  perf-profile.children.cycles-pp.pipe_double_lock
      0.00            +0.3        0.31 ± 18%  perf-profile.children.cycles-pp.memset_orig
      0.52 ± 22%      +0.3        0.85 ±  7%  perf-profile.children.cycles-pp.apparmor_file_permission
      0.00            +0.4        0.39 ± 18%  perf-profile.children.cycles-pp.__kmem_cache_free
      1.06 ±  9%      +0.4        1.46 ±  4%  perf-profile.children.cycles-pp.free_unref_page_prepare
      0.60 ± 18%      +0.4        1.00 ±  7%  perf-profile.children.cycles-pp.security_file_permission
      0.96 ± 17%      +0.5        1.48 ± 10%  perf-profile.children.cycles-pp.free_unref_page_commit
      0.24 ± 19%      +0.5        0.76 ± 11%  perf-profile.children.cycles-pp.__fsnotify_parent
      1.17 ±  7%      +0.5        1.72 ± 10%  perf-profile.children.cycles-pp.mutex_lock
      1.44 ± 18%      +0.6        2.04 ±  9%  perf-profile.children.cycles-pp.__fget_light
      1.78 ± 10%      +0.7        2.48 ±  7%  perf-profile.children.cycles-pp.entry_SYSRETQ_unsafe_stack
      0.41 ± 18%      +0.8        1.21 ±  9%  perf-profile.children.cycles-pp.vfs_splice_read
      0.00            +0.8        0.83 ± 10%  perf-profile.children.cycles-pp.__kmem_cache_alloc_node
      2.96 ± 12%      +1.0        3.97 ±  5%  perf-profile.children.cycles-pp.__entry_text_start
      0.00            +1.1        1.07 ±  8%  perf-profile.children.cycles-pp.__kmalloc
      3.46 ± 10%      +1.3        4.80 ±  7%  perf-profile.children.cycles-pp.splice_pipe_to_pipe
      3.68 ± 11%      +1.8        5.43 ±  6%  perf-profile.children.cycles-pp.free_unref_page
      0.00            +2.3        2.27 ± 11%  perf-profile.children.cycles-pp.generic_pipe_buf_release
      0.00            +2.3        2.34 ±  4%  perf-profile.children.cycles-pp.__alloc_pages_bulk
      0.00           +18.5       18.47 ± 11%  perf-profile.children.cycles-pp.iov_iter_zero
      0.00           +19.1       19.09 ± 10%  perf-profile.children.cycles-pp.read_iter_zero
     21.49 ±  9%     +22.8       44.32 ±  6%  perf-profile.children.cycles-pp.do_splice
     22.11 ±  9%     +23.2       45.27 ±  6%  perf-profile.children.cycles-pp.__do_splice
     24.25 ±  9%     +24.3       48.54 ±  6%  perf-profile.children.cycles-pp.__x64_sys_splice
      0.00           +25.2       25.24 ±  8%  perf-profile.children.cycles-pp.copy_splice_read
      0.62 ± 17%     +26.7       27.29 ±  7%  perf-profile.children.cycles-pp.splice_file_to_pipe
     32.06 ± 10%     +28.0       60.02 ±  5%  perf-profile.children.cycles-pp.splice
      8.29 ± 11%      -8.3        0.00        perf-profile.self.cycles-pp.rep_movs_alternative
      2.16 ± 17%      -2.0        0.11 ± 10%  perf-profile.self.cycles-pp.rmqueue
      1.56 ± 18%      -1.4        0.17 ± 19%  perf-profile.self.cycles-pp.__alloc_pages
      1.25 ± 17%      -1.1        0.11 ± 31%  perf-profile.self.cycles-pp.get_page_from_freelist
      2.79 ±  9%      -1.1        1.72 ±  9%  perf-profile.self.cycles-pp._raw_spin_trylock
      1.00 ± 12%      -1.0        0.05 ± 79%  perf-profile.self.cycles-pp._raw_spin_lock_irq
      0.06 ± 52%      +0.0        0.09 ± 26%  perf-profile.self.cycles-pp.__hrtimer_next_event_base
      0.06 ± 49%      +0.0        0.10 ± 17%  perf-profile.self.cycles-pp.pipe_clear_nowait
      0.08 ± 19%      +0.1        0.18 ± 16%  perf-profile.self.cycles-pp.clock_gettime
      0.28 ± 15%      +0.1        0.38 ± 13%  perf-profile.self.cycles-pp.syscall_enter_from_user_mode
      0.21 ± 18%      +0.1        0.32 ± 11%  perf-profile.self.cycles-pp.get_pipe_info
      0.33 ± 13%      +0.1        0.44 ± 10%  perf-profile.self.cycles-pp.do_syscall_64
      0.17 ±  8%      +0.1        0.31 ±  9%  perf-profile.self.cycles-pp.__list_del_entry_valid_or_report
      0.30 ± 20%      +0.1        0.43 ± 15%  perf-profile.self.cycles-pp.stress_splice
      0.16 ± 32%      +0.1        0.29 ±  8%  perf-profile.self.cycles-pp.stress_time_now_timespec
      0.30 ± 16%      +0.1        0.44 ±  9%  perf-profile.self.cycles-pp.syscall_exit_to_user_mode
      0.40 ± 14%      +0.2        0.55 ± 14%  perf-profile.self.cycles-pp.entry_SYSCALL_64_safe_stack
      0.00            +0.2        0.18 ± 12%  perf-profile.self.cycles-pp.kmalloc_slab
      0.01 ±223%      +0.2        0.19 ± 14%  perf-profile.self.cycles-pp.vfs_splice_read
      0.49 ± 19%      +0.2        0.71 ± 13%  perf-profile.self.cycles-pp.syscall_return_via_sysret
      0.00            +0.2        0.23 ± 25%  perf-profile.self.cycles-pp.kfree
      0.58 ±  9%      +0.2        0.82 ±  7%  perf-profile.self.cycles-pp.free_unref_page_prepare
      0.43 ± 17%      +0.3        0.68 ±  6%  perf-profile.self.cycles-pp.__do_splice
      0.03 ±100%      +0.3        0.31 ± 17%  perf-profile.self.cycles-pp.splice_file_to_pipe
      0.66 ±  4%      +0.3        0.94 ± 10%  perf-profile.self.cycles-pp.free_unref_page
      0.43 ± 20%      +0.3        0.72 ± 10%  perf-profile.self.cycles-pp.apparmor_file_permission
      0.50 ± 20%      +0.3        0.80 ± 11%  perf-profile.self.cycles-pp.do_splice
      0.00            +0.3        0.30 ± 16%  perf-profile.self.cycles-pp.memset_orig
      0.91 ±  9%      +0.4        1.28 ± 12%  perf-profile.self.cycles-pp.mutex_lock
      1.40 ± 14%      +0.4        1.77 ±  5%  perf-profile.self.cycles-pp.__entry_text_start
      0.00            +0.4        0.39 ± 18%  perf-profile.self.cycles-pp.__kmem_cache_free
      0.98 ± 14%      +0.4        1.38 ±  8%  perf-profile.self.cycles-pp.entry_SYSCALL_64_after_hwframe
      0.72 ± 14%      +0.4        1.13 ± 15%  perf-profile.self.cycles-pp.__x64_sys_splice
      0.84 ± 14%      +0.4        1.28 ± 10%  perf-profile.self.cycles-pp.__splice_from_pipe
      0.00            +0.4        0.44 ± 15%  perf-profile.self.cycles-pp.__kmem_cache_alloc_node
      0.83 ± 17%      +0.5        1.31 ±  9%  perf-profile.self.cycles-pp.free_unref_page_commit
      0.24 ± 20%      +0.5        0.76 ± 12%  perf-profile.self.cycles-pp.__fsnotify_parent
      1.66 ± 14%      +0.5        2.20 ±  9%  perf-profile.self.cycles-pp.splice_pipe_to_pipe
      1.42 ± 17%      +0.6        1.99 ±  9%  perf-profile.self.cycles-pp.__fget_light
      1.74 ± 10%      +0.7        2.41 ±  8%  perf-profile.self.cycles-pp.entry_SYSRETQ_unsafe_stack
      0.00            +0.7        0.68 ± 10%  perf-profile.self.cycles-pp.read_iter_zero
      1.99 ± 13%      +1.1        3.08 ±  5%  perf-profile.self.cycles-pp.splice
      0.00            +1.3        1.33 ±  7%  perf-profile.self.cycles-pp.__alloc_pages_bulk
      0.00            +2.0        2.01 ±  7%  perf-profile.self.cycles-pp.copy_splice_read
      0.00            +2.2        2.22 ± 11%  perf-profile.self.cycles-pp.generic_pipe_buf_release
      0.00           +18.4       18.38 ± 11%  perf-profile.self.cycles-pp.iov_iter_zero





Disclaimer:
Results have been estimated based on internal Intel analysis and are provided
for informational purposes only. Any difference in system hardware or software
design or configuration may affect actual performance.


-- 
0-DAY CI Kernel Test Service
https://github.com/intel/lkp-tests/wiki



^ permalink raw reply	[flat|nested] 9+ messages in thread

* Re: [linux-next:master] [drivers/char/mem]  1b057bd800: stress-ng.splice.ops_per_sec -99.8% regression
  2023-10-17 15:06 [linux-next:master] [drivers/char/mem] 1b057bd800: stress-ng.splice.ops_per_sec -99.8% regression kernel test robot
@ 2023-10-17 16:56 ` Greg Kroah-Hartman
  2023-10-18  6:31   ` Max Kellermann
  2023-10-18  7:07   ` Oliver Sang
       [not found] ` <CAKPOu+_T8xk4yd2P4KT4j3eMoFqwYmkxqDHaFtv4Hii5-XyPuA@mail.gmail.com>
  1 sibling, 2 replies; 9+ messages in thread
From: Greg Kroah-Hartman @ 2023-10-17 16:56 UTC (permalink / raw)
  To: kernel test robot
  Cc: Max Kellermann, oe-lkp, lkp, Linux Memory Management List,
	linux-kernel, ying.huang, feng.tang, fengwei.yin

On Tue, Oct 17, 2023 at 11:06:42PM +0800, kernel test robot wrote:
> 
> 
> Hello,
> 
> kernel test robot noticed a -99.8% regression of stress-ng.splice.ops_per_sec on:
> 
> 
> commit: 1b057bd800c3ea0c926191d7950cd2365eddc9bb ("drivers/char/mem: implement splice() for /dev/zero, /dev/full")
> https://git.kernel.org/cgit/linux/kernel/git/next/linux-next.git master
> 
> testcase: stress-ng
> test machine: 64 threads 2 sockets Intel(R) Xeon(R) Gold 6346 CPU @ 3.10GHz (Ice Lake) with 256G memory
> parameters:
> 
> 	nr_threads: 100%
> 	testtime: 60s
> 	class: pipe
> 	test: splice
> 	cpufreq_governor: performance
> 
> 
> In addition to that, the commit also has significant impact on the following tests:
> 
> +------------------+-------------------------------------------------------------------------------------------------+
> | testcase: change | stress-ng: stress-ng.splice.ops_per_sec 38.9% improvement                                       |

So everything now goes faster, right?  -99.8% regression means 99.8%
faster?

thanks,

greg k-h


^ permalink raw reply	[flat|nested] 9+ messages in thread

* Re: [linux-next:master] [drivers/char/mem] 1b057bd800: stress-ng.splice.ops_per_sec -99.8% regression
  2023-10-17 16:56 ` Greg Kroah-Hartman
@ 2023-10-18  6:31   ` Max Kellermann
  2023-10-18  7:07   ` Oliver Sang
  1 sibling, 0 replies; 9+ messages in thread
From: Max Kellermann @ 2023-10-18  6:31 UTC (permalink / raw)
  To: Greg Kroah-Hartman
  Cc: kernel test robot, oe-lkp, lkp, Linux Memory Management List,
	linux-kernel, ying.huang, feng.tang, fengwei.yin

On Tue, Oct 17, 2023 at 6:57 PM Greg Kroah-Hartman
<gregkh@linuxfoundation.org> wrote:
> So everything now goes faster, right?  -99.8% regression means 99.8%
> faster?

That's what I thought, too, and sounds reasonable considering this
test is described as "stress copying of /dev/zero to /dev/null",
but... it's not what that test actually does. Contrary to the
description, it doesn't use /dev/zero at all, neither does it use
/dev/full. So it shouldn't be affected by my patch at all.

strace of that test's setup:

 pipe2([4, 5], 0)                = 0
 openat(AT_FDCWD, "/dev/null", O_WRONLY) = 6

Then it loops:

 vmsplice(5, [{iov_base="\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0"...,
iov_len=65536}], 1, 0) = 65536
 splice(4, NULL, 6, NULL, 65536, SPLICE_F_MOVE) = 65536
 write(5, "\334\360U\300~\361\20jV\367\263,\221\3724\332>7\31H2|\20\254\314\212y\275\334I\304\207"...,
4096) = 4096
 vmsplice(4, [{iov_base="\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0"...,
iov_len=4096}], 1, 0) = 4096

I don't get it.


^ permalink raw reply	[flat|nested] 9+ messages in thread

* Re: [linux-next:master] [drivers/char/mem]  1b057bd800: stress-ng.splice.ops_per_sec -99.8% regression
  2023-10-17 16:56 ` Greg Kroah-Hartman
  2023-10-18  6:31   ` Max Kellermann
@ 2023-10-18  7:07   ` Oliver Sang
  2023-10-18  7:57     ` Greg Kroah-Hartman
  2023-10-18 10:01     ` Max Kellermann
  1 sibling, 2 replies; 9+ messages in thread
From: Oliver Sang @ 2023-10-18  7:07 UTC (permalink / raw)
  To: Greg Kroah-Hartman
  Cc: Max Kellermann, oe-lkp, lkp, Linux Memory Management List,
	linux-kernel, ying.huang, feng.tang, fengwei.yin, oliver.sang

hi, Greg Kroah-Hartman,

On Tue, Oct 17, 2023 at 06:56:56PM +0200, Greg Kroah-Hartman wrote:
> On Tue, Oct 17, 2023 at 11:06:42PM +0800, kernel test robot wrote:
> > 
> > 
> > Hello,
> > 
> > kernel test robot noticed a -99.8% regression of stress-ng.splice.ops_per_sec on:
> > 
> > 
> > commit: 1b057bd800c3ea0c926191d7950cd2365eddc9bb ("drivers/char/mem: implement splice() for /dev/zero, /dev/full")
> > https://git.kernel.org/cgit/linux/kernel/git/next/linux-next.git master
> > 
> > testcase: stress-ng
> > test machine: 64 threads 2 sockets Intel(R) Xeon(R) Gold 6346 CPU @ 3.10GHz (Ice Lake) with 256G memory
> > parameters:
> > 
> > 	nr_threads: 100%
> > 	testtime: 60s
> > 	class: pipe
> > 	test: splice
> > 	cpufreq_governor: performance
> > 
> > 
> > In addition to that, the commit also has significant impact on the following tests:
> > 
> > +------------------+-------------------------------------------------------------------------------------------------+
> > | testcase: change | stress-ng: stress-ng.splice.ops_per_sec 38.9% improvement                                       |
> 
> So everything now goes faster, right?  -99.8% regression means 99.8%
> faster?

let me clarify.

our auto bisect captured this commit as 'first bad commit' in two tests.

Test 1:

it found a (very big) regression comparing to parent commit.

19e3e6cdfdc73400 1b057bd800c3ea0c926191d7950 
---------------- --------------------------- 
         %stddev     %change         %stddev
             \          |                \  
  12433266           -99.8%      22893 ±  3%  stress-ng.splice.ops_per_sec

the detail data for parent in multi-runs:
  "stress-ng.splice.ops_per_sec": [
    12444442.19,
    12599010.87,
    12416009.38,
    12494132.89,
    12286766.76,
    12359235.82
  ],

for 1b057bd800:
  "stress-ng.splice.ops_per_sec": [
    24055.57,
    23235.46,
    22142.13,
    23782.13,
    21732.13,
    22415.46
  ],

so this is much slower.

the config for this Test 1 is:
testcase: stress-ng
test machine: 64 threads 2 sockets Intel(R) Xeon(R) Gold 6346 CPU @ 3.10GHz (Ice Lake) with 256G memory
parameters:
	nr_threads: 100%
	testtime: 60s
	class: pipe
	test: splice
	cpufreq_governor: performance


Test 2:

this is still a stress-ng test, but the config is different with Test 1
(the bare metal machine config, and stress-ng parameters):

testcase: stress-ng
test machine: 36 threads 1 sockets Intel(R) Core(TM) i9-10980XE CPU @ 3.00GHz (Cascade Lake) with 128G memory
parameters:
	nr_threads=1
	testtime=60s
	class=os
	test=splice
	disk=1HDD
	fs=ext4
	cpufreq_governor=performance

Test 2 shows a big improvement:

19e3e6cdfdc73400 1b057bd800c3ea0c926191d7950 
---------------- --------------------------- 
         %stddev     %change         %stddev
             \          |                \  
    171798           +38.9%     238710 ±  4%  stress-ng.splice.ops_per_sec

the detail data:
for parent:
  "stress-ng.splice.ops_per_sec": [
    173056.44,
    172030.08,
    171401.68,
    171694.23,
    171001.19,
    171606.93
  ],

for 1b057bd800:
  "stress-ng.splice.ops_per_sec": [
    244347.89,
    259085.63,
    231423.88,
    232897.93,
    226714.77,
    237792.34
  ],


there are monitoring data such like perf data in original report. FYI

> 
> thanks,
> 
> greg k-h


^ permalink raw reply	[flat|nested] 9+ messages in thread

* Re: [linux-next:master] [drivers/char/mem]  1b057bd800: stress-ng.splice.ops_per_sec -99.8% regression
  2023-10-18  7:07   ` Oliver Sang
@ 2023-10-18  7:57     ` Greg Kroah-Hartman
  2023-10-18  8:12       ` Max Kellermann
  2023-10-18 10:01     ` Max Kellermann
  1 sibling, 1 reply; 9+ messages in thread
From: Greg Kroah-Hartman @ 2023-10-18  7:57 UTC (permalink / raw)
  To: Oliver Sang
  Cc: Max Kellermann, oe-lkp, lkp, Linux Memory Management List,
	linux-kernel, ying.huang, feng.tang, fengwei.yin

On Wed, Oct 18, 2023 at 03:07:20PM +0800, Oliver Sang wrote:
> hi, Greg Kroah-Hartman,
> 
> On Tue, Oct 17, 2023 at 06:56:56PM +0200, Greg Kroah-Hartman wrote:
> > On Tue, Oct 17, 2023 at 11:06:42PM +0800, kernel test robot wrote:
> > > 
> > > 
> > > Hello,
> > > 
> > > kernel test robot noticed a -99.8% regression of stress-ng.splice.ops_per_sec on:
> > > 
> > > 
> > > commit: 1b057bd800c3ea0c926191d7950cd2365eddc9bb ("drivers/char/mem: implement splice() for /dev/zero, /dev/full")
> > > https://git.kernel.org/cgit/linux/kernel/git/next/linux-next.git master
> > > 
> > > testcase: stress-ng
> > > test machine: 64 threads 2 sockets Intel(R) Xeon(R) Gold 6346 CPU @ 3.10GHz (Ice Lake) with 256G memory
> > > parameters:
> > > 
> > > 	nr_threads: 100%
> > > 	testtime: 60s
> > > 	class: pipe
> > > 	test: splice
> > > 	cpufreq_governor: performance
> > > 
> > > 
> > > In addition to that, the commit also has significant impact on the following tests:
> > > 
> > > +------------------+-------------------------------------------------------------------------------------------------+
> > > | testcase: change | stress-ng: stress-ng.splice.ops_per_sec 38.9% improvement                                       |
> > 
> > So everything now goes faster, right?  -99.8% regression means 99.8%
> > faster?
> 
> let me clarify.
> 
> our auto bisect captured this commit as 'first bad commit' in two tests.
> 
> Test 1:
> 
> it found a (very big) regression comparing to parent commit.
> 
> 19e3e6cdfdc73400 1b057bd800c3ea0c926191d7950 
> ---------------- --------------------------- 
>          %stddev     %change         %stddev
>              \          |                \  
>   12433266           -99.8%      22893 ±  3%  stress-ng.splice.ops_per_sec
> 
> the detail data for parent in multi-runs:
>   "stress-ng.splice.ops_per_sec": [

stress-ng is a performance test?

>     12444442.19,
>     12599010.87,
>     12416009.38,
>     12494132.89,
>     12286766.76,
>     12359235.82
>   ],
> 
> for 1b057bd800:
>   "stress-ng.splice.ops_per_sec": [
>     24055.57,
>     23235.46,
>     22142.13,
>     23782.13,
>     21732.13,
>     22415.46
>   ],
> 
> so this is much slower.

That's odd given that as was pointed out, this test does not even touch
the code paths that this patch changed.

confused,

greg k-h


^ permalink raw reply	[flat|nested] 9+ messages in thread

* Re: [linux-next:master] [drivers/char/mem] 1b057bd800: stress-ng.splice.ops_per_sec -99.8% regression
  2023-10-18  7:57     ` Greg Kroah-Hartman
@ 2023-10-18  8:12       ` Max Kellermann
  0 siblings, 0 replies; 9+ messages in thread
From: Max Kellermann @ 2023-10-18  8:12 UTC (permalink / raw)
  To: Greg Kroah-Hartman
  Cc: Oliver Sang, oe-lkp, lkp, Linux Memory Management List,
	linux-kernel, ying.huang, feng.tang, fengwei.yin

On Wed, Oct 18, 2023 at 9:57 AM Greg Kroah-Hartman
<gregkh@linuxfoundation.org> wrote:
> That's odd given that as was pointed out, this test does not even touch
> the code paths that this patch changed.

I think I mixed up the "vmsplice" and the "splice" tests, and my
conclusion was wrong, sorry.
This performance regression is about the "splice" test which indeed
uses /dev/zero. I'll have a closer look and get back to you.


^ permalink raw reply	[flat|nested] 9+ messages in thread

* Re: [linux-next:master] [drivers/char/mem] 1b057bd800: stress-ng.splice.ops_per_sec -99.8% regression
  2023-10-18  7:07   ` Oliver Sang
  2023-10-18  7:57     ` Greg Kroah-Hartman
@ 2023-10-18 10:01     ` Max Kellermann
  1 sibling, 0 replies; 9+ messages in thread
From: Max Kellermann @ 2023-10-18 10:01 UTC (permalink / raw)
  To: Oliver Sang
  Cc: Greg Kroah-Hartman, oe-lkp, lkp, Linux Memory Management List,
	linux-kernel, ying.huang, feng.tang, fengwei.yin

On Wed, Oct 18, 2023 at 9:07 AM Oliver Sang <oliver.sang@intel.com> wrote:
> it found a (very big) regression comparing to parent commit.
>
> 19e3e6cdfdc73400 1b057bd800c3ea0c926191d7950
> ---------------- ---------------------------
>          %stddev     %change         %stddev
>              \          |                \
>   12433266           -99.8%      22893 ą  3%  stress-ng.splice.ops_per_sec

Oliver, how is it possible that we have three times the throughput
(MB_per_sec) at 1/500ths the ops?


^ permalink raw reply	[flat|nested] 9+ messages in thread

* Re: [linux-next:master] [drivers/char/mem] 1b057bd800: stress-ng.splice.ops_per_sec -99.8% regression
       [not found] ` <CAKPOu+_T8xk4yd2P4KT4j3eMoFqwYmkxqDHaFtv4Hii5-XyPuA@mail.gmail.com>
@ 2023-10-18 11:12   ` Max Kellermann
  2023-10-19  5:41     ` Oliver Sang
  0 siblings, 1 reply; 9+ messages in thread
From: Max Kellermann @ 2023-10-18 11:12 UTC (permalink / raw)
  To: kernel test robot
  Cc: oe-lkp, lkp, Linux Memory Management List, Greg Kroah-Hartman,
	linux-kernel, ying.huang, feng.tang, fengwei.yin

On Wed, Oct 18, 2023 at 12:21 PM Max Kellermann
<max.kellermann@ionos.com> wrote:
> I think this might be caused by a bug in stress-ng, leading to
> blocking pipe writes.

Just in case I happen to be right, I've submitted a PR for stress-ng:
https://github.com/ColinIanKing/stress-ng/pull/326


^ permalink raw reply	[flat|nested] 9+ messages in thread

* Re: [linux-next:master] [drivers/char/mem] 1b057bd800: stress-ng.splice.ops_per_sec -99.8% regression
  2023-10-18 11:12   ` Max Kellermann
@ 2023-10-19  5:41     ` Oliver Sang
  0 siblings, 0 replies; 9+ messages in thread
From: Oliver Sang @ 2023-10-19  5:41 UTC (permalink / raw)
  To: Max Kellermann
  Cc: oe-lkp, lkp, Linux Memory Management List, Greg Kroah-Hartman,
	linux-kernel, ying.huang, feng.tang, fengwei.yin, oliver.sang

hi, Max Kellermann,

On Wed, Oct 18, 2023 at 01:12:27PM +0200, Max Kellermann wrote:
> On Wed, Oct 18, 2023 at 12:21 PM Max Kellermann
> <max.kellermann@ionos.com> wrote:
> > I think this might be caused by a bug in stress-ng, leading to
> > blocking pipe writes.
> 
> Just in case I happen to be right, I've submitted a PR for stress-ng:
> https://github.com/ColinIanKing/stress-ng/pull/326
> 

we tested with this commit, and noticed a big improvement now.

19e3e6cdfdc73400 1b057bd800c3ea0c926191d7950
---------------- ---------------------------
         %stddev     %change         %stddev
             \          |                \
 7.861e+08 ±  2%    +167.9%  2.106e+09        stress-ng.splice.ops
  12886325 ±  2%    +167.9%   34526248        stress-ng.splice.ops_per_sec


detail comparison with more monitor/perf data as below FYI



19e3e6cdfdc73400 1b057bd800c3ea0c926191d7950
---------------- ---------------------------
         %stddev     %change         %stddev
             \          |                \
 2.464e+09            +0.0%  2.464e+09        cpuidle..time
   2495283            +0.4%    2504203        cpuidle..usage
    116.23            +4.4%     121.32        uptime.boot
      8523 ±  2%      +7.7%       9178        uptime.idle
     49.76 ±  2%     +10.2%      54.85        boot-time.boot
     26.14            -0.1%      26.13        boot-time.dhcp
      5927 ±  2%     +11.0%       6580        boot-time.idle
     28.99            +0.0       29.02        mpstat.cpu.all.idle%
      0.93            +0.0        0.94        mpstat.cpu.all.irq%
      0.01 ±  3%      -0.0        0.01 ±  5%  mpstat.cpu.all.soft%
     63.59           -10.6       52.95        mpstat.cpu.all.sys%
      6.47           +10.6       17.08        mpstat.cpu.all.usr%
    487.00 ±  2%     -99.5%       2.50 ± 20%  perf-c2c.DRAM.local
      3240 ± 11%     -96.8%     103.00 ±  3%  perf-c2c.DRAM.remote
      5244           -97.0%     156.00 ±  9%  perf-c2c.HITM.local
    432.50 ± 20%     -85.2%      64.00 ±  7%  perf-c2c.HITM.remote
      5676 ±  2%     -96.1%     220.00 ±  9%  perf-c2c.HITM.total
      0.00          -100.0%       0.00        numa-numastat.node0.interleave_hit
 4.611e+08 ±  2%    +129.2%  1.057e+09        numa-numastat.node0.local_node
 4.613e+08 ±  2%    +129.1%  1.057e+09        numa-numastat.node0.numa_hit
     78581 ± 47%     -63.1%      28977 ± 42%  numa-numastat.node0.other_node
      0.00          -100.0%       0.00        numa-numastat.node1.interleave_hit
 3.259e+08 ± 10%    +223.5%  1.054e+09        numa-numastat.node1.local_node
 3.264e+08 ± 10%    +223.0%  1.054e+09        numa-numastat.node1.numa_hit
     53882 ± 68%     +92.0%     103462 ± 12%  numa-numastat.node1.other_node
     31.28            +0.0%      31.29        vmstat.cpu.id
     62.75           -17.0%      52.11        vmstat.cpu.sy
      6.39 ±  2%    +159.7%      16.59        vmstat.cpu.us
      0.03            +0.0%       0.03        vmstat.io.bi
      4.00            +0.0%       4.00        vmstat.memory.buff
   3047892            -0.5%    3032948        vmstat.memory.cache
 2.256e+08            +0.0%  2.257e+08        vmstat.memory.free
     88.34            +0.0%      88.37        vmstat.procs.r
      3238 ±  2%      +5.1%       3402 ±  4%  vmstat.system.cs
    185280            +0.9%     186929        vmstat.system.in
     63.19            -0.1%      63.15        time.elapsed_time
     63.19            -0.1%      63.15        time.elapsed_time.max
      9701            +3.3%      10026 ±  6%  time.involuntary_context_switches
      0.00         +5e+101%       0.50 ±100%  time.major_page_faults
      4096            +0.0%       4096        time.maximum_resident_set_size
     11594            -2.1%      11347 ±  2%  time.minor_page_faults
      4096            +0.0%       4096        time.page_size
      9176            +0.0%       9180        time.percent_of_cpu_this_job_got
      5280           -16.7%       4397        time.system_time
    518.57          +170.2%       1401        time.user_time
      1723 ±  5%      -4.8%       1640        time.voluntary_context_switches
 7.861e+08 ±  2%    +167.9%  2.106e+09        stress-ng.splice.ops
  12886325 ±  2%    +167.9%   34526248        stress-ng.splice.ops_per_sec
     63.19            -0.1%      63.15        stress-ng.time.elapsed_time
     63.19            -0.1%      63.15        stress-ng.time.elapsed_time.max
      9701            +3.3%      10026 ±  6%  stress-ng.time.involuntary_context_switches
      0.00         +5e+101%       0.50 ±100%  stress-ng.time.major_page_faults
      4096            +0.0%       4096        stress-ng.time.maximum_resident_set_size
     11594            -2.1%      11347 ±  2%  stress-ng.time.minor_page_faults
      4096            +0.0%       4096        stress-ng.time.page_size
      9176            +0.0%       9180        stress-ng.time.percent_of_cpu_this_job_got
      5280           -16.7%       4397        stress-ng.time.system_time
    518.57          +170.2%       1401        stress-ng.time.user_time
      1723 ±  5%      -4.8%       1640        stress-ng.time.voluntary_context_switches
      1839            -0.0%       1839        turbostat.Avg_MHz
     70.90            -0.0       70.86        turbostat.Busy%
      2600            +0.0%       2600        turbostat.Bzy_MHz
   2460348            +0.4%    2470080        turbostat.C1
     29.44            +0.0       29.46        turbostat.C1%
     29.10            +0.1%      29.14        turbostat.CPU%c1
     71.00            -3.5%      68.50        turbostat.CoreTmp
      0.16 ±  3%    +163.6%       0.44        turbostat.IPC
  12324057            +0.9%   12436149        turbostat.IRQ
      0.00          +177.1      177.10        turbostat.PKG_%
    158.50 ± 19%     -68.1%      50.50 ±  2%  turbostat.POLL
     71.00            -3.5%      68.50        turbostat.PkgTmp
    362.06           +10.4%     399.82        turbostat.PkgWatt
     30.42 ±  2%      -7.5%      28.14 ±  3%  turbostat.RAMWatt
      1996            +0.0%       1996        turbostat.TSC_MHz
    180394            -5.7%     170032        meminfo.Active
    179746            -5.8%     169377        meminfo.Active(anon)
    647.52            +1.1%     654.95        meminfo.Active(file)
     80759            +1.3%      81828 ±  4%  meminfo.AnonHugePages
    403700            -0.4%     401890        meminfo.AnonPages
      4.00            +0.0%       4.00        meminfo.Buffers
   2929930            -0.5%    2915714        meminfo.Cached
 1.154e+08            +0.0%  1.154e+08        meminfo.CommitLimit
   1544192            -1.7%    1518156        meminfo.Committed_AS
 2.254e+08            -0.5%  2.244e+08        meminfo.DirectMap1G
  11029504 ±  9%      +9.5%   12079104 ±  8%  meminfo.DirectMap2M
    148584 ±  3%      -0.7%     147560 ±  6%  meminfo.DirectMap4k
      2048            +0.0%       2048        meminfo.Hugepagesize
    437427            -1.1%     432672        meminfo.Inactive
    437032            -1.1%     432286        meminfo.Inactive(anon)
    394.91            -2.4%     385.42        meminfo.Inactive(file)
    112746            -0.5%     112232        meminfo.KReclaimable
     24941            -1.2%      24647        meminfo.KernelStack
     74330            -8.9%      67748 ±  5%  meminfo.Mapped
 2.246e+08            +0.0%  2.247e+08        meminfo.MemAvailable
 2.256e+08            +0.0%  2.257e+08        meminfo.MemFree
 2.307e+08            +0.0%  2.307e+08        meminfo.MemTotal
   5077046            -0.8%    5036202        meminfo.Memused
     11.18            +0.0%      11.18        meminfo.Mlocked
      9587            -0.4%       9551        meminfo.PageTables
     59841            -0.0%      59837        meminfo.Percpu
    112746            -0.5%     112232        meminfo.SReclaimable
    259649            -0.4%     258517        meminfo.SUnreclaim
    213290            -6.7%     199078        meminfo.Shmem
    372396            -0.4%     370750        meminfo.Slab
   2715599            -0.0%    2715599        meminfo.Unevictable
 1.374e+13            +0.0%  1.374e+13        meminfo.VmallocTotal
    263718            -0.2%     263313        meminfo.VmallocUsed
   5528890            -0.4%    5504303        meminfo.max_used_kB
     34163 ±  4%     -72.1%       9514 ± 86%  numa-meminfo.node0.Active
     33835 ±  5%     -71.9%       9514 ± 86%  numa-meminfo.node0.Active(anon)
    328.00 ±100%    -100.0%       0.00        numa-meminfo.node0.Active(file)
      5064 ±  2%    +774.4%      44282 ± 53%  numa-meminfo.node0.AnonHugePages
     76171 ± 15%    +150.5%     190797 ± 29%  numa-meminfo.node0.AnonPages
    144806 ±  4%     +90.9%     276438 ± 23%  numa-meminfo.node0.AnonPages.max
   2743966           -96.1%     107423 ± 93%  numa-meminfo.node0.FilePages
     81560 ± 15%    +140.2%     195891 ± 28%  numa-meminfo.node0.Inactive
     81366 ± 15%    +140.8%     195891 ± 28%  numa-meminfo.node0.Inactive(anon)
    193.88 ±100%    -100.0%       0.00        numa-meminfo.node0.Inactive(file)
     82587 ±  3%     -65.1%      28849 ± 11%  numa-meminfo.node0.KReclaimable
     12151 ±  3%      +3.6%      12587        numa-meminfo.node0.KernelStack
     45176           -70.4%      13392 ± 18%  numa-meminfo.node0.Mapped
 1.279e+08            +2.0%  1.305e+08        numa-meminfo.node0.MemFree
 1.317e+08            +0.0%  1.317e+08        numa-meminfo.node0.MemTotal
   3753848           -67.2%    1232426 ±  4%  numa-meminfo.node0.MemUsed
      2133 ±  7%    +123.3%       4762 ± 43%  numa-meminfo.node0.PageTables
     82587 ±  3%     -65.1%      28849 ± 11%  numa-meminfo.node0.SReclaimable
    139456 ±  7%      +3.7%     144548        numa-meminfo.node0.SUnreclaim
     39137 ±  6%     -62.4%      14718 ± 59%  numa-meminfo.node0.Shmem
    222043 ±  3%     -21.9%     173397 ±  2%  numa-meminfo.node0.Slab
   2704307           -96.6%      92704 ± 99%  numa-meminfo.node0.Unevictable
    146360            +9.7%     160589 ±  6%  numa-meminfo.node1.Active
    146040            +9.5%     159934 ±  6%  numa-meminfo.node1.Active(anon)
    319.52 ±100%    +105.0%     654.95        numa-meminfo.node1.Active(file)
     75819           -50.3%      37670 ± 52%  numa-meminfo.node1.AnonHugePages
    327658 ±  3%     -35.5%     211208 ± 26%  numa-meminfo.node1.AnonPages
    411864           -34.8%     268684 ± 16%  numa-meminfo.node1.AnonPages.max
    186229         +1408.1%    2808443 ±  3%  numa-meminfo.node1.FilePages
    356143 ±  3%     -33.5%     236972 ± 24%  numa-meminfo.node1.Inactive
    355942 ±  3%     -33.5%     236586 ± 24%  numa-meminfo.node1.Inactive(anon)
    201.03 ±100%     +91.7%     385.42        numa-meminfo.node1.Inactive(file)
     30163 ± 10%    +176.4%      83380 ±  4%  numa-meminfo.node1.KReclaimable
     12792 ±  2%      -5.7%      12061        numa-meminfo.node1.KernelStack
     29717           +84.0%      54689 ±  2%  numa-meminfo.node1.Mapped
  97713848            -2.5%   95230678        numa-meminfo.node1.MemFree
  99034676            +0.0%   99034676        numa-meminfo.node1.MemTotal
   1320827 ±  5%    +188.0%    3803997        numa-meminfo.node1.MemUsed
     11.08            +0.8%      11.18        numa-meminfo.node1.Mlocked
      7406           -35.4%       4784 ± 41%  numa-meminfo.node1.PageTables
     30163 ± 10%    +176.4%      83380 ±  4%  numa-meminfo.node1.SReclaimable
    120178 ±  8%      -5.2%     113955        numa-meminfo.node1.SUnreclaim
    174414            +5.8%     184508 ±  4%  numa-meminfo.node1.Shmem
    150341 ±  5%     +31.3%     197335 ±  2%  numa-meminfo.node1.Slab
     11292 ± 29%  +23126.1%    2622894 ±  3%  numa-meminfo.node1.Unevictable
      8464 ±  5%     -71.9%       2380 ± 86%  numa-vmstat.node0.nr_active_anon
     82.00 ±100%    -100.0%       0.00        numa-vmstat.node0.nr_active_file
     19043 ± 15%    +150.4%      47693 ± 29%  numa-vmstat.node0.nr_anon_pages
      2.47 ±  2%    +774.4%      21.62 ± 53%  numa-vmstat.node0.nr_anon_transparent_hugepages
    685997           -96.1%      26858 ± 93%  numa-vmstat.node0.nr_file_pages
  31983370            +2.0%   32613637        numa-vmstat.node0.nr_free_pages
     20344 ± 15%    +140.7%      48968 ± 28%  numa-vmstat.node0.nr_inactive_anon
     48.47 ±100%    -100.0%       0.00        numa-vmstat.node0.nr_inactive_file
     12155 ±  3%      +3.5%      12579        numa-vmstat.node0.nr_kernel_stack
     11313           -70.4%       3351 ± 18%  numa-vmstat.node0.nr_mapped
    534.37 ±  6%    +121.2%       1182 ± 42%  numa-vmstat.node0.nr_page_table_pages
      9790 ±  6%     -62.4%       3682 ± 59%  numa-vmstat.node0.nr_shmem
     20646 ±  3%     -65.1%       7212 ± 11%  numa-vmstat.node0.nr_slab_reclaimable
     34863 ±  7%      +3.6%      36135        numa-vmstat.node0.nr_slab_unreclaimable
    676076           -96.6%      23176 ± 99%  numa-vmstat.node0.nr_unevictable
      8464 ±  5%     -71.9%       2380 ± 86%  numa-vmstat.node0.nr_zone_active_anon
     82.00 ±100%    -100.0%       0.00        numa-vmstat.node0.nr_zone_active_file
     20344 ± 15%    +140.7%      48968 ± 28%  numa-vmstat.node0.nr_zone_inactive_anon
     48.47 ±100%    -100.0%       0.00        numa-vmstat.node0.nr_zone_inactive_file
    676076           -96.6%      23176 ± 99%  numa-vmstat.node0.nr_zone_unevictable
 4.613e+08 ±  2%    +129.1%  1.057e+09        numa-vmstat.node0.numa_hit
      0.00          -100.0%       0.00        numa-vmstat.node0.numa_interleave
 4.611e+08 ±  2%    +129.2%  1.057e+09        numa-vmstat.node0.numa_local
     78581 ± 47%     -63.1%      28977 ± 42%  numa-vmstat.node0.numa_other
     36516            +9.5%      39989 ±  6%  numa-vmstat.node1.nr_active_anon
     79.88 ±100%    +105.0%     163.74        numa-vmstat.node1.nr_active_file
     81923 ±  3%     -35.5%      52807 ± 26%  numa-vmstat.node1.nr_anon_pages
     37.02           -50.3%      18.39 ± 52%  numa-vmstat.node1.nr_anon_transparent_hugepages
     46577         +1407.4%     702118 ±  3%  numa-vmstat.node1.nr_file_pages
  24428660            -2.5%   23807665        numa-vmstat.node1.nr_free_pages
     89004 ±  3%     -33.5%      59154 ± 24%  numa-vmstat.node1.nr_inactive_anon
     50.26 ±100%     +91.7%      96.35        numa-vmstat.node1.nr_inactive_file
      2.86 ±100%    -100.0%       0.00        numa-vmstat.node1.nr_isolated_anon
     12795 ±  2%      -5.7%      12060        numa-vmstat.node1.nr_kernel_stack
      7452           +83.7%      13693 ±  2%  numa-vmstat.node1.nr_mapped
      2.77            +0.0%       2.77        numa-vmstat.node1.nr_mlock
      1852           -35.5%       1193 ± 42%  numa-vmstat.node1.nr_page_table_pages
     43623            +5.8%      46134 ±  4%  numa-vmstat.node1.nr_shmem
      7540 ± 10%    +176.4%      20844 ±  4%  numa-vmstat.node1.nr_slab_reclaimable
     30045 ±  8%      -5.2%      28488        numa-vmstat.node1.nr_slab_unreclaimable
      2823 ± 29%  +23126.1%     655723 ±  3%  numa-vmstat.node1.nr_unevictable
     36516            +9.5%      39989 ±  6%  numa-vmstat.node1.nr_zone_active_anon
     79.88 ±100%    +105.0%     163.74        numa-vmstat.node1.nr_zone_active_file
     89005 ±  3%     -33.5%      59154 ± 24%  numa-vmstat.node1.nr_zone_inactive_anon
     50.26 ±100%     +91.7%      96.35        numa-vmstat.node1.nr_zone_inactive_file
      2823 ± 29%  +23126.1%     655723 ±  3%  numa-vmstat.node1.nr_zone_unevictable
 3.264e+08 ± 10%    +223.0%  1.054e+09        numa-vmstat.node1.numa_hit
      0.00          -100.0%       0.00        numa-vmstat.node1.numa_interleave
 3.259e+08 ± 10%    +223.5%  1.054e+09        numa-vmstat.node1.numa_local
     53882 ± 68%     +92.0%     103462 ± 12%  numa-vmstat.node1.numa_other
     54.00 ±  5%      -0.9%      53.50 ±  6%  proc-vmstat.direct_map_level2_splits
      3.00 ± 33%     +33.3%       4.00 ± 25%  proc-vmstat.direct_map_level3_splits
     44945            -5.8%      42350        proc-vmstat.nr_active_anon
    161.88            +1.1%     163.74        proc-vmstat.nr_active_file
    100932            -0.5%     100476        proc-vmstat.nr_anon_pages
     39.43            +1.3%      39.96 ±  4%  proc-vmstat.nr_anon_transparent_hugepages
   5606101            +0.0%    5607117        proc-vmstat.nr_dirty_background_threshold
  11225910            +0.0%   11227943        proc-vmstat.nr_dirty_threshold
    732499            -0.5%     728939        proc-vmstat.nr_file_pages
  56411216            +0.0%   56421388        proc-vmstat.nr_free_pages
    109272            -1.1%     108084        proc-vmstat.nr_inactive_anon
     98.73            -2.4%      96.35        proc-vmstat.nr_inactive_file
     24948            -1.2%      24644        proc-vmstat.nr_kernel_stack
     18621            -8.8%      16973 ±  5%  proc-vmstat.nr_mapped
      2.79            +0.0%       2.79        proc-vmstat.nr_mlock
      2397            -0.4%       2386        proc-vmstat.nr_page_table_pages
     53338            -6.7%      49779        proc-vmstat.nr_shmem
     28186            -0.5%      28057        proc-vmstat.nr_slab_reclaimable
     64909            -0.4%      64625        proc-vmstat.nr_slab_unreclaimable
    678900            -0.0%     678899        proc-vmstat.nr_unevictable
     44945            -5.8%      42350        proc-vmstat.nr_zone_active_anon
    161.88            +1.1%     163.74        proc-vmstat.nr_zone_active_file
    109272            -1.1%     108084        proc-vmstat.nr_zone_inactive_anon
     98.73            -2.4%      96.35        proc-vmstat.nr_zone_inactive_file
    678900            -0.0%     678899        proc-vmstat.nr_zone_unevictable
     24423 ± 26%      -5.7%      23043 ±  3%  proc-vmstat.numa_hint_faults
     14265 ± 34%     -24.4%      10782 ± 31%  proc-vmstat.numa_hint_faults_local
 7.877e+08 ±  2%    +168.0%  2.111e+09        proc-vmstat.numa_hit
     18.00            +0.0%      18.00        proc-vmstat.numa_huge_pte_updates
      0.00          -100.0%       0.00        proc-vmstat.numa_interleave
  7.87e+08 ±  2%    +168.3%  2.111e+09        proc-vmstat.numa_local
    132463            -0.0%     132440        proc-vmstat.numa_other
      8701            -0.9%       8621 ± 36%  proc-vmstat.numa_pages_migrated
     86479 ± 18%      +4.5%      90373        proc-vmstat.numa_pte_updates
     79724            -3.4%      77014        proc-vmstat.pgactivate
      0.00          -100.0%       0.00        proc-vmstat.pgalloc_dma32
 7.868e+08 ±  2%    +167.8%  2.107e+09        proc-vmstat.pgalloc_normal
    458741            -0.9%     454690        proc-vmstat.pgfault
 7.867e+08 ±  2%    +167.8%  2.107e+09        proc-vmstat.pgfree
      8701            -0.9%       8621 ± 36%  proc-vmstat.pgmigrate_success
      0.00          -100.0%       0.00        proc-vmstat.pgpgin
     18188 ±  4%      +4.5%      19011 ± 22%  proc-vmstat.pgreuse
     48.00            +8.3%      52.00 ±  7%  proc-vmstat.thp_collapse_alloc
      0.00         +5e+101%       0.50 ±100%  proc-vmstat.thp_deferred_split_page
     24.00            +2.1%      24.50 ±  2%  proc-vmstat.thp_fault_alloc
      7.50 ± 20%     -33.3%       5.00 ± 40%  proc-vmstat.thp_migration_success
      0.00         +5e+101%       0.50 ±100%  proc-vmstat.thp_split_pmd
      0.00          -100.0%       0.00        proc-vmstat.thp_zero_page_alloc
      3341            +0.0%       3341        proc-vmstat.unevictable_pgs_culled
      3.00            +0.0%       3.00        proc-vmstat.unevictable_pgs_mlocked
      3.00            +0.0%       3.00        proc-vmstat.unevictable_pgs_munlocked
      0.00          -100.0%       0.00        proc-vmstat.unevictable_pgs_rescued
    778368            -0.6%     773760 ±  2%  proc-vmstat.unevictable_pgs_scanned
      0.57 ±  7%     -97.9%       0.01        perf-stat.i.MPKI
 2.375e+10 ±  2%    +162.1%  6.225e+10        perf-stat.i.branch-instructions
      0.16            -0.0        0.15        perf-stat.i.branch-miss-rate%
  21671842          +104.9%   44401835 ±  3%  perf-stat.i.branch-misses
     22.10 ±  8%      -8.8       13.35        perf-stat.i.cache-miss-rate%
  69187553 ±  4%     -98.4%    1092227 ±  2%  perf-stat.i.cache-misses
 3.089e+08 ±  3%     -96.9%    9661190        perf-stat.i.cache-references
      2950 ±  2%      +5.4%       3110 ±  5%  perf-stat.i.context-switches
      1.97 ±  2%     -61.7%       0.75        perf-stat.i.cpi
    128091            +0.0%     128117        perf-stat.i.cpu-clock
 2.389e+11            +0.0%   2.39e+11        perf-stat.i.cpu-cycles
    181.78            +3.3%     187.70        perf-stat.i.cpu-migrations
      3442 ±  4%  +10605.2%     368536        perf-stat.i.cycles-between-cache-misses
      0.00 ±  2%      -0.0        0.00 ±  3%  perf-stat.i.dTLB-load-miss-rate%
    201135 ±  7%     +67.1%     336024 ±  4%  perf-stat.i.dTLB-load-misses
 3.456e+10 ±  2%    +160.5%  9.003e+10        perf-stat.i.dTLB-loads
      0.00            -0.0        0.00        perf-stat.i.dTLB-store-miss-rate%
     67306           +81.9%     122399        perf-stat.i.dTLB-store-misses
 2.203e+10 ±  2%    +163.6%  5.806e+10        perf-stat.i.dTLB-stores
  1.19e+11 ±  2%    +165.2%  3.156e+11        perf-stat.i.instructions
      0.53 ±  2%    +150.1%       1.33        perf-stat.i.ipc
      0.06 ± 42%     -25.0%       0.04 ± 59%  perf-stat.i.major-faults
      1.87            -0.0%       1.87        perf-stat.i.metric.GHz
    147.49 ± 10%     -46.3%      79.15        perf-stat.i.metric.K/sec
    629.80 ±  2%    +160.8%       1642        perf-stat.i.metric.M/sec
      4806 ±  2%      -1.4%       4738        perf-stat.i.minor-faults
     82.79           +15.8       98.54        perf-stat.i.node-load-miss-rate%
   7729245 ±  3%     -95.9%     318844        perf-stat.i.node-load-misses
   1661082 ± 13%     -99.7%       4701        perf-stat.i.node-loads
     96.66           -34.3       62.40 ± 12%  perf-stat.i.node-store-miss-rate%
   8408062 ± 18%     -97.8%     187945 ± 15%  perf-stat.i.node-store-misses
    266188 ±  3%     -65.0%      93251 ± 22%  perf-stat.i.node-stores
      4806 ±  2%      -1.4%       4738        perf-stat.i.page-faults
    128091            +0.0%     128117        perf-stat.i.task-clock
      0.58 ±  7%     -99.4%       0.00 ±  2%  perf-stat.overall.MPKI
      0.09            -0.0        0.07 ±  3%  perf-stat.overall.branch-miss-rate%
     22.45 ±  8%     -11.5       10.98        perf-stat.overall.cache-miss-rate%
      2.01 ±  2%     -62.3%       0.76        perf-stat.overall.cpi
      3460 ±  4%   +6240.4%     219417 ±  2%  perf-stat.overall.cycles-between-cache-misses
      0.00            -0.0        0.00 ±  4%  perf-stat.overall.dTLB-load-miss-rate%
      0.00            -0.0        0.00        perf-stat.overall.dTLB-store-miss-rate%
      0.50 ±  2%    +165.2%       1.32        perf-stat.overall.ipc
     82.39           +16.2       98.57        perf-stat.overall.node-load-miss-rate%
     96.84           -30.2       66.67 ± 12%  perf-stat.overall.node-store-miss-rate%
 2.342e+10 ±  2%    +162.0%  6.137e+10        perf-stat.ps.branch-instructions
  21165789          +105.8%   43553577 ±  3%  perf-stat.ps.branch-misses
  68249764 ±  4%     -98.4%    1074701 ±  2%  perf-stat.ps.cache-misses
 3.049e+08 ±  3%     -96.8%    9783139 ±  2%  perf-stat.ps.cache-references
      2875 ±  2%      +5.9%       3044 ±  5%  perf-stat.ps.context-switches
    125960            +0.0%     125967        perf-stat.ps.cpu-clock
 2.356e+11            -0.0%  2.356e+11        perf-stat.ps.cpu-cycles
    178.88            +2.4%     183.17        perf-stat.ps.cpu-migrations
    210651 ±  2%     +79.6%     378309 ±  4%  perf-stat.ps.dTLB-load-misses
 3.408e+10 ±  2%    +160.4%  8.875e+10        perf-stat.ps.dTLB-loads
     66365           +83.5%     121783        perf-stat.ps.dTLB-store-misses
 2.173e+10 ±  2%    +163.5%  5.724e+10        perf-stat.ps.dTLB-stores
 1.173e+11 ±  2%    +165.2%  3.111e+11        perf-stat.ps.instructions
      0.06 ± 42%     -28.5%       0.04 ± 60%  perf-stat.ps.major-faults
      4733 ±  2%      -1.8%       4649        perf-stat.ps.minor-faults
   7629849 ±  3%     -95.8%     317156 ±  2%  perf-stat.ps.node-load-misses
   1639040 ± 13%     -99.7%       4598        perf-stat.ps.node-loads
   8293657 ± 18%     -97.8%     185047 ± 16%  perf-stat.ps.node-store-misses
    263181 ±  4%     -65.4%      91130 ± 22%  perf-stat.ps.node-stores
      4733 ±  2%      -1.8%       4649        perf-stat.ps.page-faults
    125960            +0.0%     125966        perf-stat.ps.task-clock
 7.465e+12 ±  2%    +165.5%  1.982e+13        perf-stat.total.instructions
      1912        +56665.9%    1085563 ± 99%  sched_debug.cfs_rq:/.avg_vruntime.avg
     12690 ±  4%  +11390.8%    1458273 ± 98%  sched_debug.cfs_rq:/.avg_vruntime.max
     53.23 ± 19%  +21837.0%      11676 ± 99%  sched_debug.cfs_rq:/.avg_vruntime.min
      2780        +14828.8%     415155 ± 99%  sched_debug.cfs_rq:/.avg_vruntime.stddev
      0.12 ± 25%    +125.0%       0.28 ± 55%  sched_debug.cfs_rq:/.h_nr_running.avg
      2.00 ± 50%     -50.0%       1.00        sched_debug.cfs_rq:/.h_nr_running.max
      0.36 ± 18%      -2.2%       0.35 ±  5%  sched_debug.cfs_rq:/.h_nr_running.stddev
      1.88 ± 99%    -100.0%       0.00        sched_debug.cfs_rq:/.left_vruntime.avg
    240.81 ± 99%    -100.0%       0.00        sched_debug.cfs_rq:/.left_vruntime.max
      0.00            +0.0%       0.00        sched_debug.cfs_rq:/.left_vruntime.min
     21.20 ±100%    -100.0%       0.00        sched_debug.cfs_rq:/.left_vruntime.stddev
     10324 ± 83%     -63.3%       3793 ± 36%  sched_debug.cfs_rq:/.load.avg
   1078332 ± 96%     -97.0%      32285 ± 35%  sched_debug.cfs_rq:/.load.max
     96021 ± 93%     -93.6%       6192 ± 13%  sched_debug.cfs_rq:/.load.stddev
      2211 ± 65%     -78.0%     486.43 ± 56%  sched_debug.cfs_rq:/.load_avg.avg
     88390           -66.2%      29873 ± 48%  sched_debug.cfs_rq:/.load_avg.max
     12036 ± 36%     -71.4%       3445 ± 59%  sched_debug.cfs_rq:/.load_avg.stddev
      1912        +56666.1%    1085563 ± 99%  sched_debug.cfs_rq:/.min_vruntime.avg
     12690 ±  4%  +11390.8%    1458273 ± 98%  sched_debug.cfs_rq:/.min_vruntime.max
     53.23 ± 19%  +21837.0%      11676 ± 99%  sched_debug.cfs_rq:/.min_vruntime.min
      2780        +14828.7%     415155 ± 99%  sched_debug.cfs_rq:/.min_vruntime.stddev
      0.12 ± 25%    +125.0%       0.28 ± 55%  sched_debug.cfs_rq:/.nr_running.avg
      2.00 ± 50%     -50.0%       1.00        sched_debug.cfs_rq:/.nr_running.max
      0.36 ± 18%      -2.2%       0.35 ±  5%  sched_debug.cfs_rq:/.nr_running.stddev
    410.09 ± 82%     -91.8%      33.57 ± 17%  sched_debug.cfs_rq:/.removed.load_avg.avg
     44892 ± 97%     -98.3%     768.00 ± 33%  sched_debug.cfs_rq:/.removed.load_avg.max
      4035 ± 93%     -96.1%     155.41 ± 25%  sched_debug.cfs_rq:/.removed.load_avg.stddev
     28.03 ± 15%     -49.5%      14.16 ± 11%  sched_debug.cfs_rq:/.removed.runnable_avg.avg
    525.00           -25.4%     391.50 ± 32%  sched_debug.cfs_rq:/.removed.runnable_avg.max
    110.15 ±  8%     -39.5%      66.68 ± 20%  sched_debug.cfs_rq:/.removed.runnable_avg.stddev
     28.02 ± 15%     -49.5%      14.16 ± 11%  sched_debug.cfs_rq:/.removed.util_avg.avg
    525.00           -25.4%     391.50 ± 32%  sched_debug.cfs_rq:/.removed.util_avg.max
    110.13 ±  8%     -39.5%      66.68 ± 20%  sched_debug.cfs_rq:/.removed.util_avg.stddev
      1.88 ± 99%    -100.0%       0.00        sched_debug.cfs_rq:/.right_vruntime.avg
    240.92 ± 99%    -100.0%       0.00        sched_debug.cfs_rq:/.right_vruntime.max
      0.00            +0.0%       0.00        sched_debug.cfs_rq:/.right_vruntime.min
     21.21 ±100%    -100.0%       0.00        sched_debug.cfs_rq:/.right_vruntime.stddev
    280.12 ±  2%     +38.8%     388.72 ± 34%  sched_debug.cfs_rq:/.runnable_avg.avg
      1243 ±  8%      -0.3%       1239 ± 11%  sched_debug.cfs_rq:/.runnable_avg.max
    321.04 ±  3%      +5.1%     337.27 ±  7%  sched_debug.cfs_rq:/.runnable_avg.stddev
      0.00 ±100%    -100.0%       0.00        sched_debug.cfs_rq:/.spread.avg
      0.11 ±100%    -100.0%       0.00        sched_debug.cfs_rq:/.spread.max
      0.01 ±100%    -100.0%       0.00        sched_debug.cfs_rq:/.spread.stddev
    278.43 ±  2%     +39.2%     387.54 ± 34%  sched_debug.cfs_rq:/.util_avg.avg
      1242 ±  8%      -0.2%       1239 ± 11%  sched_debug.cfs_rq:/.util_avg.max
    319.84 ±  3%      +5.2%     336.35 ±  8%  sched_debug.cfs_rq:/.util_avg.stddev
     29.61 ± 17%    +203.8%      89.96 ± 64%  sched_debug.cfs_rq:/.util_est_enqueued.avg
    826.50            -9.4%     748.75 ± 10%  sched_debug.cfs_rq:/.util_est_enqueued.max
    120.67 ±  8%     +22.4%     147.72 ± 11%  sched_debug.cfs_rq:/.util_est_enqueued.stddev
    847909            +3.7%     879071        sched_debug.cpu.avg_idle.avg
   1049554 ±  4%      -4.7%    1000000        sched_debug.cpu.avg_idle.max
      4595 ± 24%    +177.9%      12771 ± 65%  sched_debug.cpu.avg_idle.min
    238818 ±  5%     -14.5%     204207        sched_debug.cpu.avg_idle.stddev
     52188 ±  2%     +38.6%      72309 ± 21%  sched_debug.cpu.clock.avg
     52194 ±  2%     +38.6%      72316 ± 21%  sched_debug.cpu.clock.max
     52179 ±  2%     +38.6%      72300 ± 21%  sched_debug.cpu.clock.min
      3.53           +12.7%       3.98 ±  9%  sched_debug.cpu.clock.stddev
     52058 ±  2%     +38.4%      72031 ± 20%  sched_debug.cpu.clock_task.avg
     52183 ±  2%     +38.3%      72170 ± 20%  sched_debug.cpu.clock_task.max
     44318 ±  2%     +44.9%      64224 ± 23%  sched_debug.cpu.clock_task.min
    694.52            +1.2%     702.85        sched_debug.cpu.clock_task.stddev
    446.71 ± 17%    +204.4%       1359 ± 63%  sched_debug.cpu.curr->pid.avg
      4227           +14.3%       4831 ± 12%  sched_debug.cpu.curr->pid.max
      1253 ±  8%     +21.9%       1528 ± 13%  sched_debug.cpu.curr->pid.stddev
    500332            -0.1%     500000        sched_debug.cpu.max_idle_balance_cost.avg
    542508 ±  2%      -7.8%     500000        sched_debug.cpu.max_idle_balance_cost.max
    500000            +0.0%     500000        sched_debug.cpu.max_idle_balance_cost.min
      3742 ± 28%    -100.0%       0.00        sched_debug.cpu.max_idle_balance_cost.stddev
      4294            +0.0%       4294        sched_debug.cpu.next_balance.avg
      4294            +0.0%       4294        sched_debug.cpu.next_balance.max
      4294            +0.0%       4294        sched_debug.cpu.next_balance.min
      0.00 ±  2%     -16.8%       0.00 ± 16%  sched_debug.cpu.next_balance.stddev
      0.12 ± 26%    +140.0%       0.28 ± 55%  sched_debug.cpu.nr_running.avg
      2.00 ± 50%     -50.0%       1.00        sched_debug.cpu.nr_running.max
      0.35 ± 19%      +0.3%       0.35 ±  5%  sched_debug.cpu.nr_running.stddev
      1738 ±  2%     +26.0%       2190 ± 18%  sched_debug.cpu.nr_switches.avg
     45570            -8.1%      41889 ± 27%  sched_debug.cpu.nr_switches.max
    135.50 ± 15%     +65.5%     224.25 ± 32%  sched_debug.cpu.nr_switches.min
      4624 ±  2%      +6.1%       4908 ±  9%  sched_debug.cpu.nr_switches.stddev
     52184 ±  2%     +38.6%      72304 ± 21%  sched_debug.cpu_clk
    996147            +0.0%     996147        sched_debug.dl_rq:.dl_bw->bw.avg
    996147            +0.0%     996147        sched_debug.dl_rq:.dl_bw->bw.max
    996147            +0.0%     996147        sched_debug.dl_rq:.dl_bw->bw.min
 4.295e+09            +0.0%  4.295e+09        sched_debug.jiffies
     50954 ±  2%     +39.5%      71075 ± 21%  sched_debug.ktime
    950.00            +0.0%     950.00        sched_debug.rt_rq:.rt_runtime.avg
    950.00            +0.0%     950.00        sched_debug.rt_rq:.rt_runtime.max
    950.00            +0.0%     950.00        sched_debug.rt_rq:.rt_runtime.min
     53048 ±  2%     +38.1%      73236 ± 20%  sched_debug.sched_clk
      1.00            +0.0%       1.00        sched_debug.sched_clock_stable()
      3.00            +0.0%       3.00        sched_debug.sysctl_sched.sysctl_sched_base_slice
  25056823            +0.0%   25056823        sched_debug.sysctl_sched.sysctl_sched_features
      1.00            +0.0%       1.00        sched_debug.sysctl_sched.sysctl_sched_tunable_scaling
      0.00          +3e+99%       0.00 ±100%  perf-sched.sch_delay.avg.ms.__cond_resched.__kmem_cache_alloc_node.kmalloc_trace.single_open.do_dentry_open
      0.01 ± 22%      +5.6%       0.01 ± 15%  perf-sched.sch_delay.avg.ms.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
      0.00 ±100%     +57.1%       0.01 ±  9%  perf-sched.sch_delay.avg.ms.__cond_resched.stop_one_cpu.sched_exec.bprm_execve.part
      0.01 ± 17%     -35.3%       0.01 ±  9%  perf-sched.sch_delay.avg.ms.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
      0.00 ±100%    -100.0%       0.00        perf-sched.sch_delay.avg.ms.d_alloc_parallel.__lookup_slow.walk_component.link_path_walk.part
      0.01            +8.3%       0.01 ±  7%  perf-sched.sch_delay.avg.ms.devkmsg_read.vfs_read.ksys_read.do_syscall_64
      0.01 ± 17%     -17.6%       0.01        perf-sched.sch_delay.avg.ms.do_nanosleep.hrtimer_nanosleep.common_nsleep.__x64_sys_clock_nanosleep
      0.01 ±  7%  +2.3e+05%      15.16 ± 99%  perf-sched.sch_delay.avg.ms.do_task_dead.do_exit.__x64_sys_exit.do_syscall_64.entry_SYSCALL_64_after_hwframe
      0.00 ± 11%      +0.0%       0.00 ± 11%  perf-sched.sch_delay.avg.ms.do_task_dead.do_exit.do_group_exit.__x64_sys_exit_group.do_syscall_64
      0.01 ±  4%     +19.0%       0.01 ± 36%  perf-sched.sch_delay.avg.ms.do_wait.kernel_wait4.__do_sys_wait4.do_syscall_64
      0.01 ±100%    -100.0%       0.00        perf-sched.sch_delay.avg.ms.futex_wait_queue.futex_wait.do_futex.__x64_sys_futex
      0.00        +1.5e+99%       0.00 ±100%  perf-sched.sch_delay.avg.ms.ipmi_thread.kthread.ret_from_fork.ret_from_fork_asm
      0.01 ±100%    -100.0%       0.00        perf-sched.sch_delay.avg.ms.kthreadd.ret_from_fork.ret_from_fork_asm
      6.18 ± 34%    +245.2%      21.34 ± 43%  perf-sched.sch_delay.avg.ms.pipe_read.vfs_read.ksys_read.do_syscall_64
      0.01 ± 14%     -14.3%       0.01        perf-sched.sch_delay.avg.ms.rcu_gp_kthread.kthread.ret_from_fork.ret_from_fork_asm
      0.01 ± 20%     -20.0%       0.00        perf-sched.sch_delay.avg.ms.schedule_hrtimeout_range_clock.do_poll.constprop.0.do_sys_poll
      0.01 ± 26%    +163.2%       0.02 ± 56%  perf-sched.sch_delay.avg.ms.schedule_hrtimeout_range_clock.do_select.core_sys_select.kern_select
      0.01 ±  5%     -29.4%       0.01        perf-sched.sch_delay.avg.ms.schedule_hrtimeout_range_clock.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
      0.00        +1.5e+99%       0.00 ±100%  perf-sched.sch_delay.avg.ms.schedule_hrtimeout_range_clock.usleep_range_state.ipmi_thread.kthread
      0.00 ±100%    -100.0%       0.00        perf-sched.sch_delay.avg.ms.schedule_preempt_disabled.__mutex_lock.constprop.0.cgroup_kn_lock_live
      0.00 ±100%    -100.0%       0.00        perf-sched.sch_delay.avg.ms.schedule_preempt_disabled.kthread.ret_from_fork.ret_from_fork_asm
      0.00        +6.5e+99%       0.01 ±100%  perf-sched.sch_delay.avg.ms.schedule_timeout.__wait_for_common.__synchronize_srcu._ipmi_destroy_user
      0.00 ±100%    -100.0%       0.00        perf-sched.sch_delay.avg.ms.schedule_timeout.__wait_for_common.__synchronize_srcu.fsnotify_connector_destroy_workfn
      0.00 ±100%    -100.0%       0.00        perf-sched.sch_delay.avg.ms.schedule_timeout.__wait_for_common.__synchronize_srcu.fsnotify_mark_destroy_workfn
      0.01 ±100%     -55.0%       0.00 ±100%  perf-sched.sch_delay.avg.ms.schedule_timeout.__wait_for_common.__wait_rcu_gp.synchronize_rcu
      0.00 ±100%    -100.0%       0.00        perf-sched.sch_delay.avg.ms.schedule_timeout.__wait_for_common.wait_for_completion_killable.__kthread_create_on_node
      0.02 ± 60%     -43.3%       0.01 ± 17%  perf-sched.sch_delay.avg.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone
      0.01 ± 16%      -8.3%       0.01 ±  9%  perf-sched.sch_delay.avg.ms.schedule_timeout.kcompactd.kthread.ret_from_fork
      0.01 ± 52%    +170.6%       0.02 ± 82%  perf-sched.sch_delay.avg.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
      0.00 ±100%    -100.0%       0.00        perf-sched.sch_delay.avg.ms.schedule_timeout.tty_wait_until_sent.tty_port_close_start.tty_port_close
      0.01 ±  7%      -7.7%       0.01        perf-sched.sch_delay.avg.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
      0.01 ±  5%     +10.5%       0.01 ±  4%  perf-sched.sch_delay.avg.ms.syslog_print.do_syslog.kmsg_read.vfs_read
      0.01 ±  7%     -30.8%       0.00 ± 11%  perf-sched.sch_delay.avg.ms.wait_for_partner.fifo_open.do_dentry_open.do_open
      2.74 ± 98%     -96.6%       0.09 ±  2%  perf-sched.sch_delay.avg.ms.worker_thread.kthread.ret_from_fork.ret_from_fork_asm
      0.00          +3e+99%       0.00 ±100%  perf-sched.sch_delay.max.ms.__cond_resched.__kmem_cache_alloc_node.kmalloc_trace.single_open.do_dentry_open
      0.01 ± 48%     -44.4%       0.01 ±  6%  perf-sched.sch_delay.max.ms.__cond_resched.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr.__sched_setaffinity
      0.02 ±  2%     -34.3%       0.01 ±  4%  perf-sched.sch_delay.max.ms.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
      0.00 ±100%     +75.0%       0.01 ± 14%  perf-sched.sch_delay.max.ms.__cond_resched.stop_one_cpu.sched_exec.bprm_execve.part
      0.02 ± 33%     -66.7%       0.01        perf-sched.sch_delay.max.ms.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
      0.00 ±100%    -100.0%       0.00        perf-sched.sch_delay.max.ms.d_alloc_parallel.__lookup_slow.walk_component.link_path_walk.part
      0.01           +29.2%       0.02 ± 22%  perf-sched.sch_delay.max.ms.devkmsg_read.vfs_read.ksys_read.do_syscall_64
      0.01 ±  3%     -11.1%       0.01 ±  8%  perf-sched.sch_delay.max.ms.do_nanosleep.hrtimer_nanosleep.common_nsleep.__x64_sys_clock_nanosleep
      0.02 ± 17%  +2.9e+06%     500.11 ± 99%  perf-sched.sch_delay.max.ms.do_task_dead.do_exit.__x64_sys_exit.do_syscall_64.entry_SYSCALL_64_after_hwframe
      0.02 ± 11%      -2.9%       0.02 ± 27%  perf-sched.sch_delay.max.ms.do_task_dead.do_exit.do_group_exit.__x64_sys_exit_group.do_syscall_64
      0.05 ± 47%    +740.6%       0.45 ± 95%  perf-sched.sch_delay.max.ms.do_wait.kernel_wait4.__do_sys_wait4.do_syscall_64
      0.01 ± 15%     +26.9%       0.02 ± 21%  perf-sched.sch_delay.max.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.syscall_exit_to_user_mode.do_syscall_64
      0.01 ±100%    -100.0%       0.00        perf-sched.sch_delay.max.ms.futex_wait_queue.futex_wait.do_futex.__x64_sys_futex
      0.00        +1.5e+99%       0.00 ±100%  perf-sched.sch_delay.max.ms.ipmi_thread.kthread.ret_from_fork.ret_from_fork_asm
      0.01 ±100%    -100.0%       0.00        perf-sched.sch_delay.max.ms.kthreadd.ret_from_fork.ret_from_fork_asm
      1002            +0.1%       1003        perf-sched.sch_delay.max.ms.pipe_read.vfs_read.ksys_read.do_syscall_64
      0.02 ±  3%     -51.5%       0.01        perf-sched.sch_delay.max.ms.rcu_gp_kthread.kthread.ret_from_fork.ret_from_fork_asm
      0.02 ±  5%     -47.1%       0.01 ± 22%  perf-sched.sch_delay.max.ms.schedule_hrtimeout_range_clock.do_poll.constprop.0.do_sys_poll
      0.01 ± 30%    +303.8%       0.05 ± 73%  perf-sched.sch_delay.max.ms.schedule_hrtimeout_range_clock.do_select.core_sys_select.kern_select
      0.02           -56.8%       0.01 ±  5%  perf-sched.sch_delay.max.ms.schedule_hrtimeout_range_clock.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
      0.00          +8e+99%       0.01 ±100%  perf-sched.sch_delay.max.ms.schedule_hrtimeout_range_clock.usleep_range_state.ipmi_thread.kthread
      0.00 ±100%    -100.0%       0.00        perf-sched.sch_delay.max.ms.schedule_preempt_disabled.__mutex_lock.constprop.0.cgroup_kn_lock_live
      0.00 ±100%    -100.0%       0.00        perf-sched.sch_delay.max.ms.schedule_preempt_disabled.kthread.ret_from_fork.ret_from_fork_asm
      0.00        +6.5e+99%       0.01 ±100%  perf-sched.sch_delay.max.ms.schedule_timeout.__wait_for_common.__synchronize_srcu._ipmi_destroy_user
      0.01 ±100%    -100.0%       0.00        perf-sched.sch_delay.max.ms.schedule_timeout.__wait_for_common.__synchronize_srcu.fsnotify_connector_destroy_workfn
      0.00 ±100%    -100.0%       0.00        perf-sched.sch_delay.max.ms.schedule_timeout.__wait_for_common.__synchronize_srcu.fsnotify_mark_destroy_workfn
      0.02 ±100%     -80.9%       0.00 ±100%  perf-sched.sch_delay.max.ms.schedule_timeout.__wait_for_common.__wait_rcu_gp.synchronize_rcu
      0.00 ±100%    -100.0%       0.00        perf-sched.sch_delay.max.ms.schedule_timeout.__wait_for_common.wait_for_completion_killable.__kthread_create_on_node
      0.52 ± 96%     -78.6%       0.11 ±  9%  perf-sched.sch_delay.max.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone
      0.02 ± 37%     -48.9%       0.01 ±  4%  perf-sched.sch_delay.max.ms.schedule_timeout.kcompactd.kthread.ret_from_fork
      2.51 ± 99%    +119.0%       5.51 ± 99%  perf-sched.sch_delay.max.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
      0.01 ±100%    -100.0%       0.00        perf-sched.sch_delay.max.ms.schedule_timeout.tty_wait_until_sent.tty_port_close_start.tty_port_close
      0.03 ± 29%     -43.5%       0.02 ± 31%  perf-sched.sch_delay.max.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
      0.01 ± 25%      +4.2%       0.01 ± 12%  perf-sched.sch_delay.max.ms.syslog_print.do_syslog.kmsg_read.vfs_read
      0.02 ± 40%     -67.6%       0.01        perf-sched.sch_delay.max.ms.wait_for_partner.fifo_open.do_dentry_open.do_open
    801.97 ± 99%     -99.5%       4.00        perf-sched.sch_delay.max.ms.worker_thread.kthread.ret_from_fork.ret_from_fork_asm
      0.56 ± 54%     +65.6%       0.93 ± 18%  perf-sched.total_sch_delay.average.ms
      1300 ± 23%     -22.8%       1003        perf-sched.total_sch_delay.max.ms
    225.44           -25.2%     168.57 ± 47%  perf-sched.total_wait_and_delay.average.ms
      4145           +57.2%       6518 ± 43%  perf-sched.total_wait_and_delay.count.ms
      4925           -24.5%       3720 ± 31%  perf-sched.total_wait_and_delay.max.ms
    224.88           -25.5%     167.64 ± 48%  perf-sched.total_wait_time.average.ms
      4925           -24.5%       3720 ± 31%  perf-sched.total_wait_time.max.ms
      0.00       +1.4e+100%       0.01 ±100%  perf-sched.wait_and_delay.avg.ms.__cond_resched.__alloc_pages.__alloc_pages_bulk.copy_splice_read.splice_file_to_pipe
     11.94            -6.7%      11.14 ±  6%  perf-sched.wait_and_delay.avg.ms.__cond_resched.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr.__sched_setaffinity
      0.02 ±100%     -23.1%       0.02 ±100%  perf-sched.wait_and_delay.avg.ms.__cond_resched.mutex_lock.pipe_double_lock.splice_pipe_to_pipe.do_splice
      0.00       +1.4e+100%       0.01 ±100%  perf-sched.wait_and_delay.avg.ms.__cond_resched.read_iter_zero.copy_splice_read.splice_file_to_pipe.do_splice
    435.70 ± 14%     -70.3%     129.43 ±100%  perf-sched.wait_and_delay.avg.ms.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
    799.59            +0.0%     799.59        perf-sched.wait_and_delay.avg.ms.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
    562.97 ± 11%     -11.2%     500.03        perf-sched.wait_and_delay.avg.ms.do_nanosleep.hrtimer_nanosleep.common_nsleep.__x64_sys_clock_nanosleep
    743.65            +3.8%     772.28 ±  5%  perf-sched.wait_and_delay.avg.ms.do_task_dead.do_exit.__x64_sys_exit.do_syscall_64.entry_SYSCALL_64_after_hwframe
     29.78 ± 52%     +37.1%      40.81 ± 25%  perf-sched.wait_and_delay.avg.ms.do_task_dead.do_exit.do_group_exit.__x64_sys_exit_group.do_syscall_64
      0.45 ±  3%     -21.9%       0.35 ±  3%  perf-sched.wait_and_delay.avg.ms.do_wait.kernel_wait4.__do_sys_wait4.do_syscall_64
      0.04 ± 16%     -31.2%       0.03 ±  5%  perf-sched.wait_and_delay.avg.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.syscall_exit_to_user_mode.do_syscall_64
    283.17            +2.5%     290.38 ±  7%  perf-sched.wait_and_delay.avg.ms.pipe_read.vfs_read.ksys_read.do_syscall_64
    536.23 ±  6%     +14.6%     614.32 ±  6%  perf-sched.wait_and_delay.avg.ms.schedule_hrtimeout_range_clock.do_poll.constprop.0.do_sys_poll
    291.41 ± 30%     +26.4%     368.24        perf-sched.wait_and_delay.avg.ms.schedule_hrtimeout_range_clock.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
      0.00       +9.8e+100%       0.10 ±100%  perf-sched.wait_and_delay.avg.ms.schedule_hrtimeout_range_clock.usleep_range_state.ipmi_thread.kthread
      0.00       +5.9e+104%     592.95 ±100%  perf-sched.wait_and_delay.avg.ms.schedule_timeout.__wait_for_common.__synchronize_srcu._ipmi_destroy_user
    145.35 ±100%    -100.0%       0.00        perf-sched.wait_and_delay.avg.ms.schedule_timeout.__wait_for_common.__synchronize_srcu.fsnotify_connector_destroy_workfn
    115.31 ±100%    -100.0%       0.00        perf-sched.wait_and_delay.avg.ms.schedule_timeout.__wait_for_common.__wait_rcu_gp.synchronize_rcu
      0.83 ± 11%    +855.9%       7.95 ± 92%  perf-sched.wait_and_delay.avg.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone
    453.62            -0.0%     453.57        perf-sched.wait_and_delay.avg.ms.schedule_timeout.kcompactd.kthread.ret_from_fork
      8.16 ±  7%    +109.7%      17.11        perf-sched.wait_and_delay.avg.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
    669.80            -2.3%     654.65 ±  2%  perf-sched.wait_and_delay.avg.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
      0.01 ± 14%     -64.3%       0.00 ±100%  perf-sched.wait_and_delay.avg.ms.wait_for_partner.fifo_open.do_dentry_open.do_open
    461.67            +4.7%     483.54 ±  2%  perf-sched.wait_and_delay.avg.ms.worker_thread.kthread.ret_from_fork.ret_from_fork_asm
      0.00       +1.4e+103%      14.00 ±100%  perf-sched.wait_and_delay.count.__cond_resched.__alloc_pages.__alloc_pages_bulk.copy_splice_read.splice_file_to_pipe
    768.00            -8.3%     704.50 ±  9%  perf-sched.wait_and_delay.count.__cond_resched.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr.__sched_setaffinity
     11.00 ±100%     +68.2%      18.50 ±100%  perf-sched.wait_and_delay.count.__cond_resched.mutex_lock.pipe_double_lock.splice_pipe_to_pipe.do_splice
      0.00       +2.4e+103%      24.00 ±100%  perf-sched.wait_and_delay.count.__cond_resched.read_iter_zero.copy_splice_read.splice_file_to_pipe.do_splice
     10.00 ± 40%     -60.0%       4.00 ±100%  perf-sched.wait_and_delay.count.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
      5.00            +0.0%       5.00        perf-sched.wait_and_delay.count.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
      8.00           -25.0%       6.00        perf-sched.wait_and_delay.count.do_nanosleep.hrtimer_nanosleep.common_nsleep.__x64_sys_clock_nanosleep
     37.00            -8.1%      34.00 ±  2%  perf-sched.wait_and_delay.count.do_task_dead.do_exit.__x64_sys_exit.do_syscall_64.entry_SYSCALL_64_after_hwframe
     88.50            -6.8%      82.50 ±  5%  perf-sched.wait_and_delay.count.do_task_dead.do_exit.do_group_exit.__x64_sys_exit_group.do_syscall_64
    123.50 ±  2%      -8.1%     113.50 ±  2%  perf-sched.wait_and_delay.count.do_wait.kernel_wait4.__do_sys_wait4.do_syscall_64
    691.50           -28.9%     491.50 ±  7%  perf-sched.wait_and_delay.count.exit_to_user_mode_loop.exit_to_user_mode_prepare.syscall_exit_to_user_mode.do_syscall_64
    245.00            -2.2%     239.50 ±  4%  perf-sched.wait_and_delay.count.pipe_read.vfs_read.ksys_read.do_syscall_64
     22.50 ±  6%      -4.4%      21.50 ±  2%  perf-sched.wait_and_delay.count.schedule_hrtimeout_range_clock.do_poll.constprop.0.do_sys_poll
     23.50 ± 44%     -44.7%      13.00        perf-sched.wait_and_delay.count.schedule_hrtimeout_range_clock.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
      0.00         +3e+105%       2964 ±100%  perf-sched.wait_and_delay.count.schedule_hrtimeout_range_clock.usleep_range_state.ipmi_thread.kthread
      0.00         +5e+101%       0.50 ±100%  perf-sched.wait_and_delay.count.schedule_timeout.__wait_for_common.__synchronize_srcu._ipmi_destroy_user
      1.00 ±100%    -100.0%       0.00        perf-sched.wait_and_delay.count.schedule_timeout.__wait_for_common.__synchronize_srcu.fsnotify_connector_destroy_workfn
      1.50 ±100%    -100.0%       0.00        perf-sched.wait_and_delay.count.schedule_timeout.__wait_for_common.__wait_rcu_gp.synchronize_rcu
     88.00            -8.0%      81.00 ±  2%  perf-sched.wait_and_delay.count.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone
     20.00            +0.0%      20.00        perf-sched.wait_and_delay.count.schedule_timeout.kcompactd.kthread.ret_from_fork
    610.00 ±  7%     -52.8%     288.00        perf-sched.wait_and_delay.count.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
    971.00 ±  2%      -9.2%     881.50 ±  7%  perf-sched.wait_and_delay.count.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
     36.50           -57.5%      15.50 ±100%  perf-sched.wait_and_delay.count.wait_for_partner.fifo_open.do_dentry_open.do_open
    282.00 ±  6%      +5.9%     298.50 ±  4%  perf-sched.wait_and_delay.count.worker_thread.kthread.ret_from_fork.ret_from_fork_asm
      0.00       +2.9e+100%       0.03 ±100%  perf-sched.wait_and_delay.max.ms.__cond_resched.__alloc_pages.__alloc_pages_bulk.copy_splice_read.splice_file_to_pipe
      4925           -25.4%       3673 ± 33%  perf-sched.wait_and_delay.max.ms.__cond_resched.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr.__sched_setaffinity
      0.03 ±100%     -23.4%       0.02 ±100%  perf-sched.wait_and_delay.max.ms.__cond_resched.mutex_lock.pipe_double_lock.splice_pipe_to_pipe.do_splice
      0.00         +3e+100%       0.03 ±100%  perf-sched.wait_and_delay.max.ms.__cond_resched.read_iter_zero.copy_splice_read.splice_file_to_pipe.do_splice
      2045 ± 45%     -75.6%     498.97 ±100%  perf-sched.wait_and_delay.max.ms.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
    999.52            +0.0%     999.53        perf-sched.wait_and_delay.max.ms.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
      1000            -0.0%       1000        perf-sched.wait_and_delay.max.ms.do_nanosleep.hrtimer_nanosleep.common_nsleep.__x64_sys_clock_nanosleep
      1000           +76.9%       1769 ± 43%  perf-sched.wait_and_delay.max.ms.do_task_dead.do_exit.__x64_sys_exit.do_syscall_64.entry_SYSCALL_64_after_hwframe
    984.18           +63.5%       1609 ± 24%  perf-sched.wait_and_delay.max.ms.do_task_dead.do_exit.do_group_exit.__x64_sys_exit_group.do_syscall_64
      1.77 ±  6%     -12.9%       1.54 ± 20%  perf-sched.wait_and_delay.max.ms.do_wait.kernel_wait4.__do_sys_wait4.do_syscall_64
      0.30 ± 37%     -54.2%       0.14 ± 16%  perf-sched.wait_and_delay.max.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.syscall_exit_to_user_mode.do_syscall_64
      1533 ± 30%     +30.7%       2004        perf-sched.wait_and_delay.max.ms.pipe_read.vfs_read.ksys_read.do_syscall_64
      1000           +54.6%       1547 ± 35%  perf-sched.wait_and_delay.max.ms.schedule_hrtimeout_range_clock.do_poll.constprop.0.do_sys_poll
      1258 ± 60%     -60.3%     499.99        perf-sched.wait_and_delay.max.ms.schedule_hrtimeout_range_clock.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
      0.00       +1.1e+101%       0.11 ±100%  perf-sched.wait_and_delay.max.ms.schedule_hrtimeout_range_clock.usleep_range_state.ipmi_thread.kthread
      0.00       +5.9e+104%     592.95 ±100%  perf-sched.wait_and_delay.max.ms.schedule_timeout.__wait_for_common.__synchronize_srcu._ipmi_destroy_user
    290.69 ±100%    -100.0%       0.00        perf-sched.wait_and_delay.max.ms.schedule_timeout.__wait_for_common.__synchronize_srcu.fsnotify_connector_destroy_workfn
    249.99 ±100%    -100.0%       0.00        perf-sched.wait_and_delay.max.ms.schedule_timeout.__wait_for_common.__wait_rcu_gp.synchronize_rcu
      2.86 ± 19%  +21211.2%     608.43 ± 99%  perf-sched.wait_and_delay.max.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone
    505.00            -0.1%     504.50        perf-sched.wait_and_delay.max.ms.schedule_timeout.kcompactd.kthread.ret_from_fork
    512.50 ± 19%      -8.4%     469.51 ±  2%  perf-sched.wait_and_delay.max.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
      2568 ± 19%     -39.3%       1560 ± 31%  perf-sched.wait_and_delay.max.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
      0.02 ± 40%     -83.8%       0.00 ±100%  perf-sched.wait_and_delay.max.ms.wait_for_partner.fifo_open.do_dentry_open.do_open
      2722 ±  5%     -17.7%       2241 ±  5%  perf-sched.wait_and_delay.max.ms.worker_thread.kthread.ret_from_fork.ret_from_fork_asm
      0.00         +3e+100%       0.03 ±  3%  perf-sched.wait_time.avg.ms.__cond_resched.__alloc_pages.__alloc_pages_bulk.copy_splice_read.splice_file_to_pipe
      0.03          -100.0%       0.00        perf-sched.wait_time.avg.ms.__cond_resched.__alloc_pages.pipe_write.vfs_write.ksys_write
      0.00       +2.8e+100%       0.03 ±  5%  perf-sched.wait_time.avg.ms.__cond_resched.__kmem_cache_alloc_node.__kmalloc.copy_splice_read.splice_file_to_pipe
      0.02 ± 29%     +58.3%       0.04 ± 13%  perf-sched.wait_time.avg.ms.__cond_resched.__splice_from_pipe.splice_from_pipe.do_splice.__do_splice
     11.94            -6.7%      11.14 ±  6%  perf-sched.wait_time.avg.ms.__cond_resched.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr.__sched_setaffinity
      0.05 ± 21%     -36.0%       0.03 ±  6%  perf-sched.wait_time.avg.ms.__cond_resched.mutex_lock.pipe_double_lock.splice_pipe_to_pipe.do_splice
      0.04 ± 11%    -100.0%       0.00        perf-sched.wait_time.avg.ms.__cond_resched.mutex_lock.pipe_write.vfs_write.ksys_write
      0.00         +3e+100%       0.03 ± 52%  perf-sched.wait_time.avg.ms.__cond_resched.mutex_lock.splice_file_to_pipe.do_splice.__do_splice
      0.06 ± 43%     -45.4%       0.04 ± 12%  perf-sched.wait_time.avg.ms.__cond_resched.mutex_lock.splice_from_pipe.do_splice.__do_splice
      0.06 ± 16%     -65.3%       0.02 ± 16%  perf-sched.wait_time.avg.ms.__cond_resched.mutex_lock.splice_pipe_to_pipe.do_splice.__do_splice
      0.00       +2.6e+100%       0.03 ±  9%  perf-sched.wait_time.avg.ms.__cond_resched.read_iter_zero.copy_splice_read.splice_file_to_pipe.do_splice
    435.69 ± 14%     -70.3%     129.43 ±100%  perf-sched.wait_time.avg.ms.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
    799.58            +0.0%     799.59        perf-sched.wait_time.avg.ms.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
      2.47 ±  2%     -24.3%       1.87 ±  3%  perf-sched.wait_time.avg.ms.devkmsg_read.vfs_read.ksys_read.do_syscall_64
    562.96 ± 11%     -11.2%     500.03        perf-sched.wait_time.avg.ms.do_nanosleep.hrtimer_nanosleep.common_nsleep.__x64_sys_clock_nanosleep
    743.64            +1.8%     757.11 ±  7%  perf-sched.wait_time.avg.ms.do_task_dead.do_exit.__x64_sys_exit.do_syscall_64.entry_SYSCALL_64_after_hwframe
     29.77 ± 52%     +37.1%      40.81 ± 25%  perf-sched.wait_time.avg.ms.do_task_dead.do_exit.do_group_exit.__x64_sys_exit_group.do_syscall_64
      0.44 ±  3%     -22.9%       0.34 ±  4%  perf-sched.wait_time.avg.ms.do_wait.kernel_wait4.__do_sys_wait4.do_syscall_64
      0.04 ± 16%     -31.2%       0.03 ±  5%  perf-sched.wait_time.avg.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.syscall_exit_to_user_mode.do_syscall_64
      0.42 ±100%    -100.0%       0.00        perf-sched.wait_time.avg.ms.futex_wait_queue.futex_wait.do_futex.__x64_sys_futex
      0.00         +1e+101%       0.10 ±100%  perf-sched.wait_time.avg.ms.ipmi_thread.kthread.ret_from_fork.ret_from_fork_asm
    276.99            -2.9%     269.04 ±  4%  perf-sched.wait_time.avg.ms.pipe_read.vfs_read.ksys_read.do_syscall_64
      2.92 ±  7%     -13.9%       2.52 ± 10%  perf-sched.wait_time.avg.ms.rcu_gp_kthread.kthread.ret_from_fork.ret_from_fork_asm
    536.22 ±  6%     +14.6%     614.32 ±  6%  perf-sched.wait_time.avg.ms.schedule_hrtimeout_range_clock.do_poll.constprop.0.do_sys_poll
    291.40 ± 30%     +26.4%     368.23        perf-sched.wait_time.avg.ms.schedule_hrtimeout_range_clock.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
      0.00       +9.6e+100%       0.10 ±100%  perf-sched.wait_time.avg.ms.schedule_hrtimeout_range_clock.usleep_range_state.ipmi_thread.kthread
      0.00       +5.9e+104%     592.95 ±100%  perf-sched.wait_time.avg.ms.schedule_timeout.__wait_for_common.__synchronize_srcu._ipmi_destroy_user
    145.35 ±100%    -100.0%       0.00        perf-sched.wait_time.avg.ms.schedule_timeout.__wait_for_common.__synchronize_srcu.fsnotify_connector_destroy_workfn
      1.10 ±100%    -100.0%       0.00        perf-sched.wait_time.avg.ms.schedule_timeout.__wait_for_common.__synchronize_srcu.fsnotify_mark_destroy_workfn
    115.30 ±100%    -100.0%       0.01 ±100%  perf-sched.wait_time.avg.ms.schedule_timeout.__wait_for_common.__wait_rcu_gp.synchronize_rcu
      0.01 ±100%    -100.0%       0.00        perf-sched.wait_time.avg.ms.schedule_timeout.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr
      0.82 ± 10%    +872.5%       7.94 ± 92%  perf-sched.wait_time.avg.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone
    453.61            -0.0%     453.57        perf-sched.wait_time.avg.ms.schedule_timeout.kcompactd.kthread.ret_from_fork
      8.15 ±  7%    +109.7%      17.09        perf-sched.wait_time.avg.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
      9.38 ±100%    -100.0%       0.00        perf-sched.wait_time.avg.ms.schedule_timeout.tty_wait_until_sent.tty_port_close_start.tty_port_close
    669.79            -2.3%     654.65 ±  2%  perf-sched.wait_time.avg.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
      2.48 ±  2%     -24.3%       1.88 ±  3%  perf-sched.wait_time.avg.ms.syslog_print.do_syslog.kmsg_read.vfs_read
      0.00 ±100%      +0.0%       0.00 ±100%  perf-sched.wait_time.avg.ms.wait_for_partner.fifo_open.do_dentry_open.do_open
    458.93 ±  2%      +5.3%     483.45 ±  2%  perf-sched.wait_time.avg.ms.worker_thread.kthread.ret_from_fork.ret_from_fork_asm
      0.00       +5.1e+100%       0.05 ± 13%  perf-sched.wait_time.max.ms.__cond_resched.__alloc_pages.__alloc_pages_bulk.copy_splice_read.splice_file_to_pipe
      0.07          -100.0%       0.00        perf-sched.wait_time.max.ms.__cond_resched.__alloc_pages.pipe_write.vfs_write.ksys_write
      0.00       +4.6e+100%       0.05 ±  4%  perf-sched.wait_time.max.ms.__cond_resched.__kmem_cache_alloc_node.__kmalloc.copy_splice_read.splice_file_to_pipe
      0.05            -8.4%       0.04        perf-sched.wait_time.max.ms.__cond_resched.__splice_from_pipe.splice_from_pipe.do_splice.__do_splice
      4925           -25.4%       3673 ± 33%  perf-sched.wait_time.max.ms.__cond_resched.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr.__sched_setaffinity
      0.23 ± 72%     -79.1%       0.05        perf-sched.wait_time.max.ms.__cond_resched.mutex_lock.pipe_double_lock.splice_pipe_to_pipe.do_splice
      0.07 ± 11%    -100.0%       0.00        perf-sched.wait_time.max.ms.__cond_resched.mutex_lock.pipe_write.vfs_write.ksys_write
      0.00       +1.1e+101%       0.11 ± 60%  perf-sched.wait_time.max.ms.__cond_resched.mutex_lock.splice_file_to_pipe.do_splice.__do_splice
      0.23 ± 75%     -63.8%       0.08 ± 35%  perf-sched.wait_time.max.ms.__cond_resched.mutex_lock.splice_from_pipe.do_splice.__do_splice
      0.16           -73.6%       0.04 ±  4%  perf-sched.wait_time.max.ms.__cond_resched.mutex_lock.splice_pipe_to_pipe.do_splice.__do_splice
      0.00       +5.5e+100%       0.06 ±  7%  perf-sched.wait_time.max.ms.__cond_resched.read_iter_zero.copy_splice_read.splice_file_to_pipe.do_splice
      2045 ± 45%     -75.6%     498.96 ±100%  perf-sched.wait_time.max.ms.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
    999.51            +0.0%     999.52        perf-sched.wait_time.max.ms.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
      4.95 ±  2%     -24.3%       3.74 ±  3%  perf-sched.wait_time.max.ms.devkmsg_read.vfs_read.ksys_read.do_syscall_64
      1000            -0.0%       1000        perf-sched.wait_time.max.ms.do_nanosleep.hrtimer_nanosleep.common_nsleep.__x64_sys_clock_nanosleep
      1000           +76.9%       1768 ± 43%  perf-sched.wait_time.max.ms.do_task_dead.do_exit.__x64_sys_exit.do_syscall_64.entry_SYSCALL_64_after_hwframe
    984.18           +63.5%       1609 ± 24%  perf-sched.wait_time.max.ms.do_task_dead.do_exit.do_group_exit.__x64_sys_exit_group.do_syscall_64
      1.75 ±  6%     -35.0%       1.14 ±  5%  perf-sched.wait_time.max.ms.do_wait.kernel_wait4.__do_sys_wait4.do_syscall_64
      0.30 ± 37%     -54.2%       0.14 ± 16%  perf-sched.wait_time.max.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.syscall_exit_to_user_mode.do_syscall_64
      1.06 ±100%    -100.0%       0.00        perf-sched.wait_time.max.ms.futex_wait_queue.futex_wait.do_futex.__x64_sys_futex
      0.00         +1e+101%       0.10 ±100%  perf-sched.wait_time.max.ms.ipmi_thread.kthread.ret_from_fork.ret_from_fork_asm
      1059            +0.0%       1059        perf-sched.wait_time.max.ms.pipe_read.vfs_read.ksys_read.do_syscall_64
      4.99            -9.9%       4.49 ± 11%  perf-sched.wait_time.max.ms.rcu_gp_kthread.kthread.ret_from_fork.ret_from_fork_asm
      1000           +54.6%       1547 ± 35%  perf-sched.wait_time.max.ms.schedule_hrtimeout_range_clock.do_poll.constprop.0.do_sys_poll
      1258 ± 60%     -60.3%     499.98        perf-sched.wait_time.max.ms.schedule_hrtimeout_range_clock.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
      0.00       +1.0e+101%       0.10 ±100%  perf-sched.wait_time.max.ms.schedule_hrtimeout_range_clock.usleep_range_state.ipmi_thread.kthread
      0.00       +5.9e+104%     592.95 ±100%  perf-sched.wait_time.max.ms.schedule_timeout.__wait_for_common.__synchronize_srcu._ipmi_destroy_user
    290.68 ±100%    -100.0%       0.00        perf-sched.wait_time.max.ms.schedule_timeout.__wait_for_common.__synchronize_srcu.fsnotify_connector_destroy_workfn
      1.10 ±100%    -100.0%       0.00        perf-sched.wait_time.max.ms.schedule_timeout.__wait_for_common.__synchronize_srcu.fsnotify_mark_destroy_workfn
    249.99 ±100%    -100.0%       0.01 ±100%  perf-sched.wait_time.max.ms.schedule_timeout.__wait_for_common.__wait_rcu_gp.synchronize_rcu
      0.02 ±100%    -100.0%       0.00        perf-sched.wait_time.max.ms.schedule_timeout.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr
      2.33 ±  2%  +25962.4%     608.43 ± 99%  perf-sched.wait_time.max.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone
    504.99            -0.1%     504.50        perf-sched.wait_time.max.ms.schedule_timeout.kcompactd.kthread.ret_from_fork
    512.49 ± 19%      -8.4%     469.50 ±  2%  perf-sched.wait_time.max.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
    105.74 ±100%    -100.0%       0.00        perf-sched.wait_time.max.ms.schedule_timeout.tty_wait_until_sent.tty_port_close_start.tty_port_close
      2568 ± 19%     -39.3%       1560 ± 31%  perf-sched.wait_time.max.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
      4.97 ±  2%     -24.3%       3.76 ±  3%  perf-sched.wait_time.max.ms.syslog_print.do_syslog.kmsg_read.vfs_read
      0.01 ±100%     -45.0%       0.01 ±100%  perf-sched.wait_time.max.ms.wait_for_partner.fifo_open.do_dentry_open.do_open
      2720 ±  5%     -17.6%       2241 ±  5%  perf-sched.wait_time.max.ms.worker_thread.kthread.ret_from_fork.ret_from_fork_asm
     51.54           -51.5        0.00        perf-profile.calltrace.cycles-pp.__folio_put.__splice_from_pipe.splice_from_pipe.do_splice.__do_splice
     51.45           -51.4        0.00        perf-profile.calltrace.cycles-pp.__mem_cgroup_uncharge.__folio_put.__splice_from_pipe.splice_from_pipe.do_splice
     51.12           -51.1        0.00        perf-profile.calltrace.cycles-pp.uncharge_batch.__mem_cgroup_uncharge.__folio_put.__splice_from_pipe.splice_from_pipe
     53.42           -50.0        3.44        perf-profile.calltrace.cycles-pp.__splice_from_pipe.splice_from_pipe.do_splice.__do_splice.__x64_sys_splice
     54.16           -49.1        5.02        perf-profile.calltrace.cycles-pp.splice_from_pipe.do_splice.__do_splice.__x64_sys_splice.do_syscall_64
     48.88           -48.9        0.00        perf-profile.calltrace.cycles-pp.page_counter_uncharge.uncharge_batch.__mem_cgroup_uncharge.__folio_put.__splice_from_pipe
     68.02           -28.8       39.20        perf-profile.calltrace.cycles-pp.do_splice.__do_splice.__x64_sys_splice.do_syscall_64.entry_SYSCALL_64_after_hwframe
     69.56           -25.6       43.96        perf-profile.calltrace.cycles-pp.__do_splice.__x64_sys_splice.do_syscall_64.entry_SYSCALL_64_after_hwframe.splice
     73.24           -18.1       55.15        perf-profile.calltrace.cycles-pp.__x64_sys_splice.do_syscall_64.entry_SYSCALL_64_after_hwframe.splice
     75.82           -13.2       62.66        perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.splice
     10.86           -10.9        0.00        perf-profile.calltrace.cycles-pp.write
     77.52           -10.1       67.40        perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.splice
     10.10           -10.1        0.00        perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.write
      9.96           -10.0        0.00        perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.write
      9.72            -9.7        0.00        perf-profile.calltrace.cycles-pp.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe.write
      9.44            -9.4        0.00        perf-profile.calltrace.cycles-pp.vfs_write.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe.write
      8.60            -8.6        0.00        perf-profile.calltrace.cycles-pp.pipe_write.vfs_write.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe
      5.62 ±  3%      -5.6        0.00        perf-profile.calltrace.cycles-pp.propagate_protected_usage.page_counter_uncharge.uncharge_batch.__mem_cgroup_uncharge.__folio_put
      4.94            -4.9        0.00        perf-profile.calltrace.cycles-pp.__alloc_pages.pipe_write.vfs_write.ksys_write.do_syscall_64
      3.58            -3.6        0.00        perf-profile.calltrace.cycles-pp.__memcg_kmem_charge_page.__alloc_pages.pipe_write.vfs_write.ksys_write
      1.74            -1.7        0.00        perf-profile.calltrace.cycles-pp.try_charge_memcg.__memcg_kmem_charge_page.__alloc_pages.pipe_write.vfs_write
      1.19 ±  5%      -1.2        0.00        perf-profile.calltrace.cycles-pp.memcg_account_kmem.uncharge_batch.__mem_cgroup_uncharge.__folio_put.__splice_from_pipe
      1.14 ±  3%      -1.1        0.00        perf-profile.calltrace.cycles-pp.memcg_account_kmem.__memcg_kmem_charge_page.__alloc_pages.pipe_write.vfs_write
      0.97 ±  3%      -1.0        0.00        perf-profile.calltrace.cycles-pp.get_page_from_freelist.__alloc_pages.pipe_write.vfs_write.ksys_write
      0.94            -0.9        0.00        perf-profile.calltrace.cycles-pp.page_counter_try_charge.try_charge_memcg.__memcg_kmem_charge_page.__alloc_pages.pipe_write
      0.86            -0.9        0.00        perf-profile.calltrace.cycles-pp.copy_page_from_iter.pipe_write.vfs_write.ksys_write.do_syscall_64
      0.75 ±  2%      -0.8        0.00        perf-profile.calltrace.cycles-pp.rmqueue.get_page_from_freelist.__alloc_pages.pipe_write.vfs_write
      0.74            -0.7        0.00        perf-profile.calltrace.cycles-pp._copy_from_iter.copy_page_from_iter.pipe_write.vfs_write.ksys_write
      3.48 ±  7%      -0.6        2.84        perf-profile.calltrace.cycles-pp.mutex_unlock.splice_pipe_to_pipe.do_splice.__do_splice.__x64_sys_splice
      0.54 ±  4%      -0.5        0.00        perf-profile.calltrace.cycles-pp._raw_spin_lock_irq.pipe_write.vfs_write.ksys_write.do_syscall_64
      0.54            -0.5        0.00        perf-profile.calltrace.cycles-pp.copyin._copy_from_iter.copy_page_from_iter.pipe_write.vfs_write
      0.28 ±100%      -0.3        0.00        perf-profile.calltrace.cycles-pp._raw_spin_trylock.free_unref_page.__splice_from_pipe.splice_from_pipe.do_splice
      2.18 ±  2%      +0.1        2.33        perf-profile.calltrace.cycles-pp.mutex_lock.splice_pipe_to_pipe.do_splice.__do_splice.__x64_sys_splice
      0.00            +0.3        0.26 ±100%  perf-profile.calltrace.cycles-pp.free_unref_page_prepare.free_unref_page.__splice_from_pipe.splice_from_pipe.do_splice
      0.00            +0.5        0.52        perf-profile.calltrace.cycles-pp.kill_fasync.splice_pipe_to_pipe.do_splice.__do_splice.__x64_sys_splice
      0.00            +0.6        0.56        perf-profile.calltrace.cycles-pp.security_file_permission.vfs_splice_read.splice_file_to_pipe.do_splice.__do_splice
      0.00            +0.6        0.56        perf-profile.calltrace.cycles-pp.__kmem_cache_free.copy_splice_read.splice_file_to_pipe.do_splice.__do_splice
      0.00            +0.6        0.58        perf-profile.calltrace.cycles-pp.__cond_resched.mutex_lock.splice_pipe_to_pipe.do_splice.__do_splice
      0.00            +0.6        0.62        perf-profile.calltrace.cycles-pp.__cond_resched.mutex_lock.pipe_double_lock.splice_pipe_to_pipe.do_splice
      0.00            +0.6        0.63        perf-profile.calltrace.cycles-pp.security_file_permission.do_splice.__do_splice.__x64_sys_splice.do_syscall_64
      0.95 ±  3%      +0.6        1.58        perf-profile.calltrace.cycles-pp.free_unref_page.__splice_from_pipe.splice_from_pipe.do_splice.__do_splice
      0.00            +0.7        0.74        perf-profile.calltrace.cycles-pp.get_pipe_info.do_splice.__do_splice.__x64_sys_splice.do_syscall_64
      0.00            +0.8        0.75        perf-profile.calltrace.cycles-pp.__fsnotify_parent.vfs_splice_read.splice_file_to_pipe.do_splice.__do_splice
      1.60            +0.8        2.40        perf-profile.calltrace.cycles-pp.mutex_lock.pipe_double_lock.splice_pipe_to_pipe.do_splice.__do_splice
      0.00            +0.8        0.82        perf-profile.calltrace.cycles-pp.stress_splice_looped_pipe
      0.62 ±  5%      +0.9        1.54        perf-profile.calltrace.cycles-pp.stress_splice
      0.00            +1.0        1.02        perf-profile.calltrace.cycles-pp.syscall_exit_to_user_mode_prepare.syscall_exit_to_user_mode.do_syscall_64.entry_SYSCALL_64_after_hwframe.splice
      0.00            +1.0        1.04        perf-profile.calltrace.cycles-pp.entry_SYSRETQ_unsafe_stack.splice
      0.00            +1.1        1.12        perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_safe_stack.splice
      0.26 ±100%      +1.2        1.44        perf-profile.calltrace.cycles-pp.stress_splice_flag
      0.00            +1.3        1.26 ±  2%  perf-profile.calltrace.cycles-pp.__fdget.__x64_sys_splice.do_syscall_64.entry_SYSCALL_64_after_hwframe.splice
      0.00            +1.3        1.26        perf-profile.calltrace.cycles-pp.rmqueue.get_page_from_freelist.__alloc_pages.__alloc_pages_bulk.copy_splice_read
      1.88            +1.3        3.17        perf-profile.calltrace.cycles-pp.pipe_double_lock.splice_pipe_to_pipe.do_splice.__do_splice.__x64_sys_splice
      0.66            +1.3        1.96        perf-profile.calltrace.cycles-pp.syscall_enter_from_user_mode.do_syscall_64.entry_SYSCALL_64_after_hwframe.splice
      0.00            +1.3        1.33        perf-profile.calltrace.cycles-pp.exit_to_user_mode_prepare.syscall_exit_to_user_mode.do_syscall_64.entry_SYSCALL_64_after_hwframe.splice
      0.63 ±  3%      +1.3        1.96        perf-profile.calltrace.cycles-pp.syscall_return_via_sysret.splice
      0.00            +1.4        1.38        perf-profile.calltrace.cycles-pp.iov_iter_zero.read_iter_zero.copy_splice_read.splice_file_to_pipe.do_splice
      0.00            +1.4        1.40        perf-profile.calltrace.cycles-pp.__kmem_cache_alloc_node.__kmalloc.copy_splice_read.splice_file_to_pipe.do_splice
      0.00            +1.6        1.62        perf-profile.calltrace.cycles-pp.read_iter_zero.copy_splice_read.splice_file_to_pipe.do_splice.__do_splice
      0.00            +1.7        1.68        perf-profile.calltrace.cycles-pp.__fsnotify_parent.do_splice.__do_splice.__x64_sys_splice.do_syscall_64
      0.00            +1.7        1.71        perf-profile.calltrace.cycles-pp.__kmalloc.copy_splice_read.splice_file_to_pipe.do_splice.__do_splice
      0.00            +1.8        1.84        perf-profile.calltrace.cycles-pp.get_pipe_info.__do_splice.__x64_sys_splice.do_syscall_64.entry_SYSCALL_64_after_hwframe
      0.71 ±  4%      +1.9        2.60        perf-profile.calltrace.cycles-pp.stress_mwc1
      0.00            +1.9        1.91        perf-profile.calltrace.cycles-pp.vfs_splice_read.splice_file_to_pipe.do_splice.__do_splice.__x64_sys_splice
      0.00            +1.9        1.93        perf-profile.calltrace.cycles-pp.get_page_from_freelist.__alloc_pages.__alloc_pages_bulk.copy_splice_read.splice_file_to_pipe
      1.10 ±  2%      +2.3        3.38        perf-profile.calltrace.cycles-pp.syscall_exit_to_user_mode.do_syscall_64.entry_SYSCALL_64_after_hwframe.splice
     11.77 ±  3%      +2.4       14.20        perf-profile.calltrace.cycles-pp.splice_pipe_to_pipe.do_splice.__do_splice.__x64_sys_splice.do_syscall_64
      0.00            +2.8        2.85        perf-profile.calltrace.cycles-pp.__alloc_pages.__alloc_pages_bulk.copy_splice_read.splice_file_to_pipe.do_splice
      0.00            +3.3        3.31        perf-profile.calltrace.cycles-pp.__alloc_pages_bulk.copy_splice_read.splice_file_to_pipe.do_splice.__do_splice
      2.16 ±  4%      +4.1        6.25        perf-profile.calltrace.cycles-pp.__fget_light.__x64_sys_splice.do_syscall_64.entry_SYSCALL_64_after_hwframe.splice
     86.87            +7.0       93.85        perf-profile.calltrace.cycles-pp.splice
      0.00            +9.1        9.13        perf-profile.calltrace.cycles-pp.copy_splice_read.splice_file_to_pipe.do_splice.__do_splice.__x64_sys_splice
      0.00           +12.8       12.75        perf-profile.calltrace.cycles-pp.splice_file_to_pipe.do_splice.__do_splice.__x64_sys_splice.do_syscall_64
      6.26 ±  3%     +12.8       19.04        perf-profile.calltrace.cycles-pp.__entry_text_start.splice
     51.47           -51.3        0.14 ±  3%  perf-profile.children.cycles-pp.__mem_cgroup_uncharge
     51.56           -51.2        0.38        perf-profile.children.cycles-pp.__folio_put
     51.18           -51.2        0.00        perf-profile.children.cycles-pp.uncharge_batch
     53.46           -49.9        3.58        perf-profile.children.cycles-pp.__splice_from_pipe
     54.20           -49.1        5.12        perf-profile.children.cycles-pp.splice_from_pipe
     48.92           -48.9        0.00        perf-profile.children.cycles-pp.page_counter_uncharge
     68.27           -28.3       40.02        perf-profile.children.cycles-pp.do_splice
     69.98           -24.7       45.28        perf-profile.children.cycles-pp.__do_splice
     86.38           -22.2       64.19        perf-profile.children.cycles-pp.do_syscall_64
     87.75           -20.1       67.69        perf-profile.children.cycles-pp.entry_SYSCALL_64_after_hwframe
     73.60           -17.6       55.98        perf-profile.children.cycles-pp.__x64_sys_splice
     11.11           -11.1        0.00        perf-profile.children.cycles-pp.write
      9.76            -9.8        0.00        perf-profile.children.cycles-pp.ksys_write
      9.49            -9.5        0.00        perf-profile.children.cycles-pp.vfs_write
      8.68            -8.7        0.00        perf-profile.children.cycles-pp.pipe_write
      5.73 ±  2%      -5.7        0.00        perf-profile.children.cycles-pp.propagate_protected_usage
      3.62            -3.6        0.00        perf-profile.children.cycles-pp.__memcg_kmem_charge_page
      2.36 ±  4%      -2.4        0.00        perf-profile.children.cycles-pp.memcg_account_kmem
      4.98            -2.0        2.94        perf-profile.children.cycles-pp.__alloc_pages
      1.74            -1.7        0.00        perf-profile.children.cycles-pp.try_charge_memcg
      0.94            -0.9        0.00        perf-profile.children.cycles-pp.page_counter_try_charge
      0.89            -0.9        0.00        perf-profile.children.cycles-pp.copy_page_from_iter
      0.76            -0.8        0.00        perf-profile.children.cycles-pp._copy_from_iter
      0.58            -0.6        0.00        perf-profile.children.cycles-pp.copyin
      0.56 ±  4%      -0.6        0.00        perf-profile.children.cycles-pp._raw_spin_lock_irq
      0.54            -0.5        0.00        perf-profile.children.cycles-pp.__wake_up_common_lock
      0.47            -0.5        0.00        perf-profile.children.cycles-pp.__get_obj_cgroup_from_memcg
      4.18 ±  6%      -0.5        3.71        perf-profile.children.cycles-pp.mutex_unlock
      0.37 ±  2%      -0.4        0.00        perf-profile.children.cycles-pp.anon_pipe_buf_release
      0.33            -0.3        0.00        perf-profile.children.cycles-pp._raw_spin_lock_irqsave
      0.31 ±  3%      -0.3        0.00        perf-profile.children.cycles-pp.__count_memcg_events
      0.26            -0.3        0.00        perf-profile.children.cycles-pp.alloc_pages
      0.26            -0.3        0.00        perf-profile.children.cycles-pp.file_update_time
      0.24 ±  2%      -0.2        0.00        perf-profile.children.cycles-pp.uncharge_folio
      0.97 ±  3%      -0.2        0.75        perf-profile.children.cycles-pp._raw_spin_trylock
      0.22 ±  2%      -0.2        0.00        perf-profile.children.cycles-pp.inode_needs_update_time
      0.18 ±  2%      -0.2        0.00        perf-profile.children.cycles-pp.__fdget_pos
      0.12            -0.1        0.00        perf-profile.children.cycles-pp.memcg_check_events
      0.09            -0.1        0.00        perf-profile.children.cycles-pp.cgroup_rstat_updated
      0.09            -0.1        0.00        perf-profile.children.cycles-pp.policy_node
      0.08            -0.1        0.00        perf-profile.children.cycles-pp.timestamp_truncate
      0.08 ±  6%      -0.1        0.00        perf-profile.children.cycles-pp.__wake_up_common
      0.29 ±  3%      -0.1        0.22 ±  2%  perf-profile.children.cycles-pp.start_secondary
      0.29 ±  3%      -0.1        0.23 ±  4%  perf-profile.children.cycles-pp.secondary_startup_64_no_verify
      0.29 ±  3%      -0.1        0.23 ±  4%  perf-profile.children.cycles-pp.cpu_startup_entry
      0.29 ±  3%      -0.1        0.23 ±  4%  perf-profile.children.cycles-pp.do_idle
      0.28            -0.1        0.22 ±  2%  perf-profile.children.cycles-pp.cpuidle_idle_call
      0.06            -0.1        0.00        perf-profile.children.cycles-pp._raw_spin_unlock_irqrestore
      0.26            -0.1        0.21 ±  4%  perf-profile.children.cycles-pp.cpuidle_enter
      0.26            -0.1        0.21 ±  4%  perf-profile.children.cycles-pp.cpuidle_enter_state
      0.26            -0.0        0.20 ±  2%  perf-profile.children.cycles-pp.acpi_idle_enter
      0.26            -0.0        0.20 ±  2%  perf-profile.children.cycles-pp.acpi_safe_halt
      0.14 ±  3%      -0.0        0.14        perf-profile.children.cycles-pp.perf_adjust_freq_unthr_context
      0.10 ±  4%      -0.0        0.10        perf-profile.children.cycles-pp.__intel_pmu_enable_all
      0.35 ±  2%      +0.0        0.35 ±  5%  perf-profile.children.cycles-pp.__hrtimer_run_queues
      0.25            +0.0        0.25        perf-profile.children.cycles-pp.scheduler_tick
      0.14 ±  3%      +0.0        0.15        perf-profile.children.cycles-pp.perf_event_task_tick
      0.28            +0.0        0.28        perf-profile.children.cycles-pp.tick_sched_handle
      0.28            +0.0        0.28        perf-profile.children.cycles-pp.update_process_times
      0.30 ±  3%      +0.0        0.31 ±  6%  perf-profile.children.cycles-pp.tick_sched_timer
      0.02 ±100%      +0.0        0.05        perf-profile.children.cycles-pp.task_tick_fair
      0.00            +0.1        0.05        perf-profile.children.cycles-pp.should_fail_alloc_page
      0.00            +0.1        0.06        perf-profile.children.cycles-pp.splice_write_null
      0.00            +0.1        0.08 ±  5%  perf-profile.children.cycles-pp.__list_add_valid_or_report
      0.07            +0.1        0.16 ±  3%  perf-profile.children.cycles-pp.splice_from_pipe_next
      0.06 ±  9%      +0.1        0.15        perf-profile.children.cycles-pp.__page_cache_release
      0.00            +0.1        0.10 ±  4%  perf-profile.children.cycles-pp.iov_iter_bvec
      0.06            +0.1        0.18        perf-profile.children.cycles-pp.__list_del_entry_valid_or_report
      0.00            +0.1        0.12        perf-profile.children.cycles-pp.kmalloc_slab
      0.07 ± 14%      +0.1        0.20 ±  4%  perf-profile.children.cycles-pp.__get_task_ioprio
      0.00            +0.1        0.14 ±  3%  perf-profile.children.cycles-pp.wait_for_space
      0.00            +0.1        0.14        perf-profile.children.cycles-pp.memcg_slab_post_alloc_hook
      0.00            +0.1        0.14        perf-profile.children.cycles-pp.pipe_lock
      0.08            +0.1        0.22 ±  2%  perf-profile.children.cycles-pp.rw_verify_area
      0.08 ±  6%      +0.2        0.23 ±  4%  perf-profile.children.cycles-pp.get_pfnblock_flags_mask
      0.00            +0.2        0.20 ±  2%  perf-profile.children.cycles-pp.fsnotify_perm
      0.12 ±  4%      +0.2        0.32        perf-profile.children.cycles-pp.aa_file_perm
      0.00            +0.2        0.24 ±  2%  perf-profile.children.cycles-pp.kfree
      0.00            +0.3        0.27        perf-profile.children.cycles-pp.memset_orig
      0.16 ±  6%      +0.3        0.46        perf-profile.children.cycles-pp.free_unref_page_commit
      0.00            +0.3        0.32        perf-profile.children.cycles-pp.generic_pipe_buf_release
      0.18 ±  5%      +0.4        0.54 ±  2%  perf-profile.children.cycles-pp.free_unref_page_prepare
      0.20 ±  2%      +0.4        0.58        perf-profile.children.cycles-pp.splice@plt
      0.22 ±  2%      +0.5        0.68 ±  2%  perf-profile.children.cycles-pp.pipe_unlock
      0.24 ±  4%      +0.5        0.74        perf-profile.children.cycles-pp.pipe_clear_nowait
      0.43            +0.5        0.97        perf-profile.children.cycles-pp.apparmor_file_permission
      0.78 ±  3%      +0.6        1.34        perf-profile.children.cycles-pp.rmqueue
      0.30 ±  5%      +0.6        0.87        perf-profile.children.cycles-pp.kill_fasync
      0.00            +0.6        0.58        perf-profile.children.cycles-pp.__kmem_cache_free
      0.30            +0.6        0.89        perf-profile.children.cycles-pp.rcu_all_qs
      0.34            +0.7        1.03        perf-profile.children.cycles-pp.__fdget
      0.98 ±  3%      +0.7        1.68        perf-profile.children.cycles-pp.free_unref_page
      0.45 ±  6%      +0.7        1.14        perf-profile.children.cycles-pp.entry_SYSCALL_64_safe_stack
      0.46 ±  3%      +0.7        1.19        perf-profile.children.cycles-pp.stress_splice_looped_pipe
      0.54            +0.8        1.30        perf-profile.children.cycles-pp.security_file_permission
      0.46 ±  2%      +0.8        1.28        perf-profile.children.cycles-pp.syscall_exit_to_user_mode_prepare
      1.00 ±  3%      +1.0        1.98        perf-profile.children.cycles-pp.get_page_from_freelist
      0.58 ±  3%      +1.0        1.60        perf-profile.children.cycles-pp.exit_to_user_mode_prepare
      0.86 ±  5%      +1.4        2.21        perf-profile.children.cycles-pp.stress_splice
      0.00            +1.4        1.40        perf-profile.children.cycles-pp.iov_iter_zero
      0.81            +1.4        2.22        perf-profile.children.cycles-pp.syscall_enter_from_user_mode
      0.84 ±  2%      +1.4        2.25        perf-profile.children.cycles-pp.syscall_return_via_sysret
      1.95            +1.4        3.37        perf-profile.children.cycles-pp.pipe_double_lock
      0.73 ±  4%      +1.5        2.20        perf-profile.children.cycles-pp.stress_splice_flag
      0.00            +1.5        1.50        perf-profile.children.cycles-pp.__kmem_cache_alloc_node
      0.85            +1.5        2.38        perf-profile.children.cycles-pp.__cond_resched
      4.48            +1.6        6.08        perf-profile.children.cycles-pp.mutex_lock
      0.86 ±  4%      +1.7        2.52        perf-profile.children.cycles-pp.stress_mwc1
      0.00            +1.7        1.66        perf-profile.children.cycles-pp.read_iter_zero
      0.00            +1.8        1.79        perf-profile.children.cycles-pp.__kmalloc
      0.85 ±  3%      +1.8        2.66        perf-profile.children.cycles-pp.get_pipe_info
      0.00            +2.0        1.95        perf-profile.children.cycles-pp.vfs_splice_read
      0.32            +2.2        2.48        perf-profile.children.cycles-pp.__fsnotify_parent
      1.57 ±  2%      +2.8        4.37        perf-profile.children.cycles-pp.syscall_exit_to_user_mode
     12.04 ±  2%      +2.9       14.98        perf-profile.children.cycles-pp.splice_pipe_to_pipe
      0.00            +3.4        3.36        perf-profile.children.cycles-pp.__alloc_pages_bulk
      2.44 ±  4%      +4.2        6.68        perf-profile.children.cycles-pp.__fget_light
      3.46 ±  3%      +5.9        9.32        perf-profile.children.cycles-pp.entry_SYSRETQ_unsafe_stack
     86.88            +6.9       93.80        perf-profile.children.cycles-pp.splice
      4.12 ±  4%      +7.0       11.08        perf-profile.children.cycles-pp.__entry_text_start
      0.00            +9.3        9.29        perf-profile.children.cycles-pp.copy_splice_read
      0.00           +12.9       12.88        perf-profile.children.cycles-pp.splice_file_to_pipe
     43.14           -43.1        0.00        perf-profile.self.cycles-pp.page_counter_uncharge
      5.69 ±  2%      -5.7        0.00        perf-profile.self.cycles-pp.propagate_protected_usage
      2.27 ±  4%      -2.3        0.00        perf-profile.self.cycles-pp.memcg_account_kmem
      0.86            -0.9        0.00        perf-profile.self.cycles-pp.page_counter_try_charge
      0.78 ±  2%      -0.8        0.00        perf-profile.self.cycles-pp.try_charge_memcg
      0.64 ±  5%      -0.6        0.00        perf-profile.self.cycles-pp.uncharge_batch
      4.08 ±  6%      -0.6        3.49        perf-profile.self.cycles-pp.mutex_unlock
      0.57            -0.6        0.00        perf-profile.self.cycles-pp.copyin
      0.54 ±  4%      -0.5        0.00        perf-profile.self.cycles-pp._raw_spin_lock_irq
      0.48            -0.5        0.00        perf-profile.self.cycles-pp.pipe_write
      0.46            -0.5        0.00        perf-profile.self.cycles-pp.__get_obj_cgroup_from_memcg
      0.46            -0.5        0.00        perf-profile.self.cycles-pp.vfs_write
      0.37 ±  2%      -0.4        0.00        perf-profile.self.cycles-pp.anon_pipe_buf_release
      0.32            -0.3        0.00        perf-profile.self.cycles-pp._raw_spin_lock_irqsave
      0.30 ±  3%      -0.3        0.00        perf-profile.self.cycles-pp.__count_memcg_events
      0.27 ±  7%      -0.3        0.00        perf-profile.self.cycles-pp.__memcg_kmem_charge_page
      0.26            -0.3        0.00        perf-profile.self.cycles-pp.write
      0.94 ±  3%      -0.3        0.68        perf-profile.self.cycles-pp._raw_spin_trylock
      0.22 ±  2%      -0.2        0.00        perf-profile.self.cycles-pp.uncharge_folio
      0.18 ±  2%      -0.2        0.00        perf-profile.self.cycles-pp._copy_from_iter
      0.14 ±  3%      -0.1        0.00        perf-profile.self.cycles-pp.alloc_pages
      0.13            -0.1        0.00        perf-profile.self.cycles-pp.copy_page_from_iter
      0.11            -0.1        0.00        perf-profile.self.cycles-pp.__wake_up_common_lock
      0.10            -0.1        0.00        perf-profile.self.cycles-pp.inode_needs_update_time
      0.10 ±  5%      -0.1        0.00        perf-profile.self.cycles-pp.ksys_write
      0.09            -0.1        0.00        perf-profile.self.cycles-pp.memcg_check_events
      0.07            -0.1        0.00        perf-profile.self.cycles-pp.cgroup_rstat_updated
      0.07            -0.1        0.00        perf-profile.self.cycles-pp.timestamp_truncate
      0.06 ±  9%      -0.1        0.00        perf-profile.self.cycles-pp.__wake_up_common
      0.12 ±  4%      -0.0        0.09        perf-profile.self.cycles-pp.acpi_safe_halt
      0.05            -0.0        0.02 ±100%  perf-profile.self.cycles-pp.perf_adjust_freq_unthr_context
      0.02 ±100%      -0.0        0.00        perf-profile.self.cycles-pp.file_update_time
      0.02 ±100%      -0.0        0.00        perf-profile.self.cycles-pp.policy_node
      0.10 ±  4%      -0.0        0.10        perf-profile.self.cycles-pp.__intel_pmu_enable_all
      0.07            +0.1        0.12 ±  4%  perf-profile.self.cycles-pp.__mem_cgroup_uncharge
      0.00            +0.1        0.07        perf-profile.self.cycles-pp.__list_add_valid_or_report
      0.00            +0.1        0.08        perf-profile.self.cycles-pp.pipe_lock
      0.06 ±  7%      +0.1        0.15        perf-profile.self.cycles-pp.splice_from_pipe_next
      0.00            +0.1        0.08 ±  5%  perf-profile.self.cycles-pp.iov_iter_bvec
      0.06 ±  9%      +0.1        0.14        perf-profile.self.cycles-pp.__page_cache_release
      0.00            +0.1        0.09        perf-profile.self.cycles-pp.__folio_put
      0.00            +0.1        0.10        perf-profile.self.cycles-pp.kmalloc_slab
      0.06            +0.1        0.16 ±  3%  perf-profile.self.cycles-pp.rw_verify_area
      0.06 ±  7%      +0.1        0.18 ±  2%  perf-profile.self.cycles-pp.__get_task_ioprio
      0.05            +0.1        0.16        perf-profile.self.cycles-pp.__list_del_entry_valid_or_report
      0.00            +0.1        0.12 ±  4%  perf-profile.self.cycles-pp.wait_for_space
      0.00            +0.1        0.12        perf-profile.self.cycles-pp.memcg_slab_post_alloc_hook
      0.06 ±  7%      +0.1        0.20 ±  7%  perf-profile.self.cycles-pp.get_pfnblock_flags_mask
      0.10            +0.2        0.27        perf-profile.self.cycles-pp.aa_file_perm
      0.08 ±  5%      +0.2        0.26        perf-profile.self.cycles-pp.splice@plt
      0.00            +0.2        0.18 ±  2%  perf-profile.self.cycles-pp.fsnotify_perm
      0.10 ±  4%      +0.2        0.30        perf-profile.self.cycles-pp.free_unref_page_prepare
      0.00            +0.2        0.20 ±  2%  perf-profile.self.cycles-pp.__kmalloc
      0.00            +0.2        0.21 ±  4%  perf-profile.self.cycles-pp.kfree
      0.13            +0.2        0.36 ±  2%  perf-profile.self.cycles-pp.security_file_permission
      0.14 ±  3%      +0.2        0.37        perf-profile.self.cycles-pp.free_unref_page
      0.12 ±  4%      +0.2        0.37 ±  2%  perf-profile.self.cycles-pp.free_unref_page_commit
      0.00            +0.2        0.25        perf-profile.self.cycles-pp.memset_orig
      0.00            +0.3        0.26        perf-profile.self.cycles-pp.read_iter_zero
      0.14 ±  3%      +0.3        0.40        perf-profile.self.cycles-pp.__fdget
      0.13            +0.3        0.40        perf-profile.self.cycles-pp.pipe_unlock
      0.16 ±  3%      +0.3        0.45 ±  2%  perf-profile.self.cycles-pp.kill_fasync
      0.00            +0.3        0.30        perf-profile.self.cycles-pp.generic_pipe_buf_release
      0.16 ±  3%      +0.3        0.48        perf-profile.self.cycles-pp.pipe_clear_nowait
      0.30 ±  3%      +0.3        0.62        perf-profile.self.cycles-pp.apparmor_file_permission
      0.00            +0.3        0.32        perf-profile.self.cycles-pp.vfs_splice_read
      0.21            +0.4        0.60        perf-profile.self.cycles-pp.rcu_all_qs
      3.76            +0.4        4.15        perf-profile.self.cycles-pp.mutex_lock
      0.00            +0.4        0.42        perf-profile.self.cycles-pp.__alloc_pages_bulk
      0.22 ±  6%      +0.4        0.65        perf-profile.self.cycles-pp.get_page_from_freelist
      0.26            +0.4        0.68        perf-profile.self.cycles-pp.splice_from_pipe
      0.44            +0.5        0.91        perf-profile.self.cycles-pp.__splice_from_pipe
      0.29 ±  3%      +0.5        0.78        perf-profile.self.cycles-pp.__alloc_pages
      0.28            +0.5        0.78        perf-profile.self.cycles-pp.pipe_double_lock
      0.29            +0.5        0.78        perf-profile.self.cycles-pp.syscall_exit_to_user_mode_prepare
      0.28            +0.5        0.79        perf-profile.self.cycles-pp.rmqueue
      0.00            +0.6        0.55        perf-profile.self.cycles-pp.__kmem_cache_free
      0.00            +0.6        0.58        perf-profile.self.cycles-pp.splice_file_to_pipe
      0.35 ±  8%      +0.6        0.93        perf-profile.self.cycles-pp.stress_splice_looped_pipe
      0.44 ±  5%      +0.7        1.13        perf-profile.self.cycles-pp.entry_SYSCALL_64_safe_stack
      0.38 ±  2%      +0.8        1.16        perf-profile.self.cycles-pp.stress_splice_flag
      0.48 ±  4%      +0.8        1.33        perf-profile.self.cycles-pp.exit_to_user_mode_prepare
      0.00            +0.9        0.92        perf-profile.self.cycles-pp.__kmem_cache_alloc_node
      0.54            +0.9        1.48        perf-profile.self.cycles-pp.__cond_resched
      0.50 ±  2%      +1.1        1.58        perf-profile.self.cycles-pp.get_pipe_info
      0.72 ±  9%      +1.1        1.83        perf-profile.self.cycles-pp.stress_splice
      0.62 ±  3%      +1.1        1.76        perf-profile.self.cycles-pp.syscall_exit_to_user_mode
      0.70            +1.2        1.90        perf-profile.self.cycles-pp.syscall_enter_from_user_mode
      0.68 ±  4%      +1.3        1.97        perf-profile.self.cycles-pp.stress_mwc1
      0.00            +1.4        1.38        perf-profile.self.cycles-pp.iov_iter_zero
      0.84 ±  2%      +1.4        2.25        perf-profile.self.cycles-pp.syscall_return_via_sysret
      4.14            +1.5        5.62        perf-profile.self.cycles-pp.splice_pipe_to_pipe
      0.00            +1.5        1.50        perf-profile.self.cycles-pp.copy_splice_read
      0.94 ±  2%      +1.6        2.56        perf-profile.self.cycles-pp.do_syscall_64
      1.15 ±  5%      +2.0        3.12        perf-profile.self.cycles-pp.__entry_text_start
      0.32 ±  3%      +2.1        2.42        perf-profile.self.cycles-pp.__fsnotify_parent
      1.46 ±  3%      +2.3        3.73        perf-profile.self.cycles-pp.entry_SYSCALL_64_after_hwframe
      1.23 ±  2%      +2.5        3.72        perf-profile.self.cycles-pp.__x64_sys_splice
      1.22 ±  3%      +2.5        3.76        perf-profile.self.cycles-pp.__do_splice
      1.20 ±  2%      +2.8        3.95        perf-profile.self.cycles-pp.do_splice
      2.29 ±  3%      +4.0        6.26        perf-profile.self.cycles-pp.__fget_light
      3.35 ±  3%      +5.7        9.03        perf-profile.self.cycles-pp.entry_SYSRETQ_unsafe_stack
      3.66 ±  2%      +6.7       10.31        perf-profile.self.cycles-pp.splice
      2491 ±  2%      -8.4%       2283 ±  3%  slabinfo.Acpi-State.active_objs
     48.86 ±  2%      -8.4%      44.77 ±  3%  slabinfo.Acpi-State.active_slabs
      2491 ±  2%      -8.4%       2283 ±  3%  slabinfo.Acpi-State.num_objs
     48.86 ±  2%      -8.4%      44.77 ±  3%  slabinfo.Acpi-State.num_slabs
     36.00            +0.0%      36.00        slabinfo.DCCP.active_objs
      2.00            +0.0%       2.00        slabinfo.DCCP.active_slabs
     36.00            +0.0%      36.00        slabinfo.DCCP.num_objs
      2.00            +0.0%       2.00        slabinfo.DCCP.num_slabs
     34.00            +0.0%      34.00        slabinfo.DCCPv6.active_objs
      2.00            +0.0%       2.00        slabinfo.DCCPv6.active_slabs
     34.00            +0.0%      34.00        slabinfo.DCCPv6.num_objs
      2.00            +0.0%       2.00        slabinfo.DCCPv6.num_slabs
    272.00 ±  5%      +0.0%     272.00 ±  5%  slabinfo.RAW.active_objs
      8.50 ±  5%      +0.0%       8.50 ±  5%  slabinfo.RAW.active_slabs
    272.00 ±  5%      +0.0%     272.00 ±  5%  slabinfo.RAW.num_objs
      8.50 ±  5%      +0.0%       8.50 ±  5%  slabinfo.RAW.num_slabs
    208.00            +0.0%     208.00        slabinfo.RAWv6.active_objs
      8.00            +0.0%       8.00        slabinfo.RAWv6.active_slabs
    208.00            +0.0%     208.00        slabinfo.RAWv6.num_objs
      8.00            +0.0%       8.00        slabinfo.RAWv6.num_slabs
     69.46            -9.5%      62.90 ± 10%  slabinfo.TCP.active_objs
      4.96            -9.5%       4.49 ± 10%  slabinfo.TCP.active_slabs
     69.46            -9.5%      62.90 ± 10%  slabinfo.TCP.num_objs
      4.96            -9.5%       4.49 ± 10%  slabinfo.TCP.num_slabs
     39.00            +0.0%      39.00        slabinfo.TCPv6.active_objs
      3.00            +0.0%       3.00        slabinfo.TCPv6.active_slabs
     39.00            +0.0%      39.00        slabinfo.TCPv6.num_objs
      3.00            +0.0%       3.00        slabinfo.TCPv6.num_slabs
    120.00            +0.0%     120.00        slabinfo.UDPv6.active_objs
      5.00            +0.0%       5.00        slabinfo.UDPv6.active_slabs
    120.00            +0.0%     120.00        slabinfo.UDPv6.num_objs
      5.00            +0.0%       5.00        slabinfo.UDPv6.num_slabs
      1845 ±  7%      +1.0%       1865 ±  6%  slabinfo.UNIX.active_objs
     61.53 ±  7%      +1.0%      62.18 ±  6%  slabinfo.UNIX.active_slabs
      1845 ±  7%      +1.0%       1865 ±  6%  slabinfo.UNIX.num_objs
     61.53 ±  7%      +1.0%      62.18 ±  6%  slabinfo.UNIX.num_slabs
     21238            +5.8%      22466        slabinfo.anon_vma.active_objs
    548.62            +6.0%     581.43        slabinfo.anon_vma.active_slabs
     21396            +6.0%      22675        slabinfo.anon_vma.num_objs
    548.62            +6.0%     581.43        slabinfo.anon_vma.num_slabs
     30177            -2.6%      29406        slabinfo.anon_vma_chain.active_objs
    473.61            -2.2%     463.15        slabinfo.anon_vma_chain.active_slabs
     30310            -2.2%      29641        slabinfo.anon_vma_chain.num_objs
    473.61            -2.2%     463.15        slabinfo.anon_vma_chain.num_slabs
     90.00 ± 11%     +11.1%     100.00 ± 20%  slabinfo.bdev_cache.active_objs
      4.50 ± 11%     +11.1%       5.00 ± 20%  slabinfo.bdev_cache.active_slabs
     90.00 ± 11%     +11.1%     100.00 ± 20%  slabinfo.bdev_cache.num_objs
      4.50 ± 11%     +11.1%       5.00 ± 20%  slabinfo.bdev_cache.num_slabs
    800.00 ±  4%      -4.0%     768.00        slabinfo.bio-120.active_objs
     12.50 ±  4%      -4.0%      12.00        slabinfo.bio-120.active_slabs
    800.00 ±  4%      -4.0%     768.00        slabinfo.bio-120.num_objs
     12.50 ±  4%      -4.0%      12.00        slabinfo.bio-120.num_slabs
    693.00 ±  9%      +3.0%     714.00 ±  5%  slabinfo.bio-184.active_objs
     16.50 ±  9%      +3.0%      17.00 ±  5%  slabinfo.bio-184.active_slabs
    693.00 ±  9%      +3.0%     714.00 ±  5%  slabinfo.bio-184.num_objs
     16.50 ±  9%      +3.0%      17.00 ±  5%  slabinfo.bio-184.num_slabs
    128.00            +0.0%     128.00        slabinfo.bio-248.active_objs
      2.00            +0.0%       2.00        slabinfo.bio-248.active_slabs
    128.00            +0.0%     128.00        slabinfo.bio-248.num_objs
      2.00            +0.0%       2.00        slabinfo.bio-248.num_slabs
     51.00            +0.0%      51.00        slabinfo.bio-296.active_objs
      1.00            +0.0%       1.00        slabinfo.bio-296.active_slabs
     51.00            +0.0%      51.00        slabinfo.bio-296.num_objs
      1.00            +0.0%       1.00        slabinfo.bio-296.num_slabs
    168.00            +0.0%     168.00        slabinfo.bio-360.active_objs
      4.00            +0.0%       4.00        slabinfo.bio-360.active_slabs
    168.00            +0.0%     168.00        slabinfo.bio-360.num_objs
      4.00            +0.0%       4.00        slabinfo.bio-360.num_slabs
     42.00            +0.0%      42.00        slabinfo.bio-376.active_objs
      1.00            +0.0%       1.00        slabinfo.bio-376.active_slabs
     42.00            +0.0%      42.00        slabinfo.bio-376.num_objs
      1.00            +0.0%       1.00        slabinfo.bio-376.num_slabs
     36.00            +0.0%      36.00        slabinfo.bio-432.active_objs
      1.00            +0.0%       1.00        slabinfo.bio-432.active_slabs
     36.00            +0.0%      36.00        slabinfo.bio-432.num_objs
      1.00            +0.0%       1.00        slabinfo.bio-432.num_slabs
    170.00            +0.0%     170.00        slabinfo.bio_post_read_ctx.active_objs
      2.00            +0.0%       2.00        slabinfo.bio_post_read_ctx.active_slabs
    170.00            +0.0%     170.00        slabinfo.bio_post_read_ctx.num_objs
      2.00            +0.0%       2.00        slabinfo.bio_post_read_ctx.num_slabs
     32.00            +0.0%      32.00        slabinfo.biovec-128.active_objs
      2.00            +0.0%       2.00        slabinfo.biovec-128.active_slabs
     32.00            +0.0%      32.00        slabinfo.biovec-128.num_objs
      2.00            +0.0%       2.00        slabinfo.biovec-128.num_slabs
    352.00 ± 18%     +22.7%     432.00 ± 11%  slabinfo.biovec-64.active_objs
     11.00 ± 18%     +22.7%      13.50 ± 11%  slabinfo.biovec-64.active_slabs
    352.00 ± 18%     +22.7%     432.00 ± 11%  slabinfo.biovec-64.num_objs
     11.00 ± 18%     +22.7%      13.50 ± 11%  slabinfo.biovec-64.num_slabs
     56.00            +0.0%      56.00        slabinfo.biovec-max.active_objs
      7.00            +0.0%       7.00        slabinfo.biovec-max.active_slabs
     56.00            +0.0%      56.00        slabinfo.biovec-max.num_objs
      7.00            +0.0%       7.00        slabinfo.biovec-max.num_slabs
    204.00            +0.0%     204.00        slabinfo.btrfs_extent_buffer.active_objs
      3.00            +0.0%       3.00        slabinfo.btrfs_extent_buffer.active_slabs
    204.00            +0.0%     204.00        slabinfo.btrfs_extent_buffer.num_objs
      3.00            +0.0%       3.00        slabinfo.btrfs_extent_buffer.num_slabs
     39.00            +0.0%      39.00        slabinfo.btrfs_free_space.active_objs
      1.00            +0.0%       1.00        slabinfo.btrfs_free_space.active_slabs
     39.00            +0.0%      39.00        slabinfo.btrfs_free_space.num_objs
      1.00            +0.0%       1.00        slabinfo.btrfs_free_space.num_slabs
    101.50 ± 14%     +14.3%     116.00        slabinfo.btrfs_inode.active_objs
      3.50 ± 14%     +14.3%       4.00        slabinfo.btrfs_inode.active_slabs
    101.50 ± 14%     +14.3%     116.00        slabinfo.btrfs_inode.num_objs
      3.50 ± 14%     +14.3%       4.00        slabinfo.btrfs_inode.num_slabs
    269.45 ±  6%      +6.7%     287.45        slabinfo.btrfs_path.active_objs
      7.48 ±  6%      +6.7%       7.98        slabinfo.btrfs_path.active_slabs
    269.45 ±  6%      +6.7%     287.45        slabinfo.btrfs_path.num_objs
      7.48 ±  6%      +6.7%       7.98        slabinfo.btrfs_path.num_slabs
    253.50 ±  7%      +7.7%     273.00        slabinfo.buffer_head.active_objs
      6.50 ±  7%      +7.7%       7.00        slabinfo.buffer_head.active_slabs
    253.50 ±  7%      +7.7%     273.00        slabinfo.buffer_head.num_objs
      6.50 ±  7%      +7.7%       7.00        slabinfo.buffer_head.num_slabs
      8066 ±  7%      -5.4%       7628        slabinfo.cred_jar.active_objs
    192.07 ±  7%      -5.4%     181.62        slabinfo.cred_jar.active_slabs
      8066 ±  7%      -5.4%       7628        slabinfo.cred_jar.num_objs
    192.07 ±  7%      -5.4%     181.62        slabinfo.cred_jar.num_slabs
     39.00            +0.0%      39.00        slabinfo.dax_cache.active_objs
      1.00            +0.0%       1.00        slabinfo.dax_cache.active_slabs
     39.00            +0.0%      39.00        slabinfo.dax_cache.num_objs
      1.00            +0.0%       1.00        slabinfo.dax_cache.num_slabs
    117830            -1.7%     115870        slabinfo.dentry.active_objs
      2856            -1.6%       2809        slabinfo.dentry.active_slabs
    119957            -1.6%     117994        slabinfo.dentry.num_objs
      2856            -1.6%       2809        slabinfo.dentry.num_slabs
     30.00            +0.0%      30.00        slabinfo.dmaengine-unmap-128.active_objs
      1.00            +0.0%       1.00        slabinfo.dmaengine-unmap-128.active_slabs
     30.00            +0.0%      30.00        slabinfo.dmaengine-unmap-128.num_objs
      1.00            +0.0%       1.00        slabinfo.dmaengine-unmap-128.num_slabs
     64.00            +0.0%      64.00        slabinfo.dmaengine-unmap-2.active_objs
      1.00            +0.0%       1.00        slabinfo.dmaengine-unmap-2.active_slabs
     64.00            +0.0%      64.00        slabinfo.dmaengine-unmap-2.num_objs
      1.00            +0.0%       1.00        slabinfo.dmaengine-unmap-2.num_slabs
     15.00            +0.0%      15.00        slabinfo.dmaengine-unmap-256.active_objs
      1.00            +0.0%       1.00        slabinfo.dmaengine-unmap-256.active_slabs
     15.00            +0.0%      15.00        slabinfo.dmaengine-unmap-256.num_objs
      1.00            +0.0%       1.00        slabinfo.dmaengine-unmap-256.num_slabs
     12558 ±  2%     +26.9%      15930 ±  9%  slabinfo.ep_head.active_objs
     49.06 ±  2%     +26.9%      62.23 ±  9%  slabinfo.ep_head.active_slabs
     12558 ±  2%     +26.9%      15930 ±  9%  slabinfo.ep_head.num_objs
     49.06 ±  2%     +26.9%      62.23 ±  9%  slabinfo.ep_head.num_slabs
    864.17 ±  7%     -12.2%     758.50 ±  2%  slabinfo.file_lock_cache.active_objs
     23.36 ±  7%     -12.2%      20.50 ±  2%  slabinfo.file_lock_cache.active_slabs
    864.17 ±  7%     -12.2%     758.50 ±  2%  slabinfo.file_lock_cache.num_objs
     23.36 ±  7%     -12.2%      20.50 ±  2%  slabinfo.file_lock_cache.num_slabs
      5586 ±  2%      +0.4%       5607        slabinfo.files_cache.active_objs
    121.45 ±  2%      +0.4%     121.91        slabinfo.files_cache.active_slabs
      5586 ±  2%      +0.4%       5607        slabinfo.files_cache.num_objs
    121.45 ±  2%      +0.4%     121.91        slabinfo.files_cache.num_slabs
     22890            -1.9%      22452        slabinfo.filp.active_objs
    376.06 ±  2%      -0.0%     375.89        slabinfo.filp.active_slabs
     24067 ±  2%      -0.0%      24056        slabinfo.filp.num_objs
    376.06 ±  2%      -0.0%     375.89        slabinfo.filp.num_slabs
      2797 ±  9%     -21.0%       2210 ±  7%  slabinfo.fsnotify_mark_connector.active_objs
     21.86 ±  9%     -21.0%      17.27 ±  7%  slabinfo.fsnotify_mark_connector.active_slabs
      2797 ±  9%     -21.0%       2210 ±  7%  slabinfo.fsnotify_mark_connector.num_objs
     21.86 ±  9%     -21.0%      17.27 ±  7%  slabinfo.fsnotify_mark_connector.num_slabs
      8723            -1.7%       8577 ±  3%  slabinfo.ftrace_event_field.active_objs
    119.50            -1.7%     117.50 ±  3%  slabinfo.ftrace_event_field.active_slabs
      8723            -1.7%       8577 ±  3%  slabinfo.ftrace_event_field.num_objs
    119.50            -1.7%     117.50 ±  3%  slabinfo.ftrace_event_field.num_slabs
    168.00            +0.0%     168.00        slabinfo.fuse_request.active_objs
      3.00            +0.0%       3.00        slabinfo.fuse_request.active_slabs
    168.00            +0.0%     168.00        slabinfo.fuse_request.num_objs
      3.00            +0.0%       3.00        slabinfo.fuse_request.num_slabs
     98.00            +0.0%      98.00        slabinfo.hugetlbfs_inode_cache.active_objs
      2.00            +0.0%       2.00        slabinfo.hugetlbfs_inode_cache.active_slabs
     98.00            +0.0%      98.00        slabinfo.hugetlbfs_inode_cache.num_objs
      2.00            +0.0%       2.00        slabinfo.hugetlbfs_inode_cache.num_slabs
     84504            -0.1%      84410        slabinfo.inode_cache.active_objs
      1661            -0.2%       1658        slabinfo.inode_cache.active_slabs
     84715            -0.2%      84560        slabinfo.inode_cache.num_objs
      1661            -0.2%       1658        slabinfo.inode_cache.num_slabs
    182.50 ± 20%     +20.0%     219.00        slabinfo.ip_fib_alias.active_objs
      2.50 ± 20%     +20.0%       3.00        slabinfo.ip_fib_alias.active_slabs
    182.50 ± 20%     +20.0%     219.00        slabinfo.ip_fib_alias.num_objs
      2.50 ± 20%     +20.0%       3.00        slabinfo.ip_fib_alias.num_slabs
    212.50 ± 20%     +20.0%     255.00        slabinfo.ip_fib_trie.active_objs
      2.50 ± 20%     +20.0%       3.00        slabinfo.ip_fib_trie.active_slabs
    212.50 ± 20%     +20.0%     255.00        slabinfo.ip_fib_trie.num_objs
      2.50 ± 20%     +20.0%       3.00        slabinfo.ip_fib_trie.num_slabs
    102444            +0.2%     102635        slabinfo.kernfs_node_cache.active_objs
      1600            +0.2%       1603        slabinfo.kernfs_node_cache.active_slabs
    102444            +0.2%     102635        slabinfo.kernfs_node_cache.num_objs
      1600            +0.2%       1603        slabinfo.kernfs_node_cache.num_slabs
      4579 ±  6%     -10.6%       4093 ± 10%  slabinfo.khugepaged_mm_slot.active_objs
     44.90 ±  6%     -10.6%      40.13 ± 10%  slabinfo.khugepaged_mm_slot.active_slabs
      4579 ±  6%     -10.6%       4093 ± 10%  slabinfo.khugepaged_mm_slot.num_objs
     44.90 ±  6%     -10.6%      40.13 ± 10%  slabinfo.khugepaged_mm_slot.num_slabs
     12242            -1.3%      12080        slabinfo.kmalloc-128.active_objs
    193.24            -1.2%     190.91        slabinfo.kmalloc-128.active_slabs
     12367            -1.2%      12218        slabinfo.kmalloc-128.num_objs
    193.24            -1.2%     190.91        slabinfo.kmalloc-128.num_slabs
     50881            +0.6%      51173        slabinfo.kmalloc-16.active_objs
    199.00            +0.5%     199.99        slabinfo.kmalloc-16.active_slabs
     50944            +0.5%      51198        slabinfo.kmalloc-16.num_objs
    199.00            +0.5%     199.99        slabinfo.kmalloc-16.num_slabs
      8518            -1.0%       8434        slabinfo.kmalloc-192.active_objs
    202.82            -1.0%     200.82        slabinfo.kmalloc-192.active_slabs
      8518            -1.0%       8434        slabinfo.kmalloc-192.num_objs
    202.82            -1.0%     200.82        slabinfo.kmalloc-192.num_slabs
      6877            -0.7%       6829        slabinfo.kmalloc-1k.active_objs
    216.60            -1.1%     214.30        slabinfo.kmalloc-1k.active_slabs
      6931            -1.1%       6857        slabinfo.kmalloc-1k.num_objs
    216.60            -1.1%     214.30        slabinfo.kmalloc-1k.num_slabs
     11098            +0.1%      11111        slabinfo.kmalloc-256.active_objs
    176.92            -0.5%     175.96        slabinfo.kmalloc-256.active_slabs
     11322            -0.5%      11261        slabinfo.kmalloc-256.num_objs
    176.92            -0.5%     175.96        slabinfo.kmalloc-256.num_slabs
      5325            -0.4%       5305        slabinfo.kmalloc-2k.active_objs
    338.52            -0.4%     337.16        slabinfo.kmalloc-2k.active_slabs
      5416            -0.4%       5394        slabinfo.kmalloc-2k.num_objs
    338.52            -0.4%     337.16        slabinfo.kmalloc-2k.num_slabs
     53093            -1.3%      52419        slabinfo.kmalloc-32.active_objs
    415.41            -1.3%     409.98        slabinfo.kmalloc-32.active_slabs
     53172            -1.3%      52477        slabinfo.kmalloc-32.num_objs
    415.41            -1.3%     409.98        slabinfo.kmalloc-32.num_slabs
      2048            +0.6%       2060        slabinfo.kmalloc-4k.active_objs
    259.69            +0.3%     260.39        slabinfo.kmalloc-4k.active_slabs
      2077            +0.3%       2083        slabinfo.kmalloc-4k.num_objs
    259.69            +0.3%     260.39        slabinfo.kmalloc-4k.num_slabs
     21135            +1.8%      21522        slabinfo.kmalloc-512.active_objs
    332.32            +2.0%     338.86        slabinfo.kmalloc-512.active_slabs
     21268            +2.0%      21687        slabinfo.kmalloc-512.num_objs
    332.32            +2.0%     338.86        slabinfo.kmalloc-512.num_slabs
     54000            -0.3%      53842        slabinfo.kmalloc-64.active_objs
    844.55            -0.3%     841.89        slabinfo.kmalloc-64.active_slabs
     54051            -0.3%      53881        slabinfo.kmalloc-64.num_objs
    844.55            -0.3%     841.89        slabinfo.kmalloc-64.num_slabs
     86635 ±  2%      -0.8%      85976        slabinfo.kmalloc-8.active_objs
    174.83            -0.7%     173.64        slabinfo.kmalloc-8.active_slabs
     89515            -0.7%      88901        slabinfo.kmalloc-8.num_objs
    174.83            -0.7%     173.64        slabinfo.kmalloc-8.num_slabs
      1279            +0.3%       1283        slabinfo.kmalloc-8k.active_objs
    322.16            +0.4%     323.51        slabinfo.kmalloc-8k.active_slabs
      1288            +0.4%       1294        slabinfo.kmalloc-8k.num_objs
    322.16            +0.4%     323.51        slabinfo.kmalloc-8k.num_slabs
     28911 ±  2%      +0.5%      29046        slabinfo.kmalloc-96.active_objs
    706.20 ±  2%      +0.9%     712.75        slabinfo.kmalloc-96.active_slabs
     29660 ±  2%      +0.9%      29935        slabinfo.kmalloc-96.num_objs
    706.20 ±  2%      +0.9%     712.75        slabinfo.kmalloc-96.num_slabs
      1216           +10.5%       1344 ±  4%  slabinfo.kmalloc-cg-128.active_objs
     19.00           +10.5%      21.00 ±  4%  slabinfo.kmalloc-cg-128.active_slabs
      1216           +10.5%       1344 ±  4%  slabinfo.kmalloc-cg-128.num_objs
     19.00           +10.5%      21.00 ±  4%  slabinfo.kmalloc-cg-128.num_slabs
      4187 ±  3%      +0.9%       4224 ±  9%  slabinfo.kmalloc-cg-16.active_objs
     16.36 ±  3%      +0.9%      16.50 ±  9%  slabinfo.kmalloc-cg-16.active_slabs
      4187 ±  3%      +0.9%       4224 ±  9%  slabinfo.kmalloc-cg-16.num_objs
     16.36 ±  3%      +0.9%      16.50 ±  9%  slabinfo.kmalloc-cg-16.num_slabs
      5777            -0.3%       5760        slabinfo.kmalloc-cg-192.active_objs
    137.55            -0.3%     137.15        slabinfo.kmalloc-cg-192.active_slabs
      5777            -0.3%       5760        slabinfo.kmalloc-cg-192.num_objs
    137.55            -0.3%     137.15        slabinfo.kmalloc-cg-192.num_slabs
      4473            +0.4%       4493        slabinfo.kmalloc-cg-1k.active_objs
    139.81            +0.4%     140.42        slabinfo.kmalloc-cg-1k.active_slabs
      4473            +0.4%       4493        slabinfo.kmalloc-cg-1k.num_objs
    139.81            +0.4%     140.42        slabinfo.kmalloc-cg-1k.num_slabs
      1056 ±  9%      +9.1%       1152        slabinfo.kmalloc-cg-256.active_objs
     16.50 ±  9%      +9.1%      18.00        slabinfo.kmalloc-cg-256.active_slabs
      1056 ±  9%      +9.1%       1152        slabinfo.kmalloc-cg-256.num_objs
     16.50 ±  9%      +9.1%      18.00        slabinfo.kmalloc-cg-256.num_slabs
      1806            -1.0%       1788 ±  2%  slabinfo.kmalloc-cg-2k.active_objs
    112.93            -1.0%     111.76 ±  2%  slabinfo.kmalloc-cg-2k.active_slabs
      1806            -1.0%       1788 ±  2%  slabinfo.kmalloc-cg-2k.num_objs
    112.93            -1.0%     111.76 ±  2%  slabinfo.kmalloc-cg-2k.num_slabs
     16717            -3.8%      16088        slabinfo.kmalloc-cg-32.active_objs
    130.61            -3.8%     125.69        slabinfo.kmalloc-cg-32.active_slabs
     16717            -3.8%      16088        slabinfo.kmalloc-cg-32.num_objs
    130.61            -3.8%     125.69        slabinfo.kmalloc-cg-32.num_slabs
      1444            +0.5%       1451        slabinfo.kmalloc-cg-4k.active_objs
    187.69            +1.2%     189.94        slabinfo.kmalloc-cg-4k.active_slabs
      1501            +1.2%       1519        slabinfo.kmalloc-cg-4k.num_objs
    187.69            +1.2%     189.94        slabinfo.kmalloc-cg-4k.num_slabs
      8224            +0.8%       8288        slabinfo.kmalloc-cg-512.active_objs
    128.50            +0.8%     129.50        slabinfo.kmalloc-cg-512.active_slabs
      8224            +0.8%       8288        slabinfo.kmalloc-cg-512.num_objs
    128.50            +0.8%     129.50        slabinfo.kmalloc-cg-512.num_slabs
      2795            -1.6%       2749 ±  4%  slabinfo.kmalloc-cg-64.active_objs
     43.67            -1.6%      42.96 ±  4%  slabinfo.kmalloc-cg-64.active_slabs
      2795            -1.6%       2749 ±  4%  slabinfo.kmalloc-cg-64.num_objs
     43.67            -1.6%      42.96 ±  4%  slabinfo.kmalloc-cg-64.num_slabs
     64713            -0.3%      64518        slabinfo.kmalloc-cg-8.active_objs
    126.39            -0.3%     126.01        slabinfo.kmalloc-cg-8.active_slabs
     64713            -0.3%      64518        slabinfo.kmalloc-cg-8.num_objs
    126.39            -0.3%     126.01        slabinfo.kmalloc-cg-8.num_slabs
     42.43 ±  6%      +3.6%      43.94        slabinfo.kmalloc-cg-8k.active_objs
     10.61 ±  6%      +3.6%      10.98        slabinfo.kmalloc-cg-8k.active_slabs
     42.43 ±  6%      +3.6%      43.94        slabinfo.kmalloc-cg-8k.num_objs
     10.61 ±  6%      +3.6%      10.98        slabinfo.kmalloc-cg-8k.num_slabs
      1627            -7.9%       1498 ±  2%  slabinfo.kmalloc-cg-96.active_objs
     38.75            -7.9%      35.68 ±  2%  slabinfo.kmalloc-cg-96.active_slabs
      1627            -7.9%       1498 ±  2%  slabinfo.kmalloc-cg-96.num_objs
     38.75            -7.9%      35.68 ±  2%  slabinfo.kmalloc-cg-96.num_slabs
    448.00 ± 14%     +14.3%     512.00        slabinfo.kmalloc-rcl-128.active_objs
      7.00 ± 14%     +14.3%       8.00        slabinfo.kmalloc-rcl-128.active_slabs
    448.00 ± 14%     +14.3%     512.00        slabinfo.kmalloc-rcl-128.num_objs
      7.00 ± 14%     +14.3%       8.00        slabinfo.kmalloc-rcl-128.num_slabs
    147.00 ± 14%     +14.3%     168.00        slabinfo.kmalloc-rcl-192.active_objs
      3.50 ± 14%     +14.3%       4.00        slabinfo.kmalloc-rcl-192.active_slabs
    147.00 ± 14%     +14.3%     168.00        slabinfo.kmalloc-rcl-192.num_objs
      3.50 ± 14%     +14.3%       4.00        slabinfo.kmalloc-rcl-192.num_slabs
      8069            +1.1%       8162        slabinfo.kmalloc-rcl-64.active_objs
    126.23            +1.0%     127.53        slabinfo.kmalloc-rcl-64.active_slabs
      8078            +1.0%       8162        slabinfo.kmalloc-rcl-64.num_objs
    126.23            +1.0%     127.53        slabinfo.kmalloc-rcl-64.num_slabs
      1547 ± 22%      -4.3%       1481 ±  3%  slabinfo.kmalloc-rcl-96.active_objs
     36.86 ± 22%      -4.3%      35.27 ±  3%  slabinfo.kmalloc-rcl-96.active_slabs
      1547 ± 22%      -4.3%       1481 ±  3%  slabinfo.kmalloc-rcl-96.num_objs
     36.86 ± 22%      -4.3%      35.27 ±  3%  slabinfo.kmalloc-rcl-96.num_slabs
      1056 ±  9%     +12.1%       1184 ±  8%  slabinfo.kmem_cache.active_objs
     16.50 ±  9%     +12.1%      18.50 ±  8%  slabinfo.kmem_cache.active_slabs
      1056 ±  9%     +12.1%       1184 ±  8%  slabinfo.kmem_cache.num_objs
     16.50 ±  9%     +12.1%      18.50 ±  8%  slabinfo.kmem_cache.num_slabs
      1242 ±  7%     +10.3%       1370 ±  7%  slabinfo.kmem_cache_node.active_objs
     19.50 ±  7%     +10.3%      21.50 ±  6%  slabinfo.kmem_cache_node.active_slabs
      1248 ±  7%     +10.3%       1376 ±  6%  slabinfo.kmem_cache_node.num_objs
     19.50 ±  7%     +10.3%      21.50 ±  6%  slabinfo.kmem_cache_node.num_slabs
     25810 ±  2%      -1.9%      25322        slabinfo.lsm_file_cache.active_objs
    154.80 ±  2%      -1.7%     152.13        slabinfo.lsm_file_cache.active_slabs
     26316 ±  2%      -1.7%      25862        slabinfo.lsm_file_cache.num_objs
    154.80 ±  2%      -1.7%     152.13        slabinfo.lsm_file_cache.num_slabs
     21081            -1.6%      20753        slabinfo.maple_node.active_objs
    335.68            -1.5%     330.72        slabinfo.maple_node.active_slabs
     21483            -1.5%      21165        slabinfo.maple_node.num_objs
    335.68            -1.5%     330.72        slabinfo.maple_node.num_slabs
      3229            -0.2%       3221        slabinfo.mm_struct.active_objs
    134.56            -0.2%     134.24        slabinfo.mm_struct.active_slabs
      3229            -0.2%       3221        slabinfo.mm_struct.num_objs
    134.56            -0.2%     134.24        slabinfo.mm_struct.num_slabs
      1224 ±  8%      +0.0%       1224 ±  4%  slabinfo.mnt_cache.active_objs
     24.00 ±  8%      +0.0%      24.00 ±  4%  slabinfo.mnt_cache.active_slabs
      1224 ±  8%      +0.0%       1224 ±  4%  slabinfo.mnt_cache.num_objs
     24.00 ±  8%      +0.0%      24.00 ±  4%  slabinfo.mnt_cache.num_slabs
     34.00            +0.0%      34.00        slabinfo.mqueue_inode_cache.active_objs
      1.00            +0.0%       1.00        slabinfo.mqueue_inode_cache.active_slabs
     34.00            +0.0%      34.00        slabinfo.mqueue_inode_cache.num_objs
      1.00            +0.0%       1.00        slabinfo.mqueue_inode_cache.num_slabs
      1024            +0.0%       1024        slabinfo.names_cache.active_objs
    128.00            +0.0%     128.00        slabinfo.names_cache.active_slabs
      1024            +0.0%       1024        slabinfo.names_cache.num_objs
    128.00            +0.0%     128.00        slabinfo.names_cache.num_slabs
      7.00            +0.0%       7.00        slabinfo.net_namespace.active_objs
      1.00            +0.0%       1.00        slabinfo.net_namespace.active_slabs
      7.00            +0.0%       7.00        slabinfo.net_namespace.num_objs
      1.00            +0.0%       1.00        slabinfo.net_namespace.num_slabs
     46.00            +0.0%      46.00        slabinfo.nfs_commit_data.active_objs
      1.00            +0.0%       1.00        slabinfo.nfs_commit_data.active_slabs
     46.00            +0.0%      46.00        slabinfo.nfs_commit_data.num_objs
      1.00            +0.0%       1.00        slabinfo.nfs_commit_data.num_slabs
     36.00            +0.0%      36.00        slabinfo.nfs_read_data.active_objs
      1.00            +0.0%       1.00        slabinfo.nfs_read_data.active_slabs
     36.00            +0.0%      36.00        slabinfo.nfs_read_data.num_objs
      1.00            +0.0%       1.00        slabinfo.nfs_read_data.num_slabs
    348.54 ±  5%      -7.2%     323.55 ±  2%  slabinfo.nsproxy.active_objs
      6.22 ±  5%      -7.2%       5.78 ±  2%  slabinfo.nsproxy.active_slabs
    348.54 ±  5%      -7.2%     323.55 ±  2%  slabinfo.nsproxy.num_objs
      6.22 ±  5%      -7.2%       5.78 ±  2%  slabinfo.nsproxy.num_slabs
    180.00            +0.0%     180.00        slabinfo.numa_policy.active_objs
      3.00            +0.0%       3.00        slabinfo.numa_policy.active_slabs
    180.00            +0.0%     180.00        slabinfo.numa_policy.num_objs
      3.00            +0.0%       3.00        slabinfo.numa_policy.num_slabs
      5104            +0.3%       5118        slabinfo.perf_event.active_objs
    208.93            +0.5%     209.90        slabinfo.perf_event.active_slabs
      5223            +0.5%       5247        slabinfo.perf_event.num_objs
    208.93            +0.5%     209.90        slabinfo.perf_event.num_slabs
      8986 ±  3%      -0.4%       8953        slabinfo.pid.active_objs
    140.41 ±  3%      -0.4%     139.90        slabinfo.pid.active_slabs
      8986 ±  3%      -0.4%       8953        slabinfo.pid.num_objs
    140.41 ±  3%      -0.4%     139.90        slabinfo.pid.num_slabs
      7933            -0.4%       7904 ±  2%  slabinfo.pool_workqueue.active_objs
    124.43            +0.0%     124.45 ±  2%  slabinfo.pool_workqueue.active_slabs
      7963            +0.0%       7965 ±  2%  slabinfo.pool_workqueue.num_objs
    124.43            +0.0%     124.45 ±  2%  slabinfo.pool_workqueue.num_slabs
      6762            +0.6%       6804        slabinfo.proc_dir_entry.active_objs
    161.00            +0.6%     162.00        slabinfo.proc_dir_entry.active_slabs
      6762            +0.6%       6804        slabinfo.proc_dir_entry.num_objs
    161.00            +0.6%     162.00        slabinfo.proc_dir_entry.num_slabs
     17426            -2.5%      16992        slabinfo.proc_inode_cache.active_objs
    379.09            -2.5%     369.55        slabinfo.proc_inode_cache.active_slabs
     17437            -2.5%      16999        slabinfo.proc_inode_cache.num_objs
    379.09            -2.5%     369.55        slabinfo.proc_inode_cache.num_slabs
     34728            -0.1%      34679        slabinfo.radix_tree_node.active_objs
    621.22            -0.3%     619.51        slabinfo.radix_tree_node.active_slabs
     34788            -0.3%      34692        slabinfo.radix_tree_node.num_objs
    621.22            -0.3%     619.51        slabinfo.radix_tree_node.num_slabs
    332.50 ±  5%     -15.8%     280.00        slabinfo.request_queue.active_objs
     14.50 ±  3%     -10.3%      13.00        slabinfo.request_queue.active_slabs
    507.50 ±  3%     -10.3%     455.00        slabinfo.request_queue.num_objs
     14.50 ±  3%     -10.3%      13.00        slabinfo.request_queue.num_slabs
     46.00            +0.0%      46.00        slabinfo.rpc_inode_cache.active_objs
      1.00            +0.0%       1.00        slabinfo.rpc_inode_cache.active_slabs
     46.00            +0.0%      46.00        slabinfo.rpc_inode_cache.num_objs
      1.00            +0.0%       1.00        slabinfo.rpc_inode_cache.num_slabs
      1504 ±  2%      +0.0%       1504 ±  2%  slabinfo.scsi_sense_cache.active_objs
     24.50 ±  2%      +0.0%      24.50 ±  2%  slabinfo.scsi_sense_cache.active_slabs
      1568 ±  2%      +0.0%       1568 ±  2%  slabinfo.scsi_sense_cache.num_objs
     24.50 ±  2%      +0.0%      24.50 ±  2%  slabinfo.scsi_sense_cache.num_slabs
      9104            -0.7%       9038        slabinfo.seq_file.active_objs
    133.89            -0.7%     132.92        slabinfo.seq_file.active_slabs
      9104            -0.7%       9038        slabinfo.seq_file.num_objs
    133.89            -0.7%     132.92        slabinfo.seq_file.num_slabs
     23757            +0.4%      23842        slabinfo.shared_policy_node.active_objs
    279.50            +0.4%     280.50        slabinfo.shared_policy_node.active_slabs
     23757            +0.4%      23842        slabinfo.shared_policy_node.num_objs
    279.50            +0.4%     280.50        slabinfo.shared_policy_node.num_slabs
      5808 ±  2%      -2.0%       5694        slabinfo.shmem_inode_cache.active_objs
    141.66 ±  2%      -2.0%     138.90        slabinfo.shmem_inode_cache.active_slabs
      5808 ±  2%      -2.0%       5694        slabinfo.shmem_inode_cache.num_objs
    141.66 ±  2%      -2.0%     138.90        slabinfo.shmem_inode_cache.num_slabs
      3258            +0.4%       3270 ±  2%  slabinfo.sighand_cache.active_objs
    217.31            +0.4%     218.09 ±  2%  slabinfo.sighand_cache.active_slabs
      3259            +0.4%       3271 ±  2%  slabinfo.sighand_cache.num_objs
    217.31            +0.4%     218.09 ±  2%  slabinfo.sighand_cache.num_slabs
      5256 ±  4%      -2.2%       5139 ±  2%  slabinfo.signal_cache.active_objs
    187.78 ±  4%      -2.3%     183.54 ±  2%  slabinfo.signal_cache.active_slabs
      5257 ±  4%      -2.3%       5139 ±  2%  slabinfo.signal_cache.num_objs
    187.78 ±  4%      -2.3%     183.54 ±  2%  slabinfo.signal_cache.num_slabs
      6952            +1.0%       7023        slabinfo.sigqueue.active_objs
    136.31            +1.0%     137.71        slabinfo.sigqueue.active_slabs
      6952            +1.0%       7023        slabinfo.sigqueue.num_objs
    136.31            +1.0%     137.71        slabinfo.sigqueue.num_slabs
    554.58            +5.9%     587.36 ±  7%  slabinfo.skbuff_ext_cache.active_objs
     13.20            +5.9%      13.98 ±  7%  slabinfo.skbuff_ext_cache.active_slabs
    554.58            +5.9%     587.36 ±  7%  slabinfo.skbuff_ext_cache.num_objs
     13.20            +5.9%      13.98 ±  7%  slabinfo.skbuff_ext_cache.num_slabs
      8971            +1.6%       9118        slabinfo.skbuff_head_cache.active_objs
    140.18            +1.6%     142.47        slabinfo.skbuff_head_cache.active_slabs
      8971            +1.6%       9118        slabinfo.skbuff_head_cache.num_objs
    140.18            +1.6%     142.47        slabinfo.skbuff_head_cache.num_slabs
      6109 ±  4%      +9.5%       6692        slabinfo.skbuff_small_head.active_objs
    119.79 ±  4%      +9.5%     131.23        slabinfo.skbuff_small_head.active_slabs
      6109 ±  4%      +9.5%       6692        slabinfo.skbuff_small_head.num_objs
    119.79 ±  4%      +9.5%     131.23        slabinfo.skbuff_small_head.num_slabs
      3252 ±  4%      +9.1%       3547        slabinfo.sock_inode_cache.active_objs
     83.40 ±  4%      +9.1%      90.95        slabinfo.sock_inode_cache.active_slabs
      3252 ±  4%      +9.1%       3547        slabinfo.sock_inode_cache.num_objs
     83.40 ±  4%      +9.1%      90.95        slabinfo.sock_inode_cache.num_slabs
      1469 ±  8%      -2.6%       1431 ±  4%  slabinfo.task_group.active_objs
     28.82 ±  8%      -2.6%      28.08 ±  4%  slabinfo.task_group.active_slabs
      1469 ±  8%      -2.6%       1431 ±  4%  slabinfo.task_group.num_objs
     28.82 ±  8%      -2.6%      28.08 ±  4%  slabinfo.task_group.num_slabs
      2331            -3.8%       2244        slabinfo.task_struct.active_objs
      2334            -3.8%       2245        slabinfo.task_struct.active_slabs
      2334            -3.8%       2245        slabinfo.task_struct.num_objs
      2334            -3.8%       2245        slabinfo.task_struct.num_slabs
    299.10            +6.4%     318.17 ±  5%  slabinfo.taskstats.active_objs
      8.08            +6.4%       8.60 ±  5%  slabinfo.taskstats.active_slabs
    299.10            +6.4%     318.17 ±  5%  slabinfo.taskstats.num_objs
      8.08            +6.4%       8.60 ±  5%  slabinfo.taskstats.num_slabs
      2369            +1.9%       2415        slabinfo.trace_event_file.active_objs
     51.50            +1.9%      52.50        slabinfo.trace_event_file.active_slabs
      2369            +1.9%       2415        slabinfo.trace_event_file.num_objs
     51.50            +1.9%      52.50        slabinfo.trace_event_file.num_slabs
      1585 ±  2%      +1.2%       1603        slabinfo.tracefs_inode_cache.active_objs
     31.70 ±  2%      +1.2%      32.07        slabinfo.tracefs_inode_cache.active_slabs
      1585 ±  2%      +1.2%       1603        slabinfo.tracefs_inode_cache.num_objs
     31.70 ±  2%      +1.2%      32.07        slabinfo.tracefs_inode_cache.num_slabs
    116.34            +0.0%     116.34        slabinfo.tw_sock_TCP.active_objs
      1.94            +0.0%       1.94        slabinfo.tw_sock_TCP.active_slabs
    116.34            +0.0%     116.34        slabinfo.tw_sock_TCP.num_objs
      1.94            +0.0%       1.94        slabinfo.tw_sock_TCP.num_slabs
    111.00            +0.0%     111.00        slabinfo.uts_namespace.active_objs
      3.00            +0.0%       3.00        slabinfo.uts_namespace.active_slabs
    111.00            +0.0%     111.00        slabinfo.uts_namespace.num_objs
      3.00            +0.0%       3.00        slabinfo.uts_namespace.num_slabs
     32868            -1.8%      32279        slabinfo.vm_area_struct.active_objs
    750.95            -1.8%     737.25        slabinfo.vm_area_struct.active_slabs
     33041            -1.8%      32439        slabinfo.vm_area_struct.num_objs
    750.95            -1.8%     737.25        slabinfo.vm_area_struct.num_slabs
     42944            +0.7%      43254        slabinfo.vma_lock.active_objs
    423.60            +1.0%     427.95        slabinfo.vma_lock.active_slabs
     43207            +1.0%      43651        slabinfo.vma_lock.num_objs
    423.60            +1.0%     427.95        slabinfo.vma_lock.num_slabs
    201997            -0.2%     201641        slabinfo.vmap_area.active_objs
      3612            -0.2%       3605        slabinfo.vmap_area.active_slabs
    202318            -0.2%     201897        slabinfo.vmap_area.num_objs
      3612            -0.2%       3605        slabinfo.vmap_area.num_slabs



^ permalink raw reply	[flat|nested] 9+ messages in thread

end of thread, other threads:[~2023-10-19  5:41 UTC | newest]

Thread overview: 9+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2023-10-17 15:06 [linux-next:master] [drivers/char/mem] 1b057bd800: stress-ng.splice.ops_per_sec -99.8% regression kernel test robot
2023-10-17 16:56 ` Greg Kroah-Hartman
2023-10-18  6:31   ` Max Kellermann
2023-10-18  7:07   ` Oliver Sang
2023-10-18  7:57     ` Greg Kroah-Hartman
2023-10-18  8:12       ` Max Kellermann
2023-10-18 10:01     ` Max Kellermann
     [not found] ` <CAKPOu+_T8xk4yd2P4KT4j3eMoFqwYmkxqDHaFtv4Hii5-XyPuA@mail.gmail.com>
2023-10-18 11:12   ` Max Kellermann
2023-10-19  5:41     ` Oliver Sang

This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox