linux-mm.kvack.org archive mirror
 help / color / mirror / Atom feed
* [linus:master] [mm, pcp]  6ccdcb6d3a:  stress-ng.judy.ops_per_sec -4.7% regression
@ 2023-11-23  5:03 kernel test robot
  2023-11-23  5:40 ` Huang, Ying
  0 siblings, 1 reply; 3+ messages in thread
From: kernel test robot @ 2023-11-23  5:03 UTC (permalink / raw)
  To: Huang Ying
  Cc: oe-lkp, lkp, linux-kernel, Andrew Morton, Mel Gorman,
	Vlastimil Babka, David Hildenbrand, Johannes Weiner, Dave Hansen,
	Michal Hocko, Pavel Tatashin, Matthew Wilcox, Christoph Lameter,
	Arjan van de Ven, Sudeep Holla, linux-mm, ying.huang, feng.tang,
	fengwei.yin, oliver.sang



Hello,

kernel test robot noticed a -4.7% regression of stress-ng.judy.ops_per_sec on:


commit: 6ccdcb6d3a741c4e005ca6ffd4a62ddf8b5bead3 ("mm, pcp: reduce detecting time of consecutive high order page freeing")
https://git.kernel.org/cgit/linux/kernel/git/torvalds/linux.git master

testcase: stress-ng
test machine: 224 threads 2 sockets Intel(R) Xeon(R) Platinum 8480CTDX (Sapphire Rapids) with 512G memory
parameters:

	nr_threads: 100%
	testtime: 60s
	class: cpu-cache
	test: judy
	disk: 1SSD
	cpufreq_governor: performance


In addition to that, the commit also has significant impact on the following tests:

+------------------+-------------------------------------------------------------------------------------------------+
| testcase: change | lmbench3: lmbench3.TCP.socket.bandwidth.10MB.MB/sec 23.7% improvement                           |
| test machine     | 48 threads 2 sockets Intel(R) Xeon(R) CPU E5-2697 v2 @ 2.70GHz (Ivy Bridge-EP) with 112G memory |
| test parameters  | cpufreq_governor=performance                                                                    |
|                  | mode=development                                                                                |
|                  | nr_threads=100%                                                                                 |
|                  | test=TCP                                                                                        |
|                  | test_memory_size=50%                                                                            |
+------------------+-------------------------------------------------------------------------------------------------+
| testcase: change | stress-ng: stress-ng.file-ioctl.ops_per_sec -6.6% regression                                    |
| test machine     | 36 threads 1 sockets Intel(R) Core(TM) i9-9980XE CPU @ 3.00GHz (Skylake) with 32G memory        |
| test parameters  | class=filesystem                                                                                |
|                  | cpufreq_governor=performance                                                                    |
|                  | disk=1SSD                                                                                       |
|                  | fs=btrfs                                                                                        |
|                  | nr_threads=10%                                                                                  |
|                  | test=file-ioctl                                                                                 |
|                  | testtime=60s                                                                                    |
+------------------+-------------------------------------------------------------------------------------------------+


If you fix the issue in a separate patch/commit (i.e. not just a new version of
the same patch/commit), kindly add following tags
| Reported-by: kernel test robot <oliver.sang@intel.com>
| Closes: https://lore.kernel.org/oe-lkp/202311231029.3aa790-oliver.sang@intel.com


Details are as below:
-------------------------------------------------------------------------------------------------->


The kernel config and materials to reproduce are available at:
https://download.01.org/0day-ci/archive/20231123/202311231029.3aa790-oliver.sang@intel.com

=========================================================================================
class/compiler/cpufreq_governor/disk/kconfig/nr_threads/rootfs/tbox_group/test/testcase/testtime:
  cpu-cache/gcc-12/performance/1SSD/x86_64-rhel-8.3/100%/debian-11.1-x86_64-20220510.cgz/lkp-spr-2sp4/judy/stress-ng/60s

commit: 
  57c0419c5f ("mm, pcp: decrease PCP high if free pages < high watermark")
  6ccdcb6d3a ("mm, pcp: reduce detecting time of consecutive high order page freeing")

57c0419c5f0ea2cc 6ccdcb6d3a741c4e005ca6ffd4a 
---------------- --------------------------- 
         %stddev     %change         %stddev
             \          |                \  
      4.57 ±  5%     +46.8%       6.71 ± 17%  iostat.cpu.system
      2842            +1.0%       2871        turbostat.Bzy_MHz
      0.12 ±  3%      +0.4        0.55 ± 26%  mpstat.cpu.all.soft%
      3.05 ±  6%      +1.8        4.86 ± 20%  mpstat.cpu.all.sys%
  81120642            -2.9%   78746159        proc-vmstat.numa_hit
  80886548            -2.9%   78513494        proc-vmstat.numa_local
  82771023            -2.9%   80399459        proc-vmstat.pgalloc_normal
  82356596            -2.9%   79991041        proc-vmstat.pgfree
  12325708 ±  3%      +5.3%   12974746        perf-stat.i.dTLB-load-misses
      0.38 ± 44%     +27.2%       0.48        perf-stat.overall.cpi
    668.74 ± 44%     +24.7%     834.02        perf-stat.overall.cycles-between-cache-misses
      0.00 ± 45%      +0.0        0.01 ± 10%  perf-stat.overall.dTLB-load-miss-rate%
  10040254 ± 44%     +26.0%   12650801        perf-stat.ps.dTLB-load-misses
   7036371 ±  3%      -2.8%    6842720        stress-ng.judy.Judy_delete_operations_per_sec
   9244466 ±  3%      -7.8%    8524505 ±  3%  stress-ng.judy.Judy_insert_operations_per_sec
      2912 ±  3%      -4.7%       2774        stress-ng.judy.ops_per_sec
     13316 ±  8%     +22.8%      16355 ± 13%  stress-ng.time.maximum_resident_set_size
    445.86 ±  5%     +64.2%     732.21 ± 15%  stress-ng.time.system_time
     40885 ± 40%    +373.8%     193712 ± 11%  sched_debug.cfs_rq:/.left_vruntime.avg
    465264 ± 31%    +142.5%    1128399 ±  5%  sched_debug.cfs_rq:/.left_vruntime.stddev
      8322 ± 34%    +140.8%      20039 ± 17%  sched_debug.cfs_rq:/.load.avg
     40886 ± 40%    +373.8%     193713 ± 11%  sched_debug.cfs_rq:/.right_vruntime.avg
    465274 ± 31%    +142.5%    1128401 ±  5%  sched_debug.cfs_rq:/.right_vruntime.stddev
    818.77 ± 10%     +43.3%       1172 ±  5%  sched_debug.cpu.curr->pid.stddev
      0.05 ± 74%    +659.6%       0.41 ± 35%  perf-sched.sch_delay.avg.ms.do_task_dead.do_exit.do_group_exit.__x64_sys_exit_group.do_syscall_64
      0.10 ± 48%    +140.3%       0.24 ± 11%  perf-sched.sch_delay.avg.ms.do_wait.kernel_wait4.__do_sys_wait4.do_syscall_64
      0.01 ± 14%    +102.6%       0.03 ± 29%  perf-sched.sch_delay.avg.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
      0.05 ±122%   +1322.6%       0.65 ± 20%  perf-sched.sch_delay.avg.ms.wait_for_partner.fifo_open.do_dentry_open.do_open
      1.70 ± 79%    +729.3%      14.10 ± 48%  perf-sched.sch_delay.max.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
      1.08 ±101%    +233.4%       3.60 ±  7%  perf-sched.sch_delay.max.ms.wait_for_partner.fifo_open.do_dentry_open.do_open
      0.01 ±  8%     +54.7%       0.02 ± 18%  perf-sched.total_sch_delay.average.ms
      0.18 ±  5%    +555.7%       1.20 ± 38%  perf-sched.wait_and_delay.avg.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt
      0.21 ±  4%    +524.6%       1.29 ± 47%  perf-sched.wait_and_delay.avg.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_reschedule_ipi
    235.65 ± 31%     -57.0%     101.40 ± 17%  perf-sched.wait_and_delay.avg.ms.pipe_read.vfs_read.ksys_read.do_syscall_64
    127.50 ±100%    +126.3%     288.50 ±  9%  perf-sched.wait_and_delay.count.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_call_function_single
    125.83 ±144%    +407.2%     638.17 ± 27%  perf-sched.wait_and_delay.count.exit_to_user_mode_loop.exit_to_user_mode_prepare.syscall_exit_to_user_mode.do_syscall_64
    344.50 ± 36%    +114.6%     739.33 ± 24%  perf-sched.wait_and_delay.count.pipe_read.vfs_read.ksys_read.do_syscall_64
      0.92 ±114%    +482.2%       5.38 ± 47%  perf-sched.wait_and_delay.max.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_call_function_single
      3.22 ± 89%    +223.9%      10.44 ± 50%  perf-sched.wait_and_delay.max.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_reschedule_ipi
      0.18 ± 43%    +471.8%       1.01 ± 36%  perf-sched.wait_time.avg.ms.__cond_resched.__alloc_pages.__folio_alloc.vma_alloc_folio.do_anonymous_page
     34.39 ± 46%     +88.8%      64.95 ± 18%  perf-sched.wait_time.avg.ms.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
      0.21 ± 13%    +813.6%       1.95 ± 38%  perf-sched.wait_time.avg.ms.__cond_resched.tlb_batch_pages_flush.tlb_finish_mmu.unmap_region.constprop
      0.18 ± 15%    +457.1%       1.02 ± 58%  perf-sched.wait_time.avg.ms.__cond_resched.unmap_vmas.unmap_region.constprop.0
    417.61 ± 68%     -87.6%      51.85 ±146%  perf-sched.wait_time.avg.ms.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
      0.22 ± 25%    +614.2%       1.57 ± 71%  perf-sched.wait_time.avg.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_exc_page_fault
      0.18 ±  5%    +556.3%       1.20 ± 38%  perf-sched.wait_time.avg.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt
      0.21 ±  4%    +524.6%       1.29 ± 47%  perf-sched.wait_time.avg.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_reschedule_ipi
     38.72 ± 39%     -53.1%      18.17 ± 30%  perf-sched.wait_time.avg.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.syscall_exit_to_user_mode.do_syscall_64
    235.60 ± 31%     -57.0%     101.37 ± 17%  perf-sched.wait_time.avg.ms.pipe_read.vfs_read.ksys_read.do_syscall_64
      2.17 ± 30%     +45.3%       3.16 ± 13%  perf-sched.wait_time.avg.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone
      1.02 ±131%    +574.3%       6.90 ± 52%  perf-sched.wait_time.max.ms.__cond_resched.__alloc_pages.__folio_alloc.vma_alloc_folio.do_anonymous_page
      0.18 ±191%  +92359.0%     169.05 ±219%  perf-sched.wait_time.max.ms.__cond_resched.dput.__fput.__x64_sys_close.do_syscall_64
     69.64 ± 44%     +33.2%      92.76 ±  4%  perf-sched.wait_time.max.ms.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
      0.64 ± 67%    +653.6%       4.82 ± 54%  perf-sched.wait_time.max.ms.__cond_resched.unmap_vmas.unmap_region.constprop.0
      1.75 ± 49%    +206.5%       5.38 ± 47%  perf-sched.wait_time.max.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_call_function_single
      3.22 ± 89%    +223.9%      10.44 ± 50%  perf-sched.wait_time.max.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_reschedule_ipi


***************************************************************************************************
lkp-ivb-2ep1: 48 threads 2 sockets Intel(R) Xeon(R) CPU E5-2697 v2 @ 2.70GHz (Ivy Bridge-EP) with 112G memory
=========================================================================================
compiler/cpufreq_governor/kconfig/mode/nr_threads/rootfs/tbox_group/test/test_memory_size/testcase:
  gcc-12/performance/x86_64-rhel-8.3/development/100%/debian-11.1-x86_64-20220510.cgz/lkp-ivb-2ep1/TCP/50%/lmbench3

commit: 
  57c0419c5f ("mm, pcp: decrease PCP high if free pages < high watermark")
  6ccdcb6d3a ("mm, pcp: reduce detecting time of consecutive high order page freeing")

57c0419c5f0ea2cc 6ccdcb6d3a741c4e005ca6ffd4a 
---------------- --------------------------- 
         %stddev     %change         %stddev
             \          |                \  
      0.07 ± 38%    +105.0%       0.14 ± 32%  perf-sched.sch_delay.max.ms.wait_for_partner.fifo_open.do_dentry_open.do_open
     26.75            -4.9%      25.45        turbostat.RAMWatt
    678809            +7.2%     727594 ±  2%  vmstat.system.cs
  97929782           -13.1%   85054266        numa-numastat.node0.local_node
  97933343           -13.1%   85056081        numa-numastat.node0.numa_hit
  97933344           -13.1%   85055901        numa-vmstat.node0.numa_hit
  97929783           -13.1%   85054086        numa-vmstat.node0.numa_local
     32188           +23.7%      39813        lmbench3.TCP.socket.bandwidth.10MB.MB/sec
    652.63            -4.4%     624.04        lmbench3.time.elapsed_time
    652.63            -4.4%     624.04        lmbench3.time.elapsed_time.max
      8597            -5.9%       8092        lmbench3.time.system_time
      0.88 ±  7%      -0.1        0.76 ±  5%  perf-profile.calltrace.cycles-pp.security_file_permission.vfs_read.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe
      0.71 ± 10%      -0.1        0.61 ±  7%  perf-profile.calltrace.cycles-pp.apparmor_file_permission.security_file_permission.vfs_read.ksys_read.do_syscall_64
      0.78 ±  3%      -0.1        0.70 ±  6%  perf-profile.children.cycles-pp.security_socket_recvmsg
      0.36 ±  9%      +0.1        0.42 ± 11%  perf-profile.children.cycles-pp.skb_page_frag_refill
      0.40 ± 10%      +0.1        0.48 ± 12%  perf-profile.children.cycles-pp.sk_page_frag_refill
      0.51 ±  4%      -0.1        0.44 ± 13%  perf-profile.self.cycles-pp.sock_read_iter
      0.36 ± 10%      +0.1        0.42 ± 11%  perf-profile.self.cycles-pp.skb_page_frag_refill
    158897 ±  2%      -6.8%     148107        proc-vmstat.nr_anon_pages
    160213 ±  2%      -6.8%     149290        proc-vmstat.nr_inactive_anon
    160213 ±  2%      -6.8%     149290        proc-vmstat.nr_zone_inactive_anon
 1.715e+08            -7.1%  1.593e+08        proc-vmstat.numa_hit
 1.715e+08            -7.1%  1.592e+08        proc-vmstat.numa_local
 1.367e+09            -7.1%   1.27e+09        proc-vmstat.pgalloc_normal
   2324641            -2.7%    2261187        proc-vmstat.pgfault
 1.367e+09            -7.1%   1.27e+09        proc-vmstat.pgfree
     77011            -4.4%      73597        proc-vmstat.pgreuse
      5.99 ±  3%     -29.9%       4.20 ±  4%  perf-stat.i.MPKI
 7.914e+09 ±  2%      +4.5%  8.271e+09        perf-stat.i.branch-instructions
  1.51e+08            +4.6%  1.579e+08        perf-stat.i.branch-misses
      7.65 ±  4%      -0.9        6.73 ±  3%  perf-stat.i.cache-miss-rate%
  66394790 ±  2%     -21.9%   51865866 ±  3%  perf-stat.i.cache-misses
    682132            +7.2%     731279 ±  2%  perf-stat.i.context-switches
      4.01           -16.0%       3.37        perf-stat.i.cpi
     71772 ±  4%     +11.5%      80055 ±  8%  perf-stat.i.cycles-between-cache-misses
 9.368e+09 ±  2%      +3.6%  9.706e+09        perf-stat.i.dTLB-stores
  33695419 ±  2%      +7.1%   36096466 ±  2%  perf-stat.i.iTLB-load-misses
    573897 ± 35%     -38.6%     352477 ± 19%  perf-stat.i.iTLB-loads
  4.09e+10 ±  2%      +4.5%  4.273e+10        perf-stat.i.instructions
      0.37            +4.3%       0.39        perf-stat.i.ipc
      0.09 ± 22%     -44.0%       0.05 ± 26%  perf-stat.i.major-faults
    490.16 ±  2%      -8.6%     448.21 ±  2%  perf-stat.i.metric.K/sec
    635.38 ±  2%      +3.5%     657.46        perf-stat.i.metric.M/sec
     37.54            +2.3       39.84        perf-stat.i.node-load-miss-rate%
   8300835 ±  2%     -10.8%    7406820 ±  2%  perf-stat.i.node-load-misses
  76993977 ±  3%      -6.6%   71936169 ±  3%  perf-stat.i.node-loads
     26.58 ±  4%      +4.1       30.71 ±  3%  perf-stat.i.node-store-miss-rate%
   2341211 ±  4%     -29.6%    1648802 ±  3%  perf-stat.i.node-store-misses
  34198780 ±  3%     -33.2%   22857201 ±  3%  perf-stat.i.node-stores
      1.63           -25.5%       1.21 ±  3%  perf-stat.overall.MPKI
     10.67            -2.3        8.36        perf-stat.overall.cache-miss-rate%
      2.83            -5.2%       2.69        perf-stat.overall.cpi
      1740           +27.3%       2216 ±  3%  perf-stat.overall.cycles-between-cache-misses
      0.35            +5.5%       0.37        perf-stat.overall.ipc
      9.73            -0.4        9.34        perf-stat.overall.node-load-miss-rate%
      6.39            +0.3        6.72        perf-stat.overall.node-store-miss-rate%
 7.914e+09 ±  2%      +4.6%  8.276e+09        perf-stat.ps.branch-instructions
 1.509e+08            +4.7%  1.579e+08        perf-stat.ps.branch-misses
  66615187 ±  2%     -22.1%   51881477 ±  3%  perf-stat.ps.cache-misses
    679734            +7.2%     729007 ±  2%  perf-stat.ps.context-switches
 9.369e+09 ±  2%      +3.7%  9.712e+09        perf-stat.ps.dTLB-stores
  33673038 ±  2%      +7.2%   36098564 ±  2%  perf-stat.ps.iTLB-load-misses
  4.09e+10 ±  2%      +4.6%  4.276e+10        perf-stat.ps.instructions
      0.09 ± 23%     -44.4%       0.05 ± 26%  perf-stat.ps.major-faults
   8328473 ±  2%     -11.0%    7410272 ±  2%  perf-stat.ps.node-load-misses
  77301667 ±  3%      -6.9%   71997671 ±  3%  perf-stat.ps.node-loads
   2344250 ±  4%     -29.7%    1647553 ±  3%  perf-stat.ps.node-store-misses
  34315831 ±  3%     -33.4%   22865994 ±  3%  perf-stat.ps.node-stores



***************************************************************************************************
lkp-skl-d08: 36 threads 1 sockets Intel(R) Core(TM) i9-9980XE CPU @ 3.00GHz (Skylake) with 32G memory
=========================================================================================
class/compiler/cpufreq_governor/disk/fs/kconfig/nr_threads/rootfs/tbox_group/test/testcase/testtime:
  filesystem/gcc-12/performance/1SSD/btrfs/x86_64-rhel-8.3/10%/debian-11.1-x86_64-20220510.cgz/lkp-skl-d08/file-ioctl/stress-ng/60s

commit: 
  57c0419c5f ("mm, pcp: decrease PCP high if free pages < high watermark")
  6ccdcb6d3a ("mm, pcp: reduce detecting time of consecutive high order page freeing")

57c0419c5f0ea2cc 6ccdcb6d3a741c4e005ca6ffd4a 
---------------- --------------------------- 
         %stddev     %change         %stddev
             \          |                \  
    127.00 ± 10%     +36.1%     172.83 ± 15%  perf-c2c.HITM.local
      0.00 ± 72%    +130.4%       0.01 ± 30%  perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc.alloc_extent_state.__clear_extent_bit.btrfs_clone_files
     14.83 ± 19%     +33.7%      19.83 ± 10%  sched_debug.cpu.nr_uninterruptible.max
    339939            -6.6%     317593        stress-ng.file-ioctl.ops
      5665            -6.6%       5293        stress-ng.file-ioctl.ops_per_sec
      6444 ±  4%     -25.2%       4820 ±  5%  stress-ng.time.involuntary_context_switches
  89198237            -6.5%   83411572        proc-vmstat.numa_hit
  89117176            -6.8%   83056324        proc-vmstat.numa_local
  92833230            -6.6%   86743293        proc-vmstat.pgalloc_normal
  92791999            -6.6%   86700599        proc-vmstat.pgfree
      0.25 ± 56%    +110.2%       0.53 ± 12%  perf-stat.i.major-faults
    127575 ± 27%    +138.3%     303957 ±  3%  perf-stat.i.node-stores
      0.25 ± 56%    +110.2%       0.52 ± 12%  perf-stat.ps.major-faults
    125751 ± 27%    +138.3%     299653 ±  3%  perf-stat.ps.node-stores
 1.199e+12            -2.1%  1.174e+12        perf-stat.total.instructions
     15.80            -0.7       15.14        perf-profile.calltrace.cycles-pp.filemap_read_folio.do_read_cache_folio.vfs_dedupe_file_range_compare.__generic_remap_file_range_prep.generic_remap_file_range_prep
     15.46            -0.6       14.84        perf-profile.calltrace.cycles-pp.btrfs_read_folio.filemap_read_folio.do_read_cache_folio.vfs_dedupe_file_range_compare.__generic_remap_file_range_prep
      9.84            -0.5        9.32        perf-profile.calltrace.cycles-pp.memcmp.vfs_dedupe_file_range_compare.__generic_remap_file_range_prep.generic_remap_file_range_prep.btrfs_remap_file_range
     11.95            -0.4       11.52        perf-profile.calltrace.cycles-pp.btrfs_do_readpage.btrfs_read_folio.filemap_read_folio.do_read_cache_folio.vfs_dedupe_file_range_compare
      8.72 ±  2%      -0.4        8.28        perf-profile.calltrace.cycles-pp.filemap_add_folio.do_read_cache_folio.vfs_dedupe_file_range_compare.__generic_remap_file_range_prep.generic_remap_file_range_prep
      5.56 ±  2%      -0.4        5.18        perf-profile.calltrace.cycles-pp.__filemap_add_folio.filemap_add_folio.do_read_cache_folio.vfs_dedupe_file_range_compare.__generic_remap_file_range_prep
      0.64 ± 10%      -0.3        0.36 ± 71%  perf-profile.calltrace.cycles-pp.find_free_extent.btrfs_reserve_extent.__btrfs_prealloc_file_range.btrfs_prealloc_file_range.btrfs_fallocate
      2.57 ±  5%      -0.3        2.29 ±  2%  perf-profile.calltrace.cycles-pp.ioctl_preallocate.__x64_sys_ioctl.do_syscall_64.entry_SYSCALL_64_after_hwframe.ioctl
      2.44 ±  6%      -0.3        2.17 ±  2%  perf-profile.calltrace.cycles-pp.btrfs_fallocate.vfs_fallocate.ioctl_preallocate.__x64_sys_ioctl.do_syscall_64
      2.53 ±  5%      -0.3        2.26 ±  2%  perf-profile.calltrace.cycles-pp.vfs_fallocate.ioctl_preallocate.__x64_sys_ioctl.do_syscall_64.entry_SYSCALL_64_after_hwframe
      0.66 ±  9%      -0.2        0.46 ± 45%  perf-profile.calltrace.cycles-pp.btrfs_reserve_extent.__btrfs_prealloc_file_range.btrfs_prealloc_file_range.btrfs_fallocate.vfs_fallocate
      1.42 ±  3%      -0.1        1.31 ±  4%  perf-profile.calltrace.cycles-pp.clear_state_bit.__clear_extent_bit.btrfs_invalidate_folio.truncate_cleanup_folio.truncate_inode_pages_range
      0.70 ±  4%      -0.1        0.62 ±  2%  perf-profile.calltrace.cycles-pp.__mod_lruvec_page_state.__filemap_add_folio.filemap_add_folio.do_read_cache_folio.vfs_dedupe_file_range_compare
      0.69 ±  4%      -0.1        0.63 ±  4%  perf-profile.calltrace.cycles-pp.btrfs_punch_hole.btrfs_fallocate.vfs_fallocate.ioctl_preallocate.__x64_sys_ioctl
     29.90            +0.6       30.49        perf-profile.calltrace.cycles-pp.do_read_cache_folio.vfs_dedupe_file_range_compare.__generic_remap_file_range_prep.generic_remap_file_range_prep.btrfs_remap_file_range
      0.00            +0.9        0.86 ±  6%  perf-profile.calltrace.cycles-pp.__list_del_entry_valid_or_report.rmqueue_bulk.__rmqueue_pcplist.rmqueue.get_page_from_freelist
     68.10            +1.2       69.29        perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.ioctl.stress_run
     68.47            +1.2       69.68        perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.ioctl.stress_run
     67.35            +1.2       68.59        perf-profile.calltrace.cycles-pp.__x64_sys_ioctl.do_syscall_64.entry_SYSCALL_64_after_hwframe.ioctl.stress_run
     21.54 ±  3%      +1.5       23.02        perf-profile.calltrace.cycles-pp.ioctl_file_clone.do_vfs_ioctl.__x64_sys_ioctl.do_syscall_64.entry_SYSCALL_64_after_hwframe
     21.51 ±  3%      +1.5       23.00        perf-profile.calltrace.cycles-pp.do_clone_file_range.vfs_clone_file_range.ioctl_file_clone.do_vfs_ioctl.__x64_sys_ioctl
     21.46 ±  3%      +1.5       22.94        perf-profile.calltrace.cycles-pp.btrfs_remap_file_range.do_clone_file_range.vfs_clone_file_range.ioctl_file_clone.do_vfs_ioctl
     21.53 ±  3%      +1.5       23.01        perf-profile.calltrace.cycles-pp.vfs_clone_file_range.ioctl_file_clone.do_vfs_ioctl.__x64_sys_ioctl.do_syscall_64
      0.00            +1.5        1.49 ±  3%  perf-profile.calltrace.cycles-pp.__free_one_page.free_pcppages_bulk.free_unref_page_commit.free_unref_page.btrfs_clone
     21.15 ±  3%      +1.5       22.66        perf-profile.calltrace.cycles-pp.btrfs_clone_files.btrfs_remap_file_range.do_clone_file_range.vfs_clone_file_range.ioctl_file_clone
     64.61            +1.5       66.16        perf-profile.calltrace.cycles-pp.do_vfs_ioctl.__x64_sys_ioctl.do_syscall_64.entry_SYSCALL_64_after_hwframe.ioctl
      2.66 ±  2%      +1.8        4.51 ±  3%  perf-profile.calltrace.cycles-pp.folio_alloc.do_read_cache_folio.vfs_dedupe_file_range_compare.__generic_remap_file_range_prep.generic_remap_file_range_prep
      0.97 ±  3%      +1.8        2.82 ±  5%  perf-profile.calltrace.cycles-pp.rmqueue.get_page_from_freelist.__alloc_pages.folio_alloc.do_read_cache_folio
      2.02 ±  3%      +1.9        3.90 ±  4%  perf-profile.calltrace.cycles-pp.__alloc_pages.folio_alloc.do_read_cache_folio.vfs_dedupe_file_range_compare.__generic_remap_file_range_prep
      1.27 ±  2%      +1.9        3.17 ±  4%  perf-profile.calltrace.cycles-pp.get_page_from_freelist.__alloc_pages.folio_alloc.do_read_cache_folio.vfs_dedupe_file_range_compare
      0.35 ± 70%      +2.0        2.31 ±  5%  perf-profile.calltrace.cycles-pp.__rmqueue_pcplist.rmqueue.get_page_from_freelist.__alloc_pages.folio_alloc
      0.00            +2.0        2.00 ±  4%  perf-profile.calltrace.cycles-pp.rmqueue_bulk.__rmqueue_pcplist.rmqueue.get_page_from_freelist.__alloc_pages
      1.72 ±  2%      +2.1        3.78        perf-profile.calltrace.cycles-pp.btrfs_clone.btrfs_clone_files.btrfs_remap_file_range.do_clone_file_range.vfs_clone_file_range
      0.00            +2.1        2.09 ±  2%  perf-profile.calltrace.cycles-pp.free_pcppages_bulk.free_unref_page_commit.free_unref_page.btrfs_clone.btrfs_clone_files
      0.00            +2.1        2.12 ±  2%  perf-profile.calltrace.cycles-pp.free_unref_page_commit.free_unref_page.btrfs_clone.btrfs_clone_files.btrfs_remap_file_range
      0.00            +2.1        2.14 ±  2%  perf-profile.calltrace.cycles-pp.free_unref_page.btrfs_clone.btrfs_clone_files.btrfs_remap_file_range.do_clone_file_range
     15.81            -0.7       15.15        perf-profile.children.cycles-pp.filemap_read_folio
     15.47            -0.6       14.86        perf-profile.children.cycles-pp.btrfs_read_folio
      9.89            -0.5        9.38        perf-profile.children.cycles-pp.memcmp
     11.98            -0.4       11.54        perf-profile.children.cycles-pp.btrfs_do_readpage
      8.74 ±  2%      -0.4        8.30        perf-profile.children.cycles-pp.filemap_add_folio
      9.73 ±  3%      -0.4        9.35        perf-profile.children.cycles-pp.__clear_extent_bit
      5.66 ±  2%      -0.4        5.30        perf-profile.children.cycles-pp.__filemap_add_folio
      2.45 ±  6%      -0.3        2.17 ±  2%  perf-profile.children.cycles-pp.btrfs_fallocate
      2.57 ±  5%      -0.3        2.29 ±  2%  perf-profile.children.cycles-pp.ioctl_preallocate
      2.53 ±  5%      -0.3        2.26 ±  2%  perf-profile.children.cycles-pp.vfs_fallocate
      4.67 ±  2%      -0.3        4.41 ±  3%  perf-profile.children.cycles-pp.__set_extent_bit
      4.83 ±  2%      -0.3        4.58 ±  3%  perf-profile.children.cycles-pp.lock_extent
      5.06 ±  2%      -0.2        4.82 ±  2%  perf-profile.children.cycles-pp.alloc_extent_state
      4.11 ±  2%      -0.2        3.94 ±  2%  perf-profile.children.cycles-pp.kmem_cache_alloc
      1.37 ±  4%      -0.1        1.25 ±  2%  perf-profile.children.cycles-pp.__mod_lruvec_page_state
      0.66 ±  9%      -0.1        0.54 ±  6%  perf-profile.children.cycles-pp.btrfs_reserve_extent
      0.64 ± 10%      -0.1        0.53 ±  6%  perf-profile.children.cycles-pp.find_free_extent
      0.96 ±  4%      -0.1        0.87 ±  6%  perf-profile.children.cycles-pp.__wake_up
      0.62 ±  4%      -0.1        0.54 ±  6%  perf-profile.children.cycles-pp.__cond_resched
      1.20 ±  4%      -0.1        1.12 ±  3%  perf-profile.children.cycles-pp.free_extent_state
      0.99 ±  3%      -0.1        0.92 ±  4%  perf-profile.children.cycles-pp.syscall_return_via_sysret
      0.89 ±  3%      -0.1        0.81 ±  5%  perf-profile.children.cycles-pp.syscall_exit_to_user_mode
      0.69 ±  4%      -0.1        0.64 ±  4%  perf-profile.children.cycles-pp.btrfs_punch_hole
      0.12 ± 10%      -0.0        0.09 ± 10%  perf-profile.children.cycles-pp.__fget_light
      0.02 ±141%      +0.0        0.06 ± 13%  perf-profile.children.cycles-pp.calc_available_free_space
      0.29 ±  8%      +0.1        0.39 ±  6%  perf-profile.children.cycles-pp.__mod_zone_page_state
      0.09 ± 17%      +0.2        0.25 ±  6%  perf-profile.children.cycles-pp.__kmalloc_node
      0.09 ± 15%      +0.2        0.25 ±  4%  perf-profile.children.cycles-pp.kvmalloc_node
      0.08 ± 11%      +0.2        0.24 ±  4%  perf-profile.children.cycles-pp.__kmalloc_large_node
      0.24 ± 13%      +0.2        0.41 ±  4%  perf-profile.children.cycles-pp.__list_add_valid_or_report
      0.32 ± 15%      +0.6        0.91 ±  4%  perf-profile.children.cycles-pp.native_queued_spin_lock_slowpath
     30.03            +0.6       30.64        perf-profile.children.cycles-pp.do_read_cache_folio
      1.10 ±  4%      +0.6        1.72 ±  2%  perf-profile.children.cycles-pp._raw_spin_lock_irqsave
      0.58 ±  6%      +0.9        1.50 ±  5%  perf-profile.children.cycles-pp.__list_del_entry_valid_or_report
     67.36            +1.2       68.60        perf-profile.children.cycles-pp.__x64_sys_ioctl
     21.52 ±  3%      +1.5       23.00        perf-profile.children.cycles-pp.do_clone_file_range
     21.54 ±  3%      +1.5       23.02        perf-profile.children.cycles-pp.ioctl_file_clone
     21.53 ±  3%      +1.5       23.01        perf-profile.children.cycles-pp.vfs_clone_file_range
     21.16 ±  3%      +1.5       22.66        perf-profile.children.cycles-pp.btrfs_clone_files
      0.00            +1.5        1.52 ±  3%  perf-profile.children.cycles-pp.__free_one_page
     64.61            +1.5       66.16        perf-profile.children.cycles-pp.do_vfs_ioctl
     64.16            +1.5       65.71        perf-profile.children.cycles-pp.btrfs_remap_file_range
      2.68 ±  3%      +1.8        4.52 ±  3%  perf-profile.children.cycles-pp.folio_alloc
      0.54 ±  6%      +2.0        2.51 ±  5%  perf-profile.children.cycles-pp.__rmqueue_pcplist
      1.03 ±  3%      +2.0        3.04 ±  5%  perf-profile.children.cycles-pp.rmqueue
      2.16 ±  3%      +2.0        4.19 ±  4%  perf-profile.children.cycles-pp.__alloc_pages
      1.32 ±  2%      +2.1        3.42 ±  4%  perf-profile.children.cycles-pp.get_page_from_freelist
      0.00            +2.1        2.10 ±  2%  perf-profile.children.cycles-pp.free_pcppages_bulk
      2.66 ±  2%      +2.1        4.77        perf-profile.children.cycles-pp.btrfs_clone
      0.03 ±100%      +2.1        2.17 ±  2%  perf-profile.children.cycles-pp.free_unref_page
      0.40 ±  6%      +2.2        2.55 ±  2%  perf-profile.children.cycles-pp.free_unref_page_commit
      0.00            +2.2        2.21 ±  4%  perf-profile.children.cycles-pp.rmqueue_bulk
      9.82            -0.5        9.32        perf-profile.self.cycles-pp.memcmp
      0.84 ±  5%      -0.1        0.76 ±  6%  perf-profile.self.cycles-pp.syscall_exit_to_user_mode
      1.13 ±  4%      -0.1        1.05 ±  2%  perf-profile.self.cycles-pp.free_extent_state
      0.99 ±  3%      -0.1        0.92 ±  4%  perf-profile.self.cycles-pp.syscall_return_via_sysret
      0.22 ±  8%      -0.1        0.16 ± 13%  perf-profile.self.cycles-pp.find_free_extent
      0.38 ±  4%      -0.1        0.32 ±  8%  perf-profile.self.cycles-pp.__cond_resched
      0.12 ± 10%      -0.0        0.08 ± 11%  perf-profile.self.cycles-pp.__fget_light
      0.06 ±  7%      -0.0        0.04 ± 45%  perf-profile.self.cycles-pp.__x64_sys_ioctl
      0.07 ± 15%      +0.0        0.10 ±  9%  perf-profile.self.cycles-pp.folio_alloc
      0.28 ± 10%      +0.1        0.36 ±  7%  perf-profile.self.cycles-pp.get_page_from_freelist
      0.26 ±  8%      +0.1        0.36 ±  4%  perf-profile.self.cycles-pp.__mod_zone_page_state
      0.22 ± 14%      +0.2        0.38 ±  5%  perf-profile.self.cycles-pp.__list_add_valid_or_report
      0.00            +0.2        0.24 ±  6%  perf-profile.self.cycles-pp.free_pcppages_bulk
      0.32 ± 15%      +0.6        0.91 ±  4%  perf-profile.self.cycles-pp.native_queued_spin_lock_slowpath
      0.00            +0.6        0.62 ± 10%  perf-profile.self.cycles-pp.rmqueue_bulk
      0.55 ±  6%      +0.9        1.46 ±  5%  perf-profile.self.cycles-pp.__list_del_entry_valid_or_report
      0.00            +1.3        1.32 ±  4%  perf-profile.self.cycles-pp.__free_one_page





Disclaimer:
Results have been estimated based on internal Intel analysis and are provided
for informational purposes only. Any difference in system hardware or software
design or configuration may affect actual performance.


-- 
0-DAY CI Kernel Test Service
https://github.com/intel/lkp-tests/wiki



^ permalink raw reply	[flat|nested] 3+ messages in thread

* Re: [linus:master] [mm, pcp]  6ccdcb6d3a: stress-ng.judy.ops_per_sec -4.7% regression
  2023-11-23  5:03 [linus:master] [mm, pcp] 6ccdcb6d3a: stress-ng.judy.ops_per_sec -4.7% regression kernel test robot
@ 2023-11-23  5:40 ` Huang, Ying
  2023-11-24  6:53   ` Oliver Sang
  0 siblings, 1 reply; 3+ messages in thread
From: Huang, Ying @ 2023-11-23  5:40 UTC (permalink / raw)
  To: kernel test robot
  Cc: oe-lkp, lkp, linux-kernel, Andrew Morton, Mel Gorman,
	Vlastimil Babka, David Hildenbrand, Johannes Weiner, Dave Hansen,
	Michal Hocko, Pavel Tatashin, Matthew Wilcox, Christoph Lameter,
	Arjan van de Ven, Sudeep Holla, linux-mm, feng.tang, fengwei.yin

Hi,

Thanks for test!

kernel test robot <oliver.sang@intel.com> writes:

> Hello,
>
> kernel test robot noticed a -4.7% regression of stress-ng.judy.ops_per_sec on:
>
>
> commit: 6ccdcb6d3a741c4e005ca6ffd4a62ddf8b5bead3 ("mm, pcp: reduce detecting time of consecutive high order page freeing")
> https://git.kernel.org/cgit/linux/kernel/git/torvalds/linux.git master
>
> testcase: stress-ng
> test machine: 224 threads 2 sockets Intel(R) Xeon(R) Platinum 8480CTDX (Sapphire Rapids) with 512G memory
> parameters:
>
> 	nr_threads: 100%
> 	testtime: 60s
> 	class: cpu-cache
> 	test: judy
> 	disk: 1SSD
> 	cpufreq_governor: performance
>
>
> In addition to that, the commit also has significant impact on the following tests:
>
> +------------------+-------------------------------------------------------------------------------------------------+
> | testcase: change | lmbench3: lmbench3.TCP.socket.bandwidth.10MB.MB/sec 23.7% improvement                           |
> | test machine     | 48 threads 2 sockets Intel(R) Xeon(R) CPU E5-2697 v2 @ 2.70GHz (Ivy Bridge-EP) with 112G memory |
> | test parameters  | cpufreq_governor=performance                                                                    |
> |                  | mode=development                                                                                |
> |                  | nr_threads=100%                                                                                 |
> |                  | test=TCP                                                                                        |
> |                  | test_memory_size=50%                                                                            |
> +------------------+-------------------------------------------------------------------------------------------------+
> | testcase: change | stress-ng: stress-ng.file-ioctl.ops_per_sec -6.6% regression                                    |
> | test machine     | 36 threads 1 sockets Intel(R) Core(TM) i9-9980XE CPU @ 3.00GHz (Skylake) with 32G memory        |
> | test parameters  | class=filesystem                                                                                |
> |                  | cpufreq_governor=performance                                                                    |
> |                  | disk=1SSD                                                                                       |
> |                  | fs=btrfs                                                                                        |
> |                  | nr_threads=10%                                                                                  |
> |                  | test=file-ioctl                                                                                 |
> |                  | testtime=60s                                                                                    |
> +------------------+-------------------------------------------------------------------------------------------------+

It's expected that this commit will benefit some workload (mainly
network, inter-process communication related) and hurt some workload.
But the  whole series should have no much regression.  Can you try the
whole series for the regression test cases?  The series start from
commit ca71fe1ad922 ("mm, pcp: avoid to drain PCP when process exit") to
commit 6ccdcb6d3a74 ("mm, pcp: reduce detecting time of consecutive high
order page freeing").

--
Best Regards,
Huang, Ying



^ permalink raw reply	[flat|nested] 3+ messages in thread

* Re: [linus:master] [mm, pcp]  6ccdcb6d3a: stress-ng.judy.ops_per_sec -4.7% regression
  2023-11-23  5:40 ` Huang, Ying
@ 2023-11-24  6:53   ` Oliver Sang
  0 siblings, 0 replies; 3+ messages in thread
From: Oliver Sang @ 2023-11-24  6:53 UTC (permalink / raw)
  To: Huang, Ying
  Cc: oe-lkp, lkp, linux-kernel, Andrew Morton, Mel Gorman,
	Vlastimil Babka, David Hildenbrand, Johannes Weiner, Dave Hansen,
	Michal Hocko, Pavel Tatashin, Matthew Wilcox, Christoph Lameter,
	Arjan van de Ven, Sudeep Holla, linux-mm, feng.tang, fengwei.yin,
	oliver.sang

[-- Attachment #1: Type: text/plain, Size: 6263 bytes --]

hi, Huang Ying,

On Thu, Nov 23, 2023 at 01:40:02PM +0800, Huang, Ying wrote:
> Hi,
> 
> Thanks for test!
> 
> kernel test robot <oliver.sang@intel.com> writes:
> 
> > Hello,
> >
> > kernel test robot noticed a -4.7% regression of stress-ng.judy.ops_per_sec on:
> >
> >
> > commit: 6ccdcb6d3a741c4e005ca6ffd4a62ddf8b5bead3 ("mm, pcp: reduce detecting time of consecutive high order page freeing")
> > https://git.kernel.org/cgit/linux/kernel/git/torvalds/linux.git master
> >
> > testcase: stress-ng
> > test machine: 224 threads 2 sockets Intel(R) Xeon(R) Platinum 8480CTDX (Sapphire Rapids) with 512G memory
> > parameters:
> >
> > 	nr_threads: 100%
> > 	testtime: 60s
> > 	class: cpu-cache
> > 	test: judy
> > 	disk: 1SSD
> > 	cpufreq_governor: performance
> >
> >
> > In addition to that, the commit also has significant impact on the following tests:
> >
> > +------------------+-------------------------------------------------------------------------------------------------+
> > | testcase: change | lmbench3: lmbench3.TCP.socket.bandwidth.10MB.MB/sec 23.7% improvement                           |
> > | test machine     | 48 threads 2 sockets Intel(R) Xeon(R) CPU E5-2697 v2 @ 2.70GHz (Ivy Bridge-EP) with 112G memory |
> > | test parameters  | cpufreq_governor=performance                                                                    |
> > |                  | mode=development                                                                                |
> > |                  | nr_threads=100%                                                                                 |
> > |                  | test=TCP                                                                                        |
> > |                  | test_memory_size=50%                                                                            |
> > +------------------+-------------------------------------------------------------------------------------------------+
> > | testcase: change | stress-ng: stress-ng.file-ioctl.ops_per_sec -6.6% regression                                    |
> > | test machine     | 36 threads 1 sockets Intel(R) Core(TM) i9-9980XE CPU @ 3.00GHz (Skylake) with 32G memory        |
> > | test parameters  | class=filesystem                                                                                |
> > |                  | cpufreq_governor=performance                                                                    |
> > |                  | disk=1SSD                                                                                       |
> > |                  | fs=btrfs                                                                                        |
> > |                  | nr_threads=10%                                                                                  |
> > |                  | test=file-ioctl                                                                                 |
> > |                  | testtime=60s                                                                                    |
> > +------------------+-------------------------------------------------------------------------------------------------+
> 
> It's expected that this commit will benefit some workload (mainly
> network, inter-process communication related) and hurt some workload.
> But the  whole series should have no much regression.  Can you try the
> whole series for the regression test cases?  The series start from
> commit ca71fe1ad922 ("mm, pcp: avoid to drain PCP when process exit") to
> commit 6ccdcb6d3a74 ("mm, pcp: reduce detecting time of consecutive high
> order page freeing").

since:
* 6ccdcb6d3a741 mm, pcp: reduce detecting time of consecutive high order page freeing
* 57c0419c5f0ea mm, pcp: decrease PCP high if free pages < high watermark
* 51a755c56dc05 mm: tune PCP high automatically
* 90b41691b9881 mm: add framework for PCP high auto-tuning
* c0a242394cb98 mm, page_alloc: scale the number of pages that are batch allocated
* 52166607ecc98 mm: restrict the pcp batch scale factor to avoid too long latency
* 362d37a106dd3 mm, pcp: reduce lock contention for draining high-order pages
* 94a3bfe4073cd cacheinfo: calculate size of per-CPU data cache slice
* ca71fe1ad9221 mm, pcp: avoid to drain PCP when process exit
* 1f4f7f0f8845d mm/oom_killer: simplify OOM killer info dump helper

I tested 1f4f7f0f8845d vs 6ccdcb6d3a741.

for stress-ng.judy.ops_per_sec, there is a smaller regression (-2.0%):
(full comparison is attached as ncompare-judy)

=========================================================================================
class/compiler/cpufreq_governor/disk/kconfig/nr_threads/rootfs/tbox_group/test/testcase/testtime:
  cpu-cache/gcc-12/performance/1SSD/x86_64-rhel-8.3/100%/debian-11.1-x86_64-20220510.cgz/lkp-spr-2sp4/judy/stress-ng/60s

1f4f7f0f8845dbac 6ccdcb6d3a741c4e005ca6ffd4a
---------------- ---------------------------
         %stddev     %change         %stddev
             \          |                \
   6925490            -0.9%    6862477        stress-ng.judy.Judy_delete_operations_per_sec
  22515488            -0.4%   22420191        stress-ng.judy.Judy_find_operations_per_sec
   9036524            -3.9%    8685310 ±  3%  stress-ng.judy.Judy_insert_operations_per_sec
    171299            -2.0%     167905        stress-ng.judy.ops
      2853            -2.0%       2796        stress-ng.judy.ops_per_sec


for stress-ng.file-ioctl.ops_per_sec, there is a similar regression (-6.9%):
(full comparison is attached as ncompare-file-ioctl)

=========================================================================================
class/compiler/cpufreq_governor/disk/fs/kconfig/nr_threads/rootfs/tbox_group/test/testcase/testtime:
  filesystem/gcc-12/performance/1SSD/btrfs/x86_64-rhel-8.3/10%/debian-11.1-x86_64-20220510.cgz/lkp-skl-d08/file-ioctl/stress-ng/60s

1f4f7f0f8845dbac 6ccdcb6d3a741c4e005ca6ffd4a
---------------- ---------------------------
         %stddev     %change         %stddev
             \          |                \
    340971            -6.9%     317411        stress-ng.file-ioctl.ops
      5682            -6.9%       5290        stress-ng.file-ioctl.ops_per_sec

> 
> --
> Best Regards,
> Huang, Ying
> 

[-- Attachment #2: ncompare-judy --]
[-- Type: text/plain, Size: 182481 bytes --]


xsang@inn:~$ lkp ncompare -s commit=1f4f7f0f8845d/tbox_group=lkp-spr-2sp4/testcase=stress-ng -o commit=6ccdcb6d3a741  -a | grep -v interrupts | grep -v softirq
=========================================================================================
class/compiler/cpufreq_governor/disk/kconfig/nr_threads/rootfs/tbox_group/test/testcase/testtime:
  cpu-cache/gcc-12/performance/1SSD/x86_64-rhel-8.3/100%/debian-11.1-x86_64-20220510.cgz/lkp-spr-2sp4/judy/stress-ng/60s

commit:
  1f4f7f0f8845dbac40289cc3d50b81314c5a12b8
  6ccdcb6d3a741c4e005ca6ffd4a62ddf8b5bead3

1f4f7f0f8845dbac 6ccdcb6d3a741c4e005ca6ffd4a
---------------- ---------------------------
       fail:runs  %reproduction    fail:runs
           |             |             |
         10:10           0%          10:10    stress-ng.judy.Judy_delete_operations_per_sec.pass
         10:10           0%          10:10    stress-ng.judy.Judy_find_operations_per_sec.pass
         10:10           0%          10:10    stress-ng.judy.Judy_insert_operations_per_sec.pass
         10:10           0%          10:10    stress-ng.judy.pass
         %stddev     %change         %stddev
             \          |                \
    151.27 ± 56%      +6.6%     161.22 ± 52%  dmesg.timestamp:last
    124.42 ± 11%      +1.1%     125.79 ±  9%  kmsg.timestamp:last
 1.107e+09 ± 29%     +67.3%  1.852e+09 ± 75%  cpuidle..time
   1084691 ± 30%     +71.0%    1855097 ± 77%  cpuidle..usage
    125.81 ± 10%      +3.3%     129.98 ± 11%  uptime.boot
     13667 ± 21%      +6.8%      14591 ± 23%  uptime.idle
     53.76 ±  5%      +0.6%      54.06 ±  4%  boot-time.boot
     30.41            -0.2%      30.36        boot-time.dhcp
     11000 ±  5%      +0.6%      11066 ±  4%  boot-time.idle
     10.11 ± 18%     +35.2%      13.67 ± 51%  iostat.cpu.idle
      4.96 ±  2%     +25.9%       6.25 ± 17%  iostat.cpu.system
     84.92 ±  2%      -5.7%      80.08 ±  8%  iostat.cpu.user
      7.37 ± 27%      +3.8       11.16 ± 66%  mpstat.cpu.all.idle%
      1.57 ±  3%      -0.1        1.51 ±  5%  mpstat.cpu.all.irq%
      0.14 ±  3%      +0.4        0.49 ± 33%  mpstat.cpu.all.soft%
      3.38 ±  2%      +1.0        4.41 ± 22%  mpstat.cpu.all.sys%
     87.54 ±  2%      -5.1       82.43 ±  8%  mpstat.cpu.all.usr%
     17731 ± 33%      +7.7%      19098 ± 26%  perf-c2c.DRAM.local
    227.40 ± 39%     +19.3%     271.30 ± 26%  perf-c2c.DRAM.remote
    396.50 ± 33%     +21.0%     479.80 ± 29%  perf-c2c.HITM.local
     54.60 ± 51%     +12.1%      61.20 ± 35%  perf-c2c.HITM.remote
    451.10 ± 35%     +19.9%     541.00 ± 29%  perf-c2c.HITM.total
      0.00          -100.0%       0.00        numa-numastat.node0.interleave_hit
  42179114           -12.9%   36727757 ±  8%  numa-numastat.node0.local_node
  42322224           -12.9%   36881294 ±  8%  numa-numastat.node0.numa_hit
    141985 ± 51%      +8.1%     153548 ± 39%  numa-numastat.node0.other_node
      0.00          -100.0%       0.00        numa-numastat.node1.interleave_hit
  41820129           -12.8%   36449421 ±  9%  numa-numastat.node1.local_node
  41909301           -12.8%   36528507 ±  9%  numa-numastat.node1.numa_hit
     89842 ± 81%     -12.8%      78315 ± 77%  numa-numastat.node1.other_node
      9.94 ± 20%     +37.6%      13.68 ± 50%  vmstat.cpu.id
      4.91 ±  2%     +27.3%       6.25 ± 17%  vmstat.cpu.sy
     85.15 ±  2%      -6.0%      80.06 ±  8%  vmstat.cpu.us
      0.02 ±  2%      -4.2%       0.01 ±  8%  vmstat.io.bi
      4.00            +0.0%       4.00        vmstat.memory.buff
   3275886            +0.7%    3300023        vmstat.memory.cache
 5.199e+08            -0.1%  5.196e+08        vmstat.memory.free
    202.27 ±  2%      -3.2%     195.88 ±  7%  vmstat.procs.r
      9290            -4.0%       8915 ±  6%  vmstat.system.cs
    366738            -2.3%     358141 ±  2%  vmstat.system.in
     62.20            -0.0%      62.20        time.elapsed_time
     62.20            -0.0%      62.20        time.elapsed_time.max
    258360            -2.6%     251676        time.involuntary_context_switches
      0.30 ±152%    +166.7%       0.80 ±108%  time.major_page_faults
     10893 ±  7%     +23.0%      13397 ± 29%  time.maximum_resident_set_size
  82141065           -13.1%   71375734 ±  9%  time.minor_page_faults
      4096            +0.0%       4096        time.page_size
     21055            -1.2%      20804        time.percent_of_cpu_this_job_got
    480.34           +35.5%     650.87 ± 21%  time.system_time
     12618            -2.6%      12291        time.user_time
    563.40 ±  2%      -0.7%     559.50 ±  3%  time.voluntary_context_switches
      2608 ±  2%      -3.3%       2521 ±  7%  turbostat.Avg_MHz
     92.38 ±  2%      -3.8       88.53 ±  8%  turbostat.Busy%
      2839            +0.9%       2864        turbostat.Bzy_MHz
      1414 ± 18%     +32.0%       1866 ± 35%  turbostat.C1
    527910 ± 28%     +64.1%     866385 ± 73%  turbostat.C1E
      3.33 ± 26%      +1.7        5.05 ± 64%  turbostat.C1E%
    612375 ± 28%     +67.9%    1028102 ± 76%  turbostat.C6
      4.26 ± 24%      +2.1        6.39 ± 62%  turbostat.C6%
      7.50 ± 25%     +51.2%      11.34 ± 63%  turbostat.CPU%c1
      0.12 ± 17%      +1.6%       0.13 ± 21%  turbostat.CPU%c6
     59.70 ±  3%      -2.0%      58.50 ±  4%  turbostat.CoreTmp
      1.25            -2.6%       1.22        turbostat.IPC
  24279278            +2.3%   24843604 ±  5%  turbostat.IRQ
    180.02 ±  2%     -11.0      169.01 ±  8%  turbostat.PKG_%
      2789 ± 31%     +33.1%       3712 ± 47%  turbostat.POLL
     60.00 ±  3%      -2.3%      58.60 ±  4%  turbostat.PkgTmp
    672.54            -1.8%     660.52 ±  3%  turbostat.PkgWatt
     37.66            -2.4%      36.77 ±  2%  turbostat.RAMWatt
      1989            -0.0%       1988        turbostat.TSC_MHz
   6925490            -0.9%    6862477        stress-ng.judy.Judy_delete_operations_per_sec
  22515488            -0.4%   22420191        stress-ng.judy.Judy_find_operations_per_sec
   9036524            -3.9%    8685310 ±  3%  stress-ng.judy.Judy_insert_operations_per_sec
    171299            -2.0%     167905        stress-ng.judy.ops
      2853            -2.0%       2796        stress-ng.judy.ops_per_sec
     62.20            -0.0%      62.20        stress-ng.time.elapsed_time
     62.20            -0.0%      62.20        stress-ng.time.elapsed_time.max
    258360            -2.6%     251676        stress-ng.time.involuntary_context_switches
      0.30 ±152%    +166.7%       0.80 ±108%  stress-ng.time.major_page_faults
     10893 ±  7%     +23.0%      13397 ± 29%  stress-ng.time.maximum_resident_set_size
  82141065           -13.1%   71375734 ±  9%  stress-ng.time.minor_page_faults
      4096            +0.0%       4096        stress-ng.time.page_size
     21055            -1.2%      20804        stress-ng.time.percent_of_cpu_this_job_got
    480.34           +35.5%     650.87 ± 21%  stress-ng.time.system_time
     12618            -2.6%      12291        stress-ng.time.user_time
    563.40 ±  2%      -0.7%     559.50 ±  3%  stress-ng.time.voluntary_context_switches
     46768 ± 18%     +73.1%      80961 ± 56%  meminfo.Active
     46689 ± 18%     +73.2%      80881 ± 56%  meminfo.Active(anon)
     78.40 ±  6%      +2.0%      79.98        meminfo.Active(file)
    127245 ± 16%      +5.8%     134611 ± 14%  meminfo.AnonHugePages
   1451846            -2.1%    1422053 ±  6%  meminfo.AnonPages
      4.00            +0.0%       4.00        meminfo.Buffers
   3121206            +0.7%    3144477        meminfo.Cached
 2.637e+08            +0.0%  2.637e+08        meminfo.CommitLimit
   4368713            -2.2%    4271782 ±  5%  meminfo.Committed_AS
 5.245e+08            -0.1%   5.24e+08        meminfo.DirectMap1G
  11274240 ± 12%      +4.5%   11782144 ± 12%  meminfo.DirectMap2M
    198052 ±  5%      +8.3%     214436 ±  5%  meminfo.DirectMap4k
      2048            +0.0%       2048        meminfo.Hugepagesize
   1800847            -2.1%    1763898 ±  5%  meminfo.Inactive
   1800650            -2.1%    1763721 ±  5%  meminfo.Inactive(anon)
    197.33 ±  2%     -10.5%     176.70 ±  8%  meminfo.Inactive(file)
    148140            +0.2%     148422        meminfo.KReclaimable
     41102            -0.2%      41026        meminfo.KernelStack
    517263 ±  2%      +0.3%     518808 ±  2%  meminfo.Mapped
 5.178e+08            -0.1%  5.175e+08        meminfo.MemAvailable
 5.199e+08            -0.1%  5.196e+08        meminfo.MemFree
 5.274e+08            +0.0%  5.274e+08        meminfo.MemTotal
   7517161            +4.3%    7841550 ±  4%  meminfo.Memused
     10.95 ±  2%      -4.0%      10.51 ±  8%  meminfo.Mlocked
     25594 ±  3%      -1.9%      25112 ±  7%  meminfo.PageTables
    109277            +0.3%     109597        meminfo.Percpu
    148140            +0.2%     148422        meminfo.SReclaimable
    462907            +0.2%     463832        meminfo.SUnreclaim
    405085 ±  2%      +5.8%     428379 ±  8%  meminfo.Shmem
    611047            +0.2%     612255        meminfo.Slab
   2715849            -0.0%    2715845        meminfo.Unevictable
 1.374e+13            +0.0%  1.374e+13        meminfo.VmallocTotal
    292661            +0.0%     292707        meminfo.VmallocUsed
   7721333            +6.6%    8231493 ±  4%  meminfo.max_used_kB
      6639 ± 70%     +25.7%       8344 ±160%  numa-meminfo.node0.Active
      6609 ± 70%     +25.5%       8296 ±161%  numa-meminfo.node0.Active(anon)
     30.40 ±123%     +57.9%      48.00 ± 81%  numa-meminfo.node0.Active(file)
     65249 ± 56%     -24.1%      49545 ± 74%  numa-meminfo.node0.AnonHugePages
    698546 ± 10%     -10.1%     627683 ± 11%  numa-meminfo.node0.AnonPages
    783098 ±  9%      -4.6%     746765 ± 10%  numa-meminfo.node0.AnonPages.max
   1849056 ± 60%     -27.0%    1350657 ± 82%  numa-meminfo.node0.FilePages
    708083 ± 10%      -8.9%     645187 ± 13%  numa-meminfo.node0.Inactive
    708005 ± 10%      -8.9%     645081 ± 13%  numa-meminfo.node0.Inactive(anon)
     77.48 ±122%     +37.3%     106.35 ± 82%  numa-meminfo.node0.Inactive(file)
     82324 ± 32%      -8.0%      75702 ± 27%  numa-meminfo.node0.KReclaimable
     20512 ±  7%      +1.4%      20790 ±  7%  numa-meminfo.node0.KernelStack
    175959 ±  9%      -0.1%     175730 ±  7%  numa-meminfo.node0.Mapped
 2.594e+08            +0.2%  2.599e+08        numa-meminfo.node0.MemFree
 2.635e+08            +0.0%  2.635e+08        numa-meminfo.node0.MemTotal
   4076783 ± 28%     -10.2%    3660809 ± 31%  numa-meminfo.node0.MemUsed
      6.66 ± 81%      -2.3%       6.51 ± 81%  numa-meminfo.node0.Mlocked
      9901 ± 61%      +4.5%      10346 ± 49%  numa-meminfo.node0.PageTables
     82324 ± 32%      -8.0%      75702 ± 27%  numa-meminfo.node0.SReclaimable
    248888 ±  3%      +0.5%     250078 ±  4%  numa-meminfo.node0.SUnreclaim
     20568 ± 40%     +41.9%      29194 ± 98%  numa-meminfo.node0.Shmem
    331213 ±  9%      -1.6%     325780 ±  7%  numa-meminfo.node0.Slab
   1828379 ± 61%     -27.7%    1321308 ± 85%  numa-meminfo.node0.Unevictable
     40129 ± 22%     +83.1%      73478 ± 53%  numa-meminfo.node1.Active
     40081 ± 22%     +83.2%      73447 ± 53%  numa-meminfo.node1.Active(anon)
     48.00 ± 81%     -33.4%      31.98 ±122%  numa-meminfo.node1.Active(file)
     62048 ± 50%     +37.2%      85152 ± 49%  numa-meminfo.node1.AnonHugePages
    756214 ±  9%      +3.8%     784679 ± 12%  numa-meminfo.node1.AnonPages
    924235 ±  6%      +6.7%     986345 ±  7%  numa-meminfo.node1.AnonPages.max
   1271716 ± 88%     +41.1%    1793782 ± 63%  numa-meminfo.node1.FilePages
   1095798 ±  5%      +1.1%    1108364 ± 10%  numa-meminfo.node1.Inactive
   1095678 ±  5%      +1.2%    1108294 ± 10%  numa-meminfo.node1.Inactive(anon)
    119.85 ± 77%     -41.3%      70.35 ±118%  numa-meminfo.node1.Inactive(file)
     65822 ± 40%     +10.5%      72715 ± 29%  numa-meminfo.node1.KReclaimable
     20590 ±  7%      -1.7%      20233 ±  8%  numa-meminfo.node1.KernelStack
    341090 ±  4%      +0.0%     341144 ±  4%  numa-meminfo.node1.Mapped
 2.605e+08            -0.3%  2.597e+08        numa-meminfo.node1.MemFree
 2.639e+08            +0.0%  2.639e+08        numa-meminfo.node1.MemTotal
   3445580 ± 33%     +21.0%    4169560 ± 27%  numa-meminfo.node1.MemUsed
      4.33 ±122%      -8.0%       3.98 ±123%  numa-meminfo.node1.Mlocked
     15703 ± 37%      -6.1%      14739 ± 38%  numa-meminfo.node1.PageTables
     65822 ± 40%     +10.5%      72715 ± 29%  numa-meminfo.node1.SReclaimable
    214196 ±  4%      -0.2%     213692 ±  5%  numa-meminfo.node1.SUnreclaim
    384078 ±  3%      +3.9%     399143 ±  7%  numa-meminfo.node1.Shmem
    280018 ± 11%      +2.3%     286407 ±  9%  numa-meminfo.node1.Slab
    887469 ±126%     +57.1%    1394536 ± 80%  numa-meminfo.node1.Unevictable
      1655 ± 70%     +28.3%       2123 ±163%  numa-vmstat.node0.nr_active_anon
      7.60 ±123%     +57.9%      12.00 ± 81%  numa-vmstat.node0.nr_active_file
    173571 ± 10%      -9.6%     156867 ± 11%  numa-vmstat.node0.nr_anon_pages
     31.86 ± 57%     -24.0%      24.20 ± 74%  numa-vmstat.node0.nr_anon_transparent_hugepages
    462268 ± 60%     -27.0%     337669 ± 82%  numa-vmstat.node0.nr_file_pages
  64861249            +0.2%   64962956        numa-vmstat.node0.nr_free_pages
    176031 ±  9%      -8.5%     161129 ± 12%  numa-vmstat.node0.nr_inactive_anon
     19.37 ±122%     +37.3%      26.59 ± 82%  numa-vmstat.node0.nr_inactive_file
      0.00 ±300%      +1.6%       0.00 ±300%  numa-vmstat.node0.nr_isolated_anon
     20512 ±  7%      +1.4%      20790 ±  7%  numa-vmstat.node0.nr_kernel_stack
     43945 ±  9%      -0.2%      43864 ±  7%  numa-vmstat.node0.nr_mapped
      1.66 ± 81%      -2.5%       1.62 ± 81%  numa-vmstat.node0.nr_mlock
      2475 ± 61%      +4.5%       2585 ± 49%  numa-vmstat.node0.nr_page_table_pages
      5146 ± 40%     +41.9%       7303 ± 98%  numa-vmstat.node0.nr_shmem
     20581 ± 32%      -8.0%      18925 ± 27%  numa-vmstat.node0.nr_slab_reclaimable
     62224 ±  3%      +0.5%      62514 ±  4%  numa-vmstat.node0.nr_slab_unreclaimable
    457094 ± 61%     -27.7%     330327 ± 85%  numa-vmstat.node0.nr_unevictable
      1655 ± 70%     +28.3%       2123 ±163%  numa-vmstat.node0.nr_zone_active_anon
      7.60 ±123%     +57.9%      12.00 ± 81%  numa-vmstat.node0.nr_zone_active_file
    176029 ±  9%      -8.5%     161125 ± 12%  numa-vmstat.node0.nr_zone_inactive_anon
     19.37 ±122%     +37.3%      26.59 ± 82%  numa-vmstat.node0.nr_zone_inactive_file
    457094 ± 61%     -27.7%     330327 ± 85%  numa-vmstat.node0.nr_zone_unevictable
  42322270           -12.9%   36881741 ±  8%  numa-vmstat.node0.numa_hit
      0.00          -100.0%       0.00        numa-vmstat.node0.numa_interleave
  42179160           -12.9%   36728210 ±  8%  numa-vmstat.node0.numa_local
    141985 ± 51%      +8.1%     153548 ± 39%  numa-vmstat.node0.numa_other
     10088 ± 22%     +82.5%      18410 ± 53%  numa-vmstat.node1.nr_active_anon
     12.00 ± 81%     -33.4%       7.99 ±122%  numa-vmstat.node1.nr_active_file
    187803 ±  9%      +4.4%     196071 ± 12%  numa-vmstat.node1.nr_anon_pages
     30.30 ± 50%     +37.2%      41.58 ± 49%  numa-vmstat.node1.nr_anon_transparent_hugepages
    317887 ± 88%     +41.1%     448582 ± 63%  numa-vmstat.node1.nr_file_pages
  65116967            -0.3%   64935325        numa-vmstat.node1.nr_free_pages
    272583 ±  6%      +1.6%     276951 ± 10%  numa-vmstat.node1.nr_inactive_anon
     29.96 ± 77%     -41.3%      17.59 ±118%  numa-vmstat.node1.nr_inactive_file
      0.00       +5.9e+101%       0.59 ±300%  numa-vmstat.node1.nr_isolated_anon
     20596 ±  7%      -1.8%      20229 ±  8%  numa-vmstat.node1.nr_kernel_stack
     85156 ±  4%      +0.3%      85436 ±  4%  numa-vmstat.node1.nr_mapped
      1.08 ±122%      -8.0%       1.00 ±123%  numa-vmstat.node1.nr_mlock
      3929 ± 37%      -6.4%       3678 ± 38%  numa-vmstat.node1.nr_page_table_pages
     95977 ±  3%      +4.1%      99922 ±  7%  numa-vmstat.node1.nr_shmem
     16455 ± 40%     +10.5%      18180 ± 29%  numa-vmstat.node1.nr_slab_reclaimable
     53550 ±  4%      -0.2%      53423 ±  5%  numa-vmstat.node1.nr_slab_unreclaimable
    221867 ±126%     +57.1%     348634 ± 80%  numa-vmstat.node1.nr_unevictable
     10088 ± 22%     +82.5%      18410 ± 53%  numa-vmstat.node1.nr_zone_active_anon
     12.00 ± 81%     -33.4%       7.99 ±122%  numa-vmstat.node1.nr_zone_active_file
    272580 ±  6%      +1.6%     276963 ± 10%  numa-vmstat.node1.nr_zone_inactive_anon
     29.96 ± 77%     -41.3%      17.59 ±118%  numa-vmstat.node1.nr_zone_inactive_file
    221867 ±126%     +57.1%     348634 ± 80%  numa-vmstat.node1.nr_zone_unevictable
  41909012           -12.8%   36528578 ±  9%  numa-vmstat.node1.numa_hit
      0.00          -100.0%       0.00        numa-vmstat.node1.numa_interleave
  41819840           -12.8%   36449493 ±  9%  numa-vmstat.node1.numa_local
     89842 ± 81%     -12.8%      78314 ± 77%  numa-vmstat.node1.numa_other
     59.40 ±  9%     +14.0%      67.70 ±  9%  proc-vmstat.direct_map_level2_splits
      5.00 ± 28%     +12.0%       5.60 ± 27%  proc-vmstat.direct_map_level3_splits
     11737 ± 19%     +72.4%      20235 ± 56%  proc-vmstat.nr_active_anon
     19.60 ±  6%      +2.0%      19.99        proc-vmstat.nr_active_file
    362955            -2.2%     354917 ±  6%  proc-vmstat.nr_anon_pages
     62.13 ± 16%      +5.8%      65.73 ± 14%  proc-vmstat.nr_anon_transparent_hugepages
  12924190            -0.1%   12916030        proc-vmstat.nr_dirty_background_threshold
  25879981            -0.1%   25863641        proc-vmstat.nr_dirty_threshold
    780355            +0.7%     786173        proc-vmstat.nr_file_pages
   1.3e+08            -0.1%  1.299e+08        proc-vmstat.nr_free_pages
    450188            -2.2%     440379 ±  5%  proc-vmstat.nr_inactive_anon
     49.33 ±  2%     -10.5%      44.18 ±  8%  proc-vmstat.nr_inactive_file
      0.57 ±300%      -2.9%       0.56 ±299%  proc-vmstat.nr_isolated_anon
     41099            -0.2%      41022        proc-vmstat.nr_kernel_stack
    129179 ±  2%      +0.3%     129620 ±  2%  proc-vmstat.nr_mapped
      2.73 ±  2%      -4.2%       2.62 ±  8%  proc-vmstat.nr_mlock
      6394 ±  3%      -1.9%       6270 ±  7%  proc-vmstat.nr_page_table_pages
    101324 ±  2%      +5.7%     107147 ±  8%  proc-vmstat.nr_shmem
     37036            +0.2%      37105        proc-vmstat.nr_slab_reclaimable
    115732            +0.2%     115946        proc-vmstat.nr_slab_unreclaimable
    678962            -0.0%     678961        proc-vmstat.nr_unevictable
     11737 ± 19%     +72.4%      20235 ± 56%  proc-vmstat.nr_zone_active_anon
     19.60 ±  6%      +2.0%      19.99        proc-vmstat.nr_zone_active_file
    450188            -2.2%     440380 ±  5%  proc-vmstat.nr_zone_inactive_anon
     49.33 ±  2%     -10.5%      44.18 ±  8%  proc-vmstat.nr_zone_inactive_file
    678962            -0.0%     678961        proc-vmstat.nr_zone_unevictable
     59062 ±  5%      -2.1%      57815 ± 10%  proc-vmstat.numa_hint_faults
     55966 ±  7%      -3.9%      53755 ±  9%  proc-vmstat.numa_hint_faults_local
  84235288           -12.8%   73412158 ±  8%  proc-vmstat.numa_hit
     36.70 ±  4%      -1.9%      36.00        proc-vmstat.numa_huge_pte_updates
      0.00          -100.0%       0.00        proc-vmstat.numa_interleave
  84003235           -12.9%   73179535 ±  8%  proc-vmstat.numa_local
    231828            +0.0%     231864        proc-vmstat.numa_other
      4365 ± 37%      -0.1%       4362 ± 61%  proc-vmstat.numa_pages_migrated
    191149            +1.4%     193820 ±  4%  proc-vmstat.numa_pte_updates
     72087 ± 36%      -8.6%      65920 ± 67%  proc-vmstat.pgactivate
  85864265           -12.5%   75117942 ±  8%  proc-vmstat.pgalloc_normal
  82879163           -13.0%   72119383 ±  9%  proc-vmstat.pgfault
  85427394           -12.6%   74693640 ±  8%  proc-vmstat.pgfree
      4365 ± 37%      -0.1%       4362 ± 61%  proc-vmstat.pgmigrate_success
      0.00          -100.0%       0.00        proc-vmstat.pgpgin
     27273 ±  3%      +4.5%      28501 ± 11%  proc-vmstat.pgreuse
     46.70 ±  9%     +11.3%      52.00 ± 14%  proc-vmstat.thp_collapse_alloc
      0.20 ±200%    +200.0%       0.60 ± 81%  proc-vmstat.thp_deferred_split_page
     48.40            +0.6%      48.70        proc-vmstat.thp_fault_alloc
      4.80 ± 75%     -37.5%       3.00 ±100%  proc-vmstat.thp_migration_success
      0.20 ±200%    +200.0%       0.60 ± 81%  proc-vmstat.thp_split_pmd
      0.00          -100.0%       0.00        proc-vmstat.thp_zero_page_alloc
     35.00            +0.0%      35.00        proc-vmstat.unevictable_pgs_culled
      3.00            +0.0%       3.00        proc-vmstat.unevictable_pgs_mlocked
      3.00            +0.0%       3.00        proc-vmstat.unevictable_pgs_munlocked
      0.00          -100.0%       0.00        proc-vmstat.unevictable_pgs_rescued
   1563750 ±  2%      +3.0%    1610100 ± 12%  proc-vmstat.unevictable_pgs_scanned
      0.61 ±  3%      +0.4%       0.61        perf-stat.i.MPKI
 1.501e+11 ±  2%      -6.3%  1.407e+11 ±  8%  perf-stat.i.branch-instructions
      0.33 ±  3%      +0.0        0.33 ±  6%  perf-stat.i.branch-miss-rate%
 3.963e+08 ±  2%      -7.9%  3.649e+08 ±  9%  perf-stat.i.branch-misses
     65.13            -2.3       62.80 ±  7%  perf-stat.i.cache-miss-rate%
 7.382e+08 ±  2%      -6.3%  6.919e+08 ±  8%  perf-stat.i.cache-misses
 1.106e+09            -6.6%  1.033e+09 ±  8%  perf-stat.i.cache-references
      9256            -4.6%       8828 ±  6%  perf-stat.i.context-switches
      0.50 ±  7%     +33.3%       0.66 ± 43%  perf-stat.i.cpi
    224180            -0.0%     224171        perf-stat.i.cpu-clock
  5.93e+11 ±  2%      -3.8%  5.706e+11 ±  8%  perf-stat.i.cpu-cycles
    359.21 ±  2%      -6.1%     337.30 ±  3%  perf-stat.i.cpu-migrations
    810.86 ±  3%     +50.1%       1217 ± 61%  perf-stat.i.cycles-between-cache-misses
      0.01 ±134%      +0.1        0.13 ±176%  perf-stat.i.dTLB-load-miss-rate%
  12662250 ±  2%      +1.0%   12784614 ±  2%  perf-stat.i.dTLB-load-misses
 2.861e+11 ±  2%      -6.5%  2.677e+11 ±  8%  perf-stat.i.dTLB-loads
      0.01 ±  9%      +0.0        0.01 ± 43%  perf-stat.i.dTLB-store-miss-rate%
  13024682 ±  2%     -14.4%   11149515 ±  8%  perf-stat.i.dTLB-store-misses
 1.757e+11 ±  2%      -6.6%  1.641e+11 ±  8%  perf-stat.i.dTLB-stores
 1.275e+12 ±  2%      -6.4%  1.194e+12 ±  8%  perf-stat.i.instructions
      2.10            -6.3%       1.97 ±  7%  perf-stat.i.ipc
      0.20 ± 48%      +6.9%       0.21 ± 40%  perf-stat.i.major-faults
      1.26 ±  2%      -6.7%       1.18 ±  9%  perf-stat.i.metric.G/sec
      2.65 ±  2%      -3.8%       2.55 ±  8%  perf-stat.i.metric.GHz
    694.48            -3.0%     673.92 ±  2%  perf-stat.i.metric.K/sec
      1472            -6.2%       1380 ±  8%  perf-stat.i.metric.M/sec
   1305005 ±  2%     -17.5%    1076189 ±  9%  perf-stat.i.minor-faults
      4.59 ±  8%      +2.1        6.65 ± 51%  perf-stat.i.node-load-miss-rate%
   2077324 ±  2%      -2.2%    2032308 ± 12%  perf-stat.i.node-load-misses
 1.353e+08            -5.4%  1.281e+08 ±  8%  perf-stat.i.node-loads
   1305005 ±  2%     -17.5%    1076190 ±  9%  perf-stat.i.page-faults
    224180            -0.0%     224171        perf-stat.i.task-clock
      0.46 ± 50%     +25.1%       0.58        perf-stat.overall.MPKI
      0.21 ± 50%      +0.0        0.26        perf-stat.overall.branch-miss-rate%
     53.47 ± 50%     +13.6       67.06        perf-stat.overall.cache-miss-rate%
      0.37 ± 50%     +28.6%       0.48        perf-stat.overall.cpi
    643.14 ± 50%     +28.4%     826.00        perf-stat.overall.cycles-between-cache-misses
      0.00 ± 50%      +0.0        0.00 ± 10%  perf-stat.overall.dTLB-load-miss-rate%
      0.01 ± 50%      +0.0        0.01 ±  7%  perf-stat.overall.dTLB-store-miss-rate%
      1.72 ± 50%     +21.6%       2.09        perf-stat.overall.ipc
      1.19 ± 50%      +0.4        1.58 ± 12%  perf-stat.overall.node-load-miss-rate%
 1.173e+11 ± 50%     +17.9%  1.383e+11 ±  7%  perf-stat.ps.branch-instructions
 3.098e+08 ± 50%     +15.8%  3.586e+08 ±  8%  perf-stat.ps.branch-misses
 5.777e+08 ± 50%     +17.9%  6.809e+08 ±  7%  perf-stat.ps.cache-misses
 8.643e+08 ± 50%     +17.4%  1.015e+09 ±  7%  perf-stat.ps.cache-references
      7130 ± 50%     +19.9%       8552 ±  5%  perf-stat.ps.context-switches
    174005 ± 50%     +25.5%     218373        perf-stat.ps.cpu-clock
 4.644e+11 ± 50%     +21.0%  5.621e+11 ±  7%  perf-stat.ps.cpu-cycles
    272.51 ± 50%     +16.6%     317.85 ±  2%  perf-stat.ps.cpu-migrations
   9924949 ± 50%     +25.8%   12485559 ±  2%  perf-stat.ps.dTLB-load-misses
 2.237e+11 ± 50%     +17.6%  2.631e+11 ±  7%  perf-stat.ps.dTLB-loads
  10159493 ± 50%      +7.8%   10948561 ±  7%  perf-stat.ps.dTLB-store-misses
 1.374e+11 ± 50%     +17.4%  1.613e+11 ±  7%  perf-stat.ps.dTLB-stores
 9.964e+11 ± 50%     +17.7%  1.173e+12 ±  7%  perf-stat.ps.instructions
      0.16 ± 73%     +21.3%       0.20 ± 35%  perf-stat.ps.major-faults
   1016189 ± 50%      +3.9%    1056067 ±  9%  perf-stat.ps.minor-faults
   1612917 ± 50%     +25.8%    2029077 ± 12%  perf-stat.ps.node-load-misses
 1.067e+08 ± 50%     +18.9%  1.269e+08 ±  7%  perf-stat.ps.node-loads
   1016189 ± 50%      +3.9%    1056067 ±  9%  perf-stat.ps.page-faults
    174005 ± 50%     +25.5%     218373        perf-stat.ps.task-clock
 6.343e+13 ± 50%     +23.6%  7.843e+13        perf-stat.total.instructions
   7114718            -1.2%    7027229        sched_debug.cfs_rq:/.avg_vruntime.avg
   7180297            -1.0%    7105319        sched_debug.cfs_rq:/.avg_vruntime.max
   4657471            -3.9%    4477025 ±  2%  sched_debug.cfs_rq:/.avg_vruntime.min
    174545 ±  2%      +3.8%     181256 ±  3%  sched_debug.cfs_rq:/.avg_vruntime.stddev
      0.55            +2.0%       0.56        sched_debug.cfs_rq:/.h_nr_running.avg
      2.10 ± 20%      +9.5%       2.30 ± 17%  sched_debug.cfs_rq:/.h_nr_running.max
      0.50            +0.0%       0.50        sched_debug.cfs_rq:/.h_nr_running.min
      0.22 ±  8%     +12.7%       0.25 ± 10%  sched_debug.cfs_rq:/.h_nr_running.stddev
     43106 ± 33%    +312.8%     177940 ± 52%  sched_debug.cfs_rq:/.left_vruntime.avg
   6406425 ± 17%     +10.2%    7059947        sched_debug.cfs_rq:/.left_vruntime.max
      0.00            +0.0%       0.00        sched_debug.cfs_rq:/.left_vruntime.min
    490512 ± 23%    +111.4%    1037149 ± 28%  sched_debug.cfs_rq:/.left_vruntime.stddev
      9516 ± 26%    +105.3%      19537 ± 37%  sched_debug.cfs_rq:/.load.avg
   1009013 ± 48%     +15.6%    1166009 ± 43%  sched_debug.cfs_rq:/.load.max
      2280            -0.4%       2271        sched_debug.cfs_rq:/.load.min
     80417 ± 39%     +48.9%     119761 ± 28%  sched_debug.cfs_rq:/.load.stddev
    615.03 ± 25%     +70.0%       1045 ± 46%  sched_debug.cfs_rq:/.load_avg.avg
     43884 ±  3%      +0.4%      44060 ±  3%  sched_debug.cfs_rq:/.load_avg.max
      2.15 ± 10%      +0.0%       2.15 ± 14%  sched_debug.cfs_rq:/.load_avg.min
      4551 ± 15%     +29.7%       5903 ± 27%  sched_debug.cfs_rq:/.load_avg.stddev
   7114719            -1.2%    7027233        sched_debug.cfs_rq:/.min_vruntime.avg
   7180297            -1.0%    7105319        sched_debug.cfs_rq:/.min_vruntime.max
   4657590            -3.9%    4477122 ±  2%  sched_debug.cfs_rq:/.min_vruntime.min
    174538 ±  2%      +3.8%     181250 ±  3%  sched_debug.cfs_rq:/.min_vruntime.stddev
      0.54            +1.4%       0.54        sched_debug.cfs_rq:/.nr_running.avg
      1.80 ± 22%      +8.3%       1.95 ± 17%  sched_debug.cfs_rq:/.nr_running.max
      0.50            +0.0%       0.50        sched_debug.cfs_rq:/.nr_running.min
      0.17 ± 12%     +18.9%       0.21 ± 12%  sched_debug.cfs_rq:/.nr_running.stddev
     92.62 ±144%    +105.7%     190.49 ± 96%  sched_debug.cfs_rq:/.removed.load_avg.avg
     13672 ±147%     +96.3%      26833 ± 80%  sched_debug.cfs_rq:/.removed.load_avg.max
      1060 ±145%    +102.2%       2144 ± 84%  sched_debug.cfs_rq:/.removed.load_avg.stddev
      6.84 ± 48%      -1.3%       6.76 ± 45%  sched_debug.cfs_rq:/.removed.runnable_avg.avg
    265.30 ±  2%      -1.6%     261.00        sched_debug.cfs_rq:/.removed.runnable_avg.max
     39.85 ± 21%      -1.3%      39.33 ± 20%  sched_debug.cfs_rq:/.removed.runnable_avg.stddev
      6.84 ± 48%      -1.3%       6.76 ± 45%  sched_debug.cfs_rq:/.removed.util_avg.avg
    265.30 ±  2%      -1.6%     261.00        sched_debug.cfs_rq:/.removed.util_avg.max
     39.85 ± 21%      -1.3%      39.33 ± 20%  sched_debug.cfs_rq:/.removed.util_avg.stddev
     43106 ± 33%    +312.8%     177940 ± 52%  sched_debug.cfs_rq:/.right_vruntime.avg
   6406449 ± 17%     +10.2%    7059947        sched_debug.cfs_rq:/.right_vruntime.max
      0.00            +0.0%       0.00        sched_debug.cfs_rq:/.right_vruntime.min
    490516 ± 23%    +111.4%    1037152 ± 28%  sched_debug.cfs_rq:/.right_vruntime.stddev
    602.56            +1.7%     612.78        sched_debug.cfs_rq:/.runnable_avg.avg
      1824 ± 15%      +8.0%       1970 ± 13%  sched_debug.cfs_rq:/.runnable_avg.max
    511.80            -9.8%     461.45 ± 32%  sched_debug.cfs_rq:/.runnable_avg.min
    197.27 ±  7%      +6.7%     210.57 ±  8%  sched_debug.cfs_rq:/.runnable_avg.stddev
      0.50 ± 99%     +10.0%       0.55 ±124%  sched_debug.cfs_rq:/.spread.avg
    111.63 ± 99%      +7.6%     120.07 ±126%  sched_debug.cfs_rq:/.spread.max
      7.44 ± 99%      +7.7%       8.02 ±125%  sched_debug.cfs_rq:/.spread.stddev
    588.90            +0.3%     590.76        sched_debug.cfs_rq:/.util_avg.avg
      1267 ± 16%      +1.5%       1286 ± 17%  sched_debug.cfs_rq:/.util_avg.max
     32.70 ±290%    +205.4%      99.85 ±172%  sched_debug.cfs_rq:/.util_avg.min
    163.16 ±  4%      +0.8%     164.51 ±  9%  sched_debug.cfs_rq:/.util_avg.stddev
    261.48            +0.7%     263.22        sched_debug.cfs_rq:/.util_est_enqueued.avg
    924.20            +9.8%       1014 ± 13%  sched_debug.cfs_rq:/.util_est_enqueued.max
     16.40 ± 90%     +74.4%      28.60 ± 42%  sched_debug.cfs_rq:/.util_est_enqueued.min
     87.88 ±  8%      +5.1%      92.40 ± 10%  sched_debug.cfs_rq:/.util_est_enqueued.stddev
    895647            +2.7%     919389 ±  5%  sched_debug.cpu.avg_idle.avg
   1584773 ± 25%      -5.4%    1499839 ± 21%  sched_debug.cpu.avg_idle.max
    180026 ± 47%     -17.3%     148883 ± 47%  sched_debug.cpu.avg_idle.min
    139458 ± 12%     +17.5%     163799 ± 45%  sched_debug.cpu.avg_idle.stddev
     91837 ± 13%      +0.9%      92634 ± 10%  sched_debug.cpu.clock.avg
     91883 ± 13%      +0.9%      92682 ± 10%  sched_debug.cpu.clock.max
     91778 ± 13%      +0.9%      92573 ± 10%  sched_debug.cpu.clock.min
     28.81 ± 15%      +1.8%      29.34 ± 13%  sched_debug.cpu.clock.stddev
     91190 ± 13%      +0.9%      91995 ± 10%  sched_debug.cpu.clock_task.avg
     91362 ± 13%      +0.9%      92168 ± 10%  sched_debug.cpu.clock_task.max
     78321 ± 15%      +1.2%      79242 ± 12%  sched_debug.cpu.clock_task.min
    873.05            -1.0%     864.68        sched_debug.cpu.clock_task.stddev
      4249            -1.9%       4166        sched_debug.cpu.curr->pid.avg
      7337 ±  3%      +0.8%       7398 ±  2%  sched_debug.cpu.curr->pid.max
      1048 ± 97%     -90.8%      96.65 ± 46%  sched_debug.cpu.curr->pid.min
    836.33 ± 10%     +35.5%       1132 ± 14%  sched_debug.cpu.curr->pid.stddev
    502089            +1.2%     508197 ±  3%  sched_debug.cpu.max_idle_balance_cost.avg
    700999 ± 18%      -2.1%     685975 ± 15%  sched_debug.cpu.max_idle_balance_cost.max
    500000            +0.0%     500000        sched_debug.cpu.max_idle_balance_cost.min
     17056 ± 57%     +39.7%      23833 ±121%  sched_debug.cpu.max_idle_balance_cost.stddev
      4294            +0.0%       4294        sched_debug.cpu.next_balance.avg
      4294            +0.0%       4294        sched_debug.cpu.next_balance.max
      4294            +0.0%       4294        sched_debug.cpu.next_balance.min
      0.00 ± 57%     -20.1%       0.00 ± 36%  sched_debug.cpu.next_balance.stddev
      0.55            +1.9%       0.56 ±  2%  sched_debug.cpu.nr_running.avg
      2.45 ± 19%      +2.0%       2.50 ± 20%  sched_debug.cpu.nr_running.max
      0.50            +0.0%       0.50        sched_debug.cpu.nr_running.min
      0.24 ± 10%      +8.1%       0.26 ± 11%  sched_debug.cpu.nr_running.stddev
      2437            -0.6%       2423        sched_debug.cpu.nr_switches.avg
     46748 ± 21%      +2.2%      47779 ± 11%  sched_debug.cpu.nr_switches.max
      1195            -3.2%       1157 ±  2%  sched_debug.cpu.nr_switches.min
      3810 ± 11%      +1.6%       3869 ±  6%  sched_debug.cpu.nr_switches.stddev
      0.01 ± 56%     -54.2%       0.00 ±110%  sched_debug.cpu.nr_uninterruptible.avg
     42.85 ± 38%      -0.9%      42.45 ± 38%  sched_debug.cpu.nr_uninterruptible.max
    -17.50           +35.1%     -23.65        sched_debug.cpu.nr_uninterruptible.min
      5.78 ± 12%      +4.1%       6.02 ± 12%  sched_debug.cpu.nr_uninterruptible.stddev
     91781 ± 13%      +0.9%      92577 ± 10%  sched_debug.cpu_clk
    996147            +0.0%     996147        sched_debug.dl_rq:.dl_bw->bw.avg
    996147            +0.0%     996147        sched_debug.dl_rq:.dl_bw->bw.max
    996147            +0.0%     996147        sched_debug.dl_rq:.dl_bw->bw.min
 4.295e+09            +0.0%  4.295e+09        sched_debug.jiffies
     90574 ± 13%      +0.9%      91369 ± 10%  sched_debug.ktime
      0.00            +0.0%       0.00        sched_debug.rt_rq:.rt_nr_migratory.avg
      0.50            +0.0%       0.50        sched_debug.rt_rq:.rt_nr_migratory.max
      0.03            +0.0%       0.03        sched_debug.rt_rq:.rt_nr_migratory.stddev
      0.00            +0.0%       0.00        sched_debug.rt_rq:.rt_nr_running.avg
      0.50            +0.0%       0.50        sched_debug.rt_rq:.rt_nr_running.max
      0.03            +0.0%       0.03        sched_debug.rt_rq:.rt_nr_running.stddev
    950.00            +0.0%     950.00        sched_debug.rt_rq:.rt_runtime.avg
    950.00            +0.0%     950.00        sched_debug.rt_rq:.rt_runtime.max
    950.00            +0.0%     950.00        sched_debug.rt_rq:.rt_runtime.min
      0.00        +2.2e+98%       0.00 ±299%  sched_debug.rt_rq:.rt_throttled.avg
      0.00         +5e+100%       0.05 ±300%  sched_debug.rt_rq:.rt_throttled.max
      0.00        +3.3e+99%       0.00 ±299%  sched_debug.rt_rq:.rt_throttled.stddev
      0.91 ± 64%      -7.2%       0.84 ± 91%  sched_debug.rt_rq:.rt_time.avg
    203.22 ± 64%      -7.2%     188.67 ± 91%  sched_debug.rt_rq:.rt_time.max
     13.55 ± 64%      -7.2%      12.58 ± 91%  sched_debug.rt_rq:.rt_time.stddev
     92678 ± 12%      +0.9%      93471 ± 10%  sched_debug.sched_clk
      1.00            +0.0%       1.00        sched_debug.sched_clock_stable()
      3.00            +0.0%       3.00        sched_debug.sysctl_sched.sysctl_sched_base_slice
  25056823            +0.0%   25056823        sched_debug.sysctl_sched.sysctl_sched_features
      1.00            +0.0%       1.00        sched_debug.sysctl_sched.sysctl_sched_tunable_scaling
      0.13 ±200%      -0.1        0.00        perf-profile.calltrace.cycles-pp.tick_sched_handle.tick_sched_timer.__hrtimer_run_queues.hrtimer_interrupt.__sysvec_apic_timer_interrupt
      0.13 ±200%      -0.1        0.00        perf-profile.calltrace.cycles-pp.update_process_times.tick_sched_handle.tick_sched_timer.__hrtimer_run_queues.hrtimer_interrupt
      0.12 ±200%      -0.1        0.00        perf-profile.calltrace.cycles-pp.scheduler_tick.update_process_times.tick_sched_handle.tick_sched_timer.__hrtimer_run_queues
      0.10 ±200%      -0.1        0.05 ±300%  perf-profile.calltrace.cycles-pp.j__udyLAllocJBBJP
      0.05 ±299%      -0.1        0.00        perf-profile.calltrace.cycles-pp.get_mem_cgroup_from_mm.__mem_cgroup_charge.do_anonymous_page.__handle_mm_fault.handle_mm_fault
      0.14 ±199%      -0.0        0.10 ±200%  perf-profile.calltrace.cycles-pp.tick_sched_timer.__hrtimer_run_queues.hrtimer_interrupt.__sysvec_apic_timer_interrupt.sysvec_apic_timer_interrupt
      0.12 ±200%      +0.0        0.16 ±152%  perf-profile.calltrace.cycles-pp.page_counter_uncharge.uncharge_batch.__mem_cgroup_uncharge_list.release_pages.tlb_batch_pages_flush
      0.13 ±200%      +0.1        0.22 ±122%  perf-profile.calltrace.cycles-pp.uncharge_batch.__mem_cgroup_uncharge_list.release_pages.tlb_batch_pages_flush.tlb_finish_mmu
      0.25 ±200%      +0.1        0.40 ± 65%  perf-profile.calltrace.cycles-pp.__hrtimer_run_queues.hrtimer_interrupt.__sysvec_apic_timer_interrupt.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt
      0.00            +0.2        0.19 ±152%  perf-profile.calltrace.cycles-pp.rmqueue.get_page_from_freelist.__alloc_pages.pte_alloc_one.__pte_alloc
      0.00            +0.2        0.19 ±152%  perf-profile.calltrace.cycles-pp.__rmqueue_pcplist.rmqueue.get_page_from_freelist.__alloc_pages.pte_alloc_one
      0.00            +0.2        0.19 ±153%  perf-profile.calltrace.cycles-pp.get_page_from_freelist.__alloc_pages.pte_alloc_one.__pte_alloc.do_anonymous_page
      0.00            +0.2        0.20 ±152%  perf-profile.calltrace.cycles-pp.__alloc_pages.pte_alloc_one.__pte_alloc.do_anonymous_page.__handle_mm_fault
      0.00            +0.2        0.20 ±152%  perf-profile.calltrace.cycles-pp.__pte_alloc.do_anonymous_page.__handle_mm_fault.handle_mm_fault.do_user_addr_fault
      0.00            +0.2        0.20 ±152%  perf-profile.calltrace.cycles-pp.pte_alloc_one.__pte_alloc.do_anonymous_page.__handle_mm_fault.handle_mm_fault
      0.23 ±200%      +0.2        0.43 ±122%  perf-profile.calltrace.cycles-pp.JudyLDel@plt
      0.14 ±200%      +0.3        0.40 ± 66%  perf-profile.calltrace.cycles-pp.__mem_cgroup_uncharge_list.release_pages.tlb_batch_pages_flush.tlb_finish_mmu.unmap_region
      0.12 ±200%      +0.3        0.38 ± 65%  perf-profile.calltrace.cycles-pp.hrtimer_interrupt.__sysvec_apic_timer_interrupt.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.JudyLGet
      0.12 ±200%      +0.3        0.38 ± 65%  perf-profile.calltrace.cycles-pp.__sysvec_apic_timer_interrupt.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.JudyLGet
      0.16 ±200%      +0.3        0.42 ± 66%  perf-profile.calltrace.cycles-pp.__mem_cgroup_charge.do_anonymous_page.__handle_mm_fault.handle_mm_fault.do_user_addr_fault
      0.13 ±200%      +0.3        0.43 ± 65%  perf-profile.calltrace.cycles-pp.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.JudyLGet
      0.14 ±200%      +0.3        0.47 ± 65%  perf-profile.calltrace.cycles-pp.asm_sysvec_apic_timer_interrupt.JudyLGet
      0.17 ±200%      +0.4        0.53 ± 65%  perf-profile.calltrace.cycles-pp.hrtimer_interrupt.__sysvec_apic_timer_interrupt.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt
      0.17 ±200%      +0.4        0.54 ± 65%  perf-profile.calltrace.cycles-pp.__sysvec_apic_timer_interrupt.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt
      0.16 ±200%      +0.4        0.53 ± 68%  perf-profile.calltrace.cycles-pp.ordered_events__queue.process_simple.reader__read_event.perf_session__process_events.record__finish_output
      0.16 ±200%      +0.4        0.53 ± 68%  perf-profile.calltrace.cycles-pp.queue_event.ordered_events__queue.process_simple.reader__read_event.perf_session__process_events
      0.17 ±200%      +0.4        0.54 ± 65%  perf-profile.calltrace.cycles-pp.cfree
      0.16 ±200%      +0.4        0.53 ± 68%  perf-profile.calltrace.cycles-pp.process_simple.reader__read_event.perf_session__process_events.record__finish_output.__cmd_record
      0.16 ±200%      +0.4        0.53 ± 68%  perf-profile.calltrace.cycles-pp.__cmd_record
      0.16 ±200%      +0.4        0.53 ± 68%  perf-profile.calltrace.cycles-pp.record__finish_output.__cmd_record
      0.16 ±200%      +0.4        0.53 ± 68%  perf-profile.calltrace.cycles-pp.perf_session__process_events.record__finish_output.__cmd_record
      0.16 ±200%      +0.4        0.53 ± 68%  perf-profile.calltrace.cycles-pp.reader__read_event.perf_session__process_events.record__finish_output.__cmd_record
      0.20 ±200%      +0.4        0.65 ± 65%  perf-profile.calltrace.cycles-pp.__split_vma.do_vmi_align_munmap.__do_sys_brk.do_syscall_64.entry_SYSCALL_64_after_hwframe
      0.21 ±200%      +0.5        0.67 ± 65%  perf-profile.calltrace.cycles-pp.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt
      0.22 ±200%      +0.5        0.72 ± 65%  perf-profile.calltrace.cycles-pp.asm_sysvec_apic_timer_interrupt
      0.25 ±200%      +0.5        0.77 ± 65%  perf-profile.calltrace.cycles-pp.j__udyLCascade2
      0.24 ±200%      +0.5        0.76 ± 65%  perf-profile.calltrace.cycles-pp.j__udyLCreateBranchU
      0.27 ±200%      +0.6        0.87 ± 65%  perf-profile.calltrace.cycles-pp.JudyLGet@plt
      0.27 ±200%      +0.6        0.88 ± 65%  perf-profile.calltrace.cycles-pp.JudyLIns@plt
      0.00            +0.7        0.68 ± 92%  perf-profile.calltrace.cycles-pp._raw_spin_lock_irqsave.free_pcppages_bulk.free_unref_page_commit.free_unref_page.rcu_do_batch
      0.34 ±200%      +0.8        1.11 ± 65%  perf-profile.calltrace.cycles-pp.malloc
      0.00            +0.8        0.82 ± 79%  perf-profile.calltrace.cycles-pp.free_pcppages_bulk.free_unref_page_commit.free_unref_page.rcu_do_batch.rcu_core
      0.00            +1.1        1.13 ± 84%  perf-profile.calltrace.cycles-pp._raw_spin_lock_irqsave.free_pcppages_bulk.free_unref_page_commit.free_unref_page.__unfreeze_partials
      0.00            +1.2        1.24 ± 83%  perf-profile.calltrace.cycles-pp.free_pcppages_bulk.free_unref_page_commit.free_unref_page.__unfreeze_partials.rcu_do_batch
      0.00            +1.2        1.24 ± 82%  perf-profile.calltrace.cycles-pp.free_unref_page_commit.free_unref_page.__unfreeze_partials.rcu_do_batch.rcu_core
      0.73 ±200%      +1.6        2.29 ± 65%  perf-profile.calltrace.cycles-pp.JudyLDel
      0.00            +1.6        1.57 ±101%  perf-profile.calltrace.cycles-pp.__rmqueue_pcplist.rmqueue.get_page_from_freelist.__alloc_pages.__folio_alloc
      0.70 ±200%      +1.6        2.27 ± 65%  perf-profile.calltrace.cycles-pp.JudyLIns
      0.00            +1.6        1.58 ±100%  perf-profile.calltrace.cycles-pp.rmqueue.get_page_from_freelist.__alloc_pages.__folio_alloc.vma_alloc_folio
      0.00            +1.6        1.59 ±110%  perf-profile.calltrace.cycles-pp._raw_spin_lock_irqsave.rmqueue_bulk.__rmqueue_pcplist.rmqueue.get_page_from_freelist
      0.00            +1.6        1.59 ±110%  perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_irqsave.rmqueue_bulk.__rmqueue_pcplist.rmqueue
      0.00            +1.7        1.71 ±108%  perf-profile.calltrace.cycles-pp.rmqueue_bulk.__rmqueue_pcplist.rmqueue.get_page_from_freelist.__alloc_pages
      0.12 ±200%      +1.7        1.86 ± 92%  perf-profile.calltrace.cycles-pp.get_page_from_freelist.__alloc_pages.__folio_alloc.vma_alloc_folio.do_anonymous_page
      0.12 ±200%      +1.8        1.88 ± 92%  perf-profile.calltrace.cycles-pp.__alloc_pages.__folio_alloc.vma_alloc_folio.do_anonymous_page.__handle_mm_fault
      0.12 ±200%      +1.8        1.88 ± 92%  perf-profile.calltrace.cycles-pp.__folio_alloc.vma_alloc_folio.do_anonymous_page.__handle_mm_fault.handle_mm_fault
      0.13 ±200%      +1.8        1.90 ± 91%  perf-profile.calltrace.cycles-pp.vma_alloc_folio.do_anonymous_page.__handle_mm_fault.handle_mm_fault.do_user_addr_fault
      0.00            +1.8        1.81 ± 87%  perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_irqsave.free_pcppages_bulk.free_unref_page_commit.free_unref_page
      0.12 ±200%      +2.2        2.30 ± 73%  perf-profile.calltrace.cycles-pp._raw_spin_lock_irqsave.folio_lruvec_lock_irqsave.release_pages.tlb_batch_pages_flush.tlb_finish_mmu
      0.13 ±200%      +2.2        2.32 ± 73%  perf-profile.calltrace.cycles-pp.folio_lruvec_lock_irqsave.release_pages.tlb_batch_pages_flush.tlb_finish_mmu.unmap_region
      0.00            +2.2        2.19 ± 74%  perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_irqsave.folio_lruvec_lock_irqsave.release_pages.tlb_batch_pages_flush
      0.12 ±200%      +2.3        2.42 ± 77%  perf-profile.calltrace.cycles-pp.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
      0.12 ±200%      +2.3        2.45 ± 77%  perf-profile.calltrace.cycles-pp.kthread.ret_from_fork.ret_from_fork_asm
      0.12 ±200%      +2.3        2.45 ± 77%  perf-profile.calltrace.cycles-pp.ret_from_fork_asm
      0.12 ±200%      +2.3        2.45 ± 77%  perf-profile.calltrace.cycles-pp.ret_from_fork.ret_from_fork_asm
      1.05 ±200%      +2.4        3.42 ± 65%  perf-profile.calltrace.cycles-pp.stress_judy
      0.33 ±200%      +2.6        2.90 ± 70%  perf-profile.calltrace.cycles-pp.release_pages.tlb_batch_pages_flush.tlb_finish_mmu.unmap_region.do_vmi_align_munmap
      0.34 ±200%      +2.6        2.92 ± 70%  perf-profile.calltrace.cycles-pp.tlb_batch_pages_flush.tlb_finish_mmu.unmap_region.do_vmi_align_munmap.__do_sys_brk
      0.46 ±200%      +2.6        3.10 ± 80%  perf-profile.calltrace.cycles-pp.do_anonymous_page.__handle_mm_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault
      0.38 ±200%      +2.7        3.04 ± 69%  perf-profile.calltrace.cycles-pp.tlb_finish_mmu.unmap_region.do_vmi_align_munmap.__do_sys_brk.do_syscall_64
      0.48 ±200%      +2.7        3.15 ± 79%  perf-profile.calltrace.cycles-pp.__handle_mm_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault
      0.50 ±200%      +2.7        3.20 ± 79%  perf-profile.calltrace.cycles-pp.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault
      0.53 ±200%      +2.8        3.30 ± 78%  perf-profile.calltrace.cycles-pp.do_user_addr_fault.exc_page_fault.asm_exc_page_fault
      0.53 ±200%      +2.8        3.31 ± 78%  perf-profile.calltrace.cycles-pp.exc_page_fault.asm_exc_page_fault
      0.57 ±200%      +2.9        3.43 ± 77%  perf-profile.calltrace.cycles-pp.asm_exc_page_fault
      0.49 ±200%      +2.9        3.36 ± 68%  perf-profile.calltrace.cycles-pp.unmap_region.do_vmi_align_munmap.__do_sys_brk.do_syscall_64.entry_SYSCALL_64_after_hwframe
      0.78 ±200%      +3.5        4.26 ± 67%  perf-profile.calltrace.cycles-pp.do_vmi_align_munmap.__do_sys_brk.do_syscall_64.entry_SYSCALL_64_after_hwframe.brk
      0.81 ±200%      +3.6        4.36 ± 67%  perf-profile.calltrace.cycles-pp.__do_sys_brk.do_syscall_64.entry_SYSCALL_64_after_hwframe.brk
      0.82 ±200%      +3.6        4.38 ± 67%  perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.brk
      0.82 ±200%      +3.6        4.38 ± 67%  perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.brk
      0.83 ±200%      +3.6        4.40 ± 67%  perf-profile.calltrace.cycles-pp.brk
     13.33 ±200%     +29.8       43.13 ± 65%  perf-profile.calltrace.cycles-pp.JudyLGet
      0.02 ±200%      -0.0        0.00        perf-profile.children.cycles-pp.rcu_sched_clock_irq
      0.01 ±200%      -0.0        0.00        perf-profile.children.cycles-pp.rcu_pending
      0.01 ±201%      -0.0        0.00        perf-profile.children.cycles-pp.update_cfs_group
      0.01 ±200%      -0.0        0.00        perf-profile.children.cycles-pp.perf_rotate_context
      0.01 ±200%      -0.0        0.00        perf-profile.children.cycles-pp.check_cpu_stall
      0.01 ±200%      -0.0        0.00        perf-profile.children.cycles-pp.arch_scale_freq_tick
      0.01 ±300%      -0.0        0.00        perf-profile.children.cycles-pp.down_write
      0.01 ±200%      +0.0        0.01 ±200%  perf-profile.children.cycles-pp.__get_obj_cgroup_from_memcg
      0.01 ±300%      +0.0        0.01 ±299%  perf-profile.children.cycles-pp.mas_store_b_node
      0.01 ±300%      +0.0        0.01 ±299%  perf-profile.children.cycles-pp.cgroup_rstat_updated
      0.01 ±200%      +0.0        0.02 ±152%  perf-profile.children.cycles-pp.calc_global_load_tick
      0.02 ±200%      +0.0        0.02 ±122%  perf-profile.children.cycles-pp._compound_head
      0.00            +0.0        0.01 ±299%  perf-profile.children.cycles-pp.perf_mmap__read_head
      0.00            +0.0        0.01 ±300%  perf-profile.children.cycles-pp.__get_free_pages
      0.00            +0.0        0.01 ±300%  perf-profile.children.cycles-pp.__tlb_remove_page_size
      0.01 ±201%      +0.0        0.02 ±122%  perf-profile.children.cycles-pp.j__udyLAllocJLL3
      0.01 ±200%      +0.0        0.02 ±100%  perf-profile.children.cycles-pp.rcu_nocb_try_bypass
      0.00            +0.0        0.02 ±152%  perf-profile.children.cycles-pp.__list_add_valid_or_report
      0.02 ±200%      +0.0        0.04 ± 65%  perf-profile.children.cycles-pp.perf_trace_sched_stat_runtime
      0.00            +0.0        0.02 ±153%  perf-profile.children.cycles-pp.worker_thread
      0.00            +0.0        0.02 ±153%  perf-profile.children.cycles-pp.process_one_work
      0.01 ±200%      +0.0        0.03 ± 82%  perf-profile.children.cycles-pp.mas_find
      0.01 ±201%      +0.0        0.03 ± 81%  perf-profile.children.cycles-pp.mtree_range_walk
      0.03 ±200%      +0.0        0.05 ± 67%  perf-profile.children.cycles-pp.propagate_protected_usage
      0.01 ±200%      +0.0        0.04 ± 65%  perf-profile.children.cycles-pp.update_load_avg
      0.01 ±201%      +0.0        0.04 ± 65%  perf-profile.children.cycles-pp.mas_destroy
      0.01 ±201%      +0.0        0.04 ±100%  perf-profile.children.cycles-pp.tick_sched_do_timer
      0.01 ±200%      +0.0        0.04 ± 66%  perf-profile.children.cycles-pp.get_obj_cgroup_from_current
      0.01 ±200%      +0.0        0.04 ± 66%  perf-profile.children.cycles-pp.mem_cgroup_update_lru_size
      0.01 ±200%      +0.0        0.04 ± 66%  perf-profile.children.cycles-pp.__cond_resched
      0.01 ±200%      +0.0        0.04 ± 66%  perf-profile.children.cycles-pp.__mod_node_page_state
      0.01 ±200%      +0.0        0.04 ± 66%  perf-profile.children.cycles-pp.do_brk_flags
      0.01 ±200%      +0.0        0.04 ± 66%  perf-profile.children.cycles-pp.lock_vma_under_rcu
      0.02 ±200%      +0.0        0.05 ± 66%  perf-profile.children.cycles-pp.hrtimer_active
      0.02 ±200%      +0.0        0.04 ± 66%  perf-profile.children.cycles-pp.free_unref_page_list
      0.01 ±200%      +0.0        0.04 ± 66%  perf-profile.children.cycles-pp.unlink_anon_vmas
      0.02 ±200%      +0.0        0.05 ± 65%  perf-profile.children.cycles-pp.___perf_sw_event
      0.02 ±200%      +0.0        0.05 ± 66%  perf-profile.children.cycles-pp.__mod_lruvec_state
      0.02 ±200%      +0.0        0.05 ± 65%  perf-profile.children.cycles-pp.try_charge_memcg
      0.03 ±200%      +0.0        0.07 ± 65%  perf-profile.children.cycles-pp.update_curr
      0.02 ±200%      +0.0        0.05 ± 65%  perf-profile.children.cycles-pp.mas_walk
      0.01 ±200%      +0.0        0.05 ± 65%  perf-profile.children.cycles-pp.task_mm_cid_work
      0.00            +0.0        0.04 ± 66%  perf-profile.children.cycles-pp.note_gp_changes
      0.02 ±200%      +0.0        0.05 ± 65%  perf-profile.children.cycles-pp.free_pgtables
      0.02 ±200%      +0.0        0.05 ± 65%  perf-profile.children.cycles-pp.task_work_run
      0.02 ±200%      +0.0        0.06 ± 66%  perf-profile.children.cycles-pp.__perf_sw_event
      0.02 ±200%      +0.0        0.06 ± 66%  perf-profile.children.cycles-pp.j__udyLCascade3
      0.02 ±200%      +0.0        0.06 ± 65%  perf-profile.children.cycles-pp.j__udyLFreeJLL2
      0.02 ±200%      +0.0        0.06 ± 65%  perf-profile.children.cycles-pp.exit_to_user_mode_loop
      0.02 ±200%      +0.0        0.07 ± 67%  perf-profile.children.cycles-pp.lru_add_fn
      0.04 ±200%      +0.0        0.08 ± 66%  perf-profile.children.cycles-pp.mem_cgroup_commit_charge
      0.02 ±200%      +0.0        0.07 ± 65%  perf-profile.children.cycles-pp.j__udyLAllocJBBJP@plt
      0.02 ±200%      +0.0        0.06 ± 65%  perf-profile.children.cycles-pp.irqentry_exit_to_user_mode
      0.04 ±200%      +0.0        0.09 ± 66%  perf-profile.children.cycles-pp.__count_memcg_events
      0.02 ±200%      +0.0        0.07 ± 65%  perf-profile.children.cycles-pp.j__udyLLeaf2ToLeaf3
      0.02 ±200%      +0.0        0.07 ± 65%  perf-profile.children.cycles-pp.exit_to_user_mode_prepare
      0.02 ±200%      +0.0        0.07 ± 65%  perf-profile.children.cycles-pp.mas_commit_b_node
      0.02 ±200%      +0.0        0.07 ± 65%  perf-profile.children.cycles-pp.__call_rcu_common
      0.01 ±200%      +0.1        0.06 ± 68%  perf-profile.children.cycles-pp.generic_perform_write
      0.02 ±200%      +0.1        0.07 ± 65%  perf-profile.children.cycles-pp.j__udyLFreeJBBJP@plt
      0.03 ±200%      +0.1        0.08 ± 65%  perf-profile.children.cycles-pp.anon_vma_clone
      0.05 ±200%      +0.1        0.10 ± 66%  perf-profile.children.cycles-pp.__mod_memcg_lruvec_state
      0.01 ±200%      +0.1        0.07 ± 69%  perf-profile.children.cycles-pp.shmem_file_write_iter
      0.02 ±200%      +0.1        0.07 ± 70%  perf-profile.children.cycles-pp.vfs_write
      0.03 ±200%      +0.1        0.09 ± 65%  perf-profile.children.cycles-pp.mod_objcg_state
      0.03 ±200%      +0.1        0.08 ± 65%  perf-profile.children.cycles-pp.memcg_slab_post_alloc_hook
      0.02 ±200%      +0.1        0.07 ± 69%  perf-profile.children.cycles-pp.ksys_write
      0.00            +0.1        0.06 ± 74%  perf-profile.children.cycles-pp.allocate_slab
      0.05 ±200%      +0.1        0.10 ± 66%  perf-profile.children.cycles-pp.page_remove_rmap
      0.03 ±200%      +0.1        0.09 ± 65%  perf-profile.children.cycles-pp.j__udyLAllocJLL2
      0.03 ±200%      +0.1        0.09 ± 65%  perf-profile.children.cycles-pp.vm_area_free_rcu_cb
      0.02 ±200%      +0.1        0.08 ± 69%  perf-profile.children.cycles-pp.__libc_write
      0.04 ±200%      +0.1        0.10 ± 66%  perf-profile.children.cycles-pp.native_flush_tlb_one_user
      0.02 ±200%      +0.1        0.08 ± 70%  perf-profile.children.cycles-pp.record__pushfn
      0.02 ±200%      +0.1        0.08 ± 68%  perf-profile.children.cycles-pp.writen
      0.06 ±200%      +0.1        0.12 ± 66%  perf-profile.children.cycles-pp.folio_add_new_anon_rmap
      0.04 ±200%      +0.1        0.10 ± 66%  perf-profile.children.cycles-pp.flush_tlb_func
      0.02 ±200%      +0.1        0.08 ± 71%  perf-profile.children.cycles-pp.perf_mux_hrtimer_handler
      0.02 ±200%      +0.1        0.09 ± 68%  perf-profile.children.cycles-pp.___slab_alloc
      0.03 ±200%      +0.1        0.10 ± 67%  perf-profile.children.cycles-pp.ktime_get_update_offsets_now
      0.04 ±200%      +0.1        0.12 ± 66%  perf-profile.children.cycles-pp.flush_tlb_mm_range
      0.04 ±200%      +0.1        0.13 ± 65%  perf-profile.children.cycles-pp.sync_regs
      0.04 ±200%      +0.1        0.13 ± 65%  perf-profile.children.cycles-pp.rcu_cblist_dequeue
      0.04 ±200%      +0.1        0.13 ± 66%  perf-profile.children.cycles-pp.mas_wr_node_store
      0.03 ±200%      +0.1        0.12 ± 66%  perf-profile.children.cycles-pp.__kmem_cache_alloc_bulk
      0.04 ±200%      +0.1        0.13 ± 65%  perf-profile.children.cycles-pp.kmem_cache_free
      0.04 ±200%      +0.1        0.13 ± 65%  perf-profile.children.cycles-pp.mas_wr_bnode
      0.08 ±200%      +0.1        0.18 ± 65%  perf-profile.children.cycles-pp.task_tick_fair
      0.02 ±200%      +0.1        0.12 ± 69%  perf-profile.children.cycles-pp.perf_mmap__push
      0.03 ±199%      +0.1        0.13 ± 65%  perf-profile.children.cycles-pp.kmem_cache_alloc_bulk
      0.02 ±200%      +0.1        0.12 ± 69%  perf-profile.children.cycles-pp.record__mmap_read_evlist
      0.02 ±200%      +0.1        0.13 ± 69%  perf-profile.children.cycles-pp.cmd_record
      0.10 ±200%      +0.1        0.20 ± 66%  perf-profile.children.cycles-pp.__mod_lruvec_page_state
      0.05 ±200%      +0.1        0.16 ± 65%  perf-profile.children.cycles-pp.__slab_free
      0.05 ±200%      +0.1        0.16 ± 65%  perf-profile.children.cycles-pp.__pte_offset_map_lock
      0.05 ±200%      +0.1        0.16 ± 65%  perf-profile.children.cycles-pp.vm_area_dup
      0.03 ±200%      +0.1        0.14 ± 68%  perf-profile.children.cycles-pp.__libc_start_main
      0.03 ±200%      +0.1        0.14 ± 68%  perf-profile.children.cycles-pp.main
      0.03 ±200%      +0.1        0.14 ± 68%  perf-profile.children.cycles-pp.run_builtin
      0.00            +0.1        0.11 ± 68%  perf-profile.children.cycles-pp.__free_one_page
      0.05 ±200%      +0.1        0.16 ± 65%  perf-profile.children.cycles-pp.mas_wr_store_entry
      0.05 ±200%      +0.1        0.16 ± 65%  perf-profile.children.cycles-pp.free@plt
      0.05 ±200%      +0.1        0.16 ± 65%  perf-profile.children.cycles-pp.JudyFree
      0.05 ±200%      +0.1        0.16 ± 65%  perf-profile.children.cycles-pp.malloc@plt
      0.05 ±200%      +0.1        0.17 ± 65%  perf-profile.children.cycles-pp.JudyMalloc
      0.08 ±200%      +0.1        0.19 ± 67%  perf-profile.children.cycles-pp.zap_pte_range
      0.06 ±200%      +0.1        0.18 ± 65%  perf-profile.children.cycles-pp.mas_store_gfp
      0.06 ±200%      +0.1        0.18 ± 65%  perf-profile.children.cycles-pp.mas_store_prealloc
      0.08 ±200%      +0.1        0.20 ± 66%  perf-profile.children.cycles-pp.native_irq_return_iret
      0.06 ±200%      +0.1        0.18 ± 65%  perf-profile.children.cycles-pp.vma_complete
      0.08 ±200%      +0.1        0.20 ± 67%  perf-profile.children.cycles-pp.zap_pmd_range
      0.05 ±200%      +0.1        0.18 ± 68%  perf-profile.children.cycles-pp.folio_add_lru_vma
      0.06 ±200%      +0.1        0.19 ± 65%  perf-profile.children.cycles-pp.__intel_pmu_enable_all
      0.05 ±200%      +0.1        0.18 ± 68%  perf-profile.children.cycles-pp.folio_batch_move_lru
      0.08 ±200%      +0.1        0.22 ± 67%  perf-profile.children.cycles-pp.unmap_page_range
      0.05 ±200%      +0.1        0.18 ± 65%  perf-profile.children.cycles-pp.mas_alloc_nodes
      0.05 ±200%      +0.1        0.19 ± 65%  perf-profile.children.cycles-pp.mas_preallocate
      0.09 ±200%      +0.1        0.23 ± 67%  perf-profile.children.cycles-pp.unmap_vmas
      0.05 ±200%      +0.1        0.20 ± 65%  perf-profile.children.cycles-pp.__irq_exit_rcu
      0.07 ±200%      +0.2        0.22 ± 65%  perf-profile.children.cycles-pp.kmem_cache_alloc
      0.00            +0.2        0.16 ± 70%  perf-profile.children.cycles-pp.clockevents_program_event
      0.05 ±200%      +0.2        0.21 ± 66%  perf-profile.children.cycles-pp.__list_del_entry_valid_or_report
      0.08 ±200%      +0.2        0.24 ± 65%  perf-profile.children.cycles-pp.JudyFree@plt
      0.06 ±200%      +0.2        0.23 ± 66%  perf-profile.children.cycles-pp.clear_page_erms
      0.08 ±200%      +0.2        0.25 ± 65%  perf-profile.children.cycles-pp.JudyMalloc@plt
      0.10 ±200%      +0.2        0.29 ± 66%  perf-profile.children.cycles-pp.get_mem_cgroup_from_mm
      0.10 ±200%      +0.2        0.31 ± 65%  perf-profile.children.cycles-pp.perf_event_task_tick
      0.10 ±200%      +0.2        0.31 ± 65%  perf-profile.children.cycles-pp.perf_adjust_freq_unthr_context
      0.00            +0.2        0.20 ± 68%  perf-profile.children.cycles-pp.ktime_get
      0.12 ±200%      +0.2        0.33 ± 66%  perf-profile.children.cycles-pp.page_counter_uncharge
      0.13 ±200%      +0.2        0.37 ± 66%  perf-profile.children.cycles-pp.uncharge_batch
      0.14 ±200%      +0.3        0.40 ± 66%  perf-profile.children.cycles-pp.__mem_cgroup_uncharge_list
      0.16 ±200%      +0.3        0.43 ± 66%  perf-profile.children.cycles-pp.__mem_cgroup_charge
      0.00            +0.3        0.28 ± 94%  perf-profile.children.cycles-pp.__pte_alloc
      0.00            +0.3        0.28 ± 94%  perf-profile.children.cycles-pp.pte_alloc_one
      0.12 ±200%      +0.3        0.40 ± 65%  perf-profile.children.cycles-pp.j__udyLFreeJBBJP
      0.23 ±200%      +0.4        0.58 ± 65%  perf-profile.children.cycles-pp.scheduler_tick
      0.16 ±200%      +0.4        0.53 ± 65%  perf-profile.children.cycles-pp.j__udyLAllocJBBJP
      0.16 ±200%      +0.4        0.53 ± 68%  perf-profile.children.cycles-pp.ordered_events__queue
      0.16 ±200%      +0.4        0.53 ± 68%  perf-profile.children.cycles-pp.queue_event
      0.16 ±200%      +0.4        0.53 ± 68%  perf-profile.children.cycles-pp.process_simple
      0.16 ±200%      +0.4        0.53 ± 68%  perf-profile.children.cycles-pp.record__finish_output
      0.16 ±200%      +0.4        0.53 ± 68%  perf-profile.children.cycles-pp.perf_session__process_events
      0.16 ±200%      +0.4        0.53 ± 68%  perf-profile.children.cycles-pp.reader__read_event
      0.17 ±200%      +0.4        0.55 ± 65%  perf-profile.children.cycles-pp.JudyLDel@plt
      0.26 ±200%      +0.4        0.64 ± 65%  perf-profile.children.cycles-pp.update_process_times
      0.26 ±200%      +0.4        0.64 ± 65%  perf-profile.children.cycles-pp.tick_sched_handle
      0.27 ±200%      +0.4        0.70 ± 65%  perf-profile.children.cycles-pp.tick_sched_timer
      0.20 ±200%      +0.4        0.63 ± 65%  perf-profile.children.cycles-pp.cfree
      0.20 ±200%      +0.4        0.65 ± 65%  perf-profile.children.cycles-pp.__split_vma
      0.20 ±200%      +0.5        0.67 ± 65%  perf-profile.children.cycles-pp.JudyLIns@plt
      0.18 ±200%      +0.5        0.66 ± 67%  perf-profile.children.cycles-pp.__cmd_record
      0.29 ±200%      +0.5        0.80 ± 65%  perf-profile.children.cycles-pp.__hrtimer_run_queues
      0.25 ±200%      +0.5        0.77 ± 65%  perf-profile.children.cycles-pp.j__udyLCascade2
      0.24 ±199%      +0.5        0.77 ± 65%  perf-profile.children.cycles-pp.j__udyLCreateBranchU
      0.32 ±200%      +0.7        1.02 ± 65%  perf-profile.children.cycles-pp.JudyLGet@plt
      0.33 ±200%      +0.7        1.07 ± 65%  perf-profile.children.cycles-pp.hrtimer_interrupt
      0.33 ±200%      +0.7        1.08 ± 65%  perf-profile.children.cycles-pp.__sysvec_apic_timer_interrupt
      0.38 ±200%      +0.9        1.25 ± 65%  perf-profile.children.cycles-pp.malloc
      0.39 ±200%      +0.9        1.28 ± 65%  perf-profile.children.cycles-pp.sysvec_apic_timer_interrupt
      0.42 ±200%      +1.0        1.39 ± 65%  perf-profile.children.cycles-pp.asm_sysvec_apic_timer_interrupt
      0.00            +1.3        1.27 ± 82%  perf-profile.children.cycles-pp.__unfreeze_partials
      0.13 ±200%      +1.8        1.89 ± 91%  perf-profile.children.cycles-pp.__folio_alloc
      0.13 ±200%      +1.8        1.91 ± 91%  perf-profile.children.cycles-pp.vma_alloc_folio
      0.00            +1.8        1.84 ±100%  perf-profile.children.cycles-pp.rmqueue_bulk
      0.05 ±200%      +1.9        1.91 ± 98%  perf-profile.children.cycles-pp.rmqueue
      0.84 ±200%      +1.9        2.72 ± 65%  perf-profile.children.cycles-pp.JudyLIns
      0.00            +1.9        1.89 ± 99%  perf-profile.children.cycles-pp.__rmqueue_pcplist
      0.92 ±200%      +2.0        2.88 ± 65%  perf-profile.children.cycles-pp.JudyLDel
      0.12 ±200%      +2.1        2.18 ± 92%  perf-profile.children.cycles-pp.get_page_from_freelist
      0.13 ±200%      +2.1        2.21 ± 91%  perf-profile.children.cycles-pp.__alloc_pages
      0.01 ±300%      +2.1        2.09 ± 81%  perf-profile.children.cycles-pp.free_unref_page
      0.01 ±300%      +2.1        2.09 ± 81%  perf-profile.children.cycles-pp.free_pcppages_bulk
      0.00            +2.1        2.10 ± 81%  perf-profile.children.cycles-pp.free_unref_page_commit
      0.15 ±200%      +2.3        2.40 ± 73%  perf-profile.children.cycles-pp.folio_lruvec_lock_irqsave
      0.12 ±200%      +2.3        2.42 ± 77%  perf-profile.children.cycles-pp.smpboot_thread_fn
      0.12 ±200%      +2.3        2.45 ± 77%  perf-profile.children.cycles-pp.kthread
      0.12 ±200%      +2.3        2.45 ± 77%  perf-profile.children.cycles-pp.ret_from_fork_asm
      0.12 ±200%      +2.3        2.45 ± 77%  perf-profile.children.cycles-pp.ret_from_fork
      0.14 ±199%      +2.4        2.50 ± 76%  perf-profile.children.cycles-pp.rcu_do_batch
      0.16 ±200%      +2.4        2.56 ± 75%  perf-profile.children.cycles-pp.rcu_core
      0.34 ±200%      +2.6        2.92 ± 70%  perf-profile.children.cycles-pp.tlb_batch_pages_flush
      0.34 ±200%      +2.6        2.92 ± 70%  perf-profile.children.cycles-pp.release_pages
      0.46 ±200%      +2.6        3.11 ± 80%  perf-profile.children.cycles-pp.do_anonymous_page
      0.38 ±200%      +2.7        3.04 ± 69%  perf-profile.children.cycles-pp.tlb_finish_mmu
      0.48 ±200%      +2.7        3.16 ± 79%  perf-profile.children.cycles-pp.__handle_mm_fault
      0.50 ±200%      +2.7        3.21 ± 79%  perf-profile.children.cycles-pp.handle_mm_fault
      0.53 ±200%      +2.8        3.31 ± 78%  perf-profile.children.cycles-pp.do_user_addr_fault
      0.54 ±200%      +2.8        3.32 ± 78%  perf-profile.children.cycles-pp.exc_page_fault
      0.58 ±200%      +2.9        3.44 ± 77%  perf-profile.children.cycles-pp.asm_exc_page_fault
      0.49 ±200%      +2.9        3.36 ± 68%  perf-profile.children.cycles-pp.unmap_region
      1.44 ±200%      +3.2        4.66 ± 65%  perf-profile.children.cycles-pp.stress_judy
      0.78 ±200%      +3.5        4.27 ± 67%  perf-profile.children.cycles-pp.do_vmi_align_munmap
      0.81 ±200%      +3.6        4.37 ± 67%  perf-profile.children.cycles-pp.__do_sys_brk
      0.83 ±199%      +3.6        4.42 ± 67%  perf-profile.children.cycles-pp.brk
      0.84 ±200%      +3.6        4.48 ± 67%  perf-profile.children.cycles-pp.do_syscall_64
      0.84 ±200%      +3.6        4.49 ± 67%  perf-profile.children.cycles-pp.entry_SYSCALL_64_after_hwframe
      0.12 ±200%      +5.7        5.86 ± 83%  perf-profile.children.cycles-pp.native_queued_spin_lock_slowpath
      0.16 ±200%      +5.8        6.00 ± 83%  perf-profile.children.cycles-pp._raw_spin_lock_irqsave
     13.47 ±200%     +30.1       43.58 ± 65%  perf-profile.children.cycles-pp.JudyLGet
      0.02 ±200%      -0.0        0.01 ±300%  perf-profile.self.cycles-pp.perf_trace_sched_stat_runtime
      0.01 ±201%      -0.0        0.00        perf-profile.self.cycles-pp.update_cfs_group
      0.01 ±200%      -0.0        0.00        perf-profile.self.cycles-pp.update_curr
      0.01 ±200%      -0.0        0.00        perf-profile.self.cycles-pp.memcg_slab_post_alloc_hook
      0.01 ±200%      -0.0        0.00        perf-profile.self.cycles-pp.check_cpu_stall
      0.01 ±200%      -0.0        0.00        perf-profile.self.cycles-pp.arch_scale_freq_tick
      0.01 ±300%      -0.0        0.00        perf-profile.self.cycles-pp.cgroup_rstat_updated
      0.01 ±200%      +0.0        0.01 ±200%  perf-profile.self.cycles-pp.__get_obj_cgroup_from_memcg
      0.01 ±200%      +0.0        0.02 ±152%  perf-profile.self.cycles-pp.calc_global_load_tick
      0.00            +0.0        0.01 ±299%  perf-profile.self.cycles-pp.rcu_nocb_try_bypass
      0.00            +0.0        0.01 ±300%  perf-profile.self.cycles-pp.zap_pte_range
      0.01 ±200%      +0.0        0.02 ±152%  perf-profile.self.cycles-pp.__handle_mm_fault
      0.01 ±200%      +0.0        0.02 ±122%  perf-profile.self.cycles-pp._compound_head
      0.01 ±201%      +0.0        0.02 ±122%  perf-profile.self.cycles-pp.mtree_range_walk
      0.01 ±201%      +0.0        0.02 ±123%  perf-profile.self.cycles-pp.try_charge_memcg
      0.01 ±200%      +0.0        0.03 ±100%  perf-profile.self.cycles-pp.__mod_node_page_state
      0.02 ±200%      +0.0        0.05 ± 68%  perf-profile.self.cycles-pp.propagate_protected_usage
      0.01 ±200%      +0.0        0.04 ± 66%  perf-profile.self.cycles-pp.mem_cgroup_update_lru_size
      0.01 ±300%      +0.0        0.03 ±101%  perf-profile.self.cycles-pp.tick_sched_do_timer
      0.02 ±200%      +0.0        0.04 ± 66%  perf-profile.self.cycles-pp.hrtimer_active
      0.01 ±200%      +0.0        0.04 ± 66%  perf-profile.self.cycles-pp.___perf_sw_event
      0.02 ±200%      +0.0        0.05 ± 66%  perf-profile.self.cycles-pp.__kmem_cache_alloc_bulk
      0.01 ±200%      +0.0        0.05 ± 66%  perf-profile.self.cycles-pp.task_mm_cid_work
      0.01 ±300%      +0.0        0.04 ± 66%  perf-profile.self.cycles-pp.get_page_from_freelist
      0.04 ±201%      +0.0        0.08 ± 66%  perf-profile.self.cycles-pp.__count_memcg_events
      0.02 ±200%      +0.0        0.05 ± 65%  perf-profile.self.cycles-pp.j__udyLFreeJLL2
      0.02 ±200%      +0.0        0.06 ± 66%  perf-profile.self.cycles-pp.release_pages
      0.02 ±200%      +0.0        0.06 ± 65%  perf-profile.self.cycles-pp.kmem_cache_alloc
      0.02 ±200%      +0.0        0.06 ± 65%  perf-profile.self.cycles-pp.j__udyLCascade3
      0.02 ±200%      +0.0        0.06 ± 66%  perf-profile.self.cycles-pp.j__udyLAllocJLL2
      0.05 ±200%      +0.0        0.10 ± 66%  perf-profile.self.cycles-pp.__mod_memcg_lruvec_state
      0.02 ±200%      +0.0        0.06 ± 66%  perf-profile.self.cycles-pp.mas_wr_node_store
      0.02 ±200%      +0.0        0.07 ± 65%  perf-profile.self.cycles-pp.j__udyLFreeJBBJP@plt
      0.02 ±200%      +0.0        0.07 ± 65%  perf-profile.self.cycles-pp.j__udyLAllocJBBJP@plt
      0.02 ±200%      +0.0        0.07 ± 65%  perf-profile.self.cycles-pp.mod_objcg_state
      0.02 ±200%      +0.0        0.07 ± 65%  perf-profile.self.cycles-pp.j__udyLLeaf2ToLeaf3
      0.02 ±200%      +0.1        0.08 ± 65%  perf-profile.self.cycles-pp.malloc@plt
      0.03 ±200%      +0.1        0.08 ± 65%  perf-profile.self.cycles-pp.free@plt
      0.03 ±200%      +0.1        0.08 ± 65%  perf-profile.self.cycles-pp.JudyFree@plt
      0.02 ±200%      +0.1        0.08 ± 65%  perf-profile.self.cycles-pp.JudyFree
      0.03 ±200%      +0.1        0.08 ± 65%  perf-profile.self.cycles-pp.kmem_cache_free
      0.02 ±200%      +0.1        0.08 ± 65%  perf-profile.self.cycles-pp.JudyMalloc
      0.05 ±200%      +0.1        0.11 ± 66%  perf-profile.self.cycles-pp.__mod_lruvec_page_state
      0.03 ±200%      +0.1        0.08 ± 65%  perf-profile.self.cycles-pp.JudyMalloc@plt
      0.04 ±200%      +0.1        0.10 ± 66%  perf-profile.self.cycles-pp.native_flush_tlb_one_user
      0.03 ±200%      +0.1        0.10 ± 67%  perf-profile.self.cycles-pp.ktime_get_update_offsets_now
      0.00            +0.1        0.07 ± 76%  perf-profile.self.cycles-pp.rmqueue_bulk
      0.05 ±200%      +0.1        0.13 ± 65%  perf-profile.self.cycles-pp.perf_adjust_freq_unthr_context
      0.04 ±200%      +0.1        0.13 ± 65%  perf-profile.self.cycles-pp.sync_regs
      0.04 ±200%      +0.1        0.13 ± 65%  perf-profile.self.cycles-pp.rcu_cblist_dequeue
      0.04 ±200%      +0.1        0.14 ± 65%  perf-profile.self.cycles-pp.__pte_offset_map_lock
      0.00            +0.1        0.09 ± 69%  perf-profile.self.cycles-pp.__free_one_page
      0.04 ±200%      +0.1        0.14 ± 65%  perf-profile.self.cycles-pp._raw_spin_lock_irqsave
      0.05 ±200%      +0.1        0.15 ± 65%  perf-profile.self.cycles-pp.__slab_free
      0.08 ±200%      +0.1        0.20 ± 66%  perf-profile.self.cycles-pp.native_irq_return_iret
      0.06 ±200%      +0.1        0.19 ± 65%  perf-profile.self.cycles-pp.JudyLDel@plt
      0.06 ±200%      +0.1        0.19 ± 65%  perf-profile.self.cycles-pp.__intel_pmu_enable_all
      0.07 ±200%      +0.2        0.23 ± 65%  perf-profile.self.cycles-pp.JudyLIns@plt
      0.05 ±200%      +0.2        0.21 ± 66%  perf-profile.self.cycles-pp.__list_del_entry_valid_or_report
      0.06 ±200%      +0.2        0.23 ± 66%  perf-profile.self.cycles-pp.clear_page_erms
      0.10 ±200%      +0.2        0.28 ± 66%  perf-profile.self.cycles-pp.get_mem_cgroup_from_mm
      0.09 ±200%      +0.2        0.29 ± 66%  perf-profile.self.cycles-pp.page_counter_uncharge
      0.00            +0.2        0.20 ± 68%  perf-profile.self.cycles-pp.ktime_get
      0.10 ±200%      +0.2        0.33 ± 65%  perf-profile.self.cycles-pp.j__udyLFreeJBBJP
      0.12 ±200%      +0.3        0.39 ± 65%  perf-profile.self.cycles-pp.JudyLGet@plt
      0.12 ±200%      +0.3        0.40 ± 65%  perf-profile.self.cycles-pp.j__udyLAllocJBBJP
      0.15 ±200%      +0.4        0.52 ± 68%  perf-profile.self.cycles-pp.queue_event
      0.17 ±200%      +0.4        0.54 ± 65%  perf-profile.self.cycles-pp.cfree
      0.25 ±200%      +0.5        0.76 ± 65%  perf-profile.self.cycles-pp.j__udyLCascade2
      0.23 ±200%      +0.5        0.75 ± 65%  perf-profile.self.cycles-pp.j__udyLCreateBranchU
      0.36 ±200%      +0.8        1.15 ± 65%  perf-profile.self.cycles-pp.malloc
      0.78 ±200%      +1.7        2.44 ± 65%  perf-profile.self.cycles-pp.JudyLDel
      0.76 ±200%      +1.7        2.48 ± 65%  perf-profile.self.cycles-pp.JudyLIns
      1.25 ±200%      +2.8        4.04 ± 65%  perf-profile.self.cycles-pp.stress_judy
      0.12 ±200%      +5.7        5.86 ± 83%  perf-profile.self.cycles-pp.native_queued_spin_lock_slowpath
     13.20 ±200%     +29.5       42.71 ± 65%  perf-profile.self.cycles-pp.JudyLGet
      2886 ±  8%      +3.7%       2993 ± 12%  slabinfo.Acpi-State.active_objs
     56.60 ±  8%      +3.7%      58.70 ± 12%  slabinfo.Acpi-State.active_slabs
      2886 ±  8%      +3.7%       2993 ± 12%  slabinfo.Acpi-State.num_objs
     56.60 ±  8%      +3.7%      58.70 ± 12%  slabinfo.Acpi-State.num_slabs
     36.00            +0.0%      36.00        slabinfo.DCCP.active_objs
      2.00            +0.0%       2.00        slabinfo.DCCP.active_slabs
     36.00            +0.0%      36.00        slabinfo.DCCP.num_objs
      2.00            +0.0%       2.00        slabinfo.DCCP.num_slabs
     34.00            +0.0%      34.00        slabinfo.DCCPv6.active_objs
      2.00            +0.0%       2.00        slabinfo.DCCPv6.active_slabs
     34.00            +0.0%      34.00        slabinfo.DCCPv6.num_objs
      2.00            +0.0%       2.00        slabinfo.DCCPv6.num_slabs
    451.20 ±  2%      -0.7%     448.00        slabinfo.RAW.active_objs
     14.10 ±  2%      -0.7%      14.00        slabinfo.RAW.active_slabs
    451.20 ±  2%      -0.7%     448.00        slabinfo.RAW.num_objs
     14.10 ±  2%      -0.7%      14.00        slabinfo.RAW.num_slabs
    278.20 ±  4%      +0.0%     278.20 ±  4%  slabinfo.RAWv6.active_objs
     10.70 ±  4%      +0.0%      10.70 ±  4%  slabinfo.RAWv6.active_slabs
    278.20 ±  4%      +0.0%     278.20 ±  4%  slabinfo.RAWv6.num_objs
     10.70 ±  4%      +0.0%      10.70 ±  4%  slabinfo.RAWv6.num_slabs
     55.02 ±  7%      +3.5%      56.97        slabinfo.TCP.active_objs
      3.93 ±  7%      +3.5%       4.07        slabinfo.TCP.active_slabs
     55.02 ±  7%      +3.5%      56.97        slabinfo.TCP.num_objs
      3.93 ±  7%      +3.5%       4.07        slabinfo.TCP.num_slabs
     37.70 ± 10%      +3.4%      39.00        slabinfo.TCPv6.active_objs
      2.90 ± 10%      +3.4%       3.00        slabinfo.TCPv6.active_slabs
     37.70 ± 10%      +3.4%      39.00        slabinfo.TCPv6.num_objs
      2.90 ± 10%      +3.4%       3.00        slabinfo.TCPv6.num_slabs
    110.40 ± 10%      +6.5%     117.60 ±  6%  slabinfo.UDPv6.active_objs
      4.60 ± 10%      +6.5%       4.90 ±  6%  slabinfo.UDPv6.active_slabs
    110.40 ± 10%      +6.5%     117.60 ±  6%  slabinfo.UDPv6.num_objs
      4.60 ± 10%      +6.5%       4.90 ±  6%  slabinfo.UDPv6.num_slabs
      2115 ±  8%      -1.2%       2089 ± 10%  slabinfo.UNIX.active_objs
     70.51 ±  8%      -1.2%      69.65 ± 10%  slabinfo.UNIX.active_slabs
      2115 ±  8%      -1.2%       2089 ± 10%  slabinfo.UNIX.num_objs
     70.51 ±  8%      -1.2%      69.65 ± 10%  slabinfo.UNIX.num_slabs
     40729 ±  4%      +2.9%      41929 ±  3%  slabinfo.anon_vma.active_objs
      1047 ±  4%      +2.9%       1077 ±  3%  slabinfo.anon_vma.active_slabs
     40838 ±  4%      +2.9%      42028 ±  3%  slabinfo.anon_vma.num_objs
      1047 ±  4%      +2.9%       1077 ±  3%  slabinfo.anon_vma.num_slabs
     61114 ±  2%      +1.7%      62156 ±  2%  slabinfo.anon_vma_chain.active_objs
    959.75 ±  2%      +1.8%     976.69 ±  2%  slabinfo.anon_vma_chain.active_slabs
     61424 ±  2%      +1.8%      62508 ±  2%  slabinfo.anon_vma_chain.num_objs
    959.75 ±  2%      +1.8%     976.69 ±  2%  slabinfo.anon_vma_chain.num_slabs
     76.00 ± 15%     +10.5%      84.00 ± 17%  slabinfo.bdev_cache.active_objs
      3.80 ± 15%     +10.5%       4.20 ± 17%  slabinfo.bdev_cache.active_slabs
     76.00 ± 15%     +10.5%      84.00 ± 17%  slabinfo.bdev_cache.num_objs
      3.80 ± 15%     +10.5%       4.20 ± 17%  slabinfo.bdev_cache.num_slabs
    454.40 ±  4%      +1.4%     460.80 ±  5%  slabinfo.bio-120.active_objs
      7.10 ±  4%      +1.4%       7.20 ±  5%  slabinfo.bio-120.active_slabs
    454.40 ±  4%      +1.4%     460.80 ±  5%  slabinfo.bio-120.num_objs
      7.10 ±  4%      +1.4%       7.20 ±  5%  slabinfo.bio-120.num_slabs
    537.60 ± 22%      +3.1%     554.40 ±  8%  slabinfo.bio-184.active_objs
     12.80 ± 22%      +3.1%      13.20 ±  8%  slabinfo.bio-184.active_slabs
    537.60 ± 22%      +3.1%     554.40 ±  8%  slabinfo.bio-184.num_objs
     12.80 ± 22%      +3.1%      13.20 ±  8%  slabinfo.bio-184.num_slabs
    128.00            +0.0%     128.00        slabinfo.bio-248.active_objs
      2.00            +0.0%       2.00        slabinfo.bio-248.active_slabs
    128.00            +0.0%     128.00        slabinfo.bio-248.num_objs
      2.00            +0.0%       2.00        slabinfo.bio-248.num_slabs
     51.00            +0.0%      51.00        slabinfo.bio-296.active_objs
      1.00            +0.0%       1.00        slabinfo.bio-296.active_slabs
     51.00            +0.0%      51.00        slabinfo.bio-296.num_objs
      1.00            +0.0%       1.00        slabinfo.bio-296.num_slabs
    121.80 ± 10%      +3.4%     126.00        slabinfo.bio-360.active_objs
      2.90 ± 10%      +3.4%       3.00        slabinfo.bio-360.active_slabs
    121.80 ± 10%      +3.4%     126.00        slabinfo.bio-360.num_objs
      2.90 ± 10%      +3.4%       3.00        slabinfo.bio-360.num_slabs
     42.00            +0.0%      42.00        slabinfo.bio-376.active_objs
      1.00            +0.0%       1.00        slabinfo.bio-376.active_slabs
     42.00            +0.0%      42.00        slabinfo.bio-376.num_objs
      1.00            +0.0%       1.00        slabinfo.bio-376.num_slabs
     36.00            +0.0%      36.00        slabinfo.bio-432.active_objs
      1.00            +0.0%       1.00        slabinfo.bio-432.active_slabs
     36.00            +0.0%      36.00        slabinfo.bio-432.num_objs
      1.00            +0.0%       1.00        slabinfo.bio-432.num_slabs
    170.00            +0.0%     170.00        slabinfo.bio_post_read_ctx.active_objs
      2.00            +0.0%       2.00        slabinfo.bio_post_read_ctx.active_slabs
    170.00            +0.0%     170.00        slabinfo.bio_post_read_ctx.num_objs
      2.00            +0.0%       2.00        slabinfo.bio_post_read_ctx.num_slabs
     33.60 ± 14%      +9.5%      36.80 ± 19%  slabinfo.biovec-128.active_objs
      2.10 ± 14%      +9.5%       2.30 ± 19%  slabinfo.biovec-128.active_slabs
     33.60 ± 14%      +9.5%      36.80 ± 19%  slabinfo.biovec-128.num_objs
      2.10 ± 14%      +9.5%       2.30 ± 19%  slabinfo.biovec-128.num_slabs
    300.80 ±  9%      +7.4%     323.20 ± 12%  slabinfo.biovec-64.active_objs
      9.40 ±  9%      +7.4%      10.10 ± 12%  slabinfo.biovec-64.active_slabs
    300.80 ±  9%      +7.4%     323.20 ± 12%  slabinfo.biovec-64.num_objs
      9.40 ±  9%      +7.4%      10.10 ± 12%  slabinfo.biovec-64.num_slabs
     56.00            +0.0%      56.00        slabinfo.biovec-max.active_objs
      7.00            +0.0%       7.00        slabinfo.biovec-max.active_slabs
     56.00            +0.0%      56.00        slabinfo.biovec-max.num_objs
      7.00            +0.0%       7.00        slabinfo.biovec-max.num_slabs
    136.00            +0.0%     136.00        slabinfo.btrfs_extent_buffer.active_objs
      2.00            +0.0%       2.00        slabinfo.btrfs_extent_buffer.active_slabs
    136.00            +0.0%     136.00        slabinfo.btrfs_extent_buffer.num_objs
      2.00            +0.0%       2.00        slabinfo.btrfs_extent_buffer.num_slabs
      0.00       +7.8e+102%       7.80 ±200%  slabinfo.btrfs_free_space.active_objs
      0.00         +2e+101%       0.20 ±200%  slabinfo.btrfs_free_space.active_slabs
      0.00       +7.8e+102%       7.80 ±200%  slabinfo.btrfs_free_space.num_objs
      0.00         +2e+101%       0.20 ±200%  slabinfo.btrfs_free_space.num_slabs
    104.40 ± 13%     -11.1%      92.80 ± 12%  slabinfo.btrfs_inode.active_objs
      3.60 ± 13%     -11.1%       3.20 ± 12%  slabinfo.btrfs_inode.active_slabs
    104.40 ± 13%     -11.1%      92.80 ± 12%  slabinfo.btrfs_inode.num_objs
      3.60 ± 13%     -11.1%       3.20 ± 12%  slabinfo.btrfs_inode.num_slabs
    211.85 ±  5%      +1.7%     215.47        slabinfo.btrfs_path.active_objs
      5.88 ±  5%      +1.7%       5.99        slabinfo.btrfs_path.active_slabs
    211.85 ±  5%      +1.7%     215.47        slabinfo.btrfs_path.num_objs
      5.88 ±  5%      +1.7%       5.99        slabinfo.btrfs_path.num_slabs
    156.00           +15.0%     179.40 ± 10%  slabinfo.buffer_head.active_objs
      4.00           +15.0%       4.60 ± 10%  slabinfo.buffer_head.active_slabs
    156.00           +15.0%     179.40 ± 10%  slabinfo.buffer_head.num_objs
      4.00           +15.0%       4.60 ± 10%  slabinfo.buffer_head.num_slabs
     14481            +0.1%      14494        slabinfo.cred_jar.active_objs
    344.80            +0.1%     345.10        slabinfo.cred_jar.active_slabs
     14481            +0.1%      14494        slabinfo.cred_jar.num_objs
    344.80            +0.1%     345.10        slabinfo.cred_jar.num_slabs
     39.00            +0.0%      39.00        slabinfo.dax_cache.active_objs
      1.00            +0.0%       1.00        slabinfo.dax_cache.active_slabs
     39.00            +0.0%      39.00        slabinfo.dax_cache.num_objs
      1.00            +0.0%       1.00        slabinfo.dax_cache.num_slabs
    165719            +0.4%     166422        slabinfo.dentry.active_objs
      3967            +0.6%       3990        slabinfo.dentry.active_slabs
    166647            +0.6%     167586        slabinfo.dentry.num_objs
      3967            +0.6%       3990        slabinfo.dentry.num_slabs
     30.00            +0.0%      30.00        slabinfo.dmaengine-unmap-128.active_objs
      1.00            +0.0%       1.00        slabinfo.dmaengine-unmap-128.active_slabs
     30.00            +0.0%      30.00        slabinfo.dmaengine-unmap-128.num_objs
      1.00            +0.0%       1.00        slabinfo.dmaengine-unmap-128.num_slabs
     64.00            +0.0%      64.00        slabinfo.dmaengine-unmap-2.active_objs
      1.00            +0.0%       1.00        slabinfo.dmaengine-unmap-2.active_slabs
     64.00            +0.0%      64.00        slabinfo.dmaengine-unmap-2.num_objs
      1.00            +0.0%       1.00        slabinfo.dmaengine-unmap-2.num_slabs
     15.00            +0.0%      15.00        slabinfo.dmaengine-unmap-256.active_objs
      1.00            +0.0%       1.00        slabinfo.dmaengine-unmap-256.active_slabs
     15.00            +0.0%      15.00        slabinfo.dmaengine-unmap-256.num_objs
      1.00            +0.0%       1.00        slabinfo.dmaengine-unmap-256.num_slabs
     24640 ± 10%      -1.5%      24268 ±  6%  slabinfo.ep_head.active_objs
     96.25 ± 10%      -1.5%      94.80 ±  6%  slabinfo.ep_head.active_slabs
     24640 ± 10%      -1.5%      24268 ±  6%  slabinfo.ep_head.num_objs
     96.25 ± 10%      -1.5%      94.80 ±  6%  slabinfo.ep_head.num_slabs
    706.56 ±  7%      +1.7%     718.75 ±  9%  slabinfo.file_lock_cache.active_objs
     19.10 ±  7%      +1.7%      19.43 ±  9%  slabinfo.file_lock_cache.active_slabs
    706.56 ±  7%      +1.7%     718.75 ±  9%  slabinfo.file_lock_cache.num_objs
     19.10 ±  7%      +1.7%      19.43 ±  9%  slabinfo.file_lock_cache.num_slabs
      9531            +1.5%       9673 ±  2%  slabinfo.files_cache.active_objs
    207.21            +1.5%     210.29 ±  2%  slabinfo.files_cache.active_slabs
      9531            +1.5%       9673 ±  2%  slabinfo.files_cache.num_objs
    207.21            +1.5%     210.29 ±  2%  slabinfo.files_cache.num_slabs
     40884            +0.5%      41085        slabinfo.filp.active_objs
    672.40            +0.2%     674.07 ±  3%  slabinfo.filp.active_slabs
     43033            +0.2%      43140 ±  3%  slabinfo.filp.num_objs
    672.40            +0.2%     674.07 ±  3%  slabinfo.filp.num_slabs
      2316 ±  7%      +2.8%       2380 ±  8%  slabinfo.fsnotify_mark_connector.active_objs
     18.10 ±  7%      +2.8%      18.60 ±  8%  slabinfo.fsnotify_mark_connector.active_slabs
      2316 ±  7%      +2.8%       2380 ±  8%  slabinfo.fsnotify_mark_connector.num_objs
     18.10 ±  7%      +2.8%      18.60 ±  8%  slabinfo.fsnotify_mark_connector.num_slabs
      9300 ±  3%      +1.2%       9409 ±  5%  slabinfo.ftrace_event_field.active_objs
    127.40 ±  3%      +1.2%     128.90 ±  5%  slabinfo.ftrace_event_field.active_slabs
      9300 ±  3%      +1.2%       9409 ±  5%  slabinfo.ftrace_event_field.num_objs
    127.40 ±  3%      +1.2%     128.90 ±  5%  slabinfo.ftrace_event_field.num_slabs
     56.00            +0.0%      56.00        slabinfo.fuse_request.active_objs
      1.00            +0.0%       1.00        slabinfo.fuse_request.active_slabs
     56.00            +0.0%      56.00        slabinfo.fuse_request.num_objs
      1.00            +0.0%       1.00        slabinfo.fuse_request.num_slabs
     98.00            +0.0%      98.00        slabinfo.hugetlbfs_inode_cache.active_objs
      2.00            +0.0%       2.00        slabinfo.hugetlbfs_inode_cache.active_slabs
     98.00            +0.0%      98.00        slabinfo.hugetlbfs_inode_cache.num_objs
      2.00            +0.0%       2.00        slabinfo.hugetlbfs_inode_cache.num_slabs
    105824            +0.1%     105909        slabinfo.inode_cache.active_objs
      2077            +0.1%       2079        slabinfo.inode_cache.active_slabs
    105950            +0.1%     106056        slabinfo.inode_cache.num_objs
      2077            +0.1%       2079        slabinfo.inode_cache.num_slabs
    182.50 ± 20%     +16.0%     211.70 ± 10%  slabinfo.ip_fib_alias.active_objs
      2.50 ± 20%     +16.0%       2.90 ± 10%  slabinfo.ip_fib_alias.active_slabs
    182.50 ± 20%     +16.0%     211.70 ± 10%  slabinfo.ip_fib_alias.num_objs
      2.50 ± 20%     +16.0%       2.90 ± 10%  slabinfo.ip_fib_alias.num_slabs
    212.50 ± 20%     +16.0%     246.50 ± 10%  slabinfo.ip_fib_trie.active_objs
      2.50 ± 20%     +16.0%       2.90 ± 10%  slabinfo.ip_fib_trie.active_slabs
    212.50 ± 20%     +16.0%     246.50 ± 10%  slabinfo.ip_fib_trie.num_objs
      2.50 ± 20%     +16.0%       2.90 ± 10%  slabinfo.ip_fib_trie.num_slabs
    168800            +0.0%     168832        slabinfo.kernfs_node_cache.active_objs
      2637            +0.0%       2638        slabinfo.kernfs_node_cache.active_slabs
    168800            +0.0%     168832        slabinfo.kernfs_node_cache.num_objs
      2637            +0.0%       2638        slabinfo.kernfs_node_cache.num_slabs
     21600            -0.1%      21570        slabinfo.khugepaged_mm_slot.active_objs
    211.77            -0.1%     211.48        slabinfo.khugepaged_mm_slot.active_slabs
     21600            -0.1%      21570        slabinfo.khugepaged_mm_slot.num_objs
    211.77            -0.1%     211.48        slabinfo.khugepaged_mm_slot.num_slabs
     20061            +0.5%      20170        slabinfo.kmalloc-128.active_objs
    320.03            +0.0%     320.16        slabinfo.kmalloc-128.active_slabs
     20482            +0.0%      20490        slabinfo.kmalloc-128.num_objs
    320.03            +0.0%     320.16        slabinfo.kmalloc-128.num_slabs
     83404            -0.0%      83399        slabinfo.kmalloc-16.active_objs
    325.80            +0.1%     326.10        slabinfo.kmalloc-16.active_slabs
     83404            +0.1%      83481        slabinfo.kmalloc-16.num_objs
    325.80            +0.1%     326.10        slabinfo.kmalloc-16.num_slabs
     21409            +0.1%      21421        slabinfo.kmalloc-192.active_objs
    509.82            +0.1%     510.14        slabinfo.kmalloc-192.active_slabs
     21412            +0.1%      21425        slabinfo.kmalloc-192.num_objs
    509.82            +0.1%     510.14        slabinfo.kmalloc-192.num_slabs
     10382            -0.2%      10360        slabinfo.kmalloc-1k.active_objs
    325.71            -0.1%     325.33        slabinfo.kmalloc-1k.active_slabs
     10422            -0.1%      10410        slabinfo.kmalloc-1k.num_objs
    325.71            -0.1%     325.33        slabinfo.kmalloc-1k.num_slabs
     20195            +0.2%      20233        slabinfo.kmalloc-256.active_objs
    323.34 ±  2%      +0.2%     323.86        slabinfo.kmalloc-256.active_slabs
     20693 ±  2%      +0.2%      20727        slabinfo.kmalloc-256.num_objs
    323.34 ±  2%      +0.2%     323.86        slabinfo.kmalloc-256.num_slabs
      9041            -0.3%       9015        slabinfo.kmalloc-2k.active_objs
    566.68            -0.2%     565.41        slabinfo.kmalloc-2k.active_slabs
      9066            -0.2%       9046        slabinfo.kmalloc-2k.num_objs
    566.68            -0.2%     565.41        slabinfo.kmalloc-2k.num_slabs
    175334            -0.1%     175183        slabinfo.kmalloc-32.active_objs
      1370            -0.1%       1368        slabinfo.kmalloc-32.active_slabs
    175365            -0.1%     175221        slabinfo.kmalloc-32.num_objs
      1370            -0.1%       1368        slabinfo.kmalloc-32.num_slabs
      3202            -0.6%       3184        slabinfo.kmalloc-4k.active_objs
    400.65            -0.5%     398.60        slabinfo.kmalloc-4k.active_slabs
      3205            -0.5%       3188        slabinfo.kmalloc-4k.num_objs
    400.65            -0.5%     398.60        slabinfo.kmalloc-4k.num_slabs
     33467            +0.5%      33643        slabinfo.kmalloc-512.active_objs
    524.91            +0.4%     527.19        slabinfo.kmalloc-512.active_slabs
     33594            +0.4%      33740        slabinfo.kmalloc-512.num_objs
    524.91            +0.4%     527.19        slabinfo.kmalloc-512.num_slabs
    172450            -0.1%     172322        slabinfo.kmalloc-64.active_objs
      2694            -0.1%       2692        slabinfo.kmalloc-64.active_slabs
    172473            -0.1%     172339        slabinfo.kmalloc-64.num_objs
      2694            -0.1%       2692        slabinfo.kmalloc-64.num_slabs
    138463            +0.5%     139132        slabinfo.kmalloc-8.active_objs
    280.42 ±  2%      -0.1%     280.24        slabinfo.kmalloc-8.active_slabs
    143573 ±  2%      -0.1%     143480        slabinfo.kmalloc-8.num_objs
    280.42 ±  2%      -0.1%     280.24        slabinfo.kmalloc-8.num_slabs
      1429            -0.1%       1427        slabinfo.kmalloc-8k.active_objs
    357.53            -0.1%     357.17        slabinfo.kmalloc-8k.active_slabs
      1430            -0.1%       1428        slabinfo.kmalloc-8k.num_objs
    357.53            -0.1%     357.17        slabinfo.kmalloc-8k.num_slabs
     38336            +1.3%      38843 ±  2%  slabinfo.kmalloc-96.active_objs
    932.92            +1.5%     946.90 ±  2%  slabinfo.kmalloc-96.active_slabs
     39182            +1.5%      39769 ±  2%  slabinfo.kmalloc-96.num_objs
    932.92            +1.5%     946.90 ±  2%  slabinfo.kmalloc-96.num_slabs
      1542 ±  2%      -0.4%       1536 ±  5%  slabinfo.kmalloc-cg-128.active_objs
     24.10 ±  2%      -0.4%      24.00 ±  5%  slabinfo.kmalloc-cg-128.active_slabs
      1542 ±  2%      -0.4%       1536 ±  5%  slabinfo.kmalloc-cg-128.num_objs
     24.10 ±  2%      -0.4%      24.00 ±  5%  slabinfo.kmalloc-cg-128.num_slabs
      4325 ±  2%      -1.7%       4252 ±  4%  slabinfo.kmalloc-cg-16.active_objs
     16.90 ±  2%      -1.7%      16.61 ±  4%  slabinfo.kmalloc-cg-16.active_slabs
      4325 ±  2%      -1.7%       4252 ±  4%  slabinfo.kmalloc-cg-16.num_objs
     16.90 ±  2%      -1.7%      16.61 ±  4%  slabinfo.kmalloc-cg-16.num_slabs
      8798 ±  2%      +2.3%       9004        slabinfo.kmalloc-cg-192.active_objs
    209.49 ±  2%      +2.3%     214.40        slabinfo.kmalloc-cg-192.active_slabs
      8798 ±  2%      +2.3%       9004        slabinfo.kmalloc-cg-192.num_objs
    209.49 ±  2%      +2.3%     214.40        slabinfo.kmalloc-cg-192.num_slabs
      6641 ±  3%      +2.4%       6802 ±  3%  slabinfo.kmalloc-cg-1k.active_objs
    207.56 ±  3%      +2.4%     212.58 ±  3%  slabinfo.kmalloc-cg-1k.active_slabs
      6641 ±  3%      +2.4%       6802 ±  3%  slabinfo.kmalloc-cg-1k.num_objs
    207.56 ±  3%      +2.4%     212.58 ±  3%  slabinfo.kmalloc-cg-1k.num_slabs
      1472 ±  4%      -6.1%       1382 ±  5%  slabinfo.kmalloc-cg-256.active_objs
     23.00 ±  4%      -6.1%      21.60 ±  5%  slabinfo.kmalloc-cg-256.active_slabs
      1472 ±  4%      -6.1%       1382 ±  5%  slabinfo.kmalloc-cg-256.num_objs
     23.00 ±  4%      -6.1%      21.60 ±  5%  slabinfo.kmalloc-cg-256.num_slabs
      1018 ±  3%      +2.2%       1040 ±  4%  slabinfo.kmalloc-cg-2k.active_objs
     63.66 ±  3%      +2.2%      65.06 ±  4%  slabinfo.kmalloc-cg-2k.active_slabs
      1018 ±  3%      +2.2%       1040 ±  4%  slabinfo.kmalloc-cg-2k.num_objs
     63.66 ±  3%      +2.2%      65.06 ±  4%  slabinfo.kmalloc-cg-2k.num_slabs
     29379            +0.1%      29422        slabinfo.kmalloc-cg-32.active_objs
    229.52            +0.1%     229.87        slabinfo.kmalloc-cg-32.active_slabs
     29379            +0.1%      29422        slabinfo.kmalloc-cg-32.num_objs
    229.52            +0.1%     229.87        slabinfo.kmalloc-cg-32.num_slabs
      2697            -0.5%       2683        slabinfo.kmalloc-cg-4k.active_objs
    340.10            -0.6%     338.10        slabinfo.kmalloc-cg-4k.active_slabs
      2720            -0.6%       2704        slabinfo.kmalloc-cg-4k.num_objs
    340.10            -0.6%     338.10        slabinfo.kmalloc-cg-4k.num_slabs
     14431            -0.1%      14419        slabinfo.kmalloc-cg-512.active_objs
    225.50            -0.1%     225.30        slabinfo.kmalloc-cg-512.active_slabs
     14431            -0.1%      14419        slabinfo.kmalloc-cg-512.num_objs
    225.50            -0.1%     225.30        slabinfo.kmalloc-cg-512.num_slabs
      3134 ±  4%      -0.2%       3127 ±  7%  slabinfo.kmalloc-cg-64.active_objs
     48.97 ±  4%      -0.2%      48.87 ±  7%  slabinfo.kmalloc-cg-64.active_slabs
      3134 ±  4%      -0.2%       3127 ±  7%  slabinfo.kmalloc-cg-64.num_objs
     48.97 ±  4%      -0.2%      48.87 ±  7%  slabinfo.kmalloc-cg-64.num_slabs
    115595            +0.1%     115732        slabinfo.kmalloc-cg-8.active_objs
    225.77            +0.1%     226.04        slabinfo.kmalloc-cg-8.active_slabs
    115595            +0.1%     115732        slabinfo.kmalloc-cg-8.num_objs
    225.77            +0.1%     226.04        slabinfo.kmalloc-cg-8.num_slabs
     50.61 ±  3%      +0.3%      50.74 ±  3%  slabinfo.kmalloc-cg-8k.active_objs
     12.65 ±  3%      +0.3%      12.68 ±  3%  slabinfo.kmalloc-cg-8k.active_slabs
     50.61 ±  3%      +0.3%      50.74 ±  3%  slabinfo.kmalloc-cg-8k.num_objs
     12.65 ±  3%      +0.3%      12.68 ±  3%  slabinfo.kmalloc-cg-8k.num_slabs
      1931 ±  4%      +2.8%       1986 ±  7%  slabinfo.kmalloc-cg-96.active_objs
     46.00 ±  4%      +2.8%      47.30 ±  7%  slabinfo.kmalloc-cg-96.active_slabs
      1931 ±  4%      +2.8%       1986 ±  7%  slabinfo.kmalloc-cg-96.num_objs
     46.00 ±  4%      +2.8%      47.30 ±  7%  slabinfo.kmalloc-cg-96.num_slabs
    300.80 ± 13%      +2.1%     307.20 ±  8%  slabinfo.kmalloc-rcl-128.active_objs
      4.70 ± 13%      +2.1%       4.80 ±  8%  slabinfo.kmalloc-rcl-128.active_slabs
    300.80 ± 13%      +2.1%     307.20 ±  8%  slabinfo.kmalloc-rcl-128.num_objs
      4.70 ± 13%      +2.1%       4.80 ±  8%  slabinfo.kmalloc-rcl-128.num_slabs
     42.00            +0.0%      42.00        slabinfo.kmalloc-rcl-192.active_objs
      1.00            +0.0%       1.00        slabinfo.kmalloc-rcl-192.active_slabs
     42.00            +0.0%      42.00        slabinfo.kmalloc-rcl-192.num_objs
      1.00            +0.0%       1.00        slabinfo.kmalloc-rcl-192.num_slabs
      8078 ±  3%      -0.7%       8020 ±  4%  slabinfo.kmalloc-rcl-64.active_objs
    126.36 ±  3%      -0.7%     125.46 ±  3%  slabinfo.kmalloc-rcl-64.active_slabs
      8087 ±  3%      -0.7%       8029 ±  3%  slabinfo.kmalloc-rcl-64.num_objs
    126.36 ±  3%      -0.7%     125.46 ±  3%  slabinfo.kmalloc-rcl-64.num_slabs
      1247 ±  7%      +0.0%       1247 ±  7%  slabinfo.kmalloc-rcl-96.active_objs
     29.70 ±  7%      +0.0%      29.70 ±  7%  slabinfo.kmalloc-rcl-96.active_slabs
      1247 ±  7%      +0.0%       1247 ±  7%  slabinfo.kmalloc-rcl-96.num_objs
     29.70 ±  7%      +0.0%      29.70 ±  7%  slabinfo.kmalloc-rcl-96.num_slabs
      1113 ±  6%      -9.2%       1011 ± 10%  slabinfo.kmem_cache.active_objs
     17.40 ±  6%      -9.2%      15.80 ± 10%  slabinfo.kmem_cache.active_slabs
      1113 ±  6%      -9.2%       1011 ± 10%  slabinfo.kmem_cache.num_objs
     17.40 ±  6%      -9.2%      15.80 ± 10%  slabinfo.kmem_cache.num_slabs
      1297 ±  5%      -7.4%       1201 ±  8%  slabinfo.kmem_cache_node.active_objs
     20.30 ±  5%      -7.4%      18.80 ±  8%  slabinfo.kmem_cache_node.active_slabs
      1299 ±  5%      -7.4%       1203 ±  8%  slabinfo.kmem_cache_node.num_objs
     20.30 ±  5%      -7.4%      18.80 ±  8%  slabinfo.kmem_cache_node.num_slabs
     44655            -0.2%      44587        slabinfo.lsm_file_cache.active_objs
    263.38            -0.1%     263.16        slabinfo.lsm_file_cache.active_slabs
     44775            -0.1%      44737        slabinfo.lsm_file_cache.num_objs
    263.38            -0.1%     263.16        slabinfo.lsm_file_cache.num_slabs
     48621 ±  2%      +7.3%      52173 ±  4%  slabinfo.maple_node.active_objs
    772.36 ±  2%      +8.3%     836.19 ±  3%  slabinfo.maple_node.active_slabs
     49431 ±  2%      +8.3%      53516 ±  3%  slabinfo.maple_node.num_objs
    772.36 ±  2%      +8.3%     836.19 ±  3%  slabinfo.maple_node.num_slabs
      5734            +1.2%       5801        slabinfo.mm_struct.active_objs
    238.95            +1.2%     241.73        slabinfo.mm_struct.active_slabs
      5734            +1.2%       5801        slabinfo.mm_struct.num_objs
    238.95            +1.2%     241.73        slabinfo.mm_struct.num_slabs
      1116 ±  6%      -0.5%       1111 ±  5%  slabinfo.mnt_cache.active_objs
     21.90 ±  6%      -0.5%      21.80 ±  5%  slabinfo.mnt_cache.active_slabs
      1116 ±  6%      -0.5%       1111 ±  5%  slabinfo.mnt_cache.num_objs
     21.90 ±  6%      -0.5%      21.80 ±  5%  slabinfo.mnt_cache.num_slabs
     34.00            +0.0%      34.00        slabinfo.mqueue_inode_cache.active_objs
      1.00            +0.0%       1.00        slabinfo.mqueue_inode_cache.active_slabs
     34.00            +0.0%      34.00        slabinfo.mqueue_inode_cache.num_objs
      1.00            +0.0%       1.00        slabinfo.mqueue_inode_cache.num_slabs
      1792            +0.2%       1795        slabinfo.names_cache.active_objs
    224.00            +0.2%     224.40        slabinfo.names_cache.active_slabs
      1792            +0.2%       1795        slabinfo.names_cache.num_objs
    224.00            +0.2%     224.40        slabinfo.names_cache.num_slabs
      7.00            +0.0%       7.00        slabinfo.net_namespace.active_objs
      1.00            +0.0%       1.00        slabinfo.net_namespace.active_slabs
      7.00            +0.0%       7.00        slabinfo.net_namespace.num_objs
      1.00            +0.0%       1.00        slabinfo.net_namespace.num_slabs
     46.00            +0.0%      46.00        slabinfo.nfs_commit_data.active_objs
      1.00            +0.0%       1.00        slabinfo.nfs_commit_data.active_slabs
     46.00            +0.0%      46.00        slabinfo.nfs_commit_data.num_objs
      1.00            +0.0%       1.00        slabinfo.nfs_commit_data.num_slabs
     36.00            +0.0%      36.00        slabinfo.nfs_read_data.active_objs
      1.00            +0.0%       1.00        slabinfo.nfs_read_data.active_slabs
     36.00            +0.0%      36.00        slabinfo.nfs_read_data.num_objs
      1.00            +0.0%       1.00        slabinfo.nfs_read_data.num_slabs
    292.96 ±  8%      -0.9%     290.20 ±  8%  slabinfo.nsproxy.active_objs
      5.23 ±  8%      -0.9%       5.18 ±  8%  slabinfo.nsproxy.active_slabs
    292.96 ±  8%      -0.9%     290.20 ±  8%  slabinfo.nsproxy.num_objs
      5.23 ±  8%      -0.9%       5.18 ±  8%  slabinfo.nsproxy.num_slabs
    240.00            +0.0%     240.00        slabinfo.numa_policy.active_objs
      4.00            +0.0%       4.00        slabinfo.numa_policy.active_slabs
    240.00            +0.0%     240.00        slabinfo.numa_policy.num_objs
      4.00            +0.0%       4.00        slabinfo.numa_policy.num_slabs
     10075            -0.5%      10024        slabinfo.perf_event.active_objs
    404.18            -0.4%     402.52        slabinfo.perf_event.active_slabs
     10104            -0.4%      10062        slabinfo.perf_event.num_objs
    404.18            -0.4%     402.52        slabinfo.perf_event.num_slabs
     16201            -0.7%      16089        slabinfo.pid.active_objs
    253.20            -0.7%     251.40        slabinfo.pid.active_slabs
     16204            -0.7%      16089        slabinfo.pid.num_objs
    253.20            -0.7%     251.40        slabinfo.pid.num_slabs
     13300            -1.0%      13163        slabinfo.pool_workqueue.active_objs
    208.12            -1.1%     205.93        slabinfo.pool_workqueue.active_slabs
     13319            -1.1%      13179        slabinfo.pool_workqueue.num_objs
    208.12            -1.1%     205.93        slabinfo.pool_workqueue.num_slabs
      4792            -1.1%       4741        slabinfo.proc_dir_entry.active_objs
    114.10            -1.1%     112.90        slabinfo.proc_dir_entry.active_slabs
      4792            -1.1%       4741        slabinfo.proc_dir_entry.num_objs
    114.10            -1.1%     112.90        slabinfo.proc_dir_entry.num_slabs
     29409            +0.4%      29525        slabinfo.proc_inode_cache.active_objs
    639.60            +0.4%     642.26        slabinfo.proc_inode_cache.active_slabs
     29421            +0.4%      29544        slabinfo.proc_inode_cache.num_objs
    639.60            +0.4%     642.26        slabinfo.proc_inode_cache.num_slabs
     40983            +0.1%      41043        slabinfo.radix_tree_node.active_objs
    732.66            +0.2%     733.78        slabinfo.radix_tree_node.active_slabs
     41029            +0.2%      41091        slabinfo.radix_tree_node.num_objs
    732.66            +0.2%     733.78        slabinfo.radix_tree_node.num_slabs
    294.00 ± 18%      +8.3%     318.50 ± 18%  slabinfo.request_queue.active_objs
     10.20 ±  9%      +4.9%      10.70 ± 16%  slabinfo.request_queue.active_slabs
    357.00 ±  9%      +4.9%     374.50 ± 16%  slabinfo.request_queue.num_objs
     10.20 ±  9%      +4.9%      10.70 ± 16%  slabinfo.request_queue.num_slabs
     46.00            +0.0%      46.00        slabinfo.rpc_inode_cache.active_objs
      1.00            +0.0%       1.00        slabinfo.rpc_inode_cache.active_slabs
     46.00            +0.0%      46.00        slabinfo.rpc_inode_cache.num_objs
      1.00            +0.0%       1.00        slabinfo.rpc_inode_cache.num_slabs
      4812            +0.3%       4825        slabinfo.scsi_sense_cache.active_objs
     76.20            +0.3%      76.40        slabinfo.scsi_sense_cache.active_slabs
      4876            +0.3%       4889        slabinfo.scsi_sense_cache.num_objs
     76.20            +0.3%      76.40        slabinfo.scsi_sense_cache.num_slabs
     15979            -0.1%      15967        slabinfo.seq_file.active_objs
    234.99            -0.1%     234.81        slabinfo.seq_file.active_slabs
     15979            -0.1%      15967        slabinfo.seq_file.num_objs
    234.99            -0.1%     234.81        slabinfo.seq_file.num_slabs
    101821            +0.2%     101974        slabinfo.shared_policy_node.active_objs
      1197            +0.2%       1199        slabinfo.shared_policy_node.active_slabs
    101821            +0.2%     101974        slabinfo.shared_policy_node.num_objs
      1197            +0.2%       1199        slabinfo.shared_policy_node.num_slabs
      8779            +0.4%       8816 ±  3%  slabinfo.shmem_inode_cache.active_objs
    204.19            +0.4%     205.04 ±  3%  slabinfo.shmem_inode_cache.active_slabs
      8779            +0.4%       8816 ±  3%  slabinfo.shmem_inode_cache.num_objs
    204.19            +0.4%     205.04 ±  3%  slabinfo.shmem_inode_cache.num_slabs
      5473            +0.8%       5519        slabinfo.sighand_cache.active_objs
    365.02            +0.9%     368.23        slabinfo.sighand_cache.active_slabs
      5475            +0.9%       5523        slabinfo.sighand_cache.num_objs
    365.02            +0.9%     368.23        slabinfo.sighand_cache.num_slabs
      8990            +0.1%       8995 ±  2%  slabinfo.signal_cache.active_objs
    321.28            +0.1%     321.61 ±  2%  slabinfo.signal_cache.active_slabs
      8995            +0.1%       9005 ±  2%  slabinfo.signal_cache.num_objs
    321.28            +0.1%     321.61 ±  2%  slabinfo.signal_cache.num_slabs
     11923            +0.2%      11946        slabinfo.sigqueue.active_objs
    233.79            +0.2%     234.24        slabinfo.sigqueue.active_slabs
     11923            +0.2%      11946        slabinfo.sigqueue.num_objs
    233.79            +0.2%     234.24        slabinfo.sigqueue.num_slabs
    510.09 ±  8%      +3.2%     526.37 ±  6%  slabinfo.skbuff_ext_cache.active_objs
     12.15 ±  8%      +3.2%      12.53 ±  6%  slabinfo.skbuff_ext_cache.active_slabs
    510.09 ±  8%      +3.2%     526.37 ±  6%  slabinfo.skbuff_ext_cache.num_objs
     12.15 ±  8%      +3.2%      12.53 ±  6%  slabinfo.skbuff_ext_cache.num_slabs
     15648            -0.4%      15590        slabinfo.skbuff_head_cache.active_objs
    245.30            -0.4%     244.20        slabinfo.skbuff_head_cache.active_slabs
     15699            -0.4%      15628        slabinfo.skbuff_head_cache.num_objs
    245.30            -0.4%     244.20        slabinfo.skbuff_head_cache.num_slabs
      8653 ±  5%      -1.9%       8491 ±  5%  slabinfo.skbuff_small_head.active_objs
    169.68 ±  5%      -1.9%     166.50 ±  5%  slabinfo.skbuff_small_head.active_slabs
      8653 ±  5%      -1.9%       8491 ±  5%  slabinfo.skbuff_small_head.num_objs
    169.68 ±  5%      -1.9%     166.50 ±  5%  slabinfo.skbuff_small_head.num_slabs
      4352 ±  5%      -2.8%       4228 ±  5%  slabinfo.sock_inode_cache.active_objs
    111.59 ±  5%      -2.8%     108.42 ±  5%  slabinfo.sock_inode_cache.active_slabs
      4352 ±  5%      -2.8%       4228 ±  5%  slabinfo.sock_inode_cache.num_objs
    111.59 ±  5%      -2.8%     108.42 ±  5%  slabinfo.sock_inode_cache.num_slabs
      1634 ±  6%      +4.9%       1714 ±  5%  slabinfo.task_group.active_objs
     32.05 ±  6%      +4.9%      33.62 ±  5%  slabinfo.task_group.active_slabs
      1634 ±  6%      +4.9%       1714 ±  5%  slabinfo.task_group.num_objs
     32.05 ±  6%      +4.9%      33.62 ±  5%  slabinfo.task_group.num_slabs
      3787            -0.3%       3777        slabinfo.task_struct.active_objs
      3790            -0.3%       3779        slabinfo.task_struct.active_slabs
      3790            -0.3%       3779        slabinfo.task_struct.num_objs
      3790            -0.3%       3779        slabinfo.task_struct.num_slabs
    262.17 ±  8%      -1.9%     257.14 ±  9%  slabinfo.taskstats.active_objs
      7.09 ±  8%      -1.9%       6.95 ±  9%  slabinfo.taskstats.active_slabs
    262.17 ±  8%      -1.9%     257.14 ±  9%  slabinfo.taskstats.num_objs
      7.09 ±  8%      -1.9%       6.95 ±  9%  slabinfo.taskstats.num_slabs
      2350            -0.4%       2341        slabinfo.trace_event_file.active_objs
     51.10            -0.4%      50.90        slabinfo.trace_event_file.active_slabs
      2350            -0.4%       2341        slabinfo.trace_event_file.num_objs
     51.10            -0.4%      50.90        slabinfo.trace_event_file.num_slabs
      2394            +0.3%       2400        slabinfo.tracefs_inode_cache.active_objs
     47.89            +0.3%      48.02        slabinfo.tracefs_inode_cache.active_slabs
      2394            +0.3%       2400        slabinfo.tracefs_inode_cache.num_objs
     47.89            +0.3%      48.02        slabinfo.tracefs_inode_cache.num_slabs
     60.00            +0.0%      60.00        slabinfo.tw_sock_TCP.active_objs
      1.00            +0.0%       1.00        slabinfo.tw_sock_TCP.active_slabs
     60.00            +0.0%      60.00        slabinfo.tw_sock_TCP.num_objs
      1.00            +0.0%       1.00        slabinfo.tw_sock_TCP.num_slabs
     74.00            +0.0%      74.00        slabinfo.uts_namespace.active_objs
      2.00            +0.0%       2.00        slabinfo.uts_namespace.active_slabs
     74.00            +0.0%      74.00        slabinfo.uts_namespace.num_objs
      2.00            +0.0%       2.00        slabinfo.uts_namespace.num_slabs
     85556            -2.1%      83786 ±  4%  slabinfo.vm_area_struct.active_objs
      1952            -1.8%       1916 ±  4%  slabinfo.vm_area_struct.active_slabs
     85920            -1.8%      84334 ±  4%  slabinfo.vm_area_struct.num_objs
      1952            -1.8%       1916 ±  4%  slabinfo.vm_area_struct.num_slabs
     98458            -0.5%      97931 ±  2%  slabinfo.vma_lock.active_objs
    969.74            -0.4%     965.79 ±  2%  slabinfo.vma_lock.active_slabs
     98913            -0.4%      98510 ±  2%  slabinfo.vma_lock.num_objs
    969.74            -0.4%     965.79 ±  2%  slabinfo.vma_lock.num_slabs
    811570            +0.0%     811625        slabinfo.vmap_area.active_objs
     14497            +0.0%      14499        slabinfo.vmap_area.active_slabs
    811855            +0.0%     811960        slabinfo.vmap_area.num_objs
     14497            +0.0%      14499        slabinfo.vmap_area.num_slabs
    276.34            +0.1%     276.48        slabinfo.x86_lbr.active_objs
     13.95            +0.0%      13.96        slabinfo.x86_lbr.active_slabs
    293.03            +0.0%     293.07        slabinfo.x86_lbr.num_objs
     13.95            +0.0%      13.96        slabinfo.x86_lbr.num_slabs
      0.00       +1.4e+100%       0.01 ±146%  perf-sched.sch_delay.avg.ms.__cond_resched.__alloc_pages.__folio_alloc.vma_alloc_folio.do_anonymous_page
      0.00       +4.1e+101%       0.41 ±300%  perf-sched.sch_delay.avg.ms.__cond_resched.__alloc_pages.__pud_alloc.__handle_mm_fault.handle_mm_fault
      0.00          +7e+98%       0.00 ±300%  perf-sched.sch_delay.avg.ms.__cond_resched.__alloc_pages.pte_alloc_one.__pte_alloc.move_page_tables
      0.01 ±300%    -100.0%       0.00        perf-sched.sch_delay.avg.ms.__cond_resched.__flush_work.isra.0.__lru_add_drain_all
      0.00 ±300%    -100.0%       0.00        perf-sched.sch_delay.avg.ms.__cond_resched.__kmem_cache_alloc_node.__kmalloc.acpi_ex_allocate_name_string.acpi_ex_get_name_string
      0.00          +7e+98%       0.00 ±300%  perf-sched.sch_delay.avg.ms.__cond_resched.__kmem_cache_alloc_node.__kmalloc.load_elf_phdrs.load_elf_binary
      0.00          +8e+99%       0.01 ±300%  perf-sched.sch_delay.avg.ms.__cond_resched.__kmem_cache_alloc_node.__kmalloc_node.memcg_alloc_slab_cgroups.allocate_slab
      0.00 ±300%    -100.0%       0.00        perf-sched.sch_delay.avg.ms.__cond_resched.__kmem_cache_alloc_node.kmalloc_node_trace.alloc_fair_sched_group.sched_create_group
      0.00 ±300%    -100.0%       0.00        perf-sched.sch_delay.avg.ms.__cond_resched.__kmem_cache_alloc_node.kmalloc_trace.perf_event_mmap_event.perf_event_mmap
      0.00        +1.4e+99%       0.00 ±300%  perf-sched.sch_delay.avg.ms.__cond_resched.__kmem_cache_alloc_node.kmalloc_trace.vmstat_start.seq_read_iter
      0.00        +1.1e+99%       0.00 ±300%  perf-sched.sch_delay.avg.ms.__cond_resched.__put_anon_vma.unlink_anon_vmas.free_pgtables.exit_mmap
      0.00        +1.4e+99%       0.00 ±300%  perf-sched.sch_delay.avg.ms.__cond_resched.__vmalloc_area_node.__vmalloc_node_range.alloc_thread_stack_node.dup_task_struct
      0.00 ± 96%     +50.0%       0.00 ± 80%  perf-sched.sch_delay.avg.ms.__cond_resched.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr.__sched_setaffinity
      0.18 ±292%     -97.9%       0.00 ±156%  perf-sched.sch_delay.avg.ms.__cond_resched.__wait_for_common.wait_for_completion_state.kernel_clone.__x64_sys_vfork
      0.00 ±300%    -100.0%       0.00        perf-sched.sch_delay.avg.ms.__cond_resched.cgroup_css_set_fork.cgroup_can_fork.copy_process.kernel_clone
      0.00       +7.8e+100%       0.08 ±300%  perf-sched.sch_delay.avg.ms.__cond_resched.change_pmd_range.change_p4d_range.change_protection_range.mprotect_fixup
      0.00          +9e+98%       0.00 ±299%  perf-sched.sch_delay.avg.ms.__cond_resched.copy_strings.isra.0.do_execveat_common
      0.00       +3.7e+101%       0.37 ±299%  perf-sched.sch_delay.avg.ms.__cond_resched.dentry_kill.dput.step_into.link_path_walk
      0.00 ±300%    -100.0%       0.00        perf-sched.sch_delay.avg.ms.__cond_resched.down_read.walk_component.path_lookupat.filename_lookup
      0.00        +1.1e+99%       0.00 ±300%  perf-sched.sch_delay.avg.ms.__cond_resched.down_write.anon_vma_clone.__split_vma.vma_modify
      0.05 ±298%    +730.6%       0.39 ±106%  perf-sched.sch_delay.avg.ms.__cond_resched.down_write_killable.exec_mmap.begin_new_exec.load_elf_binary
      0.00          +7e+98%       0.00 ±299%  perf-sched.sch_delay.avg.ms.__cond_resched.down_write_killable.setup_arg_pages.load_elf_binary.search_binary_handler
      0.00       +5.1e+100%       0.05 ±300%  perf-sched.sch_delay.avg.ms.__cond_resched.dput.__fput.__x64_sys_close.do_syscall_64
      0.00       +2.5e+101%       0.25 ±298%  perf-sched.sch_delay.avg.ms.__cond_resched.dput.step_into.link_path_walk.part
      0.00       +2.7e+101%       0.27 ±300%  perf-sched.sch_delay.avg.ms.__cond_resched.dput.step_into.path_lookupat.filename_lookup
      0.00 ±300%    -100.0%       0.00        perf-sched.sch_delay.avg.ms.__cond_resched.dput.terminate_walk.path_openat.do_filp_open
      0.00          +1e+99%       0.00 ±300%  perf-sched.sch_delay.avg.ms.__cond_resched.exit_mmap.__mmput.exec_mmap.begin_new_exec
      0.00 ±300%    -100.0%       0.00        perf-sched.sch_delay.avg.ms.__cond_resched.filemap_read.__kernel_read.search_binary_handler.exec_binprm
      0.00          +8e+98%       0.00 ±300%  perf-sched.sch_delay.avg.ms.__cond_resched.khugepaged.kthread.ret_from_fork.ret_from_fork_asm
      0.00          +1e+99%       0.00 ±300%  perf-sched.sch_delay.avg.ms.__cond_resched.kmem_cache_alloc.__anon_vma_prepare.do_anonymous_page.__handle_mm_fault
      0.00        +2.7e+99%       0.00 ±299%  perf-sched.sch_delay.avg.ms.__cond_resched.kmem_cache_alloc.acpi_ut_create_generic_state.acpi_ds_result_push.acpi_ds_exec_end_op
      0.00 ±152%    -100.0%       0.00        perf-sched.sch_delay.avg.ms.__cond_resched.kmem_cache_alloc.alloc_pid.copy_process.kernel_clone
      0.00       +2.5e+101%       0.25 ±299%  perf-sched.sch_delay.avg.ms.__cond_resched.kmem_cache_alloc.getname_flags.part.0
      0.05 ±300%    -100.0%       0.00        perf-sched.sch_delay.avg.ms.__cond_resched.kmem_cache_alloc.mas_alloc_nodes.mas_preallocate.__split_vma
      0.00       +4.2e+100%       0.04 ±300%  perf-sched.sch_delay.avg.ms.__cond_resched.kmem_cache_alloc.mas_alloc_nodes.mas_preallocate.vma_link
      0.00 ±200%    -100.0%       0.00        perf-sched.sch_delay.avg.ms.__cond_resched.kmem_cache_alloc.vm_area_alloc.alloc_bprm.do_execveat_common
      0.00        +1.3e+99%       0.00 ±299%  perf-sched.sch_delay.avg.ms.__cond_resched.kmem_cache_alloc.vm_area_dup.dup_mmap.dup_mm
      0.00        +1.6e+99%       0.00 ±300%  perf-sched.sch_delay.avg.ms.__cond_resched.mutex_lock.drm_property_free_blob.drm_gem_destroy_shadow_plane_state.drm_atomic_state_default_clear
      0.00       +5.4e+101%       0.54 ±300%  perf-sched.sch_delay.avg.ms.__cond_resched.mutex_lock_interruptible.devkmsg_read.vfs_read.ksys_read
      0.00       +9.1e+101%       0.91 ±130%  perf-sched.sch_delay.avg.ms.__cond_resched.mutex_lock_killable.pcpu_alloc.mm_init.alloc_bprm
      0.00        +7.9e+99%       0.01 ±133%  perf-sched.sch_delay.avg.ms.__cond_resched.mutex_lock_killable.pcpu_alloc.mm_init.dup_mm
      0.01 ±299%     -65.0%       0.00 ±223%  perf-sched.sch_delay.avg.ms.__cond_resched.process_one_work.worker_thread.kthread.ret_from_fork
      0.00        +2.1e+99%       0.00 ±213%  perf-sched.sch_delay.avg.ms.__cond_resched.refresh_cpu_vm_stats.vmstat_update.process_one_work.worker_thread
      0.30 ±299%    -100.0%       0.00        perf-sched.sch_delay.avg.ms.__cond_resched.remove_vma.do_vmi_align_munmap.do_vmi_munmap.__vm_munmap
      0.00       +8.1e+101%       0.81 ±155%  perf-sched.sch_delay.avg.ms.__cond_resched.shmem_get_folio_gfp.shmem_read_folio_gfp.drm_gem_get_pages.drm_gem_shmem_get_pages
      0.01 ±  3%      -5.8%       0.01 ±  4%  perf-sched.sch_delay.avg.ms.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
      1.77 ±  5%      +4.7%       1.85 ± 10%  perf-sched.sch_delay.avg.ms.__cond_resched.stop_one_cpu.sched_exec.bprm_execve.part
      0.33 ±300%    -100.0%       0.00        perf-sched.sch_delay.avg.ms.__cond_resched.task_work_run.exit_to_user_mode_loop.exit_to_user_mode_prepare.syscall_exit_to_user_mode
      0.00        +2.2e+99%       0.00 ±300%  perf-sched.sch_delay.avg.ms.__cond_resched.task_work_run.get_signal.arch_do_signal_or_restart.exit_to_user_mode_loop
      0.00       +5.9e+100%       0.06 ±300%  perf-sched.sch_delay.avg.ms.__cond_resched.tlb_batch_pages_flush.tlb_finish_mmu.exit_mmap.__mmput
      0.00       +3.1e+100%       0.03 ±201%  perf-sched.sch_delay.avg.ms.__cond_resched.unmap_vmas.unmap_region.constprop.0
      0.31 ±201%    +246.6%       1.06 ±228%  perf-sched.sch_delay.avg.ms.__cond_resched.wp_page_copy.__handle_mm_fault.handle_mm_fault.do_user_addr_fault
      1.67 ±152%     +24.2%       2.08 ±122%  perf-sched.sch_delay.avg.ms.__cond_resched.ww_mutex_lock.drm_gem_vunmap_unlocked.drm_fbdev_generic_helper_fb_dirty.drm_fb_helper_damage_work
      1.06 ±200%     +40.8%       1.50 ±153%  perf-sched.sch_delay.avg.ms.__cond_resched.ww_mutex_lock.drm_gem_vunmap_unlocked.drm_gem_fb_vunmap.drm_atomic_helper_cleanup_planes
      0.00        +4.7e+99%       0.00 ±192%  perf-sched.sch_delay.avg.ms.__cond_resched.zap_pmd_range.isra.0.unmap_page_range
      0.69 ± 83%     -69.5%       0.21 ±172%  perf-sched.sch_delay.avg.ms.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
      0.20 ±300%    -100.0%       0.00        perf-sched.sch_delay.avg.ms.cleaner_kthread.kthread.ret_from_fork.ret_from_fork_asm
      0.84 ± 69%   +1807.1%      15.92 ±108%  perf-sched.sch_delay.avg.ms.devkmsg_read.vfs_read.ksys_read.do_syscall_64
      0.42 ±119%     -48.3%       0.22 ±121%  perf-sched.sch_delay.avg.ms.do_nanosleep.hrtimer_nanosleep.common_nsleep.__x64_sys_clock_nanosleep
      0.14 ± 54%    +235.6%       0.46 ± 31%  perf-sched.sch_delay.avg.ms.do_task_dead.do_exit.do_group_exit.__x64_sys_exit_group.do_syscall_64
      0.16 ± 32%     +37.1%       0.22 ± 20%  perf-sched.sch_delay.avg.ms.do_wait.kernel_wait4.__do_sys_wait4.do_syscall_64
      0.01 ±208%      +6.1%       0.01 ±123%  perf-sched.sch_delay.avg.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_exc_page_fault
      0.02 ± 41%    +249.5%       0.07 ± 68%  perf-sched.sch_delay.avg.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.syscall_exit_to_user_mode.do_syscall_64
      0.00 ±300%    -100.0%       0.00        perf-sched.sch_delay.avg.ms.futex_wait_queue.futex_wait.do_futex.__x64_sys_futex
      0.20 ±300%    -100.0%       0.00        perf-sched.sch_delay.avg.ms.io_schedule.folio_wait_bit_common.filemap_fault.__do_fault
      0.28 ±297%     -99.2%       0.00 ±300%  perf-sched.sch_delay.avg.ms.kthreadd.ret_from_fork.ret_from_fork_asm
      0.04 ± 40%     -28.4%       0.03 ± 28%  perf-sched.sch_delay.avg.ms.pipe_read.vfs_read.ksys_read.do_syscall_64
      0.40 ± 23%     -27.9%       0.29 ± 40%  perf-sched.sch_delay.avg.ms.schedule_hrtimeout_range_clock.do_poll.constprop.0.do_sys_poll
      2.21 ± 49%      +9.3%       2.42 ± 42%  perf-sched.sch_delay.avg.ms.schedule_hrtimeout_range_clock.do_select.core_sys_select.kern_select
      0.25 ± 62%     +79.2%       0.46 ±168%  perf-sched.sch_delay.avg.ms.schedule_hrtimeout_range_clock.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
      0.00 ±202%     -28.6%       0.00 ±300%  perf-sched.sch_delay.avg.ms.schedule_preempt_disabled.kthread.ret_from_fork.ret_from_fork_asm
      0.00 ±300%    -100.0%       0.00        perf-sched.sch_delay.avg.ms.schedule_preempt_disabled.rwsem_down_read_slowpath.down_read_killable.__access_remote_vm
      2.67 ±203%     -30.4%       1.86 ±300%  perf-sched.sch_delay.avg.ms.schedule_timeout.__wait_for_common.wait_for_completion_killable.__kthread_create_on_node
      0.45 ± 23%     +14.3%       0.52 ± 16%  perf-sched.sch_delay.avg.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone
      0.02 ± 25%     -13.2%       0.02 ± 23%  perf-sched.sch_delay.avg.ms.schedule_timeout.kcompactd.kthread.ret_from_fork
      0.02 ± 45%     +11.2%       0.03 ± 39%  perf-sched.sch_delay.avg.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
      0.00 ±299%    -100.0%       0.00        perf-sched.sch_delay.avg.ms.schedule_timeout.transaction_kthread.kthread.ret_from_fork
      0.01           +77.7%       0.02 ± 32%  perf-sched.sch_delay.avg.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
      0.96 ± 77%   +1111.2%      11.68 ± 99%  perf-sched.sch_delay.avg.ms.syslog_print.do_syslog.kmsg_read.vfs_read
      0.12 ±102%    +410.5%       0.62 ± 19%  perf-sched.sch_delay.avg.ms.wait_for_partner.fifo_open.do_dentry_open.do_open
      0.02 ± 10%     +57.8%       0.04 ± 48%  perf-sched.sch_delay.avg.ms.worker_thread.kthread.ret_from_fork.ret_from_fork_asm
      0.00       +6.6e+101%       0.66 ±159%  perf-sched.sch_delay.max.ms.__cond_resched.__alloc_pages.__folio_alloc.vma_alloc_folio.do_anonymous_page
      0.00       +4.1e+101%       0.41 ±300%  perf-sched.sch_delay.max.ms.__cond_resched.__alloc_pages.__pud_alloc.__handle_mm_fault.handle_mm_fault
      0.00          +7e+98%       0.00 ±300%  perf-sched.sch_delay.max.ms.__cond_resched.__alloc_pages.pte_alloc_one.__pte_alloc.move_page_tables
      0.01 ±300%    -100.0%       0.00        perf-sched.sch_delay.max.ms.__cond_resched.__flush_work.isra.0.__lru_add_drain_all
      0.00 ±300%    -100.0%       0.00        perf-sched.sch_delay.max.ms.__cond_resched.__kmem_cache_alloc_node.__kmalloc.acpi_ex_allocate_name_string.acpi_ex_get_name_string
      0.00          +7e+98%       0.00 ±300%  perf-sched.sch_delay.max.ms.__cond_resched.__kmem_cache_alloc_node.__kmalloc.load_elf_phdrs.load_elf_binary
      0.00       +1.6e+100%       0.02 ±300%  perf-sched.sch_delay.max.ms.__cond_resched.__kmem_cache_alloc_node.__kmalloc_node.memcg_alloc_slab_cgroups.allocate_slab
      0.00 ±300%    -100.0%       0.00        perf-sched.sch_delay.max.ms.__cond_resched.__kmem_cache_alloc_node.kmalloc_node_trace.alloc_fair_sched_group.sched_create_group
      0.00 ±300%    -100.0%       0.00        perf-sched.sch_delay.max.ms.__cond_resched.__kmem_cache_alloc_node.kmalloc_trace.perf_event_mmap_event.perf_event_mmap
      0.00        +1.4e+99%       0.00 ±300%  perf-sched.sch_delay.max.ms.__cond_resched.__kmem_cache_alloc_node.kmalloc_trace.vmstat_start.seq_read_iter
      0.00        +1.1e+99%       0.00 ±300%  perf-sched.sch_delay.max.ms.__cond_resched.__put_anon_vma.unlink_anon_vmas.free_pgtables.exit_mmap
      0.00        +1.4e+99%       0.00 ±300%  perf-sched.sch_delay.max.ms.__cond_resched.__vmalloc_area_node.__vmalloc_node_range.alloc_thread_stack_node.dup_task_struct
      1.62 ± 97%     +65.5%       2.69 ± 82%  perf-sched.sch_delay.max.ms.__cond_resched.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr.__sched_setaffinity
      0.18 ±292%     -97.9%       0.00 ±156%  perf-sched.sch_delay.max.ms.__cond_resched.__wait_for_common.wait_for_completion_state.kernel_clone.__x64_sys_vfork
      0.00 ±300%    -100.0%       0.00        perf-sched.sch_delay.max.ms.__cond_resched.cgroup_css_set_fork.cgroup_can_fork.copy_process.kernel_clone
      0.00       +2.3e+101%       0.23 ±300%  perf-sched.sch_delay.max.ms.__cond_resched.change_pmd_range.change_p4d_range.change_protection_range.mprotect_fixup
      0.00          +9e+98%       0.00 ±299%  perf-sched.sch_delay.max.ms.__cond_resched.copy_strings.isra.0.do_execveat_common
      0.00       +3.7e+101%       0.37 ±299%  perf-sched.sch_delay.max.ms.__cond_resched.dentry_kill.dput.step_into.link_path_walk
      0.00 ±300%    -100.0%       0.00        perf-sched.sch_delay.max.ms.__cond_resched.down_read.walk_component.path_lookupat.filename_lookup
      0.00        +1.1e+99%       0.00 ±300%  perf-sched.sch_delay.max.ms.__cond_resched.down_write.anon_vma_clone.__split_vma.vma_modify
      0.19 ±298%    +433.1%       1.00 ± 87%  perf-sched.sch_delay.max.ms.__cond_resched.down_write_killable.exec_mmap.begin_new_exec.load_elf_binary
      0.00          +7e+98%       0.00 ±299%  perf-sched.sch_delay.max.ms.__cond_resched.down_write_killable.setup_arg_pages.load_elf_binary.search_binary_handler
      0.00       +3.6e+101%       0.36 ±299%  perf-sched.sch_delay.max.ms.__cond_resched.dput.__fput.__x64_sys_close.do_syscall_64
      0.00       +2.5e+101%       0.25 ±298%  perf-sched.sch_delay.max.ms.__cond_resched.dput.step_into.link_path_walk.part
      0.00       +2.7e+101%       0.27 ±300%  perf-sched.sch_delay.max.ms.__cond_resched.dput.step_into.path_lookupat.filename_lookup
      0.00 ±300%    -100.0%       0.00        perf-sched.sch_delay.max.ms.__cond_resched.dput.terminate_walk.path_openat.do_filp_open
      0.00          +1e+99%       0.00 ±300%  perf-sched.sch_delay.max.ms.__cond_resched.exit_mmap.__mmput.exec_mmap.begin_new_exec
      0.00 ±300%    -100.0%       0.00        perf-sched.sch_delay.max.ms.__cond_resched.filemap_read.__kernel_read.search_binary_handler.exec_binprm
      0.00        +1.6e+99%       0.00 ±300%  perf-sched.sch_delay.max.ms.__cond_resched.khugepaged.kthread.ret_from_fork.ret_from_fork_asm
      0.00          +1e+99%       0.00 ±300%  perf-sched.sch_delay.max.ms.__cond_resched.kmem_cache_alloc.__anon_vma_prepare.do_anonymous_page.__handle_mm_fault
      0.00        +2.7e+99%       0.00 ±299%  perf-sched.sch_delay.max.ms.__cond_resched.kmem_cache_alloc.acpi_ut_create_generic_state.acpi_ds_result_push.acpi_ds_exec_end_op
      0.00 ±152%    -100.0%       0.00        perf-sched.sch_delay.max.ms.__cond_resched.kmem_cache_alloc.alloc_pid.copy_process.kernel_clone
      0.00       +2.5e+101%       0.25 ±299%  perf-sched.sch_delay.max.ms.__cond_resched.kmem_cache_alloc.getname_flags.part.0
      0.26 ±300%    -100.0%       0.00        perf-sched.sch_delay.max.ms.__cond_resched.kmem_cache_alloc.mas_alloc_nodes.mas_preallocate.__split_vma
      0.00       +4.2e+100%       0.04 ±300%  perf-sched.sch_delay.max.ms.__cond_resched.kmem_cache_alloc.mas_alloc_nodes.mas_preallocate.vma_link
      0.00 ±200%    -100.0%       0.00        perf-sched.sch_delay.max.ms.__cond_resched.kmem_cache_alloc.vm_area_alloc.alloc_bprm.do_execveat_common
      0.00        +1.3e+99%       0.00 ±299%  perf-sched.sch_delay.max.ms.__cond_resched.kmem_cache_alloc.vm_area_dup.dup_mmap.dup_mm
      0.00        +1.6e+99%       0.00 ±300%  perf-sched.sch_delay.max.ms.__cond_resched.mutex_lock.drm_property_free_blob.drm_gem_destroy_shadow_plane_state.drm_atomic_state_default_clear
      0.00       +5.4e+101%       0.54 ±300%  perf-sched.sch_delay.max.ms.__cond_resched.mutex_lock_interruptible.devkmsg_read.vfs_read.ksys_read
      0.00       +1.2e+102%       1.24 ±129%  perf-sched.sch_delay.max.ms.__cond_resched.mutex_lock_killable.pcpu_alloc.mm_init.alloc_bprm
      0.00        +7.9e+99%       0.01 ±133%  perf-sched.sch_delay.max.ms.__cond_resched.mutex_lock_killable.pcpu_alloc.mm_init.dup_mm
      0.01 ±299%     -65.0%       0.00 ±223%  perf-sched.sch_delay.max.ms.__cond_resched.process_one_work.worker_thread.kthread.ret_from_fork
      0.00        +2.1e+99%       0.00 ±213%  perf-sched.sch_delay.max.ms.__cond_resched.refresh_cpu_vm_stats.vmstat_update.process_one_work.worker_thread
      0.30 ±299%    -100.0%       0.00        perf-sched.sch_delay.max.ms.__cond_resched.remove_vma.do_vmi_align_munmap.do_vmi_munmap.__vm_munmap
      0.00       +1.3e+102%       1.30 ±158%  perf-sched.sch_delay.max.ms.__cond_resched.shmem_get_folio_gfp.shmem_read_folio_gfp.drm_gem_get_pages.drm_gem_shmem_get_pages
      0.02 ± 22%      +4.5%       0.02 ± 23%  perf-sched.sch_delay.max.ms.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
      4.01 ±  3%      -0.8%       3.97 ±  2%  perf-sched.sch_delay.max.ms.__cond_resched.stop_one_cpu.sched_exec.bprm_execve.part
      0.33 ±300%    -100.0%       0.00        perf-sched.sch_delay.max.ms.__cond_resched.task_work_run.exit_to_user_mode_loop.exit_to_user_mode_prepare.syscall_exit_to_user_mode
      0.00        +2.2e+99%       0.00 ±300%  perf-sched.sch_delay.max.ms.__cond_resched.task_work_run.get_signal.arch_do_signal_or_restart.exit_to_user_mode_loop
      0.00       +1.2e+101%       0.12 ±300%  perf-sched.sch_delay.max.ms.__cond_resched.tlb_batch_pages_flush.tlb_finish_mmu.exit_mmap.__mmput
      0.00       +5.7e+101%       0.57 ±200%  perf-sched.sch_delay.max.ms.__cond_resched.unmap_vmas.unmap_region.constprop.0
      0.44 ±205%    +141.1%       1.06 ±228%  perf-sched.sch_delay.max.ms.__cond_resched.wp_page_copy.__handle_mm_fault.handle_mm_fault.do_user_addr_fault
      1.67 ±152%     +24.2%       2.08 ±122%  perf-sched.sch_delay.max.ms.__cond_resched.ww_mutex_lock.drm_gem_vunmap_unlocked.drm_fbdev_generic_helper_fb_dirty.drm_fb_helper_damage_work
      1.06 ±200%     +40.8%       1.50 ±153%  perf-sched.sch_delay.max.ms.__cond_resched.ww_mutex_lock.drm_gem_vunmap_unlocked.drm_gem_fb_vunmap.drm_atomic_helper_cleanup_planes
      0.00       +3.7e+100%       0.04 ±191%  perf-sched.sch_delay.max.ms.__cond_resched.zap_pmd_range.isra.0.unmap_page_range
      3.10 ± 59%     -60.7%       1.22 ±161%  perf-sched.sch_delay.max.ms.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
      0.20 ±300%    -100.0%       0.00        perf-sched.sch_delay.max.ms.cleaner_kthread.kthread.ret_from_fork.ret_from_fork_asm
      1.64 ± 70%    +909.5%      16.51 ±101%  perf-sched.sch_delay.max.ms.devkmsg_read.vfs_read.ksys_read.do_syscall_64
      2.08 ±108%     -37.6%       1.30 ±122%  perf-sched.sch_delay.max.ms.do_nanosleep.hrtimer_nanosleep.common_nsleep.__x64_sys_clock_nanosleep
      3.28 ± 34%     +29.2%       4.24 ± 13%  perf-sched.sch_delay.max.ms.do_task_dead.do_exit.do_group_exit.__x64_sys_exit_group.do_syscall_64
      3.32 ± 27%      -2.9%       3.22 ± 23%  perf-sched.sch_delay.max.ms.do_wait.kernel_wait4.__do_sys_wait4.do_syscall_64
      0.53 ±209%    +168.8%       1.43 ± 93%  perf-sched.sch_delay.max.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_exc_page_fault
      0.55 ±194%     +60.5%       0.88 ±201%  perf-sched.sch_delay.max.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt
      0.00 ±300%    -100.0%       0.00        perf-sched.sch_delay.max.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_reschedule_ipi
      2.77 ± 30%    +150.3%       6.93 ±142%  perf-sched.sch_delay.max.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.syscall_exit_to_user_mode.do_syscall_64
      0.00 ±300%    -100.0%       0.00        perf-sched.sch_delay.max.ms.futex_wait_queue.futex_wait.do_futex.__x64_sys_futex
      0.40 ±300%    -100.0%       0.00        perf-sched.sch_delay.max.ms.io_schedule.folio_wait_bit_common.filemap_fault.__do_fault
      0.28 ±297%     -99.2%       0.00 ±300%  perf-sched.sch_delay.max.ms.kthreadd.ret_from_fork.ret_from_fork_asm
      3.24 ± 30%     -10.1%       2.91 ± 47%  perf-sched.sch_delay.max.ms.pipe_read.vfs_read.ksys_read.do_syscall_64
      3.63 ± 18%     -14.5%       3.10 ± 33%  perf-sched.sch_delay.max.ms.schedule_hrtimeout_range_clock.do_poll.constprop.0.do_sys_poll
      3.25 ± 30%     +14.0%       3.70 ± 27%  perf-sched.sch_delay.max.ms.schedule_hrtimeout_range_clock.do_select.core_sys_select.kern_select
      2.83 ± 57%     +81.2%       5.12 ±167%  perf-sched.sch_delay.max.ms.schedule_hrtimeout_range_clock.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
      0.00 ±202%     -28.6%       0.00 ±300%  perf-sched.sch_delay.max.ms.schedule_preempt_disabled.kthread.ret_from_fork.ret_from_fork_asm
      0.00 ±300%    -100.0%       0.00        perf-sched.sch_delay.max.ms.schedule_preempt_disabled.rwsem_down_read_slowpath.down_read_killable.__access_remote_vm
      2.67 ±203%     -30.4%       1.86 ±300%  perf-sched.sch_delay.max.ms.schedule_timeout.__wait_for_common.wait_for_completion_killable.__kthread_create_on_node
      4.15 ± 10%     +17.7%       4.89 ± 23%  perf-sched.sch_delay.max.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone
      0.08 ±106%     -26.8%       0.06 ±115%  perf-sched.sch_delay.max.ms.schedule_timeout.kcompactd.kthread.ret_from_fork
      8.41 ±161%     +26.1%      10.61 ± 97%  perf-sched.sch_delay.max.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
      0.00 ±299%    -100.0%       0.00        perf-sched.sch_delay.max.ms.schedule_timeout.transaction_kthread.kthread.ret_from_fork
      4.38 ± 45%    +154.3%      11.15 ± 59%  perf-sched.sch_delay.max.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
      1.90 ± 78%    +518.9%      11.77 ± 98%  perf-sched.sch_delay.max.ms.syslog_print.do_syslog.kmsg_read.vfs_read
      2.30 ± 46%     +56.5%       3.60 ±  7%  perf-sched.sch_delay.max.ms.wait_for_partner.fifo_open.do_dentry_open.do_open
      3.74 ± 39%     +32.3%       4.95 ± 55%  perf-sched.sch_delay.max.ms.worker_thread.kthread.ret_from_fork.ret_from_fork_asm
      0.01 ±  4%     +41.1%       0.02 ± 19%  perf-sched.total_sch_delay.average.ms
     13.07 ± 84%    +113.7%      27.94 ± 42%  perf-sched.total_sch_delay.max.ms
     73.81            +1.5%      74.92        perf-sched.total_wait_and_delay.average.ms
     42853            -1.5%      42224        perf-sched.total_wait_and_delay.count.ms
      4899            -0.2%       4892        perf-sched.total_wait_and_delay.max.ms
     73.80            +1.5%      74.90        perf-sched.total_wait_time.average.ms
      4898            -0.2%       4889        perf-sched.total_wait_time.max.ms
     40.85 ±299%    -100.0%       0.00        perf-sched.wait_and_delay.avg.ms.__cond_resched.__kmem_cache_alloc_node.__kmalloc.acpi_ex_allocate_name_string.acpi_ex_get_name_string
      0.00         +1e+104%     100.01 ±300%  perf-sched.wait_and_delay.avg.ms.__cond_resched.__kmem_cache_alloc_node.kmalloc_trace.vmstat_start.seq_read_iter
      6.73            -5.2%       6.37 ± 17%  perf-sched.wait_and_delay.avg.ms.__cond_resched.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr.__sched_setaffinity
      0.00       +1.4e+103%      14.35 ±299%  perf-sched.wait_and_delay.avg.ms.__cond_resched.dput.__fput.__x64_sys_close.do_syscall_64
      0.00         +5e+103%      50.14 ±300%  perf-sched.wait_and_delay.avg.ms.__cond_resched.kmem_cache_alloc.acpi_ut_create_generic_state.acpi_ds_result_push.acpi_ds_exec_end_op
      0.00         +1e+104%     100.02 ±300%  perf-sched.wait_and_delay.avg.ms.__cond_resched.kmem_cache_alloc.vm_area_dup.dup_mmap.dup_mm
      0.00       +2.1e+103%      20.74 ±299%  perf-sched.wait_and_delay.avg.ms.__cond_resched.mutex_lock.drm_property_free_blob.drm_gem_destroy_shadow_plane_state.drm_atomic_state_default_clear
      0.00         +3e+104%     300.08 ±152%  perf-sched.wait_and_delay.avg.ms.__cond_resched.mutex_lock_killable.pcpu_alloc.mm_init.dup_mm
    204.67 ±300%      +0.1%     204.80 ±200%  perf-sched.wait_and_delay.avg.ms.__cond_resched.process_one_work.worker_thread.kthread.ret_from_fork
      0.00         +2e+104%     204.67 ±200%  perf-sched.wait_and_delay.avg.ms.__cond_resched.refresh_cpu_vm_stats.vmstat_update.process_one_work.worker_thread
      0.00       +5.3e+103%      52.66 ±161%  perf-sched.wait_and_delay.avg.ms.__cond_resched.shmem_get_folio_gfp.shmem_read_folio_gfp.drm_gem_get_pages.drm_gem_shmem_get_pages
      0.00       +7.8e+102%       7.84 ±299%  perf-sched.wait_and_delay.avg.ms.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
      0.00         +1e+104%     100.06 ±300%  perf-sched.wait_and_delay.avg.ms.__cond_resched.task_work_run.get_signal.arch_do_signal_or_restart.exit_to_user_mode_loop
      0.00       +2.6e+103%      25.66 ±202%  perf-sched.wait_and_delay.avg.ms.__cond_resched.zap_pmd_range.isra.0.unmap_page_range
    271.66 ± 68%     -70.3%      80.71 ±155%  perf-sched.wait_and_delay.avg.ms.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
    214.27 ± 86%     -29.6%     150.78 ±104%  perf-sched.wait_and_delay.avg.ms.do_nanosleep.hrtimer_nanosleep.common_nsleep.__x64_sys_clock_nanosleep
    235.93 ±  5%      +1.4%     239.24 ±  5%  perf-sched.wait_and_delay.avg.ms.do_task_dead.do_exit.do_group_exit.__x64_sys_exit_group.do_syscall_64
      0.00       +7.1e+101%       0.71 ±175%  perf-sched.wait_and_delay.avg.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_exc_page_fault
      0.24 ±  6%    +311.6%       1.00 ± 46%  perf-sched.wait_and_delay.avg.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt
      0.02 ±299%    +545.1%       0.11 ± 45%  perf-sched.wait_and_delay.avg.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_call_function_single
      0.24 ±  6%    +338.2%       1.07 ± 53%  perf-sched.wait_and_delay.avg.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_reschedule_ipi
     23.21 ± 50%      -6.7%      21.65 ± 34%  perf-sched.wait_and_delay.avg.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.syscall_exit_to_user_mode.do_syscall_64
    169.92 ± 16%     -30.1%     118.73 ± 28%  perf-sched.wait_and_delay.avg.ms.pipe_read.vfs_read.ksys_read.do_syscall_64
    552.06 ±  5%      +1.1%     558.36 ±  2%  perf-sched.wait_and_delay.avg.ms.schedule_hrtimeout_range_clock.do_poll.constprop.0.do_sys_poll
    370.29 ±  4%      +4.6%     387.38 ±  3%  perf-sched.wait_and_delay.avg.ms.schedule_hrtimeout_range_clock.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
    319.32 ±299%    -100.0%       0.00        perf-sched.wait_and_delay.avg.ms.schedule_preempt_disabled.rwsem_down_read_slowpath.down_read_killable.__access_remote_vm
    452.79            +0.1%     453.34        perf-sched.wait_and_delay.avg.ms.schedule_timeout.kcompactd.kthread.ret_from_fork
      3.65 ±  2%      -5.4%       3.45 ±  5%  perf-sched.wait_and_delay.avg.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
    121.38            +2.6%     124.58        perf-sched.wait_and_delay.avg.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
    724.27 ±  2%      +1.6%     735.88        perf-sched.wait_and_delay.avg.ms.worker_thread.kthread.ret_from_fork.ret_from_fork_asm
      0.10 ±300%    -100.0%       0.00        perf-sched.wait_and_delay.count.__cond_resched.__kmem_cache_alloc_node.__kmalloc.acpi_ex_allocate_name_string.acpi_ex_get_name_string
      0.00         +1e+101%       0.10 ±300%  perf-sched.wait_and_delay.count.__cond_resched.__kmem_cache_alloc_node.kmalloc_trace.vmstat_start.seq_read_iter
      1344            +0.0%       1344        perf-sched.wait_and_delay.count.__cond_resched.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr.__sched_setaffinity
      0.00         +7e+101%       0.70 ±300%  perf-sched.wait_and_delay.count.__cond_resched.dput.__fput.__x64_sys_close.do_syscall_64
      0.00         +1e+101%       0.10 ±300%  perf-sched.wait_and_delay.count.__cond_resched.kmem_cache_alloc.acpi_ut_create_generic_state.acpi_ds_result_push.acpi_ds_exec_end_op
      0.00         +1e+101%       0.10 ±300%  perf-sched.wait_and_delay.count.__cond_resched.kmem_cache_alloc.vm_area_dup.dup_mmap.dup_mm
      0.00         +1e+101%       0.10 ±300%  perf-sched.wait_and_delay.count.__cond_resched.mutex_lock.drm_property_free_blob.drm_gem_destroy_shadow_plane_state.drm_atomic_state_default_clear
      0.00         +3e+101%       0.30 ±152%  perf-sched.wait_and_delay.count.__cond_resched.mutex_lock_killable.pcpu_alloc.mm_init.dup_mm
      0.10 ±300%    +100.0%       0.20 ±200%  perf-sched.wait_and_delay.count.__cond_resched.process_one_work.worker_thread.kthread.ret_from_fork
      0.00         +2e+101%       0.20 ±200%  perf-sched.wait_and_delay.count.__cond_resched.refresh_cpu_vm_stats.vmstat_update.process_one_work.worker_thread
      0.00         +4e+101%       0.40 ±165%  perf-sched.wait_and_delay.count.__cond_resched.shmem_get_folio_gfp.shmem_read_folio_gfp.drm_gem_get_pages.drm_gem_shmem_get_pages
      0.00       +1.3e+103%      12.90 ±299%  perf-sched.wait_and_delay.count.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
      0.00         +1e+101%       0.10 ±300%  perf-sched.wait_and_delay.count.__cond_resched.task_work_run.get_signal.arch_do_signal_or_restart.exit_to_user_mode_loop
      0.00       +1.6e+102%       1.60 ±201%  perf-sched.wait_and_delay.count.__cond_resched.zap_pmd_range.isra.0.unmap_page_range
      8.00 ± 50%     -50.0%       4.00 ±122%  perf-sched.wait_and_delay.count.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
      4.40 ± 66%     -18.2%       3.60 ± 81%  perf-sched.wait_and_delay.count.do_nanosleep.hrtimer_nanosleep.common_nsleep.__x64_sys_clock_nanosleep
    124.30            -0.2%     124.00        perf-sched.wait_and_delay.count.do_task_dead.do_exit.do_group_exit.__x64_sys_exit_group.do_syscall_64
      0.00       +1.1e+104%     112.40 ±152%  perf-sched.wait_and_delay.count.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_exc_page_fault
     18268            -4.4%      17455 ±  3%  perf-sched.wait_and_delay.count.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt
     28.50 ±300%    +834.4%     266.30 ± 34%  perf-sched.wait_and_delay.count.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_call_function_single
      1595            +1.4%       1617 ±  3%  perf-sched.wait_and_delay.count.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_reschedule_ipi
    284.40 ± 37%     +95.5%     556.00 ± 36%  perf-sched.wait_and_delay.count.exit_to_user_mode_loop.exit_to_user_mode_prepare.syscall_exit_to_user_mode.do_syscall_64
    439.20 ± 15%     +50.7%     662.00 ± 30%  perf-sched.wait_and_delay.count.pipe_read.vfs_read.ksys_read.do_syscall_64
     21.30 ±  5%      -1.4%      21.00 ±  3%  perf-sched.wait_and_delay.count.schedule_hrtimeout_range_clock.do_poll.constprop.0.do_sys_poll
     13.20 ±  7%      -3.8%      12.70 ±  3%  perf-sched.wait_and_delay.count.schedule_hrtimeout_range_clock.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
      0.10 ±300%    -100.0%       0.00        perf-sched.wait_and_delay.count.schedule_preempt_disabled.rwsem_down_read_slowpath.down_read_killable.__access_remote_vm
     19.70 ±  3%      +1.0%      19.90        perf-sched.wait_and_delay.count.schedule_timeout.kcompactd.kthread.ret_from_fork
      1365            +6.3%       1451 ±  5%  perf-sched.wait_and_delay.count.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
     17282            -3.5%      16674 ±  2%  perf-sched.wait_and_delay.count.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
      1249 ±  2%      -1.4%       1231        perf-sched.wait_and_delay.count.worker_thread.kthread.ret_from_fork.ret_from_fork_asm
     40.85 ±299%    -100.0%       0.00        perf-sched.wait_and_delay.max.ms.__cond_resched.__kmem_cache_alloc_node.__kmalloc.acpi_ex_allocate_name_string.acpi_ex_get_name_string
      0.00         +1e+104%     100.01 ±300%  perf-sched.wait_and_delay.max.ms.__cond_resched.__kmem_cache_alloc_node.kmalloc_trace.vmstat_start.seq_read_iter
      4899            -8.0%       4506 ± 25%  perf-sched.wait_and_delay.max.ms.__cond_resched.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr.__sched_setaffinity
      0.00         +1e+104%     100.35 ±300%  perf-sched.wait_and_delay.max.ms.__cond_resched.dput.__fput.__x64_sys_close.do_syscall_64
      0.00         +5e+103%      50.14 ±300%  perf-sched.wait_and_delay.max.ms.__cond_resched.kmem_cache_alloc.acpi_ut_create_generic_state.acpi_ds_result_push.acpi_ds_exec_end_op
      0.00         +1e+104%     100.02 ±300%  perf-sched.wait_and_delay.max.ms.__cond_resched.kmem_cache_alloc.vm_area_dup.dup_mmap.dup_mm
      0.00       +2.1e+103%      20.74 ±299%  perf-sched.wait_and_delay.max.ms.__cond_resched.mutex_lock.drm_property_free_blob.drm_gem_destroy_shadow_plane_state.drm_atomic_state_default_clear
      0.00         +3e+104%     300.08 ±152%  perf-sched.wait_and_delay.max.ms.__cond_resched.mutex_lock_killable.pcpu_alloc.mm_init.dup_mm
    204.67 ±300%      +0.1%     204.80 ±200%  perf-sched.wait_and_delay.max.ms.__cond_resched.process_one_work.worker_thread.kthread.ret_from_fork
      0.00         +2e+104%     204.67 ±200%  perf-sched.wait_and_delay.max.ms.__cond_resched.refresh_cpu_vm_stats.vmstat_update.process_one_work.worker_thread
      0.00       +6.3e+103%      62.80 ±152%  perf-sched.wait_and_delay.max.ms.__cond_resched.shmem_get_folio_gfp.shmem_read_folio_gfp.drm_gem_get_pages.drm_gem_shmem_get_pages
      0.00       +9.3e+102%       9.33 ±300%  perf-sched.wait_and_delay.max.ms.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
      0.00         +1e+104%     100.06 ±300%  perf-sched.wait_and_delay.max.ms.__cond_resched.task_work_run.get_signal.arch_do_signal_or_restart.exit_to_user_mode_loop
      0.00         +2e+104%     200.44 ±200%  perf-sched.wait_and_delay.max.ms.__cond_resched.zap_pmd_range.isra.0.unmap_page_range
    804.48 ± 50%     -50.2%     400.68 ±122%  perf-sched.wait_and_delay.max.ms.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
    702.01 ± 65%     -14.4%     600.63 ± 81%  perf-sched.wait_and_delay.max.ms.do_nanosleep.hrtimer_nanosleep.common_nsleep.__x64_sys_clock_nanosleep
      1323 ± 48%     +37.2%       1815 ± 54%  perf-sched.wait_and_delay.max.ms.do_task_dead.do_exit.do_group_exit.__x64_sys_exit_group.do_syscall_64
      0.00         +1e+104%     102.23 ±292%  perf-sched.wait_and_delay.max.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_exc_page_fault
    128.86 ±226%      +2.2%     131.68 ±219%  perf-sched.wait_and_delay.max.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt
      0.25 ±299%   +1670.4%       4.39 ± 60%  perf-sched.wait_and_delay.max.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_call_function_single
      5.66 ±156%     +54.5%       8.74 ± 54%  perf-sched.wait_and_delay.max.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_reschedule_ipi
    901.41 ± 33%     +53.9%       1387 ± 83%  perf-sched.wait_and_delay.max.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.syscall_exit_to_user_mode.do_syscall_64
      1051            +0.8%       1059        perf-sched.wait_and_delay.max.ms.pipe_read.vfs_read.ksys_read.do_syscall_64
      1007            -0.1%       1006        perf-sched.wait_and_delay.max.ms.schedule_hrtimeout_range_clock.do_poll.constprop.0.do_sys_poll
    499.87            +0.0%     499.88        perf-sched.wait_and_delay.max.ms.schedule_hrtimeout_range_clock.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
    319.32 ±299%    -100.0%       0.00        perf-sched.wait_and_delay.max.ms.schedule_preempt_disabled.rwsem_down_read_slowpath.down_read_killable.__access_remote_vm
    504.61            +0.1%     505.27        perf-sched.wait_and_delay.max.ms.schedule_timeout.kcompactd.kthread.ret_from_fork
     20.67 ±122%     +10.5%      22.84 ± 81%  perf-sched.wait_and_delay.max.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
      1053           +10.8%       1167 ± 26%  perf-sched.wait_and_delay.max.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
      2046            -0.0%       2046        perf-sched.wait_and_delay.max.ms.worker_thread.kthread.ret_from_fork.ret_from_fork_asm
      0.20 ± 25%    +330.8%       0.87 ± 40%  perf-sched.wait_time.avg.ms.__cond_resched.__alloc_pages.__folio_alloc.vma_alloc_folio.do_anonymous_page
      0.30 ±300%    +166.9%       0.80 ±228%  perf-sched.wait_time.avg.ms.__cond_resched.__alloc_pages.__folio_alloc.vma_alloc_folio.shmem_alloc_folio
      0.00 ±300%    -100.0%       0.00        perf-sched.wait_time.avg.ms.__cond_resched.__alloc_pages.__folio_alloc.vma_alloc_folio.wp_page_copy
      0.00       +1.2e+102%       1.19 ±225%  perf-sched.wait_time.avg.ms.__cond_resched.__alloc_pages.__pud_alloc.alloc_new_pud.constprop
      0.40 ±300%    -100.0%       0.00        perf-sched.wait_time.avg.ms.__cond_resched.__alloc_pages.pte_alloc_one.__do_fault.do_read_fault
      0.00       +4.7e+101%       0.47 ±300%  perf-sched.wait_time.avg.ms.__cond_resched.__alloc_pages.pte_alloc_one.__pte_alloc.do_anonymous_page
      2.19 ±124%     -88.6%       0.25 ±299%  perf-sched.wait_time.avg.ms.__cond_resched.__do_fault.do_read_fault.do_fault.__handle_mm_fault
     40.85 ±300%    -100.0%       0.00        perf-sched.wait_time.avg.ms.__cond_resched.__kmem_cache_alloc_node.__kmalloc.acpi_ex_allocate_name_string.acpi_ex_get_name_string
      0.00          +2e+99%       0.00 ±300%  perf-sched.wait_time.avg.ms.__cond_resched.__kmem_cache_alloc_node.__kmalloc_node.alloc_cpumask_var_node.__sched_setaffinity
      0.00       +3.1e+101%       0.31 ±300%  perf-sched.wait_time.avg.ms.__cond_resched.__kmem_cache_alloc_node.__kmalloc_node.memcg_alloc_slab_cgroups.allocate_slab
      0.00 ±300%    +308.3%       0.00 ±200%  perf-sched.wait_time.avg.ms.__cond_resched.__kmem_cache_alloc_node.kmalloc_trace.perf_event_mmap_event.perf_event_mmap
      0.00        +3.9e+99%       0.00 ±200%  perf-sched.wait_time.avg.ms.__cond_resched.__kmem_cache_alloc_node.kmalloc_trace.single_open.single_open_size
      0.00         +1e+104%     100.01 ±299%  perf-sched.wait_time.avg.ms.__cond_resched.__kmem_cache_alloc_node.kmalloc_trace.vmstat_start.seq_read_iter
      6.72            -5.2%       6.37 ± 17%  perf-sched.wait_time.avg.ms.__cond_resched.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr.__sched_setaffinity
      0.84 ±132%     -63.0%       0.31 ±194%  perf-sched.wait_time.avg.ms.__cond_resched.__wait_for_common.wait_for_completion_state.kernel_clone.__x64_sys_vfork
      0.15 ±300%    -100.0%       0.00        perf-sched.wait_time.avg.ms.__cond_resched.cgroup_css_set_fork.cgroup_can_fork.copy_process.kernel_clone
      0.00 ±300%    +327.3%       0.01 ±299%  perf-sched.wait_time.avg.ms.__cond_resched.change_pmd_range.change_p4d_range.change_protection_range.mprotect_fixup
      0.00 ±300%    -100.0%       0.00        perf-sched.wait_time.avg.ms.__cond_resched.dentry_kill.dput.step_into.link_path_walk
      0.15 ±154%    +506.7%       0.92 ±112%  perf-sched.wait_time.avg.ms.__cond_resched.down_write.__split_vma.do_vmi_align_munmap.__do_sys_brk
      0.20 ± 41%    +161.6%       0.53 ± 98%  perf-sched.wait_time.avg.ms.__cond_resched.down_write.anon_vma_clone.__split_vma.do_vmi_align_munmap
      0.00        +2.2e+99%       0.00 ±300%  perf-sched.wait_time.avg.ms.__cond_resched.down_write.anon_vma_clone.__split_vma.vma_modify
      0.00       +2.6e+101%       0.26 ±201%  perf-sched.wait_time.avg.ms.__cond_resched.down_write.do_brk_flags.__do_sys_brk.do_syscall_64
      0.01 ±204%    -100.0%       0.00        perf-sched.wait_time.avg.ms.__cond_resched.down_write.free_pgtables.exit_mmap.__mmput
      0.00 ±300%   +1245.2%       0.04 ±299%  perf-sched.wait_time.avg.ms.__cond_resched.down_write.mmap_region.do_mmap.vm_mmap_pgoff
      0.05 ±194%    +126.6%       0.12 ±223%  perf-sched.wait_time.avg.ms.__cond_resched.down_write.unlink_anon_vmas.free_pgtables.unmap_region
      0.00 ±299%    -100.0%       0.00        perf-sched.wait_time.avg.ms.__cond_resched.down_write.unlink_file_vma.free_pgtables.exit_mmap
      0.07 ±125%     -33.7%       0.05 ±300%  perf-sched.wait_time.avg.ms.__cond_resched.down_write.vma_prepare.__split_vma.do_vmi_align_munmap
      0.02 ±300%    +723.8%       0.19 ±259%  perf-sched.wait_time.avg.ms.__cond_resched.down_write.vma_prepare.do_brk_flags.__do_sys_brk
      0.25 ± 76%    +104.2%       0.51 ±114%  perf-sched.wait_time.avg.ms.__cond_resched.down_write_killable.__do_sys_brk.do_syscall_64.entry_SYSCALL_64_after_hwframe
      0.32 ±194%     -79.5%       0.07 ±248%  perf-sched.wait_time.avg.ms.__cond_resched.down_write_killable.exec_mmap.begin_new_exec.load_elf_binary
      0.00 ±299%  +12272.7%       0.41 ±300%  perf-sched.wait_time.avg.ms.__cond_resched.down_write_killable.setup_arg_pages.load_elf_binary.search_binary_handler
      0.21 ±299%    -100.0%       0.00        perf-sched.wait_time.avg.ms.__cond_resched.down_write_killable.vm_mmap_pgoff.elf_map.load_elf_binary
      0.00 ±300%    -100.0%       0.00        perf-sched.wait_time.avg.ms.__cond_resched.down_write_killable.vm_mmap_pgoff.elf_map.load_elf_interp
      0.38 ±300%    -100.0%       0.00        perf-sched.wait_time.avg.ms.__cond_resched.down_write_killable.vm_mmap_pgoff.ksys_mmap_pgoff.do_syscall_64
      0.08 ±198%  +17398.4%      14.54 ±294%  perf-sched.wait_time.avg.ms.__cond_resched.dput.__fput.__x64_sys_close.do_syscall_64
      0.01 ±122%    +290.4%       0.04 ±282%  perf-sched.wait_time.avg.ms.__cond_resched.dput.step_into.link_path_walk.part
      0.58 ±206%     -99.7%       0.00 ±299%  perf-sched.wait_time.avg.ms.__cond_resched.dput.terminate_walk.path_openat.do_filp_open
      0.00 ±201%    -100.0%       0.00        perf-sched.wait_time.avg.ms.__cond_resched.exit_mmap.__mmput.exit_mm.do_exit
      0.00 ±299%    -100.0%       0.00        perf-sched.wait_time.avg.ms.__cond_resched.filemap_read.__kernel_read.search_binary_handler.exec_binprm
      0.00 ±200%    -100.0%       0.00        perf-sched.wait_time.avg.ms.__cond_resched.filemap_read.vfs_read.ksys_read.do_syscall_64
      0.20 ±300%   +1341.7%       2.87 ±188%  perf-sched.wait_time.avg.ms.__cond_resched.generic_perform_write.shmem_file_write_iter.vfs_write.ksys_write
      0.00 ±300%   +2110.0%       0.02 ±300%  perf-sched.wait_time.avg.ms.__cond_resched.khugepaged.kthread.ret_from_fork.ret_from_fork_asm
      0.00 ±300%    -100.0%       0.00        perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc.__anon_vma_prepare.do_anonymous_page.__handle_mm_fault
      0.00       +3.9e+101%       0.39 ±300%  perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc.__anon_vma_prepare.wp_page_copy.__handle_mm_fault
      0.00         +5e+103%      50.14 ±300%  perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc.acpi_ut_create_generic_state.acpi_ds_result_push.acpi_ds_exec_end_op
      0.42 ±154%    -100.0%       0.00        perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc.alloc_pid.copy_process.kernel_clone
      0.20 ± 30%    +241.4%       0.68 ±123%  perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc.mas_alloc_nodes.mas_preallocate.__split_vma
      0.21 ±128%    +117.3%       0.46 ±137%  perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc.mas_alloc_nodes.mas_preallocate.do_brk_flags
      0.16 ±300%    -100.0%       0.00        perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc.mas_alloc_nodes.mas_preallocate.mmap_region
      0.00        +2.4e+99%       0.00 ±300%  perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc.mas_alloc_nodes.mas_preallocate.vma_expand
      0.00 ±299%    -100.0%       0.00        perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc.security_file_alloc.init_file.alloc_empty_file
      0.00 ±300%    -100.0%       0.00        perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc.vm_area_alloc.__install_special_mapping.map_vdso
      0.32 ±300%    -100.0%       0.00        perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc.vm_area_alloc.do_brk_flags.__do_sys_brk
      0.10 ±299%    -100.0%       0.00        perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc.vm_area_alloc.mmap_region.do_mmap
      0.32 ± 65%     +94.9%       0.62 ±100%  perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc.vm_area_dup.__split_vma.do_vmi_align_munmap
      0.00       +1.7e+101%       0.17 ±300%  perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc.vm_area_dup.__split_vma.vma_modify
      0.00         +1e+104%     100.01 ±300%  perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc.vm_area_dup.dup_mmap.dup_mm
      0.12 ± 94%     -11.6%       0.10 ±152%  perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc_bulk.mas_alloc_nodes.mas_preallocate.__split_vma
      0.00 ±300%  +24600.0%       0.42 ±250%  perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc_bulk.mas_alloc_nodes.mas_preallocate.do_brk_flags
      0.00       +2.1e+103%      20.73 ±300%  perf-sched.wait_time.avg.ms.__cond_resched.mutex_lock.drm_property_free_blob.drm_gem_destroy_shadow_plane_state.drm_atomic_state_default_clear
      0.00 ±300%    -100.0%       0.00        perf-sched.wait_time.avg.ms.__cond_resched.mutex_lock.futex_exit_release.exit_mm_release.exit_mm
      0.00 ±200%     -47.7%       0.00 ±300%  perf-sched.wait_time.avg.ms.__cond_resched.mutex_lock.perf_event_ctx_lock_nested.constprop.0
      2.39 ± 81%     -18.7%       1.94 ±100%  perf-sched.wait_time.avg.ms.__cond_resched.mutex_lock.perf_poll.do_poll.constprop
      0.00       +7.7e+100%       0.08 ±299%  perf-sched.wait_time.avg.ms.__cond_resched.mutex_lock_killable.pcpu_alloc.mm_init.alloc_bprm
      0.00         +3e+104%     300.07 ±152%  perf-sched.wait_time.avg.ms.__cond_resched.mutex_lock_killable.pcpu_alloc.mm_init.dup_mm
    204.66 ±300%      +0.1%     204.80 ±200%  perf-sched.wait_time.avg.ms.__cond_resched.process_one_work.worker_thread.kthread.ret_from_fork
      0.00         +2e+104%     204.66 ±200%  perf-sched.wait_time.avg.ms.__cond_resched.refresh_cpu_vm_stats.vmstat_update.process_one_work.worker_thread
      0.20 ±122%    +125.5%       0.44 ±244%  perf-sched.wait_time.avg.ms.__cond_resched.remove_vma.do_vmi_align_munmap.__do_sys_brk.do_syscall_64
      0.00 ±300%  +1.3e+06%      57.48 ±143%  perf-sched.wait_time.avg.ms.__cond_resched.shmem_get_folio_gfp.shmem_read_folio_gfp.drm_gem_get_pages.drm_gem_shmem_get_pages
      0.15 ±300%    +362.8%       0.69 ±202%  perf-sched.wait_time.avg.ms.__cond_resched.shmem_inode_acct_blocks.shmem_alloc_and_add_folio.shmem_get_folio_gfp.shmem_write_begin
      0.00        +2.8e+99%       0.00 ±300%  perf-sched.wait_time.avg.ms.__cond_resched.slab_pre_alloc_hook.constprop.0.kmem_cache_alloc_lru
     39.66 ± 14%     +56.8%      62.18 ± 16%  perf-sched.wait_time.avg.ms.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
      0.02 ±106%    +448.4%       0.08 ±146%  perf-sched.wait_time.avg.ms.__cond_resched.stop_one_cpu.sched_exec.bprm_execve.part
      0.00 ±300%    -100.0%       0.00        perf-sched.wait_time.avg.ms.__cond_resched.switch_task_namespaces.do_exit.do_group_exit.__x64_sys_exit_group
      0.04 ± 89%     +47.0%       0.07 ±174%  perf-sched.wait_time.avg.ms.__cond_resched.task_work_run.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode
      0.00       +4.6e+100%       0.05 ±271%  perf-sched.wait_time.avg.ms.__cond_resched.task_work_run.exit_to_user_mode_loop.exit_to_user_mode_prepare.syscall_exit_to_user_mode
      0.00         +1e+104%     100.06 ±300%  perf-sched.wait_time.avg.ms.__cond_resched.task_work_run.get_signal.arch_do_signal_or_restart.exit_to_user_mode_loop
      0.40 ±295%     -99.0%       0.00 ±209%  perf-sched.wait_time.avg.ms.__cond_resched.tlb_batch_pages_flush.tlb_finish_mmu.exit_mmap.__mmput
      0.28 ± 15%    +483.6%       1.63 ± 48%  perf-sched.wait_time.avg.ms.__cond_resched.tlb_batch_pages_flush.tlb_finish_mmu.unmap_region.constprop
      0.02 ±300%   +1253.6%       0.30 ±300%  perf-sched.wait_time.avg.ms.__cond_resched.unmap_page_range.unmap_vmas.unmap_region.constprop
      0.28 ± 28%    +193.3%       0.81 ± 66%  perf-sched.wait_time.avg.ms.__cond_resched.unmap_vmas.unmap_region.constprop.0
      0.00 ±214%    -100.0%       0.00        perf-sched.wait_time.avg.ms.__cond_resched.wp_page_copy.__handle_mm_fault.handle_mm_fault.do_user_addr_fault
      5.09 ±300%    -100.0%       0.00        perf-sched.wait_time.avg.ms.__cond_resched.ww_mutex_lock.drm_gem_vunmap_unlocked.drm_fbdev_generic_helper_fb_dirty.drm_fb_helper_damage_work
      0.00       +3.7e+100%       0.04 ±253%  perf-sched.wait_time.avg.ms.__cond_resched.ww_mutex_lock.drm_gem_vunmap_unlocked.drm_gem_fb_vunmap.drm_atomic_helper_cleanup_planes
      0.27 ± 29%   +9684.8%      26.29 ±195%  perf-sched.wait_time.avg.ms.__cond_resched.zap_pmd_range.isra.0.unmap_page_range
    271.34 ± 67%     -69.9%      81.76 ±152%  perf-sched.wait_time.avg.ms.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
      2.37 ± 40%     -67.9%       0.76 ±154%  perf-sched.wait_time.avg.ms.devkmsg_read.vfs_read.ksys_read.do_syscall_64
    214.20 ± 86%     -29.4%     151.17 ±104%  perf-sched.wait_time.avg.ms.do_nanosleep.hrtimer_nanosleep.common_nsleep.__x64_sys_clock_nanosleep
    235.79 ±  5%      +1.3%     238.78 ±  5%  perf-sched.wait_time.avg.ms.do_task_dead.do_exit.do_group_exit.__x64_sys_exit_group.do_syscall_64
      2.94 ±  4%     +30.9%       3.84 ± 85%  perf-sched.wait_time.avg.ms.do_wait.kernel_wait4.__do_sys_wait4.do_syscall_64
      0.00       +2.6e+101%       0.26 ±300%  perf-sched.wait_time.avg.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_common_interrupt
      0.26 ± 22%    +375.6%       1.22 ± 80%  perf-sched.wait_time.avg.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_exc_page_fault
      0.24 ±  6%    +311.6%       1.00 ± 46%  perf-sched.wait_time.avg.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt
      0.13 ± 18%      -9.5%       0.11 ± 29%  perf-sched.wait_time.avg.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_call_function_single
      0.24 ±  6%    +338.2%       1.07 ± 53%  perf-sched.wait_time.avg.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_reschedule_ipi
     27.55 ± 36%     -21.7%      21.58 ± 34%  perf-sched.wait_time.avg.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.syscall_exit_to_user_mode.do_syscall_64
      0.73 ±300%    -100.0%       0.00        perf-sched.wait_time.avg.ms.futex_wait_queue.futex_wait.do_futex.__x64_sys_futex
      0.00 ±300%  +32366.7%       0.39 ±300%  perf-sched.wait_time.avg.ms.io_schedule.folio_wait_bit_common.filemap_fault.__do_fault
    169.88 ± 16%     -30.1%     118.71 ± 28%  perf-sched.wait_time.avg.ms.pipe_read.vfs_read.ksys_read.do_syscall_64
    551.67 ±  5%      +1.2%     558.07 ±  2%  perf-sched.wait_time.avg.ms.schedule_hrtimeout_range_clock.do_poll.constprop.0.do_sys_poll
    370.03 ±  4%      +4.6%     386.92 ±  3%  perf-sched.wait_time.avg.ms.schedule_hrtimeout_range_clock.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
    319.32 ±299%    -100.0%       0.00        perf-sched.wait_time.avg.ms.schedule_preempt_disabled.rwsem_down_read_slowpath.down_read_killable.__access_remote_vm
      0.01 ±122%      +6.5%       0.01 ±123%  perf-sched.wait_time.avg.ms.schedule_timeout.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr
      2.65 ± 10%     +17.7%       3.12 ± 11%  perf-sched.wait_time.avg.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone
    452.77            +0.1%     453.32        perf-sched.wait_time.avg.ms.schedule_timeout.kcompactd.kthread.ret_from_fork
      0.18 ±300%    +102.3%       0.36 ±299%  perf-sched.wait_time.avg.ms.schedule_timeout.khugepaged_wait_work.khugepaged.kthread
      3.63            -5.5%       3.43 ±  5%  perf-sched.wait_time.avg.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
    121.37            +2.6%     124.56        perf-sched.wait_time.avg.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
      3.83 ± 19%     -71.1%       1.11 ±123%  perf-sched.wait_time.avg.ms.syslog_print.do_syslog.kmsg_read.vfs_read
      0.44 ± 43%     -21.7%       0.34 ± 40%  perf-sched.wait_time.avg.ms.wait_for_partner.fifo_open.do_dentry_open.do_open
    724.24 ±  2%      +1.6%     735.84        perf-sched.wait_time.avg.ms.worker_thread.kthread.ret_from_fork.ret_from_fork_asm
      0.83 ± 62%    +573.9%       5.62 ± 59%  perf-sched.wait_time.max.ms.__cond_resched.__alloc_pages.__folio_alloc.vma_alloc_folio.do_anonymous_page
      0.30 ±300%    +166.9%       0.80 ±228%  perf-sched.wait_time.max.ms.__cond_resched.__alloc_pages.__folio_alloc.vma_alloc_folio.shmem_alloc_folio
      0.00 ±300%    -100.0%       0.00        perf-sched.wait_time.max.ms.__cond_resched.__alloc_pages.__folio_alloc.vma_alloc_folio.wp_page_copy
      0.00       +1.2e+102%       1.19 ±225%  perf-sched.wait_time.max.ms.__cond_resched.__alloc_pages.__pud_alloc.alloc_new_pud.constprop
      0.40 ±300%    -100.0%       0.00        perf-sched.wait_time.max.ms.__cond_resched.__alloc_pages.pte_alloc_one.__do_fault.do_read_fault
      0.00       +8.8e+101%       0.88 ±300%  perf-sched.wait_time.max.ms.__cond_resched.__alloc_pages.pte_alloc_one.__pte_alloc.do_anonymous_page
      2.19 ±124%     -77.3%       0.50 ±300%  perf-sched.wait_time.max.ms.__cond_resched.__do_fault.do_read_fault.do_fault.__handle_mm_fault
     40.85 ±300%    -100.0%       0.00        perf-sched.wait_time.max.ms.__cond_resched.__kmem_cache_alloc_node.__kmalloc.acpi_ex_allocate_name_string.acpi_ex_get_name_string
      0.00          +2e+99%       0.00 ±300%  perf-sched.wait_time.max.ms.__cond_resched.__kmem_cache_alloc_node.__kmalloc_node.alloc_cpumask_var_node.__sched_setaffinity
      0.00       +5.5e+101%       0.55 ±300%  perf-sched.wait_time.max.ms.__cond_resched.__kmem_cache_alloc_node.__kmalloc_node.memcg_alloc_slab_cgroups.allocate_slab
      0.00 ±300%    +104.2%       0.00 ±200%  perf-sched.wait_time.max.ms.__cond_resched.__kmem_cache_alloc_node.kmalloc_trace.perf_event_mmap_event.perf_event_mmap
      0.00        +3.9e+99%       0.00 ±200%  perf-sched.wait_time.max.ms.__cond_resched.__kmem_cache_alloc_node.kmalloc_trace.single_open.single_open_size
      0.00         +1e+104%     100.01 ±299%  perf-sched.wait_time.max.ms.__cond_resched.__kmem_cache_alloc_node.kmalloc_trace.vmstat_start.seq_read_iter
      4898            -8.1%       4504 ± 25%  perf-sched.wait_time.max.ms.__cond_resched.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr.__sched_setaffinity
      0.84 ±132%     -63.0%       0.31 ±194%  perf-sched.wait_time.max.ms.__cond_resched.__wait_for_common.wait_for_completion_state.kernel_clone.__x64_sys_vfork
      0.15 ±300%    -100.0%       0.00        perf-sched.wait_time.max.ms.__cond_resched.cgroup_css_set_fork.cgroup_can_fork.copy_process.kernel_clone
      0.00 ±300%   +1100.0%       0.03 ±300%  perf-sched.wait_time.max.ms.__cond_resched.change_pmd_range.change_p4d_range.change_protection_range.mprotect_fixup
      0.00 ±300%    -100.0%       0.00        perf-sched.wait_time.max.ms.__cond_resched.dentry_kill.dput.step_into.link_path_walk
      0.16 ±148%    +731.9%       1.33 ±146%  perf-sched.wait_time.max.ms.__cond_resched.down_write.__split_vma.do_vmi_align_munmap.__do_sys_brk
      0.52 ± 53%    +238.0%       1.75 ±108%  perf-sched.wait_time.max.ms.__cond_resched.down_write.anon_vma_clone.__split_vma.do_vmi_align_munmap
      0.00        +2.2e+99%       0.00 ±300%  perf-sched.wait_time.max.ms.__cond_resched.down_write.anon_vma_clone.__split_vma.vma_modify
      0.00       +2.9e+101%       0.29 ±178%  perf-sched.wait_time.max.ms.__cond_resched.down_write.do_brk_flags.__do_sys_brk.do_syscall_64
      0.01 ±212%    -100.0%       0.00        perf-sched.wait_time.max.ms.__cond_resched.down_write.free_pgtables.exit_mmap.__mmput
      0.00 ±300%   +1245.2%       0.04 ±299%  perf-sched.wait_time.max.ms.__cond_resched.down_write.mmap_region.do_mmap.vm_mmap_pgoff
      0.06 ±199%    +321.8%       0.24 ±258%  perf-sched.wait_time.max.ms.__cond_resched.down_write.unlink_anon_vmas.free_pgtables.unmap_region
      0.00 ±299%    -100.0%       0.00        perf-sched.wait_time.max.ms.__cond_resched.down_write.unlink_file_vma.free_pgtables.exit_mmap
      0.08 ±119%     -37.4%       0.05 ±300%  perf-sched.wait_time.max.ms.__cond_resched.down_write.vma_prepare.__split_vma.do_vmi_align_munmap
      0.02 ±300%    +763.4%       0.20 ±247%  perf-sched.wait_time.max.ms.__cond_resched.down_write.vma_prepare.do_brk_flags.__do_sys_brk
      0.48 ±115%     +96.0%       0.94 ± 98%  perf-sched.wait_time.max.ms.__cond_resched.down_write_killable.__do_sys_brk.do_syscall_64.entry_SYSCALL_64_after_hwframe
      0.32 ±192%     -12.3%       0.28 ±276%  perf-sched.wait_time.max.ms.__cond_resched.down_write_killable.exec_mmap.begin_new_exec.load_elf_binary
      0.00 ±299%  +12272.7%       0.41 ±300%  perf-sched.wait_time.max.ms.__cond_resched.down_write_killable.setup_arg_pages.load_elf_binary.search_binary_handler
      0.21 ±299%    -100.0%       0.00        perf-sched.wait_time.max.ms.__cond_resched.down_write_killable.vm_mmap_pgoff.elf_map.load_elf_binary
      0.00 ±300%    -100.0%       0.00        perf-sched.wait_time.max.ms.__cond_resched.down_write_killable.vm_mmap_pgoff.elf_map.load_elf_interp
      0.38 ±300%    -100.0%       0.00        perf-sched.wait_time.max.ms.__cond_resched.down_write_killable.vm_mmap_pgoff.ksys_mmap_pgoff.do_syscall_64
      0.14 ±230%  +71629.0%     101.64 ±294%  perf-sched.wait_time.max.ms.__cond_resched.dput.__fput.__x64_sys_close.do_syscall_64
      0.01 ±122%    +290.4%       0.04 ±282%  perf-sched.wait_time.max.ms.__cond_resched.dput.step_into.link_path_walk.part
      0.58 ±206%     -99.7%       0.00 ±299%  perf-sched.wait_time.max.ms.__cond_resched.dput.terminate_walk.path_openat.do_filp_open
      0.00 ±201%    -100.0%       0.00        perf-sched.wait_time.max.ms.__cond_resched.exit_mmap.__mmput.exit_mm.do_exit
      0.00 ±299%    -100.0%       0.00        perf-sched.wait_time.max.ms.__cond_resched.filemap_read.__kernel_read.search_binary_handler.exec_binprm
      0.00 ±200%    -100.0%       0.00        perf-sched.wait_time.max.ms.__cond_resched.filemap_read.vfs_read.ksys_read.do_syscall_64
      0.40 ±300%    +961.6%       4.23 ±219%  perf-sched.wait_time.max.ms.__cond_resched.generic_perform_write.shmem_file_write_iter.vfs_write.ksys_write
      0.00 ±300%   +4310.0%       0.04 ±300%  perf-sched.wait_time.max.ms.__cond_resched.khugepaged.kthread.ret_from_fork.ret_from_fork_asm
      0.00 ±300%    -100.0%       0.00        perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc.__anon_vma_prepare.do_anonymous_page.__handle_mm_fault
      0.00       +3.9e+101%       0.39 ±300%  perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc.__anon_vma_prepare.wp_page_copy.__handle_mm_fault
      0.00         +5e+103%      50.14 ±300%  perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc.acpi_ut_create_generic_state.acpi_ds_result_push.acpi_ds_exec_end_op
      0.42 ±154%    -100.0%       0.00        perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc.alloc_pid.copy_process.kernel_clone
      0.53 ±101%    +124.1%       1.19 ± 90%  perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc.mas_alloc_nodes.mas_preallocate.__split_vma
      0.28 ±154%    +291.2%       1.11 ±157%  perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc.mas_alloc_nodes.mas_preallocate.do_brk_flags
      0.16 ±300%    -100.0%       0.00        perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc.mas_alloc_nodes.mas_preallocate.mmap_region
      0.00        +2.4e+99%       0.00 ±300%  perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc.mas_alloc_nodes.mas_preallocate.vma_expand
      0.00 ±299%    -100.0%       0.00        perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc.security_file_alloc.init_file.alloc_empty_file
      0.00 ±300%    -100.0%       0.00        perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc.vm_area_alloc.__install_special_mapping.map_vdso
      0.32 ±300%    -100.0%       0.00        perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc.vm_area_alloc.do_brk_flags.__do_sys_brk
      0.10 ±299%    -100.0%       0.00        perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc.vm_area_alloc.mmap_region.do_mmap
      1.20 ± 83%     +66.0%       2.00 ±101%  perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc.vm_area_dup.__split_vma.do_vmi_align_munmap
      0.00       +1.7e+101%       0.17 ±300%  perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc.vm_area_dup.__split_vma.vma_modify
      0.00         +1e+104%     100.01 ±300%  perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc.vm_area_dup.dup_mmap.dup_mm
      0.12 ± 92%     -16.6%       0.10 ±152%  perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc_bulk.mas_alloc_nodes.mas_preallocate.__split_vma
      0.00 ±300%  +38405.9%       0.65 ±244%  perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc_bulk.mas_alloc_nodes.mas_preallocate.do_brk_flags
      0.00       +2.1e+103%      20.73 ±300%  perf-sched.wait_time.max.ms.__cond_resched.mutex_lock.drm_property_free_blob.drm_gem_destroy_shadow_plane_state.drm_atomic_state_default_clear
      0.00 ±300%    -100.0%       0.00        perf-sched.wait_time.max.ms.__cond_resched.mutex_lock.futex_exit_release.exit_mm_release.exit_mm
      0.00 ±200%     -47.7%       0.00 ±300%  perf-sched.wait_time.max.ms.__cond_resched.mutex_lock.perf_event_ctx_lock_nested.constprop.0
      2.39 ± 81%     -18.7%       1.94 ±100%  perf-sched.wait_time.max.ms.__cond_resched.mutex_lock.perf_poll.do_poll.constprop
      0.00       +4.6e+101%       0.46 ±300%  perf-sched.wait_time.max.ms.__cond_resched.mutex_lock_killable.pcpu_alloc.mm_init.alloc_bprm
      0.00         +3e+104%     300.07 ±152%  perf-sched.wait_time.max.ms.__cond_resched.mutex_lock_killable.pcpu_alloc.mm_init.dup_mm
    204.66 ±300%      +0.1%     204.80 ±200%  perf-sched.wait_time.max.ms.__cond_resched.process_one_work.worker_thread.kthread.ret_from_fork
      0.00         +2e+104%     204.66 ±200%  perf-sched.wait_time.max.ms.__cond_resched.refresh_cpu_vm_stats.vmstat_update.process_one_work.worker_thread
      0.32 ±126%     +45.0%       0.46 ±237%  perf-sched.wait_time.max.ms.__cond_resched.remove_vma.do_vmi_align_munmap.__do_sys_brk.do_syscall_64
      0.00 ±300%  +1.6e+06%      73.17 ±128%  perf-sched.wait_time.max.ms.__cond_resched.shmem_get_folio_gfp.shmem_read_folio_gfp.drm_gem_get_pages.drm_gem_shmem_get_pages
      0.30 ±300%    +244.1%       1.02 ±220%  perf-sched.wait_time.max.ms.__cond_resched.shmem_inode_acct_blocks.shmem_alloc_and_add_folio.shmem_get_folio_gfp.shmem_write_begin
      0.00        +2.8e+99%       0.00 ±300%  perf-sched.wait_time.max.ms.__cond_resched.slab_pre_alloc_hook.constprop.0.kmem_cache_alloc_lru
     86.01 ±  6%      +5.5%      90.72 ±  4%  perf-sched.wait_time.max.ms.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
      1.75 ±105%     +80.5%       3.16 ±114%  perf-sched.wait_time.max.ms.__cond_resched.stop_one_cpu.sched_exec.bprm_execve.part
      0.00 ±300%    -100.0%       0.00        perf-sched.wait_time.max.ms.__cond_resched.switch_task_namespaces.do_exit.do_group_exit.__x64_sys_exit_group
      0.10 ±100%    +248.6%       0.35 ±258%  perf-sched.wait_time.max.ms.__cond_resched.task_work_run.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode
      0.00       +4.6e+100%       0.05 ±271%  perf-sched.wait_time.max.ms.__cond_resched.task_work_run.exit_to_user_mode_loop.exit_to_user_mode_prepare.syscall_exit_to_user_mode
      0.00         +1e+104%     100.06 ±300%  perf-sched.wait_time.max.ms.__cond_resched.task_work_run.get_signal.arch_do_signal_or_restart.exit_to_user_mode_loop
      0.40 ±295%     -98.7%       0.01 ±200%  perf-sched.wait_time.max.ms.__cond_resched.tlb_batch_pages_flush.tlb_finish_mmu.exit_mmap.__mmput
      2.03 ± 23%    +227.5%       6.66 ± 48%  perf-sched.wait_time.max.ms.__cond_resched.tlb_batch_pages_flush.tlb_finish_mmu.unmap_region.constprop
      0.02 ±300%   +1253.6%       0.30 ±300%  perf-sched.wait_time.max.ms.__cond_resched.unmap_page_range.unmap_vmas.unmap_region.constprop
      1.59 ± 53%    +153.9%       4.04 ± 57%  perf-sched.wait_time.max.ms.__cond_resched.unmap_vmas.unmap_region.constprop.0
      0.00 ±200%    -100.0%       0.00        perf-sched.wait_time.max.ms.__cond_resched.wp_page_copy.__handle_mm_fault.handle_mm_fault.do_user_addr_fault
      5.09 ±300%    -100.0%       0.00        perf-sched.wait_time.max.ms.__cond_resched.ww_mutex_lock.drm_gem_vunmap_unlocked.drm_fbdev_generic_helper_fb_dirty.drm_fb_helper_damage_work
      0.00       +3.7e+100%       0.04 ±253%  perf-sched.wait_time.max.ms.__cond_resched.ww_mutex_lock.drm_gem_vunmap_unlocked.drm_gem_fb_vunmap.drm_atomic_helper_cleanup_planes
      1.18 ± 46%  +17108.0%     202.73 ±197%  perf-sched.wait_time.max.ms.__cond_resched.zap_pmd_range.isra.0.unmap_page_range
    802.26 ± 49%     -49.8%     402.51 ±120%  perf-sched.wait_time.max.ms.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
      4.73 ± 40%     -67.8%       1.52 ±154%  perf-sched.wait_time.max.ms.devkmsg_read.vfs_read.ksys_read.do_syscall_64
    702.00 ± 65%     -14.3%     601.71 ± 81%  perf-sched.wait_time.max.ms.do_nanosleep.hrtimer_nanosleep.common_nsleep.__x64_sys_clock_nanosleep
      1321 ± 48%     +37.2%       1813 ± 54%  perf-sched.wait_time.max.ms.do_task_dead.do_exit.do_group_exit.__x64_sys_exit_group.do_syscall_64
      6.90 ±  9%   +1483.8%     109.24 ±272%  perf-sched.wait_time.max.ms.do_wait.kernel_wait4.__do_sys_wait4.do_syscall_64
      0.00       +2.6e+101%       0.26 ±300%  perf-sched.wait_time.max.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_common_interrupt
      2.33 ± 39%   +4436.5%     105.58 ±282%  perf-sched.wait_time.max.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_exc_page_fault
    128.51 ±225%      +2.5%     131.68 ±219%  perf-sched.wait_time.max.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt
      2.44 ± 12%     +89.7%       4.63 ± 50%  perf-sched.wait_time.max.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_call_function_single
      5.66 ±156%     +54.5%       8.74 ± 54%  perf-sched.wait_time.max.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_reschedule_ipi
      1000           +38.5%       1386 ± 83%  perf-sched.wait_time.max.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.syscall_exit_to_user_mode.do_syscall_64
      1.41 ±300%    -100.0%       0.00        perf-sched.wait_time.max.ms.futex_wait_queue.futex_wait.do_futex.__x64_sys_futex
      0.00 ±300%  +16133.3%       0.39 ±300%  perf-sched.wait_time.max.ms.io_schedule.folio_wait_bit_common.filemap_fault.__do_fault
      1051            +0.8%       1059        perf-sched.wait_time.max.ms.pipe_read.vfs_read.ksys_read.do_syscall_64
      1004            -0.0%       1003        perf-sched.wait_time.max.ms.schedule_hrtimeout_range_clock.do_poll.constprop.0.do_sys_poll
    499.84            +0.0%     499.86        perf-sched.wait_time.max.ms.schedule_hrtimeout_range_clock.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
    319.32 ±299%    -100.0%       0.00        perf-sched.wait_time.max.ms.schedule_preempt_disabled.rwsem_down_read_slowpath.down_read_killable.__access_remote_vm
      0.01 ±122%      +4.8%       0.01 ±123%  perf-sched.wait_time.max.ms.schedule_timeout.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr
      8.87 ±  6%     +50.9%      13.38 ± 50%  perf-sched.wait_time.max.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone
    504.57            +0.1%     505.25        perf-sched.wait_time.max.ms.schedule_timeout.kcompactd.kthread.ret_from_fork
      0.18 ±300%    +102.3%       0.36 ±299%  perf-sched.wait_time.max.ms.schedule_timeout.khugepaged_wait_work.khugepaged.kthread
     14.61 ± 99%      +0.9%      14.74 ± 73%  perf-sched.wait_time.max.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
      1053           +10.8%       1167 ± 26%  perf-sched.wait_time.max.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
      7.67 ± 19%     -71.1%       2.22 ±123%  perf-sched.wait_time.max.ms.syslog_print.do_syslog.kmsg_read.vfs_read
      3.55 ± 12%     +13.4%       4.03 ± 29%  perf-sched.wait_time.max.ms.wait_for_partner.fifo_open.do_dentry_open.do_open
      2046            +0.0%       2046        perf-sched.wait_time.max.ms.worker_thread.kthread.ret_from_fork.ret_from_fork_asm

xsang@inn:~$

[-- Attachment #3: ncompare-file-ioctl --]
[-- Type: text/plain, Size: 207251 bytes --]


xsang@inn:~$ lkp ncompare -s commit=1f4f7f0f8845d/tbox_group=lkp-skl-d08/testcase=stress-ng -o commit=6ccdcb6d3a741  -a | grep -v interrupts | grep -v softirq
=========================================================================================
class/compiler/cpufreq_governor/disk/fs/kconfig/nr_threads/rootfs/tbox_group/test/testcase/testtime:
  filesystem/gcc-12/performance/1SSD/btrfs/x86_64-rhel-8.3/10%/debian-11.1-x86_64-20220510.cgz/lkp-skl-d08/file-ioctl/stress-ng/60s

commit:
  1f4f7f0f8845dbac40289cc3d50b81314c5a12b8
  6ccdcb6d3a741c4e005ca6ffd4a62ddf8b5bead3

1f4f7f0f8845dbac 6ccdcb6d3a741c4e005ca6ffd4a
---------------- ---------------------------
         %stddev     %change         %stddev
             \          |                \
     10.99            +0.0%      10.99        kmsg.timestamp:EDAC_skx:ECC_is_disabled_on_imc
     75.12            +2.0%      76.60 ±  3%  kmsg.timestamp:last
     26.34 ±  2%      +5.6%      27.82 ± 10%  kmsg.timestamp:x86/PAT:bmc-watchdog:#map_pfn_expected_mapping_type_uncached-minus_for[mem#-#],got_write-back
     75.12            +2.0%      76.60 ±  3%  dmesg.timestamp:last
 2.143e+09            +5.1%  2.252e+09 ±  9%  cpuidle..time
   3124995            +3.6%    3237684 ±  7%  cpuidle..usage
     91.83            +5.0%      96.44 ±  7%  uptime.boot
      3043            +5.3%       3205 ±  8%  uptime.idle
     24.14            -0.2%      24.08        boot-time.boot
     10.14            -0.3%      10.11        boot-time.dhcp
    826.15            -0.3%     823.94        boot-time.idle
     12.62 ± 40%     +17.2%      14.80 ± 44%  perf-c2c.DRAM.local
    130.38 ± 10%     +26.7%     165.20 ± 16%  perf-c2c.HITM.local
    130.38 ± 10%     +26.7%     165.20 ± 16%  perf-c2c.HITM.total
     91.17            +0.2%      91.37        iostat.cpu.idle
      0.05           +39.9%       0.07 ± 38%  iostat.cpu.iowait
      8.11            -1.8%       7.96 ±  5%  iostat.cpu.system
      0.67 ±  3%     -11.3%       0.59 ± 14%  iostat.cpu.user
     90.99            +0.2       91.21        mpstat.cpu.all.idle%
      0.05            +0.0        0.07 ± 38%  mpstat.cpu.all.iowait%
      0.79 ±  3%      +0.1        0.87 ± 20%  mpstat.cpu.all.irq%
      0.04 ±  5%      +0.0        0.05 ± 20%  mpstat.cpu.all.soft%
      7.48            -0.2        7.23 ±  8%  mpstat.cpu.all.sys%
      0.65 ±  3%      -0.1        0.58 ± 14%  mpstat.cpu.all.usr%
     62.04            +0.0%      62.04        time.elapsed_time
     62.04            +0.0%      62.04        time.elapsed_time.max
      2112            +0.0%       2112        time.file_system_inputs
     96.00            +0.0%      96.00        time.file_system_outputs
      6184 ±  7%     -23.0%       4761 ±  5%  time.involuntary_context_switches
      0.62 ±210%     -52.0%       0.30 ±213%  time.major_page_faults
      6912            +0.0%       6912        time.maximum_resident_set_size
      9509            -0.1%       9496        time.minor_page_faults
      4096            +0.0%       4096        time.page_size
    286.00            +0.0%     286.00        time.percent_of_cpu_this_job_got
    172.84            +0.3%     173.27        time.system_time
      4.68            -8.0%       4.31        time.user_time
    123062 ±  5%      -8.5%     112627 ±  5%  time.voluntary_context_switches
     91.20            +0.2%      91.42        vmstat.cpu.id
      8.25            -2.1%       8.08 ±  5%  vmstat.cpu.sy
      0.49 ±  5%     -12.7%       0.43 ± 20%  vmstat.cpu.us
      0.05 ± 13%     +32.8%       0.07 ± 36%  vmstat.cpu.wa
     15.91            -3.6%      15.33 ±  8%  vmstat.io.bi
      6.65 ± 76%     +49.4%       9.93 ± 69%  vmstat.io.bo
      1034            +0.1%       1035        vmstat.memory.buff
   2954092            -0.0%    2953196        vmstat.memory.cache
  28985158            +0.0%   28997431        vmstat.memory.free
      0.02 ± 77%     +79.8%       0.03 ± 55%  vmstat.procs.b
      3.19 ±  2%      -3.3%       3.08 ±  8%  vmstat.procs.r
     33719            -4.5%      32188 ±  7%  vmstat.system.cs
     51824            -1.1%      51271 ±  2%  vmstat.system.in
    340971            -6.9%     317411        stress-ng.file-ioctl.ops
      5682            -6.9%       5290        stress-ng.file-ioctl.ops_per_sec
     62.04            +0.0%      62.04        stress-ng.time.elapsed_time
     62.04            +0.0%      62.04        stress-ng.time.elapsed_time.max
      2112            +0.0%       2112        stress-ng.time.file_system_inputs
     96.00            +0.0%      96.00        stress-ng.time.file_system_outputs
      6184 ±  7%     -23.0%       4761 ±  5%  stress-ng.time.involuntary_context_switches
      0.62 ±210%     -52.0%       0.30 ±213%  stress-ng.time.major_page_faults
      6912            +0.0%       6912        stress-ng.time.maximum_resident_set_size
      9509            -0.1%       9496        stress-ng.time.minor_page_faults
      4096            +0.0%       4096        stress-ng.time.page_size
    286.00            +0.0%     286.00        stress-ng.time.percent_of_cpu_this_job_got
    172.84            +0.3%     173.27        stress-ng.time.system_time
      4.68            -8.0%       4.31        stress-ng.time.user_time
    123062 ±  5%      -8.5%     112627 ±  5%  stress-ng.time.voluntary_context_switches
    357.50            -3.4%     345.30 ±  6%  turbostat.Avg_MHz
      9.56            -0.2        9.38 ±  3%  turbostat.Busy%
      3740            -1.8%       3672 ±  3%  turbostat.Bzy_MHz
     61848            -0.6%      61459        turbostat.C1
      0.14 ±  6%      +0.0        0.14 ± 10%  turbostat.C1%
    926012            -0.1%     924627        turbostat.C1E
      2.70            -0.1        2.58 ±  6%  turbostat.C1E%
   2097927            +5.4%    2211048 ± 10%  turbostat.C6
     87.94            +0.4       88.29        turbostat.C6%
     88.61            -1.3%      87.45 ±  3%  turbostat.CPU%c1
      1.83 ± 21%     +72.6%       3.16 ± 95%  turbostat.CPU%c6
     34.12            +4.9%      35.80 ±  3%  turbostat.CoreTmp
      0.43            -2.7%       0.42 ±  2%  turbostat.IPC
   3473046            +3.2%    3585079 ±  6%  turbostat.IRQ
     35417 ±  9%      +3.2%      36563 ±  9%  turbostat.POLL
      0.00 ±264%      -0.0        0.00        turbostat.POLL%
      1.58 ± 24%     +59.7%       2.53 ± 95%  turbostat.Pkg%pc2
     33.75 ±  3%      +3.4%      34.90 ±  2%  turbostat.PkgTmp
     87.13            -2.7%      84.80 ±  5%  turbostat.PkgWatt
      0.86            +1.5%       0.87        turbostat.RAMWatt
      3001            -0.0%       3000        turbostat.TSC_MHz
     69867            -0.7%      69398 ±  5%  meminfo.Active
     66909            -0.6%      66489 ±  5%  meminfo.Active(anon)
      2958            -1.7%       2908 ±  6%  meminfo.Active(file)
     55351 ±  9%     +11.3%      61587 ± 17%  meminfo.AnonHugePages
    297646            -0.3%     296691        meminfo.AnonPages
      1034            +0.1%       1035        meminfo.Buffers
   2874923            -0.0%    2873986        meminfo.Cached
  16281020            +0.0%   16281020        meminfo.CommitLimit
    566106            -1.1%     559898        meminfo.Committed_AS
  27525120 ±  3%      -0.2%   27472691 ±  3%  meminfo.DirectMap1G
   5554944 ± 18%      +1.0%    5609062 ± 16%  meminfo.DirectMap2M
    166176 ±  7%      -1.0%     164486 ± 11%  meminfo.DirectMap4k
     62.51 ± 16%     -10.6%      55.88        meminfo.Dirty
      2048            +0.0%       2048        meminfo.Hugepagesize
    307749            -0.5%     306284        meminfo.Inactive
    305087            -0.5%     303694        meminfo.Inactive(anon)
      2662 ±  2%      -2.7%       2589 ±  3%  meminfo.Inactive(file)
     78478            +0.1%      78579        meminfo.KReclaimable
      9050            +0.3%       9075        meminfo.KernelStack
     41083            -1.2%      40591 ±  2%  meminfo.Mapped
  28759154            +0.0%   28770515        meminfo.MemAvailable
  28981905            +0.0%   28993284        meminfo.MemFree
  32562044            +0.0%   32562044        meminfo.MemTotal
   3580138            -0.3%    3568759        meminfo.Memused
     10.86            -3.3%      10.50 ±  8%  meminfo.Mlocked
      5273            -1.0%       5220 ±  2%  meminfo.PageTables
     16259            -0.0%      16253        meminfo.Percpu
     78478            +0.1%      78579        meminfo.SReclaimable
     90009            +0.3%      90246        meminfo.SUnreclaim
     74602            -1.1%      73789 ±  4%  meminfo.Shmem
    168487            +0.2%     168825        meminfo.Slab
   2795729            -0.0%    2795724        meminfo.Unevictable
 3.436e+10            +0.0%  3.436e+10        meminfo.VmallocTotal
     20749            +0.2%      20783        meminfo.VmallocUsed
      0.77 ±200%      -6.4%       0.73 ±134%  meminfo.Writeback
   3758317            +0.1%    3762244        meminfo.max_used_kB
     70.38 ±  8%      -3.2%      68.10 ± 13%  proc-vmstat.direct_map_level2_splits
      2.75 ± 35%     -20.0%       2.20 ± 34%  proc-vmstat.direct_map_level3_splits
     16707            -0.5%      16622 ±  5%  proc-vmstat.nr_active_anon
    740.83            -1.9%     727.11 ±  6%  proc-vmstat.nr_active_file
     74446            -0.4%      74173        proc-vmstat.nr_anon_pages
     26.99 ±  9%     +11.4%      30.07 ± 17%  proc-vmstat.nr_anon_transparent_hugepages
    148.88 ± 65%     +53.4%     228.40 ± 86%  proc-vmstat.nr_dirtied
     15.63 ± 16%     -10.6%      13.97        proc-vmstat.nr_dirty
    717010            +0.0%     717280        proc-vmstat.nr_dirty_background_threshold
   1435774            +0.0%    1436316        proc-vmstat.nr_dirty_threshold
    718972            -0.0%     718748        proc-vmstat.nr_file_pages
      0.00         +1e+101%       0.10 ±300%  proc-vmstat.nr_foll_pin_acquired
      0.00         +1e+101%       0.10 ±300%  proc-vmstat.nr_foll_pin_released
   7245424            +0.0%    7248168        proc-vmstat.nr_free_pages
     76303            -0.5%      75925        proc-vmstat.nr_inactive_anon
    662.04 ±  3%      -3.0%     642.06 ±  3%  proc-vmstat.nr_inactive_file
      9051            +0.3%       9076        proc-vmstat.nr_kernel_stack
     10282            -1.3%      10149 ±  2%  proc-vmstat.nr_mapped
      2.72            -3.5%       2.62 ±  8%  proc-vmstat.nr_mlock
      1318            -1.0%       1305 ±  2%  proc-vmstat.nr_page_table_pages
     18634            -1.0%      18448 ±  4%  proc-vmstat.nr_shmem
     19618            +0.1%      19644        proc-vmstat.nr_slab_reclaimable
     22503            +0.3%      22561        proc-vmstat.nr_slab_unreclaimable
    698932            -0.0%     698931        proc-vmstat.nr_unevictable
      0.19 ±200%      -6.4%       0.18 ±134%  proc-vmstat.nr_writeback
    122.62 ± 66%     +55.4%     190.60 ± 76%  proc-vmstat.nr_written
     16707            -0.5%      16622 ±  5%  proc-vmstat.nr_zone_active_anon
    740.83            -1.9%     727.11 ±  6%  proc-vmstat.nr_zone_active_file
     76303            -0.5%      75925        proc-vmstat.nr_zone_inactive_anon
    662.04 ±  3%      -3.0%     642.06 ±  3%  proc-vmstat.nr_zone_inactive_file
    698932            -0.0%     698931        proc-vmstat.nr_zone_unevictable
     15.58 ± 15%      -9.5%      14.09 ±  2%  proc-vmstat.nr_zone_write_pending
  89530411            -7.1%   83208879        proc-vmstat.numa_hit
      0.00          -100.0%       0.00        proc-vmstat.numa_interleave
  89301961            -7.0%   83009577        proc-vmstat.numa_local
     26399            -2.7%      25675        proc-vmstat.pgactivate
      0.00          -100.0%       0.00        proc-vmstat.pgalloc_dma32
  93091135            -6.9%   86691043        proc-vmstat.pgalloc_normal
      0.00          -100.0%       0.00        proc-vmstat.pgdeactivate
    212409            +1.2%     215026 ±  3%  proc-vmstat.pgfault
  93049512            -6.9%   86647969        proc-vmstat.pgfree
      1056            +0.0%       1056        proc-vmstat.pgpgin
    440.00 ± 76%     +65.1%     726.40 ± 77%  proc-vmstat.pgpgout
      8496            +2.4%       8699 ±  4%  proc-vmstat.pgreuse
      0.00          -100.0%       0.00        proc-vmstat.pgrotated
     46.75 ±  4%      +9.1%      51.00 ± 11%  proc-vmstat.thp_collapse_alloc
      0.00          -100.0%       0.00        proc-vmstat.thp_deferred_split_page
      1.50 ± 74%     -46.7%       0.80 ± 93%  proc-vmstat.thp_fault_alloc
      0.00          -100.0%       0.00        proc-vmstat.thp_split_pmd
      0.00          -100.0%       0.00        proc-vmstat.thp_zero_page_alloc
     14.00            +0.0%      14.00        proc-vmstat.unevictable_pgs_culled
      3.00            +0.0%       3.00        proc-vmstat.unevictable_pgs_mlocked
      3.00            +0.0%       3.00        proc-vmstat.unevictable_pgs_munlocked
      0.13 ± 62%    +299.7%       0.51 ±132%  perf-stat.i.MPKI
 3.868e+09            -6.1%  3.632e+09 ±  8%  perf-stat.i.branch-instructions
      0.69 ± 17%      +0.6        1.26 ± 85%  perf-stat.i.branch-miss-rate%
  21656746 ±  2%      -7.3%   20079062 ±  9%  perf-stat.i.branch-misses
      1.38 ± 12%      +0.2        1.61 ± 23%  perf-stat.i.cache-miss-rate%
   1243295 ± 18%     +17.9%    1465860 ±  9%  perf-stat.i.cache-misses
 2.181e+08 ±  2%      +0.5%  2.192e+08 ±  7%  perf-stat.i.cache-references
     35135            -4.7%      33477 ±  8%  perf-stat.i.context-switches
      0.74 ± 12%     +67.7%       1.24 ± 71%  perf-stat.i.cpi
     36006            -0.0%      36004        perf-stat.i.cpu-clock
 1.247e+10            -3.6%  1.203e+10 ±  6%  perf-stat.i.cpu-cycles
     65.35            -1.7%      64.26 ±  4%  perf-stat.i.cpu-migrations
     24611 ± 31%     -29.1%      17449 ± 17%  perf-stat.i.cycles-between-cache-misses
      0.02 ± 68%      +0.1        0.08 ±143%  perf-stat.i.dTLB-load-miss-rate%
    311590 ±  2%      +4.9%     326828 ± 14%  perf-stat.i.dTLB-load-misses
 4.523e+09            -6.5%  4.227e+09 ±  8%  perf-stat.i.dTLB-loads
      0.00 ± 92%      +0.0        0.01 ±149%  perf-stat.i.dTLB-store-miss-rate%
      9906 ±  6%     +34.2%      13292 ± 45%  perf-stat.i.dTLB-store-misses
 2.698e+09            -6.0%  2.536e+09 ±  8%  perf-stat.i.dTLB-stores
     23.04 ±  4%      +1.5       24.52 ± 17%  perf-stat.i.iTLB-load-miss-rate%
    338747 ± 21%      -9.7%     305803 ± 14%  perf-stat.i.iTLB-load-misses
   1150637 ± 14%      -9.2%    1044691 ±  8%  perf-stat.i.iTLB-loads
 1.885e+10            -5.9%  1.775e+10 ±  8%  perf-stat.i.instructions
     57534 ± 14%      +1.4%      58352 ± 11%  perf-stat.i.instructions-per-iTLB-miss
      1.50            -5.5%       1.41 ±  8%  perf-stat.i.ipc
      0.28 ± 54%     +75.9%       0.49 ± 22%  perf-stat.i.major-faults
      0.35            -3.6%       0.33 ±  6%  perf-stat.i.metric.GHz
     68.15 ± 26%    +110.8%     143.65 ± 93%  perf-stat.i.metric.K/sec
    314.01            -6.2%     294.69 ±  8%  perf-stat.i.metric.M/sec
      1925            -3.0%       1867 ±  4%  perf-stat.i.minor-faults
    109386 ± 22%      -3.9%     105144 ±  7%  perf-stat.i.node-loads
    166651 ± 31%     +85.0%     308368 ±  4%  perf-stat.i.node-stores
      1925            -3.0%       1868 ±  4%  perf-stat.i.page-faults
     36006            -0.0%      36004        perf-stat.i.task-clock
      0.07 ± 19%     +26.7%       0.08 ± 16%  perf-stat.overall.MPKI
      0.56 ±  2%      -0.0        0.55 ±  8%  perf-stat.overall.branch-miss-rate%
      0.57 ± 20%      +0.1        0.67 ± 14%  perf-stat.overall.cache-miss-rate%
      0.66            +2.6%       0.68        perf-stat.overall.cpi
     10338 ± 15%     -19.7%       8301 ± 13%  perf-stat.overall.cycles-between-cache-misses
      0.01 ±  2%      +0.0        0.01 ± 24%  perf-stat.overall.dTLB-load-miss-rate%
      0.00 ±  7%      +0.0        0.00 ± 55%  perf-stat.overall.dTLB-store-miss-rate%
     22.61 ±  3%      +0.0       22.64 ± 10%  perf-stat.overall.iTLB-load-miss-rate%
     57455 ± 14%      +2.2%      58724 ± 11%  perf-stat.overall.instructions-per-iTLB-miss
      1.51            -2.5%       1.47        perf-stat.overall.ipc
 3.808e+09            -6.1%  3.578e+09 ±  8%  perf-stat.ps.branch-instructions
  21311538 ±  2%      -7.2%   19773267 ±  9%  perf-stat.ps.branch-misses
   1223635 ± 18%     +18.0%    1443696 ±  9%  perf-stat.ps.cache-misses
 2.148e+08 ±  2%      +0.5%  2.159e+08 ±  6%  perf-stat.ps.cache-references
     34590            -4.7%      32976 ±  7%  perf-stat.ps.context-switches
     35446            +0.1%      35466        perf-stat.ps.cpu-clock
 1.228e+10            -3.5%  1.185e+10 ±  6%  perf-stat.ps.cpu-cycles
     64.33            -1.6%      63.29 ±  3%  perf-stat.ps.cpu-migrations
    306740 ±  2%      +5.0%     321949 ± 14%  perf-stat.ps.dTLB-load-misses
 4.453e+09            -6.5%  4.164e+09 ±  8%  perf-stat.ps.dTLB-loads
      9745 ±  6%     +34.3%      13089 ± 45%  perf-stat.ps.dTLB-store-misses
 2.656e+09            -5.9%  2.498e+09 ±  8%  perf-stat.ps.dTLB-stores
    333519 ± 21%      -9.7%     301231 ± 14%  perf-stat.ps.iTLB-load-misses
   1132855 ± 14%      -9.2%    1029084 ±  8%  perf-stat.ps.iTLB-loads
 1.856e+10            -5.8%  1.748e+10 ±  8%  perf-stat.ps.instructions
      0.28 ± 54%     +76.0%       0.48 ± 22%  perf-stat.ps.major-faults
      1894            -2.9%       1838 ±  4%  perf-stat.ps.minor-faults
    107669 ± 22%      -3.8%     103537 ±  7%  perf-stat.ps.node-loads
    164076 ± 31%     +85.2%     303821 ±  4%  perf-stat.ps.node-stores
      1894            -2.9%       1839 ±  4%  perf-stat.ps.page-faults
     35446            +0.1%      35466        perf-stat.ps.task-clock
 1.201e+12            -2.2%  1.175e+12        perf-stat.total.instructions
      9668 ± 38%     +21.7%      11761        sched_debug.cfs_rq:/.avg_vruntime.avg
     73804 ± 37%     +22.9%      90724 ±  8%  sched_debug.cfs_rq:/.avg_vruntime.max
    761.77 ± 37%     +36.5%       1039 ± 20%  sched_debug.cfs_rq:/.avg_vruntime.min
     15647 ± 39%     +24.9%      19545 ±  4%  sched_debug.cfs_rq:/.avg_vruntime.stddev
      0.34 ± 22%     -18.3%       0.27 ±  6%  sched_debug.cfs_rq:/.h_nr_running.avg
      1.44 ± 47%     -20.0%       1.15 ± 19%  sched_debug.cfs_rq:/.h_nr_running.max
      0.48 ± 15%      -9.2%       0.43 ±  5%  sched_debug.cfs_rq:/.h_nr_running.stddev
     19.19 ±141%    +560.7%     126.76 ±184%  sched_debug.cfs_rq:/.left_vruntime.avg
    549.96 ±132%    +700.5%       4402 ±189%  sched_debug.cfs_rq:/.left_vruntime.max
      0.00            +0.0%       0.00        sched_debug.cfs_rq:/.left_vruntime.min
     94.06 ±133%    +671.5%     725.66 ±189%  sched_debug.cfs_rq:/.left_vruntime.stddev
     40431 ± 59%     -19.3%      32628 ± 34%  sched_debug.cfs_rq:/.load.avg
    664217 ±106%     -50.6%     327942 ± 70%  sched_debug.cfs_rq:/.load.max
    133230 ± 85%     -35.2%      86300 ± 54%  sched_debug.cfs_rq:/.load.stddev
      5030 ± 34%     -20.8%       3981 ±  3%  sched_debug.cfs_rq:/.load_avg.avg
     14574 ± 32%     -20.9%      11525 ±  3%  sched_debug.cfs_rq:/.load_avg.max
      1.06 ±197%     -10.6%       0.95 ±162%  sched_debug.cfs_rq:/.load_avg.min
      6600 ± 34%     -21.3%       5193 ±  2%  sched_debug.cfs_rq:/.load_avg.stddev
      9668 ± 38%     +21.7%      11761        sched_debug.cfs_rq:/.min_vruntime.avg
     73804 ± 37%     +22.9%      90727 ±  8%  sched_debug.cfs_rq:/.min_vruntime.max
    761.77 ± 37%     +36.5%       1039 ± 20%  sched_debug.cfs_rq:/.min_vruntime.min
     15647 ± 39%     +24.9%      19545 ±  4%  sched_debug.cfs_rq:/.min_vruntime.stddev
      0.34 ± 22%     -18.3%       0.27 ±  6%  sched_debug.cfs_rq:/.nr_running.avg
      1.44 ± 47%     -20.0%       1.15 ± 19%  sched_debug.cfs_rq:/.nr_running.max
      0.48 ± 15%      -9.2%       0.43 ±  5%  sched_debug.cfs_rq:/.nr_running.stddev
     24.80 ± 79%     +82.3%      45.20 ± 59%  sched_debug.cfs_rq:/.removed.load_avg.avg
    637.94 ± 53%     -11.9%     562.10 ± 49%  sched_debug.cfs_rq:/.removed.load_avg.max
    120.18 ± 63%     +20.1%     144.33 ± 49%  sched_debug.cfs_rq:/.removed.load_avg.stddev
      7.66 ± 68%    +140.4%      18.41 ± 64%  sched_debug.cfs_rq:/.removed.runnable_avg.avg
    270.06 ± 66%      -1.4%     266.15 ± 57%  sched_debug.cfs_rq:/.removed.runnable_avg.max
     44.38 ± 66%     +36.4%      60.54 ± 56%  sched_debug.cfs_rq:/.removed.runnable_avg.stddev
      7.65 ± 68%    +140.5%      18.41 ± 64%  sched_debug.cfs_rq:/.removed.util_avg.avg
    269.94 ± 66%      -1.4%     266.10 ± 57%  sched_debug.cfs_rq:/.removed.util_avg.max
     44.36 ± 66%     +36.5%      60.53 ± 56%  sched_debug.cfs_rq:/.removed.util_avg.stddev
     19.19 ±141%    +560.6%     126.76 ±184%  sched_debug.cfs_rq:/.right_vruntime.avg
    550.03 ±132%    +700.4%       4402 ±189%  sched_debug.cfs_rq:/.right_vruntime.max
      0.00            +0.0%       0.00        sched_debug.cfs_rq:/.right_vruntime.min
     94.07 ±133%    +671.4%     725.66 ±189%  sched_debug.cfs_rq:/.right_vruntime.stddev
    480.62 ± 26%     -15.5%     406.33 ±  3%  sched_debug.cfs_rq:/.runnable_avg.avg
      1462 ± 15%     -10.4%       1310 ±  9%  sched_debug.cfs_rq:/.runnable_avg.max
     16.62 ±140%     -39.5%      10.05 ±180%  sched_debug.cfs_rq:/.runnable_avg.min
    373.33 ± 16%     -11.4%     330.95 ±  7%  sched_debug.cfs_rq:/.runnable_avg.stddev
      0.00 ±228%    -100.0%       0.00        sched_debug.cfs_rq:/.spread.avg
      0.07 ±228%    -100.0%       0.00        sched_debug.cfs_rq:/.spread.max
      0.01 ±228%    -100.0%       0.00        sched_debug.cfs_rq:/.spread.stddev
    476.46 ± 26%     -15.4%     403.03 ±  2%  sched_debug.cfs_rq:/.util_avg.avg
      1454 ± 15%     -10.4%       1303 ±  9%  sched_debug.cfs_rq:/.util_avg.max
     15.75 ±142%     -42.9%       9.00 ±201%  sched_debug.cfs_rq:/.util_avg.min
    371.95 ± 16%     -11.4%     329.56 ±  8%  sched_debug.cfs_rq:/.util_avg.stddev
     83.12 ± 15%      -5.3%      78.72 ± 13%  sched_debug.cfs_rq:/.util_est_enqueued.avg
    836.69 ±  3%      -0.7%     830.70 ±  6%  sched_debug.cfs_rq:/.util_est_enqueued.max
    215.18 ± 12%      +1.9%     219.21 ±  5%  sched_debug.cfs_rq:/.util_est_enqueued.stddev
    577964 ± 14%      +5.2%     608099 ±  3%  sched_debug.cpu.avg_idle.avg
    985615 ±  2%      +0.2%     988020 ±  2%  sched_debug.cpu.avg_idle.max
     28127 ±134%     -81.3%       5250 ± 49%  sched_debug.cpu.avg_idle.min
    274888 ± 10%      +7.8%     296303 ±  4%  sched_debug.cpu.avg_idle.stddev
     48833 ± 26%     +18.4%      57815 ±  5%  sched_debug.cpu.clock.avg
     48834 ± 26%     +18.4%      57817 ±  5%  sched_debug.cpu.clock.max
     48832 ± 26%     +18.4%      57814 ±  5%  sched_debug.cpu.clock.min
      0.80 ±  6%      +3.2%       0.82 ±  9%  sched_debug.cpu.clock.stddev
     48639 ± 25%     +18.3%      57562 ±  5%  sched_debug.cpu.clock_task.avg
     48681 ± 25%     +18.3%      57610 ±  5%  sched_debug.cpu.clock_task.max
     48275 ± 26%     +18.5%      57194 ±  5%  sched_debug.cpu.clock_task.min
     74.28 ±  8%      +6.2%      78.89 ±  2%  sched_debug.cpu.clock_task.stddev
    972.91 ± 15%      -9.7%     878.68 ±  9%  sched_debug.cpu.curr->pid.avg
      3815 ±  7%      +4.4%       3984        sched_debug.cpu.curr->pid.max
      1478 ±  2%      -1.4%       1457        sched_debug.cpu.curr->pid.stddev
    500000            +0.0%     500000        sched_debug.cpu.max_idle_balance_cost.avg
    500000            +0.0%     500000        sched_debug.cpu.max_idle_balance_cost.max
    500000            +0.0%     500000        sched_debug.cpu.max_idle_balance_cost.min
      4294            +0.0%       4294        sched_debug.cpu.next_balance.avg
      4294            +0.0%       4294        sched_debug.cpu.next_balance.max
      4294            +0.0%       4294        sched_debug.cpu.next_balance.min
      0.00 ± 19%      -6.3%       0.00 ± 21%  sched_debug.cpu.next_balance.stddev
      0.32 ± 21%     -14.4%       0.28 ±  8%  sched_debug.cpu.nr_running.avg
      1.44 ± 47%     -20.0%       1.15 ± 19%  sched_debug.cpu.nr_running.max
      0.47 ± 15%      -8.9%       0.43 ±  4%  sched_debug.cpu.nr_running.stddev
     28149 ± 48%     +26.3%      35546        sched_debug.cpu.nr_switches.avg
    117924 ± 41%     +26.2%     148828 ± 16%  sched_debug.cpu.nr_switches.max
      1171 ± 52%     +24.6%       1459 ± 38%  sched_debug.cpu.nr_switches.min
     27983 ± 44%     +25.6%      35144 ± 15%  sched_debug.cpu.nr_switches.stddev
      0.00 ±264%     -20.0%       0.00 ±300%  sched_debug.cpu.nr_uninterruptible.avg
     18.56 ± 30%     +11.5%      20.70 ± 16%  sched_debug.cpu.nr_uninterruptible.max
    -12.56            -3.3%     -12.15        sched_debug.cpu.nr_uninterruptible.min
      6.74 ± 19%      +0.4%       6.77 ±  9%  sched_debug.cpu.nr_uninterruptible.stddev
     48832 ± 26%     +18.4%      57814 ±  5%  sched_debug.cpu_clk
    996147            +0.0%     996147        sched_debug.dl_rq:.dl_bw->bw.avg
    996147            +0.0%     996147        sched_debug.dl_rq:.dl_bw->bw.max
    996147            +0.0%     996147        sched_debug.dl_rq:.dl_bw->bw.min
 4.295e+09            +0.0%  4.295e+09        sched_debug.jiffies
     48753 ± 26%     +18.4%      57735 ±  5%  sched_debug.ktime
    950.00            +0.0%     950.00        sched_debug.rt_rq:.rt_runtime.avg
    950.00            +0.0%     950.00        sched_debug.rt_rq:.rt_runtime.max
    950.00            +0.0%     950.00        sched_debug.rt_rq:.rt_runtime.min
     48853 ± 26%     +18.4%      57835 ±  5%  sched_debug.sched_clk
      1.00            +0.0%       1.00        sched_debug.sched_clock_stable()
      3.00            +0.0%       3.00        sched_debug.sysctl_sched.sysctl_sched_base_slice
  25056823            +0.0%   25056823        sched_debug.sysctl_sched.sysctl_sched_features
      1.00            +0.0%       1.00        sched_debug.sysctl_sched.sysctl_sched_tunable_scaling
      0.00 ±264%    +416.0%       0.01 ±178%  perf-sched.sch_delay.avg.ms.__cond_resched.__alloc_pages.folio_alloc.do_read_cache_folio.vfs_dedupe_file_range_compare
      0.00          +5e+98%       0.00 ±300%  perf-sched.sch_delay.avg.ms.__cond_resched.__btrfs_run_delayed_refs.btrfs_run_delayed_refs.flush_space.btrfs_preempt_reclaim_metadata_space
      0.00        +1.5e+99%       0.00 ±161%  perf-sched.sch_delay.avg.ms.__cond_resched.__flush_work.isra.0.__lru_add_drain_all
      0.00 ±264%    -100.0%       0.00        perf-sched.sch_delay.avg.ms.__cond_resched.__kmem_cache_alloc_node.kmalloc_trace.btrfs_alloc_block_rsv.btrfs_replace_file_extents
      0.00          +5e+98%       0.00 ±300%  perf-sched.sch_delay.avg.ms.__cond_resched.__kmem_cache_alloc_node.kmalloc_trace.btrfs_fallocate.vfs_fallocate
      0.00 ±264%    -100.0%       0.00        perf-sched.sch_delay.avg.ms.__cond_resched.__wait_for_common.wait_for_completion_state.kernel_clone.__x64_sys_vfork
      0.00          +9e+98%       0.00 ±300%  perf-sched.sch_delay.avg.ms.__cond_resched.btrfs_alloc_path.btrfs_del_csums.cleanup_ref_head.__btrfs_run_delayed_refs
      0.00          +6e+98%       0.00 ±300%  perf-sched.sch_delay.avg.ms.__cond_resched.btrfs_search_slot.btrfs_del_csums.cleanup_ref_head.__btrfs_run_delayed_refs
      0.00 ±264%    +140.0%       0.00 ±152%  perf-sched.sch_delay.avg.ms.__cond_resched.btrfs_search_slot.btrfs_insert_empty_items.btrfs_insert_replace_extent.btrfs_replace_file_extents
      0.00 ±264%    -100.0%       0.00        perf-sched.sch_delay.avg.ms.__cond_resched.cgroup_rstat_flush_locked.cgroup_rstat_flush.do_flush_stats.flush_memcg_stats_dwork
      0.00 ±264%    -100.0%       0.00        perf-sched.sch_delay.avg.ms.__cond_resched.down_read.walk_component.link_path_walk.part
      0.00          +5e+98%       0.00 ±300%  perf-sched.sch_delay.avg.ms.__cond_resched.down_write.__btrfs_tree_lock.btrfs_lock_root_node.btrfs_search_slot
      0.02 ±259%    -100.0%       0.00        perf-sched.sch_delay.avg.ms.__cond_resched.find_free_extent.btrfs_reserve_extent.__btrfs_prealloc_file_range.btrfs_prealloc_file_range
      0.17 ±262%     -99.8%       0.00 ±165%  perf-sched.sch_delay.avg.ms.__cond_resched.kmem_cache_alloc.alloc_extent_state.__clear_extent_bit.btrfs_clone_files
      0.00        +7.3e+99%       0.01 ±300%  perf-sched.sch_delay.avg.ms.__cond_resched.kmem_cache_alloc.alloc_extent_state.__clear_extent_bit.btrfs_do_readpage
      0.00 ±264%    -100.0%       0.00        perf-sched.sch_delay.avg.ms.__cond_resched.kmem_cache_alloc.alloc_extent_state.__clear_extent_bit.btrfs_extent_same_range
      0.00       +1.3e+100%       0.01 ±297%  perf-sched.sch_delay.avg.ms.__cond_resched.kmem_cache_alloc.alloc_extent_state.__clear_extent_bit.btrfs_invalidate_folio
      0.00 ±264%     +60.0%       0.00 ±300%  perf-sched.sch_delay.avg.ms.__cond_resched.kmem_cache_alloc.alloc_extent_state.__clear_extent_bit.btrfs_replace_file_extents
      0.00 ±264%     -73.3%       0.00 ±300%  perf-sched.sch_delay.avg.ms.__cond_resched.kmem_cache_alloc.alloc_extent_state.__set_extent_bit.set_extent_bit
      0.00          +4e+98%       0.00 ±300%  perf-sched.sch_delay.avg.ms.__cond_resched.kmem_cache_alloc.btrfs_drop_extent_map_range.btrfs_drop_extents.btrfs_replace_file_extents
      0.00          +4e+98%       0.00 ±300%  perf-sched.sch_delay.avg.ms.__cond_resched.kmem_cache_alloc.start_transaction.btrfs_fileattr_set.vfs_fileattr_set
      0.00 ±264%    -100.0%       0.00        perf-sched.sch_delay.avg.ms.__cond_resched.kmem_cache_alloc.start_transaction.btrfs_replace_file_extents.btrfs_punch_hole
      0.00          +6e+98%       0.00 ±300%  perf-sched.sch_delay.avg.ms.__cond_resched.mutex_lock.perf_poll.do_poll.constprop
      0.00          +5e+98%       0.00 ±300%  perf-sched.sch_delay.avg.ms.__cond_resched.rcu_gp_cleanup.rcu_gp_kthread.kthread.ret_from_fork
      0.00        +1.7e+99%       0.00 ±300%  perf-sched.sch_delay.avg.ms.__cond_resched.shrink_dentry_list.shrink_dcache_parent.d_invalidate.proc_invalidate_siblings_dcache
      0.00 ± 78%      +6.7%       0.01 ± 70%  perf-sched.sch_delay.avg.ms.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
      0.00          +8e+98%       0.00 ±156%  perf-sched.sch_delay.avg.ms.__cond_resched.stop_one_cpu.sched_exec.bprm_execve.part
      0.00 ±264%     -60.0%       0.00 ±300%  perf-sched.sch_delay.avg.ms.__cond_resched.task_work_run.exit_to_user_mode_loop.exit_to_user_mode_prepare.syscall_exit_to_user_mode
      0.01 ±264%    -100.0%       0.00        perf-sched.sch_delay.avg.ms.__cond_resched.truncate_inode_pages_range.btrfs_clone_files.isra.0
      0.00          +2e+98%       0.00 ±300%  perf-sched.sch_delay.avg.ms.__cond_resched.wp_page_copy.__handle_mm_fault.handle_mm_fault.do_user_addr_fault
      0.01 ± 17%      -4.0%       0.00 ± 15%  perf-sched.sch_delay.avg.ms.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
      0.01 ± 44%      +1.7%       0.01 ± 39%  perf-sched.sch_delay.avg.ms.devkmsg_read.vfs_read.ksys_read.do_syscall_64
      0.00            +0.0%       0.00        perf-sched.sch_delay.avg.ms.do_nanosleep.hrtimer_nanosleep.common_nsleep.__x64_sys_clock_nanosleep
      0.00 ± 10%      -4.0%       0.00        perf-sched.sch_delay.avg.ms.do_task_dead.do_exit.do_group_exit.__x64_sys_exit_group.do_syscall_64
      0.01            +0.0%       0.01        perf-sched.sch_delay.avg.ms.do_wait.kernel_wait4.__do_sys_wait4.do_syscall_64
      0.00 ±264%    +140.0%       0.00 ±300%  perf-sched.sch_delay.avg.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_exc_page_fault
      0.00          +2e+98%       0.00 ±200%  perf-sched.sch_delay.avg.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt
      0.00 ±133%   +6324.6%       0.10 ±295%  perf-sched.sch_delay.avg.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_reschedule_ipi
      0.08 ±260%     -99.6%       0.00 ±213%  perf-sched.sch_delay.avg.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.syscall_exit_to_user_mode.do_syscall_64
      0.00          +1e+98%       0.00 ±300%  perf-sched.sch_delay.avg.ms.handle_reserve_ticket.__reserve_bytes.btrfs_reserve_data_bytes.btrfs_fallocate
      0.00 ±111%     +10.0%       0.00 ±103%  perf-sched.sch_delay.avg.ms.io_schedule.folio_wait_bit_common.filemap_fault.__do_fault
      0.00            +3.3%       0.00 ±  9%  perf-sched.sch_delay.avg.ms.pipe_read.vfs_read.ksys_read.do_syscall_64
      0.00           +10.0%       0.00 ± 27%  perf-sched.sch_delay.avg.ms.schedule_hrtimeout_range_clock.do_poll.constprop.0.do_sys_poll
      0.01 ± 33%     -11.7%       0.01 ± 12%  perf-sched.sch_delay.avg.ms.schedule_hrtimeout_range_clock.do_select.core_sys_select.kern_select
      0.00 ± 10%     +14.6%       0.01 ± 18%  perf-sched.sch_delay.avg.ms.schedule_hrtimeout_range_clock.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
      0.00 ±264%    -100.0%       0.00        perf-sched.sch_delay.avg.ms.schedule_preempt_disabled.rwsem_down_read_slowpath.down_read.__btrfs_tree_read_lock
      0.00 ±158%     +30.9%       0.00 ±263%  perf-sched.sch_delay.avg.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write.__btrfs_tree_lock
      0.00        +1.1e+99%       0.00 ±209%  perf-sched.sch_delay.avg.ms.schedule_timeout.__wait_for_common.__flush_work.isra.0
      0.00 ± 10%      -4.0%       0.00        perf-sched.sch_delay.avg.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone
      0.00 ± 14%      +4.2%       0.00 ± 18%  perf-sched.sch_delay.avg.ms.schedule_timeout.kcompactd.kthread.ret_from_fork
      0.00        +1.1e+99%       0.00 ±124%  perf-sched.sch_delay.avg.ms.schedule_timeout.khugepaged_wait_work.khugepaged.kthread
      0.00 ± 29%      -8.1%       0.00 ±  9%  perf-sched.sch_delay.avg.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
      0.03 ± 46%      -1.7%       0.03 ± 42%  perf-sched.sch_delay.avg.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
      0.01 ± 30%     +10.9%       0.01 ± 40%  perf-sched.sch_delay.avg.ms.syslog_print.do_syslog.kmsg_read.vfs_read
      0.00 ± 15%     +17.6%       0.00 ± 36%  perf-sched.sch_delay.avg.ms.wait_for_partner.fifo_open.do_dentry_open.do_open
      0.03 ± 36%     +10.7%       0.04 ± 24%  perf-sched.sch_delay.avg.ms.worker_thread.kthread.ret_from_fork.ret_from_fork_asm
      0.16 ±261%    +395.6%       0.81 ±172%  perf-sched.sch_delay.max.ms.__cond_resched.__alloc_pages.folio_alloc.do_read_cache_folio.vfs_dedupe_file_range_compare
      0.00          +5e+98%       0.00 ±300%  perf-sched.sch_delay.max.ms.__cond_resched.__btrfs_run_delayed_refs.btrfs_run_delayed_refs.flush_space.btrfs_preempt_reclaim_metadata_space
      0.00        +1.5e+99%       0.00 ±161%  perf-sched.sch_delay.max.ms.__cond_resched.__flush_work.isra.0.__lru_add_drain_all
      0.00 ±264%    -100.0%       0.00        perf-sched.sch_delay.max.ms.__cond_resched.__kmem_cache_alloc_node.kmalloc_trace.btrfs_alloc_block_rsv.btrfs_replace_file_extents
      0.00          +5e+98%       0.00 ±300%  perf-sched.sch_delay.max.ms.__cond_resched.__kmem_cache_alloc_node.kmalloc_trace.btrfs_fallocate.vfs_fallocate
      0.01 ±  8%     +10.5%       0.01 ± 20%  perf-sched.sch_delay.max.ms.__cond_resched.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr.__sched_setaffinity
      0.00 ±264%    -100.0%       0.00        perf-sched.sch_delay.max.ms.__cond_resched.__wait_for_common.wait_for_completion_state.kernel_clone.__x64_sys_vfork
      0.00          +9e+98%       0.00 ±300%  perf-sched.sch_delay.max.ms.__cond_resched.btrfs_alloc_path.btrfs_del_csums.cleanup_ref_head.__btrfs_run_delayed_refs
      0.00          +6e+98%       0.00 ±300%  perf-sched.sch_delay.max.ms.__cond_resched.btrfs_search_slot.btrfs_del_csums.cleanup_ref_head.__btrfs_run_delayed_refs
      0.00 ±130%     -20.0%       0.00 ±157%  perf-sched.sch_delay.max.ms.__cond_resched.btrfs_search_slot.btrfs_insert_empty_items.btrfs_insert_replace_extent.btrfs_replace_file_extents
      0.00 ±264%    -100.0%       0.00        perf-sched.sch_delay.max.ms.__cond_resched.cgroup_rstat_flush_locked.cgroup_rstat_flush.do_flush_stats.flush_memcg_stats_dwork
      0.00 ±264%    -100.0%       0.00        perf-sched.sch_delay.max.ms.__cond_resched.down_read.walk_component.link_path_walk.part
      0.00          +5e+98%       0.00 ±300%  perf-sched.sch_delay.max.ms.__cond_resched.down_write.__btrfs_tree_lock.btrfs_lock_root_node.btrfs_search_slot
      0.12 ±259%    -100.0%       0.00        perf-sched.sch_delay.max.ms.__cond_resched.find_free_extent.btrfs_reserve_extent.__btrfs_prealloc_file_range.btrfs_prealloc_file_range
      0.34 ±262%     -98.6%       0.00 ±133%  perf-sched.sch_delay.max.ms.__cond_resched.kmem_cache_alloc.alloc_extent_state.__clear_extent_bit.btrfs_clone_files
      0.00 ± 86%  +10253.9%       0.43 ±297%  perf-sched.sch_delay.max.ms.__cond_resched.kmem_cache_alloc.alloc_extent_state.__clear_extent_bit.btrfs_do_readpage
      0.00 ±264%    -100.0%       0.00        perf-sched.sch_delay.max.ms.__cond_resched.kmem_cache_alloc.alloc_extent_state.__clear_extent_bit.btrfs_extent_same_range
      0.00 ±264%  +28300.0%       0.18 ±297%  perf-sched.sch_delay.max.ms.__cond_resched.kmem_cache_alloc.alloc_extent_state.__clear_extent_bit.btrfs_invalidate_folio
      0.00 ±264%     -60.0%       0.00 ±300%  perf-sched.sch_delay.max.ms.__cond_resched.kmem_cache_alloc.alloc_extent_state.__clear_extent_bit.btrfs_replace_file_extents
      0.01 ± 38%      +2.9%       0.01 ± 28%  perf-sched.sch_delay.max.ms.__cond_resched.kmem_cache_alloc.alloc_extent_state.__set_extent_bit.lock_extent
      0.00 ±264%     -85.2%       0.00 ±300%  perf-sched.sch_delay.max.ms.__cond_resched.kmem_cache_alloc.alloc_extent_state.__set_extent_bit.set_extent_bit
      0.00          +4e+98%       0.00 ±300%  perf-sched.sch_delay.max.ms.__cond_resched.kmem_cache_alloc.btrfs_drop_extent_map_range.btrfs_drop_extents.btrfs_replace_file_extents
      0.00          +4e+98%       0.00 ±300%  perf-sched.sch_delay.max.ms.__cond_resched.kmem_cache_alloc.start_transaction.btrfs_fileattr_set.vfs_fileattr_set
      0.00 ±264%    -100.0%       0.00        perf-sched.sch_delay.max.ms.__cond_resched.kmem_cache_alloc.start_transaction.btrfs_replace_file_extents.btrfs_punch_hole
      0.00          +6e+98%       0.00 ±300%  perf-sched.sch_delay.max.ms.__cond_resched.mutex_lock.perf_poll.do_poll.constprop
      0.00          +5e+98%       0.00 ±300%  perf-sched.sch_delay.max.ms.__cond_resched.rcu_gp_cleanup.rcu_gp_kthread.kthread.ret_from_fork
      0.00        +1.7e+99%       0.00 ±300%  perf-sched.sch_delay.max.ms.__cond_resched.shrink_dentry_list.shrink_dcache_parent.d_invalidate.proc_invalidate_siblings_dcache
      0.01 ± 78%     +13.2%       0.01 ± 66%  perf-sched.sch_delay.max.ms.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
      0.00          +8e+98%       0.00 ±156%  perf-sched.sch_delay.max.ms.__cond_resched.stop_one_cpu.sched_exec.bprm_execve.part
      0.00 ±264%     -60.0%       0.00 ±300%  perf-sched.sch_delay.max.ms.__cond_resched.task_work_run.exit_to_user_mode_loop.exit_to_user_mode_prepare.syscall_exit_to_user_mode
      0.31 ±262%     -99.4%       0.00 ±127%  perf-sched.sch_delay.max.ms.__cond_resched.truncate_inode_pages_range.btrfs_clone_files.isra.0
      0.00 ±141%     -60.0%       0.00 ±300%  perf-sched.sch_delay.max.ms.__cond_resched.vfs_dedupe_file_range_compare.constprop.0.__generic_remap_file_range_prep
      0.00          +2e+98%       0.00 ±300%  perf-sched.sch_delay.max.ms.__cond_resched.wp_page_copy.__handle_mm_fault.handle_mm_fault.do_user_addr_fault
      0.01 ± 27%      -6.0%       0.01 ± 36%  perf-sched.sch_delay.max.ms.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
      0.01 ± 44%      +5.9%       0.01 ± 38%  perf-sched.sch_delay.max.ms.devkmsg_read.vfs_read.ksys_read.do_syscall_64
      0.21 ±201%    +167.9%       0.56 ±270%  perf-sched.sch_delay.max.ms.do_nanosleep.hrtimer_nanosleep.common_nsleep.__x64_sys_clock_nanosleep
      0.01 ± 11%      +5.6%       0.01 ± 36%  perf-sched.sch_delay.max.ms.do_task_dead.do_exit.do_group_exit.__x64_sys_exit_group.do_syscall_64
      0.01 ± 19%      +3.5%       0.01 ± 15%  perf-sched.sch_delay.max.ms.do_wait.kernel_wait4.__do_sys_wait4.do_syscall_64
      0.00 ±264%    +140.0%       0.00 ±300%  perf-sched.sch_delay.max.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_exc_page_fault
      0.00 ±264%    +513.3%       0.00 ±148%  perf-sched.sch_delay.max.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt
      0.00 ±152%   +7456.4%       0.21 ±296%  perf-sched.sch_delay.max.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_reschedule_ipi
      0.68 ±262%     -99.8%       0.00 ±200%  perf-sched.sch_delay.max.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.syscall_exit_to_user_mode.do_syscall_64
      0.00 ±130%      -7.7%       0.00 ±225%  perf-sched.sch_delay.max.ms.handle_reserve_ticket.__reserve_bytes.btrfs_reserve_data_bytes.btrfs_fallocate
      0.00 ±102%      -5.5%       0.00 ±103%  perf-sched.sch_delay.max.ms.io_schedule.folio_wait_bit_common.filemap_fault.__do_fault
      0.02 ± 23%     -23.3%       0.01 ± 29%  perf-sched.sch_delay.max.ms.pipe_read.vfs_read.ksys_read.do_syscall_64
      0.01 ± 21%      +9.3%       0.01 ± 15%  perf-sched.sch_delay.max.ms.schedule_hrtimeout_range_clock.do_poll.constprop.0.do_sys_poll
      0.01 ± 48%     -20.0%       0.01 ± 17%  perf-sched.sch_delay.max.ms.schedule_hrtimeout_range_clock.do_select.core_sys_select.kern_select
      0.01 ± 39%      +4.9%       0.01 ± 28%  perf-sched.sch_delay.max.ms.schedule_hrtimeout_range_clock.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
      1.64 ± 85%     -71.6%       0.47 ±128%  perf-sched.sch_delay.max.ms.schedule_preempt_disabled.rwsem_down_read_slowpath.down_read.__btrfs_tree_read_lock
      0.19 ±169%     -43.9%       0.11 ±259%  perf-sched.sch_delay.max.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write.__btrfs_tree_lock
      0.00        +1.1e+99%       0.00 ±209%  perf-sched.sch_delay.max.ms.schedule_timeout.__wait_for_common.__flush_work.isra.0
      0.00 ± 35%     +29.7%       0.01 ± 40%  perf-sched.sch_delay.max.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone
      0.00 ±  9%     +43.2%       0.01 ± 30%  perf-sched.sch_delay.max.ms.schedule_timeout.kcompactd.kthread.ret_from_fork
      0.00        +1.1e+99%       0.00 ±124%  perf-sched.sch_delay.max.ms.schedule_timeout.khugepaged_wait_work.khugepaged.kthread
      0.51 ±260%     -98.1%       0.01 ± 20%  perf-sched.sch_delay.max.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
      3.09 ± 40%      -1.1%       3.06 ± 27%  perf-sched.sch_delay.max.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
      0.01 ± 30%     +20.7%       0.01 ± 36%  perf-sched.sch_delay.max.ms.syslog_print.do_syslog.kmsg_read.vfs_read
      0.00 ± 85%     +75.1%       0.01 ±158%  perf-sched.sch_delay.max.ms.wait_for_partner.fifo_open.do_dentry_open.do_open
      2.39 ± 35%      +5.1%       2.51 ± 28%  perf-sched.sch_delay.max.ms.worker_thread.kthread.ret_from_fork.ret_from_fork_asm
      0.00            +0.0%       0.00        perf-sched.total_sch_delay.average.ms
      3.72 ± 30%      +0.8%       3.75 ± 17%  perf-sched.total_sch_delay.max.ms
      4.89 ±  3%      +4.9%       5.13 ±  3%  perf-sched.total_wait_and_delay.average.ms
     89766            -3.1%      86942        perf-sched.total_wait_and_delay.count.ms
      3132 ± 20%      +5.2%       3295 ± 18%  perf-sched.total_wait_and_delay.max.ms
      4.89 ±  3%      +4.9%       5.13 ±  3%  perf-sched.total_wait_time.average.ms
      3132 ± 20%      +5.2%       3295 ± 18%  perf-sched.total_wait_time.max.ms
      0.00       +1.5e+103%      15.38 ±300%  perf-sched.wait_and_delay.avg.ms.__cond_resched.__do_fault.do_read_fault.do_fault.__handle_mm_fault
     18.52 ±  5%      +1.5%      18.81 ±  8%  perf-sched.wait_and_delay.avg.ms.__cond_resched.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr.__sched_setaffinity
      0.00       +2.3e+102%       2.33 ±300%  perf-sched.wait_and_delay.avg.ms.__cond_resched.btrfs_alloc_path.btrfs_del_csums.cleanup_ref_head.__btrfs_run_delayed_refs
      0.00         +1e+103%      10.44 ±300%  perf-sched.wait_and_delay.avg.ms.__cond_resched.btrfs_search_slot.btrfs_del_csums.cleanup_ref_head.__btrfs_run_delayed_refs
     18.90 ±264%    -100.0%       0.00        perf-sched.wait_and_delay.avg.ms.__cond_resched.cgroup_rstat_flush_locked.cgroup_rstat_flush.do_flush_stats.flush_memcg_stats_dwork
      0.00         +1e+104%     100.10 ±300%  perf-sched.wait_and_delay.avg.ms.__cond_resched.mutex_lock.perf_poll.do_poll.constprop
      0.00         +1e+104%     100.02 ±300%  perf-sched.wait_and_delay.avg.ms.__cond_resched.shrink_dentry_list.shrink_dcache_parent.d_invalidate.proc_invalidate_siblings_dcache
    250.06 ±172%     +54.6%     386.64 ±113%  perf-sched.wait_and_delay.avg.ms.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
    799.87            -0.0%     799.87        perf-sched.wait_and_delay.avg.ms.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
      0.12            +0.0%       0.12        perf-sched.wait_and_delay.avg.ms.do_nanosleep.hrtimer_nanosleep.common_nsleep.__x64_sys_clock_nanosleep
    283.11 ±  8%      +0.4%     284.37 ±  9%  perf-sched.wait_and_delay.avg.ms.do_task_dead.do_exit.do_group_exit.__x64_sys_exit_group.do_syscall_64
      0.00         +1e+104%     100.02 ±300%  perf-sched.wait_and_delay.avg.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_exc_page_fault
    286.77 ±  7%      -1.3%     283.11 ±  6%  perf-sched.wait_and_delay.avg.ms.pipe_read.vfs_read.ksys_read.do_syscall_64
    298.84            +0.4%     299.90 ±  3%  perf-sched.wait_and_delay.avg.ms.schedule_hrtimeout_range_clock.do_poll.constprop.0.do_sys_poll
    420.93 ±  7%      -8.3%     386.20 ±  5%  perf-sched.wait_and_delay.avg.ms.schedule_hrtimeout_range_clock.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
      0.01 ±  9%      +7.9%       0.01 ±  6%  perf-sched.wait_and_delay.avg.ms.schedule_preempt_disabled.rwsem_down_read_slowpath.down_read.__btrfs_tree_read_lock
    453.61            -0.2%     452.50        perf-sched.wait_and_delay.avg.ms.schedule_timeout.kcompactd.kthread.ret_from_fork
      4.01 ±  8%      -0.5%       3.99 ±  8%  perf-sched.wait_and_delay.avg.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
    312.30 ±  5%      +2.2%     319.28 ±  8%  perf-sched.wait_and_delay.avg.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
    496.24 ±  3%      +2.0%     506.30 ±  6%  perf-sched.wait_and_delay.avg.ms.worker_thread.kthread.ret_from_fork.ret_from_fork_asm
      0.00         +1e+101%       0.10 ±300%  perf-sched.wait_and_delay.count.__cond_resched.__do_fault.do_read_fault.do_fault.__handle_mm_fault
    200.50 ±  8%      +0.5%     201.50 ±  8%  perf-sched.wait_and_delay.count.__cond_resched.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr.__sched_setaffinity
      0.00         +1e+101%       0.10 ±300%  perf-sched.wait_and_delay.count.__cond_resched.btrfs_alloc_path.btrfs_del_csums.cleanup_ref_head.__btrfs_run_delayed_refs
      0.00         +1e+101%       0.10 ±300%  perf-sched.wait_and_delay.count.__cond_resched.btrfs_search_slot.btrfs_del_csums.cleanup_ref_head.__btrfs_run_delayed_refs
      0.12 ±264%    -100.0%       0.00        perf-sched.wait_and_delay.count.__cond_resched.cgroup_rstat_flush_locked.cgroup_rstat_flush.do_flush_stats.flush_memcg_stats_dwork
      0.00         +1e+101%       0.10 ±300%  perf-sched.wait_and_delay.count.__cond_resched.mutex_lock.perf_poll.do_poll.constprop
      0.00         +1e+101%       0.10 ±300%  perf-sched.wait_and_delay.count.__cond_resched.shrink_dentry_list.shrink_dcache_parent.d_invalidate.proc_invalidate_siblings_dcache
      0.50 ±141%    +160.0%       1.30 ±142%  perf-sched.wait_and_delay.count.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
     10.00            +0.0%      10.00        perf-sched.wait_and_delay.count.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
     74939            +0.0%      74972        perf-sched.wait_and_delay.count.do_nanosleep.hrtimer_nanosleep.common_nsleep.__x64_sys_clock_nanosleep
     64.00            +0.0%      64.00        perf-sched.wait_and_delay.count.do_task_dead.do_exit.do_group_exit.__x64_sys_exit_group.do_syscall_64
      0.00         +1e+101%       0.10 ±300%  perf-sched.wait_and_delay.count.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_exc_page_fault
    165.88 ±  7%      +1.2%     167.80 ±  5%  perf-sched.wait_and_delay.count.pipe_read.vfs_read.ksys_read.do_syscall_64
     56.50            -0.9%      56.00 ±  2%  perf-sched.wait_and_delay.count.schedule_hrtimeout_range_clock.do_poll.constprop.0.do_sys_poll
     11.75 ±  7%      +6.4%      12.50 ±  4%  perf-sched.wait_and_delay.count.schedule_hrtimeout_range_clock.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
     11447 ±  6%     -24.2%       8673 ±  6%  perf-sched.wait_and_delay.count.schedule_preempt_disabled.rwsem_down_read_slowpath.down_read.__btrfs_tree_read_lock
     10.00            -2.0%       9.80 ±  4%  perf-sched.wait_and_delay.count.schedule_timeout.kcompactd.kthread.ret_from_fork
      1253 ±  8%      +0.5%       1259 ±  8%  perf-sched.wait_and_delay.count.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
    674.00 ±  7%      -3.4%     651.10 ±  4%  perf-sched.wait_and_delay.count.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
    225.25 ±  5%      +5.3%     237.20 ±  4%  perf-sched.wait_and_delay.count.worker_thread.kthread.ret_from_fork.ret_from_fork_asm
      0.00       +1.5e+103%      15.38 ±300%  perf-sched.wait_and_delay.max.ms.__cond_resched.__do_fault.do_read_fault.do_fault.__handle_mm_fault
      1000            +0.0%       1000        perf-sched.wait_and_delay.max.ms.__cond_resched.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr.__sched_setaffinity
      0.00       +2.3e+102%       2.33 ±300%  perf-sched.wait_and_delay.max.ms.__cond_resched.btrfs_alloc_path.btrfs_del_csums.cleanup_ref_head.__btrfs_run_delayed_refs
      0.00         +1e+103%      10.44 ±300%  perf-sched.wait_and_delay.max.ms.__cond_resched.btrfs_search_slot.btrfs_del_csums.cleanup_ref_head.__btrfs_run_delayed_refs
     18.90 ±264%    -100.0%       0.00        perf-sched.wait_and_delay.max.ms.__cond_resched.cgroup_rstat_flush_locked.cgroup_rstat_flush.do_flush_stats.flush_memcg_stats_dwork
      0.00         +1e+104%     100.10 ±300%  perf-sched.wait_and_delay.max.ms.__cond_resched.mutex_lock.perf_poll.do_poll.constprop
      0.00         +1e+104%     100.02 ±300%  perf-sched.wait_and_delay.max.ms.__cond_resched.shrink_dentry_list.shrink_dcache_parent.d_invalidate.proc_invalidate_siblings_dcache
    251.12 ±171%    +179.6%     702.09 ±128%  perf-sched.wait_and_delay.max.ms.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
    999.87            -0.0%     999.86        perf-sched.wait_and_delay.max.ms.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
      1000            +0.0%       1000        perf-sched.wait_and_delay.max.ms.do_nanosleep.hrtimer_nanosleep.common_nsleep.__x64_sys_clock_nanosleep
      1868 ± 49%     -19.8%       1498 ± 44%  perf-sched.wait_and_delay.max.ms.do_task_dead.do_exit.do_group_exit.__x64_sys_exit_group.do_syscall_64
      0.00         +1e+104%     100.02 ±300%  perf-sched.wait_and_delay.max.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_exc_page_fault
      1008            -0.0%       1008        perf-sched.wait_and_delay.max.ms.pipe_read.vfs_read.ksys_read.do_syscall_64
      1000            -0.0%       1000        perf-sched.wait_and_delay.max.ms.schedule_hrtimeout_range_clock.do_poll.constprop.0.do_sys_poll
    531.24 ±  7%      -5.9%     499.99        perf-sched.wait_and_delay.max.ms.schedule_hrtimeout_range_clock.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
      2.63 ± 60%     -53.1%       1.23 ± 67%  perf-sched.wait_and_delay.max.ms.schedule_preempt_disabled.rwsem_down_read_slowpath.down_read.__btrfs_tree_read_lock
    504.63            +0.0%     504.71        perf-sched.wait_and_delay.max.ms.schedule_timeout.kcompactd.kthread.ret_from_fork
      5.01            +0.1%       5.01        perf-sched.wait_and_delay.max.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
      2794 ± 23%      +7.7%       3011 ± 18%  perf-sched.wait_and_delay.max.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
      2561 ± 18%     +14.4%       2930 ± 33%  perf-sched.wait_and_delay.max.ms.worker_thread.kthread.ret_from_fork.ret_from_fork_asm
      0.00 ±182%     +31.8%       0.00 ±109%  perf-sched.wait_time.avg.ms.__cond_resched.__alloc_pages.__kmalloc_large_node.__kmalloc_node.kvmalloc_node
      0.01 ± 43%      -1.9%       0.01 ± 49%  perf-sched.wait_time.avg.ms.__cond_resched.__alloc_pages.folio_alloc.do_read_cache_folio.vfs_dedupe_file_range_compare
      0.00          +8e+98%       0.00 ±300%  perf-sched.wait_time.avg.ms.__cond_resched.__btrfs_run_delayed_refs.btrfs_run_delayed_refs.flush_space.btrfs_preempt_reclaim_metadata_space
      0.00 ±174%     -64.4%       0.00 ±300%  perf-sched.wait_time.avg.ms.__cond_resched.__clear_extent_bit.btrfs_replace_file_extents.btrfs_punch_hole.btrfs_fallocate
      0.00       +1.5e+103%      15.38 ±300%  perf-sched.wait_time.avg.ms.__cond_resched.__do_fault.do_read_fault.do_fault.__handle_mm_fault
      0.00 ±174%     -20.0%       0.00 ±201%  perf-sched.wait_time.avg.ms.__cond_resched.__kmem_cache_alloc_node.__kmalloc_node_track_caller.memdup_user.do_vfs_ioctl
      0.01 ±101%     -20.0%       0.01 ± 49%  perf-sched.wait_time.avg.ms.__cond_resched.__kmem_cache_alloc_node.kmalloc_trace.btrfs_alloc_block_rsv.btrfs_replace_file_extents
      0.00 ±264%    +508.0%       0.00 ±191%  perf-sched.wait_time.avg.ms.__cond_resched.__kmem_cache_alloc_node.kmalloc_trace.btrfs_fallocate.vfs_fallocate
     18.52 ±  5%      +1.5%      18.81 ±  8%  perf-sched.wait_time.avg.ms.__cond_resched.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr.__sched_setaffinity
      0.08 ±264%    -100.0%       0.00        perf-sched.wait_time.avg.ms.__cond_resched.__wait_for_common.wait_for_completion_state.kernel_clone.__x64_sys_vfork
      0.00 ±264%     -52.9%       0.00 ±200%  perf-sched.wait_time.avg.ms.__cond_resched.btrfs_alloc_path.btrfs_clone.btrfs_clone_files.isra
      0.00        +3.8e+99%       0.00 ± 52%  perf-sched.wait_time.avg.ms.__cond_resched.btrfs_alloc_path.btrfs_clone.btrfs_extent_same_range.btrfs_remap_file_range
      0.00       +2.3e+102%       2.33 ±300%  perf-sched.wait_time.avg.ms.__cond_resched.btrfs_alloc_path.btrfs_del_csums.cleanup_ref_head.__btrfs_run_delayed_refs
      0.00 ±264%    -100.0%       0.00        perf-sched.wait_time.avg.ms.__cond_resched.btrfs_alloc_path.btrfs_get_extent.find_first_non_hole.btrfs_punch_hole
      0.00 ±174%    -100.0%       0.00        perf-sched.wait_time.avg.ms.__cond_resched.btrfs_alloc_path.btrfs_punch_hole.btrfs_fallocate.vfs_fallocate
      0.00        +2.4e+99%       0.00 ±112%  perf-sched.wait_time.avg.ms.__cond_resched.btrfs_alloc_path.insert_prealloc_file_extent.__btrfs_prealloc_file_range.btrfs_prealloc_file_range
      0.00 ±264%    -100.0%       0.00        perf-sched.wait_time.avg.ms.__cond_resched.btrfs_search_slot.btrfs_clone.btrfs_clone_files.isra
      0.00          +6e+98%       0.00 ±300%  perf-sched.wait_time.avg.ms.__cond_resched.btrfs_search_slot.btrfs_clone.btrfs_extent_same_range.btrfs_remap_file_range
      0.00         +1e+103%      10.44 ±299%  perf-sched.wait_time.avg.ms.__cond_resched.btrfs_search_slot.btrfs_del_csums.cleanup_ref_head.__btrfs_run_delayed_refs
      0.01 ± 53%    +269.2%       0.04 ±240%  perf-sched.wait_time.avg.ms.__cond_resched.btrfs_search_slot.btrfs_insert_empty_items.btrfs_insert_replace_extent.btrfs_replace_file_extents
      0.01 ±106%     -75.2%       0.00 ±152%  perf-sched.wait_time.avg.ms.__cond_resched.btrfs_search_slot.btrfs_lookup_file_extent.btrfs_drop_extents.btrfs_replace_file_extents
      0.00 ±264%    -100.0%       0.00        perf-sched.wait_time.avg.ms.__cond_resched.btrfs_search_slot.btrfs_lookup_file_extent.btrfs_get_extent.find_first_non_hole
     18.90 ±264%    -100.0%       0.00        perf-sched.wait_time.avg.ms.__cond_resched.cgroup_rstat_flush_locked.cgroup_rstat_flush.do_flush_stats.flush_memcg_stats_dwork
      0.00 ±105%     -24.4%       0.00 ±162%  perf-sched.wait_time.avg.ms.__cond_resched.down_read.__btrfs_tree_read_lock.btrfs_read_lock_root_node.btrfs_search_slot
      0.00 ±174%      -2.2%       0.00 ±201%  perf-sched.wait_time.avg.ms.__cond_resched.down_read.find_free_extent.btrfs_reserve_extent.__btrfs_prealloc_file_range
      0.00 ±264%      +0.0%       0.00 ±300%  perf-sched.wait_time.avg.ms.__cond_resched.down_read.unmap_mapping_range.truncate_pagecache_range.btrfs_punch_hole_lock_range
      0.00 ± 85%     +45.2%       0.00 ± 95%  perf-sched.wait_time.avg.ms.__cond_resched.down_write.__btrfs_tree_lock.btrfs_lock_root_node.btrfs_search_slot
      0.00          +5e+98%       0.00 ±300%  perf-sched.wait_time.avg.ms.__cond_resched.down_write.btrfs_inode_lock.btrfs_punch_hole.btrfs_fallocate
      0.00 ±264%    -100.0%       0.00        perf-sched.wait_time.avg.ms.__cond_resched.down_write.btrfs_inode_lock.btrfs_remap_file_range.do_clone_file_range
      0.00 ±264%    -100.0%       0.00        perf-sched.wait_time.avg.ms.__cond_resched.down_write.btrfs_remap_file_range.do_clone_file_range.vfs_clone_file_range
      0.00 ±174%     -70.9%       0.00 ±300%  perf-sched.wait_time.avg.ms.__cond_resched.down_write.lock_two_inodes.btrfs_remap_file_range.do_clone_file_range
      0.00 ±264%    -100.0%       0.00        perf-sched.wait_time.avg.ms.__cond_resched.down_write.lock_two_inodes.btrfs_remap_file_range.vfs_dedupe_file_range_one
      0.00          +4e+98%       0.00 ±300%  perf-sched.wait_time.avg.ms.__cond_resched.down_write.vfs_fileattr_set.do_vfs_ioctl.__x64_sys_ioctl
      0.03 ±210%     -80.2%       0.01 ±  8%  perf-sched.wait_time.avg.ms.__cond_resched.find_free_extent.btrfs_reserve_extent.__btrfs_prealloc_file_range.btrfs_prealloc_file_range
      0.00 ± 41%     -15.4%       0.00 ± 76%  perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc.alloc_extent_map.__btrfs_prealloc_file_range.btrfs_prealloc_file_range
      0.00 ±196%    -100.0%       0.00        perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc.alloc_extent_map.btrfs_get_extent.btrfs_do_readpage
      0.01 ± 83%     -18.6%       0.01 ± 34%  perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc.alloc_extent_state.__clear_extent_bit.btrfs_clone_files
      0.01 ±  8%     +10.5%       0.01 ± 12%  perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc.alloc_extent_state.__clear_extent_bit.btrfs_do_readpage
      0.00 ±102%     -45.9%       0.00 ±101%  perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc.alloc_extent_state.__clear_extent_bit.btrfs_extent_same_range
      0.00 ±105%     +11.1%       0.00 ±101%  perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc.alloc_extent_state.__clear_extent_bit.btrfs_fallocate
      0.01 ± 42%     +14.7%       0.01 ± 77%  perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc.alloc_extent_state.__clear_extent_bit.btrfs_invalidate_folio
      0.00 ±264%     -55.6%       0.00 ±300%  perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc.alloc_extent_state.__clear_extent_bit.btrfs_punch_hole
      0.01 ± 73%     -33.7%       0.01 ±105%  perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc.alloc_extent_state.__clear_extent_bit.btrfs_replace_file_extents
      0.01 ±  9%     +39.5%       0.01 ± 69%  perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc.alloc_extent_state.__set_extent_bit.lock_extent
      0.01 ± 27%     +26.8%       0.01 ± 60%  perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc.alloc_extent_state.__set_extent_bit.set_extent_bit
      0.00 ±264%     +12.0%       0.00 ±300%  perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc.btrfs_add_delayed_data_ref.btrfs_alloc_reserved_file_extent.btrfs_insert_replace_extent
      0.00 ± 79%     +12.0%       0.00 ± 83%  perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc.btrfs_add_delayed_data_ref.btrfs_drop_extents.btrfs_replace_file_extents
      0.00 ±100%     -41.8%       0.00 ±153%  perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc.btrfs_add_delayed_data_ref.btrfs_insert_replace_extent.btrfs_replace_file_extents
      0.00        +3.7e+99%       0.00 ±265%  perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc.btrfs_clone.btrfs_clone_files.isra
      0.00 ±100%     +84.0%       0.00 ± 57%  perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc.btrfs_drop_extent_map_range.btrfs_drop_extents.btrfs_replace_file_extents
      0.00 ±264%    -100.0%       0.00        perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc.btrfs_drop_extent_map_range.btrfs_replace_extent_map_range.__btrfs_prealloc_file_range
      0.00 ±264%    -100.0%       0.00        perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc.btrfs_get_extent.btrfs_do_readpage.btrfs_read_folio
      0.00          +5e+98%       0.00 ±300%  perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc.btrfs_punch_hole.btrfs_fallocate.vfs_fallocate
      0.00 ±264%    -100.0%       0.00        perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc.insert_prealloc_file_extent.__btrfs_prealloc_file_range.btrfs_prealloc_file_range
      0.00 ± 83%     -40.7%       0.00 ±122%  perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc.start_transaction.btrfs_dirty_inode.file_modified
      0.00 ±173%    +520.0%       0.01 ±230%  perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc.start_transaction.btrfs_fileattr_set.vfs_fileattr_set
      0.00 ±133%     +42.9%       0.00 ±100%  perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc.start_transaction.btrfs_replace_file_extents.btrfs_clone
      0.00 ±264%     -27.3%       0.00 ±300%  perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc.start_transaction.btrfs_replace_file_extents.btrfs_punch_hole
      0.00 ±173%    -100.0%       0.00        perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc.start_transaction.btrfs_replace_file_extents.insert_prealloc_file_extent
      0.00 ±264%     +20.0%       0.00 ±300%  perf-sched.wait_time.avg.ms.__cond_resched.mnt_want_write_file.do_vfs_ioctl.__x64_sys_ioctl.do_syscall_64
      0.00          +1e+99%       0.00 ±204%  perf-sched.wait_time.avg.ms.__cond_resched.mnt_want_write_file.vfs_dedupe_file_range_one.vfs_dedupe_file_range.do_vfs_ioctl
      0.00 ±136%     -20.0%       0.00 ±153%  perf-sched.wait_time.avg.ms.__cond_resched.mutex_lock.__btrfs_release_delayed_node.part.0
      0.00 ±264%      +0.0%       0.00 ±300%  perf-sched.wait_time.avg.ms.__cond_resched.mutex_lock.btrfs_delayed_update_inode.btrfs_update_inode.__btrfs_prealloc_file_range
      0.00 ±264%    -100.0%       0.00        perf-sched.wait_time.avg.ms.__cond_resched.mutex_lock.btrfs_delayed_update_inode.btrfs_update_inode.btrfs_dirty_inode
      0.00 ±264%    +380.0%       0.00 ±300%  perf-sched.wait_time.avg.ms.__cond_resched.mutex_lock.btrfs_delayed_update_inode.btrfs_update_inode.btrfs_fileattr_set
      0.00 ±264%    +252.0%       0.00 ±245%  perf-sched.wait_time.avg.ms.__cond_resched.mutex_lock.btrfs_delayed_update_inode.btrfs_update_inode.btrfs_punch_hole
      0.00 ±173%    +212.0%       0.00 ± 51%  perf-sched.wait_time.avg.ms.__cond_resched.mutex_lock.btrfs_delayed_update_inode.btrfs_update_inode.clone_finish_inode_update
      0.00         +1e+104%     100.10 ±299%  perf-sched.wait_time.avg.ms.__cond_resched.mutex_lock.perf_poll.do_poll.constprop
      0.00         +4e+101%       0.40 ±300%  perf-sched.wait_time.avg.ms.__cond_resched.rcu_gp_cleanup.rcu_gp_kthread.kthread.ret_from_fork
      0.00         +1e+104%     100.02 ±300%  perf-sched.wait_time.avg.ms.__cond_resched.shrink_dentry_list.shrink_dcache_parent.d_invalidate.proc_invalidate_siblings_dcache
    250.06 ±172%     +54.6%     386.64 ±113%  perf-sched.wait_time.avg.ms.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
      0.00          +4e+98%       0.00 ±300%  perf-sched.wait_time.avg.ms.__cond_resched.start_transaction.btrfs_replace_file_extents.btrfs_clone.btrfs_extent_same_range
      0.00        +5.8e+99%       0.01 ±300%  perf-sched.wait_time.avg.ms.__cond_resched.task_work_run.exit_to_user_mode_loop.exit_to_user_mode_prepare.syscall_exit_to_user_mode
      0.01 ± 22%     +21.7%       0.01 ± 67%  perf-sched.wait_time.avg.ms.__cond_resched.truncate_inode_pages_range.btrfs_clone_files.isra.0
      0.00        +1.4e+99%       0.00 ±153%  perf-sched.wait_time.avg.ms.__cond_resched.truncate_inode_pages_range.btrfs_punch_hole_lock_range.btrfs_punch_hole.btrfs_fallocate
      0.00 ± 43%     -12.5%       0.00 ± 69%  perf-sched.wait_time.avg.ms.__cond_resched.vfs_clone_file_range.ioctl_file_clone.do_vfs_ioctl.__x64_sys_ioctl
      0.01 ± 93%     -31.9%       0.01 ± 17%  perf-sched.wait_time.avg.ms.__cond_resched.vfs_dedupe_file_range_compare.constprop.0.__generic_remap_file_range_prep
      0.00 ± 39%     -38.3%       0.00 ± 82%  perf-sched.wait_time.avg.ms.__cond_resched.vfs_fallocate.ioctl_preallocate.__x64_sys_ioctl.do_syscall_64
    799.87            -0.0%     799.86        perf-sched.wait_time.avg.ms.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
      0.11 ±173%    +141.9%       0.27 ± 82%  perf-sched.wait_time.avg.ms.devkmsg_read.vfs_read.ksys_read.do_syscall_64
      0.11            +0.0%       0.11        perf-sched.wait_time.avg.ms.do_nanosleep.hrtimer_nanosleep.common_nsleep.__x64_sys_clock_nanosleep
    283.11 ±  8%      +0.4%     284.37 ±  9%  perf-sched.wait_time.avg.ms.do_task_dead.do_exit.do_group_exit.__x64_sys_exit_group.do_syscall_64
      0.28 ±  2%      +4.4%       0.29 ±  2%  perf-sched.wait_time.avg.ms.do_wait.kernel_wait4.__do_sys_wait4.do_syscall_64
      0.00         +1e+104%     100.02 ±300%  perf-sched.wait_time.avg.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_exc_page_fault
      0.00 ± 10%     +21.1%       0.01 ± 29%  perf-sched.wait_time.avg.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt
      0.08 ±205%     -69.0%       0.03 ± 77%  perf-sched.wait_time.avg.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_reschedule_ipi
      0.02 ±155%     -57.8%       0.01 ± 47%  perf-sched.wait_time.avg.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.syscall_exit_to_user_mode.do_syscall_64
      0.01 ± 12%      +1.8%       0.01 ± 26%  perf-sched.wait_time.avg.ms.handle_reserve_ticket.__reserve_bytes.btrfs_reserve_data_bytes.btrfs_fallocate
      0.00 ±185%     -46.7%       0.00 ±200%  perf-sched.wait_time.avg.ms.io_schedule.folio_wait_bit_common.filemap_fault.__do_fault
    286.76 ±  7%      -1.3%     283.11 ±  6%  perf-sched.wait_time.avg.ms.pipe_read.vfs_read.ksys_read.do_syscall_64
    298.84            +0.4%     299.90 ±  3%  perf-sched.wait_time.avg.ms.schedule_hrtimeout_range_clock.do_poll.constprop.0.do_sys_poll
    420.93 ±  7%      -8.3%     386.19 ±  5%  perf-sched.wait_time.avg.ms.schedule_hrtimeout_range_clock.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
      0.01 ±  6%      +3.4%       0.01 ±  8%  perf-sched.wait_time.avg.ms.schedule_preempt_disabled.rwsem_down_read_slowpath.down_read.__btrfs_tree_read_lock
      0.01 ±  9%    +106.5%       0.01 ±117%  perf-sched.wait_time.avg.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write.__btrfs_tree_lock
      0.00        +1.5e+99%       0.00 ±155%  perf-sched.wait_time.avg.ms.schedule_timeout.__wait_for_common.__flush_work.isra.0
      0.00 ±188%     -70.5%       0.00 ±300%  perf-sched.wait_time.avg.ms.schedule_timeout.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr
      0.42            +1.3%       0.42        perf-sched.wait_time.avg.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone
    453.61            -0.2%     452.50        perf-sched.wait_time.avg.ms.schedule_timeout.kcompactd.kthread.ret_from_fork
      0.00         +5e+101%       0.50 ±202%  perf-sched.wait_time.avg.ms.schedule_timeout.khugepaged_wait_work.khugepaged.kthread
      4.01 ±  8%      -0.5%       3.99 ±  8%  perf-sched.wait_time.avg.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
    312.27 ±  5%      +2.2%     319.25 ±  8%  perf-sched.wait_time.avg.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
      0.11 ±173%    +141.6%       0.27 ± 82%  perf-sched.wait_time.avg.ms.syslog_print.do_syslog.kmsg_read.vfs_read
      0.00          +1e+98%       0.00 ±300%  perf-sched.wait_time.avg.ms.wait_for_partner.fifo_open.do_dentry_open.do_open
    496.21 ±  3%      +2.0%     506.26 ±  6%  perf-sched.wait_time.avg.ms.worker_thread.kthread.ret_from_fork.ret_from_fork_asm
      0.00 ±178%     +24.4%       0.00 ±109%  perf-sched.wait_time.max.ms.__cond_resched.__alloc_pages.__kmalloc_large_node.__kmalloc_node.kvmalloc_node
      0.12 ±183%     -29.5%       0.08 ±218%  perf-sched.wait_time.max.ms.__cond_resched.__alloc_pages.folio_alloc.do_read_cache_folio.vfs_dedupe_file_range_compare
      0.00          +8e+98%       0.00 ±300%  perf-sched.wait_time.max.ms.__cond_resched.__btrfs_run_delayed_refs.btrfs_run_delayed_refs.flush_space.btrfs_preempt_reclaim_metadata_space
      0.00 ±174%     -64.4%       0.00 ±300%  perf-sched.wait_time.max.ms.__cond_resched.__clear_extent_bit.btrfs_replace_file_extents.btrfs_punch_hole.btrfs_fallocate
      0.00       +1.5e+103%      15.38 ±300%  perf-sched.wait_time.max.ms.__cond_resched.__do_fault.do_read_fault.do_fault.__handle_mm_fault
      0.00 ±174%     -20.0%       0.00 ±201%  perf-sched.wait_time.max.ms.__cond_resched.__kmem_cache_alloc_node.__kmalloc_node_track_caller.memdup_user.do_vfs_ioctl
      0.01 ±125%     -32.6%       0.01 ± 59%  perf-sched.wait_time.max.ms.__cond_resched.__kmem_cache_alloc_node.kmalloc_trace.btrfs_alloc_block_rsv.btrfs_replace_file_extents
      0.00 ±264%    +508.0%       0.00 ±191%  perf-sched.wait_time.max.ms.__cond_resched.__kmem_cache_alloc_node.kmalloc_trace.btrfs_fallocate.vfs_fallocate
      1000            +0.0%       1000        perf-sched.wait_time.max.ms.__cond_resched.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr.__sched_setaffinity
      0.08 ±264%    -100.0%       0.00        perf-sched.wait_time.max.ms.__cond_resched.__wait_for_common.wait_for_completion_state.kernel_clone.__x64_sys_vfork
      0.00 ±264%     -52.9%       0.00 ±200%  perf-sched.wait_time.max.ms.__cond_resched.btrfs_alloc_path.btrfs_clone.btrfs_clone_files.isra
      0.00        +3.9e+99%       0.00 ± 51%  perf-sched.wait_time.max.ms.__cond_resched.btrfs_alloc_path.btrfs_clone.btrfs_extent_same_range.btrfs_remap_file_range
      0.00       +2.3e+102%       2.33 ±300%  perf-sched.wait_time.max.ms.__cond_resched.btrfs_alloc_path.btrfs_del_csums.cleanup_ref_head.__btrfs_run_delayed_refs
      0.00 ±264%    -100.0%       0.00        perf-sched.wait_time.max.ms.__cond_resched.btrfs_alloc_path.btrfs_get_extent.find_first_non_hole.btrfs_punch_hole
      0.00 ±174%    -100.0%       0.00        perf-sched.wait_time.max.ms.__cond_resched.btrfs_alloc_path.btrfs_punch_hole.btrfs_fallocate.vfs_fallocate
      0.00        +2.6e+99%       0.00 ±120%  perf-sched.wait_time.max.ms.__cond_resched.btrfs_alloc_path.insert_prealloc_file_extent.__btrfs_prealloc_file_range.btrfs_prealloc_file_range
      0.00 ±264%    -100.0%       0.00        perf-sched.wait_time.max.ms.__cond_resched.btrfs_search_slot.btrfs_clone.btrfs_clone_files.isra
      0.00          +6e+98%       0.00 ±300%  perf-sched.wait_time.max.ms.__cond_resched.btrfs_search_slot.btrfs_clone.btrfs_extent_same_range.btrfs_remap_file_range
      0.00         +1e+103%      10.44 ±299%  perf-sched.wait_time.max.ms.__cond_resched.btrfs_search_slot.btrfs_del_csums.cleanup_ref_head.__btrfs_run_delayed_refs
      0.03 ± 85%    +731.0%       0.27 ±273%  perf-sched.wait_time.max.ms.__cond_resched.btrfs_search_slot.btrfs_insert_empty_items.btrfs_insert_replace_extent.btrfs_replace_file_extents
      0.01 ± 99%     -77.1%       0.00 ±152%  perf-sched.wait_time.max.ms.__cond_resched.btrfs_search_slot.btrfs_lookup_file_extent.btrfs_drop_extents.btrfs_replace_file_extents
      0.00 ±264%    -100.0%       0.00        perf-sched.wait_time.max.ms.__cond_resched.btrfs_search_slot.btrfs_lookup_file_extent.btrfs_get_extent.find_first_non_hole
     18.90 ±264%    -100.0%       0.00        perf-sched.wait_time.max.ms.__cond_resched.cgroup_rstat_flush_locked.cgroup_rstat_flush.do_flush_stats.flush_memcg_stats_dwork
      0.00 ±105%     -15.6%       0.00 ±170%  perf-sched.wait_time.max.ms.__cond_resched.down_read.__btrfs_tree_read_lock.btrfs_read_lock_root_node.btrfs_search_slot
      0.00 ±174%      -2.2%       0.00 ±201%  perf-sched.wait_time.max.ms.__cond_resched.down_read.find_free_extent.btrfs_reserve_extent.__btrfs_prealloc_file_range
      0.00 ±264%      +0.0%       0.00 ±300%  perf-sched.wait_time.max.ms.__cond_resched.down_read.unmap_mapping_range.truncate_pagecache_range.btrfs_punch_hole_lock_range
      0.00 ± 89%     +30.7%       0.00 ± 95%  perf-sched.wait_time.max.ms.__cond_resched.down_write.__btrfs_tree_lock.btrfs_lock_root_node.btrfs_search_slot
      0.00          +5e+98%       0.00 ±300%  perf-sched.wait_time.max.ms.__cond_resched.down_write.btrfs_inode_lock.btrfs_punch_hole.btrfs_fallocate
      0.00 ±264%    -100.0%       0.00        perf-sched.wait_time.max.ms.__cond_resched.down_write.btrfs_inode_lock.btrfs_remap_file_range.do_clone_file_range
      0.00 ±264%    -100.0%       0.00        perf-sched.wait_time.max.ms.__cond_resched.down_write.btrfs_remap_file_range.do_clone_file_range.vfs_clone_file_range
      0.00 ±174%     -70.9%       0.00 ±300%  perf-sched.wait_time.max.ms.__cond_resched.down_write.lock_two_inodes.btrfs_remap_file_range.do_clone_file_range
      0.00 ±264%    -100.0%       0.00        perf-sched.wait_time.max.ms.__cond_resched.down_write.lock_two_inodes.btrfs_remap_file_range.vfs_dedupe_file_range_one
      0.00          +4e+98%       0.00 ±300%  perf-sched.wait_time.max.ms.__cond_resched.down_write.vfs_fileattr_set.do_vfs_ioctl.__x64_sys_ioctl
      0.18 ±251%     -96.5%       0.01 ± 31%  perf-sched.wait_time.max.ms.__cond_resched.find_free_extent.btrfs_reserve_extent.__btrfs_prealloc_file_range.btrfs_prealloc_file_range
      0.00 ± 41%     -13.1%       0.00 ± 76%  perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc.alloc_extent_map.__btrfs_prealloc_file_range.btrfs_prealloc_file_range
      0.00 ±196%    -100.0%       0.00        perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc.alloc_extent_map.btrfs_get_extent.btrfs_do_readpage
      0.01 ±114%     +58.0%       0.02 ± 82%  perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc.alloc_extent_state.__clear_extent_bit.btrfs_clone_files
      0.02 ± 71%      +7.8%       0.03 ± 61%  perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc.alloc_extent_state.__clear_extent_bit.btrfs_do_readpage
      0.01 ±122%     -56.2%       0.00 ±101%  perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc.alloc_extent_state.__clear_extent_bit.btrfs_extent_same_range
      0.00 ±105%     +11.1%       0.00 ±101%  perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc.alloc_extent_state.__clear_extent_bit.btrfs_fallocate
      0.02 ±136%     +68.4%       0.04 ±226%  perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc.alloc_extent_state.__clear_extent_bit.btrfs_invalidate_folio
      0.00 ±264%     -55.6%       0.00 ±300%  perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc.alloc_extent_state.__clear_extent_bit.btrfs_punch_hole
      0.01 ± 96%     -53.5%       0.01 ±102%  perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc.alloc_extent_state.__clear_extent_bit.btrfs_replace_file_extents
      0.04 ± 44%    +210.9%       0.11 ±198%  perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc.alloc_extent_state.__set_extent_bit.lock_extent
      0.02 ± 72%      +6.8%       0.02 ± 97%  perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc.alloc_extent_state.__set_extent_bit.set_extent_bit
      0.00 ±264%     +12.0%       0.00 ±300%  perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc.btrfs_add_delayed_data_ref.btrfs_alloc_reserved_file_extent.btrfs_insert_replace_extent
      0.00 ± 79%     +24.8%       0.00 ± 91%  perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc.btrfs_add_delayed_data_ref.btrfs_drop_extents.btrfs_replace_file_extents
      0.00 ±100%     -41.8%       0.00 ±153%  perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc.btrfs_add_delayed_data_ref.btrfs_insert_replace_extent.btrfs_replace_file_extents
      0.00        +3.7e+99%       0.00 ±265%  perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc.btrfs_clone.btrfs_clone_files.isra
      0.00 ±100%     +84.0%       0.00 ± 57%  perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc.btrfs_drop_extent_map_range.btrfs_drop_extents.btrfs_replace_file_extents
      0.00 ±264%    -100.0%       0.00        perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc.btrfs_drop_extent_map_range.btrfs_replace_extent_map_range.__btrfs_prealloc_file_range
      0.00 ±264%    -100.0%       0.00        perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc.btrfs_get_extent.btrfs_do_readpage.btrfs_read_folio
      0.00          +5e+98%       0.00 ±300%  perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc.btrfs_punch_hole.btrfs_fallocate.vfs_fallocate
      0.00 ±264%    -100.0%       0.00        perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc.insert_prealloc_file_extent.__btrfs_prealloc_file_range.btrfs_prealloc_file_range
      0.00 ± 94%     -46.7%       0.00 ±122%  perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc.start_transaction.btrfs_dirty_inode.file_modified
      0.00 ±173%    +520.0%       0.01 ±230%  perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc.start_transaction.btrfs_fileattr_set.vfs_fileattr_set
      0.00 ±133%     +42.9%       0.00 ±100%  perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc.start_transaction.btrfs_replace_file_extents.btrfs_clone
      0.00 ±264%     -27.3%       0.00 ±300%  perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc.start_transaction.btrfs_replace_file_extents.btrfs_punch_hole
      0.00 ±174%    -100.0%       0.00        perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc.start_transaction.btrfs_replace_file_extents.insert_prealloc_file_extent
      0.00 ±264%     +20.0%       0.00 ±300%  perf-sched.wait_time.max.ms.__cond_resched.mnt_want_write_file.do_vfs_ioctl.__x64_sys_ioctl.do_syscall_64
      0.00          +1e+99%       0.00 ±204%  perf-sched.wait_time.max.ms.__cond_resched.mnt_want_write_file.vfs_dedupe_file_range_one.vfs_dedupe_file_range.do_vfs_ioctl
      0.00 ±136%     -20.0%       0.00 ±153%  perf-sched.wait_time.max.ms.__cond_resched.mutex_lock.__btrfs_release_delayed_node.part.0
      0.00 ±264%     +20.0%       0.00 ±300%  perf-sched.wait_time.max.ms.__cond_resched.mutex_lock.btrfs_delayed_update_inode.btrfs_update_inode.__btrfs_prealloc_file_range
      0.00 ±264%    -100.0%       0.00        perf-sched.wait_time.max.ms.__cond_resched.mutex_lock.btrfs_delayed_update_inode.btrfs_update_inode.btrfs_dirty_inode
      0.00 ±264%    +380.0%       0.00 ±300%  perf-sched.wait_time.max.ms.__cond_resched.mutex_lock.btrfs_delayed_update_inode.btrfs_update_inode.btrfs_fileattr_set
      0.00 ±264%    +252.0%       0.00 ±245%  perf-sched.wait_time.max.ms.__cond_resched.mutex_lock.btrfs_delayed_update_inode.btrfs_update_inode.btrfs_punch_hole
      0.00 ±173%    +220.0%       0.00 ± 52%  perf-sched.wait_time.max.ms.__cond_resched.mutex_lock.btrfs_delayed_update_inode.btrfs_update_inode.clone_finish_inode_update
      0.00         +1e+104%     100.10 ±299%  perf-sched.wait_time.max.ms.__cond_resched.mutex_lock.perf_poll.do_poll.constprop
      0.00         +4e+101%       0.40 ±300%  perf-sched.wait_time.max.ms.__cond_resched.rcu_gp_cleanup.rcu_gp_kthread.kthread.ret_from_fork
      0.00         +1e+104%     100.02 ±300%  perf-sched.wait_time.max.ms.__cond_resched.shrink_dentry_list.shrink_dcache_parent.d_invalidate.proc_invalidate_siblings_dcache
    251.12 ±171%    +179.6%     702.09 ±128%  perf-sched.wait_time.max.ms.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
      0.00          +4e+98%       0.00 ±300%  perf-sched.wait_time.max.ms.__cond_resched.start_transaction.btrfs_replace_file_extents.btrfs_clone.btrfs_extent_same_range
      0.00        +5.8e+99%       0.01 ±300%  perf-sched.wait_time.max.ms.__cond_resched.task_work_run.exit_to_user_mode_loop.exit_to_user_mode_prepare.syscall_exit_to_user_mode
      0.03 ± 80%    +209.9%       0.08 ±208%  perf-sched.wait_time.max.ms.__cond_resched.truncate_inode_pages_range.btrfs_clone_files.isra.0
      0.00        +1.4e+99%       0.00 ±153%  perf-sched.wait_time.max.ms.__cond_resched.truncate_inode_pages_range.btrfs_punch_hole_lock_range.btrfs_punch_hole.btrfs_fallocate
      0.00 ± 43%      -2.5%       0.00 ± 77%  perf-sched.wait_time.max.ms.__cond_resched.vfs_clone_file_range.ioctl_file_clone.do_vfs_ioctl.__x64_sys_ioctl
      0.08 ±209%     -77.4%       0.02 ±111%  perf-sched.wait_time.max.ms.__cond_resched.vfs_dedupe_file_range_compare.constprop.0.__generic_remap_file_range_prep
      0.00 ± 40%     -37.8%       0.00 ± 82%  perf-sched.wait_time.max.ms.__cond_resched.vfs_fallocate.ioctl_preallocate.__x64_sys_ioctl.do_syscall_64
    999.86            -0.0%     999.86        perf-sched.wait_time.max.ms.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
      0.22 ±173%    +141.8%       0.53 ± 82%  perf-sched.wait_time.max.ms.devkmsg_read.vfs_read.ksys_read.do_syscall_64
      1000            +0.0%       1000        perf-sched.wait_time.max.ms.do_nanosleep.hrtimer_nanosleep.common_nsleep.__x64_sys_clock_nanosleep
      1868 ± 49%     -19.8%       1498 ± 44%  perf-sched.wait_time.max.ms.do_task_dead.do_exit.do_group_exit.__x64_sys_exit_group.do_syscall_64
      0.86 ±  3%      +1.6%       0.87 ±  3%  perf-sched.wait_time.max.ms.do_wait.kernel_wait4.__do_sys_wait4.do_syscall_64
      0.00         +1e+104%     100.02 ±300%  perf-sched.wait_time.max.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_exc_page_fault
      0.01 ± 14%    +127.8%       0.01 ±127%  perf-sched.wait_time.max.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt
      0.09 ±198%     -58.3%       0.04 ± 70%  perf-sched.wait_time.max.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_reschedule_ipi
      0.08 ±214%     -83.2%       0.01 ±104%  perf-sched.wait_time.max.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.syscall_exit_to_user_mode.do_syscall_64
      0.02 ± 81%      -3.8%       0.02 ± 91%  perf-sched.wait_time.max.ms.handle_reserve_ticket.__reserve_bytes.btrfs_reserve_data_bytes.btrfs_fallocate
      0.00 ±174%     -28.9%       0.00 ±200%  perf-sched.wait_time.max.ms.io_schedule.folio_wait_bit_common.filemap_fault.__do_fault
      1008            -0.0%       1008        perf-sched.wait_time.max.ms.pipe_read.vfs_read.ksys_read.do_syscall_64
      1000            -0.0%       1000        perf-sched.wait_time.max.ms.schedule_hrtimeout_range_clock.do_poll.constprop.0.do_sys_poll
    531.24 ±  7%      -5.9%     499.98        perf-sched.wait_time.max.ms.schedule_hrtimeout_range_clock.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
      1.34 ± 92%     -23.4%       1.02 ± 82%  perf-sched.wait_time.max.ms.schedule_preempt_disabled.rwsem_down_read_slowpath.down_read.__btrfs_tree_read_lock
      0.04 ± 54%    +806.7%       0.39 ±206%  perf-sched.wait_time.max.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write.__btrfs_tree_lock
      0.00        +2.1e+99%       0.00 ±157%  perf-sched.wait_time.max.ms.schedule_timeout.__wait_for_common.__flush_work.isra.0
      0.00 ±188%     -70.5%       0.00 ±300%  perf-sched.wait_time.max.ms.schedule_timeout.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr
      0.93            +0.4%       0.94        perf-sched.wait_time.max.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone
    504.62            +0.0%     504.70        perf-sched.wait_time.max.ms.schedule_timeout.kcompactd.kthread.ret_from_fork
      0.00         +5e+101%       0.50 ±202%  perf-sched.wait_time.max.ms.schedule_timeout.khugepaged_wait_work.khugepaged.kthread
      5.00            +0.0%       5.00        perf-sched.wait_time.max.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
      2794 ± 23%      +7.7%       3011 ± 18%  perf-sched.wait_time.max.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
      0.22 ±173%    +141.7%       0.54 ± 82%  perf-sched.wait_time.max.ms.syslog_print.do_syslog.kmsg_read.vfs_read
      0.00 ±264%    +100.0%       0.00 ±200%  perf-sched.wait_time.max.ms.wait_for_partner.fifo_open.do_dentry_open.do_open
      2561 ± 18%     +14.4%       2930 ± 33%  perf-sched.wait_time.max.ms.worker_thread.kthread.ret_from_fork.ret_from_fork_asm
    860.62 ± 13%      -2.8%     836.40 ± 18%  slabinfo.Acpi-State.active_objs
     16.88 ± 13%      -2.8%      16.40 ± 18%  slabinfo.Acpi-State.active_slabs
    860.62 ± 13%      -2.8%     836.40 ± 18%  slabinfo.Acpi-State.num_objs
     16.88 ± 13%      -2.8%      16.40 ± 18%  slabinfo.Acpi-State.num_slabs
     36.00            +0.0%      36.00        slabinfo.DCCP.active_objs
      2.00            +0.0%       2.00        slabinfo.DCCP.active_slabs
     36.00            +0.0%      36.00        slabinfo.DCCP.num_objs
      2.00            +0.0%       2.00        slabinfo.DCCP.num_slabs
     34.00            +0.0%      34.00        slabinfo.DCCPv6.active_objs
      2.00            +0.0%       2.00        slabinfo.DCCPv6.active_slabs
     34.00            +0.0%      34.00        slabinfo.DCCPv6.num_objs
      2.00            +0.0%       2.00        slabinfo.DCCPv6.num_slabs
     96.00            +0.0%      96.00        slabinfo.RAW.active_objs
      3.00            +0.0%       3.00        slabinfo.RAW.active_slabs
     96.00            +0.0%      96.00        slabinfo.RAW.num_objs
      3.00            +0.0%       3.00        slabinfo.RAW.num_slabs
     78.00            +6.7%      83.20 ± 12%  slabinfo.RAWv6.active_objs
      3.00            +6.7%       3.20 ± 12%  slabinfo.RAWv6.active_slabs
     78.00            +6.7%      83.20 ± 12%  slabinfo.RAWv6.num_objs
      3.00            +6.7%       3.20 ± 12%  slabinfo.RAWv6.num_slabs
     54.83 ±  8%      +3.6%      56.80        slabinfo.TCP.active_objs
      3.92 ±  8%      +3.6%       4.06        slabinfo.TCP.active_slabs
     54.83 ±  8%      +3.6%      56.80        slabinfo.TCP.num_objs
      3.92 ±  8%      +3.6%       4.06        slabinfo.TCP.num_slabs
     37.38 ± 11%      -2.6%      36.40 ± 14%  slabinfo.TCPv6.active_objs
      2.88 ± 11%      -2.6%       2.80 ± 14%  slabinfo.TCPv6.active_slabs
     37.38 ± 11%      -2.6%      36.40 ± 14%  slabinfo.TCPv6.num_objs
      2.88 ± 11%      -2.6%       2.80 ± 14%  slabinfo.TCPv6.num_slabs
    105.00 ± 11%      -1.7%     103.20 ± 14%  slabinfo.UDPv6.active_objs
      4.38 ± 11%      -1.7%       4.30 ± 14%  slabinfo.UDPv6.active_slabs
    105.00 ± 11%      -1.7%     103.20 ± 14%  slabinfo.UDPv6.num_objs
      4.38 ± 11%      -1.7%       4.30 ± 14%  slabinfo.UDPv6.num_slabs
    986.19 ±  4%      +2.2%       1008 ±  4%  slabinfo.UNIX.active_objs
     32.87 ±  4%      +2.2%      33.60 ±  4%  slabinfo.UNIX.active_slabs
    986.19 ±  4%      +2.2%       1008 ±  4%  slabinfo.UNIX.num_objs
     32.87 ±  4%      +2.2%      33.60 ±  4%  slabinfo.UNIX.num_slabs
      6255 ±  3%      -0.8%       6207 ±  5%  slabinfo.anon_vma.active_objs
    163.36 ±  3%      -1.2%     161.37 ±  5%  slabinfo.anon_vma.active_slabs
      6371 ±  3%      -1.2%       6293 ±  5%  slabinfo.anon_vma.num_objs
    163.36 ±  3%      -1.2%     161.37 ±  5%  slabinfo.anon_vma.num_slabs
      7503            +2.4%       7681 ±  4%  slabinfo.anon_vma_chain.active_objs
    119.97            +1.6%     121.86 ±  4%  slabinfo.anon_vma_chain.active_slabs
      7678            +1.6%       7798 ±  4%  slabinfo.anon_vma_chain.num_objs
    119.97            +1.6%     121.86 ±  4%  slabinfo.anon_vma_chain.num_slabs
     70.00 ± 14%     +14.3%      80.00 ± 15%  slabinfo.bdev_cache.active_objs
      3.50 ± 14%     +14.3%       4.00 ± 15%  slabinfo.bdev_cache.active_slabs
     70.00 ± 14%     +14.3%      80.00 ± 15%  slabinfo.bdev_cache.num_objs
      3.50 ± 14%     +14.3%       4.00 ± 15%  slabinfo.bdev_cache.num_slabs
     60.55 ± 16%     +15.3%      69.81 ± 10%  slabinfo.bio-120.active_objs
      1.89 ± 16%     +15.3%       2.18 ± 10%  slabinfo.bio-120.active_slabs
     60.55 ± 16%     +15.3%      69.81 ± 10%  slabinfo.bio-120.num_objs
      1.89 ± 16%     +15.3%       2.18 ± 10%  slabinfo.bio-120.num_slabs
    990.14 ±  8%      +7.9%       1068 ±  6%  slabinfo.bio-184.active_objs
     23.57 ±  8%      +7.9%      25.45 ±  6%  slabinfo.bio-184.active_slabs
    990.14 ±  8%      +7.9%       1068 ±  6%  slabinfo.bio-184.num_objs
     23.57 ±  8%      +7.9%      25.45 ±  6%  slabinfo.bio-184.num_slabs
     64.00            +0.0%      64.00        slabinfo.bio-248.active_objs
      2.00            +0.0%       2.00        slabinfo.bio-248.active_slabs
     64.00            +0.0%      64.00        slabinfo.bio-248.num_objs
      2.00            +0.0%       2.00        slabinfo.bio-248.num_slabs
     51.00            +0.0%      51.00        slabinfo.bio-296.active_objs
      1.00            +0.0%       1.00        slabinfo.bio-296.active_slabs
     51.00            +0.0%      51.00        slabinfo.bio-296.num_objs
      1.00            +0.0%       1.00        slabinfo.bio-296.num_slabs
    176.79 ±  8%      +7.2%     189.60 ± 10%  slabinfo.bio-360.active_objs
      4.21 ±  8%      +7.2%       4.51 ± 10%  slabinfo.bio-360.active_slabs
    176.79 ±  8%      +7.2%     189.60 ± 10%  slabinfo.bio-360.num_objs
      4.21 ±  8%      +7.2%       4.51 ± 10%  slabinfo.bio-360.num_slabs
     42.00            +0.0%      42.00        slabinfo.bio-376.active_objs
      1.00            +0.0%       1.00        slabinfo.bio-376.active_slabs
     42.00            +0.0%      42.00        slabinfo.bio-376.num_objs
      1.00            +0.0%       1.00        slabinfo.bio-376.num_slabs
     36.00            +0.0%      36.00        slabinfo.bio-432.active_objs
      1.00            +0.0%       1.00        slabinfo.bio-432.active_slabs
     36.00            +0.0%      36.00        slabinfo.bio-432.num_objs
      1.00            +0.0%       1.00        slabinfo.bio-432.num_slabs
    170.00            +0.0%     170.00        slabinfo.bio_post_read_ctx.active_objs
      2.00            +0.0%       2.00        slabinfo.bio_post_read_ctx.active_slabs
    170.00            +0.0%     170.00        slabinfo.bio_post_read_ctx.num_objs
      2.00            +0.0%       2.00        slabinfo.bio_post_read_ctx.num_slabs
     52.00 ± 29%      -7.7%      48.00 ± 53%  slabinfo.biovec-128.active_objs
      3.25 ± 29%      -7.7%       3.00 ± 53%  slabinfo.biovec-128.active_slabs
     52.00 ± 29%      -7.7%      48.00 ± 53%  slabinfo.biovec-128.num_objs
      3.25 ± 29%      -7.7%       3.00 ± 53%  slabinfo.biovec-128.num_slabs
    423.27 ± 11%      -0.3%     421.86 ± 10%  slabinfo.biovec-64.active_objs
     13.23 ± 11%      -0.3%      13.18 ± 10%  slabinfo.biovec-64.active_slabs
    423.27 ± 11%      -0.3%     421.86 ± 10%  slabinfo.biovec-64.num_objs
     13.23 ± 11%      -0.3%      13.18 ± 10%  slabinfo.biovec-64.num_slabs
     56.14            +2.3%      57.45 ±  3%  slabinfo.biovec-max.active_objs
      7.02            +2.3%       7.18 ±  3%  slabinfo.biovec-max.active_slabs
     56.14            +2.3%      57.45 ±  3%  slabinfo.biovec-max.num_objs
      7.02            +2.3%       7.18 ±  3%  slabinfo.biovec-max.num_slabs
      5847 ±129%     +16.9%       6837 ±122%  slabinfo.btrfs_delayed_ref_head.active_objs
    195.09 ±129%     +16.9%     228.11 ±122%  slabinfo.btrfs_delayed_ref_head.active_slabs
      5852 ±129%     +16.9%       6843 ±122%  slabinfo.btrfs_delayed_ref_head.num_objs
    195.09 ±129%     +16.9%     228.11 ±122%  slabinfo.btrfs_delayed_ref_head.num_slabs
    217.62 ± 11%      +5.8%     230.16 ± 11%  slabinfo.btrfs_extent_buffer.active_objs
      6.40 ± 11%      +5.8%       6.77 ± 11%  slabinfo.btrfs_extent_buffer.active_slabs
    217.62 ± 11%      +5.8%     230.16 ± 11%  slabinfo.btrfs_extent_buffer.num_objs
      6.40 ± 11%      +5.8%       6.77 ± 11%  slabinfo.btrfs_extent_buffer.num_slabs
    376.75 ± 12%      +6.1%     399.82 ± 14%  slabinfo.btrfs_free_space.active_objs
      9.66 ± 12%      +6.1%      10.25 ± 14%  slabinfo.btrfs_free_space.active_slabs
    376.75 ± 12%      +6.1%     399.82 ± 14%  slabinfo.btrfs_free_space.num_objs
      9.66 ± 12%      +6.1%      10.25 ± 14%  slabinfo.btrfs_free_space.num_slabs
    212.92 ± 11%      -0.6%     211.70 ±  9%  slabinfo.btrfs_inode.active_objs
      7.34 ± 11%      -0.6%       7.30 ±  9%  slabinfo.btrfs_inode.active_slabs
    212.92 ± 11%      -0.6%     211.70 ±  9%  slabinfo.btrfs_inode.num_objs
      7.34 ± 11%      -0.6%       7.30 ±  9%  slabinfo.btrfs_inode.num_slabs
      0.65 ±174%    +958.1%       6.90 ±126%  slabinfo.btrfs_ordered_extent.active_objs
      0.02 ±174%    +958.1%       0.18 ±126%  slabinfo.btrfs_ordered_extent.active_slabs
      0.65 ±174%    +958.1%       6.90 ±126%  slabinfo.btrfs_ordered_extent.num_objs
      0.02 ±174%    +958.1%       0.18 ±126%  slabinfo.btrfs_ordered_extent.num_slabs
      1128 ±  3%      -0.8%       1119 ±  4%  slabinfo.btrfs_path.active_objs
     31.36 ±  3%      -0.8%      31.09 ±  4%  slabinfo.btrfs_path.active_slabs
      1128 ±  3%      -0.8%       1119 ±  4%  slabinfo.btrfs_path.num_objs
     31.36 ±  3%      -0.8%      31.09 ±  4%  slabinfo.btrfs_path.num_slabs
      1553 ±  3%      +0.3%       1557 ±  2%  slabinfo.buffer_head.active_objs
     39.83 ±  3%      +0.3%      39.94 ±  2%  slabinfo.buffer_head.active_slabs
      1553 ±  3%      +0.3%       1557 ±  2%  slabinfo.buffer_head.num_objs
     39.83 ±  3%      +0.3%      39.94 ±  2%  slabinfo.buffer_head.num_slabs
      2441 ±  4%      -1.9%       2393 ±  4%  slabinfo.cred_jar.active_objs
     58.13 ±  4%      -1.9%      57.00 ±  4%  slabinfo.cred_jar.active_slabs
      2441 ±  4%      -1.9%       2393 ±  4%  slabinfo.cred_jar.num_objs
     58.13 ±  4%      -1.9%      57.00 ±  4%  slabinfo.cred_jar.num_slabs
     39.00            +0.0%      39.00        slabinfo.dax_cache.active_objs
      1.00            +0.0%       1.00        slabinfo.dax_cache.active_slabs
     39.00            +0.0%      39.00        slabinfo.dax_cache.num_objs
      1.00            +0.0%       1.00        slabinfo.dax_cache.num_slabs
     75745            -0.0%      75742        slabinfo.dentry.active_objs
      1827            -0.1%       1825        slabinfo.dentry.active_slabs
     76744            -0.1%      76680        slabinfo.dentry.num_objs
      1827            -0.1%       1825        slabinfo.dentry.num_slabs
     30.00            +0.0%      30.00        slabinfo.dmaengine-unmap-128.active_objs
      1.00            +0.0%       1.00        slabinfo.dmaengine-unmap-128.active_slabs
     30.00            +0.0%      30.00        slabinfo.dmaengine-unmap-128.num_objs
      1.00            +0.0%       1.00        slabinfo.dmaengine-unmap-128.num_slabs
     64.00            +0.0%      64.00        slabinfo.dmaengine-unmap-2.active_objs
      1.00            +0.0%       1.00        slabinfo.dmaengine-unmap-2.active_slabs
     64.00            +0.0%      64.00        slabinfo.dmaengine-unmap-2.num_objs
      1.00            +0.0%       1.00        slabinfo.dmaengine-unmap-2.num_slabs
     15.00            +0.0%      15.00        slabinfo.dmaengine-unmap-256.active_objs
      1.00            +0.0%       1.00        slabinfo.dmaengine-unmap-256.active_slabs
     15.00            +0.0%      15.00        slabinfo.dmaengine-unmap-256.num_objs
      1.00            +0.0%       1.00        slabinfo.dmaengine-unmap-256.num_slabs
      7872 ±  6%      +0.8%       7936 ±  5%  slabinfo.ep_head.active_objs
     30.75 ±  6%      +0.8%      31.00 ±  5%  slabinfo.ep_head.active_slabs
      7872 ±  6%      +0.8%       7936 ±  5%  slabinfo.ep_head.num_objs
     30.75 ±  6%      +0.8%      31.00 ±  5%  slabinfo.ep_head.num_slabs
    693.75 ± 14%      -2.8%     674.48 ± 14%  slabinfo.file_lock_cache.active_objs
     18.75 ± 14%      -2.8%      18.23 ± 14%  slabinfo.file_lock_cache.active_slabs
    693.75 ± 14%      -2.8%     674.48 ± 14%  slabinfo.file_lock_cache.num_objs
     18.75 ± 14%      -2.8%      18.23 ± 14%  slabinfo.file_lock_cache.num_slabs
      1684            -0.3%       1679        slabinfo.files_cache.active_objs
     36.62            -0.3%      36.50        slabinfo.files_cache.active_slabs
      1684            -0.3%       1679        slabinfo.files_cache.num_objs
     36.62            -0.3%      36.50        slabinfo.files_cache.num_slabs
      5932 ±  2%      +0.5%       5964        slabinfo.filp.active_objs
    282.47 ±  2%      +0.4%     283.61 ±  2%  slabinfo.filp.active_slabs
      9039 ±  2%      +0.4%       9075 ±  2%  slabinfo.filp.num_objs
    282.47 ±  2%      +0.4%     283.61 ±  2%  slabinfo.filp.num_slabs
      2371 ±  8%      +7.7%       2553 ± 11%  slabinfo.fsnotify_mark_connector.active_objs
     18.53 ±  8%      +7.7%      19.95 ± 11%  slabinfo.fsnotify_mark_connector.active_slabs
      2371 ±  8%      +7.7%       2553 ± 11%  slabinfo.fsnotify_mark_connector.num_objs
     18.53 ±  8%      +7.7%      19.95 ± 11%  slabinfo.fsnotify_mark_connector.num_slabs
      6433            -0.4%       6409        slabinfo.ftrace_event_field.active_objs
     88.12            -0.4%      87.80        slabinfo.ftrace_event_field.active_slabs
      6433            -0.4%       6409        slabinfo.ftrace_event_field.num_objs
     88.12            -0.4%      87.80        slabinfo.ftrace_event_field.num_slabs
    942.93 ±  3%      -1.7%     926.82 ±  3%  slabinfo.fuse_request.active_objs
     33.68 ±  3%      -1.7%      33.10 ±  3%  slabinfo.fuse_request.active_slabs
    942.93 ±  3%      -1.7%     926.82 ±  3%  slabinfo.fuse_request.num_objs
     33.68 ±  3%      -1.7%      33.10 ±  3%  slabinfo.fuse_request.num_slabs
     98.00            +0.0%      98.00        slabinfo.hugetlbfs_inode_cache.active_objs
      2.00            +0.0%       2.00        slabinfo.hugetlbfs_inode_cache.active_slabs
     98.00            +0.0%      98.00        slabinfo.hugetlbfs_inode_cache.num_objs
      2.00            +0.0%       2.00        slabinfo.hugetlbfs_inode_cache.num_slabs
     64153            +0.1%      64202        slabinfo.inode_cache.active_objs
      1260            +0.1%       1262        slabinfo.inode_cache.active_slabs
     64287            +0.1%      64367        slabinfo.inode_cache.num_objs
      1260            +0.1%       1262        slabinfo.inode_cache.num_slabs
    191.62 ± 26%      -8.6%     175.20 ± 20%  slabinfo.ip_fib_alias.active_objs
      2.62 ± 26%      -8.6%       2.40 ± 20%  slabinfo.ip_fib_alias.active_slabs
    191.62 ± 26%      -8.6%     175.20 ± 20%  slabinfo.ip_fib_alias.num_objs
      2.62 ± 26%      -8.6%       2.40 ± 20%  slabinfo.ip_fib_alias.num_slabs
    223.12 ± 26%      -8.6%     204.00 ± 20%  slabinfo.ip_fib_trie.active_objs
      2.62 ± 26%      -8.6%       2.40 ± 20%  slabinfo.ip_fib_trie.active_slabs
    223.12 ± 26%      -8.6%     204.00 ± 20%  slabinfo.ip_fib_trie.num_objs
      2.62 ± 26%      -8.6%       2.40 ± 20%  slabinfo.ip_fib_trie.num_slabs
     47676            -0.1%      47638        slabinfo.kernfs_node_cache.active_objs
      1489            -0.1%       1488        slabinfo.kernfs_node_cache.active_slabs
     47676            -0.1%      47638        slabinfo.kernfs_node_cache.num_objs
      1489            -0.1%       1488        slabinfo.kernfs_node_cache.num_slabs
      2852 ±  5%      -1.8%       2800 ±  9%  slabinfo.khugepaged_mm_slot.active_objs
     27.96 ±  5%      -1.8%      27.46 ±  9%  slabinfo.khugepaged_mm_slot.active_slabs
      2852 ±  5%      -1.8%       2800 ±  9%  slabinfo.khugepaged_mm_slot.num_objs
     27.96 ±  5%      -1.8%      27.46 ±  9%  slabinfo.khugepaged_mm_slot.num_slabs
      2123            +0.0%       2124        slabinfo.kmalloc-128.active_objs
     66.37            +0.0%      66.39        slabinfo.kmalloc-128.active_slabs
      2123            +0.0%       2124        slabinfo.kmalloc-128.num_objs
     66.37            +0.0%      66.39        slabinfo.kmalloc-128.num_slabs
     18176            -0.7%      18048        slabinfo.kmalloc-16.active_objs
     71.00            -0.7%      70.50        slabinfo.kmalloc-16.active_slabs
     18176            -0.7%      18048        slabinfo.kmalloc-16.num_objs
     71.00            -0.7%      70.50        slabinfo.kmalloc-16.num_slabs
      2713 ±  4%      -1.0%       2685 ±  2%  slabinfo.kmalloc-192.active_objs
     64.60 ±  4%      -1.0%      63.95 ±  2%  slabinfo.kmalloc-192.active_slabs
      2713 ±  4%      -1.0%       2685 ±  2%  slabinfo.kmalloc-192.num_objs
     64.60 ±  4%      -1.0%      63.95 ±  2%  slabinfo.kmalloc-192.num_slabs
      2439            -0.4%       2428        slabinfo.kmalloc-1k.active_objs
     76.23            -0.4%      75.89        slabinfo.kmalloc-1k.active_slabs
      2439            -0.4%       2428        slabinfo.kmalloc-1k.num_objs
     76.23            -0.4%      75.89        slabinfo.kmalloc-1k.num_slabs
      4185 ±  2%      +2.5%       4291        slabinfo.kmalloc-256.active_objs
    131.05 ±  2%      +2.7%     134.64        slabinfo.kmalloc-256.active_slabs
      4193 ±  2%      +2.7%       4308        slabinfo.kmalloc-256.num_objs
    131.05 ±  2%      +2.7%     134.64        slabinfo.kmalloc-256.num_slabs
      1734 ±  2%      -0.5%       1725 ±  2%  slabinfo.kmalloc-2k.active_objs
    111.17 ±  2%      -0.8%     110.32 ±  2%  slabinfo.kmalloc-2k.active_slabs
      1778 ±  2%      -0.8%       1765 ±  2%  slabinfo.kmalloc-2k.num_objs
    111.17 ±  2%      -0.8%     110.32 ±  2%  slabinfo.kmalloc-2k.num_slabs
     16923            +1.2%      17130        slabinfo.kmalloc-32.active_objs
    132.22            +1.3%     133.91        slabinfo.kmalloc-32.active_slabs
     16923            +1.3%      17140        slabinfo.kmalloc-32.num_objs
    132.22            +1.3%     133.91        slabinfo.kmalloc-32.num_slabs
    649.65            -0.5%     646.70        slabinfo.kmalloc-4k.active_objs
     82.06            -0.3%      81.81        slabinfo.kmalloc-4k.active_slabs
    656.49            -0.3%     654.49        slabinfo.kmalloc-4k.num_objs
     82.06            -0.3%      81.81        slabinfo.kmalloc-4k.num_slabs
      5183 ±  2%      +0.9%       5231 ±  2%  slabinfo.kmalloc-512.active_objs
    162.48 ±  2%      +0.9%     163.93 ±  2%  slabinfo.kmalloc-512.active_slabs
      5199 ±  2%      +0.9%       5245 ±  2%  slabinfo.kmalloc-512.num_objs
    162.48 ±  2%      +0.9%     163.93 ±  2%  slabinfo.kmalloc-512.num_slabs
     19017            -0.1%      18996        slabinfo.kmalloc-64.active_objs
    297.15            -0.1%     296.82        slabinfo.kmalloc-64.active_slabs
     19017            -0.1%      18996        slabinfo.kmalloc-64.num_objs
    297.15            -0.1%     296.82        slabinfo.kmalloc-64.num_slabs
     33657            -0.2%      33580        slabinfo.kmalloc-8.active_objs
     65.75            -0.2%      65.60        slabinfo.kmalloc-8.active_slabs
     33664            -0.2%      33587        slabinfo.kmalloc-8.num_objs
     65.75            -0.2%      65.60        slabinfo.kmalloc-8.num_slabs
    197.38            -0.5%     196.45        slabinfo.kmalloc-8k.active_objs
     50.72            -0.1%      50.65        slabinfo.kmalloc-8k.active_slabs
    202.86            -0.1%     202.60        slabinfo.kmalloc-8k.num_objs
     50.72            -0.1%      50.65        slabinfo.kmalloc-8k.num_slabs
     20981            -0.6%      20845        slabinfo.kmalloc-96.active_objs
    517.73            -0.8%     513.39        slabinfo.kmalloc-96.active_slabs
     21744            -0.8%      21562        slabinfo.kmalloc-96.num_objs
    517.73            -0.8%     513.39        slabinfo.kmalloc-96.num_slabs
    447.63 ± 10%     +10.6%     495.16 ±  6%  slabinfo.kmalloc-cg-128.active_objs
     13.99 ± 10%     +10.6%      15.47 ±  6%  slabinfo.kmalloc-cg-128.active_slabs
    447.63 ± 10%     +10.6%     495.16 ±  6%  slabinfo.kmalloc-cg-128.num_objs
     13.99 ± 10%     +10.6%      15.47 ±  6%  slabinfo.kmalloc-cg-128.num_slabs
      3257 ± 10%      +8.9%       3547 ±  8%  slabinfo.kmalloc-cg-16.active_objs
     12.73 ± 10%      +8.9%      13.86 ±  8%  slabinfo.kmalloc-cg-16.active_slabs
      3257 ± 10%      +8.9%       3547 ±  8%  slabinfo.kmalloc-cg-16.num_objs
     12.73 ± 10%      +8.9%      13.86 ±  8%  slabinfo.kmalloc-cg-16.num_slabs
      1638            +0.5%       1646        slabinfo.kmalloc-cg-192.active_objs
     39.00            +0.5%      39.20        slabinfo.kmalloc-cg-192.active_slabs
      1638            +0.5%       1646        slabinfo.kmalloc-cg-192.num_objs
     39.00            +0.5%      39.20        slabinfo.kmalloc-cg-192.num_slabs
      1568 ±  3%      +3.5%       1622 ±  5%  slabinfo.kmalloc-cg-1k.active_objs
     49.00 ±  3%      +3.5%      50.70 ±  5%  slabinfo.kmalloc-cg-1k.active_slabs
      1568 ±  3%      +3.5%       1622 ±  5%  slabinfo.kmalloc-cg-1k.num_objs
     49.00 ±  3%      +3.5%      50.70 ±  5%  slabinfo.kmalloc-cg-1k.num_slabs
    280.00 ±  9%      -0.6%     278.40 ±  8%  slabinfo.kmalloc-cg-256.active_objs
      8.75 ±  9%      -0.6%       8.70 ±  8%  slabinfo.kmalloc-cg-256.active_slabs
    280.00 ±  9%      -0.6%     278.40 ±  8%  slabinfo.kmalloc-cg-256.num_objs
      8.75 ±  9%      -0.6%       8.70 ±  8%  slabinfo.kmalloc-cg-256.num_slabs
    530.09 ±  7%      -0.7%     526.12 ±  6%  slabinfo.kmalloc-cg-2k.active_objs
     33.13 ±  7%      -0.7%      32.88 ±  6%  slabinfo.kmalloc-cg-2k.active_slabs
    530.09 ±  7%      -0.7%     526.12 ±  6%  slabinfo.kmalloc-cg-2k.num_objs
     33.13 ±  7%      -0.7%      32.88 ±  6%  slabinfo.kmalloc-cg-2k.num_slabs
      5010 ±  2%      -0.4%       4991        slabinfo.kmalloc-cg-32.active_objs
     39.14 ±  2%      -0.4%      39.00        slabinfo.kmalloc-cg-32.active_slabs
      5010 ±  2%      -0.4%       4991        slabinfo.kmalloc-cg-32.num_objs
     39.14 ±  2%      -0.4%      39.00        slabinfo.kmalloc-cg-32.num_slabs
    498.48            -1.8%     489.34 ±  3%  slabinfo.kmalloc-cg-4k.active_objs
     65.83            -1.4%      64.93 ±  2%  slabinfo.kmalloc-cg-4k.active_slabs
    526.61            -1.4%     519.45 ±  2%  slabinfo.kmalloc-cg-4k.num_objs
     65.83            -1.4%      64.93 ±  2%  slabinfo.kmalloc-cg-4k.num_slabs
      1188            +0.5%       1193        slabinfo.kmalloc-cg-512.active_objs
     37.12            +0.5%      37.30        slabinfo.kmalloc-cg-512.active_slabs
      1188            +0.5%       1193        slabinfo.kmalloc-cg-512.num_objs
     37.12            +0.5%      37.30        slabinfo.kmalloc-cg-512.num_slabs
      1572 ±  8%      +1.5%       1596 ±  7%  slabinfo.kmalloc-cg-64.active_objs
     24.58 ±  8%      +1.5%      24.94 ±  7%  slabinfo.kmalloc-cg-64.active_slabs
      1572 ±  8%      +1.5%       1596 ±  7%  slabinfo.kmalloc-cg-64.num_objs
     24.58 ±  8%      +1.5%      24.94 ±  7%  slabinfo.kmalloc-cg-64.num_slabs
     18432            +0.0%      18432        slabinfo.kmalloc-cg-8.active_objs
     36.00            +0.0%      36.00        slabinfo.kmalloc-cg-8.active_slabs
     18432            +0.0%      18432        slabinfo.kmalloc-cg-8.num_objs
     36.00            +0.0%      36.00        slabinfo.kmalloc-cg-8.num_slabs
     59.11 ±  6%      -4.1%      56.71 ±  7%  slabinfo.kmalloc-cg-8k.active_objs
     14.78 ±  6%      -4.1%      14.18 ±  7%  slabinfo.kmalloc-cg-8k.active_slabs
     59.11 ±  6%      -4.1%      56.71 ±  7%  slabinfo.kmalloc-cg-8k.num_objs
     14.78 ±  6%      -4.1%      14.18 ±  7%  slabinfo.kmalloc-cg-8k.num_slabs
    969.83 ± 11%      -0.1%     969.09 ±  6%  slabinfo.kmalloc-cg-96.active_objs
     23.09 ± 11%      -0.1%      23.07 ±  6%  slabinfo.kmalloc-cg-96.active_slabs
    969.83 ± 11%      -0.1%     969.09 ±  6%  slabinfo.kmalloc-cg-96.num_objs
     23.09 ± 11%      -0.1%      23.07 ±  6%  slabinfo.kmalloc-cg-96.num_slabs
     88.00 ± 24%      +9.1%      96.00 ± 21%  slabinfo.kmalloc-rcl-128.active_objs
      2.75 ± 24%      +9.1%       3.00 ± 21%  slabinfo.kmalloc-rcl-128.active_slabs
     88.00 ± 24%      +9.1%      96.00 ± 21%  slabinfo.kmalloc-rcl-128.num_objs
      2.75 ± 24%      +9.1%       3.00 ± 21%  slabinfo.kmalloc-rcl-128.num_slabs
     42.00            +0.0%      42.00        slabinfo.kmalloc-rcl-192.active_objs
      1.00            +0.0%       1.00        slabinfo.kmalloc-rcl-192.active_slabs
     42.00            +0.0%      42.00        slabinfo.kmalloc-rcl-192.num_objs
      1.00            +0.0%       1.00        slabinfo.kmalloc-rcl-192.num_slabs
      6310 ±  2%      +1.7%       6416 ±  2%  slabinfo.kmalloc-rcl-64.active_objs
    100.99 ±  2%      +0.4%     101.41 ±  2%  slabinfo.kmalloc-rcl-64.active_slabs
      6463 ±  2%      +0.4%       6490 ±  2%  slabinfo.kmalloc-rcl-64.num_objs
    100.99 ±  2%      +0.4%     101.41 ±  2%  slabinfo.kmalloc-rcl-64.num_slabs
      1386 ±  7%      -0.9%       1373 ±  7%  slabinfo.kmalloc-rcl-96.active_objs
     33.00 ±  7%      -0.9%      32.70 ±  7%  slabinfo.kmalloc-rcl-96.active_slabs
      1386 ±  7%      -0.9%       1373 ±  7%  slabinfo.kmalloc-rcl-96.num_objs
     33.00 ±  7%      -0.9%      32.70 ±  7%  slabinfo.kmalloc-rcl-96.num_slabs
    416.00 ±  9%      -2.3%     406.40 ±  9%  slabinfo.kmem_cache.active_objs
     13.00 ±  9%      -2.3%      12.70 ±  9%  slabinfo.kmem_cache.active_slabs
    416.00 ±  9%      -2.3%     406.40 ±  9%  slabinfo.kmem_cache.num_objs
     13.00 ±  9%      -2.3%      12.70 ±  9%  slabinfo.kmem_cache.num_slabs
    632.00 ± 11%      -1.8%     620.80 ± 13%  slabinfo.kmem_cache_node.active_objs
      9.88 ± 11%      -1.8%       9.70 ± 13%  slabinfo.kmem_cache_node.active_slabs
    632.00 ± 11%      -1.8%     620.80 ± 13%  slabinfo.kmem_cache_node.num_objs
      9.88 ± 11%      -1.8%       9.70 ± 13%  slabinfo.kmem_cache_node.num_slabs
      9711 ± 77%      +4.6%      10157 ± 81%  slabinfo.kvm_async_pf.active_objs
    324.09 ± 77%      +4.5%     338.80 ± 81%  slabinfo.kvm_async_pf.active_slabs
      9722 ± 77%      +4.5%      10163 ± 81%  slabinfo.kvm_async_pf.num_objs
    324.09 ± 77%      +4.5%     338.80 ± 81%  slabinfo.kvm_async_pf.num_slabs
      7974 ±  5%      -0.8%       7909 ±  4%  slabinfo.lsm_file_cache.active_objs
     47.18 ±  4%      -1.0%      46.72 ±  4%  slabinfo.lsm_file_cache.active_slabs
      8021 ±  4%      -1.0%       7941 ±  4%  slabinfo.lsm_file_cache.num_objs
     47.18 ±  4%      -1.0%      46.72 ±  4%  slabinfo.lsm_file_cache.num_slabs
      4530            -1.2%       4476 ±  3%  slabinfo.maple_node.active_objs
    164.60 ±  2%      -2.0%     161.37 ±  5%  slabinfo.maple_node.active_slabs
      5267 ±  2%      -2.0%       5163 ±  5%  slabinfo.maple_node.num_objs
    164.60 ±  2%      -2.0%     161.37 ±  5%  slabinfo.maple_node.num_slabs
    903.00            +0.5%     907.20        slabinfo.mm_struct.active_objs
     37.62            +0.5%      37.80        slabinfo.mm_struct.active_slabs
    903.00            +0.5%     907.20        slabinfo.mm_struct.num_objs
     37.62            +0.5%      37.80        slabinfo.mm_struct.num_slabs
    930.36 ±  6%      +4.0%     967.46 ±  8%  slabinfo.mnt_cache.active_objs
     18.24 ±  6%      +4.0%      18.97 ±  8%  slabinfo.mnt_cache.active_slabs
    930.36 ±  6%      +4.0%     967.46 ±  8%  slabinfo.mnt_cache.num_objs
     18.24 ±  6%      +4.0%      18.97 ±  8%  slabinfo.mnt_cache.num_slabs
     34.00            +0.0%      34.00        slabinfo.mqueue_inode_cache.active_objs
      1.00            +0.0%       1.00        slabinfo.mqueue_inode_cache.active_slabs
     34.00            +0.0%      34.00        slabinfo.mqueue_inode_cache.num_objs
      1.00            +0.0%       1.00        slabinfo.mqueue_inode_cache.num_slabs
    290.00            -0.7%     288.00        slabinfo.names_cache.active_objs
     36.25            -0.7%      36.00        slabinfo.names_cache.active_slabs
    290.00            -0.7%     288.00        slabinfo.names_cache.num_objs
     36.25            -0.7%      36.00        slabinfo.names_cache.num_slabs
      7.00            +0.0%       7.00        slabinfo.net_namespace.active_objs
      1.00            +0.0%       1.00        slabinfo.net_namespace.active_slabs
      7.00            +0.0%       7.00        slabinfo.net_namespace.num_objs
      1.00            +0.0%       1.00        slabinfo.net_namespace.num_slabs
     46.00            +0.0%      46.00        slabinfo.nfs_commit_data.active_objs
      1.00            +0.0%       1.00        slabinfo.nfs_commit_data.active_slabs
     46.00            +0.0%      46.00        slabinfo.nfs_commit_data.num_objs
      1.00            +0.0%       1.00        slabinfo.nfs_commit_data.num_slabs
     36.00            +0.0%      36.00        slabinfo.nfs_read_data.active_objs
      1.00            +0.0%       1.00        slabinfo.nfs_read_data.active_slabs
     36.00            +0.0%      36.00        slabinfo.nfs_read_data.num_objs
      1.00            +0.0%       1.00        slabinfo.nfs_read_data.num_slabs
    338.26 ±  9%      -5.3%     320.40 ± 13%  slabinfo.nsproxy.active_objs
      6.04 ±  9%      -5.3%       5.72 ± 13%  slabinfo.nsproxy.active_slabs
    338.26 ±  9%      -5.3%     320.40 ± 13%  slabinfo.nsproxy.num_objs
      6.04 ±  9%      -5.3%       5.72 ± 13%  slabinfo.nsproxy.num_slabs
     60.00            +0.0%      60.00        slabinfo.numa_policy.active_objs
      2.00            +0.0%       2.00        slabinfo.numa_policy.active_slabs
     60.00            +0.0%      60.00        slabinfo.numa_policy.num_objs
      2.00            +0.0%       2.00        slabinfo.numa_policy.num_slabs
      1489            +0.0%       1489        slabinfo.perf_event.active_objs
     62.34 ±  2%      +0.1%      62.38        slabinfo.perf_event.active_slabs
      1558 ±  2%      +0.1%       1559        slabinfo.perf_event.num_objs
     62.34 ±  2%      +0.1%      62.38        slabinfo.perf_event.num_slabs
      1822 ±  3%      +0.4%       1829 ±  2%  slabinfo.pid.active_objs
     56.95 ±  3%      +0.4%      57.16 ±  2%  slabinfo.pid.active_slabs
      1822 ±  3%      +0.4%       1829 ±  2%  slabinfo.pid.num_objs
     56.95 ±  3%      +0.4%      57.16 ±  2%  slabinfo.pid.num_slabs
      2692 ±  2%      +0.7%       2711        slabinfo.pool_workqueue.active_objs
     84.15 ±  2%      +0.7%      84.75        slabinfo.pool_workqueue.active_slabs
      2692 ±  2%      +0.7%       2711        slabinfo.pool_workqueue.num_objs
     84.15 ±  2%      +0.7%      84.75        slabinfo.pool_workqueue.num_slabs
      1013 ±  2%      +4.9%       1062 ±  5%  slabinfo.proc_dir_entry.active_objs
     24.12 ±  2%      +4.9%      25.30 ±  5%  slabinfo.proc_dir_entry.active_slabs
      1013 ±  2%      +4.9%       1062 ±  5%  slabinfo.proc_dir_entry.num_objs
     24.12 ±  2%      +4.9%      25.30 ±  5%  slabinfo.proc_dir_entry.num_slabs
      5273            +0.8%       5313        slabinfo.proc_inode_cache.active_objs
    114.97            +0.8%     115.94        slabinfo.proc_inode_cache.active_slabs
      5288            +0.8%       5333        slabinfo.proc_inode_cache.num_objs
    114.97            +0.8%     115.94        slabinfo.proc_inode_cache.num_slabs
     29682            -0.0%      29682        slabinfo.radix_tree_node.active_objs
      1062            +0.0%       1062        slabinfo.radix_tree_node.active_slabs
     29749            +0.0%      29752        slabinfo.radix_tree_node.num_objs
      1062            +0.0%       1062        slabinfo.radix_tree_node.num_slabs
     87.50 ± 20%      +4.0%      91.00 ± 18%  slabinfo.request_queue.active_objs
      2.50 ± 20%      +4.0%       2.60 ± 18%  slabinfo.request_queue.active_slabs
     87.50 ± 20%      +4.0%      91.00 ± 18%  slabinfo.request_queue.num_objs
      2.50 ± 20%      +4.0%       2.60 ± 18%  slabinfo.request_queue.num_slabs
     46.00            +0.0%      46.00        slabinfo.rpc_inode_cache.active_objs
      1.00            +0.0%       1.00        slabinfo.rpc_inode_cache.active_slabs
     46.00            +0.0%      46.00        slabinfo.rpc_inode_cache.num_objs
      1.00            +0.0%       1.00        slabinfo.rpc_inode_cache.num_slabs
    352.00            +0.0%     352.00        slabinfo.scsi_sense_cache.active_objs
     11.00            +0.0%      11.00        slabinfo.scsi_sense_cache.active_slabs
    352.00            +0.0%     352.00        slabinfo.scsi_sense_cache.num_objs
     11.00            +0.0%      11.00        slabinfo.scsi_sense_cache.num_slabs
      1407            -0.7%       1397        slabinfo.seq_file.active_objs
     41.40            -0.7%      41.11        slabinfo.seq_file.active_slabs
      1407            -0.7%       1397        slabinfo.seq_file.num_objs
     41.40            -0.7%      41.11        slabinfo.seq_file.num_slabs
      8021            -0.2%       8007        slabinfo.shared_policy_node.active_objs
     94.38            -0.2%      94.20        slabinfo.shared_policy_node.active_slabs
      8021            -0.2%       8007        slabinfo.shared_policy_node.num_objs
     94.38            -0.2%      94.20        slabinfo.shared_policy_node.num_slabs
      2346 ±  2%      -2.4%       2290 ±  3%  slabinfo.shmem_inode_cache.active_objs
     54.58 ±  2%      -2.4%      53.26 ±  3%  slabinfo.shmem_inode_cache.active_slabs
      2346 ±  2%      -2.4%       2290 ±  3%  slabinfo.shmem_inode_cache.num_objs
     54.58 ±  2%      -2.4%      53.26 ±  3%  slabinfo.shmem_inode_cache.num_slabs
      1053            -0.9%       1043        slabinfo.sighand_cache.active_objs
     70.23            -0.9%      69.58        slabinfo.sighand_cache.active_slabs
      1053            -0.9%       1043        slabinfo.sighand_cache.num_objs
     70.23            -0.9%      69.58        slabinfo.sighand_cache.num_slabs
      1711 ±  3%      -1.0%       1694 ±  3%  slabinfo.signal_cache.active_objs
     61.14 ±  3%      -0.9%      60.60 ±  3%  slabinfo.signal_cache.active_slabs
      1711 ±  3%      -0.9%       1696 ±  3%  slabinfo.signal_cache.num_objs
     61.14 ±  3%      -0.9%      60.60 ±  3%  slabinfo.signal_cache.num_slabs
      2441 ±  2%      +1.1%       2468 ±  3%  slabinfo.sigqueue.active_objs
     47.88 ±  2%      +1.1%      48.40 ±  3%  slabinfo.sigqueue.active_slabs
      2441 ±  2%      +1.1%       2468 ±  3%  slabinfo.sigqueue.num_objs
     47.88 ±  2%      +1.1%      48.40 ±  3%  slabinfo.sigqueue.num_slabs
    326.54 ±  8%      +2.0%     333.10 ± 14%  slabinfo.skbuff_ext_cache.active_objs
      7.77 ±  8%      +2.0%       7.93 ± 14%  slabinfo.skbuff_ext_cache.active_slabs
    326.54 ±  8%      +2.0%     333.10 ± 14%  slabinfo.skbuff_ext_cache.num_objs
      7.77 ±  8%      +2.0%       7.93 ± 14%  slabinfo.skbuff_ext_cache.num_slabs
      2389 ±  7%      +5.9%       2531 ±  7%  slabinfo.skbuff_head_cache.active_objs
     76.38 ±  7%      +4.2%      79.60 ±  7%  slabinfo.skbuff_head_cache.active_slabs
      2444 ±  7%      +4.2%       2547 ±  7%  slabinfo.skbuff_head_cache.num_objs
     76.38 ±  7%      +4.2%      79.60 ±  7%  slabinfo.skbuff_head_cache.num_slabs
      2454 ±  4%      +1.4%       2488 ±  7%  slabinfo.skbuff_small_head.active_objs
     49.62 ±  3%      +1.8%      50.50 ±  6%  slabinfo.skbuff_small_head.active_slabs
      2530 ±  3%      +1.8%       2575 ±  6%  slabinfo.skbuff_small_head.num_objs
     49.62 ±  3%      +1.8%      50.50 ±  6%  slabinfo.skbuff_small_head.num_slabs
      1506 ±  2%      +1.2%       1524 ±  2%  slabinfo.sock_inode_cache.active_objs
     38.62 ±  2%      +1.2%      39.10 ±  2%  slabinfo.sock_inode_cache.active_slabs
      1506 ±  2%      +1.2%       1524 ±  2%  slabinfo.sock_inode_cache.num_objs
     38.62 ±  2%      +1.2%      39.10 ±  2%  slabinfo.sock_inode_cache.num_slabs
      1045 ± 10%      +1.6%       1062 ±  8%  slabinfo.task_group.active_objs
     20.49 ± 10%      +1.6%      20.83 ±  8%  slabinfo.task_group.active_slabs
      1045 ± 10%      +1.6%       1062 ±  8%  slabinfo.task_group.num_objs
     20.49 ± 10%      +1.6%      20.83 ±  8%  slabinfo.task_group.num_slabs
    825.90            +1.0%     833.92        slabinfo.task_struct.active_objs
    826.76            +1.0%     834.85        slabinfo.task_struct.active_slabs
    826.76            +1.0%     834.85        slabinfo.task_struct.num_objs
    826.76            +1.0%     834.85        slabinfo.task_struct.num_slabs
    293.53 ±  6%      -5.6%     277.11 ±  9%  slabinfo.taskstats.active_objs
      7.93 ±  6%      -5.6%       7.49 ±  9%  slabinfo.taskstats.active_slabs
    293.53 ±  6%      -5.6%     277.11 ±  9%  slabinfo.taskstats.num_objs
      7.93 ±  6%      -5.6%       7.49 ±  9%  slabinfo.taskstats.num_slabs
      2294            +0.1%       2295        slabinfo.trace_event_file.active_objs
     49.88            +0.1%      49.90        slabinfo.trace_event_file.active_slabs
      2294            +0.1%       2295        slabinfo.trace_event_file.num_objs
     49.88            +0.1%      49.90        slabinfo.trace_event_file.num_slabs
    788.76 ±  4%      -0.2%     787.16 ±  6%  slabinfo.tracefs_inode_cache.active_objs
     15.78 ±  4%      -0.2%      15.74 ±  6%  slabinfo.tracefs_inode_cache.active_slabs
    788.76 ±  4%      -0.2%     787.16 ±  6%  slabinfo.tracefs_inode_cache.num_objs
     15.78 ±  4%      -0.2%      15.74 ±  6%  slabinfo.tracefs_inode_cache.num_slabs
     30.00            +0.0%      30.00        slabinfo.tw_sock_TCP.active_objs
      1.00            +0.0%       1.00        slabinfo.tw_sock_TCP.active_slabs
     30.00            +0.0%      30.00        slabinfo.tw_sock_TCP.num_objs
      1.00            +0.0%       1.00        slabinfo.tw_sock_TCP.num_slabs
     74.00            -5.0%      70.30 ± 15%  slabinfo.uts_namespace.active_objs
      2.00            -5.0%       1.90 ± 15%  slabinfo.uts_namespace.active_slabs
     74.00            -5.0%      70.30 ± 15%  slabinfo.uts_namespace.num_objs
      2.00            -5.0%       1.90 ± 15%  slabinfo.uts_namespace.num_slabs
      8533 ±  2%      -0.8%       8462        slabinfo.vm_area_struct.active_objs
    197.16 ±  2%      -0.8%     195.49        slabinfo.vm_area_struct.active_slabs
      8674 ±  2%      -0.8%       8601        slabinfo.vm_area_struct.num_objs
    197.16 ±  2%      -0.8%     195.49        slabinfo.vm_area_struct.num_slabs
     11310 ±  2%      +0.5%      11367 ±  3%  slabinfo.vma_lock.active_objs
    113.38 ±  2%      +0.0%     113.38 ±  3%  slabinfo.vma_lock.active_slabs
     11564 ±  2%      +0.0%      11565 ±  3%  slabinfo.vma_lock.num_objs
    113.38 ±  2%      +0.0%     113.38 ±  3%  slabinfo.vma_lock.num_slabs
     50524            +0.1%      50562        slabinfo.vmap_area.active_objs
    903.07            +0.0%     903.44        slabinfo.vmap_area.active_slabs
     50571            +0.0%      50592        slabinfo.vmap_area.num_objs
    903.07            +0.0%     903.44        slabinfo.vmap_area.num_slabs
    143.69            +0.2%     144.05        slabinfo.xfrm_dst_cache.active_objs
      2.82            +0.2%       2.82        slabinfo.xfrm_dst_cache.active_slabs
    143.69            +0.2%     144.05        slabinfo.xfrm_dst_cache.num_objs
      2.82            +0.2%       2.82        slabinfo.xfrm_dst_cache.num_slabs
     20.08            -1.5       18.61        perf-profile.calltrace.cycles-pp.truncate_inode_pages_range.btrfs_clone_files.btrfs_remap_file_range.do_clone_file_range.vfs_clone_file_range
     11.79 ±  2%      -0.9       10.85        perf-profile.calltrace.cycles-pp.truncate_cleanup_folio.truncate_inode_pages_range.btrfs_clone_files.btrfs_remap_file_range.do_clone_file_range
     11.45 ±  2%      -0.9       10.53        perf-profile.calltrace.cycles-pp.btrfs_invalidate_folio.truncate_cleanup_folio.truncate_inode_pages_range.btrfs_clone_files.btrfs_remap_file_range
     16.01 ±  2%      -0.8       15.17 ±  3%  perf-profile.calltrace.cycles-pp.filemap_read_folio.do_read_cache_folio.vfs_dedupe_file_range_compare.__generic_remap_file_range_prep.generic_remap_file_range_prep
     15.67 ±  2%      -0.8       14.87 ±  3%  perf-profile.calltrace.cycles-pp.btrfs_read_folio.filemap_read_folio.do_read_cache_folio.vfs_dedupe_file_range_compare.__generic_remap_file_range_prep
      9.93 ±  2%      -0.6        9.38 ±  2%  perf-profile.calltrace.cycles-pp.memcmp.vfs_dedupe_file_range_compare.__generic_remap_file_range_prep.generic_remap_file_range_prep.btrfs_remap_file_range
     12.06 ±  2%      -0.6       11.51 ±  3%  perf-profile.calltrace.cycles-pp.btrfs_do_readpage.btrfs_read_folio.filemap_read_folio.do_read_cache_folio.vfs_dedupe_file_range_compare
      4.16 ±  3%      -0.4        3.76 ±  3%  perf-profile.calltrace.cycles-pp.__clear_extent_bit.btrfs_invalidate_folio.truncate_cleanup_folio.truncate_inode_pages_range.btrfs_clone_files
     27.14 ±  4%      -0.4       26.75 ±  6%  perf-profile.calltrace.cycles-pp.do_idle.cpu_startup_entry.start_secondary.secondary_startup_64_no_verify
     27.15 ±  4%      -0.4       26.77 ±  6%  perf-profile.calltrace.cycles-pp.cpu_startup_entry.start_secondary.secondary_startup_64_no_verify
     27.15 ±  4%      -0.4       26.77 ±  6%  perf-profile.calltrace.cycles-pp.start_secondary.secondary_startup_64_no_verify
     26.86 ±  4%      -0.4       26.51 ±  6%  perf-profile.calltrace.cycles-pp.cpuidle_idle_call.do_idle.cpu_startup_entry.start_secondary.secondary_startup_64_no_verify
     26.47 ±  4%      -0.3       26.13 ±  6%  perf-profile.calltrace.cycles-pp.cpuidle_enter.cpuidle_idle_call.do_idle.cpu_startup_entry.start_secondary
      8.63 ±  2%      -0.3        8.29 ±  2%  perf-profile.calltrace.cycles-pp.filemap_add_folio.do_read_cache_folio.vfs_dedupe_file_range_compare.__generic_remap_file_range_prep.generic_remap_file_range_prep
      0.33 ± 77%      -0.3        0.07 ±300%  perf-profile.calltrace.cycles-pp.alloc_pages.folio_alloc.do_read_cache_folio.vfs_dedupe_file_range_compare.__generic_remap_file_range_prep
      3.44 ±  4%      -0.3        3.18 ±  3%  perf-profile.calltrace.cycles-pp.btrfs_lock_and_flush_ordered_range.btrfs_read_folio.filemap_read_folio.do_read_cache_folio.vfs_dedupe_file_range_compare
     23.72 ±  4%      -0.3       23.46 ±  5%  perf-profile.calltrace.cycles-pp.intel_idle_ibrs.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call.do_idle
      3.23 ±  3%      -0.3        2.98 ±  3%  perf-profile.calltrace.cycles-pp.delete_from_page_cache_batch.truncate_inode_pages_range.btrfs_clone_files.btrfs_remap_file_range.do_clone_file_range
      0.57 ±  3%      -0.2        0.32 ± 81%  perf-profile.calltrace.cycles-pp.kmem_cache_alloc.alloc_extent_state.__clear_extent_bit.try_release_extent_mapping.btrfs_invalidate_folio
     27.67 ±  4%      -0.2       27.44 ±  6%  perf-profile.calltrace.cycles-pp.secondary_startup_64_no_verify
      0.61 ±  4%      -0.2        0.38 ± 65%  perf-profile.calltrace.cycles-pp.find_free_extent.btrfs_reserve_extent.__btrfs_prealloc_file_range.btrfs_prealloc_file_range.btrfs_fallocate
      0.62 ±  9%      -0.2        0.40 ± 66%  perf-profile.calltrace.cycles-pp.clock_nanosleep
      0.26 ±100%      -0.2        0.05 ±300%  perf-profile.calltrace.cycles-pp.folio_mark_accessed.do_read_cache_folio.vfs_dedupe_file_range_compare.__generic_remap_file_range_prep.generic_remap_file_range_prep
      3.78 ±  2%      -0.2        3.58 ±  3%  perf-profile.calltrace.cycles-pp.__folio_batch_release.truncate_inode_pages_range.btrfs_clone_files.btrfs_remap_file_range.do_clone_file_range
      3.68 ±  2%      -0.2        3.48 ±  3%  perf-profile.calltrace.cycles-pp.release_pages.__folio_batch_release.truncate_inode_pages_range.btrfs_clone_files.btrfs_remap_file_range
      5.42 ±  2%      -0.2        5.22 ±  2%  perf-profile.calltrace.cycles-pp.__filemap_add_folio.filemap_add_folio.do_read_cache_folio.vfs_dedupe_file_range_compare.__generic_remap_file_range_prep
      2.47 ±  2%      -0.2        2.28 ±  5%  perf-profile.calltrace.cycles-pp.lock_extent.btrfs_invalidate_folio.truncate_cleanup_folio.truncate_inode_pages_range.btrfs_clone_files
      0.62 ±  7%      -0.2        0.43 ± 67%  perf-profile.calltrace.cycles-pp.intel_idle.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call.do_idle
      0.66 ±  9%      -0.2        0.48 ± 51%  perf-profile.calltrace.cycles-pp.set_page_extent_mapped.btrfs_do_readpage.btrfs_read_folio.filemap_read_folio.do_read_cache_folio
      0.62 ±  4%      -0.2        0.45 ± 50%  perf-profile.calltrace.cycles-pp.btrfs_reserve_extent.__btrfs_prealloc_file_range.btrfs_prealloc_file_range.btrfs_fallocate.vfs_fallocate
      1.50 ±  5%      -0.2        1.32 ±  5%  perf-profile.calltrace.cycles-pp.clear_state_bit.__clear_extent_bit.btrfs_invalidate_folio.truncate_cleanup_folio.truncate_inode_pages_range
      2.50 ±  4%      -0.2        2.33 ±  4%  perf-profile.calltrace.cycles-pp.ioctl_preallocate.__x64_sys_ioctl.do_syscall_64.entry_SYSCALL_64_after_hwframe.ioctl
      2.37 ±  4%      -0.2        2.20 ±  4%  perf-profile.calltrace.cycles-pp.lock_extent.btrfs_lock_and_flush_ordered_range.btrfs_read_folio.filemap_read_folio.do_read_cache_folio
      2.28            -0.2        2.11 ±  5%  perf-profile.calltrace.cycles-pp.__set_extent_bit.lock_extent.btrfs_invalidate_folio.truncate_cleanup_folio.truncate_inode_pages_range
      2.46 ±  4%      -0.2        2.29 ±  3%  perf-profile.calltrace.cycles-pp.vfs_fallocate.ioctl_preallocate.__x64_sys_ioctl.do_syscall_64.entry_SYSCALL_64_after_hwframe
      2.20 ±  5%      -0.2        2.04 ±  5%  perf-profile.calltrace.cycles-pp.__set_extent_bit.lock_extent.btrfs_lock_and_flush_ordered_range.btrfs_read_folio.filemap_read_folio
      2.37 ±  3%      -0.2        2.21 ±  4%  perf-profile.calltrace.cycles-pp.btrfs_fallocate.vfs_fallocate.ioctl_preallocate.__x64_sys_ioctl.do_syscall_64
      5.28            -0.1        5.14 ±  2%  perf-profile.calltrace.cycles-pp.memset_orig.btrfs_do_readpage.btrfs_read_folio.filemap_read_folio.do_read_cache_folio
      2.86 ±  3%      -0.1        2.72 ±  3%  perf-profile.calltrace.cycles-pp.folio_add_lru.filemap_add_folio.do_read_cache_folio.vfs_dedupe_file_range_compare.__generic_remap_file_range_prep
      3.71 ±  2%      -0.1        3.57 ±  4%  perf-profile.calltrace.cycles-pp.__clear_extent_bit.btrfs_do_readpage.btrfs_read_folio.filemap_read_folio.do_read_cache_folio
      2.59 ±  4%      -0.1        2.46 ±  3%  perf-profile.calltrace.cycles-pp.folio_batch_move_lru.folio_add_lru.filemap_add_folio.do_read_cache_folio.vfs_dedupe_file_range_compare
      0.13 ±173%      -0.1        0.00        perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.clock_nanosleep
      1.14 ±  2%      -0.1        1.03 ±  3%  perf-profile.calltrace.cycles-pp.alloc_extent_state.__clear_extent_bit.btrfs_invalidate_folio.truncate_cleanup_folio.truncate_inode_pages_range
      0.72 ±  5%      -0.1        0.61 ±  3%  perf-profile.calltrace.cycles-pp.__mod_lruvec_page_state.__filemap_add_folio.filemap_add_folio.do_read_cache_folio.vfs_dedupe_file_range_compare
      0.95 ±  6%      -0.1        0.84 ±  7%  perf-profile.calltrace.cycles-pp.btrfs_get_extent.btrfs_do_readpage.btrfs_read_folio.filemap_read_folio.do_read_cache_folio
      1.19 ±  3%      -0.1        1.09 ±  5%  perf-profile.calltrace.cycles-pp.btrfs_prealloc_file_range.btrfs_fallocate.vfs_fallocate.ioctl_preallocate.__x64_sys_ioctl
      1.19 ±  3%      -0.1        1.09 ±  5%  perf-profile.calltrace.cycles-pp.__btrfs_prealloc_file_range.btrfs_prealloc_file_range.btrfs_fallocate.vfs_fallocate.ioctl_preallocate
      1.52 ±  5%      -0.1        1.42 ±  4%  perf-profile.calltrace.cycles-pp.xas_store.delete_from_page_cache_batch.truncate_inode_pages_range.btrfs_clone_files.btrfs_remap_file_range
      2.27 ±  5%      -0.1        2.17 ±  3%  perf-profile.calltrace.cycles-pp.try_release_extent_mapping.btrfs_invalidate_folio.truncate_cleanup_folio.truncate_inode_pages_range.btrfs_clone_files
     26.65 ±  4%      -0.1       26.56 ±  6%  perf-profile.calltrace.cycles-pp.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call.do_idle.cpu_startup_entry
      1.88 ±  6%      -0.1        1.79 ±  4%  perf-profile.calltrace.cycles-pp.__filemap_get_folio.do_read_cache_folio.vfs_dedupe_file_range_compare.__generic_remap_file_range_prep.generic_remap_file_range_prep
      0.41 ± 57%      -0.1        0.32 ± 81%  perf-profile.calltrace.cycles-pp.xa_get_order.__filemap_add_folio.filemap_add_folio.do_read_cache_folio.vfs_dedupe_file_range_compare
      0.84 ±  3%      -0.1        0.76 ±  3%  perf-profile.calltrace.cycles-pp.kmem_cache_alloc.alloc_extent_state.__clear_extent_bit.btrfs_invalidate_folio.truncate_cleanup_folio
      1.75 ±  5%      -0.1        1.67 ±  3%  perf-profile.calltrace.cycles-pp.__clear_extent_bit.try_release_extent_mapping.btrfs_invalidate_folio.truncate_cleanup_folio.truncate_inode_pages_range
      1.49 ±  5%      -0.1        1.40 ±  4%  perf-profile.calltrace.cycles-pp.lru_add_fn.folio_batch_move_lru.folio_add_lru.filemap_add_folio.do_read_cache_folio
      0.72 ±  5%      -0.1        0.63 ±  6%  perf-profile.calltrace.cycles-pp.clear_page_extent_mapped.btrfs_invalidate_folio.truncate_cleanup_folio.truncate_inode_pages_range.btrfs_clone_files
      0.87 ±  3%      -0.1        0.79 ±  5%  perf-profile.calltrace.cycles-pp.filemap_unaccount_folio.delete_from_page_cache_batch.truncate_inode_pages_range.btrfs_clone_files.btrfs_remap_file_range
      0.98 ±  5%      -0.1        0.90 ±  6%  perf-profile.calltrace.cycles-pp.alloc_extent_state.__set_extent_bit.lock_extent.btrfs_lock_and_flush_ordered_range.btrfs_read_folio
      0.65 ±  6%      -0.1        0.57 ±  7%  perf-profile.calltrace.cycles-pp.__mod_lruvec_page_state.filemap_unaccount_folio.delete_from_page_cache_batch.truncate_inode_pages_range.btrfs_clone_files
      1.60 ±  7%      -0.1        1.53 ±  4%  perf-profile.calltrace.cycles-pp.filemap_get_entry.__filemap_get_folio.do_read_cache_folio.vfs_dedupe_file_range_compare.__generic_remap_file_range_prep
      0.07 ±264%      -0.1        0.00        perf-profile.calltrace.cycles-pp.end_page_read.btrfs_do_readpage.btrfs_read_folio.filemap_read_folio.do_read_cache_folio
      1.06 ±  2%      -0.1        1.00 ±  6%  perf-profile.calltrace.cycles-pp.alloc_extent_state.__set_extent_bit.lock_extent.btrfs_invalidate_folio.truncate_cleanup_folio
      0.06 ±264%      -0.1        0.00        perf-profile.calltrace.cycles-pp.__wake_up.clear_state_bit.__clear_extent_bit.btrfs_invalidate_folio.truncate_cleanup_folio
      1.62 ±  3%      -0.1        1.56 ±  5%  perf-profile.calltrace.cycles-pp.clear_state_bit.__clear_extent_bit.btrfs_do_readpage.btrfs_read_folio.filemap_read_folio
      0.06 ±264%      -0.1        0.00        perf-profile.calltrace.cycles-pp.__wake_up.clear_state_bit.__clear_extent_bit.btrfs_do_readpage.btrfs_read_folio
      0.06 ±264%      -0.1        0.00        perf-profile.calltrace.cycles-pp.folio_unlock.vfs_dedupe_file_range_compare.__generic_remap_file_range_prep.generic_remap_file_range_prep.btrfs_remap_file_range
      0.68 ± 16%      -0.1        0.62 ±  8%  perf-profile.calltrace.cycles-pp.xas_load.filemap_get_entry.__filemap_get_folio.do_read_cache_folio.vfs_dedupe_file_range_compare
      0.67 ±  8%      -0.1        0.62 ±  7%  perf-profile.calltrace.cycles-pp.btrfs_replace_file_extents.btrfs_clone.btrfs_extent_same_range.btrfs_remap_file_range.vfs_dedupe_file_range_one
      0.79 ±  3%      -0.1        0.74 ±  6%  perf-profile.calltrace.cycles-pp.alloc_extent_state.__clear_extent_bit.try_release_extent_mapping.btrfs_invalidate_folio.truncate_cleanup_folio
      0.72 ±  7%      -0.0        0.67 ±  6%  perf-profile.calltrace.cycles-pp.kmem_cache_alloc.alloc_extent_state.__set_extent_bit.lock_extent.btrfs_lock_and_flush_ordered_range
      0.78 ±  4%      -0.0        0.74 ±  5%  perf-profile.calltrace.cycles-pp.kmem_cache_alloc.alloc_extent_state.__set_extent_bit.lock_extent.btrfs_invalidate_folio
      0.74 ±  5%      -0.0        0.70 ±  6%  perf-profile.calltrace.cycles-pp.syscall_return_via_sysret.ioctl.stress_run
      0.65 ±  4%      -0.0        0.61 ±  2%  perf-profile.calltrace.cycles-pp.btrfs_lookup_ordered_range.btrfs_lock_and_flush_ordered_range.btrfs_read_folio.filemap_read_folio.do_read_cache_folio
      0.67 ±  4%      -0.0        0.63 ±  7%  perf-profile.calltrace.cycles-pp.syscall_exit_to_user_mode.do_syscall_64.entry_SYSCALL_64_after_hwframe.ioctl.stress_run
      0.20 ±129%      -0.0        0.16 ±152%  perf-profile.calltrace.cycles-pp.btrfs_insert_replace_extent.btrfs_replace_file_extents.btrfs_clone.btrfs_clone_files.btrfs_remap_file_range
     41.84 ±  2%      -0.0       41.80 ±  2%  perf-profile.calltrace.cycles-pp.generic_remap_file_range_prep.btrfs_remap_file_range.vfs_dedupe_file_range_one.vfs_dedupe_file_range.do_vfs_ioctl
     41.84 ±  2%      -0.0       41.80 ±  2%  perf-profile.calltrace.cycles-pp.__generic_remap_file_range_prep.generic_remap_file_range_prep.btrfs_remap_file_range.vfs_dedupe_file_range_one.vfs_dedupe_file_range
     41.78 ±  2%      -0.0       41.75 ±  2%  perf-profile.calltrace.cycles-pp.vfs_dedupe_file_range_compare.__generic_remap_file_range_prep.generic_remap_file_range_prep.btrfs_remap_file_range.vfs_dedupe_file_range_one
      0.93 ±  4%      -0.0        0.90 ±  5%  perf-profile.calltrace.cycles-pp.find_lock_entries.truncate_inode_pages_range.btrfs_clone_files.btrfs_remap_file_range.do_clone_file_range
      1.23 ±  4%      -0.0        1.20 ±  2%  perf-profile.calltrace.cycles-pp.free_unref_page_list.release_pages.__folio_batch_release.truncate_inode_pages_range.btrfs_clone_files
      1.12 ±  3%      -0.0        1.10 ±  4%  perf-profile.calltrace.cycles-pp.xas_store.__filemap_add_folio.filemap_add_folio.do_read_cache_folio.vfs_dedupe_file_range_compare
      1.26 ±  4%      -0.0        1.23 ±  4%  perf-profile.calltrace.cycles-pp.btrfs_replace_file_extents.btrfs_clone.btrfs_clone_files.btrfs_remap_file_range.do_clone_file_range
      0.90 ±  3%      -0.0        0.90 ±  7%  perf-profile.calltrace.cycles-pp.alloc_extent_state.__clear_extent_bit.btrfs_do_readpage.btrfs_read_folio.filemap_read_folio
      1.27 ±  2%      -0.0        1.27 ±  6%  perf-profile.calltrace.cycles-pp.__mem_cgroup_charge.__filemap_add_folio.filemap_add_folio.do_read_cache_folio.vfs_dedupe_file_range_compare
      0.65 ±  6%      +0.0        0.65 ±  5%  perf-profile.calltrace.cycles-pp.btrfs_punch_hole.btrfs_fallocate.vfs_fallocate.ioctl_preallocate.__x64_sys_ioctl
      1.12 ±  6%      +0.0        1.14 ± 10%  perf-profile.calltrace.cycles-pp.__hrtimer_run_queues.hrtimer_interrupt.__sysvec_apic_timer_interrupt.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt
      0.66 ±  3%      +0.0        0.67 ±  6%  perf-profile.calltrace.cycles-pp.kmem_cache_alloc.alloc_extent_state.__clear_extent_bit.btrfs_do_readpage.btrfs_read_folio
      1.39 ±  7%      +0.0        1.42 ± 10%  perf-profile.calltrace.cycles-pp.hrtimer_interrupt.__sysvec_apic_timer_interrupt.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.cpuidle_enter_state
      0.13 ±173%      +0.0        0.16 ±152%  perf-profile.calltrace.cycles-pp.perf_adjust_freq_unthr_context.perf_event_task_tick.scheduler_tick.update_process_times.tick_sched_handle
      1.40 ±  7%      +0.0        1.44 ± 10%  perf-profile.calltrace.cycles-pp.__sysvec_apic_timer_interrupt.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.cpuidle_enter_state.cpuidle_enter
      0.60 ±  7%      +0.0        0.64 ±  9%  perf-profile.calltrace.cycles-pp.scheduler_tick.update_process_times.tick_sched_handle.tick_sched_timer.__hrtimer_run_queues
      0.76 ±  4%      +0.0        0.80 ±  9%  perf-profile.calltrace.cycles-pp.tick_sched_timer.__hrtimer_run_queues.hrtimer_interrupt.__sysvec_apic_timer_interrupt.sysvec_apic_timer_interrupt
     42.89 ±  2%      +0.0       42.94 ±  2%  perf-profile.calltrace.cycles-pp.btrfs_remap_file_range.vfs_dedupe_file_range_one.vfs_dedupe_file_range.do_vfs_ioctl.__x64_sys_ioctl
      0.68 ±  5%      +0.0        0.73 ±  9%  perf-profile.calltrace.cycles-pp.update_process_times.tick_sched_handle.tick_sched_timer.__hrtimer_run_queues.hrtimer_interrupt
      0.69 ±  5%      +0.1        0.74 ±  9%  perf-profile.calltrace.cycles-pp.tick_sched_handle.tick_sched_timer.__hrtimer_run_queues.hrtimer_interrupt.__sysvec_apic_timer_interrupt
      2.02 ±  7%      +0.1        2.07 ±  9%  perf-profile.calltrace.cycles-pp.asm_sysvec_apic_timer_interrupt.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call.do_idle
      0.00            +0.1        0.05 ±300%  perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_irqsave.rmqueue_bulk.__rmqueue_pcplist.rmqueue
      0.00            +0.1        0.05 ±300%  perf-profile.calltrace.cycles-pp.mem_cgroup_commit_charge.__mem_cgroup_charge.__filemap_add_folio.filemap_add_folio.do_read_cache_folio
     42.93 ±  2%      +0.1       42.98 ±  2%  perf-profile.calltrace.cycles-pp.vfs_dedupe_file_range.do_vfs_ioctl.__x64_sys_ioctl.do_syscall_64.entry_SYSCALL_64_after_hwframe
     42.92 ±  2%      +0.1       42.97 ±  2%  perf-profile.calltrace.cycles-pp.vfs_dedupe_file_range_one.vfs_dedupe_file_range.do_vfs_ioctl.__x64_sys_ioctl.do_syscall_64
      0.49 ± 38%      +0.1        0.55 ±  5%  perf-profile.calltrace.cycles-pp.btrfs_drop_extents.btrfs_replace_file_extents.btrfs_clone.btrfs_clone_files.btrfs_remap_file_range
      1.85 ±  9%      +0.1        1.91 ±  9%  perf-profile.calltrace.cycles-pp.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call
      0.13 ±173%      +0.1        0.21 ±122%  perf-profile.calltrace.cycles-pp.perf_event_task_tick.scheduler_tick.update_process_times.tick_sched_handle.tick_sched_timer
      1.00 ±  7%      +0.1        1.08 ±  6%  perf-profile.calltrace.cycles-pp.btrfs_extent_same_range.btrfs_remap_file_range.vfs_dedupe_file_range_one.vfs_dedupe_file_range.do_vfs_ioctl
      0.92 ±  7%      +0.1        1.01 ±  6%  perf-profile.calltrace.cycles-pp.btrfs_clone.btrfs_extent_same_range.btrfs_remap_file_range.vfs_dedupe_file_range_one.vfs_dedupe_file_range
      0.06 ±264%      +0.1        0.16 ±152%  perf-profile.calltrace.cycles-pp.__mem_cgroup_uncharge_list.release_pages.__folio_batch_release.truncate_inode_pages_range.btrfs_clone_files
      0.00            +0.1        0.10 ±200%  perf-profile.calltrace.cycles-pp._raw_spin_lock_irqsave.rmqueue_bulk.__rmqueue_pcplist.rmqueue.get_page_from_freelist
      0.34 ±195%      +0.2        0.58 ±114%  perf-profile.calltrace.cycles-pp.cpuidle_enter.cpuidle_idle_call.do_idle.cpu_startup_entry.rest_init
      0.34 ±195%      +0.3        0.59 ±114%  perf-profile.calltrace.cycles-pp.cpuidle_idle_call.do_idle.cpu_startup_entry.rest_init.arch_call_rest_init
      0.34 ±195%      +0.3        0.60 ±115%  perf-profile.calltrace.cycles-pp.cpu_startup_entry.rest_init.arch_call_rest_init.start_kernel.x86_64_start_reservations
      0.34 ±195%      +0.3        0.60 ±115%  perf-profile.calltrace.cycles-pp.do_idle.cpu_startup_entry.rest_init.arch_call_rest_init.start_kernel
      0.34 ±195%      +0.3        0.60 ±115%  perf-profile.calltrace.cycles-pp.x86_64_start_kernel.secondary_startup_64_no_verify
      0.34 ±195%      +0.3        0.60 ±115%  perf-profile.calltrace.cycles-pp.x86_64_start_reservations.x86_64_start_kernel.secondary_startup_64_no_verify
      0.34 ±195%      +0.3        0.60 ±115%  perf-profile.calltrace.cycles-pp.start_kernel.x86_64_start_reservations.x86_64_start_kernel.secondary_startup_64_no_verify
      0.34 ±195%      +0.3        0.60 ±115%  perf-profile.calltrace.cycles-pp.arch_call_rest_init.start_kernel.x86_64_start_reservations.x86_64_start_kernel.secondary_startup_64_no_verify
      0.34 ±195%      +0.3        0.60 ±115%  perf-profile.calltrace.cycles-pp.rest_init.arch_call_rest_init.start_kernel.x86_64_start_reservations.x86_64_start_kernel
     70.99            +0.3       71.33 ±  2%  perf-profile.calltrace.cycles-pp.stress_run
     70.59            +0.4       70.96 ±  2%  perf-profile.calltrace.cycles-pp.ioctl.stress_run
     69.39            +0.5       69.86 ±  2%  perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.ioctl.stress_run
     68.99            +0.5       69.46 ±  2%  perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.ioctl.stress_run
     68.25            +0.5       68.76 ±  2%  perf-profile.calltrace.cycles-pp.__x64_sys_ioctl.do_syscall_64.entry_SYSCALL_64_after_hwframe.ioctl.stress_run
     22.27            +0.6       22.89 ±  2%  perf-profile.calltrace.cycles-pp.btrfs_remap_file_range.do_clone_file_range.vfs_clone_file_range.ioctl_file_clone.do_vfs_ioctl
     22.35            +0.6       22.97 ±  2%  perf-profile.calltrace.cycles-pp.ioctl_file_clone.do_vfs_ioctl.__x64_sys_ioctl.do_syscall_64.entry_SYSCALL_64_after_hwframe
     22.32            +0.6       22.94 ±  2%  perf-profile.calltrace.cycles-pp.do_clone_file_range.vfs_clone_file_range.ioctl_file_clone.do_vfs_ioctl.__x64_sys_ioctl
     21.97            +0.6       22.59 ±  2%  perf-profile.calltrace.cycles-pp.btrfs_clone_files.btrfs_remap_file_range.do_clone_file_range.vfs_clone_file_range.ioctl_file_clone
     22.34            +0.6       22.96 ±  2%  perf-profile.calltrace.cycles-pp.vfs_clone_file_range.ioctl_file_clone.do_vfs_ioctl.__x64_sys_ioctl.do_syscall_64
     29.94 ±  2%      +0.6       30.58 ±  3%  perf-profile.calltrace.cycles-pp.do_read_cache_folio.vfs_dedupe_file_range_compare.__generic_remap_file_range_prep.generic_remap_file_range_prep.btrfs_remap_file_range
     65.60            +0.7       66.27 ±  2%  perf-profile.calltrace.cycles-pp.do_vfs_ioctl.__x64_sys_ioctl.do_syscall_64.entry_SYSCALL_64_after_hwframe.ioctl
      0.00            +0.8        0.85 ±  8%  perf-profile.calltrace.cycles-pp.__list_del_entry_valid_or_report.rmqueue_bulk.__rmqueue_pcplist.rmqueue.get_page_from_freelist
      0.00            +1.5        1.51 ±  6%  perf-profile.calltrace.cycles-pp.__free_one_page.free_pcppages_bulk.free_unref_page_commit.free_unref_page.btrfs_clone
      0.89 ±  7%      +1.9        2.80 ±  5%  perf-profile.calltrace.cycles-pp.rmqueue.get_page_from_freelist.__alloc_pages.folio_alloc.do_read_cache_folio
      2.57 ±  3%      +2.0        4.54 ±  5%  perf-profile.calltrace.cycles-pp.folio_alloc.do_read_cache_folio.vfs_dedupe_file_range_compare.__generic_remap_file_range_prep.generic_remap_file_range_prep
      1.19 ±  6%      +2.0        3.17 ±  5%  perf-profile.calltrace.cycles-pp.get_page_from_freelist.__alloc_pages.folio_alloc.do_read_cache_folio.vfs_dedupe_file_range_compare
      1.91 ±  3%      +2.0        3.90 ±  4%  perf-profile.calltrace.cycles-pp.__alloc_pages.folio_alloc.do_read_cache_folio.vfs_dedupe_file_range_compare.__generic_remap_file_range_prep
      0.00            +2.0        1.99 ±  5%  perf-profile.calltrace.cycles-pp.rmqueue_bulk.__rmqueue_pcplist.rmqueue.get_page_from_freelist.__alloc_pages
      1.67 ±  4%      +2.1        3.76 ±  3%  perf-profile.calltrace.cycles-pp.btrfs_clone.btrfs_clone_files.btrfs_remap_file_range.do_clone_file_range.vfs_clone_file_range
      0.00            +2.1        2.09 ±  4%  perf-profile.calltrace.cycles-pp.free_pcppages_bulk.free_unref_page_commit.free_unref_page.btrfs_clone.btrfs_clone_files
      0.00            +2.1        2.12 ±  4%  perf-profile.calltrace.cycles-pp.free_unref_page_commit.free_unref_page.btrfs_clone.btrfs_clone_files.btrfs_remap_file_range
      0.00            +2.1        2.14 ±  4%  perf-profile.calltrace.cycles-pp.free_unref_page.btrfs_clone.btrfs_clone_files.btrfs_remap_file_range.do_clone_file_range
      0.00            +2.3        2.29 ±  6%  perf-profile.calltrace.cycles-pp.__rmqueue_pcplist.rmqueue.get_page_from_freelist.__alloc_pages.folio_alloc
     20.14            -1.5       18.68        perf-profile.children.cycles-pp.truncate_inode_pages_range
     11.79 ±  2%      -0.9       10.85        perf-profile.children.cycles-pp.truncate_cleanup_folio
     11.57 ±  2%      -0.9       10.64        perf-profile.children.cycles-pp.btrfs_invalidate_folio
     16.02 ±  2%      -0.8       15.18 ±  3%  perf-profile.children.cycles-pp.filemap_read_folio
     15.69 ±  2%      -0.8       14.89 ±  3%  perf-profile.children.cycles-pp.btrfs_read_folio
     10.00 ±  2%      -0.7        9.34 ±  2%  perf-profile.children.cycles-pp.__clear_extent_bit
      9.99 ±  2%      -0.6        9.43 ±  2%  perf-profile.children.cycles-pp.memcmp
     12.08 ±  2%      -0.5       11.53 ±  3%  perf-profile.children.cycles-pp.btrfs_do_readpage
     27.15 ±  4%      -0.4       26.77 ±  6%  perf-profile.children.cycles-pp.start_secondary
      5.00 ±  2%      -0.4        4.63 ±  3%  perf-profile.children.cycles-pp.lock_extent
      4.82 ±  2%      -0.4        4.46 ±  3%  perf-profile.children.cycles-pp.__set_extent_bit
      5.19 ±  2%      -0.3        4.84 ±  3%  perf-profile.children.cycles-pp.alloc_extent_state
      8.65 ±  2%      -0.3        8.32 ±  2%  perf-profile.children.cycles-pp.filemap_add_folio
      3.27 ±  3%      -0.3        3.01 ±  3%  perf-profile.children.cycles-pp.delete_from_page_cache_batch
      4.23 ±  2%      -0.3        3.97 ±  3%  perf-profile.children.cycles-pp.kmem_cache_alloc
      3.45 ±  4%      -0.3        3.20 ±  3%  perf-profile.children.cycles-pp.btrfs_lock_and_flush_ordered_range
      3.40 ±  2%      -0.2        3.15 ±  4%  perf-profile.children.cycles-pp.clear_state_bit
     27.67 ±  4%      -0.2       27.44 ±  6%  perf-profile.children.cycles-pp.secondary_startup_64_no_verify
     27.67 ±  4%      -0.2       27.44 ±  6%  perf-profile.children.cycles-pp.cpu_startup_entry
     27.66 ±  4%      -0.2       27.43 ±  6%  perf-profile.children.cycles-pp.do_idle
     23.81 ±  4%      -0.2       23.59 ±  5%  perf-profile.children.cycles-pp.intel_idle_ibrs
      3.97 ±  2%      -0.2        3.76 ±  3%  perf-profile.children.cycles-pp.release_pages
      5.54 ±  2%      -0.2        5.33 ±  2%  perf-profile.children.cycles-pp.__filemap_add_folio
     27.38 ±  4%      -0.2       27.18 ±  6%  perf-profile.children.cycles-pp.cpuidle_idle_call
      3.78 ±  2%      -0.2        3.58 ±  3%  perf-profile.children.cycles-pp.__folio_batch_release
     26.97 ±  4%      -0.2       26.78 ±  6%  perf-profile.children.cycles-pp.cpuidle_enter
     26.97 ±  4%      -0.2       26.78 ±  6%  perf-profile.children.cycles-pp.cpuidle_enter_state
      1.42 ±  3%      -0.2        1.24 ±  3%  perf-profile.children.cycles-pp.__mod_lruvec_page_state
      2.50 ±  4%      -0.2        2.33 ±  4%  perf-profile.children.cycles-pp.ioctl_preallocate
      0.86 ±  6%      -0.2        0.69 ±  7%  perf-profile.children.cycles-pp.__mod_memcg_lruvec_state
      2.46 ±  4%      -0.2        2.30 ±  3%  perf-profile.children.cycles-pp.vfs_fallocate
      2.38 ±  3%      -0.2        2.21 ±  4%  perf-profile.children.cycles-pp.btrfs_fallocate
      5.31            -0.1        5.16 ±  2%  perf-profile.children.cycles-pp.memset_orig
      2.88 ±  3%      -0.1        2.74 ±  3%  perf-profile.children.cycles-pp.folio_add_lru
      2.67 ±  4%      -0.1        2.53 ±  3%  perf-profile.children.cycles-pp.folio_batch_move_lru
      1.26 ±  2%      -0.1        1.13 ±  3%  perf-profile.children.cycles-pp.free_extent_state
      2.70 ±  4%      -0.1        2.58 ±  3%  perf-profile.children.cycles-pp.xas_store
      1.10 ±  5%      -0.1        0.99 ±  7%  perf-profile.children.cycles-pp.btrfs_get_extent
      1.19 ±  3%      -0.1        1.09 ±  5%  perf-profile.children.cycles-pp.btrfs_prealloc_file_range
      1.19 ±  3%      -0.1        1.09 ±  5%  perf-profile.children.cycles-pp.__btrfs_prealloc_file_range
      2.74 ±  2%      -0.1        2.64 ±  3%  perf-profile.children.cycles-pp.btrfs_replace_file_extents
      0.96 ±  3%      -0.1        0.87 ±  5%  perf-profile.children.cycles-pp.__wake_up
      2.28 ±  5%      -0.1        2.18 ±  2%  perf-profile.children.cycles-pp.try_release_extent_mapping
      1.93 ±  5%      -0.1        1.83 ±  3%  perf-profile.children.cycles-pp.__filemap_get_folio
      1.56 ±  5%      -0.1        1.46 ±  3%  perf-profile.children.cycles-pp.lru_add_fn
      0.62 ±  9%      -0.1        0.53 ± 17%  perf-profile.children.cycles-pp.clock_nanosleep
      1.33 ±  9%      -0.1        1.24 ±  4%  perf-profile.children.cycles-pp.xas_load
      0.72 ±  5%      -0.1        0.63 ±  6%  perf-profile.children.cycles-pp.clear_page_extent_mapped
      0.66 ±  9%      -0.1        0.58 ± 13%  perf-profile.children.cycles-pp.set_page_extent_mapped
      0.61 ±  4%      -0.1        0.53 ±  7%  perf-profile.children.cycles-pp.find_free_extent
      0.62 ±  4%      -0.1        0.54 ±  7%  perf-profile.children.cycles-pp.btrfs_reserve_extent
      0.88 ±  3%      -0.1        0.81 ±  4%  perf-profile.children.cycles-pp.filemap_unaccount_folio
      0.63 ±  4%      -0.1        0.55 ±  7%  perf-profile.children.cycles-pp.__cond_resched
      1.63 ±  7%      -0.1        1.56 ±  4%  perf-profile.children.cycles-pp.filemap_get_entry
      0.46 ±  8%      -0.1        0.39 ±  8%  perf-profile.children.cycles-pp.end_page_read
      1.59 ±  4%      -0.1        1.52 ±  5%  perf-profile.children.cycles-pp.kmem_cache_free
      0.89 ±  4%      -0.1        0.83 ±  5%  perf-profile.children.cycles-pp.syscall_exit_to_user_mode
      0.97 ±  3%      -0.1        0.92 ±  3%  perf-profile.children.cycles-pp.syscall_return_via_sysret
      0.64 ±  9%      -0.1        0.58 ±  7%  perf-profile.children.cycles-pp.cache_state_if_flags
      0.47 ±  6%      -0.1        0.41 ±  8%  perf-profile.children.cycles-pp.__entry_text_start
      1.01 ±  3%      -0.1        0.96 ±  6%  perf-profile.children.cycles-pp.btrfs_insert_replace_extent
      0.81 ±  4%      -0.1        0.76 ±  5%  perf-profile.children.cycles-pp.folio_unlock
      0.33 ±  4%      -0.0        0.28 ± 12%  perf-profile.children.cycles-pp.try_charge_memcg
      0.27 ± 11%      -0.0        0.22 ±  9%  perf-profile.children.cycles-pp._raw_read_lock
      0.33 ± 11%      -0.0        0.28 ± 15%  perf-profile.children.cycles-pp.__x64_sys_clock_nanosleep
      0.68 ±  4%      -0.0        0.63 ±  2%  perf-profile.children.cycles-pp.btrfs_lookup_ordered_range
      0.31 ± 17%      -0.0        0.26 ± 16%  perf-profile.children.cycles-pp.cgroup_rstat_updated
      0.30 ± 13%      -0.0        0.26 ± 14%  perf-profile.children.cycles-pp.common_nsleep
     41.92 ±  2%      -0.0       41.88 ±  2%  perf-profile.children.cycles-pp.__generic_remap_file_range_prep
      2.01 ±  3%      -0.0        1.96 ±  4%  perf-profile.children.cycles-pp._raw_spin_lock
      0.43 ± 10%      -0.0        0.38 ± 13%  perf-profile.children.cycles-pp.__schedule
     41.96 ±  2%      -0.0       41.92 ±  2%  perf-profile.children.cycles-pp.generic_remap_file_range_prep
      0.42 ±  6%      -0.0        0.38 ±  4%  perf-profile.children.cycles-pp.memcg_slab_post_alloc_hook
      0.81 ±  7%      -0.0        0.77 ±  6%  perf-profile.children.cycles-pp.__mod_lruvec_state
      0.30 ± 14%      -0.0        0.26 ± 14%  perf-profile.children.cycles-pp.hrtimer_nanosleep
      0.63 ±  7%      -0.0        0.59 ± 18%  perf-profile.children.cycles-pp.intel_idle
      0.29 ±  5%      -0.0        0.25 ±  9%  perf-profile.children.cycles-pp.btrfs_update_inode
      0.59 ±  7%      -0.0        0.55 ±  7%  perf-profile.children.cycles-pp.folio_lruvec_lock_irqsave
     41.82 ±  2%      -0.0       41.78 ±  2%  perf-profile.children.cycles-pp.vfs_dedupe_file_range_compare
      0.22 ± 12%      -0.0        0.18 ± 11%  perf-profile.children.cycles-pp.btrfs_page_clear_checked
      0.71 ±  8%      -0.0        0.68 ±  3%  perf-profile.children.cycles-pp._raw_spin_lock_irq
      0.28 ±  9%      -0.0        0.25 ± 11%  perf-profile.children.cycles-pp.schedule
      0.79 ± 12%      -0.0        0.76 ±  7%  perf-profile.children.cycles-pp.xas_descend
      0.47 ±  3%      -0.0        0.44 ±  6%  perf-profile.children.cycles-pp.insert_prealloc_file_extent
      0.35 ± 11%      -0.0        0.32 ± 10%  perf-profile.children.cycles-pp.__init_waitqueue_head
      0.34 ± 10%      -0.0        0.31 ±  6%  perf-profile.children.cycles-pp.btrfs_lookup_first_ordered_range
      0.53 ±  6%      -0.0        0.50 ± 12%  perf-profile.children.cycles-pp.alloc_pages
      0.28 ± 13%      -0.0        0.24 ± 16%  perf-profile.children.cycles-pp.do_nanosleep
      0.31 ± 11%      -0.0        0.28 ±  8%  perf-profile.children.cycles-pp.xas_clear_mark
      0.12 ± 11%      -0.0        0.09 ± 41%  perf-profile.children.cycles-pp.activate_task
      0.23 ±  8%      -0.0        0.20 ± 10%  perf-profile.children.cycles-pp.filemap_free_folio
      0.96 ±  4%      -0.0        0.93 ±  5%  perf-profile.children.cycles-pp.find_lock_entries
      0.28 ± 11%      -0.0        0.25 ± 12%  perf-profile.children.cycles-pp.__lookup_extent_mapping
      0.03 ±102%      -0.0        0.00        perf-profile.children.cycles-pp.btrfs_free_path
      0.28 ± 11%      -0.0        0.25 ± 11%  perf-profile.children.cycles-pp.btrfs_page_set_uptodate
      0.28 ± 11%      -0.0        0.25 ± 11%  perf-profile.children.cycles-pp.btrfs_wait_ordered_range
      0.22 ±  9%      -0.0        0.19 ± 10%  perf-profile.children.cycles-pp.rcu_all_qs
      0.31 ± 10%      -0.0        0.28 ± 10%  perf-profile.children.cycles-pp.btrfs_del_items
      0.16 ±  7%      -0.0        0.13 ± 14%  perf-profile.children.cycles-pp.__tree_search
      0.37 ±  6%      -0.0        0.34 ± 12%  perf-profile.children.cycles-pp.setup_items_for_insert
      0.22 ±  9%      -0.0        0.19 ±  8%  perf-profile.children.cycles-pp.btrfs_delayed_update_inode
      0.15 ± 11%      -0.0        0.12 ± 15%  perf-profile.children.cycles-pp.read_tsc
      0.25 ± 12%      -0.0        0.23 ±  6%  perf-profile.children.cycles-pp._raw_spin_trylock
      0.19 ±  8%      -0.0        0.16 ±  8%  perf-profile.children.cycles-pp.clone_finish_inode_update
      0.20 ±  9%      -0.0        0.18 ± 16%  perf-profile.children.cycles-pp.try_to_wake_up
      0.11 ± 12%      -0.0        0.09 ± 41%  perf-profile.children.cycles-pp.enqueue_task_fair
      0.11 ± 13%      -0.0        0.09 ± 22%  perf-profile.children.cycles-pp.update_load_avg
      0.06 ± 13%      -0.0        0.04 ± 66%  perf-profile.children.cycles-pp.__do_sys_newfstat
      0.04 ± 58%      -0.0        0.02 ±153%  perf-profile.children.cycles-pp.xas_nomem
      0.10 ± 11%      -0.0        0.08 ± 41%  perf-profile.children.cycles-pp.enqueue_entity
      0.44 ±  6%      -0.0        0.42 ±  2%  perf-profile.children.cycles-pp.free_unref_page_prepare
      0.11 ± 13%      -0.0        0.09 ± 18%  perf-profile.children.cycles-pp.btrfs_put_block_group
      0.10 ± 18%      -0.0        0.08 ± 18%  perf-profile.children.cycles-pp.entry_SYSCALL_64_safe_stack
      0.26 ±  3%      -0.0        0.23 ±  7%  perf-profile.children.cycles-pp.xas_find
      0.48 ±  7%      -0.0        0.45 ±  6%  perf-profile.children.cycles-pp.xas_find_conflict
      0.14 ± 15%      -0.0        0.12 ± 28%  perf-profile.children.cycles-pp.hrtimer_wakeup
      0.14 ± 15%      -0.0        0.12 ± 18%  perf-profile.children.cycles-pp.vfs_fileattr_set
      0.54 ±  6%      -0.0        0.52 ±  6%  perf-profile.children.cycles-pp.xa_get_order
      0.32 ±  7%      -0.0        0.30 ±  9%  perf-profile.children.cycles-pp.btrfs_is_subpage
      0.02 ±100%      -0.0        0.01 ±300%  perf-profile.children.cycles-pp.__update_load_avg_cfs_rq
      0.26 ±  7%      -0.0        0.24 ±  8%  perf-profile.children.cycles-pp.get_pfnblock_flags_mask
      0.06 ± 63%      -0.0        0.04 ±111%  perf-profile.children.cycles-pp.cleanup_ref_head
      0.15 ± 12%      -0.0        0.13 ± 12%  perf-profile.children.cycles-pp.page_counter_uncharge
      0.13 ± 11%      -0.0        0.11 ± 28%  perf-profile.children.cycles-pp.ttwu_do_activate
      0.37 ± 11%      -0.0        0.35 ±  8%  perf-profile.children.cycles-pp.start_transaction
      0.62 ±  8%      -0.0        0.60 ±  8%  perf-profile.children.cycles-pp.__mod_node_page_state
      0.52 ±  6%      -0.0        0.50 ±  8%  perf-profile.children.cycles-pp.btrfs_insert_empty_items
      0.14 ± 15%      -0.0        0.12 ± 17%  perf-profile.children.cycles-pp.intel_idle_irq
      0.06 ± 18%      -0.0        0.04 ± 66%  perf-profile.children.cycles-pp.wake_up_q
      0.51 ±  6%      -0.0        0.49 ±  7%  perf-profile.children.cycles-pp.folio_mark_accessed
      0.08 ± 30%      -0.0        0.07 ± 74%  perf-profile.children.cycles-pp.process_one_work
      0.07 ± 64%      -0.0        0.05 ± 96%  perf-profile.children.cycles-pp.btrfs_preempt_reclaim_metadata_space
      0.07 ± 64%      -0.0        0.05 ± 96%  perf-profile.children.cycles-pp.flush_space
      0.07 ± 64%      -0.0        0.05 ± 96%  perf-profile.children.cycles-pp.btrfs_run_delayed_refs
      0.07 ± 64%      -0.0        0.05 ± 96%  perf-profile.children.cycles-pp.__btrfs_run_delayed_refs
      0.03 ±101%      -0.0        0.01 ±206%  perf-profile.children.cycles-pp.tick_sched_do_timer
      1.24 ±  4%      -0.0        1.22 ±  3%  perf-profile.children.cycles-pp.free_unref_page_list
      0.36 ±  9%      -0.0        0.35 ±  7%  perf-profile.children.cycles-pp.set_state_bits
      0.15 ±  8%      -0.0        0.13 ± 11%  perf-profile.children.cycles-pp.__fxstat64
      0.05 ± 38%      -0.0        0.04 ± 65%  perf-profile.children.cycles-pp.mutex_unlock
      0.36 ±  7%      -0.0        0.34 ±  7%  perf-profile.children.cycles-pp.should_failslab
      0.10 ± 16%      -0.0        0.08 ± 10%  perf-profile.children.cycles-pp.memmove_extent_buffer
      0.39 ± 10%      -0.0        0.37 ±  3%  perf-profile.children.cycles-pp.test_range_bit
      0.06 ± 15%      -0.0        0.05 ± 53%  perf-profile.children.cycles-pp.btrfs_get_64
      0.12 ± 15%      -0.0        0.10 ±  9%  perf-profile.children.cycles-pp.file_modified
      0.12 ± 14%      -0.0        0.10 ± 16%  perf-profile.children.cycles-pp.btrfs_get_block_group
      1.20 ±  5%      -0.0        1.18 ±  5%  perf-profile.children.cycles-pp.btrfs_drop_extents
      0.19 ± 11%      -0.0        0.17 ±  9%  perf-profile.children.cycles-pp.btrfs_read_lock_root_node
      0.03 ±100%      -0.0        0.01 ±201%  perf-profile.children.cycles-pp.free_extent_buffer
      1.52 ±  6%      -0.0        1.51 ±  7%  perf-profile.children.cycles-pp.__hrtimer_run_queues
      0.12 ± 13%      -0.0        0.10 ± 13%  perf-profile.children.cycles-pp.dequeue_task_fair
      0.13 ± 11%      -0.0        0.12 ± 13%  perf-profile.children.cycles-pp._raw_spin_unlock_irqrestore
      0.11 ± 18%      -0.0        0.10 ± 15%  perf-profile.children.cycles-pp.__btrfs_tree_lock
      0.02 ±131%      -0.0        0.01 ±201%  perf-profile.children.cycles-pp.btrfs_lookup_first_ordered_extent
      0.09 ± 22%      -0.0        0.08 ± 62%  perf-profile.children.cycles-pp.worker_thread
      0.18 ±  9%      -0.0        0.16 ±  8%  perf-profile.children.cycles-pp.__write_extent_buffer
      0.07 ± 17%      -0.0        0.06 ± 37%  perf-profile.children.cycles-pp.btrfs_duplicate_item
      0.26 ± 10%      -0.0        0.24 ±  6%  perf-profile.children.cycles-pp.add_delayed_ref_head
      0.18 ± 18%      -0.0        0.17 ± 14%  perf-profile.children.cycles-pp.llseek
      0.14 ± 21%      -0.0        0.12 ± 22%  perf-profile.children.cycles-pp.tick_nohz_get_sleep_length
      0.07 ± 14%      -0.0        0.05 ± 35%  perf-profile.children.cycles-pp.memcpy_extent_buffer
      0.07 ± 21%      -0.0        0.06 ± 51%  perf-profile.children.cycles-pp.hrtimer_start_range_ns
      0.56 ±  5%      -0.0        0.55 ±  5%  perf-profile.children.cycles-pp.btrfs_search_slot
      0.14 ± 13%      -0.0        0.12 ± 11%  perf-profile.children.cycles-pp.__btrfs_qgroup_release_data
      0.02 ±129%      -0.0        0.01 ±300%  perf-profile.children.cycles-pp.policy_nodemask
      0.02 ±129%      -0.0        0.01 ±300%  perf-profile.children.cycles-pp.set_next_entity
      0.12 ±  9%      -0.0        0.11 ± 14%  perf-profile.children.cycles-pp.__btrfs_release_delayed_node
      0.01 ±173%      -0.0        0.00        perf-profile.children.cycles-pp.sched_ttwu_pending
      0.01 ±173%      -0.0        0.00        perf-profile.children.cycles-pp.inode_needs_update_time
      0.02 ±130%      -0.0        0.01 ±200%  perf-profile.children.cycles-pp.set_extent_buffer_dirty
      0.52 ±  3%      -0.0        0.50 ±  6%  perf-profile.children.cycles-pp.__mem_cgroup_uncharge_list
      0.17 ± 16%      -0.0        0.16 ± 22%  perf-profile.children.cycles-pp.perf_mux_hrtimer_handler
      0.09 ± 10%      -0.0        0.08 ± 20%  perf-profile.children.cycles-pp.btrfs_set_token_32
      0.05 ± 40%      -0.0        0.04 ± 67%  perf-profile.children.cycles-pp.work_busy
      0.18 ± 13%      -0.0        0.17 ± 16%  perf-profile.children.cycles-pp.btrfs_drop_extent_map_range
      0.14 ± 15%      -0.0        0.13 ± 13%  perf-profile.children.cycles-pp.filemap_alloc_folio
      0.19 ±  9%      -0.0        0.18 ± 13%  perf-profile.children.cycles-pp.btrfs_set_delalloc_extent
      0.11 ± 17%      -0.0        0.10 ± 20%  perf-profile.children.cycles-pp.btrfs_fileattr_set
      0.12 ± 18%      -0.0        0.11 ± 11%  perf-profile.children.cycles-pp.should_fail_alloc_page
      0.11 ± 13%      -0.0        0.10 ± 20%  perf-profile.children.cycles-pp.down_read
      0.22 ±  5%      -0.0        0.21 ±  7%  perf-profile.children.cycles-pp.uncharge_batch
      0.12 ± 18%      -0.0        0.11 ± 14%  perf-profile.children.cycles-pp.btrfs_lock_root_node
      0.02 ±129%      -0.0        0.01 ±200%  perf-profile.children.cycles-pp.__hrtimer_next_event_base
      0.20 ±  8%      -0.0        0.19 ± 14%  perf-profile.children.cycles-pp.policy_node
      0.14 ± 10%      -0.0        0.13 ± 11%  perf-profile.children.cycles-pp.btrfs_alloc_reserved_file_extent
      0.14 ±  9%      -0.0        0.13 ±  8%  perf-profile.children.cycles-pp.kmem_cache_alloc_lru
      0.07 ± 16%      -0.0        0.06 ± 18%  perf-profile.children.cycles-pp.schedule_preempt_disabled
      0.06 ± 13%      -0.0        0.05 ± 36%  perf-profile.children.cycles-pp.update_rq_clock_task
      0.08 ± 13%      -0.0        0.08 ± 17%  perf-profile.children.cycles-pp.rwsem_down_read_slowpath
      0.08 ± 13%      -0.0        0.07 ± 53%  perf-profile.children.cycles-pp.perf_rotate_context
      0.02 ±129%      -0.0        0.01 ±200%  perf-profile.children.cycles-pp.read
      0.11 ± 11%      -0.0        0.10 ± 19%  perf-profile.children.cycles-pp.__btrfs_tree_read_lock
      0.08 ± 14%      -0.0        0.07 ± 16%  perf-profile.children.cycles-pp.__filemap_fdatawrite_range
      0.29 ±  9%      -0.0        0.28 ±  9%  perf-profile.children.cycles-pp.mem_cgroup_update_lru_size
      0.40 ±  9%      -0.0        0.39 ±  7%  perf-profile.children.cycles-pp.xas_start
      0.16 ± 16%      -0.0        0.15 ± 18%  perf-profile.children.cycles-pp.schedule_idle
      0.08 ± 27%      -0.0        0.07 ± 42%  perf-profile.children.cycles-pp.tick_nohz_next_event
      0.08 ± 19%      -0.0        0.07 ± 42%  perf-profile.children.cycles-pp.sched_clock
      0.10 ± 18%      -0.0        0.10 ± 12%  perf-profile.children.cycles-pp.dequeue_entity
      0.06 ± 17%      -0.0        0.05 ± 55%  perf-profile.children.cycles-pp.update_sg_lb_stats
      0.07 ± 23%      -0.0        0.06 ± 18%  perf-profile.children.cycles-pp.btrfs_fdatawrite_range
      0.07 ± 16%      -0.0        0.06 ± 40%  perf-profile.children.cycles-pp.rwsem_wake
      0.01 ±173%      -0.0        0.01 ±299%  perf-profile.children.cycles-pp.__switch_to_asm
      0.01 ±264%      -0.0        0.00        perf-profile.children.cycles-pp.update_rq_clock
      0.01 ±264%      -0.0        0.00        perf-profile.children.cycles-pp.rep_movs_alternative
      0.11 ± 20%      -0.0        0.10 ± 17%  perf-profile.children.cycles-pp.rebalance_domains
      0.11 ± 23%      -0.0        0.10 ± 24%  perf-profile.children.cycles-pp.pick_next_task_fair
      0.23 ±  9%      -0.0        0.22 ±  6%  perf-profile.children.cycles-pp.btrfs_lookup_file_extent
      0.12 ± 20%      -0.0        0.12 ± 38%  perf-profile.children.cycles-pp.ret_from_fork_asm
      0.06 ± 39%      -0.0        0.05 ± 54%  perf-profile.children.cycles-pp.task_tick_fair
      0.40 ±  5%      -0.0        0.40 ±  6%  perf-profile.children.cycles-pp.wait_subpage_spinlock
      0.12 ± 21%      -0.0        0.12 ± 38%  perf-profile.children.cycles-pp.ret_from_fork
      0.03 ±100%      -0.0        0.02 ±123%  perf-profile.children.cycles-pp.hrtimer_next_event_without
      0.12 ± 20%      -0.0        0.12 ± 40%  perf-profile.children.cycles-pp.kthread
      0.07 ± 19%      -0.0        0.07 ± 22%  perf-profile.children.cycles-pp.entry_SYSRETQ_unsafe_stack
      0.07 ± 19%      -0.0        0.06 ± 12%  perf-profile.children.cycles-pp.slab_pre_alloc_hook
      0.01 ±264%      -0.0        0.00        perf-profile.children.cycles-pp.btrfs_del_csums
      0.01 ±264%      -0.0        0.00        perf-profile.children.cycles-pp.ksys_read
      0.01 ±264%      -0.0        0.00        perf-profile.children.cycles-pp.btrfs_get_8
      0.01 ±264%      -0.0        0.00        perf-profile.children.cycles-pp.__switch_to
      0.01 ±264%      -0.0        0.00        perf-profile.children.cycles-pp.hrtimer_update_next_event
      0.01 ±264%      -0.0        0.00        perf-profile.children.cycles-pp.assert_eb_page_uptodate
      0.01 ±264%      -0.0        0.00        perf-profile.children.cycles-pp.__slab_free
      0.18 ± 14%      -0.0        0.17 ± 28%  perf-profile.children.cycles-pp.btrfs_clear_delalloc_extent
      0.15 ±  8%      -0.0        0.14 ±  6%  perf-profile.children.cycles-pp.xas_alloc
      0.12 ±  5%      -0.0        0.12 ± 10%  perf-profile.children.cycles-pp.__filemap_fdatawait_range
      0.08 ± 19%      -0.0        0.08 ± 22%  perf-profile.children.cycles-pp.find_busiest_group
      0.10 ±  7%      -0.0        0.09 ± 20%  perf-profile.children.cycles-pp._find_first_bit
      0.04 ± 78%      -0.0        0.03 ±100%  perf-profile.children.cycles-pp.read_cache_folio
      0.09 ± 14%      -0.0        0.09 ± 16%  perf-profile.children.cycles-pp.perf_pmu_nop_void
      0.03 ±100%      -0.0        0.02 ±123%  perf-profile.children.cycles-pp.__kmem_cache_alloc_node
      0.03 ± 78%      -0.0        0.03 ±100%  perf-profile.children.cycles-pp.read_extent_buffer
      0.09 ± 16%      -0.0        0.08 ± 16%  perf-profile.children.cycles-pp.rwsem_optimistic_spin
      0.04 ± 58%      -0.0        0.04 ± 82%  perf-profile.children.cycles-pp.xas_find_marked
      0.05 ± 60%      -0.0        0.04 ± 50%  perf-profile.children.cycles-pp.btrfs_update_root_times
      0.11 ± 13%      -0.0        0.11 ± 15%  perf-profile.children.cycles-pp.filemap_fdatawait_range
      0.17 ±  7%      -0.0        0.16 ± 13%  perf-profile.children.cycles-pp.__intel_pmu_enable_all
      0.21 ±  7%      -0.0        0.20 ±  6%  perf-profile.children.cycles-pp.rb_next
      0.08 ± 20%      -0.0        0.08 ± 25%  perf-profile.children.cycles-pp.update_sd_lb_stats
      1.85 ±  7%      -0.0        1.85 ±  7%  perf-profile.children.cycles-pp.__sysvec_apic_timer_interrupt
      0.09 ± 14%      -0.0        0.08 ± 14%  perf-profile.children.cycles-pp.__memmove
      0.07 ± 12%      -0.0        0.06 ± 17%  perf-profile.children.cycles-pp.btrfs_alloc_block_rsv
      0.06 ± 20%      -0.0        0.06 ± 10%  perf-profile.children.cycles-pp.lru_add_drain_cpu
      0.08 ± 17%      -0.0        0.08 ± 19%  perf-profile.children.cycles-pp.leaf_space_used
      0.08 ± 13%      -0.0        0.08 ± 16%  perf-profile.children.cycles-pp.filemap_get_folios_tag
      0.02 ±129%      -0.0        0.02 ±153%  perf-profile.children.cycles-pp.__flush_smp_call_function_queue
      0.09 ± 15%      -0.0        0.08 ± 13%  perf-profile.children.cycles-pp.rcu_do_batch
      0.01 ±174%      -0.0        0.01 ±200%  perf-profile.children.cycles-pp.mutex_lock
      0.04 ± 79%      -0.0        0.04 ± 67%  perf-profile.children.cycles-pp._copy_from_user
      0.10 ± 28%      -0.0        0.10 ± 11%  perf-profile.children.cycles-pp.memcg_check_events
      0.02 ±129%      -0.0        0.02 ±153%  perf-profile.children.cycles-pp.filemap_write_and_wait_range
      0.16 ± 18%      -0.0        0.16 ± 15%  perf-profile.children.cycles-pp.btrfs_release_path
      0.09 ± 16%      -0.0        0.08 ± 13%  perf-profile.children.cycles-pp.rwsem_down_write_slowpath
      1.29 ±  2%      -0.0        1.29 ±  6%  perf-profile.children.cycles-pp.__mem_cgroup_charge
      1.84 ±  7%      -0.0        1.83 ±  7%  perf-profile.children.cycles-pp.hrtimer_interrupt
      0.13 ± 10%      -0.0        0.13 ± 13%  perf-profile.children.cycles-pp.btrfs_get_token_32
      0.08 ± 20%      -0.0        0.08 ± 40%  perf-profile.children.cycles-pp.sched_clock_cpu
      0.11 ± 19%      -0.0        0.11 ± 21%  perf-profile.children.cycles-pp.load_balance
      0.27 ±  6%      -0.0        0.27 ±  8%  perf-profile.children.cycles-pp.uncharge_folio
      0.08 ± 16%      -0.0        0.08 ± 21%  perf-profile.children.cycles-pp.page_counter_try_charge
      0.01 ±173%      -0.0        0.01 ±201%  perf-profile.children.cycles-pp.enqueue_hrtimer
      0.08 ± 17%      -0.0        0.08 ±  7%  perf-profile.children.cycles-pp.btrfs_root_node
      0.01 ±264%      -0.0        0.01 ±299%  perf-profile.children.cycles-pp.folio_wait_writeback
      0.01 ±264%      -0.0        0.01 ±300%  perf-profile.children.cycles-pp.btrfs_free_block_rsv
      0.12 ± 19%      -0.0        0.12 ± 14%  perf-profile.children.cycles-pp.up_write
      0.29 ± 16%      -0.0        0.29 ± 14%  perf-profile.children.cycles-pp.menu_select
      0.61 ±  8%      -0.0        0.61 ±  9%  perf-profile.children.cycles-pp.perf_event_task_tick
      0.06 ± 40%      -0.0        0.06 ± 38%  perf-profile.children.cycles-pp.btrfs_leaf_free_space
      0.08 ± 20%      -0.0        0.08 ± 15%  perf-profile.children.cycles-pp.apparmor_file_permission
      0.01 ±264%      -0.0        0.01 ±300%  perf-profile.children.cycles-pp.prepare_task_switch
      0.04 ± 58%      +0.0        0.04 ± 68%  perf-profile.children.cycles-pp.btrfs_dirty_inode
      0.02 ±173%      +0.0        0.02 ±152%  perf-profile.children.cycles-pp.kmalloc_trace
      0.09 ± 15%      +0.0        0.09 ± 18%  perf-profile.children.cycles-pp.find_first_non_hole
      0.15 ± 10%      +0.0        0.15 ±  9%  perf-profile.children.cycles-pp.down_write
      0.01 ±264%      +0.0        0.01 ±300%  perf-profile.children.cycles-pp.timekeeping_max_deferment
      0.04 ± 79%      +0.0        0.04 ± 65%  perf-profile.children.cycles-pp.join_transaction
      0.46 ±  5%      +0.0        0.47 ±  5%  perf-profile.children.cycles-pp.workingset_update_node
      0.09 ± 20%      +0.0        0.09 ± 16%  perf-profile.children.cycles-pp.security_file_permission
      0.19 ± 10%      +0.0        0.19 ± 12%  perf-profile.children.cycles-pp.__reserve_bytes
      0.09 ± 22%      +0.0        0.09 ± 16%  perf-profile.children.cycles-pp.set_extent_bit
      0.05 ± 61%      +0.0        0.05 ± 51%  perf-profile.children.cycles-pp.__libc_start_main
      0.05 ± 61%      +0.0        0.05 ± 51%  perf-profile.children.cycles-pp.main
      0.05 ± 61%      +0.0        0.05 ± 51%  perf-profile.children.cycles-pp.run_builtin
      0.11 ± 17%      +0.0        0.11 ± 11%  perf-profile.children.cycles-pp.rcu_core
      0.65 ±  5%      +0.0        0.65 ±  5%  perf-profile.children.cycles-pp.btrfs_punch_hole
      0.08 ± 15%      +0.0        0.09 ±  8%  perf-profile.children.cycles-pp.btrfs_double_extent_lock
      0.18 ± 12%      +0.0        0.19 ± 12%  perf-profile.children.cycles-pp.btrfs_reserve_metadata_bytes
      0.09 ± 21%      +0.0        0.09 ± 15%  perf-profile.children.cycles-pp.native_irq_return_iret
      0.10 ± 15%      +0.0        0.11 ±  5%  perf-profile.children.cycles-pp.__wake_up_common
      0.10 ± 14%      +0.0        0.11 ± 13%  perf-profile.children.cycles-pp.btrfs_get_32
      0.02 ±129%      +0.0        0.02 ±123%  perf-profile.children.cycles-pp.vfs_fstat
      0.01 ±174%      +0.0        0.02 ±153%  perf-profile.children.cycles-pp.exit_to_user_mode_prepare
      0.01 ±174%      +0.0        0.02 ±153%  perf-profile.children.cycles-pp.newidle_balance
      0.10 ± 10%      +0.0        0.11 ± 13%  perf-profile.children.cycles-pp.rb_erase
      0.20 ± 11%      +0.0        0.20 ±  8%  perf-profile.children.cycles-pp.clockevents_program_event
      0.09 ± 20%      +0.0        0.09 ± 23%  perf-profile.children.cycles-pp.btrfs_bin_search
      0.01 ±173%      +0.0        0.02 ±153%  perf-profile.children.cycles-pp.update_irq_load_avg
      0.01 ±173%      +0.0        0.02 ±153%  perf-profile.children.cycles-pp.security_capable
      0.01 ±173%      +0.0        0.02 ±153%  perf-profile.children.cycles-pp.asm_exc_page_fault
      0.09 ± 19%      +0.0        0.09 ± 17%  perf-profile.children.cycles-pp.xas_expand
      0.01 ±264%      +0.0        0.01 ±200%  perf-profile.children.cycles-pp.stress_file_ioctl
      0.37 ± 15%      +0.0        0.38 ± 10%  perf-profile.children.cycles-pp.__irq_exit_rcu
      0.08 ± 13%      +0.0        0.08 ± 19%  perf-profile.children.cycles-pp.btrfs_punch_hole_lock_range
      0.09 ± 24%      +0.0        0.10 ± 15%  perf-profile.children.cycles-pp.__fget_light
      0.06 ± 40%      +0.0        0.06 ± 57%  perf-profile.children.cycles-pp.ktime_get_update_offsets_now
      0.60 ±  8%      +0.0        0.60 ±  9%  perf-profile.children.cycles-pp.perf_adjust_freq_unthr_context
      0.00            +0.0        0.01 ±299%  perf-profile.children.cycles-pp.finish_task_switch
      0.00            +0.0        0.01 ±299%  perf-profile.children.cycles-pp.add_extent_mapping
      0.00            +0.0        0.01 ±299%  perf-profile.children.cycles-pp.btrfs_trans_release_metadata
      0.00            +0.0        0.01 ±299%  perf-profile.children.cycles-pp.folio_mapping
      0.00            +0.0        0.01 ±299%  perf-profile.children.cycles-pp.vfs_write
      0.00            +0.0        0.01 ±299%  perf-profile.children.cycles-pp.ksys_write
      0.00            +0.0        0.01 ±300%  perf-profile.children.cycles-pp.cpuidle_governor_latency_req
      0.00            +0.0        0.01 ±300%  perf-profile.children.cycles-pp.execve
      0.00            +0.0        0.01 ±300%  perf-profile.children.cycles-pp.__x64_sys_execve
      0.00            +0.0        0.01 ±300%  perf-profile.children.cycles-pp.do_execveat_common
      0.00            +0.0        0.01 ±300%  perf-profile.children.cycles-pp.cmd_record
      0.00            +0.0        0.01 ±300%  perf-profile.children.cycles-pp.__cmd_record
      0.27 ± 30%      +0.0        0.28 ± 39%  perf-profile.children.cycles-pp.ktime_get
      0.02 ±129%      +0.0        0.03 ±100%  perf-profile.children.cycles-pp.capable
      0.01 ±174%      +0.0        0.02 ±123%  perf-profile.children.cycles-pp.update_existing_head_ref
      0.09 ± 20%      +0.0        0.09 ± 24%  perf-profile.children.cycles-pp.native_sched_clock
      0.15 ± 16%      +0.0        0.16 ± 22%  perf-profile.children.cycles-pp.__count_memcg_events
      0.07 ± 22%      +0.0        0.08 ± 17%  perf-profile.children.cycles-pp.btrfs_block_rsv_release
      0.01 ±264%      +0.0        0.02 ±152%  perf-profile.children.cycles-pp.local_clock_noinstr
      0.28 ±  7%      +0.0        0.29 ±  6%  perf-profile.children.cycles-pp.xas_create
      0.00            +0.0        0.01 ±200%  perf-profile.children.cycles-pp.btrfs_inode_lock
      0.00            +0.0        0.01 ±200%  perf-profile.children.cycles-pp.need_preemptive_reclaim
      0.51 ±  7%      +0.0        0.52 ±  2%  perf-profile.children.cycles-pp.btrfs_add_delayed_data_ref
      0.00            +0.0        0.01 ±201%  perf-profile.children.cycles-pp.tick_nohz_irq_exit
      0.02 ±129%      +0.0        0.03 ±102%  perf-profile.children.cycles-pp.poll_idle
      0.07 ± 17%      +0.0        0.09 ± 20%  perf-profile.children.cycles-pp.__btrfs_end_transaction
      0.23 ±  9%      +0.0        0.24 ± 13%  perf-profile.children.cycles-pp.free_extent_map
      0.06 ± 44%      +0.0        0.07 ± 18%  perf-profile.children.cycles-pp.update_blocked_averages
      0.06 ± 45%      +0.0        0.08 ± 42%  perf-profile.children.cycles-pp.arch_scale_freq_tick
      0.07 ± 15%      +0.0        0.09 ± 21%  perf-profile.children.cycles-pp.irqtime_account_irq
      0.04 ± 60%      +0.0        0.06 ± 35%  perf-profile.children.cycles-pp.rb_insert_color
      0.13 ± 21%      +0.0        0.15 ±  7%  perf-profile.children.cycles-pp.lapic_next_deadline
      0.01 ±173%      +0.0        0.03 ±104%  perf-profile.children.cycles-pp.btrfs_find_space_for_alloc
      0.02 ±131%      +0.0        0.04 ± 66%  perf-profile.children.cycles-pp.btrfs_reduce_alloc_profile
      0.06 ± 43%      +0.0        0.08 ± 18%  perf-profile.children.cycles-pp.run_rebalance_domains
      0.05 ± 65%      +0.0        0.06 ± 12%  perf-profile.children.cycles-pp.insert_delayed_ref
      1.07 ±  5%      +0.0        1.09 ±  7%  perf-profile.children.cycles-pp.tick_sched_timer
      0.38 ± 14%      +0.0        0.40 ±  9%  perf-profile.children.cycles-pp.get_mem_cgroup_from_mm
      0.84 ±  5%      +0.0        0.86 ±  7%  perf-profile.children.cycles-pp.scheduler_tick
      0.01 ±264%      +0.0        0.03 ±100%  perf-profile.children.cycles-pp.syscall_enter_from_user_mode
      0.00            +0.0        0.02 ±122%  perf-profile.children.cycles-pp.___slab_alloc
      2.45 ±  9%      +0.0        2.47 ±  7%  perf-profile.children.cycles-pp.sysvec_apic_timer_interrupt
      0.02 ±131%      +0.0        0.04 ± 51%  perf-profile.children.cycles-pp.calc_available_free_space
      0.01 ±174%      +0.0        0.04 ± 66%  perf-profile.children.cycles-pp.update_curr
      0.02 ±134%      +0.0        0.05 ± 52%  perf-profile.children.cycles-pp.btrfs_can_overcommit
      2.69 ±  7%      +0.0        2.71 ±  7%  perf-profile.children.cycles-pp.asm_sysvec_apic_timer_interrupt
      0.98 ±  5%      +0.0        1.01 ±  7%  perf-profile.children.cycles-pp.tick_sched_handle
      0.42 ±  7%      +0.0        0.45 ±  9%  perf-profile.children.cycles-pp.mem_cgroup_commit_charge
      0.13 ± 74%      +0.0        0.16 ± 73%  perf-profile.children.cycles-pp.irq_enter_rcu
      0.04 ± 80%      +0.0        0.07 ± 18%  perf-profile.children.cycles-pp.rcu_sched_clock_irq
      0.96 ±  5%      +0.0        1.00 ±  7%  perf-profile.children.cycles-pp.update_process_times
      0.01 ±264%      +0.0        0.04 ± 67%  perf-profile.children.cycles-pp.rcu_pending
      0.11 ± 86%      +0.0        0.15 ± 77%  perf-profile.children.cycles-pp.tick_irq_enter
     42.93 ±  2%      +0.1       42.98 ±  2%  perf-profile.children.cycles-pp.vfs_dedupe_file_range
     42.92 ±  2%      +0.1       42.97 ±  2%  perf-profile.children.cycles-pp.vfs_dedupe_file_range_one
      1.00 ±  7%      +0.1        1.08 ±  5%  perf-profile.children.cycles-pp.btrfs_extent_same_range
      0.29 ± 16%      +0.1        0.39 ±  6%  perf-profile.children.cycles-pp.__mod_zone_page_state
      0.52 ±113%      +0.2        0.67 ± 96%  perf-profile.children.cycles-pp.x86_64_start_kernel
      0.52 ±113%      +0.2        0.67 ± 96%  perf-profile.children.cycles-pp.x86_64_start_reservations
      0.52 ±113%      +0.2        0.67 ± 96%  perf-profile.children.cycles-pp.start_kernel
      0.52 ±113%      +0.2        0.67 ± 96%  perf-profile.children.cycles-pp.arch_call_rest_init
      0.52 ±113%      +0.2        0.67 ± 96%  perf-profile.children.cycles-pp.rest_init
      0.26 ±  8%      +0.2        0.41 ±  7%  perf-profile.children.cycles-pp.__list_add_valid_or_report
      0.08 ± 21%      +0.2        0.25 ±  8%  perf-profile.children.cycles-pp.__kmalloc_node
      0.08 ± 22%      +0.2        0.24 ±  8%  perf-profile.children.cycles-pp.__kmalloc_large_node
      0.08 ± 21%      +0.2        0.25 ±  6%  perf-profile.children.cycles-pp.kvmalloc_node
     70.99            +0.3       71.33 ±  2%  perf-profile.children.cycles-pp.stress_run
     70.65            +0.4       71.02 ±  2%  perf-profile.children.cycles-pp.ioctl
     70.37            +0.4       70.77 ±  2%  perf-profile.children.cycles-pp.entry_SYSCALL_64_after_hwframe
     69.83            +0.4       70.25 ±  2%  perf-profile.children.cycles-pp.do_syscall_64
     68.25            +0.5       68.76 ±  2%  perf-profile.children.cycles-pp.__x64_sys_ioctl
      1.13 ±  3%      +0.6        1.69 ±  5%  perf-profile.children.cycles-pp._raw_spin_lock_irqsave
      0.31 ± 11%      +0.6        0.90 ±  6%  perf-profile.children.cycles-pp.native_queued_spin_lock_slowpath
     22.35            +0.6       22.97 ±  2%  perf-profile.children.cycles-pp.ioctl_file_clone
     21.97            +0.6       22.59 ±  2%  perf-profile.children.cycles-pp.btrfs_clone_files
     22.34            +0.6       22.96 ±  2%  perf-profile.children.cycles-pp.vfs_clone_file_range
     22.32            +0.6       22.94 ±  2%  perf-profile.children.cycles-pp.do_clone_file_range
     30.09 ±  2%      +0.6       30.73 ±  3%  perf-profile.children.cycles-pp.do_read_cache_folio
     65.16            +0.7       65.82 ±  2%  perf-profile.children.cycles-pp.btrfs_remap_file_range
     65.61            +0.7       66.28 ±  2%  perf-profile.children.cycles-pp.do_vfs_ioctl
      0.52 ±  4%      +1.0        1.48 ±  6%  perf-profile.children.cycles-pp.__list_del_entry_valid_or_report
      0.00            +1.5        1.54 ±  6%  perf-profile.children.cycles-pp.__free_one_page
      2.58 ±  3%      +2.0        4.55 ±  5%  perf-profile.children.cycles-pp.folio_alloc
      0.93 ±  6%      +2.1        3.02 ±  5%  perf-profile.children.cycles-pp.rmqueue
      0.00            +2.1        2.10 ±  4%  perf-profile.children.cycles-pp.free_pcppages_bulk
      0.01 ±264%      +2.2        2.17 ±  4%  perf-profile.children.cycles-pp.free_unref_page
      2.03 ±  3%      +2.2        4.19 ±  4%  perf-profile.children.cycles-pp.__alloc_pages
      1.25 ±  6%      +2.2        3.42 ±  5%  perf-profile.children.cycles-pp.get_page_from_freelist
      0.37 ± 10%      +2.2        2.54 ±  4%  perf-profile.children.cycles-pp.free_unref_page_commit
      2.60 ±  4%      +2.2        4.77 ±  3%  perf-profile.children.cycles-pp.btrfs_clone
      0.00            +2.2        2.20 ±  6%  perf-profile.children.cycles-pp.rmqueue_bulk
      0.00            +2.5        2.49 ±  6%  perf-profile.children.cycles-pp.__rmqueue_pcplist
      9.91 ±  2%      -0.5        9.38 ±  2%  perf-profile.self.cycles-pp.memcmp
     23.80 ±  4%      -0.2       23.59 ±  5%  perf-profile.self.cycles-pp.intel_idle_ibrs
      3.24 ±  2%      -0.2        3.05 ±  3%  perf-profile.self.cycles-pp.kmem_cache_alloc
      1.16 ±  3%      -0.1        1.02 ±  4%  perf-profile.self.cycles-pp.__set_extent_bit
      5.27            -0.1        5.12 ±  2%  perf-profile.self.cycles-pp.memset_orig
      0.63 ±  6%      -0.1        0.50 ±  7%  perf-profile.self.cycles-pp.__mod_memcg_lruvec_state
      1.19 ±  2%      -0.1        1.06 ±  3%  perf-profile.self.cycles-pp.free_extent_state
      1.46 ±  4%      -0.1        1.36 ±  4%  perf-profile.self.cycles-pp.xas_store
      0.92 ±  3%      -0.1        0.84 ±  6%  perf-profile.self.cycles-pp.alloc_extent_state
      1.45 ±  3%      -0.1        1.37 ±  4%  perf-profile.self.cycles-pp.__clear_extent_bit
      0.58 ±  7%      -0.1        0.51 ± 11%  perf-profile.self.cycles-pp.set_page_extent_mapped
      1.15 ±  4%      -0.1        1.08 ±  9%  perf-profile.self.cycles-pp.vfs_dedupe_file_range_compare
      0.62 ±  5%      -0.1        0.55 ±  6%  perf-profile.self.cycles-pp.clear_page_extent_mapped
      0.84 ±  4%      -0.1        0.78 ±  5%  perf-profile.self.cycles-pp.syscall_exit_to_user_mode
      0.88 ±  4%      -0.1        0.82 ±  4%  perf-profile.self.cycles-pp.btrfs_invalidate_folio
      0.94 ±  8%      -0.1        0.88 ±  7%  perf-profile.self.cycles-pp.clear_state_bit
      0.82 ±  7%      -0.1        0.76 ±  4%  perf-profile.self.cycles-pp.lru_add_fn
      0.97 ±  3%      -0.1        0.92 ±  3%  perf-profile.self.cycles-pp.syscall_return_via_sysret
      0.60 ±  9%      -0.1        0.54 ±  7%  perf-profile.self.cycles-pp.cache_state_if_flags
      1.51 ±  4%      -0.1        1.46 ±  5%  perf-profile.self.cycles-pp.kmem_cache_free
      0.38 ±  7%      -0.1        0.33 ±  9%  perf-profile.self.cycles-pp.__cond_resched
      0.42 ±  7%      -0.1        0.37 ±  9%  perf-profile.self.cycles-pp.__entry_text_start
      0.79 ±  4%      -0.1        0.74 ±  5%  perf-profile.self.cycles-pp.folio_unlock
      1.11 ±  4%      -0.0        1.06 ±  4%  perf-profile.self.cycles-pp.release_pages
      0.27 ± 11%      -0.0        0.22 ±  9%  perf-profile.self.cycles-pp._raw_read_lock
      0.41 ±  5%      -0.0        0.36 ±  5%  perf-profile.self.cycles-pp.memcg_slab_post_alloc_hook
      0.25 ±  7%      -0.0        0.20 ± 12%  perf-profile.self.cycles-pp.try_charge_memcg
      0.33 ±  4%      -0.0        0.29 ± 13%  perf-profile.self.cycles-pp.filemap_read_folio
      0.53 ±  2%      -0.0        0.48 ±  7%  perf-profile.self.cycles-pp.delete_from_page_cache_batch
      0.23 ±  8%      -0.0        0.19 ± 12%  perf-profile.self.cycles-pp.xas_find_conflict
      1.92 ±  4%      -0.0        1.88 ±  5%  perf-profile.self.cycles-pp._raw_spin_lock
      0.63 ±  7%      -0.0        0.59 ± 18%  perf-profile.self.cycles-pp.intel_idle
      0.49 ±  7%      -0.0        0.46 ±  9%  perf-profile.self.cycles-pp.xas_load
      0.50 ±  8%      -0.0        0.47 ±  8%  perf-profile.self.cycles-pp.__mod_lruvec_page_state
      0.28 ± 18%      -0.0        0.24 ± 14%  perf-profile.self.cycles-pp.cgroup_rstat_updated
      0.15 ± 16%      -0.0        0.12 ± 22%  perf-profile.self.cycles-pp.end_page_read
      0.21 ± 11%      -0.0        0.18 ± 11%  perf-profile.self.cycles-pp.btrfs_page_clear_checked
      0.53 ±  5%      -0.0        0.50 ±  6%  perf-profile.self.cycles-pp.folio_batch_move_lru
      0.29 ± 11%      -0.0        0.26 ±  8%  perf-profile.self.cycles-pp.xas_clear_mark
      0.40 ±  7%      -0.0        0.37 ±  5%  perf-profile.self.cycles-pp.btrfs_lock_and_flush_ordered_range
      0.34 ±  9%      -0.0        0.31 ±  6%  perf-profile.self.cycles-pp.btrfs_lookup_ordered_range
      0.28 ± 11%      -0.0        0.25 ± 12%  perf-profile.self.cycles-pp.__lookup_extent_mapping
      0.41 ±  7%      -0.0        0.38 ±  8%  perf-profile.self.cycles-pp.free_unref_page_list
      0.42 ±  7%      -0.0        0.39 ±  9%  perf-profile.self.cycles-pp.do_read_cache_folio
      0.20 ± 11%      -0.0        0.17 ±  9%  perf-profile.self.cycles-pp.__mod_lruvec_state
      0.23 ±  8%      -0.0        0.20 ± 10%  perf-profile.self.cycles-pp.filemap_free_folio
      0.24 ± 13%      -0.0        0.21 ±  6%  perf-profile.self.cycles-pp._raw_spin_trylock
      0.67 ±  8%      -0.0        0.64 ±  4%  perf-profile.self.cycles-pp._raw_spin_lock_irq
      0.03 ± 78%      -0.0        0.01 ±300%  perf-profile.self.cycles-pp.cpuidle_idle_call
      0.28 ± 11%      -0.0        0.25 ± 11%  perf-profile.self.cycles-pp.btrfs_page_set_uptodate
      0.17 ±  6%      -0.0        0.14 ± 11%  perf-profile.self.cycles-pp.xas_find
      0.17 ±  5%      -0.0        0.14 ± 15%  perf-profile.self.cycles-pp.rcu_all_qs
      0.29 ± 12%      -0.0        0.27 ± 10%  perf-profile.self.cycles-pp.__init_waitqueue_head
      0.46 ±  8%      -0.0        0.43 ±  8%  perf-profile.self.cycles-pp.folio_mark_accessed
      0.14 ± 12%      -0.0        0.12 ± 17%  perf-profile.self.cycles-pp.read_tsc
      0.12 ± 23%      -0.0        0.09 ± 14%  perf-profile.self.cycles-pp.btrfs_lookup_first_ordered_range
      1.16 ±  5%      -0.0        1.14 ±  4%  perf-profile.self.cycles-pp.__filemap_add_folio
      0.19 ±  6%      -0.0        0.17 ± 12%  perf-profile.self.cycles-pp.find_free_extent
      0.30 ±  6%      -0.0        0.28 ± 10%  perf-profile.self.cycles-pp.__filemap_get_folio
      0.13 ±  9%      -0.0        0.11 ± 14%  perf-profile.self.cycles-pp.__tree_search
      0.32 ±  7%      -0.0        0.30 ±  9%  perf-profile.self.cycles-pp.btrfs_is_subpage
      0.24 ±  7%      -0.0        0.22 ± 11%  perf-profile.self.cycles-pp.__wake_up
      0.03 ±100%      -0.0        0.01 ±200%  perf-profile.self.cycles-pp.__mem_cgroup_uncharge_list
      0.11 ± 12%      -0.0        0.09 ± 18%  perf-profile.self.cycles-pp.btrfs_put_block_group
      0.55 ±  7%      -0.0        0.53 ± 12%  perf-profile.self.cycles-pp.entry_SYSCALL_64_after_hwframe
      0.81 ±  3%      -0.0        0.79 ±  5%  perf-profile.self.cycles-pp._raw_spin_lock_irqsave
      0.02 ±100%      -0.0        0.01 ±300%  perf-profile.self.cycles-pp.__update_load_avg_cfs_rq
      0.14 ± 12%      -0.0        0.12 ±  9%  perf-profile.self.cycles-pp.page_counter_uncharge
      0.22 ±  7%      -0.0        0.20 ±  8%  perf-profile.self.cycles-pp.truncate_cleanup_folio
      0.60 ±  9%      -0.0        0.58 ±  7%  perf-profile.self.cycles-pp.__mod_node_page_state
      0.25 ±  7%      -0.0        0.23 ±  9%  perf-profile.self.cycles-pp.get_pfnblock_flags_mask
      0.70 ± 13%      -0.0        0.69 ±  8%  perf-profile.self.cycles-pp.xas_descend
      0.30 ±  6%      -0.0        0.28 ± 14%  perf-profile.self.cycles-pp.alloc_pages
      0.02 ±129%      -0.0        0.01 ±299%  perf-profile.self.cycles-pp.entry_SYSCALL_64_safe_stack
      0.06 ± 14%      -0.0        0.04 ± 68%  perf-profile.self.cycles-pp.btrfs_get_64
      0.13 ±  9%      -0.0        0.11 ± 16%  perf-profile.self.cycles-pp._raw_spin_unlock_irqrestore
      0.12 ± 14%      -0.0        0.10 ± 16%  perf-profile.self.cycles-pp.btrfs_get_block_group
      0.40 ±  8%      -0.0        0.39 ±  8%  perf-profile.self.cycles-pp.btrfs_get_extent
      0.18 ±  9%      -0.0        0.16 ± 11%  perf-profile.self.cycles-pp.free_unref_page_prepare
      0.02 ±129%      -0.0        0.01 ±299%  perf-profile.self.cycles-pp.set_extent_buffer_dirty
      0.20 ± 10%      -0.0        0.18 ±  5%  perf-profile.self.cycles-pp.add_delayed_ref_head
      0.05 ± 38%      -0.0        0.04 ± 65%  perf-profile.self.cycles-pp.mutex_unlock
      0.09 ±  7%      -0.0        0.08 ± 23%  perf-profile.self.cycles-pp._find_first_bit
      0.02 ±129%      -0.0        0.01 ±300%  perf-profile.self.cycles-pp.tick_sched_do_timer
      0.09 ±  7%      -0.0        0.08 ± 26%  perf-profile.self.cycles-pp.btrfs_set_token_32
      0.22 ±  8%      -0.0        0.21 ± 12%  perf-profile.self.cycles-pp.wait_subpage_spinlock
      0.15 ± 18%      -0.0        0.14 ± 15%  perf-profile.self.cycles-pp.test_range_bit
      0.25 ± 10%      -0.0        0.24 ± 12%  perf-profile.self.cycles-pp.cpuidle_enter_state
      0.18 ± 14%      -0.0        0.17 ± 15%  perf-profile.self.cycles-pp.xa_get_order
      0.05 ± 40%      -0.0        0.04 ± 83%  perf-profile.self.cycles-pp.intel_idle_irq
      0.03 ±100%      -0.0        0.02 ±153%  perf-profile.self.cycles-pp.update_load_avg
      0.78 ±  6%      -0.0        0.76 ±  7%  perf-profile.self.cycles-pp.find_lock_entries
      0.03 ± 77%      -0.0        0.02 ±122%  perf-profile.self.cycles-pp.read_cache_folio
      0.12 ± 12%      -0.0        0.11 ± 11%  perf-profile.self.cycles-pp.__btrfs_qgroup_release_data
      0.06 ± 13%      -0.0        0.05 ± 53%  perf-profile.self.cycles-pp.__write_extent_buffer
      0.03 ±100%      -0.0        0.02 ±155%  perf-profile.self.cycles-pp.hrtimer_interrupt
      0.10 ± 19%      -0.0        0.08 ± 18%  perf-profile.self.cycles-pp.filemap_alloc_folio
      0.02 ±129%      -0.0        0.01 ±200%  perf-profile.self.cycles-pp._copy_from_user
      0.26 ±  6%      -0.0        0.25 ±  8%  perf-profile.self.cycles-pp.lock_extent
      0.19 ± 10%      -0.0        0.18 ± 13%  perf-profile.self.cycles-pp.btrfs_set_delalloc_extent
      0.03 ±100%      -0.0        0.02 ±123%  perf-profile.self.cycles-pp.__schedule
      0.02 ±129%      -0.0        0.01 ±201%  perf-profile.self.cycles-pp.__kmem_cache_alloc_node
      0.03 ±100%      -0.0        0.02 ±154%  perf-profile.self.cycles-pp.__filemap_fdatawait_range
      0.27 ±  3%      -0.0        0.26 ± 14%  perf-profile.self.cycles-pp.folio_add_lru
      0.03 ±102%      -0.0        0.02 ±123%  perf-profile.self.cycles-pp.start_transaction
      0.06 ± 15%      -0.0        0.06 ± 16%  perf-profile.self.cycles-pp.entry_SYSRETQ_unsafe_stack
      0.01 ±173%      -0.0        0.01 ±299%  perf-profile.self.cycles-pp.__switch_to_asm
      0.01 ±264%      -0.0        0.00        perf-profile.self.cycles-pp.mutex_lock
      0.10 ± 15%      -0.0        0.09 ± 13%  perf-profile.self.cycles-pp.rb_next
      0.04 ± 80%      -0.0        0.03 ±103%  perf-profile.self.cycles-pp.__filemap_fdatawrite_range
      0.09 ± 12%      -0.0        0.08 ± 19%  perf-profile.self.cycles-pp.perf_pmu_nop_void
      0.07 ± 16%      -0.0        0.07 ± 39%  perf-profile.self.cycles-pp.page_counter_try_charge
      0.01 ±264%      -0.0        0.00        perf-profile.self.cycles-pp.btrfs_wait_ordered_range
      0.01 ±264%      -0.0        0.00        perf-profile.self.cycles-pp.xas_nomem
      0.01 ±264%      -0.0        0.00        perf-profile.self.cycles-pp.btrfs_insert_replace_extent
      0.01 ±264%      -0.0        0.00        perf-profile.self.cycles-pp.btrfs_get_8
      0.01 ±264%      -0.0        0.00        perf-profile.self.cycles-pp.assert_eb_page_uptodate
      0.01 ±264%      -0.0        0.00        perf-profile.self.cycles-pp.__slab_free
      0.01 ±264%      -0.0        0.00        perf-profile.self.cycles-pp.uncharge_batch
      0.26 ±  9%      -0.0        0.26 ±  9%  perf-profile.self.cycles-pp.mem_cgroup_update_lru_size
      0.05 ± 62%      -0.0        0.05 ± 69%  perf-profile.self.cycles-pp.btrfs_drop_extents
      0.07 ± 17%      -0.0        0.07 ± 15%  perf-profile.self.cycles-pp.setup_items_for_insert
      0.17 ±  7%      -0.0        0.16 ± 13%  perf-profile.self.cycles-pp.__intel_pmu_enable_all
      0.09 ± 14%      -0.0        0.08 ± 14%  perf-profile.self.cycles-pp.__memmove
      0.13 ± 11%      -0.0        0.12 ± 12%  perf-profile.self.cycles-pp.btrfs_get_token_32
      0.57 ± 15%      -0.0        0.57 ± 14%  perf-profile.self.cycles-pp.btrfs_do_readpage
      0.04 ± 58%      -0.0        0.04 ± 82%  perf-profile.self.cycles-pp.xas_find_marked
      0.03 ±101%      -0.0        0.03 ±100%  perf-profile.self.cycles-pp.btrfs_fallocate
      0.08 ± 17%      -0.0        0.08 ± 10%  perf-profile.self.cycles-pp.btrfs_root_node
      0.20 ± 13%      -0.0        0.19 ±  9%  perf-profile.self.cycles-pp.filemap_unaccount_folio
      0.10 ± 12%      -0.0        0.10 ± 13%  perf-profile.self.cycles-pp.btrfs_get_32
      0.04 ± 58%      -0.0        0.04 ± 67%  perf-profile.self.cycles-pp.btrfs_clone
      0.35 ± 10%      -0.0        0.35 ±  6%  perf-profile.self.cycles-pp.xas_start
      0.07 ± 15%      -0.0        0.07 ±  6%  perf-profile.self.cycles-pp.ioctl
      0.07 ± 22%      -0.0        0.07 ± 17%  perf-profile.self.cycles-pp.apparmor_file_permission
      0.07 ± 45%      -0.0        0.07 ± 14%  perf-profile.self.cycles-pp.memcg_check_events
      0.46 ±  5%      -0.0        0.46 ±  5%  perf-profile.self.cycles-pp.workingset_update_node
      0.10 ± 15%      -0.0        0.10 ±  7%  perf-profile.self.cycles-pp.__wake_up_common
      0.08 ± 22%      -0.0        0.08 ± 39%  perf-profile.self.cycles-pp.native_sched_clock
      0.03 ±100%      -0.0        0.03 ±100%  perf-profile.self.cycles-pp.read_extent_buffer
      0.06 ± 13%      +0.0        0.06 ± 15%  perf-profile.self.cycles-pp.should_fail_alloc_page
      0.01 ±264%      +0.0        0.01 ±300%  perf-profile.self.cycles-pp.timekeeping_max_deferment
      0.05 ± 41%      +0.0        0.05 ± 40%  perf-profile.self.cycles-pp.rwsem_optimistic_spin
      0.35 ±  7%      +0.0        0.35 ±  6%  perf-profile.self.cycles-pp.rmqueue
      0.16 ± 17%      +0.0        0.16 ± 32%  perf-profile.self.cycles-pp.btrfs_clear_delalloc_extent
      0.15 ± 21%      +0.0        0.15 ± 10%  perf-profile.self.cycles-pp.set_state_bits
      0.25 ±  5%      +0.0        0.25 ±  8%  perf-profile.self.cycles-pp.uncharge_folio
      0.09 ± 21%      +0.0        0.09 ± 15%  perf-profile.self.cycles-pp.native_irq_return_iret
      0.86 ±  4%      +0.0        0.86 ±  6%  perf-profile.self.cycles-pp.filemap_get_entry
      0.10 ± 11%      +0.0        0.10 ± 14%  perf-profile.self.cycles-pp.rb_erase
      0.21 ±  9%      +0.0        0.21 ±  8%  perf-profile.self.cycles-pp.should_failslab
      0.02 ±129%      +0.0        0.02 ±124%  perf-profile.self.cycles-pp.btrfs_remap_file_range
      0.09 ± 19%      +0.0        0.09 ± 24%  perf-profile.self.cycles-pp.btrfs_bin_search
      0.01 ±173%      +0.0        0.02 ±153%  perf-profile.self.cycles-pp.update_irq_load_avg
      0.01 ±173%      +0.0        0.02 ±153%  perf-profile.self.cycles-pp.folio_lruvec_lock_irqsave
      0.10 ± 17%      +0.0        0.10 ± 15%  perf-profile.self.cycles-pp.policy_node
      0.01 ±264%      +0.0        0.01 ±201%  perf-profile.self.cycles-pp.__generic_remap_file_range_prep
      0.00            +0.0        0.01 ±299%  perf-profile.self.cycles-pp.vfs_fallocate
      0.00            +0.0        0.01 ±299%  perf-profile.self.cycles-pp.syscall_enter_from_user_mode
      0.00            +0.0        0.01 ±299%  perf-profile.self.cycles-pp.policy_nodemask
      0.00            +0.0        0.01 ±300%  perf-profile.self.cycles-pp.try_to_wake_up
      0.00            +0.0        0.01 ±300%  perf-profile.self.cycles-pp.__btrfs_end_transaction
      0.00            +0.0        0.01 ±300%  perf-profile.self.cycles-pp.exit_to_user_mode_prepare
      0.01 ±264%      +0.0        0.01 ±200%  perf-profile.self.cycles-pp.update_sg_lb_stats
      0.09 ± 24%      +0.0        0.09 ± 15%  perf-profile.self.cycles-pp.btrfs_search_slot
      0.09 ± 27%      +0.0        0.10 ± 17%  perf-profile.self.cycles-pp.__fget_light
      0.05 ± 39%      +0.0        0.06 ± 22%  perf-profile.self.cycles-pp.do_syscall_64
      0.04 ± 78%      +0.0        0.04 ± 51%  perf-profile.self.cycles-pp.btrfs_add_delayed_data_ref
      0.07 ± 42%      +0.0        0.08 ± 22%  perf-profile.self.cycles-pp.truncate_inode_pages_range
      0.12 ± 13%      +0.0        0.13 ± 18%  perf-profile.self.cycles-pp.menu_select
      0.03 ±130%      +0.0        0.04 ± 85%  perf-profile.self.cycles-pp.__hrtimer_run_queues
      0.10 ± 20%      +0.0        0.12 ± 23%  perf-profile.self.cycles-pp.__count_memcg_events
      0.64 ±  6%      +0.0        0.65 ±  5%  perf-profile.self.cycles-pp.__alloc_pages
      0.22 ±  5%      +0.0        0.24 ±  8%  perf-profile.self.cycles-pp.filemap_add_folio
      0.00            +0.0        0.01 ±201%  perf-profile.self.cycles-pp.__btrfs_release_delayed_node
      0.01 ±174%      +0.0        0.02 ±100%  perf-profile.self.cycles-pp.btrfs_replace_file_extents
      0.17 ±  9%      +0.0        0.18 ±  8%  perf-profile.self.cycles-pp.__mem_cgroup_charge
      0.11 ± 22%      +0.0        0.12 ± 15%  perf-profile.self.cycles-pp.try_release_extent_mapping
      0.22 ±  9%      +0.0        0.24 ± 13%  perf-profile.self.cycles-pp.free_extent_map
      0.04 ± 60%      +0.0        0.06 ± 35%  perf-profile.self.cycles-pp.rb_insert_color
      0.06 ± 45%      +0.0        0.08 ± 42%  perf-profile.self.cycles-pp.arch_scale_freq_tick
      0.02 ±129%      +0.0        0.03 ±104%  perf-profile.self.cycles-pp.ktime_get_update_offsets_now
      0.10 ± 12%      +0.0        0.12 ±  9%  perf-profile.self.cycles-pp.xas_create
      0.01 ±174%      +0.0        0.03 ±100%  perf-profile.self.cycles-pp.down_write
      0.01 ±264%      +0.0        0.02 ±123%  perf-profile.self.cycles-pp.irqtime_account_irq
      0.01 ±174%      +0.0        0.03 ±102%  perf-profile.self.cycles-pp.poll_idle
      0.14 ± 15%      +0.0        0.15 ± 14%  perf-profile.self.cycles-pp.btrfs_read_folio
      0.13 ± 21%      +0.0        0.15 ±  7%  perf-profile.self.cycles-pp.lapic_next_deadline
      0.37 ± 14%      +0.0        0.39 ± 10%  perf-profile.self.cycles-pp.get_mem_cgroup_from_mm
      0.02 ±132%      +0.0        0.04 ± 51%  perf-profile.self.cycles-pp.up_write
      0.02 ±130%      +0.0        0.04 ± 52%  perf-profile.self.cycles-pp.btrfs_del_items
      0.16 ± 53%      +0.0        0.18 ± 58%  perf-profile.self.cycles-pp.ktime_get
      0.03 ±104%      +0.0        0.05 ± 38%  perf-profile.self.cycles-pp.__x64_sys_ioctl
      0.07 ± 18%      +0.0        0.10 ± 12%  perf-profile.self.cycles-pp.folio_alloc
      0.34 ±  7%      +0.0        0.37 ±  8%  perf-profile.self.cycles-pp.perf_adjust_freq_unthr_context
      0.04 ± 79%      +0.0        0.06 ± 15%  perf-profile.self.cycles-pp.do_vfs_ioctl
      0.20 ± 11%      +0.0        0.23 ± 10%  perf-profile.self.cycles-pp.mem_cgroup_commit_charge
      0.31 ± 12%      +0.1        0.36 ±  6%  perf-profile.self.cycles-pp.free_unref_page_commit
      0.32 ±  9%      +0.1        0.37 ±  9%  perf-profile.self.cycles-pp.get_page_from_freelist
      0.26 ± 19%      +0.1        0.36 ±  5%  perf-profile.self.cycles-pp.__mod_zone_page_state
      0.24 ±  8%      +0.1        0.38 ±  6%  perf-profile.self.cycles-pp.__list_add_valid_or_report
      0.00            +0.2        0.15 ± 10%  perf-profile.self.cycles-pp.__rmqueue_pcplist
      0.00            +0.2        0.24 ±  8%  perf-profile.self.cycles-pp.free_pcppages_bulk
      0.31 ± 11%      +0.6        0.90 ±  6%  perf-profile.self.cycles-pp.native_queued_spin_lock_slowpath
      0.00            +0.6        0.62 ± 13%  perf-profile.self.cycles-pp.rmqueue_bulk
      0.51 ±  4%      +0.9        1.44 ±  6%  perf-profile.self.cycles-pp.__list_del_entry_valid_or_report
      0.00            +1.3        1.32 ±  6%  perf-profile.self.cycles-pp.__free_one_page

xsang@inn:~$

^ permalink raw reply	[flat|nested] 3+ messages in thread

end of thread, other threads:[~2023-11-24  6:53 UTC | newest]

Thread overview: 3+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2023-11-23  5:03 [linus:master] [mm, pcp] 6ccdcb6d3a: stress-ng.judy.ops_per_sec -4.7% regression kernel test robot
2023-11-23  5:40 ` Huang, Ying
2023-11-24  6:53   ` Oliver Sang

This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox