linux-mm.kvack.org archive mirror
 help / color / mirror / Atom feed
* [linux-next:master] [mm/rmap]  bfc2b13b05: stress-ng.daemon.ops_per_sec 14.0% improvement
@ 2026-02-06 10:03 kernel test robot
  0 siblings, 0 replies; only message in thread
From: kernel test robot @ 2026-02-06 10:03 UTC (permalink / raw)
  To: Lorenzo Stoakes
  Cc: oe-lkp, lkp, Andrew Morton, Suren Baghdasaryan, Liam R. Howlett,
	Barry Song, Chris Li, David Hildenbrand, Harry Yoo, Jann Horn,
	Michal Hocko, Mike Rapoport, Pedro Falcato, Rik van Riel,
	Shakeel Butt, Vlastimil Babka, linux-mm, oliver.sang



Hello,

kernel test robot noticed a 14.0% improvement of stress-ng.daemon.ops_per_sec on:


commit: bfc2b13b05a1343bb60a85d840fd8956731866c5 ("mm/rmap: allocate anon_vma_chain objects unlocked when possible")
https://git.kernel.org/cgit/linux/kernel/git/next/linux-next.git master


testcase: stress-ng
config: x86_64-rhel-9.4
compiler: gcc-14
test machine: 256 threads 4 sockets INTEL(R) XEON(R) PLATINUM 8592+ (Emerald Rapids) with 256G memory
parameters:

	nr_threads: 100%
	testtime: 60s
	test: daemon
	cpufreq_governor: performance


In addition to that, the commit also has significant impact on the following tests:

+------------------+----------------------------------------------------------------------------------------------+
| testcase: change | stress-ng: stress-ng.netlink-proc.ops_per_sec 25.7% improvement                              |
| test machine     | 192 threads 2 sockets Intel(R) Xeon(R) 6740E  CPU @ 2.4GHz (Sierra Forest) with 256G memory  |
| test parameters  | cpufreq_governor=performance                                                                 |
|                  | nr_threads=100%                                                                              |
|                  | test=netlink-proc                                                                            |
|                  | testtime=60s                                                                                 |
+------------------+----------------------------------------------------------------------------------------------+


Details are as below:
-------------------------------------------------------------------------------------------------->


The kernel config and materials to reproduce are available at:
https://download.01.org/0day-ci/archive/20260206/202602061747.855f053f-lkp@intel.com

=========================================================================================
compiler/cpufreq_governor/kconfig/nr_threads/rootfs/tbox_group/test/testcase/testtime:
  gcc-14/performance/x86_64-rhel-9.4/100%/debian-13-x86_64-20250902.cgz/lkp-emr-2sp1/daemon/stress-ng/60s

commit: 
  85f03a8631 ("mm/mmap_lock: add vma_is_attached() helper")
  bfc2b13b05 ("mm/rmap: allocate anon_vma_chain objects unlocked when possible")

85f03a86318c4172 bfc2b13b05a1343bb60a85d840f 
---------------- --------------------------- 
         %stddev     %change         %stddev
             \          |                \  
      8054 ± 16%     +54.6%      12455 ± 16%  perf-c2c.DRAM.local
      0.13 ±  8%     +10.4%       0.15 ±  3%  perf-sched.sch_delay.avg.ms.[unknown].[unknown].[unknown].[unknown].[unknown]
      0.13 ±  8%     +10.4%       0.15 ±  3%  perf-sched.total_sch_delay.average.ms
     38.55            +6.0       44.57 ±  3%  mpstat.cpu.all.idle%
      0.47            -0.0        0.42 ±  2%  mpstat.cpu.all.irq%
     59.36            -6.1       53.28 ±  2%  mpstat.cpu.all.sys%
      1.21            +0.1        1.30        mpstat.cpu.all.usr%
   2817245           +14.1%    3214286 ±  2%  stress-ng.daemon.ops
     47149           +14.0%      53750 ±  2%  stress-ng.daemon.ops_per_sec
     49.50 ±  3%     +18.2%      58.50 ±  2%  stress-ng.time.percent_of_cpu_this_job_got
   2759543           +13.2%    3123624 ±  2%  stress-ng.time.voluntary_context_switches
      0.15 ± 46%      -0.1        0.05 ± 41%  turbostat.C1%
      3.58 ±  5%      +9.0%       3.90 ±  3%  turbostat.CPU%c1
    647.37            +1.9%     659.97        turbostat.PkgWatt
     28.80 ±  2%      +7.1%      30.83        turbostat.RAMWatt
     25720            +7.4%      27633 ±  2%  sched_debug.cfs_rq:/.avg_vruntime.avg
     25656            +7.6%      27595 ±  2%  sched_debug.cfs_rq:/.zero_vruntime.avg
      2214 ±  3%     +51.1%       3345 ±  6%  sched_debug.cpu.clock.stddev
      2726 ±  2%     +40.8%       3838 ±  5%  sched_debug.cpu.clock_task.stddev
   1420062           +14.0%    1619139 ±  2%  sched_debug.cpu.curr->pid.max
      0.00 ±  3%     +51.0%       0.00 ±  6%  sched_debug.cpu.next_balance.stddev
    182592            +3.7%     189367        meminfo.KReclaimable
    260530 ±  2%      -8.3%     238883        meminfo.Mapped
  18519139           +18.4%   21919824 ±  3%  meminfo.Memused
    310935 ±  4%      +9.2%     339441 ±  2%  meminfo.PageTables
    182592            +3.7%     189367        meminfo.SReclaimable
   6259412 ±  3%     +52.7%    9561140 ±  6%  meminfo.SUnreclaim
   6442004 ±  3%     +51.4%    9750508 ±  6%  meminfo.Slab
  18519311           +18.4%   21919824 ±  3%  meminfo.max_used_kB
  25317761 ±  4%     +14.8%   29072594 ±  3%  numa-numastat.node0.local_node
  25429773 ±  4%     +14.8%   29188651 ±  3%  numa-numastat.node0.numa_hit
  25578903 ±  4%     +15.2%   29473906 ±  3%  numa-numastat.node1.local_node
  25681455 ±  4%     +15.2%   29588761 ±  3%  numa-numastat.node1.numa_hit
  25472065 ±  4%     +19.2%   30373362 ±  2%  numa-numastat.node2.local_node
  25579152 ±  4%     +19.1%   30461365 ±  2%  numa-numastat.node2.numa_hit
  25502240 ±  4%     +18.2%   30147059 ±  2%  numa-numastat.node3.local_node
  25588119 ±  4%     +18.2%   30236863 ±  2%  numa-numastat.node3.numa_hit
    208008            +2.0%     212227        proc-vmstat.nr_anon_pages
   6090914            -1.4%    6006008        proc-vmstat.nr_dirty_background_threshold
  12196722            -1.4%   12026702        proc-vmstat.nr_dirty_threshold
  61229635            -1.4%   60379327        proc-vmstat.nr_free_pages
  61094372            -1.4%   60257849        proc-vmstat.nr_free_pages_blocks
     65170 ±  2%      -8.3%      59741        proc-vmstat.nr_mapped
     77887 ±  4%      +9.1%      85010 ±  3%  proc-vmstat.nr_page_table_pages
     45655            +3.7%      47350        proc-vmstat.nr_slab_reclaimable
   1565161 ±  3%     +52.8%    2390881 ±  6%  proc-vmstat.nr_slab_unreclaimable
 1.023e+08 ±  4%     +16.8%  1.195e+08 ±  2%  proc-vmstat.numa_hit
 1.019e+08 ±  4%     +16.9%  1.191e+08 ±  2%  proc-vmstat.numa_local
 1.109e+08 ±  3%     +17.3%  1.301e+08 ±  2%  proc-vmstat.pgalloc_normal
 1.748e+08 ±  3%     +15.3%  2.015e+08 ±  4%  proc-vmstat.pgfault
 1.064e+08 ±  3%     +16.2%  1.236e+08 ±  2%  proc-vmstat.pgfree
   4897546 ± 28%     +25.4%    6140878 ± 25%  numa-meminfo.node0.MemUsed
     74758 ±  7%     +15.2%      86146 ±  7%  numa-meminfo.node0.PageTables
   1585565 ±  3%     +49.9%    2376337 ±  6%  numa-meminfo.node0.SUnreclaim
   1644278 ±  3%     +48.5%    2442550 ±  5%  numa-meminfo.node0.Slab
    612346 ± 43%     -69.9%     184504 ± 45%  numa-meminfo.node1.Active
    612342 ± 43%     -69.9%     184500 ± 45%  numa-meminfo.node1.Active(anon)
   1674209 ± 85%     -78.1%     366914 ± 42%  numa-meminfo.node1.FilePages
     75408 ±  5%     +15.6%      87173 ±  6%  numa-meminfo.node1.PageTables
   1573523 ±  4%     +52.0%    2391230 ±  7%  numa-meminfo.node1.SUnreclaim
    465316 ± 27%     -72.7%     127017 ± 75%  numa-meminfo.node1.Shmem
   1626985 ±  4%     +49.0%    2424297 ±  7%  numa-meminfo.node1.Slab
   4023313 ± 31%     +58.4%    6371888 ± 27%  numa-meminfo.node2.MemUsed
   1571082 ±  3%     +55.3%    2440472 ±  5%  numa-meminfo.node2.SUnreclaim
   1604428 ±  3%     +55.3%    2492401 ±  6%  numa-meminfo.node2.Slab
   1548894 ±  4%     +54.2%    2388081 ±  6%  numa-meminfo.node3.SUnreclaim
   1586066 ±  4%     +53.0%    2427265 ±  6%  numa-meminfo.node3.Slab
     18610 ±  6%     +16.4%      21662 ±  7%  numa-vmstat.node0.nr_page_table_pages
    396573 ±  3%     +49.9%     594571 ±  6%  numa-vmstat.node0.nr_slab_unreclaimable
  25430185 ±  4%     +14.8%   29188938 ±  3%  numa-vmstat.node0.numa_hit
  25318163 ±  4%     +14.8%   29072880 ±  3%  numa-vmstat.node0.numa_local
    153177 ± 43%     -69.8%      46194 ± 45%  numa-vmstat.node1.nr_active_anon
    418628 ± 85%     -78.1%      91778 ± 42%  numa-vmstat.node1.nr_file_pages
     18770 ±  5%     +16.7%      21911 ±  7%  numa-vmstat.node1.nr_page_table_pages
    116405 ± 28%     -72.7%      31803 ± 75%  numa-vmstat.node1.nr_shmem
    393591 ±  4%     +52.0%     598339 ±  7%  numa-vmstat.node1.nr_slab_unreclaimable
    153176 ± 43%     -69.8%      46192 ± 45%  numa-vmstat.node1.nr_zone_active_anon
  25681692 ±  4%     +15.2%   29589055 ±  3%  numa-vmstat.node1.numa_hit
  25579141 ±  4%     +15.2%   29474199 ±  3%  numa-vmstat.node1.numa_local
    392991 ±  3%     +55.4%     610669 ±  5%  numa-vmstat.node2.nr_slab_unreclaimable
  25579351 ±  4%     +19.1%   30461531 ±  2%  numa-vmstat.node2.numa_hit
  25472265 ±  4%     +19.2%   30373528 ±  2%  numa-vmstat.node2.numa_local
    387461 ±  4%     +54.2%     597608 ±  6%  numa-vmstat.node3.nr_slab_unreclaimable
  25588424 ±  4%     +18.2%   30237199 ±  2%  numa-vmstat.node3.numa_hit
  25502545 ±  4%     +18.2%   30147394 ±  2%  numa-vmstat.node3.numa_local
      5.21 ±  3%     +11.9%       5.83 ±  2%  perf-stat.i.MPKI
      0.47 ±  2%      +0.1        0.55 ±  3%  perf-stat.i.branch-miss-rate%
 1.854e+08           +18.3%  2.194e+08 ±  4%  perf-stat.i.branch-misses
 1.039e+09 ±  2%     +15.1%  1.196e+09 ±  3%  perf-stat.i.cache-misses
 1.747e+09 ±  2%     +15.0%  2.008e+09 ±  3%  perf-stat.i.cache-references
      2.90            -1.9%       2.85        perf-stat.i.cpi
     67064 ±  3%     +10.5%      74100 ±  4%  perf-stat.i.cpu-migrations
    557.14 ±  3%     -12.3%     488.69 ±  3%  perf-stat.i.cycles-between-cache-misses
 1.994e+11            +2.9%  2.051e+11        perf-stat.i.instructions
      0.34            +1.9%       0.35        perf-stat.i.ipc
    252.47 ± 31%    +100.4%     505.88 ± 16%  perf-stat.i.major-faults
     24.13 ±  2%     +14.2%      27.55 ±  4%  perf-stat.i.metric.K/sec
   2906494 ±  3%     +15.6%    3358742 ±  3%  perf-stat.i.minor-faults
   2906746 ±  3%     +15.6%    3359248 ±  3%  perf-stat.i.page-faults
      5.21 ±  3%     +11.9%       5.83 ±  2%  perf-stat.overall.MPKI
      0.47 ±  2%      +0.1        0.55 ±  3%  perf-stat.overall.branch-miss-rate%
      2.90            -1.9%       2.85        perf-stat.overall.cpi
    557.08 ±  3%     -12.3%     488.29 ±  3%  perf-stat.overall.cycles-between-cache-misses
      0.34            +1.9%       0.35        perf-stat.overall.ipc
 1.808e+08           +18.6%  2.144e+08 ±  4%  perf-stat.ps.branch-misses
 1.015e+09 ±  2%     +15.4%  1.171e+09 ±  3%  perf-stat.ps.cache-misses
 1.706e+09           +15.2%  1.966e+09 ±  3%  perf-stat.ps.cache-references
     65460 ±  3%     +10.8%      72507 ±  4%  perf-stat.ps.cpu-migrations
 1.947e+11            +3.1%  2.008e+11        perf-stat.ps.instructions
    246.17 ± 31%    +100.9%     494.53 ± 16%  perf-stat.ps.major-faults
   2838136 ±  3%     +15.8%    3287669 ±  3%  perf-stat.ps.minor-faults
   2838383 ±  3%     +15.8%    3288164 ±  3%  perf-stat.ps.page-faults
 1.169e+13            +2.8%  1.201e+13        perf-stat.total.instructions
     35.08            -3.1       31.96 ±  2%  perf-profile.calltrace.cycles-pp.rwsem_down_write_slowpath.down_write.unlink_anon_vmas.free_pgtables.exit_mmap
     35.27            -3.1       32.16 ±  2%  perf-profile.calltrace.cycles-pp.down_write.unlink_anon_vmas.free_pgtables.exit_mmap.__mmput
     39.70            -2.3       37.35        perf-profile.calltrace.cycles-pp.unlink_anon_vmas.free_pgtables.exit_mmap.__mmput.exit_mm
     39.54            -2.3       37.20        perf-profile.calltrace.cycles-pp.anon_vma_clone.anon_vma_fork.dup_mmap.dup_mm.copy_process
     40.13            -2.2       37.91        perf-profile.calltrace.cycles-pp.anon_vma_fork.dup_mmap.dup_mm.copy_process.kernel_clone
     40.88            -2.2       38.70        perf-profile.calltrace.cycles-pp.free_pgtables.exit_mmap.__mmput.exit_mm.do_exit
     31.32 ±  2%      -2.2       29.16 ±  3%  perf-profile.calltrace.cycles-pp.osq_lock.rwsem_down_write_slowpath.down_write.unlink_anon_vmas.free_pgtables
     33.10            -2.1       31.01 ±  2%  perf-profile.calltrace.cycles-pp.rwsem_down_write_slowpath.down_write.anon_vma_clone.anon_vma_fork.dup_mmap
     33.39            -2.1       31.31 ±  2%  perf-profile.calltrace.cycles-pp.down_write.anon_vma_clone.anon_vma_fork.dup_mmap.dup_mm
     44.86            -1.4       43.49        perf-profile.calltrace.cycles-pp.dup_mmap.dup_mm.copy_process.kernel_clone.__do_sys_clone
     45.45            -1.3       44.18        perf-profile.calltrace.cycles-pp.dup_mm.copy_process.kernel_clone.__do_sys_clone.do_syscall_64
     45.91            -1.2       44.70        perf-profile.calltrace.cycles-pp.copy_process.kernel_clone.__do_sys_clone.do_syscall_64.entry_SYSCALL_64_after_hwframe
     46.43            -1.1       45.30        perf-profile.calltrace.cycles-pp.__do_sys_clone.do_syscall_64.entry_SYSCALL_64_after_hwframe
     46.43            -1.1       45.30        perf-profile.calltrace.cycles-pp.kernel_clone.__do_sys_clone.do_syscall_64.entry_SYSCALL_64_after_hwframe
     94.97            -1.0       93.97        perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe
     94.97            -1.0       93.98        perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe
      3.30 ±  6%      -0.9        2.38 ±  8%  perf-profile.calltrace.cycles-pp.rwsem_spin_on_owner.rwsem_down_write_slowpath.down_write.unlink_anon_vmas.free_pgtables
      3.36 ±  2%      -0.7        2.66 ±  4%  perf-profile.calltrace.cycles-pp.anon_vma_interval_tree_insert.anon_vma_clone.anon_vma_fork.dup_mmap.dup_mm
      2.67 ±  7%      -0.7        1.98 ±  9%  perf-profile.calltrace.cycles-pp.rwsem_spin_on_owner.rwsem_down_write_slowpath.down_write.anon_vma_clone.anon_vma_fork
      0.60            +0.1        0.68        perf-profile.calltrace.cycles-pp.kmem_cache_alloc_noprof.vm_area_dup.dup_mmap.dup_mm.copy_process
      0.65            +0.1        0.74 ±  3%  perf-profile.calltrace.cycles-pp.__vma_start_write.free_pgtables.exit_mmap.__mmput.exit_mm
      0.64            +0.1        0.74        perf-profile.calltrace.cycles-pp.vm_area_dup.dup_mmap.dup_mm.copy_process.kernel_clone
      0.77 ±  2%      +0.1        0.87 ±  3%  perf-profile.calltrace.cycles-pp.__vma_start_write.dup_mmap.dup_mm.copy_process.kernel_clone
      0.58 ±  5%      +0.1        0.68 ±  5%  perf-profile.calltrace.cycles-pp.__memcg_slab_free_hook.kmem_cache_free.unlink_anon_vmas.free_pgtables.exit_mmap
      0.43 ± 44%      +0.2        0.61 ±  3%  perf-profile.calltrace.cycles-pp.mm_init.dup_mm.copy_process.kernel_clone.__do_sys_clone
      1.05 ±  5%      +0.2        1.26 ±  7%  perf-profile.calltrace.cycles-pp.kmem_cache_alloc_noprof.anon_vma_clone.anon_vma_fork.dup_mmap.dup_mm
      1.23 ±  4%      +0.2        1.46 ±  5%  perf-profile.calltrace.cycles-pp.__slab_free.kmem_cache_free.unlink_anon_vmas.free_pgtables.exit_mmap
      0.34 ± 70%      +0.2        0.58 ±  2%  perf-profile.calltrace.cycles-pp.wake_up_new_task.kernel_clone.__do_sys_clone.do_syscall_64.entry_SYSCALL_64_after_hwframe
      1.39 ±  3%      +0.2        1.63 ±  5%  perf-profile.calltrace.cycles-pp.__anon_vma_interval_tree_remove.unlink_anon_vmas.free_pgtables.exit_mmap.__mmput
      1.02 ±  9%      +0.2        1.27 ±  5%  perf-profile.calltrace.cycles-pp.copy_present_ptes.copy_pte_range.copy_p4d_range.copy_page_range.dup_mmap
      1.42 ±  6%      +0.3        1.74 ±  4%  perf-profile.calltrace.cycles-pp.copy_pte_range.copy_p4d_range.copy_page_range.dup_mmap.dup_mm
      1.59 ±  5%      +0.3        1.93 ±  4%  perf-profile.calltrace.cycles-pp.copy_p4d_range.copy_page_range.dup_mmap.dup_mm.copy_process
      1.62 ±  4%      +0.4        1.97 ±  3%  perf-profile.calltrace.cycles-pp.copy_page_range.dup_mmap.dup_mm.copy_process.kernel_clone
      0.18 ±141%      +0.4        0.56 ±  6%  perf-profile.calltrace.cycles-pp.up_write.unlink_anon_vmas.free_pgtables.exit_mmap.__mmput
      2.04 ±  4%      +0.4        2.43 ±  5%  perf-profile.calltrace.cycles-pp.kmem_cache_free.unlink_anon_vmas.free_pgtables.exit_mmap.__mmput
      1.08 ± 12%      +0.5        1.54 ±  9%  perf-profile.calltrace.cycles-pp.folios_put_refs.free_pages_and_swap_cache.__tlb_batch_free_encoded_pages.tlb_finish_mmu.exit_mmap
      1.09 ± 13%      +0.5        1.58 ±  7%  perf-profile.calltrace.cycles-pp.folio_remove_rmap_ptes.zap_present_ptes.zap_pte_range.zap_pmd_range.unmap_page_range
      0.08 ±223%      +0.5        0.57 ±  4%  perf-profile.calltrace.cycles-pp.up_write.anon_vma_clone.anon_vma_fork.dup_mmap.dup_mm
      0.00            +0.6        0.55        perf-profile.calltrace.cycles-pp.kmem_cache_free.exit_mmap.__mmput.exit_mm.do_exit
      1.24 ±  3%      +0.6        1.79 ±  8%  perf-profile.calltrace.cycles-pp.next_uptodate_folio.filemap_map_pages.do_read_fault.do_fault.__handle_mm_fault
      1.45 ± 10%      +0.6        2.03 ±  8%  perf-profile.calltrace.cycles-pp.free_pages_and_swap_cache.__tlb_batch_free_encoded_pages.tlb_finish_mmu.exit_mmap.__mmput
      1.46 ± 10%      +0.6        2.05 ±  7%  perf-profile.calltrace.cycles-pp.__tlb_batch_free_encoded_pages.tlb_finish_mmu.exit_mmap.__mmput.exit_mm
      1.48 ± 10%      +0.6        2.07 ±  8%  perf-profile.calltrace.cycles-pp.tlb_finish_mmu.exit_mmap.__mmput.exit_mm.do_exit
      0.08 ±223%      +0.6        0.71 ±  8%  perf-profile.calltrace.cycles-pp.free_pages_and_swap_cache.__tlb_batch_free_encoded_pages.tlb_flush_mmu.zap_pte_range.zap_pmd_range
      0.08 ±223%      +0.6        0.71 ±  8%  perf-profile.calltrace.cycles-pp.__tlb_batch_free_encoded_pages.tlb_flush_mmu.zap_pte_range.zap_pmd_range.unmap_page_range
      0.08 ±223%      +0.6        0.71 ±  8%  perf-profile.calltrace.cycles-pp.tlb_flush_mmu.zap_pte_range.zap_pmd_range.unmap_page_range.unmap_vmas
      1.85 ±  3%      +0.7        2.58 ±  7%  perf-profile.calltrace.cycles-pp.filemap_map_pages.do_read_fault.do_fault.__handle_mm_fault.handle_mm_fault
      1.94 ±  3%      +0.8        2.70 ±  7%  perf-profile.calltrace.cycles-pp.do_fault.__handle_mm_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault
      1.93 ±  3%      +0.8        2.69 ±  7%  perf-profile.calltrace.cycles-pp.do_read_fault.do_fault.__handle_mm_fault.handle_mm_fault.do_user_addr_fault
      2.27 ±  3%      +0.8        3.08 ±  6%  perf-profile.calltrace.cycles-pp.__handle_mm_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault
      2.34 ±  3%      +0.8        3.16 ±  6%  perf-profile.calltrace.cycles-pp.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault
      2.68 ±  3%      +0.9        3.57 ±  6%  perf-profile.calltrace.cycles-pp.exc_page_fault.asm_exc_page_fault
      2.66 ±  3%      +0.9        3.56 ±  6%  perf-profile.calltrace.cycles-pp.do_user_addr_fault.exc_page_fault.asm_exc_page_fault
      2.88 ±  3%      +0.9        3.80 ±  6%  perf-profile.calltrace.cycles-pp.asm_exc_page_fault
      2.56 ±  7%      +1.0        3.51 ±  5%  perf-profile.calltrace.cycles-pp.zap_present_ptes.zap_pte_range.zap_pmd_range.unmap_page_range.unmap_vmas
      3.40 ±  6%      +1.3        4.72 ±  5%  perf-profile.calltrace.cycles-pp.zap_pte_range.zap_pmd_range.unmap_page_range.unmap_vmas.exit_mmap
      3.45 ±  6%      +1.3        4.78 ±  5%  perf-profile.calltrace.cycles-pp.zap_pmd_range.unmap_page_range.unmap_vmas.exit_mmap.__mmput
      3.50 ±  6%      +1.3        4.85 ±  5%  perf-profile.calltrace.cycles-pp.unmap_page_range.unmap_vmas.exit_mmap.__mmput.exit_mm
      3.63 ±  6%      +1.4        4.98 ±  5%  perf-profile.calltrace.cycles-pp.unmap_vmas.exit_mmap.__mmput.exit_mm.do_exit
     68.28            -5.1       63.15 ±  2%  perf-profile.children.cycles-pp.rwsem_down_write_slowpath
     68.93            -5.1       63.84 ±  2%  perf-profile.children.cycles-pp.down_write
     61.34 ±  2%      -3.5       57.82 ±  3%  perf-profile.children.cycles-pp.osq_lock
     39.70            -2.3       37.37        perf-profile.children.cycles-pp.unlink_anon_vmas
     39.56            -2.3       37.23        perf-profile.children.cycles-pp.anon_vma_clone
     40.14            -2.2       37.92        perf-profile.children.cycles-pp.anon_vma_fork
     40.88            -2.2       38.71        perf-profile.children.cycles-pp.free_pgtables
      6.01 ±  6%      -1.6        4.41 ±  9%  perf-profile.children.cycles-pp.rwsem_spin_on_owner
     44.93            -1.4       43.56        perf-profile.children.cycles-pp.dup_mmap
     45.45            -1.3       44.18        perf-profile.children.cycles-pp.dup_mm
     45.91            -1.2       44.70        perf-profile.children.cycles-pp.copy_process
     46.43            -1.1       45.30        perf-profile.children.cycles-pp.__do_sys_clone
     46.43            -1.1       45.30        perf-profile.children.cycles-pp.kernel_clone
     95.09            -1.0       94.09        perf-profile.children.cycles-pp.do_syscall_64
     95.09            -1.0       94.09        perf-profile.children.cycles-pp.entry_SYSCALL_64_after_hwframe
      3.37 ±  2%      -0.7        2.67 ±  4%  perf-profile.children.cycles-pp.anon_vma_interval_tree_insert
      0.05            +0.0        0.06        perf-profile.children.cycles-pp.autoremove_wake_function
      0.05            +0.0        0.06        perf-profile.children.cycles-pp.lru_add_drain_cpu
      0.05            +0.0        0.06        perf-profile.children.cycles-pp.mas_dup_alloc
      0.05            +0.0        0.06        perf-profile.children.cycles-pp.sched_move_task
      0.07            +0.0        0.08        perf-profile.children.cycles-pp.process_one_work
      0.06            +0.0        0.07        perf-profile.children.cycles-pp.__wake_up_common
      0.06            +0.0        0.07        perf-profile.children.cycles-pp.delayed_vfree_work
      0.06            +0.0        0.07        perf-profile.children.cycles-pp.vfree
      0.10            +0.0        0.11 ±  3%  perf-profile.children.cycles-pp.__mt_dup
      0.07            +0.0        0.08 ±  4%  perf-profile.children.cycles-pp.anon_pipe_write
      0.05            +0.0        0.06 ±  6%  perf-profile.children.cycles-pp.folio_batch_move_lru
      0.05            +0.0        0.06 ±  6%  perf-profile.children.cycles-pp.lru_add_drain
      0.06            +0.0        0.07 ±  5%  perf-profile.children.cycles-pp.__wake_up_sync_key
      0.08            +0.0        0.09 ±  4%  perf-profile.children.cycles-pp.__wp_page_copy_user
      0.08            +0.0        0.09 ±  5%  perf-profile.children.cycles-pp.__put_user_4
      0.07 ±  9%      +0.0        0.08 ±  5%  perf-profile.children.cycles-pp.__rmqueue_pcplist
      0.08 ±  5%      +0.0        0.10 ±  4%  perf-profile.children.cycles-pp.vm_normal_page
      0.11 ±  4%      +0.0        0.12 ±  3%  perf-profile.children.cycles-pp.vfs_write
      0.15 ±  2%      +0.0        0.16 ±  2%  perf-profile.children.cycles-pp.vma_mark_detached
      0.08 ±  6%      +0.0        0.09 ±  4%  perf-profile.children.cycles-pp.do_wp_page
      0.09 ±  5%      +0.0        0.10 ±  3%  perf-profile.children.cycles-pp.kernel_init_pages
      0.12 ±  3%      +0.0        0.14 ±  2%  perf-profile.children.cycles-pp.worker_thread
      0.08 ±  5%      +0.0        0.10 ±  3%  perf-profile.children.cycles-pp.__vmalloc_node_noprof
      0.08 ±  5%      +0.0        0.10 ±  3%  perf-profile.children.cycles-pp.__vmalloc_node_range_noprof
      0.12            +0.0        0.14 ±  3%  perf-profile.children.cycles-pp.acct_collect
      0.12 ±  4%      +0.0        0.13 ±  2%  perf-profile.children.cycles-pp.ksys_write
      0.05 ± 13%      +0.0        0.07 ± 14%  perf-profile.children.cycles-pp.__pmd_alloc
      0.10 ±  3%      +0.0        0.12 ±  5%  perf-profile.children.cycles-pp.rb_next
      0.17 ±  4%      +0.0        0.19 ±  2%  perf-profile.children.cycles-pp.sched_balance_find_src_group
      0.11 ±  6%      +0.0        0.13 ±  3%  perf-profile.children.cycles-pp.free_frozen_page_commit
      0.09 ±  7%      +0.0        0.11 ±  4%  perf-profile.children.cycles-pp.prep_new_page
      0.10 ±  4%      +0.0        0.12 ±  4%  perf-profile.children.cycles-pp.rmqueue
      0.16 ±  2%      +0.0        0.18 ±  3%  perf-profile.children.cycles-pp.kthread
      0.08 ±  4%      +0.0        0.10 ±  4%  perf-profile.children.cycles-pp._copy_from_user
      0.23 ±  5%      +0.0        0.26 ±  2%  perf-profile.children.cycles-pp.sched_balance_rq
      0.17 ±  2%      +0.0        0.20 ±  3%  perf-profile.children.cycles-pp.sync_regs
      0.08 ±  4%      +0.0        0.11 ±  4%  perf-profile.children.cycles-pp.__x64_sys_rt_sigprocmask
      0.12 ±  6%      +0.0        0.14 ±  4%  perf-profile.children.cycles-pp.__memcg_kmem_charge_page
      0.21            +0.0        0.24 ±  3%  perf-profile.children.cycles-pp.mas_next_slot
      0.09 ±  4%      +0.0        0.12 ±  5%  perf-profile.children.cycles-pp.__put_anon_vma
      0.26            +0.0        0.28 ±  2%  perf-profile.children.cycles-pp.mas_find
      0.16 ±  6%      +0.0        0.20 ±  2%  perf-profile.children.cycles-pp.__free_frozen_pages
      0.16 ±  2%      +0.0        0.20 ±  2%  perf-profile.children.cycles-pp.__pte_offset_map_lock
      0.18 ±  7%      +0.0        0.22 ±  2%  perf-profile.children.cycles-pp.tlb_remove_table_rcu
      0.23            +0.0        0.26 ±  2%  perf-profile.children.cycles-pp.lock_vma_under_rcu
      0.11 ±  6%      +0.0        0.14 ± 10%  perf-profile.children.cycles-pp.__put_partials
      0.24 ±  2%      +0.0        0.27 ±  3%  perf-profile.children.cycles-pp.__pi_memset
      0.22 ±  5%      +0.0        0.25 ±  2%  perf-profile.children.cycles-pp.wp_page_copy
      0.16 ±  3%      +0.0        0.20 ±  3%  perf-profile.children.cycles-pp.remove_vma
      0.14 ±  3%      +0.0        0.18 ±  5%  perf-profile.children.cycles-pp.fput
      0.09 ±  4%      +0.0        0.13 ±  7%  perf-profile.children.cycles-pp.getname_flags
      0.20 ±  6%      +0.0        0.24        perf-profile.children.cycles-pp.pte_alloc_one
      0.09 ±  4%      +0.0        0.13 ±  7%  perf-profile.children.cycles-pp.strncpy_from_user
      0.20 ±  4%      +0.0        0.24 ±  3%  perf-profile.children.cycles-pp.get_page_from_freelist
      0.10 ±  4%      +0.0        0.13 ±  7%  perf-profile.children.cycles-pp.user_path_at
      0.10 ±  3%      +0.0        0.14 ±  7%  perf-profile.children.cycles-pp.__x64_sys_chdir
      0.22 ±  3%      +0.0        0.26 ±  3%  perf-profile.children.cycles-pp.native_irq_return_iret
      0.29            +0.0        0.32 ±  2%  perf-profile.children.cycles-pp.__percpu_counter_sum
      0.36 ±  2%      +0.0        0.40 ±  2%  perf-profile.children.cycles-pp.__kmem_cache_alloc_bulk
      0.33 ±  3%      +0.0        0.37 ±  2%  perf-profile.children.cycles-pp.__account_obj_stock
      0.35 ±  2%      +0.0        0.40        perf-profile.children.cycles-pp.__pcs_replace_empty_main
      0.18 ±  2%      +0.0        0.22 ±  4%  perf-profile.children.cycles-pp.do_task_dead
      0.24 ±  3%      +0.0        0.29 ±  2%  perf-profile.children.cycles-pp.schedule_tail
      1.28            +0.0        1.32        perf-profile.children.cycles-pp.sysvec_apic_timer_interrupt
      0.37            +0.0        0.41 ±  3%  perf-profile.children.cycles-pp.update_sg_wakeup_stats
      0.25 ±  2%      +0.0        0.30 ±  3%  perf-profile.children.cycles-pp.__percpu_counter_init_many
      0.18 ±  6%      +0.0        0.24 ±  8%  perf-profile.children.cycles-pp.get_partial_node
      0.00            +0.1        0.05        perf-profile.children.cycles-pp.vm_area_init_from
      1.32            +0.1        1.37        perf-profile.children.cycles-pp.asm_sysvec_apic_timer_interrupt
      0.40 ±  3%      +0.1        0.45 ±  3%  perf-profile.children.cycles-pp.rcu_core
      0.38 ±  3%      +0.1        0.43 ±  2%  perf-profile.children.cycles-pp.rcu_do_batch
      0.44            +0.1        0.50 ±  2%  perf-profile.children.cycles-pp.mas_store
      0.30 ±  3%      +0.1        0.35        perf-profile.children.cycles-pp.vfs_read
      0.25 ±  3%      +0.1        0.30 ±  2%  perf-profile.children.cycles-pp.anon_pipe_read
      0.33            +0.1        0.38        perf-profile.children.cycles-pp.finish_task_switch
      0.30 ±  2%      +0.1        0.36        perf-profile.children.cycles-pp.ksys_read
      0.38 ±  2%      +0.1        0.43 ±  3%  perf-profile.children.cycles-pp.sched_balance_find_dst_group
      0.36 ±  2%      +0.1        0.42 ±  2%  perf-profile.children.cycles-pp.__mmdrop
      0.37 ±  3%      +0.1        0.42 ±  2%  perf-profile.children.cycles-pp.refill_obj_stock
      0.57 ±  2%      +0.1        0.63 ±  2%  perf-profile.children.cycles-pp.__irq_exit_rcu
      0.36            +0.1        0.42 ±  4%  perf-profile.children.cycles-pp.unlink_file_vma_batch_add
      0.44 ±  3%      +0.1        0.50 ±  2%  perf-profile.children.cycles-pp._raw_spin_lock
      0.32 ±  5%      +0.1        0.38 ±  2%  perf-profile.children.cycles-pp.alloc_pages_noprof
      0.60 ±  3%      +0.1        0.66 ±  3%  perf-profile.children.cycles-pp.handle_softirqs
      0.48            +0.1        0.54 ±  3%  perf-profile.children.cycles-pp.select_task_rq_fair
      0.41 ±  5%      +0.1        0.48 ±  4%  perf-profile.children.cycles-pp.__rb_insert_augmented
      0.38 ±  2%      +0.1        0.44 ±  3%  perf-profile.children.cycles-pp.unlink_file_vma_batch_process
      0.34 ±  5%      +0.1        0.41 ±  2%  perf-profile.children.cycles-pp.alloc_pages_mpol
      0.33 ±  5%      +0.1        0.40 ±  2%  perf-profile.children.cycles-pp.__alloc_frozen_pages_noprof
      0.42            +0.1        0.49 ±  2%  perf-profile.children.cycles-pp.ret_from_fork
      0.42            +0.1        0.49 ±  2%  perf-profile.children.cycles-pp.ret_from_fork_asm
      0.50            +0.1        0.58 ±  2%  perf-profile.children.cycles-pp.wake_up_new_task
      0.34 ±  8%      +0.1        0.42 ±  8%  perf-profile.children.cycles-pp._raw_spin_lock_irqsave
      0.20 ±  3%      +0.1        0.28 ±  6%  perf-profile.children.cycles-pp.folio_add_file_rmap_ptes
      0.35 ±  5%      +0.1        0.42 ±  9%  perf-profile.children.cycles-pp.native_queued_spin_lock_slowpath
      0.46 ±  2%      +0.1        0.55 ±  3%  perf-profile.children.cycles-pp.pcpu_alloc_noprof
      0.31 ±  3%      +0.1        0.40 ±  5%  perf-profile.children.cycles-pp.set_pte_range
      0.65            +0.1        0.74        perf-profile.children.cycles-pp.vm_area_dup
      0.74 ±  2%      +0.1        0.84 ±  3%  perf-profile.children.cycles-pp.__vma_enter_locked
      0.51 ±  2%      +0.1        0.61 ±  3%  perf-profile.children.cycles-pp.mm_init
      0.97 ±  4%      +0.2        1.14 ±  3%  perf-profile.children.cycles-pp.__memcg_slab_free_hook
      1.43 ±  2%      +0.2        1.62 ±  3%  perf-profile.children.cycles-pp.__vma_start_write
      1.40 ±  3%      +0.2        1.64 ±  5%  perf-profile.children.cycles-pp.__anon_vma_interval_tree_remove
      1.04 ±  9%      +0.2        1.28 ±  4%  perf-profile.children.cycles-pp.copy_present_ptes
      1.46 ±  3%      +0.2        1.70 ±  5%  perf-profile.children.cycles-pp.__slab_free
      1.20 ±  7%      +0.3        1.45 ±  4%  perf-profile.children.cycles-pp.up_write
      0.42 ± 12%      +0.3        0.71 ±  8%  perf-profile.children.cycles-pp.tlb_flush_mmu
      1.44 ±  6%      +0.3        1.75 ±  4%  perf-profile.children.cycles-pp.copy_pte_range
      1.60 ±  4%      +0.3        1.94 ±  4%  perf-profile.children.cycles-pp.copy_p4d_range
      1.63 ±  4%      +0.3        1.97 ±  4%  perf-profile.children.cycles-pp.copy_page_range
      2.10 ±  3%      +0.4        2.46 ±  3%  perf-profile.children.cycles-pp.kmem_cache_alloc_noprof
      1.11 ± 13%      +0.5        1.60 ±  7%  perf-profile.children.cycles-pp.folio_remove_rmap_ptes
      2.82 ±  3%      +0.5        3.32 ±  4%  perf-profile.children.cycles-pp.kmem_cache_free
      1.32 ±  3%      +0.6        1.90 ±  8%  perf-profile.children.cycles-pp.next_uptodate_folio
      1.48 ± 10%      +0.6        2.07 ±  8%  perf-profile.children.cycles-pp.tlb_finish_mmu
      1.41 ± 12%      +0.7        2.10 ±  6%  perf-profile.children.cycles-pp.folios_put_refs
      1.98 ±  3%      +0.8        2.76 ±  7%  perf-profile.children.cycles-pp.filemap_map_pages
      2.05 ±  3%      +0.8        2.86 ±  7%  perf-profile.children.cycles-pp.do_read_fault
      2.05 ±  3%      +0.8        2.87 ±  7%  perf-profile.children.cycles-pp.do_fault
      2.43 ±  3%      +0.9        3.31 ±  6%  perf-profile.children.cycles-pp.__handle_mm_fault
      1.88 ± 10%      +0.9        2.76 ±  6%  perf-profile.children.cycles-pp.__tlb_batch_free_encoded_pages
      1.88 ± 10%      +0.9        2.76 ±  6%  perf-profile.children.cycles-pp.free_pages_and_swap_cache
      2.51 ±  3%      +0.9        3.40 ±  6%  perf-profile.children.cycles-pp.handle_mm_fault
      2.88 ±  3%      +1.0        3.84 ±  6%  perf-profile.children.cycles-pp.exc_page_fault
      2.87 ±  3%      +1.0        3.83 ±  6%  perf-profile.children.cycles-pp.do_user_addr_fault
      2.61 ±  7%      +1.0        3.58 ±  5%  perf-profile.children.cycles-pp.zap_present_ptes
      3.12 ±  3%      +1.0        4.11 ±  6%  perf-profile.children.cycles-pp.asm_exc_page_fault
      3.41 ±  6%      +1.3        4.74 ±  5%  perf-profile.children.cycles-pp.zap_pte_range
      3.45 ±  6%      +1.3        4.79 ±  5%  perf-profile.children.cycles-pp.zap_pmd_range
      3.51 ±  6%      +1.3        4.85 ±  5%  perf-profile.children.cycles-pp.unmap_page_range
      3.63 ±  6%      +1.4        4.99 ±  5%  perf-profile.children.cycles-pp.unmap_vmas
     60.76 ±  2%      -3.5       57.25 ±  3%  perf-profile.self.cycles-pp.osq_lock
      5.95 ±  6%      -1.6        4.36 ±  9%  perf-profile.self.cycles-pp.rwsem_spin_on_owner
      3.32 ±  2%      -0.7        2.62 ±  4%  perf-profile.self.cycles-pp.anon_vma_interval_tree_insert
      0.05            +0.0        0.06        perf-profile.self.cycles-pp._find_next_and_bit
      0.06 ±  7%      +0.0        0.08 ±  6%  perf-profile.self.cycles-pp.vm_normal_page
      0.09 ±  5%      +0.0        0.10        perf-profile.self.cycles-pp.kernel_init_pages
      0.07 ±  5%      +0.0        0.08 ±  5%  perf-profile.self.cycles-pp.__memcg_kmem_charge_page
      0.14 ±  3%      +0.0        0.16 ±  2%  perf-profile.self.cycles-pp.vma_mark_detached
      0.10 ±  3%      +0.0        0.12 ±  4%  perf-profile.self.cycles-pp.set_pte_range
      0.07 ±  7%      +0.0        0.09 ± 11%  perf-profile.self.cycles-pp.do_user_addr_fault
      0.15 ±  3%      +0.0        0.18 ±  2%  perf-profile.self.cycles-pp.lock_vma_under_rcu
      0.17 ±  3%      +0.0        0.19 ±  3%  perf-profile.self.cycles-pp.sync_regs
      0.24 ±  3%      +0.0        0.27 ±  3%  perf-profile.self.cycles-pp.__account_obj_stock
      0.14 ±  3%      +0.0        0.17 ±  6%  perf-profile.self.cycles-pp.get_partial_node
      0.18 ±  2%      +0.0        0.20 ±  3%  perf-profile.self.cycles-pp.refill_obj_stock
      0.22 ±  2%      +0.0        0.24 ±  3%  perf-profile.self.cycles-pp.__percpu_counter_sum
      0.25            +0.0        0.28        perf-profile.self.cycles-pp.__kmem_cache_alloc_bulk
      0.22 ±  3%      +0.0        0.25 ±  3%  perf-profile.self.cycles-pp.__pi_memset
      0.14 ±  3%      +0.0        0.18 ±  4%  perf-profile.self.cycles-pp.fput
      0.30            +0.0        0.34        perf-profile.self.cycles-pp.mas_store
      0.22 ±  3%      +0.0        0.26 ±  3%  perf-profile.self.cycles-pp.native_irq_return_iret
      0.30            +0.0        0.33 ±  4%  perf-profile.self.cycles-pp.update_sg_wakeup_stats
      0.37 ±  4%      +0.0        0.41 ±  4%  perf-profile.self.cycles-pp.rwsem_down_write_slowpath
      0.28 ±  6%      +0.0        0.32 ±  4%  perf-profile.self.cycles-pp.zap_pte_range
      0.16 ±  3%      +0.0        0.20 ±  3%  perf-profile.self.cycles-pp.pcpu_alloc_noprof
      0.28 ±  3%      +0.0        0.33 ±  6%  perf-profile.self.cycles-pp.unlink_anon_vmas
      0.27            +0.0        0.31 ±  5%  perf-profile.self.cycles-pp.kmem_cache_free
      0.38            +0.0        0.43        perf-profile.self.cycles-pp._raw_spin_lock
      0.00            +0.1        0.05        perf-profile.self.cycles-pp.cpu_util
      0.00            +0.1        0.05        perf-profile.self.cycles-pp.free_pcppages_bulk
      0.00            +0.1        0.05        perf-profile.self.cycles-pp.mas_find
      0.40 ±  5%      +0.1        0.46 ±  4%  perf-profile.self.cycles-pp.__rb_insert_augmented
      0.20 ±  3%      +0.1        0.27 ±  6%  perf-profile.self.cycles-pp.folio_add_file_rmap_ptes
      0.35 ±  5%      +0.1        0.42 ±  9%  perf-profile.self.cycles-pp.native_queued_spin_lock_slowpath
      0.68 ±  2%      +0.1        0.78 ±  2%  perf-profile.self.cycles-pp.__vma_start_write
      0.33 ±  3%      +0.1        0.43 ±  7%  perf-profile.self.cycles-pp.filemap_map_pages
      0.54            +0.1        0.64 ±  3%  perf-profile.self.cycles-pp.dup_mmap
      0.71 ±  2%      +0.1        0.81 ±  4%  perf-profile.self.cycles-pp.__vma_enter_locked
      0.63 ±  5%      +0.1        0.74 ±  5%  perf-profile.self.cycles-pp.__memcg_slab_free_hook
      0.46 ±  5%      +0.2        0.64 ±  5%  perf-profile.self.cycles-pp.free_pages_and_swap_cache
      1.28 ±  4%      +0.2        1.51 ±  5%  perf-profile.self.cycles-pp.__anon_vma_interval_tree_remove
      0.99 ±  9%      +0.2        1.22 ±  5%  perf-profile.self.cycles-pp.copy_present_ptes
      1.43 ±  3%      +0.2        1.67 ±  5%  perf-profile.self.cycles-pp.__slab_free
      0.30 ±  2%      +0.3        0.57 ±  6%  perf-profile.self.cycles-pp.kmem_cache_alloc_noprof
      0.96 ±  3%      +0.3        1.24 ±  2%  perf-profile.self.cycles-pp.up_write
      1.36 ±  3%      +0.5        1.82 ±  6%  perf-profile.self.cycles-pp.zap_present_ptes
      1.08 ± 13%      +0.5        1.56 ±  7%  perf-profile.self.cycles-pp.folio_remove_rmap_ptes
      1.24 ±  3%      +0.6        1.81 ±  8%  perf-profile.self.cycles-pp.next_uptodate_folio
      1.33 ± 12%      +0.7        2.01 ±  6%  perf-profile.self.cycles-pp.folios_put_refs


***************************************************************************************************

=========================================================================================
compiler/cpufreq_governor/kconfig/nr_threads/rootfs/tbox_group/test/testcase/testtime:
  gcc-14/performance/x86_64-rhel-9.4/100%/debian-13-x86_64-20250902.cgz/lkp-srf-2sp3/netlink-proc/stress-ng/60s

commit: 
  85f03a8631 ("mm/mmap_lock: add vma_is_attached() helper")
  bfc2b13b05 ("mm/rmap: allocate anon_vma_chain objects unlocked when possible")

85f03a86318c4172 bfc2b13b05a1343bb60a85d840f 
---------------- --------------------------- 
         %stddev     %change         %stddev
             \          |                \  
   9243133            -5.2%    8760646        cpuidle..usage
    346361            -6.4%     324082        meminfo.Mapped
   1529351            +6.6%    1630248        meminfo.Shmem
      0.35 ±  2%      +0.1        0.44        mpstat.cpu.all.soft%
      0.98            +0.1        1.09        mpstat.cpu.all.usr%
    298266            -2.9%     289575        vmstat.system.cs
    577247            +2.4%     590966        vmstat.system.in
   7105876            +7.0%    7601087        turbostat.NMI
    332.10            +2.8%     341.32        turbostat.PkgWatt
     25.50            +6.4%      27.14        turbostat.RAMWatt
  37440414 ±  9%     +26.7%   47453343 ±  7%  numa-numastat.node0.local_node
  37552968 ±  9%     +26.6%   47551257 ±  7%  numa-numastat.node0.numa_hit
  35626288 ±  2%     +30.3%   46417467 ±  2%  numa-numastat.node1.local_node
  35715809 ±  2%     +30.3%   46520561 ±  2%  numa-numastat.node1.numa_hit
  37554292 ±  9%     +26.6%   47552046 ±  7%  numa-vmstat.node0.numa_hit
  37441738 ±  9%     +26.7%   47454132 ±  7%  numa-vmstat.node0.numa_local
  35716707 ±  2%     +30.2%   46520870 ±  2%  numa-vmstat.node1.numa_hit
  35627089 ±  2%     +30.3%   46417777 ±  2%  numa-vmstat.node1.numa_local
     18927 ± 22%     +25.0%      23668 ±  2%  perf-c2c.DRAM.local
    147147 ± 19%     +29.2%     190098 ±  5%  perf-c2c.DRAM.remote
    263763 ± 20%     +48.3%     391131 ±  5%  perf-c2c.HITM.local
    118030 ± 19%     +30.4%     153956 ±  5%  perf-c2c.HITM.remote
    381793 ± 20%     +42.8%     545088 ±  4%  perf-c2c.HITM.total
    341231           +25.7%     429025        stress-ng.netlink-proc.ops
      5697           +25.7%       7159        stress-ng.netlink-proc.ops_per_sec
    726075           +27.3%     924191        stress-ng.time.involuntary_context_switches
      7082 ±119%    +175.4%      19501 ± 62%  stress-ng.time.major_page_faults
 1.257e+08           +27.8%  1.607e+08 ±  2%  stress-ng.time.minor_page_faults
     42.26           +26.2%      53.32        stress-ng.time.user_time
   6952244           -12.2%    6106739        stress-ng.time.voluntary_context_switches
      0.09           +11.7%       0.10        perf-sched.sch_delay.avg.ms.[unknown].[unknown].[unknown].[unknown].[unknown]
     76.37 ± 18%     -20.8%      60.45 ± 12%  perf-sched.sch_delay.max.ms.[unknown].[unknown].[unknown].[unknown].[unknown]
      0.09           +11.7%       0.10        perf-sched.total_sch_delay.average.ms
     76.37 ± 18%     -20.8%      60.45 ± 12%  perf-sched.total_sch_delay.max.ms
      2571 ± 16%     +44.4%       3714 ± 11%  perf-sched.total_wait_and_delay.max.ms
      2571 ± 16%     +44.4%       3714 ± 11%  perf-sched.total_wait_time.max.ms
      2571 ± 16%     +44.4%       3714 ± 11%  perf-sched.wait_and_delay.max.ms.[unknown].[unknown].[unknown].[unknown].[unknown]
      2571 ± 16%     +44.4%       3714 ± 11%  perf-sched.wait_time.max.ms.[unknown].[unknown].[unknown].[unknown].[unknown]
    601141            +4.8%     629898        proc-vmstat.nr_active_anon
   1307815            +1.9%    1333006        proc-vmstat.nr_file_pages
     86624            -6.4%      81044        proc-vmstat.nr_mapped
    382407            +6.6%     407598        proc-vmstat.nr_shmem
    601141            +4.8%     629898        proc-vmstat.nr_zone_active_anon
  73272349 ±  4%     +28.4%   94075287 ±  2%  proc-vmstat.numa_hit
  73070343 ±  4%     +28.5%   93874463 ±  2%  proc-vmstat.numa_local
  76707662 ±  4%     +28.2%   98366347 ±  2%  proc-vmstat.pgalloc_normal
 1.262e+08           +27.8%  1.613e+08 ±  2%  proc-vmstat.pgfault
  75478722 ±  4%     +28.6%   97070551 ±  2%  proc-vmstat.pgfree
   9954110 ±  2%     +37.2%   13659353 ±  4%  proc-vmstat.pgreuse
    806263           +76.7%    1424830 ± 91%  sched_debug.cfs_rq:/.left_deadline.max
    126768 ± 18%     +47.1%     186420 ± 34%  sched_debug.cfs_rq:/.left_deadline.stddev
    806262           +76.7%    1424828 ± 91%  sched_debug.cfs_rq:/.left_vruntime.max
    126766 ± 18%     +47.1%     186417 ± 34%  sched_debug.cfs_rq:/.left_vruntime.stddev
    806262           +76.7%    1424828 ± 91%  sched_debug.cfs_rq:/.right_vruntime.max
    126766 ± 18%     +47.1%     186418 ± 34%  sched_debug.cfs_rq:/.right_vruntime.stddev
      1083 ±  8%     +12.3%       1217 ±  4%  sched_debug.cfs_rq:/.runnable_avg.max
      1054 ±  7%     +12.4%       1185 ±  4%  sched_debug.cfs_rq:/.util_avg.max
    767399 ±  3%     +18.5%     909720 ±  9%  sched_debug.cpu.curr->pid.avg
   1022954           +25.4%    1283282        sched_debug.cpu.curr->pid.max
    438214 ±  3%     +30.7%     572602 ±  6%  sched_debug.cpu.curr->pid.stddev
      3.56           +17.3%       4.17        perf-stat.i.MPKI
  2.27e+10            +5.1%  2.385e+10        perf-stat.i.branch-instructions
      0.55            +0.1        0.66        perf-stat.i.branch-miss-rate%
 1.231e+08           +26.3%  1.555e+08        perf-stat.i.branch-misses
 4.045e+08           +24.6%  5.038e+08        perf-stat.i.cache-misses
 1.242e+09           +25.5%  1.559e+09        perf-stat.i.cache-references
    308863            -3.2%     298957        perf-stat.i.context-switches
      4.20            -4.9%       3.99        perf-stat.i.cpi
     49492           +20.0%      59366        perf-stat.i.cpu-migrations
      1181           -19.0%     957.01        perf-stat.i.cycles-between-cache-misses
 1.138e+11            +6.0%  1.207e+11        perf-stat.i.instructions
      0.24            +5.1%       0.25        perf-stat.i.ipc
    116.41 ±120%    +174.4%     319.44 ± 62%  perf-stat.i.major-faults
     23.17           +25.7%      29.11 ±  2%  perf-stat.i.metric.K/sec
   2070571           +27.8%    2646416 ±  2%  perf-stat.i.minor-faults
   2070688           +27.8%    2646736 ±  2%  perf-stat.i.page-faults
      3.56           +17.3%       4.17        perf-stat.overall.MPKI
      0.54            +0.1        0.65        perf-stat.overall.branch-miss-rate%
      4.21            -5.0%       4.00        perf-stat.overall.cpi
      1184           -19.0%     960.02        perf-stat.overall.cycles-between-cache-misses
      0.24            +5.2%       0.25        perf-stat.overall.ipc
 2.229e+10            +5.2%  2.344e+10        perf-stat.ps.branch-instructions
 1.212e+08           +26.3%   1.53e+08        perf-stat.ps.branch-misses
 3.974e+08           +24.4%  4.944e+08        perf-stat.ps.cache-misses
 1.222e+09           +25.5%  1.533e+09        perf-stat.ps.cache-references
    303699            -3.2%     294063        perf-stat.ps.context-switches
 4.706e+11            +0.8%  4.746e+11        perf-stat.ps.cpu-cycles
     48643           +19.9%      58334        perf-stat.ps.cpu-migrations
 1.117e+11            +6.1%  1.186e+11        perf-stat.ps.instructions
    114.44 ±119%    +175.5%     315.27 ± 62%  perf-stat.ps.major-faults
   2036962           +27.8%    2602997 ±  2%  perf-stat.ps.minor-faults
   2037076           +27.8%    2603313 ±  2%  perf-stat.ps.page-faults
 6.832e+12            +5.6%  7.215e+12        perf-stat.total.instructions
     48.07           -48.1        0.00        perf-profile.calltrace.cycles-pp._Fork
     47.54           -47.5        0.00        perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe._Fork
     47.54           -47.5        0.00        perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe._Fork
     47.51           -47.5        0.00        perf-profile.calltrace.cycles-pp.__do_sys_clone.do_syscall_64.entry_SYSCALL_64_after_hwframe._Fork
     47.51           -47.5        0.00        perf-profile.calltrace.cycles-pp.kernel_clone.__do_sys_clone.do_syscall_64.entry_SYSCALL_64_after_hwframe._Fork
     47.04           -47.0        0.00        perf-profile.calltrace.cycles-pp.copy_process.kernel_clone.__do_sys_clone.do_syscall_64.entry_SYSCALL_64_after_hwframe
     44.60           -44.6        0.00        perf-profile.calltrace.cycles-pp.dup_mm.copy_process.kernel_clone.__do_sys_clone.do_syscall_64
     43.93           -43.9        0.00        perf-profile.calltrace.cycles-pp.dup_mmap.dup_mm.copy_process.kernel_clone.__do_sys_clone
     41.28           -41.3        0.00        perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe
     41.27           -41.3        0.00        perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe
     40.48           -40.5        0.00        perf-profile.calltrace.cycles-pp.__x64_sys_exit_group.x64_sys_call.do_syscall_64.entry_SYSCALL_64_after_hwframe
     40.48           -40.5        0.00        perf-profile.calltrace.cycles-pp.do_exit.do_group_exit.__x64_sys_exit_group.x64_sys_call.do_syscall_64
     40.48           -40.5        0.00        perf-profile.calltrace.cycles-pp.do_group_exit.__x64_sys_exit_group.x64_sys_call.do_syscall_64.entry_SYSCALL_64_after_hwframe
     40.48           -40.5        0.00        perf-profile.calltrace.cycles-pp.x64_sys_call.do_syscall_64.entry_SYSCALL_64_after_hwframe
     38.17           -38.2        0.00        perf-profile.calltrace.cycles-pp.exit_mm.do_exit.do_group_exit.__x64_sys_exit_group.x64_sys_call
     38.09           -38.1        0.00        perf-profile.calltrace.cycles-pp.__mmput.exit_mm.do_exit.do_group_exit.__x64_sys_exit_group
     38.07           -38.1        0.00        perf-profile.calltrace.cycles-pp.exit_mmap.__mmput.exit_mm.do_exit.do_group_exit
     36.56           -36.6        0.00        perf-profile.calltrace.cycles-pp.anon_vma_fork.dup_mmap.dup_mm.copy_process.kernel_clone
     30.73           -30.7        0.00        perf-profile.calltrace.cycles-pp.free_pgtables.exit_mmap.__mmput.exit_mm.do_exit
     28.47           -28.5        0.00        perf-profile.calltrace.cycles-pp.unlink_anon_vmas.free_pgtables.exit_mmap.__mmput.exit_mm
     23.42           -23.4        0.00        perf-profile.calltrace.cycles-pp.anon_vma_clone.anon_vma_fork.dup_mmap.dup_mm.copy_process
     15.86           -15.9        0.00        perf-profile.calltrace.cycles-pp.down_write.anon_vma_clone.anon_vma_fork.dup_mmap.dup_mm
     15.74           -15.7        0.00        perf-profile.calltrace.cycles-pp.down_write.unlink_anon_vmas.free_pgtables.exit_mmap.__mmput
     15.57           -15.6        0.00        perf-profile.calltrace.cycles-pp.rwsem_down_write_slowpath.down_write.unlink_anon_vmas.free_pgtables.exit_mmap
     15.53           -15.5        0.00        perf-profile.calltrace.cycles-pp.rwsem_down_write_slowpath.down_write.anon_vma_clone.anon_vma_fork.dup_mmap
     12.70           -12.7        0.00        perf-profile.calltrace.cycles-pp.osq_lock.rwsem_down_write_slowpath.down_write.unlink_anon_vmas.free_pgtables
     12.58           -12.6        0.00        perf-profile.calltrace.cycles-pp.osq_lock.rwsem_down_write_slowpath.down_write.anon_vma_clone.anon_vma_fork
     12.49           -12.5        0.00        perf-profile.calltrace.cycles-pp.down_write.anon_vma_fork.dup_mmap.dup_mm.copy_process
     12.36           -12.4        0.00        perf-profile.calltrace.cycles-pp.rwsem_down_write_slowpath.down_write.anon_vma_fork.dup_mmap.dup_mm
     10.13           -10.1        0.00        perf-profile.calltrace.cycles-pp.osq_lock.rwsem_down_write_slowpath.down_write.anon_vma_fork.dup_mmap
      8.19            -8.2        0.00        perf-profile.calltrace.cycles-pp.__put_anon_vma.unlink_anon_vmas.free_pgtables.exit_mmap.__mmput
      7.76            -7.8        0.00        perf-profile.calltrace.cycles-pp.down_write.__put_anon_vma.unlink_anon_vmas.free_pgtables.exit_mmap
      7.74            -7.7        0.00        perf-profile.calltrace.cycles-pp.rwsem_down_write_slowpath.down_write.__put_anon_vma.unlink_anon_vmas.free_pgtables
      6.72            -6.7        0.00        perf-profile.calltrace.cycles-pp.osq_lock.rwsem_down_write_slowpath.down_write.__put_anon_vma.unlink_anon_vmas
      5.41            -5.4        0.00        perf-profile.calltrace.cycles-pp.prctl
      5.34            -5.3        0.00        perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.prctl
      5.34            -5.3        0.00        perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.prctl
      5.33            -5.3        0.00        perf-profile.calltrace.cycles-pp.__do_sys_prctl.do_syscall_64.entry_SYSCALL_64_after_hwframe.prctl
      5.27            -5.3        0.00        perf-profile.calltrace.cycles-pp.proc_comm_connector.__do_sys_prctl.do_syscall_64.entry_SYSCALL_64_after_hwframe.prctl
      5.25            -5.3        0.00        perf-profile.calltrace.cycles-pp.netlink_broadcast_filtered.proc_comm_connector.__do_sys_prctl.do_syscall_64.entry_SYSCALL_64_after_hwframe
      5.20            -5.2        0.00        perf-profile.calltrace.cycles-pp.do_one_broadcast.netlink_broadcast_filtered.proc_comm_connector.__do_sys_prctl.do_syscall_64
     94.62           -94.6        0.00        perf-profile.children.cycles-pp.do_syscall_64
     94.62           -94.6        0.00        perf-profile.children.cycles-pp.entry_SYSCALL_64_after_hwframe
     52.82           -52.8        0.00        perf-profile.children.cycles-pp.down_write
     51.93           -51.9        0.00        perf-profile.children.cycles-pp.rwsem_down_write_slowpath
     48.12           -48.1        0.00        perf-profile.children.cycles-pp._Fork
     47.60           -47.6        0.00        perf-profile.children.cycles-pp.kernel_clone
     47.59           -47.6        0.00        perf-profile.children.cycles-pp.__do_sys_clone
     47.12           -47.1        0.00        perf-profile.children.cycles-pp.copy_process
     44.67           -44.7        0.00        perf-profile.children.cycles-pp.dup_mm
     44.01           -44.0        0.00        perf-profile.children.cycles-pp.dup_mmap
     42.44           -42.4        0.00        perf-profile.children.cycles-pp.osq_lock
     40.80           -40.8        0.00        perf-profile.children.cycles-pp.x64_sys_call
     40.79           -40.8        0.00        perf-profile.children.cycles-pp.__x64_sys_exit_group
     40.79           -40.8        0.00        perf-profile.children.cycles-pp.do_group_exit
     40.78           -40.8        0.00        perf-profile.children.cycles-pp.do_exit
     38.18           -38.2        0.00        perf-profile.children.cycles-pp.exit_mm
     38.09           -38.1        0.00        perf-profile.children.cycles-pp.__mmput
     38.07           -38.1        0.00        perf-profile.children.cycles-pp.exit_mmap
     36.63           -36.6        0.00        perf-profile.children.cycles-pp.anon_vma_fork
     30.74           -30.7        0.00        perf-profile.children.cycles-pp.free_pgtables
     28.47           -28.5        0.00        perf-profile.children.cycles-pp.unlink_anon_vmas
     23.47           -23.5        0.00        perf-profile.children.cycles-pp.anon_vma_clone
      9.00            -9.0        0.00        perf-profile.children.cycles-pp.netlink_broadcast_filtered
      8.94            -8.9        0.00        perf-profile.children.cycles-pp.do_one_broadcast
      8.19            -8.2        0.00        perf-profile.children.cycles-pp.__put_anon_vma
      7.89            -7.9        0.00        perf-profile.children.cycles-pp.rwsem_spin_on_owner
      5.42            -5.4        0.00        perf-profile.children.cycles-pp.prctl
      5.34            -5.3        0.00        perf-profile.children.cycles-pp.__do_sys_prctl
      5.28            -5.3        0.00        perf-profile.children.cycles-pp.proc_comm_connector
     41.94           -41.9        0.00        perf-profile.self.cycles-pp.osq_lock
      7.80            -7.8        0.00        perf-profile.self.cycles-pp.rwsem_spin_on_owner
      6.29            -6.3        0.00        perf-profile.self.cycles-pp.do_one_broadcast





Disclaimer:
Results have been estimated based on internal Intel analysis and are provided
for informational purposes only. Any difference in system hardware or software
design or configuration may affect actual performance.


-- 
0-DAY CI Kernel Test Service
https://github.com/intel/lkp-tests/wiki



^ permalink raw reply	[flat|nested] only message in thread

only message in thread, other threads:[~2026-02-06 10:04 UTC | newest]

Thread overview: (only message) (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2026-02-06 10:03 [linux-next:master] [mm/rmap] bfc2b13b05: stress-ng.daemon.ops_per_sec 14.0% improvement kernel test robot

This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox