linux-mm.kvack.org archive mirror
 help / color / mirror / Atom feed
From: kernel test robot <oliver.sang@intel.com>
To: Lorenzo Stoakes <lorenzo.stoakes@oracle.com>
Cc: <oe-lkp@lists.linux.dev>, <lkp@intel.com>,
	Andrew Morton <akpm@linux-foundation.org>,
	Suren Baghdasaryan <surenb@google.com>,
	"Liam R. Howlett" <Liam.Howlett@oracle.com>,
	Barry Song <v-songbaohua@oppo.com>,
	"Chris Li" <chriscli@google.com>,
	David Hildenbrand <david@kernel.org>,
	Harry Yoo <harry.yoo@oracle.com>, Jann Horn <jannh@google.com>,
	Michal Hocko <mhocko@suse.com>, Mike Rapoport <rppt@kernel.org>,
	Pedro Falcato <pfalcato@suse.de>, Rik van Riel <riel@surriel.com>,
	Shakeel Butt <shakeel.butt@linux.dev>,
	Vlastimil Babka <vbabka@suse.cz>, <linux-mm@kvack.org>,
	<oliver.sang@intel.com>
Subject: [linux-next:master] [mm/rmap]  bfc2b13b05: stress-ng.daemon.ops_per_sec 14.0% improvement
Date: Fri, 6 Feb 2026 18:03:44 +0800	[thread overview]
Message-ID: <202602061747.855f053f-lkp@intel.com> (raw)



Hello,

kernel test robot noticed a 14.0% improvement of stress-ng.daemon.ops_per_sec on:


commit: bfc2b13b05a1343bb60a85d840fd8956731866c5 ("mm/rmap: allocate anon_vma_chain objects unlocked when possible")
https://git.kernel.org/cgit/linux/kernel/git/next/linux-next.git master


testcase: stress-ng
config: x86_64-rhel-9.4
compiler: gcc-14
test machine: 256 threads 4 sockets INTEL(R) XEON(R) PLATINUM 8592+ (Emerald Rapids) with 256G memory
parameters:

	nr_threads: 100%
	testtime: 60s
	test: daemon
	cpufreq_governor: performance


In addition to that, the commit also has significant impact on the following tests:

+------------------+----------------------------------------------------------------------------------------------+
| testcase: change | stress-ng: stress-ng.netlink-proc.ops_per_sec 25.7% improvement                              |
| test machine     | 192 threads 2 sockets Intel(R) Xeon(R) 6740E  CPU @ 2.4GHz (Sierra Forest) with 256G memory  |
| test parameters  | cpufreq_governor=performance                                                                 |
|                  | nr_threads=100%                                                                              |
|                  | test=netlink-proc                                                                            |
|                  | testtime=60s                                                                                 |
+------------------+----------------------------------------------------------------------------------------------+


Details are as below:
-------------------------------------------------------------------------------------------------->


The kernel config and materials to reproduce are available at:
https://download.01.org/0day-ci/archive/20260206/202602061747.855f053f-lkp@intel.com

=========================================================================================
compiler/cpufreq_governor/kconfig/nr_threads/rootfs/tbox_group/test/testcase/testtime:
  gcc-14/performance/x86_64-rhel-9.4/100%/debian-13-x86_64-20250902.cgz/lkp-emr-2sp1/daemon/stress-ng/60s

commit: 
  85f03a8631 ("mm/mmap_lock: add vma_is_attached() helper")
  bfc2b13b05 ("mm/rmap: allocate anon_vma_chain objects unlocked when possible")

85f03a86318c4172 bfc2b13b05a1343bb60a85d840f 
---------------- --------------------------- 
         %stddev     %change         %stddev
             \          |                \  
      8054 ± 16%     +54.6%      12455 ± 16%  perf-c2c.DRAM.local
      0.13 ±  8%     +10.4%       0.15 ±  3%  perf-sched.sch_delay.avg.ms.[unknown].[unknown].[unknown].[unknown].[unknown]
      0.13 ±  8%     +10.4%       0.15 ±  3%  perf-sched.total_sch_delay.average.ms
     38.55            +6.0       44.57 ±  3%  mpstat.cpu.all.idle%
      0.47            -0.0        0.42 ±  2%  mpstat.cpu.all.irq%
     59.36            -6.1       53.28 ±  2%  mpstat.cpu.all.sys%
      1.21            +0.1        1.30        mpstat.cpu.all.usr%
   2817245           +14.1%    3214286 ±  2%  stress-ng.daemon.ops
     47149           +14.0%      53750 ±  2%  stress-ng.daemon.ops_per_sec
     49.50 ±  3%     +18.2%      58.50 ±  2%  stress-ng.time.percent_of_cpu_this_job_got
   2759543           +13.2%    3123624 ±  2%  stress-ng.time.voluntary_context_switches
      0.15 ± 46%      -0.1        0.05 ± 41%  turbostat.C1%
      3.58 ±  5%      +9.0%       3.90 ±  3%  turbostat.CPU%c1
    647.37            +1.9%     659.97        turbostat.PkgWatt
     28.80 ±  2%      +7.1%      30.83        turbostat.RAMWatt
     25720            +7.4%      27633 ±  2%  sched_debug.cfs_rq:/.avg_vruntime.avg
     25656            +7.6%      27595 ±  2%  sched_debug.cfs_rq:/.zero_vruntime.avg
      2214 ±  3%     +51.1%       3345 ±  6%  sched_debug.cpu.clock.stddev
      2726 ±  2%     +40.8%       3838 ±  5%  sched_debug.cpu.clock_task.stddev
   1420062           +14.0%    1619139 ±  2%  sched_debug.cpu.curr->pid.max
      0.00 ±  3%     +51.0%       0.00 ±  6%  sched_debug.cpu.next_balance.stddev
    182592            +3.7%     189367        meminfo.KReclaimable
    260530 ±  2%      -8.3%     238883        meminfo.Mapped
  18519139           +18.4%   21919824 ±  3%  meminfo.Memused
    310935 ±  4%      +9.2%     339441 ±  2%  meminfo.PageTables
    182592            +3.7%     189367        meminfo.SReclaimable
   6259412 ±  3%     +52.7%    9561140 ±  6%  meminfo.SUnreclaim
   6442004 ±  3%     +51.4%    9750508 ±  6%  meminfo.Slab
  18519311           +18.4%   21919824 ±  3%  meminfo.max_used_kB
  25317761 ±  4%     +14.8%   29072594 ±  3%  numa-numastat.node0.local_node
  25429773 ±  4%     +14.8%   29188651 ±  3%  numa-numastat.node0.numa_hit
  25578903 ±  4%     +15.2%   29473906 ±  3%  numa-numastat.node1.local_node
  25681455 ±  4%     +15.2%   29588761 ±  3%  numa-numastat.node1.numa_hit
  25472065 ±  4%     +19.2%   30373362 ±  2%  numa-numastat.node2.local_node
  25579152 ±  4%     +19.1%   30461365 ±  2%  numa-numastat.node2.numa_hit
  25502240 ±  4%     +18.2%   30147059 ±  2%  numa-numastat.node3.local_node
  25588119 ±  4%     +18.2%   30236863 ±  2%  numa-numastat.node3.numa_hit
    208008            +2.0%     212227        proc-vmstat.nr_anon_pages
   6090914            -1.4%    6006008        proc-vmstat.nr_dirty_background_threshold
  12196722            -1.4%   12026702        proc-vmstat.nr_dirty_threshold
  61229635            -1.4%   60379327        proc-vmstat.nr_free_pages
  61094372            -1.4%   60257849        proc-vmstat.nr_free_pages_blocks
     65170 ±  2%      -8.3%      59741        proc-vmstat.nr_mapped
     77887 ±  4%      +9.1%      85010 ±  3%  proc-vmstat.nr_page_table_pages
     45655            +3.7%      47350        proc-vmstat.nr_slab_reclaimable
   1565161 ±  3%     +52.8%    2390881 ±  6%  proc-vmstat.nr_slab_unreclaimable
 1.023e+08 ±  4%     +16.8%  1.195e+08 ±  2%  proc-vmstat.numa_hit
 1.019e+08 ±  4%     +16.9%  1.191e+08 ±  2%  proc-vmstat.numa_local
 1.109e+08 ±  3%     +17.3%  1.301e+08 ±  2%  proc-vmstat.pgalloc_normal
 1.748e+08 ±  3%     +15.3%  2.015e+08 ±  4%  proc-vmstat.pgfault
 1.064e+08 ±  3%     +16.2%  1.236e+08 ±  2%  proc-vmstat.pgfree
   4897546 ± 28%     +25.4%    6140878 ± 25%  numa-meminfo.node0.MemUsed
     74758 ±  7%     +15.2%      86146 ±  7%  numa-meminfo.node0.PageTables
   1585565 ±  3%     +49.9%    2376337 ±  6%  numa-meminfo.node0.SUnreclaim
   1644278 ±  3%     +48.5%    2442550 ±  5%  numa-meminfo.node0.Slab
    612346 ± 43%     -69.9%     184504 ± 45%  numa-meminfo.node1.Active
    612342 ± 43%     -69.9%     184500 ± 45%  numa-meminfo.node1.Active(anon)
   1674209 ± 85%     -78.1%     366914 ± 42%  numa-meminfo.node1.FilePages
     75408 ±  5%     +15.6%      87173 ±  6%  numa-meminfo.node1.PageTables
   1573523 ±  4%     +52.0%    2391230 ±  7%  numa-meminfo.node1.SUnreclaim
    465316 ± 27%     -72.7%     127017 ± 75%  numa-meminfo.node1.Shmem
   1626985 ±  4%     +49.0%    2424297 ±  7%  numa-meminfo.node1.Slab
   4023313 ± 31%     +58.4%    6371888 ± 27%  numa-meminfo.node2.MemUsed
   1571082 ±  3%     +55.3%    2440472 ±  5%  numa-meminfo.node2.SUnreclaim
   1604428 ±  3%     +55.3%    2492401 ±  6%  numa-meminfo.node2.Slab
   1548894 ±  4%     +54.2%    2388081 ±  6%  numa-meminfo.node3.SUnreclaim
   1586066 ±  4%     +53.0%    2427265 ±  6%  numa-meminfo.node3.Slab
     18610 ±  6%     +16.4%      21662 ±  7%  numa-vmstat.node0.nr_page_table_pages
    396573 ±  3%     +49.9%     594571 ±  6%  numa-vmstat.node0.nr_slab_unreclaimable
  25430185 ±  4%     +14.8%   29188938 ±  3%  numa-vmstat.node0.numa_hit
  25318163 ±  4%     +14.8%   29072880 ±  3%  numa-vmstat.node0.numa_local
    153177 ± 43%     -69.8%      46194 ± 45%  numa-vmstat.node1.nr_active_anon
    418628 ± 85%     -78.1%      91778 ± 42%  numa-vmstat.node1.nr_file_pages
     18770 ±  5%     +16.7%      21911 ±  7%  numa-vmstat.node1.nr_page_table_pages
    116405 ± 28%     -72.7%      31803 ± 75%  numa-vmstat.node1.nr_shmem
    393591 ±  4%     +52.0%     598339 ±  7%  numa-vmstat.node1.nr_slab_unreclaimable
    153176 ± 43%     -69.8%      46192 ± 45%  numa-vmstat.node1.nr_zone_active_anon
  25681692 ±  4%     +15.2%   29589055 ±  3%  numa-vmstat.node1.numa_hit
  25579141 ±  4%     +15.2%   29474199 ±  3%  numa-vmstat.node1.numa_local
    392991 ±  3%     +55.4%     610669 ±  5%  numa-vmstat.node2.nr_slab_unreclaimable
  25579351 ±  4%     +19.1%   30461531 ±  2%  numa-vmstat.node2.numa_hit
  25472265 ±  4%     +19.2%   30373528 ±  2%  numa-vmstat.node2.numa_local
    387461 ±  4%     +54.2%     597608 ±  6%  numa-vmstat.node3.nr_slab_unreclaimable
  25588424 ±  4%     +18.2%   30237199 ±  2%  numa-vmstat.node3.numa_hit
  25502545 ±  4%     +18.2%   30147394 ±  2%  numa-vmstat.node3.numa_local
      5.21 ±  3%     +11.9%       5.83 ±  2%  perf-stat.i.MPKI
      0.47 ±  2%      +0.1        0.55 ±  3%  perf-stat.i.branch-miss-rate%
 1.854e+08           +18.3%  2.194e+08 ±  4%  perf-stat.i.branch-misses
 1.039e+09 ±  2%     +15.1%  1.196e+09 ±  3%  perf-stat.i.cache-misses
 1.747e+09 ±  2%     +15.0%  2.008e+09 ±  3%  perf-stat.i.cache-references
      2.90            -1.9%       2.85        perf-stat.i.cpi
     67064 ±  3%     +10.5%      74100 ±  4%  perf-stat.i.cpu-migrations
    557.14 ±  3%     -12.3%     488.69 ±  3%  perf-stat.i.cycles-between-cache-misses
 1.994e+11            +2.9%  2.051e+11        perf-stat.i.instructions
      0.34            +1.9%       0.35        perf-stat.i.ipc
    252.47 ± 31%    +100.4%     505.88 ± 16%  perf-stat.i.major-faults
     24.13 ±  2%     +14.2%      27.55 ±  4%  perf-stat.i.metric.K/sec
   2906494 ±  3%     +15.6%    3358742 ±  3%  perf-stat.i.minor-faults
   2906746 ±  3%     +15.6%    3359248 ±  3%  perf-stat.i.page-faults
      5.21 ±  3%     +11.9%       5.83 ±  2%  perf-stat.overall.MPKI
      0.47 ±  2%      +0.1        0.55 ±  3%  perf-stat.overall.branch-miss-rate%
      2.90            -1.9%       2.85        perf-stat.overall.cpi
    557.08 ±  3%     -12.3%     488.29 ±  3%  perf-stat.overall.cycles-between-cache-misses
      0.34            +1.9%       0.35        perf-stat.overall.ipc
 1.808e+08           +18.6%  2.144e+08 ±  4%  perf-stat.ps.branch-misses
 1.015e+09 ±  2%     +15.4%  1.171e+09 ±  3%  perf-stat.ps.cache-misses
 1.706e+09           +15.2%  1.966e+09 ±  3%  perf-stat.ps.cache-references
     65460 ±  3%     +10.8%      72507 ±  4%  perf-stat.ps.cpu-migrations
 1.947e+11            +3.1%  2.008e+11        perf-stat.ps.instructions
    246.17 ± 31%    +100.9%     494.53 ± 16%  perf-stat.ps.major-faults
   2838136 ±  3%     +15.8%    3287669 ±  3%  perf-stat.ps.minor-faults
   2838383 ±  3%     +15.8%    3288164 ±  3%  perf-stat.ps.page-faults
 1.169e+13            +2.8%  1.201e+13        perf-stat.total.instructions
     35.08            -3.1       31.96 ±  2%  perf-profile.calltrace.cycles-pp.rwsem_down_write_slowpath.down_write.unlink_anon_vmas.free_pgtables.exit_mmap
     35.27            -3.1       32.16 ±  2%  perf-profile.calltrace.cycles-pp.down_write.unlink_anon_vmas.free_pgtables.exit_mmap.__mmput
     39.70            -2.3       37.35        perf-profile.calltrace.cycles-pp.unlink_anon_vmas.free_pgtables.exit_mmap.__mmput.exit_mm
     39.54            -2.3       37.20        perf-profile.calltrace.cycles-pp.anon_vma_clone.anon_vma_fork.dup_mmap.dup_mm.copy_process
     40.13            -2.2       37.91        perf-profile.calltrace.cycles-pp.anon_vma_fork.dup_mmap.dup_mm.copy_process.kernel_clone
     40.88            -2.2       38.70        perf-profile.calltrace.cycles-pp.free_pgtables.exit_mmap.__mmput.exit_mm.do_exit
     31.32 ±  2%      -2.2       29.16 ±  3%  perf-profile.calltrace.cycles-pp.osq_lock.rwsem_down_write_slowpath.down_write.unlink_anon_vmas.free_pgtables
     33.10            -2.1       31.01 ±  2%  perf-profile.calltrace.cycles-pp.rwsem_down_write_slowpath.down_write.anon_vma_clone.anon_vma_fork.dup_mmap
     33.39            -2.1       31.31 ±  2%  perf-profile.calltrace.cycles-pp.down_write.anon_vma_clone.anon_vma_fork.dup_mmap.dup_mm
     44.86            -1.4       43.49        perf-profile.calltrace.cycles-pp.dup_mmap.dup_mm.copy_process.kernel_clone.__do_sys_clone
     45.45            -1.3       44.18        perf-profile.calltrace.cycles-pp.dup_mm.copy_process.kernel_clone.__do_sys_clone.do_syscall_64
     45.91            -1.2       44.70        perf-profile.calltrace.cycles-pp.copy_process.kernel_clone.__do_sys_clone.do_syscall_64.entry_SYSCALL_64_after_hwframe
     46.43            -1.1       45.30        perf-profile.calltrace.cycles-pp.__do_sys_clone.do_syscall_64.entry_SYSCALL_64_after_hwframe
     46.43            -1.1       45.30        perf-profile.calltrace.cycles-pp.kernel_clone.__do_sys_clone.do_syscall_64.entry_SYSCALL_64_after_hwframe
     94.97            -1.0       93.97        perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe
     94.97            -1.0       93.98        perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe
      3.30 ±  6%      -0.9        2.38 ±  8%  perf-profile.calltrace.cycles-pp.rwsem_spin_on_owner.rwsem_down_write_slowpath.down_write.unlink_anon_vmas.free_pgtables
      3.36 ±  2%      -0.7        2.66 ±  4%  perf-profile.calltrace.cycles-pp.anon_vma_interval_tree_insert.anon_vma_clone.anon_vma_fork.dup_mmap.dup_mm
      2.67 ±  7%      -0.7        1.98 ±  9%  perf-profile.calltrace.cycles-pp.rwsem_spin_on_owner.rwsem_down_write_slowpath.down_write.anon_vma_clone.anon_vma_fork
      0.60            +0.1        0.68        perf-profile.calltrace.cycles-pp.kmem_cache_alloc_noprof.vm_area_dup.dup_mmap.dup_mm.copy_process
      0.65            +0.1        0.74 ±  3%  perf-profile.calltrace.cycles-pp.__vma_start_write.free_pgtables.exit_mmap.__mmput.exit_mm
      0.64            +0.1        0.74        perf-profile.calltrace.cycles-pp.vm_area_dup.dup_mmap.dup_mm.copy_process.kernel_clone
      0.77 ±  2%      +0.1        0.87 ±  3%  perf-profile.calltrace.cycles-pp.__vma_start_write.dup_mmap.dup_mm.copy_process.kernel_clone
      0.58 ±  5%      +0.1        0.68 ±  5%  perf-profile.calltrace.cycles-pp.__memcg_slab_free_hook.kmem_cache_free.unlink_anon_vmas.free_pgtables.exit_mmap
      0.43 ± 44%      +0.2        0.61 ±  3%  perf-profile.calltrace.cycles-pp.mm_init.dup_mm.copy_process.kernel_clone.__do_sys_clone
      1.05 ±  5%      +0.2        1.26 ±  7%  perf-profile.calltrace.cycles-pp.kmem_cache_alloc_noprof.anon_vma_clone.anon_vma_fork.dup_mmap.dup_mm
      1.23 ±  4%      +0.2        1.46 ±  5%  perf-profile.calltrace.cycles-pp.__slab_free.kmem_cache_free.unlink_anon_vmas.free_pgtables.exit_mmap
      0.34 ± 70%      +0.2        0.58 ±  2%  perf-profile.calltrace.cycles-pp.wake_up_new_task.kernel_clone.__do_sys_clone.do_syscall_64.entry_SYSCALL_64_after_hwframe
      1.39 ±  3%      +0.2        1.63 ±  5%  perf-profile.calltrace.cycles-pp.__anon_vma_interval_tree_remove.unlink_anon_vmas.free_pgtables.exit_mmap.__mmput
      1.02 ±  9%      +0.2        1.27 ±  5%  perf-profile.calltrace.cycles-pp.copy_present_ptes.copy_pte_range.copy_p4d_range.copy_page_range.dup_mmap
      1.42 ±  6%      +0.3        1.74 ±  4%  perf-profile.calltrace.cycles-pp.copy_pte_range.copy_p4d_range.copy_page_range.dup_mmap.dup_mm
      1.59 ±  5%      +0.3        1.93 ±  4%  perf-profile.calltrace.cycles-pp.copy_p4d_range.copy_page_range.dup_mmap.dup_mm.copy_process
      1.62 ±  4%      +0.4        1.97 ±  3%  perf-profile.calltrace.cycles-pp.copy_page_range.dup_mmap.dup_mm.copy_process.kernel_clone
      0.18 ±141%      +0.4        0.56 ±  6%  perf-profile.calltrace.cycles-pp.up_write.unlink_anon_vmas.free_pgtables.exit_mmap.__mmput
      2.04 ±  4%      +0.4        2.43 ±  5%  perf-profile.calltrace.cycles-pp.kmem_cache_free.unlink_anon_vmas.free_pgtables.exit_mmap.__mmput
      1.08 ± 12%      +0.5        1.54 ±  9%  perf-profile.calltrace.cycles-pp.folios_put_refs.free_pages_and_swap_cache.__tlb_batch_free_encoded_pages.tlb_finish_mmu.exit_mmap
      1.09 ± 13%      +0.5        1.58 ±  7%  perf-profile.calltrace.cycles-pp.folio_remove_rmap_ptes.zap_present_ptes.zap_pte_range.zap_pmd_range.unmap_page_range
      0.08 ±223%      +0.5        0.57 ±  4%  perf-profile.calltrace.cycles-pp.up_write.anon_vma_clone.anon_vma_fork.dup_mmap.dup_mm
      0.00            +0.6        0.55        perf-profile.calltrace.cycles-pp.kmem_cache_free.exit_mmap.__mmput.exit_mm.do_exit
      1.24 ±  3%      +0.6        1.79 ±  8%  perf-profile.calltrace.cycles-pp.next_uptodate_folio.filemap_map_pages.do_read_fault.do_fault.__handle_mm_fault
      1.45 ± 10%      +0.6        2.03 ±  8%  perf-profile.calltrace.cycles-pp.free_pages_and_swap_cache.__tlb_batch_free_encoded_pages.tlb_finish_mmu.exit_mmap.__mmput
      1.46 ± 10%      +0.6        2.05 ±  7%  perf-profile.calltrace.cycles-pp.__tlb_batch_free_encoded_pages.tlb_finish_mmu.exit_mmap.__mmput.exit_mm
      1.48 ± 10%      +0.6        2.07 ±  8%  perf-profile.calltrace.cycles-pp.tlb_finish_mmu.exit_mmap.__mmput.exit_mm.do_exit
      0.08 ±223%      +0.6        0.71 ±  8%  perf-profile.calltrace.cycles-pp.free_pages_and_swap_cache.__tlb_batch_free_encoded_pages.tlb_flush_mmu.zap_pte_range.zap_pmd_range
      0.08 ±223%      +0.6        0.71 ±  8%  perf-profile.calltrace.cycles-pp.__tlb_batch_free_encoded_pages.tlb_flush_mmu.zap_pte_range.zap_pmd_range.unmap_page_range
      0.08 ±223%      +0.6        0.71 ±  8%  perf-profile.calltrace.cycles-pp.tlb_flush_mmu.zap_pte_range.zap_pmd_range.unmap_page_range.unmap_vmas
      1.85 ±  3%      +0.7        2.58 ±  7%  perf-profile.calltrace.cycles-pp.filemap_map_pages.do_read_fault.do_fault.__handle_mm_fault.handle_mm_fault
      1.94 ±  3%      +0.8        2.70 ±  7%  perf-profile.calltrace.cycles-pp.do_fault.__handle_mm_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault
      1.93 ±  3%      +0.8        2.69 ±  7%  perf-profile.calltrace.cycles-pp.do_read_fault.do_fault.__handle_mm_fault.handle_mm_fault.do_user_addr_fault
      2.27 ±  3%      +0.8        3.08 ±  6%  perf-profile.calltrace.cycles-pp.__handle_mm_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault
      2.34 ±  3%      +0.8        3.16 ±  6%  perf-profile.calltrace.cycles-pp.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault
      2.68 ±  3%      +0.9        3.57 ±  6%  perf-profile.calltrace.cycles-pp.exc_page_fault.asm_exc_page_fault
      2.66 ±  3%      +0.9        3.56 ±  6%  perf-profile.calltrace.cycles-pp.do_user_addr_fault.exc_page_fault.asm_exc_page_fault
      2.88 ±  3%      +0.9        3.80 ±  6%  perf-profile.calltrace.cycles-pp.asm_exc_page_fault
      2.56 ±  7%      +1.0        3.51 ±  5%  perf-profile.calltrace.cycles-pp.zap_present_ptes.zap_pte_range.zap_pmd_range.unmap_page_range.unmap_vmas
      3.40 ±  6%      +1.3        4.72 ±  5%  perf-profile.calltrace.cycles-pp.zap_pte_range.zap_pmd_range.unmap_page_range.unmap_vmas.exit_mmap
      3.45 ±  6%      +1.3        4.78 ±  5%  perf-profile.calltrace.cycles-pp.zap_pmd_range.unmap_page_range.unmap_vmas.exit_mmap.__mmput
      3.50 ±  6%      +1.3        4.85 ±  5%  perf-profile.calltrace.cycles-pp.unmap_page_range.unmap_vmas.exit_mmap.__mmput.exit_mm
      3.63 ±  6%      +1.4        4.98 ±  5%  perf-profile.calltrace.cycles-pp.unmap_vmas.exit_mmap.__mmput.exit_mm.do_exit
     68.28            -5.1       63.15 ±  2%  perf-profile.children.cycles-pp.rwsem_down_write_slowpath
     68.93            -5.1       63.84 ±  2%  perf-profile.children.cycles-pp.down_write
     61.34 ±  2%      -3.5       57.82 ±  3%  perf-profile.children.cycles-pp.osq_lock
     39.70            -2.3       37.37        perf-profile.children.cycles-pp.unlink_anon_vmas
     39.56            -2.3       37.23        perf-profile.children.cycles-pp.anon_vma_clone
     40.14            -2.2       37.92        perf-profile.children.cycles-pp.anon_vma_fork
     40.88            -2.2       38.71        perf-profile.children.cycles-pp.free_pgtables
      6.01 ±  6%      -1.6        4.41 ±  9%  perf-profile.children.cycles-pp.rwsem_spin_on_owner
     44.93            -1.4       43.56        perf-profile.children.cycles-pp.dup_mmap
     45.45            -1.3       44.18        perf-profile.children.cycles-pp.dup_mm
     45.91            -1.2       44.70        perf-profile.children.cycles-pp.copy_process
     46.43            -1.1       45.30        perf-profile.children.cycles-pp.__do_sys_clone
     46.43            -1.1       45.30        perf-profile.children.cycles-pp.kernel_clone
     95.09            -1.0       94.09        perf-profile.children.cycles-pp.do_syscall_64
     95.09            -1.0       94.09        perf-profile.children.cycles-pp.entry_SYSCALL_64_after_hwframe
      3.37 ±  2%      -0.7        2.67 ±  4%  perf-profile.children.cycles-pp.anon_vma_interval_tree_insert
      0.05            +0.0        0.06        perf-profile.children.cycles-pp.autoremove_wake_function
      0.05            +0.0        0.06        perf-profile.children.cycles-pp.lru_add_drain_cpu
      0.05            +0.0        0.06        perf-profile.children.cycles-pp.mas_dup_alloc
      0.05            +0.0        0.06        perf-profile.children.cycles-pp.sched_move_task
      0.07            +0.0        0.08        perf-profile.children.cycles-pp.process_one_work
      0.06            +0.0        0.07        perf-profile.children.cycles-pp.__wake_up_common
      0.06            +0.0        0.07        perf-profile.children.cycles-pp.delayed_vfree_work
      0.06            +0.0        0.07        perf-profile.children.cycles-pp.vfree
      0.10            +0.0        0.11 ±  3%  perf-profile.children.cycles-pp.__mt_dup
      0.07            +0.0        0.08 ±  4%  perf-profile.children.cycles-pp.anon_pipe_write
      0.05            +0.0        0.06 ±  6%  perf-profile.children.cycles-pp.folio_batch_move_lru
      0.05            +0.0        0.06 ±  6%  perf-profile.children.cycles-pp.lru_add_drain
      0.06            +0.0        0.07 ±  5%  perf-profile.children.cycles-pp.__wake_up_sync_key
      0.08            +0.0        0.09 ±  4%  perf-profile.children.cycles-pp.__wp_page_copy_user
      0.08            +0.0        0.09 ±  5%  perf-profile.children.cycles-pp.__put_user_4
      0.07 ±  9%      +0.0        0.08 ±  5%  perf-profile.children.cycles-pp.__rmqueue_pcplist
      0.08 ±  5%      +0.0        0.10 ±  4%  perf-profile.children.cycles-pp.vm_normal_page
      0.11 ±  4%      +0.0        0.12 ±  3%  perf-profile.children.cycles-pp.vfs_write
      0.15 ±  2%      +0.0        0.16 ±  2%  perf-profile.children.cycles-pp.vma_mark_detached
      0.08 ±  6%      +0.0        0.09 ±  4%  perf-profile.children.cycles-pp.do_wp_page
      0.09 ±  5%      +0.0        0.10 ±  3%  perf-profile.children.cycles-pp.kernel_init_pages
      0.12 ±  3%      +0.0        0.14 ±  2%  perf-profile.children.cycles-pp.worker_thread
      0.08 ±  5%      +0.0        0.10 ±  3%  perf-profile.children.cycles-pp.__vmalloc_node_noprof
      0.08 ±  5%      +0.0        0.10 ±  3%  perf-profile.children.cycles-pp.__vmalloc_node_range_noprof
      0.12            +0.0        0.14 ±  3%  perf-profile.children.cycles-pp.acct_collect
      0.12 ±  4%      +0.0        0.13 ±  2%  perf-profile.children.cycles-pp.ksys_write
      0.05 ± 13%      +0.0        0.07 ± 14%  perf-profile.children.cycles-pp.__pmd_alloc
      0.10 ±  3%      +0.0        0.12 ±  5%  perf-profile.children.cycles-pp.rb_next
      0.17 ±  4%      +0.0        0.19 ±  2%  perf-profile.children.cycles-pp.sched_balance_find_src_group
      0.11 ±  6%      +0.0        0.13 ±  3%  perf-profile.children.cycles-pp.free_frozen_page_commit
      0.09 ±  7%      +0.0        0.11 ±  4%  perf-profile.children.cycles-pp.prep_new_page
      0.10 ±  4%      +0.0        0.12 ±  4%  perf-profile.children.cycles-pp.rmqueue
      0.16 ±  2%      +0.0        0.18 ±  3%  perf-profile.children.cycles-pp.kthread
      0.08 ±  4%      +0.0        0.10 ±  4%  perf-profile.children.cycles-pp._copy_from_user
      0.23 ±  5%      +0.0        0.26 ±  2%  perf-profile.children.cycles-pp.sched_balance_rq
      0.17 ±  2%      +0.0        0.20 ±  3%  perf-profile.children.cycles-pp.sync_regs
      0.08 ±  4%      +0.0        0.11 ±  4%  perf-profile.children.cycles-pp.__x64_sys_rt_sigprocmask
      0.12 ±  6%      +0.0        0.14 ±  4%  perf-profile.children.cycles-pp.__memcg_kmem_charge_page
      0.21            +0.0        0.24 ±  3%  perf-profile.children.cycles-pp.mas_next_slot
      0.09 ±  4%      +0.0        0.12 ±  5%  perf-profile.children.cycles-pp.__put_anon_vma
      0.26            +0.0        0.28 ±  2%  perf-profile.children.cycles-pp.mas_find
      0.16 ±  6%      +0.0        0.20 ±  2%  perf-profile.children.cycles-pp.__free_frozen_pages
      0.16 ±  2%      +0.0        0.20 ±  2%  perf-profile.children.cycles-pp.__pte_offset_map_lock
      0.18 ±  7%      +0.0        0.22 ±  2%  perf-profile.children.cycles-pp.tlb_remove_table_rcu
      0.23            +0.0        0.26 ±  2%  perf-profile.children.cycles-pp.lock_vma_under_rcu
      0.11 ±  6%      +0.0        0.14 ± 10%  perf-profile.children.cycles-pp.__put_partials
      0.24 ±  2%      +0.0        0.27 ±  3%  perf-profile.children.cycles-pp.__pi_memset
      0.22 ±  5%      +0.0        0.25 ±  2%  perf-profile.children.cycles-pp.wp_page_copy
      0.16 ±  3%      +0.0        0.20 ±  3%  perf-profile.children.cycles-pp.remove_vma
      0.14 ±  3%      +0.0        0.18 ±  5%  perf-profile.children.cycles-pp.fput
      0.09 ±  4%      +0.0        0.13 ±  7%  perf-profile.children.cycles-pp.getname_flags
      0.20 ±  6%      +0.0        0.24        perf-profile.children.cycles-pp.pte_alloc_one
      0.09 ±  4%      +0.0        0.13 ±  7%  perf-profile.children.cycles-pp.strncpy_from_user
      0.20 ±  4%      +0.0        0.24 ±  3%  perf-profile.children.cycles-pp.get_page_from_freelist
      0.10 ±  4%      +0.0        0.13 ±  7%  perf-profile.children.cycles-pp.user_path_at
      0.10 ±  3%      +0.0        0.14 ±  7%  perf-profile.children.cycles-pp.__x64_sys_chdir
      0.22 ±  3%      +0.0        0.26 ±  3%  perf-profile.children.cycles-pp.native_irq_return_iret
      0.29            +0.0        0.32 ±  2%  perf-profile.children.cycles-pp.__percpu_counter_sum
      0.36 ±  2%      +0.0        0.40 ±  2%  perf-profile.children.cycles-pp.__kmem_cache_alloc_bulk
      0.33 ±  3%      +0.0        0.37 ±  2%  perf-profile.children.cycles-pp.__account_obj_stock
      0.35 ±  2%      +0.0        0.40        perf-profile.children.cycles-pp.__pcs_replace_empty_main
      0.18 ±  2%      +0.0        0.22 ±  4%  perf-profile.children.cycles-pp.do_task_dead
      0.24 ±  3%      +0.0        0.29 ±  2%  perf-profile.children.cycles-pp.schedule_tail
      1.28            +0.0        1.32        perf-profile.children.cycles-pp.sysvec_apic_timer_interrupt
      0.37            +0.0        0.41 ±  3%  perf-profile.children.cycles-pp.update_sg_wakeup_stats
      0.25 ±  2%      +0.0        0.30 ±  3%  perf-profile.children.cycles-pp.__percpu_counter_init_many
      0.18 ±  6%      +0.0        0.24 ±  8%  perf-profile.children.cycles-pp.get_partial_node
      0.00            +0.1        0.05        perf-profile.children.cycles-pp.vm_area_init_from
      1.32            +0.1        1.37        perf-profile.children.cycles-pp.asm_sysvec_apic_timer_interrupt
      0.40 ±  3%      +0.1        0.45 ±  3%  perf-profile.children.cycles-pp.rcu_core
      0.38 ±  3%      +0.1        0.43 ±  2%  perf-profile.children.cycles-pp.rcu_do_batch
      0.44            +0.1        0.50 ±  2%  perf-profile.children.cycles-pp.mas_store
      0.30 ±  3%      +0.1        0.35        perf-profile.children.cycles-pp.vfs_read
      0.25 ±  3%      +0.1        0.30 ±  2%  perf-profile.children.cycles-pp.anon_pipe_read
      0.33            +0.1        0.38        perf-profile.children.cycles-pp.finish_task_switch
      0.30 ±  2%      +0.1        0.36        perf-profile.children.cycles-pp.ksys_read
      0.38 ±  2%      +0.1        0.43 ±  3%  perf-profile.children.cycles-pp.sched_balance_find_dst_group
      0.36 ±  2%      +0.1        0.42 ±  2%  perf-profile.children.cycles-pp.__mmdrop
      0.37 ±  3%      +0.1        0.42 ±  2%  perf-profile.children.cycles-pp.refill_obj_stock
      0.57 ±  2%      +0.1        0.63 ±  2%  perf-profile.children.cycles-pp.__irq_exit_rcu
      0.36            +0.1        0.42 ±  4%  perf-profile.children.cycles-pp.unlink_file_vma_batch_add
      0.44 ±  3%      +0.1        0.50 ±  2%  perf-profile.children.cycles-pp._raw_spin_lock
      0.32 ±  5%      +0.1        0.38 ±  2%  perf-profile.children.cycles-pp.alloc_pages_noprof
      0.60 ±  3%      +0.1        0.66 ±  3%  perf-profile.children.cycles-pp.handle_softirqs
      0.48            +0.1        0.54 ±  3%  perf-profile.children.cycles-pp.select_task_rq_fair
      0.41 ±  5%      +0.1        0.48 ±  4%  perf-profile.children.cycles-pp.__rb_insert_augmented
      0.38 ±  2%      +0.1        0.44 ±  3%  perf-profile.children.cycles-pp.unlink_file_vma_batch_process
      0.34 ±  5%      +0.1        0.41 ±  2%  perf-profile.children.cycles-pp.alloc_pages_mpol
      0.33 ±  5%      +0.1        0.40 ±  2%  perf-profile.children.cycles-pp.__alloc_frozen_pages_noprof
      0.42            +0.1        0.49 ±  2%  perf-profile.children.cycles-pp.ret_from_fork
      0.42            +0.1        0.49 ±  2%  perf-profile.children.cycles-pp.ret_from_fork_asm
      0.50            +0.1        0.58 ±  2%  perf-profile.children.cycles-pp.wake_up_new_task
      0.34 ±  8%      +0.1        0.42 ±  8%  perf-profile.children.cycles-pp._raw_spin_lock_irqsave
      0.20 ±  3%      +0.1        0.28 ±  6%  perf-profile.children.cycles-pp.folio_add_file_rmap_ptes
      0.35 ±  5%      +0.1        0.42 ±  9%  perf-profile.children.cycles-pp.native_queued_spin_lock_slowpath
      0.46 ±  2%      +0.1        0.55 ±  3%  perf-profile.children.cycles-pp.pcpu_alloc_noprof
      0.31 ±  3%      +0.1        0.40 ±  5%  perf-profile.children.cycles-pp.set_pte_range
      0.65            +0.1        0.74        perf-profile.children.cycles-pp.vm_area_dup
      0.74 ±  2%      +0.1        0.84 ±  3%  perf-profile.children.cycles-pp.__vma_enter_locked
      0.51 ±  2%      +0.1        0.61 ±  3%  perf-profile.children.cycles-pp.mm_init
      0.97 ±  4%      +0.2        1.14 ±  3%  perf-profile.children.cycles-pp.__memcg_slab_free_hook
      1.43 ±  2%      +0.2        1.62 ±  3%  perf-profile.children.cycles-pp.__vma_start_write
      1.40 ±  3%      +0.2        1.64 ±  5%  perf-profile.children.cycles-pp.__anon_vma_interval_tree_remove
      1.04 ±  9%      +0.2        1.28 ±  4%  perf-profile.children.cycles-pp.copy_present_ptes
      1.46 ±  3%      +0.2        1.70 ±  5%  perf-profile.children.cycles-pp.__slab_free
      1.20 ±  7%      +0.3        1.45 ±  4%  perf-profile.children.cycles-pp.up_write
      0.42 ± 12%      +0.3        0.71 ±  8%  perf-profile.children.cycles-pp.tlb_flush_mmu
      1.44 ±  6%      +0.3        1.75 ±  4%  perf-profile.children.cycles-pp.copy_pte_range
      1.60 ±  4%      +0.3        1.94 ±  4%  perf-profile.children.cycles-pp.copy_p4d_range
      1.63 ±  4%      +0.3        1.97 ±  4%  perf-profile.children.cycles-pp.copy_page_range
      2.10 ±  3%      +0.4        2.46 ±  3%  perf-profile.children.cycles-pp.kmem_cache_alloc_noprof
      1.11 ± 13%      +0.5        1.60 ±  7%  perf-profile.children.cycles-pp.folio_remove_rmap_ptes
      2.82 ±  3%      +0.5        3.32 ±  4%  perf-profile.children.cycles-pp.kmem_cache_free
      1.32 ±  3%      +0.6        1.90 ±  8%  perf-profile.children.cycles-pp.next_uptodate_folio
      1.48 ± 10%      +0.6        2.07 ±  8%  perf-profile.children.cycles-pp.tlb_finish_mmu
      1.41 ± 12%      +0.7        2.10 ±  6%  perf-profile.children.cycles-pp.folios_put_refs
      1.98 ±  3%      +0.8        2.76 ±  7%  perf-profile.children.cycles-pp.filemap_map_pages
      2.05 ±  3%      +0.8        2.86 ±  7%  perf-profile.children.cycles-pp.do_read_fault
      2.05 ±  3%      +0.8        2.87 ±  7%  perf-profile.children.cycles-pp.do_fault
      2.43 ±  3%      +0.9        3.31 ±  6%  perf-profile.children.cycles-pp.__handle_mm_fault
      1.88 ± 10%      +0.9        2.76 ±  6%  perf-profile.children.cycles-pp.__tlb_batch_free_encoded_pages
      1.88 ± 10%      +0.9        2.76 ±  6%  perf-profile.children.cycles-pp.free_pages_and_swap_cache
      2.51 ±  3%      +0.9        3.40 ±  6%  perf-profile.children.cycles-pp.handle_mm_fault
      2.88 ±  3%      +1.0        3.84 ±  6%  perf-profile.children.cycles-pp.exc_page_fault
      2.87 ±  3%      +1.0        3.83 ±  6%  perf-profile.children.cycles-pp.do_user_addr_fault
      2.61 ±  7%      +1.0        3.58 ±  5%  perf-profile.children.cycles-pp.zap_present_ptes
      3.12 ±  3%      +1.0        4.11 ±  6%  perf-profile.children.cycles-pp.asm_exc_page_fault
      3.41 ±  6%      +1.3        4.74 ±  5%  perf-profile.children.cycles-pp.zap_pte_range
      3.45 ±  6%      +1.3        4.79 ±  5%  perf-profile.children.cycles-pp.zap_pmd_range
      3.51 ±  6%      +1.3        4.85 ±  5%  perf-profile.children.cycles-pp.unmap_page_range
      3.63 ±  6%      +1.4        4.99 ±  5%  perf-profile.children.cycles-pp.unmap_vmas
     60.76 ±  2%      -3.5       57.25 ±  3%  perf-profile.self.cycles-pp.osq_lock
      5.95 ±  6%      -1.6        4.36 ±  9%  perf-profile.self.cycles-pp.rwsem_spin_on_owner
      3.32 ±  2%      -0.7        2.62 ±  4%  perf-profile.self.cycles-pp.anon_vma_interval_tree_insert
      0.05            +0.0        0.06        perf-profile.self.cycles-pp._find_next_and_bit
      0.06 ±  7%      +0.0        0.08 ±  6%  perf-profile.self.cycles-pp.vm_normal_page
      0.09 ±  5%      +0.0        0.10        perf-profile.self.cycles-pp.kernel_init_pages
      0.07 ±  5%      +0.0        0.08 ±  5%  perf-profile.self.cycles-pp.__memcg_kmem_charge_page
      0.14 ±  3%      +0.0        0.16 ±  2%  perf-profile.self.cycles-pp.vma_mark_detached
      0.10 ±  3%      +0.0        0.12 ±  4%  perf-profile.self.cycles-pp.set_pte_range
      0.07 ±  7%      +0.0        0.09 ± 11%  perf-profile.self.cycles-pp.do_user_addr_fault
      0.15 ±  3%      +0.0        0.18 ±  2%  perf-profile.self.cycles-pp.lock_vma_under_rcu
      0.17 ±  3%      +0.0        0.19 ±  3%  perf-profile.self.cycles-pp.sync_regs
      0.24 ±  3%      +0.0        0.27 ±  3%  perf-profile.self.cycles-pp.__account_obj_stock
      0.14 ±  3%      +0.0        0.17 ±  6%  perf-profile.self.cycles-pp.get_partial_node
      0.18 ±  2%      +0.0        0.20 ±  3%  perf-profile.self.cycles-pp.refill_obj_stock
      0.22 ±  2%      +0.0        0.24 ±  3%  perf-profile.self.cycles-pp.__percpu_counter_sum
      0.25            +0.0        0.28        perf-profile.self.cycles-pp.__kmem_cache_alloc_bulk
      0.22 ±  3%      +0.0        0.25 ±  3%  perf-profile.self.cycles-pp.__pi_memset
      0.14 ±  3%      +0.0        0.18 ±  4%  perf-profile.self.cycles-pp.fput
      0.30            +0.0        0.34        perf-profile.self.cycles-pp.mas_store
      0.22 ±  3%      +0.0        0.26 ±  3%  perf-profile.self.cycles-pp.native_irq_return_iret
      0.30            +0.0        0.33 ±  4%  perf-profile.self.cycles-pp.update_sg_wakeup_stats
      0.37 ±  4%      +0.0        0.41 ±  4%  perf-profile.self.cycles-pp.rwsem_down_write_slowpath
      0.28 ±  6%      +0.0        0.32 ±  4%  perf-profile.self.cycles-pp.zap_pte_range
      0.16 ±  3%      +0.0        0.20 ±  3%  perf-profile.self.cycles-pp.pcpu_alloc_noprof
      0.28 ±  3%      +0.0        0.33 ±  6%  perf-profile.self.cycles-pp.unlink_anon_vmas
      0.27            +0.0        0.31 ±  5%  perf-profile.self.cycles-pp.kmem_cache_free
      0.38            +0.0        0.43        perf-profile.self.cycles-pp._raw_spin_lock
      0.00            +0.1        0.05        perf-profile.self.cycles-pp.cpu_util
      0.00            +0.1        0.05        perf-profile.self.cycles-pp.free_pcppages_bulk
      0.00            +0.1        0.05        perf-profile.self.cycles-pp.mas_find
      0.40 ±  5%      +0.1        0.46 ±  4%  perf-profile.self.cycles-pp.__rb_insert_augmented
      0.20 ±  3%      +0.1        0.27 ±  6%  perf-profile.self.cycles-pp.folio_add_file_rmap_ptes
      0.35 ±  5%      +0.1        0.42 ±  9%  perf-profile.self.cycles-pp.native_queued_spin_lock_slowpath
      0.68 ±  2%      +0.1        0.78 ±  2%  perf-profile.self.cycles-pp.__vma_start_write
      0.33 ±  3%      +0.1        0.43 ±  7%  perf-profile.self.cycles-pp.filemap_map_pages
      0.54            +0.1        0.64 ±  3%  perf-profile.self.cycles-pp.dup_mmap
      0.71 ±  2%      +0.1        0.81 ±  4%  perf-profile.self.cycles-pp.__vma_enter_locked
      0.63 ±  5%      +0.1        0.74 ±  5%  perf-profile.self.cycles-pp.__memcg_slab_free_hook
      0.46 ±  5%      +0.2        0.64 ±  5%  perf-profile.self.cycles-pp.free_pages_and_swap_cache
      1.28 ±  4%      +0.2        1.51 ±  5%  perf-profile.self.cycles-pp.__anon_vma_interval_tree_remove
      0.99 ±  9%      +0.2        1.22 ±  5%  perf-profile.self.cycles-pp.copy_present_ptes
      1.43 ±  3%      +0.2        1.67 ±  5%  perf-profile.self.cycles-pp.__slab_free
      0.30 ±  2%      +0.3        0.57 ±  6%  perf-profile.self.cycles-pp.kmem_cache_alloc_noprof
      0.96 ±  3%      +0.3        1.24 ±  2%  perf-profile.self.cycles-pp.up_write
      1.36 ±  3%      +0.5        1.82 ±  6%  perf-profile.self.cycles-pp.zap_present_ptes
      1.08 ± 13%      +0.5        1.56 ±  7%  perf-profile.self.cycles-pp.folio_remove_rmap_ptes
      1.24 ±  3%      +0.6        1.81 ±  8%  perf-profile.self.cycles-pp.next_uptodate_folio
      1.33 ± 12%      +0.7        2.01 ±  6%  perf-profile.self.cycles-pp.folios_put_refs


***************************************************************************************************

=========================================================================================
compiler/cpufreq_governor/kconfig/nr_threads/rootfs/tbox_group/test/testcase/testtime:
  gcc-14/performance/x86_64-rhel-9.4/100%/debian-13-x86_64-20250902.cgz/lkp-srf-2sp3/netlink-proc/stress-ng/60s

commit: 
  85f03a8631 ("mm/mmap_lock: add vma_is_attached() helper")
  bfc2b13b05 ("mm/rmap: allocate anon_vma_chain objects unlocked when possible")

85f03a86318c4172 bfc2b13b05a1343bb60a85d840f 
---------------- --------------------------- 
         %stddev     %change         %stddev
             \          |                \  
   9243133            -5.2%    8760646        cpuidle..usage
    346361            -6.4%     324082        meminfo.Mapped
   1529351            +6.6%    1630248        meminfo.Shmem
      0.35 ±  2%      +0.1        0.44        mpstat.cpu.all.soft%
      0.98            +0.1        1.09        mpstat.cpu.all.usr%
    298266            -2.9%     289575        vmstat.system.cs
    577247            +2.4%     590966        vmstat.system.in
   7105876            +7.0%    7601087        turbostat.NMI
    332.10            +2.8%     341.32        turbostat.PkgWatt
     25.50            +6.4%      27.14        turbostat.RAMWatt
  37440414 ±  9%     +26.7%   47453343 ±  7%  numa-numastat.node0.local_node
  37552968 ±  9%     +26.6%   47551257 ±  7%  numa-numastat.node0.numa_hit
  35626288 ±  2%     +30.3%   46417467 ±  2%  numa-numastat.node1.local_node
  35715809 ±  2%     +30.3%   46520561 ±  2%  numa-numastat.node1.numa_hit
  37554292 ±  9%     +26.6%   47552046 ±  7%  numa-vmstat.node0.numa_hit
  37441738 ±  9%     +26.7%   47454132 ±  7%  numa-vmstat.node0.numa_local
  35716707 ±  2%     +30.2%   46520870 ±  2%  numa-vmstat.node1.numa_hit
  35627089 ±  2%     +30.3%   46417777 ±  2%  numa-vmstat.node1.numa_local
     18927 ± 22%     +25.0%      23668 ±  2%  perf-c2c.DRAM.local
    147147 ± 19%     +29.2%     190098 ±  5%  perf-c2c.DRAM.remote
    263763 ± 20%     +48.3%     391131 ±  5%  perf-c2c.HITM.local
    118030 ± 19%     +30.4%     153956 ±  5%  perf-c2c.HITM.remote
    381793 ± 20%     +42.8%     545088 ±  4%  perf-c2c.HITM.total
    341231           +25.7%     429025        stress-ng.netlink-proc.ops
      5697           +25.7%       7159        stress-ng.netlink-proc.ops_per_sec
    726075           +27.3%     924191        stress-ng.time.involuntary_context_switches
      7082 ±119%    +175.4%      19501 ± 62%  stress-ng.time.major_page_faults
 1.257e+08           +27.8%  1.607e+08 ±  2%  stress-ng.time.minor_page_faults
     42.26           +26.2%      53.32        stress-ng.time.user_time
   6952244           -12.2%    6106739        stress-ng.time.voluntary_context_switches
      0.09           +11.7%       0.10        perf-sched.sch_delay.avg.ms.[unknown].[unknown].[unknown].[unknown].[unknown]
     76.37 ± 18%     -20.8%      60.45 ± 12%  perf-sched.sch_delay.max.ms.[unknown].[unknown].[unknown].[unknown].[unknown]
      0.09           +11.7%       0.10        perf-sched.total_sch_delay.average.ms
     76.37 ± 18%     -20.8%      60.45 ± 12%  perf-sched.total_sch_delay.max.ms
      2571 ± 16%     +44.4%       3714 ± 11%  perf-sched.total_wait_and_delay.max.ms
      2571 ± 16%     +44.4%       3714 ± 11%  perf-sched.total_wait_time.max.ms
      2571 ± 16%     +44.4%       3714 ± 11%  perf-sched.wait_and_delay.max.ms.[unknown].[unknown].[unknown].[unknown].[unknown]
      2571 ± 16%     +44.4%       3714 ± 11%  perf-sched.wait_time.max.ms.[unknown].[unknown].[unknown].[unknown].[unknown]
    601141            +4.8%     629898        proc-vmstat.nr_active_anon
   1307815            +1.9%    1333006        proc-vmstat.nr_file_pages
     86624            -6.4%      81044        proc-vmstat.nr_mapped
    382407            +6.6%     407598        proc-vmstat.nr_shmem
    601141            +4.8%     629898        proc-vmstat.nr_zone_active_anon
  73272349 ±  4%     +28.4%   94075287 ±  2%  proc-vmstat.numa_hit
  73070343 ±  4%     +28.5%   93874463 ±  2%  proc-vmstat.numa_local
  76707662 ±  4%     +28.2%   98366347 ±  2%  proc-vmstat.pgalloc_normal
 1.262e+08           +27.8%  1.613e+08 ±  2%  proc-vmstat.pgfault
  75478722 ±  4%     +28.6%   97070551 ±  2%  proc-vmstat.pgfree
   9954110 ±  2%     +37.2%   13659353 ±  4%  proc-vmstat.pgreuse
    806263           +76.7%    1424830 ± 91%  sched_debug.cfs_rq:/.left_deadline.max
    126768 ± 18%     +47.1%     186420 ± 34%  sched_debug.cfs_rq:/.left_deadline.stddev
    806262           +76.7%    1424828 ± 91%  sched_debug.cfs_rq:/.left_vruntime.max
    126766 ± 18%     +47.1%     186417 ± 34%  sched_debug.cfs_rq:/.left_vruntime.stddev
    806262           +76.7%    1424828 ± 91%  sched_debug.cfs_rq:/.right_vruntime.max
    126766 ± 18%     +47.1%     186418 ± 34%  sched_debug.cfs_rq:/.right_vruntime.stddev
      1083 ±  8%     +12.3%       1217 ±  4%  sched_debug.cfs_rq:/.runnable_avg.max
      1054 ±  7%     +12.4%       1185 ±  4%  sched_debug.cfs_rq:/.util_avg.max
    767399 ±  3%     +18.5%     909720 ±  9%  sched_debug.cpu.curr->pid.avg
   1022954           +25.4%    1283282        sched_debug.cpu.curr->pid.max
    438214 ±  3%     +30.7%     572602 ±  6%  sched_debug.cpu.curr->pid.stddev
      3.56           +17.3%       4.17        perf-stat.i.MPKI
  2.27e+10            +5.1%  2.385e+10        perf-stat.i.branch-instructions
      0.55            +0.1        0.66        perf-stat.i.branch-miss-rate%
 1.231e+08           +26.3%  1.555e+08        perf-stat.i.branch-misses
 4.045e+08           +24.6%  5.038e+08        perf-stat.i.cache-misses
 1.242e+09           +25.5%  1.559e+09        perf-stat.i.cache-references
    308863            -3.2%     298957        perf-stat.i.context-switches
      4.20            -4.9%       3.99        perf-stat.i.cpi
     49492           +20.0%      59366        perf-stat.i.cpu-migrations
      1181           -19.0%     957.01        perf-stat.i.cycles-between-cache-misses
 1.138e+11            +6.0%  1.207e+11        perf-stat.i.instructions
      0.24            +5.1%       0.25        perf-stat.i.ipc
    116.41 ±120%    +174.4%     319.44 ± 62%  perf-stat.i.major-faults
     23.17           +25.7%      29.11 ±  2%  perf-stat.i.metric.K/sec
   2070571           +27.8%    2646416 ±  2%  perf-stat.i.minor-faults
   2070688           +27.8%    2646736 ±  2%  perf-stat.i.page-faults
      3.56           +17.3%       4.17        perf-stat.overall.MPKI
      0.54            +0.1        0.65        perf-stat.overall.branch-miss-rate%
      4.21            -5.0%       4.00        perf-stat.overall.cpi
      1184           -19.0%     960.02        perf-stat.overall.cycles-between-cache-misses
      0.24            +5.2%       0.25        perf-stat.overall.ipc
 2.229e+10            +5.2%  2.344e+10        perf-stat.ps.branch-instructions
 1.212e+08           +26.3%   1.53e+08        perf-stat.ps.branch-misses
 3.974e+08           +24.4%  4.944e+08        perf-stat.ps.cache-misses
 1.222e+09           +25.5%  1.533e+09        perf-stat.ps.cache-references
    303699            -3.2%     294063        perf-stat.ps.context-switches
 4.706e+11            +0.8%  4.746e+11        perf-stat.ps.cpu-cycles
     48643           +19.9%      58334        perf-stat.ps.cpu-migrations
 1.117e+11            +6.1%  1.186e+11        perf-stat.ps.instructions
    114.44 ±119%    +175.5%     315.27 ± 62%  perf-stat.ps.major-faults
   2036962           +27.8%    2602997 ±  2%  perf-stat.ps.minor-faults
   2037076           +27.8%    2603313 ±  2%  perf-stat.ps.page-faults
 6.832e+12            +5.6%  7.215e+12        perf-stat.total.instructions
     48.07           -48.1        0.00        perf-profile.calltrace.cycles-pp._Fork
     47.54           -47.5        0.00        perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe._Fork
     47.54           -47.5        0.00        perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe._Fork
     47.51           -47.5        0.00        perf-profile.calltrace.cycles-pp.__do_sys_clone.do_syscall_64.entry_SYSCALL_64_after_hwframe._Fork
     47.51           -47.5        0.00        perf-profile.calltrace.cycles-pp.kernel_clone.__do_sys_clone.do_syscall_64.entry_SYSCALL_64_after_hwframe._Fork
     47.04           -47.0        0.00        perf-profile.calltrace.cycles-pp.copy_process.kernel_clone.__do_sys_clone.do_syscall_64.entry_SYSCALL_64_after_hwframe
     44.60           -44.6        0.00        perf-profile.calltrace.cycles-pp.dup_mm.copy_process.kernel_clone.__do_sys_clone.do_syscall_64
     43.93           -43.9        0.00        perf-profile.calltrace.cycles-pp.dup_mmap.dup_mm.copy_process.kernel_clone.__do_sys_clone
     41.28           -41.3        0.00        perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe
     41.27           -41.3        0.00        perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe
     40.48           -40.5        0.00        perf-profile.calltrace.cycles-pp.__x64_sys_exit_group.x64_sys_call.do_syscall_64.entry_SYSCALL_64_after_hwframe
     40.48           -40.5        0.00        perf-profile.calltrace.cycles-pp.do_exit.do_group_exit.__x64_sys_exit_group.x64_sys_call.do_syscall_64
     40.48           -40.5        0.00        perf-profile.calltrace.cycles-pp.do_group_exit.__x64_sys_exit_group.x64_sys_call.do_syscall_64.entry_SYSCALL_64_after_hwframe
     40.48           -40.5        0.00        perf-profile.calltrace.cycles-pp.x64_sys_call.do_syscall_64.entry_SYSCALL_64_after_hwframe
     38.17           -38.2        0.00        perf-profile.calltrace.cycles-pp.exit_mm.do_exit.do_group_exit.__x64_sys_exit_group.x64_sys_call
     38.09           -38.1        0.00        perf-profile.calltrace.cycles-pp.__mmput.exit_mm.do_exit.do_group_exit.__x64_sys_exit_group
     38.07           -38.1        0.00        perf-profile.calltrace.cycles-pp.exit_mmap.__mmput.exit_mm.do_exit.do_group_exit
     36.56           -36.6        0.00        perf-profile.calltrace.cycles-pp.anon_vma_fork.dup_mmap.dup_mm.copy_process.kernel_clone
     30.73           -30.7        0.00        perf-profile.calltrace.cycles-pp.free_pgtables.exit_mmap.__mmput.exit_mm.do_exit
     28.47           -28.5        0.00        perf-profile.calltrace.cycles-pp.unlink_anon_vmas.free_pgtables.exit_mmap.__mmput.exit_mm
     23.42           -23.4        0.00        perf-profile.calltrace.cycles-pp.anon_vma_clone.anon_vma_fork.dup_mmap.dup_mm.copy_process
     15.86           -15.9        0.00        perf-profile.calltrace.cycles-pp.down_write.anon_vma_clone.anon_vma_fork.dup_mmap.dup_mm
     15.74           -15.7        0.00        perf-profile.calltrace.cycles-pp.down_write.unlink_anon_vmas.free_pgtables.exit_mmap.__mmput
     15.57           -15.6        0.00        perf-profile.calltrace.cycles-pp.rwsem_down_write_slowpath.down_write.unlink_anon_vmas.free_pgtables.exit_mmap
     15.53           -15.5        0.00        perf-profile.calltrace.cycles-pp.rwsem_down_write_slowpath.down_write.anon_vma_clone.anon_vma_fork.dup_mmap
     12.70           -12.7        0.00        perf-profile.calltrace.cycles-pp.osq_lock.rwsem_down_write_slowpath.down_write.unlink_anon_vmas.free_pgtables
     12.58           -12.6        0.00        perf-profile.calltrace.cycles-pp.osq_lock.rwsem_down_write_slowpath.down_write.anon_vma_clone.anon_vma_fork
     12.49           -12.5        0.00        perf-profile.calltrace.cycles-pp.down_write.anon_vma_fork.dup_mmap.dup_mm.copy_process
     12.36           -12.4        0.00        perf-profile.calltrace.cycles-pp.rwsem_down_write_slowpath.down_write.anon_vma_fork.dup_mmap.dup_mm
     10.13           -10.1        0.00        perf-profile.calltrace.cycles-pp.osq_lock.rwsem_down_write_slowpath.down_write.anon_vma_fork.dup_mmap
      8.19            -8.2        0.00        perf-profile.calltrace.cycles-pp.__put_anon_vma.unlink_anon_vmas.free_pgtables.exit_mmap.__mmput
      7.76            -7.8        0.00        perf-profile.calltrace.cycles-pp.down_write.__put_anon_vma.unlink_anon_vmas.free_pgtables.exit_mmap
      7.74            -7.7        0.00        perf-profile.calltrace.cycles-pp.rwsem_down_write_slowpath.down_write.__put_anon_vma.unlink_anon_vmas.free_pgtables
      6.72            -6.7        0.00        perf-profile.calltrace.cycles-pp.osq_lock.rwsem_down_write_slowpath.down_write.__put_anon_vma.unlink_anon_vmas
      5.41            -5.4        0.00        perf-profile.calltrace.cycles-pp.prctl
      5.34            -5.3        0.00        perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.prctl
      5.34            -5.3        0.00        perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.prctl
      5.33            -5.3        0.00        perf-profile.calltrace.cycles-pp.__do_sys_prctl.do_syscall_64.entry_SYSCALL_64_after_hwframe.prctl
      5.27            -5.3        0.00        perf-profile.calltrace.cycles-pp.proc_comm_connector.__do_sys_prctl.do_syscall_64.entry_SYSCALL_64_after_hwframe.prctl
      5.25            -5.3        0.00        perf-profile.calltrace.cycles-pp.netlink_broadcast_filtered.proc_comm_connector.__do_sys_prctl.do_syscall_64.entry_SYSCALL_64_after_hwframe
      5.20            -5.2        0.00        perf-profile.calltrace.cycles-pp.do_one_broadcast.netlink_broadcast_filtered.proc_comm_connector.__do_sys_prctl.do_syscall_64
     94.62           -94.6        0.00        perf-profile.children.cycles-pp.do_syscall_64
     94.62           -94.6        0.00        perf-profile.children.cycles-pp.entry_SYSCALL_64_after_hwframe
     52.82           -52.8        0.00        perf-profile.children.cycles-pp.down_write
     51.93           -51.9        0.00        perf-profile.children.cycles-pp.rwsem_down_write_slowpath
     48.12           -48.1        0.00        perf-profile.children.cycles-pp._Fork
     47.60           -47.6        0.00        perf-profile.children.cycles-pp.kernel_clone
     47.59           -47.6        0.00        perf-profile.children.cycles-pp.__do_sys_clone
     47.12           -47.1        0.00        perf-profile.children.cycles-pp.copy_process
     44.67           -44.7        0.00        perf-profile.children.cycles-pp.dup_mm
     44.01           -44.0        0.00        perf-profile.children.cycles-pp.dup_mmap
     42.44           -42.4        0.00        perf-profile.children.cycles-pp.osq_lock
     40.80           -40.8        0.00        perf-profile.children.cycles-pp.x64_sys_call
     40.79           -40.8        0.00        perf-profile.children.cycles-pp.__x64_sys_exit_group
     40.79           -40.8        0.00        perf-profile.children.cycles-pp.do_group_exit
     40.78           -40.8        0.00        perf-profile.children.cycles-pp.do_exit
     38.18           -38.2        0.00        perf-profile.children.cycles-pp.exit_mm
     38.09           -38.1        0.00        perf-profile.children.cycles-pp.__mmput
     38.07           -38.1        0.00        perf-profile.children.cycles-pp.exit_mmap
     36.63           -36.6        0.00        perf-profile.children.cycles-pp.anon_vma_fork
     30.74           -30.7        0.00        perf-profile.children.cycles-pp.free_pgtables
     28.47           -28.5        0.00        perf-profile.children.cycles-pp.unlink_anon_vmas
     23.47           -23.5        0.00        perf-profile.children.cycles-pp.anon_vma_clone
      9.00            -9.0        0.00        perf-profile.children.cycles-pp.netlink_broadcast_filtered
      8.94            -8.9        0.00        perf-profile.children.cycles-pp.do_one_broadcast
      8.19            -8.2        0.00        perf-profile.children.cycles-pp.__put_anon_vma
      7.89            -7.9        0.00        perf-profile.children.cycles-pp.rwsem_spin_on_owner
      5.42            -5.4        0.00        perf-profile.children.cycles-pp.prctl
      5.34            -5.3        0.00        perf-profile.children.cycles-pp.__do_sys_prctl
      5.28            -5.3        0.00        perf-profile.children.cycles-pp.proc_comm_connector
     41.94           -41.9        0.00        perf-profile.self.cycles-pp.osq_lock
      7.80            -7.8        0.00        perf-profile.self.cycles-pp.rwsem_spin_on_owner
      6.29            -6.3        0.00        perf-profile.self.cycles-pp.do_one_broadcast





Disclaimer:
Results have been estimated based on internal Intel analysis and are provided
for informational purposes only. Any difference in system hardware or software
design or configuration may affect actual performance.


-- 
0-DAY CI Kernel Test Service
https://github.com/intel/lkp-tests/wiki



                 reply	other threads:[~2026-02-06 10:04 UTC|newest]

Thread overview: [no followups] expand[flat|nested]  mbox.gz  Atom feed

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=202602061747.855f053f-lkp@intel.com \
    --to=oliver.sang@intel.com \
    --cc=Liam.Howlett@oracle.com \
    --cc=akpm@linux-foundation.org \
    --cc=chriscli@google.com \
    --cc=david@kernel.org \
    --cc=harry.yoo@oracle.com \
    --cc=jannh@google.com \
    --cc=linux-mm@kvack.org \
    --cc=lkp@intel.com \
    --cc=lorenzo.stoakes@oracle.com \
    --cc=mhocko@suse.com \
    --cc=oe-lkp@lists.linux.dev \
    --cc=pfalcato@suse.de \
    --cc=riel@surriel.com \
    --cc=rppt@kernel.org \
    --cc=shakeel.butt@linux.dev \
    --cc=surenb@google.com \
    --cc=v-songbaohua@oppo.com \
    --cc=vbabka@suse.cz \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox