(4) Intel(R) Xeon(R) Platinum 8380H CPU @ 2.90GHz (Cooper Lake) with memory: 192G ========================================================================================= compiler/cpufreq_governor/kconfig/mode/nr_task/rootfs/tbox_group/test/testcase: gcc-12/performance/x86_64-rhel-8.3/process/50%/debian-11.1-x86_64-20220510.cgz/lkp-cpl-4sp2/malloc1/will-it-scale 1803d0c5ee1a3bbe efa7df3e3bb5da8e6abbe377274 d8d7b1dae6f0311d528b289cda7 ---------------- --------------------------- --------------------------- %stddev %change %stddev %change %stddev \ | \ | \ 3161 +46.4% 4627 +47.5% 4662 vmstat.system.cs 0.58 ± 2% +0.7 1.27 +0.7 1.26 mpstat.cpu.all.irq% 0.55 ± 3% -0.5 0.09 ± 2% -0.5 0.09 ± 2% mpstat.cpu.all.soft% 1.00 ± 13% -0.7 0.29 -0.7 0.28 mpstat.cpu.all.usr% 1231431 -86.7% 164315 -86.7% 163624 will-it-scale.112.processes 10994 -86.7% 1466 -86.7% 1460 will-it-scale.per_process_ops 1231431 -86.7% 164315 -86.7% 163624 will-it-scale.workload 0.03 -66.7% 0.01 -66.7% 0.01 turbostat.IPC 81.38 -2.8% 79.12 -2.2% 79.62 turbostat.PkgTmp 764.02 +17.1% 894.78 +17.0% 893.81 turbostat.PkgWatt 19.80 +135.4% 46.59 +135.1% 46.53 turbostat.RAMWatt 771.38 ± 5% +249.5% 2696 ± 14% +231.9% 2560 ± 10% perf-c2c.DRAM.local 3050 ± 5% -69.8% 922.75 ± 6% -71.5% 869.88 ± 8% perf-c2c.DRAM.remote 11348 ± 4% -90.2% 1107 ± 5% -90.6% 1065 ± 3% perf-c2c.HITM.local 357.50 ± 21% -44.0% 200.38 ± 7% -48.2% 185.25 ± 13% perf-c2c.HITM.remote 11706 ± 4% -88.8% 1307 ± 4% -89.3% 1250 ± 3% perf-c2c.HITM.total 1.717e+08 ± 9% -85.5% 24955542 -85.5% 24880885 numa-numastat.node0.local_node 1.718e+08 ± 9% -85.4% 25046901 -85.5% 24972867 numa-numastat.node0.numa_hit 1.945e+08 ± 7% -87.0% 25203631 -87.1% 25104844 numa-numastat.node1.local_node 1.946e+08 ± 7% -87.0% 25300536 -87.1% 25180465 numa-numastat.node1.numa_hit 2.001e+08 ± 2% -87.5% 25098699 -87.5% 25011079 numa-numastat.node2.local_node 2.002e+08 ± 2% -87.4% 25173132 -87.5% 25119438 numa-numastat.node2.numa_hit 1.956e+08 ± 6% -87.3% 24922332 -87.3% 24784408 numa-numastat.node3.local_node 1.957e+08 ± 6% -87.2% 25008002 -87.3% 24874399 numa-numastat.node3.numa_hit 766959 -45.9% 414816 -46.2% 412898 meminfo.Active 766881 -45.9% 414742 -46.2% 412824 meminfo.Active(anon) 391581 +12.1% 438946 +8.4% 424669 meminfo.AnonPages 421982 +20.7% 509155 +14.8% 484430 meminfo.Inactive 421800 +20.7% 508969 +14.8% 484244 meminfo.Inactive(anon) 68496 ± 7% +88.9% 129357 ± 2% +82.9% 125252 ± 2% meminfo.Mapped 569270 -24.0% 432709 -24.1% 431884 meminfo.SUnreclaim 797185 -40.2% 476420 -40.8% 471912 meminfo.Shmem 730111 -18.8% 593041 -18.9% 592400 meminfo.Slab 148082 ± 2% -20.3% 118055 ± 4% -21.7% 115994 ± 6% numa-meminfo.node0.SUnreclaim 197311 ± 16% -22.5% 152829 ± 19% -29.8% 138546 ± 9% numa-meminfo.node0.Slab 144635 ± 5% -25.8% 107254 ± 4% -25.3% 107973 ± 6% numa-meminfo.node1.SUnreclaim 137974 ± 2% -24.5% 104205 ± 6% -25.7% 102563 ± 4% numa-meminfo.node2.SUnreclaim 167889 ± 13% -26.1% 124127 ± 9% -15.0% 142771 ± 18% numa-meminfo.node2.Slab 607639 ± 20% -46.2% 326998 ± 15% -46.8% 323458 ± 13% numa-meminfo.node3.Active 607611 ± 20% -46.2% 326968 ± 15% -46.8% 323438 ± 13% numa-meminfo.node3.Active(anon) 679476 ± 21% -31.3% 466619 ± 19% -38.5% 418074 ± 16% numa-meminfo.node3.FilePages 20150 ± 22% +128.4% 46020 ± 11% +123.0% 44932 ± 8% numa-meminfo.node3.Mapped 138148 ± 2% -25.3% 103148 ± 4% -23.8% 105326 ± 7% numa-meminfo.node3.SUnreclaim 631930 ± 20% -40.9% 373456 ± 15% -41.5% 369883 ± 13% numa-meminfo.node3.Shmem 166777 ± 7% -19.6% 134013 ± 9% -20.7% 132332 ± 7% numa-meminfo.node3.Slab 37030 ± 2% -20.3% 29511 ± 4% -21.7% 28993 ± 6% numa-vmstat.node0.nr_slab_unreclaimable 1.718e+08 ± 9% -85.4% 25047066 -85.5% 24973455 numa-vmstat.node0.numa_hit 1.717e+08 ± 9% -85.5% 24955707 -85.5% 24881472 numa-vmstat.node0.numa_local 36158 ± 5% -25.8% 26811 ± 4% -25.4% 26990 ± 6% numa-vmstat.node1.nr_slab_unreclaimable 1.946e+08 ± 7% -87.0% 25300606 -87.1% 25181038 numa-vmstat.node1.numa_hit 1.945e+08 ± 7% -87.0% 25203699 -87.1% 25105417 numa-vmstat.node1.numa_local 34499 ± 2% -24.5% 26050 ± 6% -25.7% 25638 ± 4% numa-vmstat.node2.nr_slab_unreclaimable 2.002e+08 ± 2% -87.4% 25173363 -87.5% 25119830 numa-vmstat.node2.numa_hit 2.001e+08 ± 2% -87.5% 25098930 -87.5% 25011471 numa-vmstat.node2.numa_local 151851 ± 20% -46.2% 81720 ± 15% -46.8% 80848 ± 13% numa-vmstat.node3.nr_active_anon 169827 ± 21% -31.3% 116645 ± 19% -38.5% 104502 ± 16% numa-vmstat.node3.nr_file_pages 4991 ± 23% +131.5% 11555 ± 11% +125.4% 11249 ± 8% numa-vmstat.node3.nr_mapped 157941 ± 20% -40.9% 93355 ± 15% -41.5% 92454 ± 13% numa-vmstat.node3.nr_shmem 34570 ± 2% -25.4% 25780 ± 4% -23.8% 26327 ± 7% numa-vmstat.node3.nr_slab_unreclaimable 151851 ± 20% -46.2% 81720 ± 15% -46.8% 80848 ± 13% numa-vmstat.node3.nr_zone_active_anon 1.957e+08 ± 6% -87.2% 25008117 -87.3% 24874649 numa-vmstat.node3.numa_hit 1.956e+08 ± 6% -87.3% 24922447 -87.3% 24784657 numa-vmstat.node3.numa_local 191746 -45.9% 103734 -46.2% 103228 proc-vmstat.nr_active_anon 97888 +12.1% 109757 +8.5% 106185 proc-vmstat.nr_anon_pages 947825 -8.5% 867659 -8.6% 866533 proc-vmstat.nr_file_pages 105444 +20.7% 127227 +14.9% 121113 proc-vmstat.nr_inactive_anon 17130 ± 7% +88.9% 32365 ± 2% +83.4% 31420 ± 2% proc-vmstat.nr_mapped 4007 +4.2% 4176 +4.1% 4170 proc-vmstat.nr_page_table_pages 199322 -40.2% 119155 -40.8% 118031 proc-vmstat.nr_shmem 142294 -24.0% 108161 -24.1% 107954 proc-vmstat.nr_slab_unreclaimable 191746 -45.9% 103734 -46.2% 103228 proc-vmstat.nr_zone_active_anon 105444 +20.7% 127223 +14.9% 121106 proc-vmstat.nr_zone_inactive_anon 40186 ± 13% +65.0% 66320 ± 5% +60.2% 64374 ± 13% proc-vmstat.numa_hint_faults 20248 ± 39% +108.3% 42185 ± 12% +102.6% 41033 ± 10% proc-vmstat.numa_hint_faults_local 7.623e+08 -86.8% 1.005e+08 -86.9% 1.002e+08 proc-vmstat.numa_hit 7.62e+08 -86.9% 1.002e+08 -86.9% 99786408 proc-vmstat.numa_local 181538 ± 6% +49.5% 271428 ± 3% +48.9% 270328 ± 6% proc-vmstat.numa_pte_updates 152652 ± 7% -28.6% 108996 -29.6% 107396 proc-vmstat.pgactivate 7.993e+08 +3068.4% 2.533e+10 +3055.6% 2.522e+10 proc-vmstat.pgalloc_normal 3.72e+08 -86.4% 50632612 -86.4% 50429200 proc-vmstat.pgfault 7.99e+08 +3069.7% 2.533e+10 +3056.9% 2.522e+10 proc-vmstat.pgfree 48.75 ± 2% +1e+08% 49362627 +1e+08% 49162408 proc-vmstat.thp_fault_alloc 21789703 ± 10% -20.1% 17410551 ± 7% -18.9% 17673460 ± 4% sched_debug.cfs_rq:/.avg_vruntime.max 427573 ± 99% +1126.7% 5245182 ± 17% +1104.4% 5149659 ± 13% sched_debug.cfs_rq:/.avg_vruntime.min 4757464 ± 10% -48.3% 2458136 ± 19% -46.6% 2539001 ± 11% sched_debug.cfs_rq:/.avg_vruntime.stddev 0.44 ± 2% -15.9% 0.37 ± 2% -16.6% 0.37 ± 3% sched_debug.cfs_rq:/.h_nr_running.stddev 299205 ± 38% +59.3% 476493 ± 27% +50.6% 450561 ± 42% sched_debug.cfs_rq:/.load.max 21789703 ± 10% -20.1% 17410551 ± 7% -18.9% 17673460 ± 4% sched_debug.cfs_rq:/.min_vruntime.max 427573 ± 99% +1126.7% 5245182 ± 17% +1104.4% 5149659 ± 13% sched_debug.cfs_rq:/.min_vruntime.min 4757464 ± 10% -48.3% 2458136 ± 19% -46.6% 2539001 ± 11% sched_debug.cfs_rq:/.min_vruntime.stddev 0.44 ± 2% -16.0% 0.37 ± 2% -17.2% 0.36 ± 2% sched_debug.cfs_rq:/.nr_running.stddev 446.75 ± 2% -18.4% 364.71 ± 2% -19.3% 360.46 ± 2% sched_debug.cfs_rq:/.runnable_avg.stddev 445.25 ± 2% -18.4% 363.46 ± 2% -19.3% 359.33 ± 2% sched_debug.cfs_rq:/.util_avg.stddev 946.71 ± 3% -14.7% 807.54 ± 4% -15.4% 800.58 ± 7% sched_debug.cfs_rq:/.util_est_enqueued.max 281.39 ± 7% -31.2% 193.63 ± 4% -32.0% 191.24 ± 7% sched_debug.cfs_rq:/.util_est_enqueued.stddev 1131635 ± 7% +73.7% 1965577 ± 6% +76.5% 1997455 ± 7% sched_debug.cpu.avg_idle.max 223539 ± 16% +165.4% 593172 ± 7% +146.0% 549906 ± 11% sched_debug.cpu.avg_idle.min 83325 ± 4% +64.3% 136927 ± 9% +69.7% 141399 ± 11% sched_debug.cpu.avg_idle.stddev 17.57 ± 6% +594.5% 122.01 ± 3% +588.0% 120.88 ± 3% sched_debug.cpu.clock.stddev 873.33 -11.1% 776.19 -11.8% 770.20 sched_debug.cpu.clock_task.stddev 2870 -18.1% 2351 -17.4% 2371 sched_debug.cpu.curr->pid.avg 3003 -12.5% 2627 -12.4% 2630 sched_debug.cpu.curr->pid.stddev 550902 ± 6% +74.4% 960871 ± 6% +79.8% 990291 ± 8% sched_debug.cpu.max_idle_balance_cost.max 4451 ± 59% +1043.9% 50917 ± 15% +1129.4% 54721 ± 15% sched_debug.cpu.max_idle_balance_cost.stddev 0.00 ± 17% +385.8% 0.00 ± 34% +315.7% 0.00 ± 3% sched_debug.cpu.next_balance.stddev 0.43 -17.5% 0.35 -16.8% 0.35 sched_debug.cpu.nr_running.avg 1.15 ± 8% +25.0% 1.44 ± 8% +30.4% 1.50 ± 13% sched_debug.cpu.nr_running.max 0.45 -14.4% 0.39 -14.2% 0.39 ± 2% sched_debug.cpu.nr_running.stddev 3280 ± 5% +32.5% 4345 +34.5% 4412 sched_debug.cpu.nr_switches.avg 846.82 ± 11% +109.9% 1777 ± 12% +112.4% 1799 ± 4% sched_debug.cpu.nr_switches.min 0.03 ±173% +887.2% 0.30 ± 73% +521.1% 0.19 ± 35% sched_debug.rt_rq:.rt_time.avg 6.79 ±173% +887.2% 67.01 ± 73% +521.1% 42.16 ± 35% sched_debug.rt_rq:.rt_time.max 0.45 ±173% +887.2% 4.47 ± 73% +521.1% 2.81 ± 35% sched_debug.rt_rq:.rt_time.stddev 4.65 +28.0% 5.96 +28.5% 5.98 perf-stat.i.MPKI 8.721e+09 -71.0% 2.532e+09 -71.1% 2.523e+09 perf-stat.i.branch-instructions 0.34 +0.1 0.48 +0.1 0.48 perf-stat.i.branch-miss-rate% 30145441 -58.6% 12471062 -58.6% 12487542 perf-stat.i.branch-misses 33.52 -15.3 18.20 -15.2 18.27 perf-stat.i.cache-miss-rate% 1.819e+08 -58.8% 74947458 -58.8% 74903072 perf-stat.i.cache-misses 5.429e+08 ± 2% -24.1% 4.123e+08 -24.4% 4.103e+08 perf-stat.i.cache-references 3041 +48.6% 4518 +49.7% 4552 perf-stat.i.context-switches 10.96 +212.9% 34.28 +214.1% 34.41 perf-stat.i.cpi 309.29 -11.2% 274.59 -11.3% 274.20 perf-stat.i.cpu-migrations 2354 +144.6% 5758 +144.7% 5761 perf-stat.i.cycles-between-cache-misses 0.13 -0.1 0.01 ± 3% -0.1 0.01 ± 3% perf-stat.i.dTLB-load-miss-rate% 12852209 ± 2% -98.0% 261197 ± 3% -97.9% 263864 ± 3% perf-stat.i.dTLB-load-misses 9.56e+09 -69.3% 2.932e+09 -69.4% 2.922e+09 perf-stat.i.dTLB-loads 0.12 -0.1 0.03 -0.1 0.03 perf-stat.i.dTLB-store-miss-rate% 5083186 -86.3% 693971 -86.4% 690328 perf-stat.i.dTLB-store-misses 4.209e+09 -44.9% 2.317e+09 -45.2% 2.308e+09 perf-stat.i.dTLB-stores 76.33 -39.7 36.61 -39.7 36.59 perf-stat.i.iTLB-load-miss-rate% 18717931 -80.1% 3715941 -80.2% 3698121 perf-stat.i.iTLB-load-misses 5758034 +7.7% 6202790 +7.4% 6183041 perf-stat.i.iTLB-loads 3.914e+10 -67.8% 1.261e+10 -67.9% 1.256e+10 perf-stat.i.instructions 2107 +73.9% 3663 +73.6% 3658 perf-stat.i.instructions-per-iTLB-miss 0.09 -67.9% 0.03 -68.1% 0.03 perf-stat.i.ipc 269.39 +10.6% 297.91 +10.7% 298.33 perf-stat.i.metric.K/sec 102.78 -64.5% 36.54 -64.6% 36.40 perf-stat.i.metric.M/sec 1234832 -86.4% 167556 -86.5% 166848 perf-stat.i.minor-faults 87.25 -41.9 45.32 -42.2 45.09 perf-stat.i.node-load-miss-rate% 25443233 -83.0% 4326696 ± 3% -83.4% 4227985 ± 2% perf-stat.i.node-load-misses 3723342 ± 3% +45.4% 5414430 +44.3% 5372545 perf-stat.i.node-loads 79.20 -74.4 4.78 -74.5 4.74 perf-stat.i.node-store-miss-rate% 14161911 ± 2% -83.1% 2394469 -83.2% 2382317 perf-stat.i.node-store-misses 3727955 ± 3% +1181.6% 47776544 +1188.5% 48035797 perf-stat.i.node-stores 1234832 -86.4% 167556 -86.5% 166849 perf-stat.i.page-faults 4.65 +28.0% 5.95 +28.4% 5.97 perf-stat.overall.MPKI 0.35 +0.1 0.49 +0.1 0.49 perf-stat.overall.branch-miss-rate% 33.51 -15.3 18.19 -15.3 18.26 perf-stat.overall.cache-miss-rate% 10.94 +212.3% 34.16 +213.4% 34.28 perf-stat.overall.cpi 2354 +143.9% 5741 +144.1% 5746 perf-stat.overall.cycles-between-cache-misses 0.13 -0.1 0.01 ± 3% -0.1 0.01 ± 5% perf-stat.overall.dTLB-load-miss-rate% 0.12 -0.1 0.03 -0.1 0.03 perf-stat.overall.dTLB-store-miss-rate% 76.49 -39.2 37.31 -39.2 37.29 perf-stat.overall.iTLB-load-miss-rate% 2090 +63.4% 3416 +63.5% 3417 perf-stat.overall.instructions-per-iTLB-miss 0.09 -68.0% 0.03 -68.1% 0.03 perf-stat.overall.ipc 87.22 -43.1 44.12 ± 2% -43.5 43.76 perf-stat.overall.node-load-miss-rate% 79.16 -74.4 4.77 -74.4 4.72 perf-stat.overall.node-store-miss-rate% 9549728 +140.9% 23005172 +141.1% 23022843 perf-stat.overall.path-length 8.691e+09 -71.0% 2.519e+09 -71.1% 2.51e+09 perf-stat.ps.branch-instructions 30118940 -59.1% 12319517 -59.1% 12327993 perf-stat.ps.branch-misses 1.813e+08 -58.8% 74623919 -58.9% 74563289 perf-stat.ps.cache-misses 5.41e+08 ± 2% -24.2% 4.103e+08 -24.5% 4.085e+08 perf-stat.ps.cache-references 3031 +47.9% 4485 +49.1% 4519 perf-stat.ps.context-switches 307.72 -12.7% 268.59 -12.7% 268.66 perf-stat.ps.cpu-migrations 12806734 ± 2% -98.0% 260740 ± 4% -97.9% 267782 ± 5% perf-stat.ps.dTLB-load-misses 9.528e+09 -69.4% 2.917e+09 -69.5% 2.907e+09 perf-stat.ps.dTLB-loads 5063992 -86.4% 690720 -86.4% 687415 perf-stat.ps.dTLB-store-misses 4.195e+09 -45.0% 2.306e+09 -45.2% 2.297e+09 perf-stat.ps.dTLB-stores 18661026 -80.3% 3672024 -80.4% 3658006 perf-stat.ps.iTLB-load-misses 5735379 +7.6% 6169096 +7.3% 6151755 perf-stat.ps.iTLB-loads 3.901e+10 -67.8% 1.254e+10 -68.0% 1.25e+10 perf-stat.ps.instructions 1230175 -86.4% 166708 -86.5% 166045 perf-stat.ps.minor-faults 25346347 -83.0% 4299946 ± 2% -83.4% 4203636 ± 2% perf-stat.ps.node-load-misses 3713652 ± 3% +46.6% 5444481 +45.5% 5401831 perf-stat.ps.node-loads 14107969 ± 2% -83.1% 2381707 -83.2% 2368146 perf-stat.ps.node-store-misses 3716359 ± 3% +1179.6% 47556224 +1186.1% 47797289 perf-stat.ps.node-stores 1230175 -86.4% 166708 -86.5% 166046 perf-stat.ps.page-faults 1.176e+13 -67.9% 3.78e+12 -68.0% 3.767e+12 perf-stat.total.instructions 0.01 ± 42% +385.1% 0.03 ± 8% +566.0% 0.04 ± 42% perf-sched.sch_delay.avg.ms.__cond_resched.stop_one_cpu.sched_exec.bprm_execve.part 0.01 ± 17% +354.3% 0.05 ± 8% +402.1% 0.06 ± 8% perf-sched.sch_delay.avg.ms.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown] 0.01 ± 19% +323.1% 0.06 ± 27% +347.1% 0.06 ± 17% perf-sched.sch_delay.avg.ms.devkmsg_read.vfs_read.ksys_read.do_syscall_64 0.01 ± 14% +2.9e+05% 25.06 ±172% +1.6e+05% 13.94 ±263% perf-sched.sch_delay.avg.ms.do_nanosleep.hrtimer_nanosleep.common_nsleep.__x64_sys_clock_nanosleep 0.00 ±129% +7133.3% 0.03 ± 7% +7200.0% 0.03 ± 4% perf-sched.sch_delay.avg.ms.do_task_dead.do_exit.do_group_exit.__x64_sys_exit_group.do_syscall_64 0.01 ± 8% +396.8% 0.06 ± 2% +402.1% 0.06 ± 2% perf-sched.sch_delay.avg.ms.do_wait.kernel_wait4.__do_sys_wait4.do_syscall_64 0.01 ± 9% +256.9% 0.03 ± 10% +232.8% 0.02 ± 13% perf-sched.sch_delay.avg.ms.schedule_hrtimeout_range_clock.do_poll.constprop.0.do_sys_poll 0.01 ± 15% +324.0% 0.05 ± 17% +320.8% 0.05 ± 17% perf-sched.sch_delay.avg.ms.schedule_hrtimeout_range_clock.do_select.core_sys_select.kern_select 0.01 ± 19% +338.6% 0.06 ± 7% +305.0% 0.05 ± 8% perf-sched.sch_delay.avg.ms.schedule_hrtimeout_range_clock.ep_poll.do_epoll_wait.__x64_sys_epoll_wait 0.01 ± 9% +298.4% 0.03 ± 2% +304.8% 0.03 perf-sched.sch_delay.avg.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone 0.01 ± 7% +265.8% 0.03 ± 5% +17282.9% 1.65 ±258% perf-sched.sch_delay.avg.ms.schedule_timeout.kcompactd.kthread.ret_from_fork 0.19 ± 11% -89.3% 0.02 ± 10% -89.4% 0.02 ± 10% perf-sched.sch_delay.avg.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm 0.01 ± 28% +319.8% 0.05 ± 19% +303.0% 0.05 ± 18% perf-sched.sch_delay.avg.ms.syslog_print.do_syslog.kmsg_read.vfs_read 0.01 ± 14% +338.9% 0.03 ± 9% +318.5% 0.03 ± 4% perf-sched.sch_delay.avg.ms.wait_for_partner.fifo_open.do_dentry_open.do_open 0.02 ± 20% +674.2% 0.12 ±137% +267.5% 0.06 ± 15% perf-sched.sch_delay.max.ms.__cond_resched.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr.__sched_setaffinity 0.01 ± 46% +256.9% 0.03 ± 11% +1095.8% 0.11 ±112% perf-sched.sch_delay.max.ms.__cond_resched.stop_one_cpu.sched_exec.bprm_execve.part 0.02 ± 28% +324.6% 0.07 ± 8% +353.2% 0.07 ± 9% perf-sched.sch_delay.max.ms.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown] 0.02 ± 21% +318.4% 0.07 ± 25% +389.6% 0.08 ± 26% perf-sched.sch_delay.max.ms.devkmsg_read.vfs_read.ksys_read.do_syscall_64 0.01 ± 26% +1.9e+06% 250.13 ±173% +9.7e+05% 125.09 ±264% perf-sched.sch_delay.max.ms.do_nanosleep.hrtimer_nanosleep.common_nsleep.__x64_sys_clock_nanosleep 0.02 ± 25% +585.6% 0.11 ± 63% +454.5% 0.09 ± 31% perf-sched.sch_delay.max.ms.do_task_dead.do_exit.do_group_exit.__x64_sys_exit_group.do_syscall_64 0.04 ± 39% +159.0% 0.11 ± 6% +190.0% 0.13 ± 10% perf-sched.sch_delay.max.ms.do_wait.kernel_wait4.__do_sys_wait4.do_syscall_64 0.01 ± 29% +312.9% 0.06 ± 19% +401.7% 0.07 ± 13% perf-sched.sch_delay.max.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.syscall_exit_to_user_mode.do_syscall_64 0.02 ± 25% +216.8% 0.06 ± 36% +166.4% 0.05 ± 7% perf-sched.sch_delay.max.ms.schedule_hrtimeout_range_clock.do_poll.constprop.0.do_sys_poll 0.01 ± 21% +345.8% 0.07 ± 26% +298.3% 0.06 ± 18% perf-sched.sch_delay.max.ms.schedule_hrtimeout_range_clock.do_select.core_sys_select.kern_select 0.03 ± 35% +190.2% 0.07 ± 16% +187.8% 0.07 ± 11% perf-sched.sch_delay.max.ms.schedule_hrtimeout_range_clock.ep_poll.do_epoll_wait.__x64_sys_epoll_wait 0.02 ± 19% +220.8% 0.07 ± 23% +2.9e+05% 63.06 ±263% perf-sched.sch_delay.max.ms.schedule_timeout.kcompactd.kthread.ret_from_fork 4.60 ± 5% -10.7% 4.11 ± 8% -13.4% 3.99 perf-sched.sch_delay.max.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm 0.02 ± 32% +368.0% 0.07 ± 25% +346.9% 0.07 ± 20% perf-sched.sch_delay.max.ms.syslog_print.do_syslog.kmsg_read.vfs_read 189.60 -32.9% 127.16 -33.0% 126.98 perf-sched.total_wait_and_delay.average.ms 11265 ± 3% +73.7% 19568 ± 3% +71.1% 19274 perf-sched.total_wait_and_delay.count.ms 189.18 -32.9% 126.97 -33.0% 126.81 perf-sched.total_wait_time.average.ms 0.50 ± 20% -100.0% 0.00 -100.0% 0.00 perf-sched.wait_and_delay.avg.ms.__cond_resched.__anon_vma_prepare.do_anonymous_page.__handle_mm_fault.handle_mm_fault 0.50 ± 11% -100.0% 0.00 -100.0% 0.00 perf-sched.wait_and_delay.avg.ms.__cond_resched.tlb_batch_pages_flush.tlb_finish_mmu.unmap_region.constprop 0.43 ± 16% -100.0% 0.00 -100.0% 0.00 perf-sched.wait_and_delay.avg.ms.__cond_resched.unmap_vmas.unmap_region.constprop.0 52.33 ± 31% +223.4% 169.23 ± 7% +226.5% 170.86 ± 2% perf-sched.wait_and_delay.avg.ms.do_task_dead.do_exit.do_group_exit.__x64_sys_exit_group.do_syscall_64 0.51 ± 18% -100.0% 0.00 -100.0% 0.00 perf-sched.wait_and_delay.avg.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_exc_page_fault 28.05 ± 4% +27.8% 35.84 ± 4% +26.0% 35.34 ± 8% perf-sched.wait_and_delay.avg.ms.pipe_read.vfs_read.ksys_read.do_syscall_64 2.08 ± 3% +33.2% 2.76 +32.9% 2.76 ± 2% perf-sched.wait_and_delay.avg.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone 491.80 -53.6% 227.96 ± 3% -53.5% 228.58 ± 2% perf-sched.wait_and_delay.avg.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm 222.00 ± 9% -100.0% 0.00 -100.0% 0.00 perf-sched.wait_and_delay.count.__cond_resched.__anon_vma_prepare.do_anonymous_page.__handle_mm_fault.handle_mm_fault 8.75 ± 33% -84.3% 1.38 ±140% -82.9% 1.50 ± 57% perf-sched.wait_and_delay.count.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm 1065 ± 3% -100.0% 0.00 -100.0% 0.00 perf-sched.wait_and_delay.count.__cond_resched.tlb_batch_pages_flush.tlb_finish_mmu.unmap_region.constprop 538.25 ± 9% -100.0% 0.00 -100.0% 0.00 perf-sched.wait_and_delay.count.__cond_resched.unmap_vmas.unmap_region.constprop.0 307.75 ± 6% -100.0% 0.00 -100.0% 0.00 perf-sched.wait_and_delay.count.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_exc_page_fault 2458 ± 3% -20.9% 1944 ± 4% -20.5% 1954 ± 7% perf-sched.wait_and_delay.count.pipe_read.vfs_read.ksys_read.do_syscall_64 2577 ± 5% +168.6% 6921 ± 4% +165.0% 6829 ± 2% perf-sched.wait_and_delay.count.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm 7.07 ±172% -100.0% 0.00 -100.0% 0.00 perf-sched.wait_and_delay.max.ms.__cond_resched.__anon_vma_prepare.do_anonymous_page.__handle_mm_fault.handle_mm_fault 1730 ± 24% -77.9% 382.66 ±117% -50.1% 862.68 ± 89% perf-sched.wait_and_delay.max.ms.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm 34.78 ± 43% -100.0% 0.00 -100.0% 0.00 perf-sched.wait_and_delay.max.ms.__cond_resched.tlb_batch_pages_flush.tlb_finish_mmu.unmap_region.constprop 8.04 ±179% -100.0% 0.00 -100.0% 0.00 perf-sched.wait_and_delay.max.ms.__cond_resched.unmap_vmas.unmap_region.constprop.0 9.47 ±134% -100.0% 0.00 -100.0% 0.00 perf-sched.wait_and_delay.max.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_exc_page_fault 3.96 ± 6% +60.6% 6.36 ± 5% +58.3% 6.27 ± 6% perf-sched.wait_and_delay.max.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone 0.42 ± 27% -100.0% 0.00 -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.__alloc_pages.alloc_pages_mpol.pte_alloc_one.__pte_alloc 0.50 ± 20% -100.0% 0.00 -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.__anon_vma_prepare.do_anonymous_page.__handle_mm_fault.handle_mm_fault 0.51 ± 17% -100.0% 0.00 -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.down_write.__anon_vma_prepare.do_anonymous_page.__handle_mm_fault 0.59 ± 17% -100.0% 0.00 -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc.__anon_vma_prepare.do_anonymous_page.__handle_mm_fault 0.46 ± 31% -63.3% 0.17 ± 18% -67.7% 0.15 ± 15% perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc.vm_area_alloc.mmap_region.do_mmap 0.50 ± 11% -67.8% 0.16 ± 8% -67.6% 0.16 ± 4% perf-sched.wait_time.avg.ms.__cond_resched.tlb_batch_pages_flush.tlb_finish_mmu.unmap_region.constprop 0.43 ± 16% -63.5% 0.16 ± 10% -62.6% 0.16 ± 4% perf-sched.wait_time.avg.ms.__cond_resched.unmap_vmas.unmap_region.constprop.0 0.50 ± 19% -67.0% 0.17 ± 5% -69.0% 0.16 ± 11% perf-sched.wait_time.avg.ms.__cond_resched.zap_pmd_range.isra.0.unmap_page_range 1.71 ± 5% +55.9% 2.66 ± 3% +47.3% 2.52 ± 6% perf-sched.wait_time.avg.ms.devkmsg_read.vfs_read.ksys_read.do_syscall_64 52.33 ± 31% +223.4% 169.20 ± 7% +226.5% 170.83 ± 2% perf-sched.wait_time.avg.ms.do_task_dead.do_exit.do_group_exit.__x64_sys_exit_group.do_syscall_64 0.51 ± 18% -67.7% 0.16 ± 5% -68.0% 0.16 ± 6% perf-sched.wait_time.avg.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_exc_page_fault 0.53 ± 17% -65.4% 0.18 ± 56% -66.5% 0.18 ± 10% perf-sched.wait_time.avg.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt 27.63 ± 4% +29.7% 35.83 ± 4% +27.6% 35.27 ± 8% perf-sched.wait_time.avg.ms.pipe_read.vfs_read.ksys_read.do_syscall_64 2.07 ± 3% +32.1% 2.73 +31.9% 2.73 ± 2% perf-sched.wait_time.avg.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone 491.61 -53.6% 227.94 ± 3% -53.5% 228.56 ± 2% perf-sched.wait_time.avg.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm 1.72 ± 5% +58.1% 2.73 ± 3% +50.4% 2.59 ± 7% perf-sched.wait_time.avg.ms.syslog_print.do_syslog.kmsg_read.vfs_read 1.42 ± 21% -100.0% 0.00 -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.__alloc_pages.alloc_pages_mpol.pte_alloc_one.__pte_alloc 7.07 ±172% -100.0% 0.00 -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.__anon_vma_prepare.do_anonymous_page.__handle_mm_fault.handle_mm_fault 1.66 ± 27% -100.0% 0.00 -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.down_write.__anon_vma_prepare.do_anonymous_page.__handle_mm_fault 2.05 ± 57% -100.0% 0.00 -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc.__anon_vma_prepare.do_anonymous_page.__handle_mm_fault 1.69 ± 20% -84.6% 0.26 ± 25% -86.0% 0.24 ± 6% perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc.vm_area_alloc.mmap_region.do_mmap 1730 ± 24% -76.3% 409.21 ±104% -50.1% 862.65 ± 89% perf-sched.wait_time.max.ms.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm 34.78 ± 43% -98.9% 0.38 ± 12% -98.8% 0.41 ± 10% perf-sched.wait_time.max.ms.__cond_resched.tlb_batch_pages_flush.tlb_finish_mmu.unmap_region.constprop 8.04 ±179% -96.0% 0.32 ± 18% -95.7% 0.35 ± 19% perf-sched.wait_time.max.ms.__cond_resched.unmap_vmas.unmap_region.constprop.0 4.68 ±155% -93.4% 0.31 ± 24% -93.9% 0.28 ± 21% perf-sched.wait_time.max.ms.__cond_resched.zap_pmd_range.isra.0.unmap_page_range 3.42 ± 5% +55.9% 5.33 ± 3% +47.3% 5.03 ± 6% perf-sched.wait_time.max.ms.devkmsg_read.vfs_read.ksys_read.do_syscall_64 9.47 ±134% -96.3% 0.35 ± 17% -96.1% 0.37 ± 8% perf-sched.wait_time.max.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_exc_page_fault 1.87 ± 10% -60.9% 0.73 ±164% -85.3% 0.28 ± 24% perf-sched.wait_time.max.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt 2.39 ±185% -97.8% 0.05 ±165% -98.0% 0.05 ±177% perf-sched.wait_time.max.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_reschedule_ipi 3.95 ± 6% +59.9% 6.32 ± 5% +57.6% 6.23 ± 6% perf-sched.wait_time.max.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone 3.45 ± 5% +58.1% 5.45 ± 3% +50.4% 5.19 ± 7% perf-sched.wait_time.max.ms.syslog_print.do_syslog.kmsg_read.vfs_read 56.55 ± 2% -55.1 1.45 ± 2% -55.1 1.44 ± 2% perf-profile.calltrace.cycles-pp.__munmap 56.06 ± 2% -55.1 0.96 ± 2% -55.1 0.96 ± 2% perf-profile.calltrace.cycles-pp.unmap_region.do_vmi_align_munmap.do_vmi_munmap.__vm_munmap.__x64_sys_munmap 56.50 ± 2% -55.1 1.44 -55.1 1.44 ± 2% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.__munmap 56.50 ± 2% -55.1 1.44 ± 2% -55.1 1.43 ± 2% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.__munmap 56.47 ± 2% -55.0 1.43 -55.0 1.42 ± 2% perf-profile.calltrace.cycles-pp.__vm_munmap.__x64_sys_munmap.do_syscall_64.entry_SYSCALL_64_after_hwframe.__munmap 56.48 ± 2% -55.0 1.44 ± 2% -55.0 1.43 ± 2% perf-profile.calltrace.cycles-pp.__x64_sys_munmap.do_syscall_64.entry_SYSCALL_64_after_hwframe.__munmap 56.45 ± 2% -55.0 1.42 -55.0 1.42 ± 2% perf-profile.calltrace.cycles-pp.do_vmi_munmap.__vm_munmap.__x64_sys_munmap.do_syscall_64.entry_SYSCALL_64_after_hwframe 56.40 ± 2% -55.0 1.40 ± 2% -55.0 1.39 ± 2% perf-profile.calltrace.cycles-pp.do_vmi_align_munmap.do_vmi_munmap.__vm_munmap.__x64_sys_munmap.do_syscall_64 35.28 -34.6 0.66 -34.6 0.66 perf-profile.calltrace.cycles-pp.tlb_finish_mmu.unmap_region.do_vmi_align_munmap.do_vmi_munmap.__vm_munmap 35.17 -34.6 0.57 -34.6 0.57 ± 2% perf-profile.calltrace.cycles-pp.tlb_batch_pages_flush.tlb_finish_mmu.unmap_region.do_vmi_align_munmap.do_vmi_munmap 35.11 -34.5 0.57 -34.5 0.56 perf-profile.calltrace.cycles-pp.release_pages.tlb_batch_pages_flush.tlb_finish_mmu.unmap_region.do_vmi_align_munmap 18.40 ± 7% -18.4 0.00 -18.4 0.00 perf-profile.calltrace.cycles-pp.do_anonymous_page.__handle_mm_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault 17.42 ± 7% -17.4 0.00 -17.4 0.00 perf-profile.calltrace.cycles-pp.lru_add_drain.unmap_region.do_vmi_align_munmap.do_vmi_munmap.__vm_munmap 17.42 ± 7% -17.4 0.00 -17.4 0.00 perf-profile.calltrace.cycles-pp.lru_add_drain_cpu.lru_add_drain.unmap_region.do_vmi_align_munmap.do_vmi_munmap 17.41 ± 7% -17.4 0.00 -17.4 0.00 perf-profile.calltrace.cycles-pp.folio_batch_move_lru.lru_add_drain_cpu.lru_add_drain.unmap_region.do_vmi_align_munmap 17.23 ± 6% -17.2 0.00 -17.2 0.00 perf-profile.calltrace.cycles-pp.__mem_cgroup_uncharge_list.release_pages.tlb_batch_pages_flush.tlb_finish_mmu.unmap_region 16.09 ± 8% -16.1 0.00 -16.1 0.00 perf-profile.calltrace.cycles-pp.folio_lruvec_lock_irqsave.release_pages.tlb_batch_pages_flush.tlb_finish_mmu.unmap_region 16.02 ± 8% -16.0 0.00 -16.0 0.00 perf-profile.calltrace.cycles-pp.folio_lruvec_lock_irqsave.folio_batch_move_lru.lru_add_drain_cpu.lru_add_drain.unmap_region 15.95 ± 8% -16.0 0.00 -16.0 0.00 perf-profile.calltrace.cycles-pp._raw_spin_lock_irqsave.folio_lruvec_lock_irqsave.release_pages.tlb_batch_pages_flush.tlb_finish_mmu 15.89 ± 8% -15.9 0.00 -15.9 0.00 perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_irqsave.folio_lruvec_lock_irqsave.release_pages.tlb_batch_pages_flush 15.86 ± 8% -15.9 0.00 -15.9 0.00 perf-profile.calltrace.cycles-pp._raw_spin_lock_irqsave.folio_lruvec_lock_irqsave.folio_batch_move_lru.lru_add_drain_cpu.lru_add_drain 15.82 ± 8% -15.8 0.00 -15.8 0.00 perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_irqsave.folio_lruvec_lock_irqsave.folio_batch_move_lru.lru_add_drain_cpu 9.32 ± 9% -9.3 0.00 -9.3 0.00 perf-profile.calltrace.cycles-pp.uncharge_folio.__mem_cgroup_uncharge_list.release_pages.tlb_batch_pages_flush.tlb_finish_mmu 8.52 ± 8% -8.5 0.00 -8.5 0.00 perf-profile.calltrace.cycles-pp.__mem_cgroup_charge.do_anonymous_page.__handle_mm_fault.handle_mm_fault.do_user_addr_fault 7.90 ± 4% -7.9 0.00 -7.9 0.00 perf-profile.calltrace.cycles-pp.uncharge_batch.__mem_cgroup_uncharge_list.release_pages.tlb_batch_pages_flush.tlb_finish_mmu 7.56 ± 6% -7.6 0.00 -7.6 0.00 perf-profile.calltrace.cycles-pp.__pte_alloc.do_anonymous_page.__handle_mm_fault.handle_mm_fault.do_user_addr_fault 7.55 ± 6% -7.6 0.00 -7.6 0.00 perf-profile.calltrace.cycles-pp.pte_alloc_one.__pte_alloc.do_anonymous_page.__handle_mm_fault.handle_mm_fault 6.51 ± 8% -6.5 0.00 -6.5 0.00 perf-profile.calltrace.cycles-pp.alloc_pages_mpol.pte_alloc_one.__pte_alloc.do_anonymous_page.__handle_mm_fault 6.51 ± 8% -6.5 0.00 -6.5 0.00 perf-profile.calltrace.cycles-pp.__alloc_pages.alloc_pages_mpol.pte_alloc_one.__pte_alloc.do_anonymous_page 6.41 ± 8% -6.4 0.00 -6.4 0.00 perf-profile.calltrace.cycles-pp.__memcg_kmem_charge_page.__alloc_pages.alloc_pages_mpol.pte_alloc_one.__pte_alloc 0.00 +0.5 0.54 ± 4% +0.6 0.55 ± 3% perf-profile.calltrace.cycles-pp.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.clear_page_erms.clear_huge_page.__do_huge_pmd_anonymous_page 0.00 +0.7 0.70 ± 3% +0.7 0.71 ± 2% perf-profile.calltrace.cycles-pp.asm_sysvec_apic_timer_interrupt.clear_page_erms.clear_huge_page.__do_huge_pmd_anonymous_page.__handle_mm_fault 0.00 +1.4 1.39 +1.4 1.38 ± 3% perf-profile.calltrace.cycles-pp.__cond_resched.clear_huge_page.__do_huge_pmd_anonymous_page.__handle_mm_fault.handle_mm_fault 19.16 ± 6% +57.0 76.21 +57.5 76.66 perf-profile.calltrace.cycles-pp.asm_exc_page_fault 19.09 ± 6% +57.1 76.16 +57.5 76.61 perf-profile.calltrace.cycles-pp.do_user_addr_fault.exc_page_fault.asm_exc_page_fault 19.10 ± 6% +57.1 76.17 +57.5 76.61 perf-profile.calltrace.cycles-pp.exc_page_fault.asm_exc_page_fault 18.99 ± 6% +57.1 76.14 +57.6 76.58 perf-profile.calltrace.cycles-pp.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault 18.43 ± 7% +57.7 76.11 +58.1 76.56 perf-profile.calltrace.cycles-pp.__handle_mm_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault 0.00 +73.0 73.00 +73.5 73.46 perf-profile.calltrace.cycles-pp.clear_page_erms.clear_huge_page.__do_huge_pmd_anonymous_page.__handle_mm_fault.handle_mm_fault 0.00 +75.1 75.15 +75.6 75.60 perf-profile.calltrace.cycles-pp.clear_huge_page.__do_huge_pmd_anonymous_page.__handle_mm_fault.handle_mm_fault.do_user_addr_fault 0.00 +75.9 75.92 +76.4 76.37 perf-profile.calltrace.cycles-pp.__do_huge_pmd_anonymous_page.__handle_mm_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault 58.03 ± 2% -56.0 2.05 -56.0 2.03 perf-profile.children.cycles-pp.entry_SYSCALL_64_after_hwframe 58.02 ± 2% -56.0 2.04 -56.0 2.02 perf-profile.children.cycles-pp.do_syscall_64 56.57 ± 2% -55.1 1.45 ± 2% -55.1 1.45 ± 2% perf-profile.children.cycles-pp.__munmap 56.06 ± 2% -55.1 0.97 -55.1 0.96 perf-profile.children.cycles-pp.unmap_region 56.51 ± 2% -55.1 1.43 -55.1 1.42 ± 2% perf-profile.children.cycles-pp.do_vmi_munmap 56.48 ± 2% -55.0 1.43 ± 2% -55.0 1.43 ± 2% perf-profile.children.cycles-pp.__vm_munmap 56.48 ± 2% -55.0 1.44 ± 2% -55.0 1.43 ± 2% perf-profile.children.cycles-pp.__x64_sys_munmap 56.40 ± 2% -55.0 1.40 -55.0 1.39 ± 2% perf-profile.children.cycles-pp.do_vmi_align_munmap 35.28 -34.6 0.66 -34.6 0.66 perf-profile.children.cycles-pp.tlb_finish_mmu 35.18 -34.6 0.58 -34.6 0.57 perf-profile.children.cycles-pp.tlb_batch_pages_flush 35.16 -34.6 0.57 -34.6 0.57 perf-profile.children.cycles-pp.release_pages 32.12 ± 8% -32.1 0.05 -32.1 0.04 ± 37% perf-profile.children.cycles-pp.folio_lruvec_lock_irqsave 31.85 ± 8% -31.8 0.06 -31.8 0.06 ± 5% perf-profile.children.cycles-pp._raw_spin_lock_irqsave 31.74 ± 8% -31.7 0.00 -31.7 0.00 perf-profile.children.cycles-pp.native_queued_spin_lock_slowpath 18.40 ± 7% -18.4 0.00 -18.4 0.00 perf-profile.children.cycles-pp.do_anonymous_page 17.43 ± 7% -17.4 0.00 -17.4 0.00 perf-profile.children.cycles-pp.lru_add_drain 17.43 ± 7% -17.4 0.00 -17.4 0.00 perf-profile.children.cycles-pp.lru_add_drain_cpu 17.43 ± 7% -17.3 0.10 ± 5% -17.3 0.10 ± 3% perf-profile.children.cycles-pp.folio_batch_move_lru 17.23 ± 6% -17.2 0.00 -17.2 0.00 perf-profile.children.cycles-pp.__mem_cgroup_uncharge_list 9.32 ± 9% -9.3 0.00 -9.3 0.00 perf-profile.children.cycles-pp.uncharge_folio 8.57 ± 8% -8.4 0.16 ± 4% -8.4 0.15 ± 4% perf-profile.children.cycles-pp.__mem_cgroup_charge 7.90 ± 4% -7.8 0.14 ± 5% -7.8 0.14 ± 4% perf-profile.children.cycles-pp.uncharge_batch 7.57 ± 6% -7.6 0.00 -7.6 0.00 perf-profile.children.cycles-pp.__pte_alloc 7.55 ± 6% -7.4 0.16 ± 3% -7.4 0.16 ± 3% perf-profile.children.cycles-pp.pte_alloc_one 6.54 ± 2% -6.5 0.00 -6.5 0.00 perf-profile.children.cycles-pp.__mod_memcg_lruvec_state 6.59 ± 8% -6.4 0.22 ± 2% -6.4 0.22 ± 3% perf-profile.children.cycles-pp.alloc_pages_mpol 6.58 ± 8% -6.4 0.21 ± 2% -6.4 0.22 ± 2% perf-profile.children.cycles-pp.__alloc_pages 6.41 ± 8% -6.3 0.07 ± 5% -6.3 0.07 ± 5% perf-profile.children.cycles-pp.__memcg_kmem_charge_page 4.48 ± 2% -4.3 0.18 ± 4% -4.3 0.18 ± 3% perf-profile.children.cycles-pp.__mod_lruvec_page_state 3.08 ± 4% -3.0 0.09 ± 7% -3.0 0.09 ± 6% perf-profile.children.cycles-pp.page_counter_uncharge 1.74 ± 8% -1.6 0.10 -1.6 0.10 ± 4% perf-profile.children.cycles-pp.kmem_cache_alloc 1.72 ± 2% -1.5 0.23 ± 2% -1.5 0.23 ± 4% perf-profile.children.cycles-pp.unmap_vmas 1.71 ± 2% -1.5 0.22 ± 3% -1.5 0.22 ± 4% perf-profile.children.cycles-pp.unmap_page_range 1.70 ± 2% -1.5 0.21 ± 3% -1.5 0.21 ± 4% perf-profile.children.cycles-pp.zap_pmd_range 1.36 ± 16% -1.3 0.09 ± 4% -1.3 0.09 ± 4% perf-profile.children.cycles-pp.native_irq_return_iret 1.18 ± 2% -1.1 0.08 ± 7% -1.1 0.08 ± 5% perf-profile.children.cycles-pp.page_remove_rmap 1.16 ± 2% -1.1 0.08 ± 4% -1.1 0.07 ± 6% perf-profile.children.cycles-pp.folio_add_new_anon_rmap 1.45 ± 6% -1.0 0.44 ± 2% -1.0 0.44 ± 2% perf-profile.children.cycles-pp.__mmap 1.05 -1.0 0.06 ± 7% -1.0 0.06 ± 7% perf-profile.children.cycles-pp.lru_add_fn 1.03 ± 7% -1.0 0.04 ± 37% -1.0 0.04 ± 37% perf-profile.children.cycles-pp.__anon_vma_prepare 1.38 ± 6% -1.0 0.42 ± 3% -1.0 0.42 ± 2% perf-profile.children.cycles-pp.vm_mmap_pgoff 1.33 ± 6% -0.9 0.40 ± 2% -0.9 0.40 ± 2% perf-profile.children.cycles-pp.do_mmap 0.93 ± 11% -0.9 0.03 ± 77% -0.9 0.02 ±100% perf-profile.children.cycles-pp.memcg_slab_post_alloc_hook 1.17 ± 7% -0.8 0.34 ± 2% -0.8 0.34 ± 2% perf-profile.children.cycles-pp.mmap_region 0.87 ± 5% -0.8 0.06 ± 5% -0.8 0.06 ± 9% perf-profile.children.cycles-pp.kmem_cache_free 0.89 ± 5% -0.7 0.19 ± 4% -0.7 0.20 ± 2% perf-profile.children.cycles-pp.rcu_do_batch 0.89 ± 5% -0.7 0.20 ± 4% -0.7 0.20 ± 3% perf-profile.children.cycles-pp.rcu_core 0.90 ± 5% -0.7 0.21 ± 4% -0.7 0.21 ± 2% perf-profile.children.cycles-pp.__do_softirq 0.74 ± 6% -0.7 0.06 ± 5% -0.7 0.06 ± 8% perf-profile.children.cycles-pp.irq_exit_rcu 0.72 ± 10% -0.7 0.06 ± 5% -0.7 0.06 ± 7% perf-profile.children.cycles-pp.vm_area_alloc 1.01 ± 4% -0.4 0.61 ± 4% -0.4 0.61 ± 2% perf-profile.children.cycles-pp.sysvec_apic_timer_interrupt 0.14 ± 5% -0.1 0.02 ±100% -0.1 0.02 ±100% perf-profile.children.cycles-pp.arch_get_unmapped_area_topdown 0.16 ± 9% -0.1 0.07 ± 7% -0.1 0.07 perf-profile.children.cycles-pp.__slab_free 0.15 ± 3% -0.1 0.06 ± 5% -0.1 0.06 ± 5% perf-profile.children.cycles-pp.get_unmapped_area 0.08 ± 22% -0.0 0.05 ± 41% -0.0 0.04 ± 37% perf-profile.children.cycles-pp.generic_perform_write 0.08 ± 22% -0.0 0.05 ± 41% -0.0 0.04 ± 38% perf-profile.children.cycles-pp.shmem_file_write_iter 0.09 ± 22% -0.0 0.05 ± 43% -0.0 0.05 ± 9% perf-profile.children.cycles-pp.record__pushfn 0.09 ± 22% -0.0 0.05 ± 43% -0.0 0.05 ± 9% perf-profile.children.cycles-pp.writen 0.09 ± 22% -0.0 0.05 ± 43% -0.0 0.05 ± 9% perf-profile.children.cycles-pp.__libc_write 0.11 ± 8% -0.0 0.07 ± 6% -0.0 0.08 ± 6% perf-profile.children.cycles-pp.rcu_cblist_dequeue 0.16 ± 7% -0.0 0.13 ± 4% -0.0 0.13 ± 3% perf-profile.children.cycles-pp.try_charge_memcg 0.09 ± 22% -0.0 0.07 ± 18% -0.0 0.06 ± 8% perf-profile.children.cycles-pp.vfs_write 0.09 ± 22% -0.0 0.07 ± 18% -0.0 0.06 ± 11% perf-profile.children.cycles-pp.ksys_write 0.15 ± 4% -0.0 0.13 ± 3% -0.0 0.13 ± 2% perf-profile.children.cycles-pp.get_page_from_freelist 0.09 -0.0 0.08 ± 4% -0.0 0.08 perf-profile.children.cycles-pp.flush_tlb_mm_range 0.06 +0.0 0.09 ± 4% +0.0 0.08 ± 5% perf-profile.children.cycles-pp.rcu_all_qs 0.17 ± 6% +0.0 0.20 ± 4% +0.0 0.20 ± 3% perf-profile.children.cycles-pp.kthread 0.17 ± 6% +0.0 0.20 ± 4% +0.0 0.20 ± 3% perf-profile.children.cycles-pp.ret_from_fork_asm 0.17 ± 6% +0.0 0.20 ± 4% +0.0 0.20 ± 3% perf-profile.children.cycles-pp.ret_from_fork 0.12 ± 4% +0.0 0.16 ± 3% +0.0 0.16 ± 2% perf-profile.children.cycles-pp.mas_store_prealloc 0.08 ± 6% +0.0 0.12 ± 2% +0.0 0.12 ± 4% perf-profile.children.cycles-pp.vma_alloc_folio 0.00 +0.0 0.04 ± 37% +0.1 0.05 perf-profile.children.cycles-pp.memcg_check_events 0.00 +0.0 0.04 ± 37% +0.1 0.05 perf-profile.children.cycles-pp.thp_get_unmapped_area 0.00 +0.1 0.05 +0.0 0.04 ± 57% perf-profile.children.cycles-pp.free_tail_page_prepare 0.00 +0.1 0.05 +0.1 0.05 perf-profile.children.cycles-pp.mas_destroy 0.00 +0.1 0.05 ± 9% +0.1 0.05 ± 9% perf-profile.children.cycles-pp.update_load_avg 0.00 +0.1 0.06 ± 7% +0.1 0.07 ± 7% perf-profile.children.cycles-pp.native_flush_tlb_one_user 0.00 +0.1 0.07 ± 7% +0.1 0.07 ± 6% perf-profile.children.cycles-pp.__page_cache_release 0.00 +0.1 0.07 ± 4% +0.1 0.07 ± 5% perf-profile.children.cycles-pp.mas_topiary_replace 0.08 ± 5% +0.1 0.16 ± 3% +0.1 0.15 ± 3% perf-profile.children.cycles-pp.mas_alloc_nodes 0.00 +0.1 0.08 ± 4% +0.1 0.08 ± 6% perf-profile.children.cycles-pp.prep_compound_page 0.08 ± 6% +0.1 0.17 ± 5% +0.1 0.18 ± 5% perf-profile.children.cycles-pp.task_tick_fair 0.00 +0.1 0.10 ± 5% +0.1 0.10 ± 4% perf-profile.children.cycles-pp.folio_add_lru_vma 0.00 +0.1 0.11 ± 4% +0.1 0.11 ± 5% perf-profile.children.cycles-pp.__kmem_cache_alloc_bulk 0.00 +0.1 0.12 ± 2% +0.1 0.12 ± 3% perf-profile.children.cycles-pp.kmem_cache_alloc_bulk 0.00 +0.1 0.13 ± 3% +0.1 0.13 ± 2% perf-profile.children.cycles-pp.mas_split 0.00 +0.1 0.13 +0.1 0.13 ± 3% perf-profile.children.cycles-pp._raw_spin_lock 0.11 ± 4% +0.1 0.24 ± 3% +0.1 0.25 ± 4% perf-profile.children.cycles-pp.scheduler_tick 0.00 +0.1 0.14 ± 4% +0.1 0.14 ± 5% perf-profile.children.cycles-pp.__mem_cgroup_uncharge 0.00 +0.1 0.14 ± 3% +0.1 0.14 ± 3% perf-profile.children.cycles-pp.mas_wr_bnode 0.00 +0.1 0.14 ± 5% +0.1 0.14 ± 3% perf-profile.children.cycles-pp.destroy_large_folio 0.00 +0.1 0.15 ± 4% +0.1 0.15 ± 4% perf-profile.children.cycles-pp.mas_spanning_rebalance 0.00 +0.1 0.15 ± 2% +0.2 0.15 ± 4% perf-profile.children.cycles-pp.zap_huge_pmd 0.00 +0.2 0.17 ± 3% +0.2 0.17 ± 3% perf-profile.children.cycles-pp.do_huge_pmd_anonymous_page 0.19 ± 3% +0.2 0.38 +0.2 0.38 ± 2% perf-profile.children.cycles-pp.mas_store_gfp 0.00 +0.2 0.19 ± 3% +0.2 0.18 ± 4% perf-profile.children.cycles-pp.__mod_node_page_state 0.00 +0.2 0.20 ± 3% +0.2 0.20 ± 4% perf-profile.children.cycles-pp.__mod_lruvec_state 0.12 ± 3% +0.2 0.35 +0.2 0.36 ± 3% perf-profile.children.cycles-pp.update_process_times 0.12 ± 3% +0.2 0.36 ± 2% +0.2 0.36 ± 2% perf-profile.children.cycles-pp.tick_sched_handle 0.14 ± 3% +0.2 0.39 +0.3 0.40 ± 4% perf-profile.children.cycles-pp.tick_nohz_highres_handler 0.27 ± 2% +0.3 0.52 ± 3% +0.3 0.52 ± 3% perf-profile.children.cycles-pp.hrtimer_interrupt 0.27 ± 2% +0.3 0.52 ± 4% +0.3 0.53 ± 3% perf-profile.children.cycles-pp.__sysvec_apic_timer_interrupt 0.21 ± 4% +0.3 0.48 ± 3% +0.3 0.48 ± 2% perf-profile.children.cycles-pp.__hrtimer_run_queues 0.00 +0.3 0.31 ± 2% +0.3 0.31 ± 3% perf-profile.children.cycles-pp.mas_wr_spanning_store 0.00 +0.4 0.38 +0.4 0.38 ± 2% perf-profile.children.cycles-pp.free_unref_page_prepare 0.00 +0.4 0.39 +0.4 0.40 perf-profile.children.cycles-pp.free_unref_page 0.13 ± 4% +1.3 1.42 +1.3 1.41 ± 3% perf-profile.children.cycles-pp.__cond_resched 19.19 ± 6% +57.0 76.23 +57.5 76.68 perf-profile.children.cycles-pp.asm_exc_page_fault 19.11 ± 6% +57.1 76.18 +57.5 76.63 perf-profile.children.cycles-pp.exc_page_fault 19.10 ± 6% +57.1 76.18 +57.5 76.62 perf-profile.children.cycles-pp.do_user_addr_fault 19.00 ± 6% +57.1 76.15 +57.6 76.59 perf-profile.children.cycles-pp.handle_mm_fault 18.44 ± 7% +57.7 76.12 +58.1 76.57 perf-profile.children.cycles-pp.__handle_mm_fault 0.06 ± 9% +73.3 73.38 +73.8 73.84 perf-profile.children.cycles-pp.clear_page_erms 0.00 +75.2 75.25 +75.7 75.70 perf-profile.children.cycles-pp.clear_huge_page 0.00 +75.9 75.92 +76.4 76.37 perf-profile.children.cycles-pp.__do_huge_pmd_anonymous_page 31.74 ± 8% -31.7 0.00 -31.7 0.00 perf-profile.self.cycles-pp.native_queued_spin_lock_slowpath 9.22 ± 9% -9.2 0.00 -9.2 0.00 perf-profile.self.cycles-pp.uncharge_folio 6.50 ± 2% -6.5 0.00 -6.5 0.00 perf-profile.self.cycles-pp.__mod_memcg_lruvec_state 5.56 ± 9% -5.6 0.00 -5.6 0.00 perf-profile.self.cycles-pp.__memcg_kmem_charge_page 1.94 ± 4% -1.9 0.08 ± 8% -1.9 0.08 ± 7% perf-profile.self.cycles-pp.page_counter_uncharge 1.36 ± 16% -1.3 0.09 ± 4% -1.3 0.09 ± 4% perf-profile.self.cycles-pp.native_irq_return_iret 0.16 ± 9% -0.1 0.07 ± 7% -0.1 0.07 perf-profile.self.cycles-pp.__slab_free 0.10 ± 8% -0.0 0.07 ± 6% -0.0 0.08 ± 6% perf-profile.self.cycles-pp.rcu_cblist_dequeue 0.07 ± 7% +0.0 0.08 ± 5% +0.0 0.08 ± 7% perf-profile.self.cycles-pp.page_counter_try_charge 0.00 +0.1 0.06 ± 7% +0.1 0.07 ± 7% perf-profile.self.cycles-pp.native_flush_tlb_one_user 0.01 ±264% +0.1 0.07 ± 4% +0.1 0.07 perf-profile.self.cycles-pp.rcu_all_qs 0.00 +0.1 0.07 ± 4% +0.1 0.07 ± 4% perf-profile.self.cycles-pp.__do_huge_pmd_anonymous_page 0.00 +0.1 0.08 ± 6% +0.1 0.08 ± 6% perf-profile.self.cycles-pp.prep_compound_page 0.00 +0.1 0.08 ± 5% +0.1 0.08 ± 6% perf-profile.self.cycles-pp.__kmem_cache_alloc_bulk 0.00 +0.1 0.13 ± 2% +0.1 0.13 ± 2% perf-profile.self.cycles-pp._raw_spin_lock 0.00 +0.2 0.18 ± 3% +0.2 0.18 ± 4% perf-profile.self.cycles-pp.__mod_node_page_state 0.00 +0.3 0.30 ± 2% +0.3 0.30 perf-profile.self.cycles-pp.free_unref_page_prepare 0.00 +0.6 0.58 ± 3% +0.6 0.58 ± 5% perf-profile.self.cycles-pp.clear_huge_page 0.08 ± 4% +1.2 1.25 +1.2 1.24 ± 4% perf-profile.self.cycles-pp.__cond_resched 0.05 ± 9% +72.8 72.81 +73.2 73.26 perf-profile.self.cycles-pp.clear_page_erms