* [linus:master] [mm, swap] 7277433096: swapin.throughput 33.0% regression
@ 2025-04-24 8:39 kernel test robot
2025-04-24 9:38 ` Kairui Song
0 siblings, 1 reply; 2+ messages in thread
From: kernel test robot @ 2025-04-24 8:39 UTC (permalink / raw)
To: Kairui Song
Cc: oe-lkp, lkp, linux-kernel, Andrew Morton, Chris Li, Huang, Ying,
Baoquan He, Barry Song, Hugh Dickens, Johannes Weiner,
Kalesh Singh, Nhat Pham, Ryan Roberts, Yosry Ahmed, linux-mm,
oliver.sang
Hello,
note:
from commit message, this regression should be expected. we still make out this
report FYI what's the possible impact of this change.
below details just FYI.
kernel test robot noticed a 33.0% regression of swapin.throughput on:
commit: 7277433096f6ce4a84a1620529ac4ba3e1041ee1 ("mm, swap: remove old allocation path for HDD")
https://git.kernel.org/cgit/linux/kernel/git/torvalds/linux.git master
[still regression on linus/master bc3372351d0c8b2726b7d4229b878342e3e6b0e8]
[still regression on linux-next/master 6ac908f24cd7ddae52c496bbc888e97ee7b033ac]
testcase: swapin
config: x86_64-rhel-9.4
compiler: gcc-12
test machine: 16 threads 1 sockets Intel(R) Xeon(R) E-2278G CPU @ 3.40GHz (Coffee Lake-E) with 32G memory
parameters:
disk: 1HDD
size: 8G
nr_task: 8
cpufreq_governor: performance
If you fix the issue in a separate patch/commit (i.e. not just a new version of
the same patch/commit), kindly add following tags
| Reported-by: kernel test robot <oliver.sang@intel.com>
| Closes: https://lore.kernel.org/oe-lkp/202504241621.f27743ec-lkp@intel.com
Details are as below:
-------------------------------------------------------------------------------------------------->
The kernel config and materials to reproduce are available at:
https://download.01.org/0day-ci/archive/20250424/202504241621.f27743ec-lkp@intel.com
=========================================================================================
compiler/cpufreq_governor/disk/kconfig/nr_task/rootfs/size/tbox_group/testcase:
gcc-12/performance/1HDD/x86_64-rhel-9.4/8/debian-12-x86_64-20240206.cgz/8G/lkp-cfl-e1/swapin
commit:
e027ec414f ("mm, swap: fold swap_info_get_cont in the only caller")
7277433096 ("mm, swap: remove old allocation path for HDD")
e027ec414fe8f540 7277433096f6ce4a84a1620529a
---------------- ---------------------------
%stddev %change %stddev
\ | \
5.409e+09 +51.8% 8.212e+09 ± 5% cpuidle..time
1037299 +18.5% 1228926 ± 12% cpuidle..usage
68.75 +9.2% 75.07 ± 2% iostat.cpu.idle
30.53 ± 2% -19.8% 24.47 ± 8% iostat.cpu.iowait
797.42 ± 4% +22.0% 972.57 ± 7% uptime.boot
9703 ± 4% +26.4% 12265 ± 8% uptime.idle
20243 ± 57% +248.6% 70559 ± 30% meminfo.Inactive
20238 ± 57% +248.6% 70559 ± 30% meminfo.Inactive(anon)
24668 +11.3% 27444 meminfo.Shmem
0.01 ± 14% +33.3% 0.02 ± 5% perf-sched.sch_delay.avg.ms.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
7023 ± 3% -13.1% 6105 ± 8% perf-sched.total_wait_and_delay.count.ms
4615 ± 4% -18.8% 3747 ± 12% perf-sched.wait_and_delay.count.io_schedule.folio_wait_bit_common.__folio_lock_or_retry.do_swap_page
68.63 +6.4 75.02 ± 2% mpstat.cpu.all.idle%
30.66 ± 2% -6.1 24.54 ± 8% mpstat.cpu.all.iowait%
0.09 -0.0 0.06 ± 4% mpstat.cpu.all.irq%
0.05 -0.0 0.04 ± 3% mpstat.cpu.all.soft%
0.28 -0.1 0.20 ± 5% mpstat.cpu.all.sys%
0.29 ± 21% -0.1 0.15 ± 4% mpstat.cpu.all.usr%
0.09 ± 12% -8.7% 0.08 swapin.free_time
3435 ± 2% -33.7% 2277 ± 5% swapin.median
47.13 ± 6% +8.4 55.50 ± 7% swapin.stddev%
26855 ± 2% -33.0% 18001 ± 5% swapin.throughput
339.27 +51.6% 514.32 ± 5% swapin.time.elapsed_time
339.27 +51.6% 514.32 ± 5% swapin.time.elapsed_time.max
12785 +51.4% 19352 ± 5% swapin.time.minor_page_faults
68.81 +9.1% 75.06 ± 2% vmstat.cpu.id
30.57 ± 2% -19.6% 24.56 ± 8% vmstat.cpu.wa
24857 -33.8% 16465 ± 5% vmstat.io.bi
1.30 ± 5% -7.1% 1.21 ± 5% vmstat.procs.r
24857 -33.8% 16465 ± 5% vmstat.swap.si
7.12 ± 7% -31.6% 4.87 ± 14% vmstat.swap.so
2401 -28.3% 1721 ± 3% vmstat.system.cs
2399 -24.5% 1811 ± 2% vmstat.system.in
936464 +6.4% 996818 proc-vmstat.nr_active_anon
922127 +8.3% 998395 proc-vmstat.nr_anon_pages
612143 -1.2% 605002 proc-vmstat.nr_dirty_background_threshold
1225784 -1.2% 1211485 proc-vmstat.nr_dirty_threshold
1803299 +3.7% 1869638 proc-vmstat.nr_file_pages
6195937 -1.2% 6124422 proc-vmstat.nr_free_pages
5059 ± 57% +248.6% 17639 ± 30% proc-vmstat.nr_inactive_anon
5486 -1.4% 5410 proc-vmstat.nr_page_table_pages
6172 +11.2% 6866 proc-vmstat.nr_shmem
14607 +5.3% 15384 proc-vmstat.nr_slab_unreclaimable
920671 +7.1% 986314 ± 2% proc-vmstat.nr_swapcached
4413040 +1.8% 4493858 proc-vmstat.nr_vmscan_write
936464 +6.4% 996818 proc-vmstat.nr_zone_active_anon
5059 ± 57% +248.6% 17639 ± 30% proc-vmstat.nr_zone_inactive_anon
2535302 +5.8% 2682756 proc-vmstat.numa_hit
2535395 +5.8% 2681644 proc-vmstat.numa_local
2543613 +5.9% 2694193 proc-vmstat.pgalloc_normal
2629484 +8.0% 2838539 proc-vmstat.pgfault
2520197 +6.0% 2672439 proc-vmstat.pgfree
293008 -6.8% 273047 proc-vmstat.pgmajfault
25587 ± 2% +41.3% 36143 ± 5% proc-vmstat.pgreuse
933820 +8.8% 1015820 proc-vmstat.swap_ra
919094 +9.8% 1008935 proc-vmstat.swap_ra_hit
16587 -4.3% 15880 proc-vmstat.workingset_nodes
1088601 +7.4% 1169575 proc-vmstat.workingset_refault_anon
2.328e+08 ± 24% -43.2% 1.323e+08 ± 30% sched_debug.cfs_rq:/.avg_vruntime.max
62177217 ± 36% -43.0% 35460373 ± 30% sched_debug.cfs_rq:/.avg_vruntime.stddev
2.328e+08 ± 24% -43.2% 1.323e+08 ± 30% sched_debug.cfs_rq:/.min_vruntime.max
62177217 ± 36% -43.0% 35460373 ± 30% sched_debug.cfs_rq:/.min_vruntime.stddev
210.52 ± 63% +130.6% 485.44 ± 32% sched_debug.cfs_rq:/.removed.load_avg.max
4.68 ± 79% +185.3% 13.36 ± 37% sched_debug.cfs_rq:/.removed.runnable_avg.avg
69.33 ± 72% +195.8% 205.12 ± 38% sched_debug.cfs_rq:/.removed.runnable_avg.max
17.15 ± 74% +192.5% 50.18 ± 38% sched_debug.cfs_rq:/.removed.runnable_avg.stddev
4.58 ± 84% +188.0% 13.19 ± 37% sched_debug.cfs_rq:/.removed.util_avg.avg
67.72 ± 77% +200.5% 203.49 ± 37% sched_debug.cfs_rq:/.removed.util_avg.max
16.76 ± 79% +196.5% 49.70 ± 37% sched_debug.cfs_rq:/.removed.util_avg.stddev
35.31 ± 24% -40.4% 21.03 ± 31% sched_debug.cfs_rq:/.runnable_avg.min
167.49 ± 2% -23.7% 127.74 ± 12% sched_debug.cfs_rq:/.util_avg.avg
602113 ± 5% +16.6% 702121 ± 8% sched_debug.cpu.clock.avg
602114 ± 5% +16.6% 702121 ± 8% sched_debug.cpu.clock.max
602112 ± 5% +16.6% 702120 ± 8% sched_debug.cpu.clock.min
601358 ± 5% +16.6% 701428 ± 8% sched_debug.cpu.clock_task.avg
601962 ± 5% +16.6% 701955 ± 8% sched_debug.cpu.clock_task.max
597107 ± 5% +16.9% 698032 ± 8% sched_debug.cpu.clock_task.min
1227 ± 5% -16.4% 1026 ± 10% sched_debug.cpu.clock_task.stddev
4030 ± 3% +33.1% 5364 ± 2% sched_debug.cpu.curr->pid.max
1037 ± 5% +28.9% 1336 ± 3% sched_debug.cpu.curr->pid.stddev
46279 ± 5% -19.9% 37088 ± 2% sched_debug.cpu.nr_switches.avg
602113 ± 5% +16.6% 702121 ± 8% sched_debug.cpu_clk
601390 ± 5% +16.6% 701418 ± 8% sched_debug.ktime
602514 ± 5% +16.6% 702513 ± 8% sched_debug.sched_clk
50.92 ± 3% +13.3% 57.67 ± 2% perf-stat.i.MPKI
96852867 ± 12% -41.7% 56470262 ± 4% perf-stat.i.branch-instructions
3.59 ± 2% -0.2 3.39 perf-stat.i.branch-miss-rate%
4517791 ± 19% -46.8% 2401426 ± 4% perf-stat.i.branch-misses
45.08 +3.6 48.70 perf-stat.i.cache-miss-rate%
11444503 -25.3% 8552014 ± 5% perf-stat.i.cache-misses
26646448 -30.4% 18535433 ± 3% perf-stat.i.cache-references
2380 -28.5% 1703 ± 3% perf-stat.i.context-switches
1.99 ± 2% +4.4% 2.08 perf-stat.i.cpi
6.453e+08 ± 7% -36.1% 4.124e+08 ± 4% perf-stat.i.cpu-cycles
37.44 ± 3% -29.0% 26.60 ± 3% perf-stat.i.cpu-migrations
74.34 ± 4% -18.6% 60.51 ± 4% perf-stat.i.cycles-between-cache-misses
4.335e+08 ± 14% -42.5% 2.493e+08 ± 4% perf-stat.i.instructions
0.56 ± 3% -7.2% 0.52 perf-stat.i.ipc
865.19 -38.4% 532.92 ± 5% perf-stat.i.major-faults
6605 -27.4% 4795 ± 3% perf-stat.i.minor-faults
7470 -28.7% 5328 ± 4% perf-stat.i.page-faults
4.62 ± 6% -1.1 3.55 ± 44% perf-stat.overall.branch-miss-rate%
56.43 ± 6% -28.5% 40.36 ± 44% perf-stat.overall.cycles-between-cache-misses
0.67 ± 6% -24.7% 0.50 ± 44% perf-stat.overall.ipc
96637819 ± 12% -51.6% 46743598 ± 45% perf-stat.ps.branch-instructions
4506778 ± 19% -55.9% 1988844 ± 45% perf-stat.ps.branch-misses
11410417 -38.1% 7059671 ± 45% perf-stat.ps.cache-misses
26570647 -42.3% 15341807 ± 44% perf-stat.ps.cache-references
2373 -40.5% 1411 ± 44% perf-stat.ps.context-switches
6.442e+08 ± 7% -47.0% 3.417e+08 ± 44% perf-stat.ps.cpu-cycles
37.34 ± 3% -41.0% 22.03 ± 44% perf-stat.ps.cpu-migrations
4.326e+08 ± 14% -52.3% 2.064e+08 ± 45% perf-stat.ps.instructions
862.62 -49.0% 440.19 ± 45% perf-stat.ps.major-faults
6586 -39.8% 3967 ± 44% perf-stat.ps.minor-faults
7449 -40.8% 4407 ± 44% perf-stat.ps.page-faults
6.51 ± 7% -1.2 5.26 ± 5% perf-profile.calltrace.cycles-pp.do_access
5.54 ± 6% -1.0 4.54 ± 5% perf-profile.calltrace.cycles-pp.asm_exc_page_fault.do_access
4.94 ± 5% -0.9 4.03 ± 6% perf-profile.calltrace.cycles-pp.exc_page_fault.asm_exc_page_fault.do_access
4.90 ± 5% -0.9 4.02 ± 6% perf-profile.calltrace.cycles-pp.do_user_addr_fault.exc_page_fault.asm_exc_page_fault.do_access
4.62 ± 6% -0.9 3.74 ± 7% perf-profile.calltrace.cycles-pp.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault.do_access
4.46 ± 6% -0.8 3.62 ± 7% perf-profile.calltrace.cycles-pp.__handle_mm_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault
4.33 ± 7% -0.8 3.54 ± 8% perf-profile.calltrace.cycles-pp.do_swap_page.__handle_mm_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault
2.25 ± 11% -0.4 1.86 ± 10% perf-profile.calltrace.cycles-pp.handle_edge_irq.__common_interrupt.common_interrupt.asm_common_interrupt.cpuidle_enter_state
2.26 ± 11% -0.4 1.87 ± 10% perf-profile.calltrace.cycles-pp.__common_interrupt.common_interrupt.asm_common_interrupt.cpuidle_enter_state.cpuidle_enter
2.20 ± 12% -0.4 1.83 ± 10% perf-profile.calltrace.cycles-pp.handle_irq_event.handle_edge_irq.__common_interrupt.common_interrupt.asm_common_interrupt
0.82 ± 13% -0.1 0.69 ± 7% perf-profile.calltrace.cycles-pp.ahci_handle_port_intr.ahci_single_level_irq_intr.__handle_irq_event_percpu.handle_irq_event.handle_edge_irq
4.14 ± 3% +0.2 4.39 perf-profile.calltrace.cycles-pp.cpuidle_idle_call.do_idle.cpu_startup_entry.rest_init.start_kernel
4.22 ± 3% +0.3 4.49 perf-profile.calltrace.cycles-pp.cpu_startup_entry.rest_init.start_kernel.x86_64_start_reservations.x86_64_start_kernel
4.22 ± 3% +0.3 4.49 perf-profile.calltrace.cycles-pp.rest_init.start_kernel.x86_64_start_reservations.x86_64_start_kernel.common_startup_64
4.22 ± 3% +0.3 4.49 perf-profile.calltrace.cycles-pp.start_kernel.x86_64_start_reservations.x86_64_start_kernel.common_startup_64
4.22 ± 3% +0.3 4.49 perf-profile.calltrace.cycles-pp.x86_64_start_kernel.common_startup_64
4.22 ± 3% +0.3 4.49 perf-profile.calltrace.cycles-pp.x86_64_start_reservations.x86_64_start_kernel.common_startup_64
4.22 ± 3% +0.3 4.49 perf-profile.calltrace.cycles-pp.do_idle.cpu_startup_entry.rest_init.start_kernel.x86_64_start_reservations
3.73 ± 5% +0.3 4.05 ± 2% perf-profile.calltrace.cycles-pp.cpuidle_enter.cpuidle_idle_call.do_idle.cpu_startup_entry.rest_init
83.48 +1.2 84.68 perf-profile.calltrace.cycles-pp.do_idle.cpu_startup_entry.start_secondary.common_startup_64
83.50 +1.2 84.74 perf-profile.calltrace.cycles-pp.start_secondary.common_startup_64
83.50 +1.2 84.74 perf-profile.calltrace.cycles-pp.cpu_startup_entry.start_secondary.common_startup_64
87.73 +1.5 89.22 perf-profile.calltrace.cycles-pp.common_startup_64
7.69 ± 7% -1.4 6.28 ± 6% perf-profile.children.cycles-pp.do_access
6.07 ± 5% -1.0 5.08 ± 3% perf-profile.children.cycles-pp.asm_exc_page_fault
5.40 ± 5% -0.9 4.52 ± 3% perf-profile.children.cycles-pp.exc_page_fault
5.37 ± 5% -0.9 4.50 ± 3% perf-profile.children.cycles-pp.do_user_addr_fault
4.44 ± 7% -0.8 3.60 ± 7% perf-profile.children.cycles-pp.do_swap_page
5.05 ± 6% -0.8 4.22 ± 4% perf-profile.children.cycles-pp.handle_mm_fault
4.88 ± 6% -0.8 4.07 ± 5% perf-profile.children.cycles-pp.__handle_mm_fault
3.56 ± 6% -0.6 2.95 ± 7% perf-profile.children.cycles-pp.common_interrupt
3.56 ± 6% -0.6 2.96 ± 7% perf-profile.children.cycles-pp.asm_common_interrupt
2.48 ± 11% -0.5 1.97 ± 10% perf-profile.children.cycles-pp.handle_edge_irq
2.49 ± 11% -0.5 1.98 ± 10% perf-profile.children.cycles-pp.__common_interrupt
2.42 ± 11% -0.5 1.93 ± 10% perf-profile.children.cycles-pp.handle_irq_event
2.39 ± 12% -0.5 1.91 ± 9% perf-profile.children.cycles-pp.__handle_irq_event_percpu
2.37 ± 13% -0.5 1.90 ± 9% perf-profile.children.cycles-pp.ahci_single_level_irq_intr
1.17 ± 12% -0.3 0.90 ± 11% perf-profile.children.cycles-pp.__schedule
0.74 ± 8% -0.2 0.49 ± 14% perf-profile.children.cycles-pp.__folio_lock_or_retry
0.78 ± 13% -0.2 0.55 ± 12% perf-profile.children.cycles-pp.schedule
0.64 ± 9% -0.2 0.43 ± 15% perf-profile.children.cycles-pp.io_schedule
0.68 ± 7% -0.2 0.47 ± 15% perf-profile.children.cycles-pp.folio_wait_bit_common
0.93 ± 12% -0.2 0.74 ± 5% perf-profile.children.cycles-pp.ahci_handle_port_intr
0.89 ± 9% -0.2 0.74 ± 7% perf-profile.children.cycles-pp.scsi_end_request
0.89 ± 9% -0.1 0.74 ± 8% perf-profile.children.cycles-pp.scsi_io_completion
0.44 ± 21% -0.1 0.30 ± 13% perf-profile.children.cycles-pp.blk_mq_dispatch_rq_list
0.20 ± 20% -0.1 0.09 ± 33% perf-profile.children.cycles-pp.xas_load
0.20 ± 24% -0.1 0.10 ± 29% perf-profile.children.cycles-pp.filemap_get_entry
0.25 ± 14% -0.1 0.16 ± 19% perf-profile.children.cycles-pp.ahci_handle_port_interrupt
0.24 ± 12% -0.1 0.15 ± 17% perf-profile.children.cycles-pp.sata_async_notification
0.23 ± 12% -0.1 0.14 ± 17% perf-profile.children.cycles-pp.ahci_scr_read
0.20 ± 21% -0.1 0.11 ± 25% perf-profile.children.cycles-pp.ata_scsi_queuecmd
0.20 ± 24% -0.1 0.12 ± 23% perf-profile.children.cycles-pp.scsi_dispatch_cmd
0.22 ± 22% -0.1 0.14 ± 27% perf-profile.children.cycles-pp.seq_read_iter
0.21 ± 11% +0.1 0.29 ± 19% perf-profile.children.cycles-pp.__mmput
0.17 ± 15% +0.1 0.26 ± 15% perf-profile.children.cycles-pp.do_pte_missing
0.14 ± 20% +0.1 0.23 ± 20% perf-profile.children.cycles-pp.filemap_map_pages
0.20 ± 13% +0.1 0.29 ± 19% perf-profile.children.cycles-pp.exit_mmap
0.20 ± 18% +0.1 0.29 ± 14% perf-profile.children.cycles-pp.leave_mm
0.15 ± 18% +0.1 0.24 ± 23% perf-profile.children.cycles-pp.do_read_fault
0.24 ± 25% +0.1 0.39 ± 19% perf-profile.children.cycles-pp.rb_next
0.40 ± 35% +0.2 0.57 ± 22% perf-profile.children.cycles-pp.ret_from_fork_asm
0.63 ± 11% +0.2 0.86 ± 12% perf-profile.children.cycles-pp.__hrtimer_next_event_base
4.22 ± 3% +0.3 4.49 perf-profile.children.cycles-pp.rest_init
4.22 ± 3% +0.3 4.49 perf-profile.children.cycles-pp.start_kernel
4.22 ± 3% +0.3 4.49 perf-profile.children.cycles-pp.x86_64_start_kernel
4.22 ± 3% +0.3 4.49 perf-profile.children.cycles-pp.x86_64_start_reservations
83.50 +1.2 84.74 perf-profile.children.cycles-pp.start_secondary
85.97 +1.4 87.32 perf-profile.children.cycles-pp.cpuidle_idle_call
87.73 +1.5 89.22 perf-profile.children.cycles-pp.common_startup_64
87.73 +1.5 89.22 perf-profile.children.cycles-pp.cpu_startup_entry
87.72 +1.5 89.22 perf-profile.children.cycles-pp.do_idle
2.06 ± 6% -0.3 1.71 ± 12% perf-profile.self.cycles-pp.do_access
0.23 ± 12% -0.1 0.14 ± 17% perf-profile.self.cycles-pp.ahci_scr_read
0.12 ± 29% -0.1 0.04 ±101% perf-profile.self.cycles-pp.xas_load
0.22 ± 17% +0.1 0.28 ± 12% perf-profile.self.cycles-pp.update_rq_clock_task
0.24 ± 26% +0.1 0.38 ± 19% perf-profile.self.cycles-pp.rb_next
Disclaimer:
Results have been estimated based on internal Intel analysis and are provided
for informational purposes only. Any difference in system hardware or software
design or configuration may affect actual performance.
--
0-DAY CI Kernel Test Service
https://github.com/intel/lkp-tests/wiki
^ permalink raw reply [flat|nested] 2+ messages in thread* Re: [linus:master] [mm, swap] 7277433096: swapin.throughput 33.0% regression
2025-04-24 8:39 [linus:master] [mm, swap] 7277433096: swapin.throughput 33.0% regression kernel test robot
@ 2025-04-24 9:38 ` Kairui Song
0 siblings, 0 replies; 2+ messages in thread
From: Kairui Song @ 2025-04-24 9:38 UTC (permalink / raw)
To: kernel test robot
Cc: oe-lkp, lkp, linux-kernel, Andrew Morton, Chris Li, Huang, Ying,
Baoquan He, Barry Song, Hugh Dickens, Johannes Weiner,
Kalesh Singh, Nhat Pham, Ryan Roberts, Yosry Ahmed, linux-mm
On Thu, Apr 24, 2025 at 4:40 PM kernel test robot <oliver.sang@intel.com> wrote:
>
>
>
> Hello,
>
>
> note:
> from commit message, this regression should be expected. we still make out this
> report FYI what's the possible impact of this change.
>
> below details just FYI.
Hi,
For most workloads with random swap io, there shouldn't be any
regression. I checked the test and it's doing a sequential swapout
with 8 workers (after a fresh swapon?). I was expecting
bae8a4ef3efb56bb7e83bafd3c0856845aeaf605 ("mm, swap: use a global swap
cluster for non-rotation devices") to optimize it. Seems that's not
enough?
I guess we might also need to disable the cluster shuffle in
setup_clusters for HDD. The shuffle was meant to reduce lock
contention on swap cache, the contention is clearly not an issue for
HDD, the bottleneck for this test is IO, that might be making the IO
more randomized. Or maybe clusters will better be sorted by offset for
HDD.
It seems this test can be easily reproduced with `$USEMEM -n $nr_task
-W -d -p $TMP/pidfile $((size / nr_task))`, I think it's not a real
problem in productions, but I'll try fix this if have time.
^ permalink raw reply [flat|nested] 2+ messages in thread
end of thread, other threads:[~2025-04-24 9:39 UTC | newest]
Thread overview: 2+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2025-04-24 8:39 [linus:master] [mm, swap] 7277433096: swapin.throughput 33.0% regression kernel test robot
2025-04-24 9:38 ` Kairui Song
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox