* [BUG REPORT] mm/damon: softlockup when kdamond walk page with cpu hotplug @ 2025-09-18 3:00 Xinyu Zheng 2025-09-18 8:52 ` SeongJae Park 0 siblings, 1 reply; 8+ messages in thread From: Xinyu Zheng @ 2025-09-18 3:00 UTC (permalink / raw) To: SeongJae Park, Andrew Morton, Paul E . McKenney, Peter Zijlstra Cc: damon, linux-mm, linux-kernel, zouyipeng, zhengxinyu6 A softlockup issue was found with stress test: watchdog: BUG: soft lockup - CPU#0 stuck for 26s! [migration/0:957] CPU: 0 PID: 957 Comm: migration/0 Kdump: loaded Tainted: Stopper: multi_cpu_stop+0x0/0x1e8 <- __stop_cpus.constprop.0+0x5c/0xb0 pstate: 61400009 (nZCv daif +PAN -UAO -TCO +DIT -SSBS BTYPE=--) pc : rcu_momentary_dyntick_idle+0x4c/0xa0 lr : multi_cpu_stop+0x10c/0x1e8 sp : ffff800086013d60 x29: ffff800086013d60 x28: 0000000000000001 x27: 0000000000000000 x26: 0000000000000000 x25: 00000000ffffffff x24: 0000000000000000 x23: 0000000000000001 x22: ffffab8f02977e00 x21: ffff8000b44ebb84 x20: ffff8000b44ebb60 x19: 0000000000000001 x18: 0000000000000000 x17: 000000040044ffff x16: 004000b5b5503510 x15: 0000000000000800 x14: ffff081003921440 x13: ffff5c907c75d000 x12: a34000013454d99d x11: 0000000000000000 x10: 0000000000000f90 x9 : ffffab8f01b657bc x8 : ffff081005e060f0 x7 : ffff081f7fd7b610 x6 : 0000009e0bb34c91 x5 : 00000000480fd060 x4 : ffff081f7fd7b508 x3 : ffff5c907c75d000 x2 : ffff800086013d60 x1 : 00000000b8ccb304 x0 : 00000000b8ccb30c Call trace: rcu_momentary_dyntick_idle+0x4c/0xa0 multi_cpu_stop+0x10c/0x1e8 cpu_stopper_thread+0xdc/0x1c0 smpboot_thread_fn+0x140/0x190 kthread+0xec/0x100 ret_from_fork+0x10/0x20 watchdog: BUG: soft lockup - CPU#18 stuck for 26s! [kdamond.0:408949] CPU: 18 PID: 408949 Comm: kdamond.0 Kdump: loaded Tainted: pstate: 61400009 (nZCv daif +PAN -UAO -TCO +DIT -SSBS BTYPE=--) pc : damon_mkold_pmd_entry+0x138/0x1d8 lr : damon_mkold_pmd_entry+0x68/0x1d8 sp : ffff8000c384bb00 x29: ffff8000c384bb10 x28: 0000ffff6e2a4a9b x27: 0000ffff6e2a4a9b x26: ffff080090fdeb88 x25: 0000ffff6e2a4a9b x24: ffffab8f029a9020 x23: ffff08013eb8dfe8 x22: 0000ffff6e2a4a9c x21: 0000ffff6e2a4a9b x20: ffff8000c384bd08 x19: 0000000000000000 x18: 0000000000000014 x17: 00000000f90a2272 x16: 0000000004c87773 x15: 000000004524349f x14: 00000000ee10aa21 x13: 0000000000000000 x12: ffffab8f02af4818 x11: 0000ffff7e7fffff x10: 0000ffff62700000 x9 : ffffab8f01d2f628 x8 : ffff0800879fbc0c x7 : ffff0800879fbc00 x6 : ffff0800c41c7d88 x5 : 0000000000000171 x4 : ffff08100aab0000 x3 : 00003081088800c0 x2 : 0000000000000000 x1 : 0000000000000000 x0 : 0000000000000000 Call trace: damon_mkold_pmd_entry+0x138/0x1d8 walk_pmd_range.isra.0+0x1ac/0x3a8 walk_pud_range+0x120/0x190 walk_pgd_range+0x170/0x1b8 __walk_page_range+0x184/0x198 walk_page_range+0x124/0x1f0 damon_va_prepare_access_checks+0xb4/0x1b8 kdamond_fn+0x11c/0x690 kthread+0xec/0x100 ret_from_fork+0x10/0x20 The stress test enable numa balance and kdamond, operation involves CPU hotplug and page fault with migration. CPU0 CPU18 events =============================== ====================== =============== page_fault(user task invoke) migrate_pages(pmd page migrate) __schedule kdamond_fn walk_pmd_range damon_mkold_pmd_entry <= cpu hotplug stop_machine_cpuslocked // infinite loop queue_stop_cpus_work // waiting CPU 0 user task multi_cpu_stop(migration/0) // to be scheduled // infinite loop waiting for // cpu 18 ACK Detail explanation: 1. When shutdown one cpu, a state machine in multi_cpu_stop() will wait for other cpu's migration thread reach to same state. In this case, all cpus are doing migration except cpu 18. 2. A user task which is bind on cpu 0 is allocating page and invoke page fault to migrate page. Kdamond is looping between damon_mkold_pmd_entry () and walk_pmd_range(), since target page is a migration entry. Kdamond can end the loop until user task is scheduled on CPU 0. But CPU 0 is running migration/0. 3. CONFIG_PREEMPT_NONE is enable. So all cpu are in a infinite loop. I found a similar softlockup issue which is also invoked by a memory operation with cpu hotplug. To fix that issue, add a cond_resched() to avoid block migration task. https://lore.kernel.org/all/20250211081819.33307-1-chenridong@huaweicloud.com/#t May I ask if there is any solution we can fix this issue? Such as add a cond_sched() in kdamond process. Or is there any chance to do some yield in stop machine process? Probably next time there is another different case running with cpu hotplug can cause the same softlockup. Xinyu Zheng ^ permalink raw reply [flat|nested] 8+ messages in thread
* Re: [BUG REPORT] mm/damon: softlockup when kdamond walk page with cpu hotplug 2025-09-18 3:00 [BUG REPORT] mm/damon: softlockup when kdamond walk page with cpu hotplug Xinyu Zheng @ 2025-09-18 8:52 ` SeongJae Park 2025-09-18 8:57 ` SeongJae Park 2025-09-20 3:56 ` Hugh Dickins 0 siblings, 2 replies; 8+ messages in thread From: SeongJae Park @ 2025-09-18 8:52 UTC (permalink / raw) To: Xinyu Zheng Cc: SeongJae Park, Andrew Morton, Paul E . McKenney, Peter Zijlstra, damon, linux-mm, linux-kernel, zouyipeng, Hugh Dickins Hello, On Thu, 18 Sep 2025 03:00:29 +0000 Xinyu Zheng <zhengxinyu6@huawei.com> wrote: > A softlockup issue was found with stress test: Thank you for sharing this great report! > > watchdog: BUG: soft lockup - CPU#0 stuck for 26s! [migration/0:957] > CPU: 0 PID: 957 Comm: migration/0 Kdump: loaded Tainted: > Stopper: multi_cpu_stop+0x0/0x1e8 <- __stop_cpus.constprop.0+0x5c/0xb0 > pstate: 61400009 (nZCv daif +PAN -UAO -TCO +DIT -SSBS BTYPE=--) > pc : rcu_momentary_dyntick_idle+0x4c/0xa0 > lr : multi_cpu_stop+0x10c/0x1e8 > sp : ffff800086013d60 > x29: ffff800086013d60 x28: 0000000000000001 x27: 0000000000000000 > x26: 0000000000000000 x25: 00000000ffffffff x24: 0000000000000000 > x23: 0000000000000001 x22: ffffab8f02977e00 x21: ffff8000b44ebb84 > x20: ffff8000b44ebb60 x19: 0000000000000001 x18: 0000000000000000 > x17: 000000040044ffff x16: 004000b5b5503510 x15: 0000000000000800 > x14: ffff081003921440 x13: ffff5c907c75d000 x12: a34000013454d99d > x11: 0000000000000000 x10: 0000000000000f90 x9 : ffffab8f01b657bc > x8 : ffff081005e060f0 x7 : ffff081f7fd7b610 x6 : 0000009e0bb34c91 > x5 : 00000000480fd060 x4 : ffff081f7fd7b508 x3 : ffff5c907c75d000 > x2 : ffff800086013d60 x1 : 00000000b8ccb304 x0 : 00000000b8ccb30c > Call trace: > rcu_momentary_dyntick_idle+0x4c/0xa0 > multi_cpu_stop+0x10c/0x1e8 > cpu_stopper_thread+0xdc/0x1c0 > smpboot_thread_fn+0x140/0x190 > kthread+0xec/0x100 > ret_from_fork+0x10/0x20 > > watchdog: BUG: soft lockup - CPU#18 stuck for 26s! [kdamond.0:408949] > CPU: 18 PID: 408949 Comm: kdamond.0 Kdump: loaded Tainted: > pstate: 61400009 (nZCv daif +PAN -UAO -TCO +DIT -SSBS BTYPE=--) > pc : damon_mkold_pmd_entry+0x138/0x1d8 > lr : damon_mkold_pmd_entry+0x68/0x1d8 > sp : ffff8000c384bb00 > x29: ffff8000c384bb10 x28: 0000ffff6e2a4a9b x27: 0000ffff6e2a4a9b > x26: ffff080090fdeb88 x25: 0000ffff6e2a4a9b x24: ffffab8f029a9020 > x23: ffff08013eb8dfe8 x22: 0000ffff6e2a4a9c x21: 0000ffff6e2a4a9b > x20: ffff8000c384bd08 x19: 0000000000000000 x18: 0000000000000014 > x17: 00000000f90a2272 x16: 0000000004c87773 x15: 000000004524349f > x14: 00000000ee10aa21 x13: 0000000000000000 x12: ffffab8f02af4818 > x11: 0000ffff7e7fffff x10: 0000ffff62700000 x9 : ffffab8f01d2f628 > x8 : ffff0800879fbc0c x7 : ffff0800879fbc00 x6 : ffff0800c41c7d88 > x5 : 0000000000000171 x4 : ffff08100aab0000 x3 : 00003081088800c0 > x2 : 0000000000000000 x1 : 0000000000000000 x0 : 0000000000000000 > Call trace: > damon_mkold_pmd_entry+0x138/0x1d8 > walk_pmd_range.isra.0+0x1ac/0x3a8 > walk_pud_range+0x120/0x190 > walk_pgd_range+0x170/0x1b8 > __walk_page_range+0x184/0x198 > walk_page_range+0x124/0x1f0 > damon_va_prepare_access_checks+0xb4/0x1b8 > kdamond_fn+0x11c/0x690 > kthread+0xec/0x100 > ret_from_fork+0x10/0x20 > > The stress test enable numa balance and kdamond, operation > involves CPU hotplug and page fault with migration. > > CPU0 CPU18 events > =============================== ====================== =============== > page_fault(user task invoke) > migrate_pages(pmd page migrate) > __schedule > kdamond_fn > walk_pmd_range > damon_mkold_pmd_entry <= cpu hotplug > stop_machine_cpuslocked // infinite loop > queue_stop_cpus_work // waiting CPU 0 user task > multi_cpu_stop(migration/0) // to be scheduled > // infinite loop waiting for > // cpu 18 ACK > > Detail explanation: > 1. When shutdown one cpu, a state machine in multi_cpu_stop() > will wait for other cpu's migration thread reach to same state. > In this case, all cpus are doing migration except cpu 18. > 2. A user task which is bind on cpu 0 is allocating page and > invoke page fault to migrate page. Kdamond is looping between > damon_mkold_pmd_entry () and walk_pmd_range(), damon_mkold_pmd_entry() calls pte_offset_map_lock(). If the call returns an error, damon_mkold_pmd_entry() sets walk->action as ACTION_AGAIN, to retry. If the pte_offset_map_lock() continues fails, infinite loop can happen. I understand the loop you mentioned above is this case. The error handling (retrying) was introduced by commit 7780d04046a2 ("mm/pagewalkers: ACTION_AGAIN if pte_offset_map_lock() fails"). According to the commit message, it is assumed to be safe to retry pte_offset_map_lock() inside walk_page_range(), but it is not for this corner case. And the commit introduced the error handling pattern in not only DAMON but also a few other pte_offset_map_lock() callers, so I think same issue can happen on those, too? So for the long term, I'm wondering if we should update pte_offset_map_lock() or whole pte_offset_map_lock() error handling inside walk_page_range() callbacks to deal with this corner case. Or, other involved parts like CPU hotplugging handling? > since target page > is a migration entry. Kdamond can end the loop until user task is > scheduled on CPU 0. But CPU 0 is running migration/0. > 3. CONFIG_PREEMPT_NONE is enable. So all cpu are in a infinite loop. > > I found a similar softlockup issue which is also invoked by a memory > operation with cpu hotplug. To fix that issue, add a cond_resched() > to avoid block migration task. > https://lore.kernel.org/all/20250211081819.33307-1-chenridong@huaweicloud.com/#t > > May I ask if there is any solution we can fix this issue? Such as add a > cond_sched() in kdamond process. I think adding cond_resched() on the error handling part of damon_mkold_pmd_entry() is doable. Since DAMON is a best-effort approach, just returning without setting ACTION_AGAIN would also be ok for DAMON. It will simply make the behavior same to that before the commit 7780d04046a2 ("mm/pagewalkers: ACTION_AGAIN if pte_offset_map_lock() fails"). If this is a real issue making your products bleeding, please feel free to send such a fix for DAMON (I'd prefer the second approach - just return without setting ACTION_AGAIN) as a hotfix. Nonetheless, for long term, as I mentioned above, I'm wondering if we should update pte_offset_map_lock() internal, or add similar error handling modification on every similar pte_offset_map_lock() error handling. > Or is there any chance to do some yield > in stop machine process? Probably next time there is another different case > running with cpu hotplug can cause the same softlockup. I'm not familiar with stop machine process, so I have no good idea here but this might also be an option? > > Xinyu Zheng Thanks, SJ ^ permalink raw reply [flat|nested] 8+ messages in thread
* Re: [BUG REPORT] mm/damon: softlockup when kdamond walk page with cpu hotplug 2025-09-18 8:52 ` SeongJae Park @ 2025-09-18 8:57 ` SeongJae Park 2025-09-20 3:56 ` Hugh Dickins 1 sibling, 0 replies; 8+ messages in thread From: SeongJae Park @ 2025-09-18 8:57 UTC (permalink / raw) To: SeongJae Park Cc: Xinyu Zheng, Andrew Morton, Paul E . McKenney, Peter Zijlstra, damon, linux-mm, linux-kernel, zouyipeng, Hugh Dickins On Thu, 18 Sep 2025 01:52:49 -0700 SeongJae Park <sj@kernel.org> wrote: > Hello, > > On Thu, 18 Sep 2025 03:00:29 +0000 Xinyu Zheng <zhengxinyu6@huawei.com> wrote: > > > A softlockup issue was found with stress test: [...] > damon_mkold_pmd_entry() calls pte_offset_map_lock(). If the call returns an > error, damon_mkold_pmd_entry() sets walk->action as ACTION_AGAIN, to retry. If > the pte_offset_map_lock() continues fails, infinite loop can happen. I > understand the loop you mentioned above is this case. > > The error handling (retrying) was introduced by commit 7780d04046a2 > ("mm/pagewalkers: ACTION_AGAIN if pte_offset_map_lock() fails"). I added Hugh to this thread from the original mail of this reply, because he is the author of the commit. I should clarified this on the original mail but forgot saying that, sorry. Thanks, SJ [...] ^ permalink raw reply [flat|nested] 8+ messages in thread
* Re: [BUG REPORT] mm/damon: softlockup when kdamond walk page with cpu hotplug 2025-09-18 8:52 ` SeongJae Park 2025-09-18 8:57 ` SeongJae Park @ 2025-09-20 3:56 ` Hugh Dickins 2025-09-20 10:42 ` SeongJae Park 1 sibling, 1 reply; 8+ messages in thread From: Hugh Dickins @ 2025-09-20 3:56 UTC (permalink / raw) To: SeongJae Park Cc: Xinyu Zheng, Andrew Morton, Paul E . McKenney, Peter Zijlstra, damon, linux-mm, linux-kernel, zouyipeng, Hugh Dickins On Thu, 18 Sep 2025, SeongJae Park wrote: > Hello, > > On Thu, 18 Sep 2025 03:00:29 +0000 Xinyu Zheng <zhengxinyu6@huawei.com> wrote: > > > A softlockup issue was found with stress test: > > Thank you for sharing this great report! > > > > > watchdog: BUG: soft lockup - CPU#0 stuck for 26s! [migration/0:957] > > CPU: 0 PID: 957 Comm: migration/0 Kdump: loaded Tainted: > > Stopper: multi_cpu_stop+0x0/0x1e8 <- __stop_cpus.constprop.0+0x5c/0xb0 > > pstate: 61400009 (nZCv daif +PAN -UAO -TCO +DIT -SSBS BTYPE=--) > > pc : rcu_momentary_dyntick_idle+0x4c/0xa0 > > lr : multi_cpu_stop+0x10c/0x1e8 > > sp : ffff800086013d60 > > x29: ffff800086013d60 x28: 0000000000000001 x27: 0000000000000000 > > x26: 0000000000000000 x25: 00000000ffffffff x24: 0000000000000000 > > x23: 0000000000000001 x22: ffffab8f02977e00 x21: ffff8000b44ebb84 > > x20: ffff8000b44ebb60 x19: 0000000000000001 x18: 0000000000000000 > > x17: 000000040044ffff x16: 004000b5b5503510 x15: 0000000000000800 > > x14: ffff081003921440 x13: ffff5c907c75d000 x12: a34000013454d99d > > x11: 0000000000000000 x10: 0000000000000f90 x9 : ffffab8f01b657bc > > x8 : ffff081005e060f0 x7 : ffff081f7fd7b610 x6 : 0000009e0bb34c91 > > x5 : 00000000480fd060 x4 : ffff081f7fd7b508 x3 : ffff5c907c75d000 > > x2 : ffff800086013d60 x1 : 00000000b8ccb304 x0 : 00000000b8ccb30c > > Call trace: > > rcu_momentary_dyntick_idle+0x4c/0xa0 > > multi_cpu_stop+0x10c/0x1e8 > > cpu_stopper_thread+0xdc/0x1c0 > > smpboot_thread_fn+0x140/0x190 > > kthread+0xec/0x100 > > ret_from_fork+0x10/0x20 > > > > watchdog: BUG: soft lockup - CPU#18 stuck for 26s! [kdamond.0:408949] > > CPU: 18 PID: 408949 Comm: kdamond.0 Kdump: loaded Tainted: > > pstate: 61400009 (nZCv daif +PAN -UAO -TCO +DIT -SSBS BTYPE=--) > > pc : damon_mkold_pmd_entry+0x138/0x1d8 > > lr : damon_mkold_pmd_entry+0x68/0x1d8 > > sp : ffff8000c384bb00 > > x29: ffff8000c384bb10 x28: 0000ffff6e2a4a9b x27: 0000ffff6e2a4a9b > > x26: ffff080090fdeb88 x25: 0000ffff6e2a4a9b x24: ffffab8f029a9020 > > x23: ffff08013eb8dfe8 x22: 0000ffff6e2a4a9c x21: 0000ffff6e2a4a9b > > x20: ffff8000c384bd08 x19: 0000000000000000 x18: 0000000000000014 > > x17: 00000000f90a2272 x16: 0000000004c87773 x15: 000000004524349f > > x14: 00000000ee10aa21 x13: 0000000000000000 x12: ffffab8f02af4818 > > x11: 0000ffff7e7fffff x10: 0000ffff62700000 x9 : ffffab8f01d2f628 > > x8 : ffff0800879fbc0c x7 : ffff0800879fbc00 x6 : ffff0800c41c7d88 > > x5 : 0000000000000171 x4 : ffff08100aab0000 x3 : 00003081088800c0 > > x2 : 0000000000000000 x1 : 0000000000000000 x0 : 0000000000000000 > > Call trace: > > damon_mkold_pmd_entry+0x138/0x1d8 > > walk_pmd_range.isra.0+0x1ac/0x3a8 > > walk_pud_range+0x120/0x190 > > walk_pgd_range+0x170/0x1b8 > > __walk_page_range+0x184/0x198 > > walk_page_range+0x124/0x1f0 > > damon_va_prepare_access_checks+0xb4/0x1b8 > > kdamond_fn+0x11c/0x690 > > kthread+0xec/0x100 > > ret_from_fork+0x10/0x20 > > > > The stress test enable numa balance and kdamond, operation > > involves CPU hotplug and page fault with migration. > > > > CPU0 CPU18 events > > =============================== ====================== =============== > > page_fault(user task invoke) > > migrate_pages(pmd page migrate) > > __schedule > > kdamond_fn > > walk_pmd_range > > damon_mkold_pmd_entry <= cpu hotplug > > stop_machine_cpuslocked // infinite loop > > queue_stop_cpus_work // waiting CPU 0 user task > > multi_cpu_stop(migration/0) // to be scheduled > > // infinite loop waiting for > > // cpu 18 ACK > > > > Detail explanation: > > 1. When shutdown one cpu, a state machine in multi_cpu_stop() > > will wait for other cpu's migration thread reach to same state. > > In this case, all cpus are doing migration except cpu 18. > > 2. A user task which is bind on cpu 0 is allocating page and > > invoke page fault to migrate page. Kdamond is looping between > > damon_mkold_pmd_entry () and walk_pmd_range(), > > damon_mkold_pmd_entry() calls pte_offset_map_lock(). If the call returns an > error, damon_mkold_pmd_entry() sets walk->action as ACTION_AGAIN, to retry. If > the pte_offset_map_lock() continues fails, infinite loop can happen. I > understand the loop you mentioned above is this case. > > The error handling (retrying) was introduced by commit 7780d04046a2 > ("mm/pagewalkers: ACTION_AGAIN if pte_offset_map_lock() fails"). According to > the commit message, it is assumed to be safe to retry pte_offset_map_lock() > inside walk_page_range(), but it is not for this corner case. And the commit > introduced the error handling pattern in not only DAMON but also a few other > pte_offset_map_lock() callers, so I think same issue can happen on those, too? > > So for the long term, I'm wondering if we should update pte_offset_map_lock() > or whole pte_offset_map_lock() error handling inside walk_page_range() > callbacks to deal with this corner case. Or, other involved parts like CPU > hotplugging handling? > > > since target page > > is a migration entry. Kdamond can end the loop until user task is > > scheduled on CPU 0. But CPU 0 is running migration/0. > > 3. CONFIG_PREEMPT_NONE is enable. So all cpu are in a infinite loop. > > > > I found a similar softlockup issue which is also invoked by a memory > > operation with cpu hotplug. To fix that issue, add a cond_resched() > > to avoid block migration task. > > https://lore.kernel.org/all/20250211081819.33307-1-chenridong@huaweicloud.com/#t > > > > May I ask if there is any solution we can fix this issue? Such as add a > > cond_sched() in kdamond process. > > I think adding cond_resched() on the error handling part of > damon_mkold_pmd_entry() is doable. Since DAMON is a best-effort approach, just > returning without setting ACTION_AGAIN would also be ok for DAMON. It will > simply make the behavior same to that before the commit 7780d04046a2 > ("mm/pagewalkers: ACTION_AGAIN if pte_offset_map_lock() fails"). > > If this is a real issue making your products bleeding, please feel free to send > such a fix for DAMON (I'd prefer the second approach - just return without > setting ACTION_AGAIN) as a hotfix. > > Nonetheless, for long term, as I mentioned above, I'm wondering if we should > update pte_offset_map_lock() internal, or add similar error handling > modification on every similar pte_offset_map_lock() error handling. > > > Or is there any chance to do some yield > > in stop machine process? Probably next time there is another different case > > running with cpu hotplug can cause the same softlockup. > > I'm not familiar with stop machine process, so I have no good idea here but > this might also be an option? This had me worried for a while: thought we might be needing to change lots of other places, and scatter cond_resched()s here and there. But no: no need for cond_resched()'s, this is all just a confusion about where pmd migration entries are handled: a pmd migration entry is accepted by pmd_trans_huge_lock(), but is not accepted by pmd_trans_huge(). See fs/proc/task_mmu.c for mm_walk examples of trying pmd_trans_huge_lock(), then pte_offset_map_lock() if it failed, or ACTION_AGAIN if that failed too. When I ACTION_AGAINed damon_mkold_pmd_entry() and damon_young_pmd_entry() in 6.5, I didn't realize that the pmd migration entries were reaching the pte_offset_map_lock(), with corrupt results (or did pmd_bad() filter them out? I didn't think so, but it'll take me too long now to work out whether a pmd migration entry counts as pmd_bad or not); but knew that the new pte_offset_map_lock() filtered them out safely if there was a race. But they've been reaching it without any race, so yes the ACTION_AGAIN would send the mm_walk back again and again for as long as the pmd migration entry remained there: not good, and Xinyu finds a lockup when hotplugging CPU without preemption. My suggested patch below (please take it over SJ, and do with it what you will), converting damon_mkold_pmd_entry() and damon_young_pmd_entry() to use pmd_trans_huge_lock() as I'd been expecting, so handling the pmd migration entry up in that block. (Side note: patch against 6.17-rc, but I see mm.git adds also a damos_va_stat_pmd_entry(), which would better be converted to the same pmd_trans_huge_lock() pattern - though I notice you're not setting ACTION_AGAIN in that one.) But I have to admit, there's very little gained by using ACTION_AGAIN in these functions: it helps not to miss the range when racing against THP collapse or split, but you're already content to miss the extent if it has a pmd migration entry, and there can still be an instant when the range which used to have a page table does not yet show the THP. So if you prefer a smaller fix (but a larger source file!), just dropping the walk->action = ACTION_AGAIN lines should be good enough. Hugh p.s. I believe it would be possible to do the old/young business on migration entries, but I didn't have the patience to work out the conversions needed; and it shouldn't be part of a fix anyway. --- mm/damon/vaddr.c | 39 +++++++++------------------------------ 1 file changed, 9 insertions(+), 30 deletions(-) diff --git a/mm/damon/vaddr.c b/mm/damon/vaddr.c index 87e825349bdf..c6b51a52fca0 100644 --- a/mm/damon/vaddr.c +++ b/mm/damon/vaddr.c @@ -307,24 +307,14 @@ static int damon_mkold_pmd_entry(pmd_t *pmd, unsigned long addr, unsigned long next, struct mm_walk *walk) { pte_t *pte; - pmd_t pmde; spinlock_t *ptl; - if (pmd_trans_huge(pmdp_get(pmd))) { - ptl = pmd_lock(walk->mm, pmd); - pmde = pmdp_get(pmd); - - if (!pmd_present(pmde)) { - spin_unlock(ptl); - return 0; - } - - if (pmd_trans_huge(pmde)) { + ptl = pmd_trans_huge_lock(pmd, walk->vma); + if (ptl) { + if (pmd_present(pmdp_get(pmd))) damon_pmdp_mkold(pmd, walk->vma, addr); - spin_unlock(ptl); - return 0; - } spin_unlock(ptl); + return 0; } pte = pte_offset_map_lock(walk->mm, pmd, addr, &ptl); @@ -448,21 +438,12 @@ static int damon_young_pmd_entry(pmd_t *pmd, unsigned long addr, struct damon_young_walk_private *priv = walk->private; #ifdef CONFIG_TRANSPARENT_HUGEPAGE - if (pmd_trans_huge(pmdp_get(pmd))) { - pmd_t pmde; + ptl = pmd_trans_huge_lock(pmd, walk->vma); + if (ptl) { + pmd_t pmde = pmdp_get(pmd); - ptl = pmd_lock(walk->mm, pmd); - pmde = pmdp_get(pmd); - - if (!pmd_present(pmde)) { - spin_unlock(ptl); - return 0; - } - - if (!pmd_trans_huge(pmde)) { - spin_unlock(ptl); - goto regular_page; - } + if (!pmd_present(pmde)) + goto huge_out; folio = damon_get_folio(pmd_pfn(pmde)); if (!folio) goto huge_out; @@ -476,8 +457,6 @@ static int damon_young_pmd_entry(pmd_t *pmd, unsigned long addr, spin_unlock(ptl); return 0; } - -regular_page: #endif /* CONFIG_TRANSPARENT_HUGEPAGE */ pte = pte_offset_map_lock(walk->mm, pmd, addr, &ptl); -- 2.51.0 ^ permalink raw reply [flat|nested] 8+ messages in thread
* Re: [BUG REPORT] mm/damon: softlockup when kdamond walk page with cpu hotplug 2025-09-20 3:56 ` Hugh Dickins @ 2025-09-20 10:42 ` SeongJae Park 2025-09-22 3:48 ` Xinyu Zheng 2025-09-25 11:11 ` Xinyu Zheng 0 siblings, 2 replies; 8+ messages in thread From: SeongJae Park @ 2025-09-20 10:42 UTC (permalink / raw) To: Hugh Dickins Cc: SeongJae Park, Xinyu Zheng, Andrew Morton, Paul E . McKenney, Peter Zijlstra, damon, linux-mm, linux-kernel, zouyipeng On Fri, 19 Sep 2025 20:56:56 -0700 (PDT) Hugh Dickins <hughd@google.com> wrote: > On Thu, 18 Sep 2025, SeongJae Park wrote: > > > Hello, > > > > On Thu, 18 Sep 2025 03:00:29 +0000 Xinyu Zheng <zhengxinyu6@huawei.com> wrote: > > > > > A softlockup issue was found with stress test: [...] > This had me worried for a while: thought we might be needing to change > lots of other places, and scatter cond_resched()s here and there. > > But no: no need for cond_resched()'s, this is all just a confusion about > where pmd migration entries are handled: a pmd migration entry is accepted > by pmd_trans_huge_lock(), but is not accepted by pmd_trans_huge(). > > See fs/proc/task_mmu.c for mm_walk examples of trying pmd_trans_huge_lock(), > then pte_offset_map_lock() if it failed, or ACTION_AGAIN if that failed too. > > When I ACTION_AGAINed damon_mkold_pmd_entry() and damon_young_pmd_entry() > in 6.5, I didn't realize that the pmd migration entries were reaching the > pte_offset_map_lock(), with corrupt results (or did pmd_bad() filter them > out? I didn't think so, but it'll take me too long now to work out whether > a pmd migration entry counts as pmd_bad or not); but knew that the new > pte_offset_map_lock() filtered them out safely if there was a race. > > But they've been reaching it without any race, so yes the ACTION_AGAIN > would send the mm_walk back again and again for as long as the pmd > migration entry remained there: not good, and Xinyu finds a lockup > when hotplugging CPU without preemption. Thank you for your detailed and kind explanation, Hugh! > > My suggested patch below (please take it over SJ, and do with it what > you will), converting damon_mkold_pmd_entry() and damon_young_pmd_entry() > to use pmd_trans_huge_lock() as I'd been expecting, so handling the > pmd migration entry up in that block. (Side note: patch against 6.17-rc, > but I see mm.git adds also a damos_va_stat_pmd_entry(), which would > better be converted to the same pmd_trans_huge_lock() pattern - > though I notice you're not setting ACTION_AGAIN in that one.) > > But I have to admit, there's very little gained by using ACTION_AGAIN > in these functions: it helps not to miss the range when racing against > THP collapse or split, but you're already content to miss the extent > if it has a pmd migration entry, and there can still be an instant when > the range which used to have a page table does not yet show the THP. > > So if you prefer a smaller fix (but a larger source file!), just > dropping the walk->action = ACTION_AGAIN lines should be good enough. I agree all your points. I'd prefer the smaller source file following your suggested change below (using pmd_trans_huge_lock()) in long term. But, for a short term, I'd prefer the smaller fix (dropping walk->action = ACTION_AGAIN) since it should also be merged into stable@, up to 6.5.y. So, I'd like to suggest as following. Let's drop the 'walk->action = ACTION_AGAIN' like the below attached one, for now. After it is confirmed to fix the issue and merged into relevant trees including stable trees, let's revisit the code to cleanup following pmd_trans_huge_lock() pattern. Please let me know if I'm missing something, or you have other opinions. Xinyu, could you please test if the below attached patch fixes your issue and let us know the result? If Xinyu confirms the validity of the fix and no one objects to the above plan, I will post the fix as a formal one with a better commit message. Thanks, SJ [...] ==== >8 ==== From 743cafda8982624229541741dbfe5ff252328ac0 Mon Sep 17 00:00:00 2001 From: SeongJae Park <sj@kernel.org> Date: Sat, 20 Sep 2025 03:35:34 -0700 Subject: [PATCH] mm/damon/vaddr: do not try page table walk again For a quick fix of a softlockup issue: https://lore.kernel.org/20250918030029.2652607-1-zhengxinyu6@huawei.com Signed-off-by: SeongJae Park <sj@kernel.org> --- From 743cafda8982624229541741dbfe5ff252328ac0 Mon Sep 17 00:00:00 2001 From: SeongJae Park <sj@kernel.org> Date: Sat, 20 Sep 2025 03:35:34 -0700 Subject: [PATCH] mm/damon/vaddr: do not try page table walk again For a quick fix of a softlockup issue: https://lore.kernel.org/20250918030029.2652607-1-zhengxinyu6@huawei.com Signed-off-by: SeongJae Park <sj@kernel.org> --- mm/damon/vaddr.c | 8 ++------ 1 file changed, 2 insertions(+), 6 deletions(-) diff --git a/mm/damon/vaddr.c b/mm/damon/vaddr.c index 8c048f9b129e..7e834467b2d8 100644 --- a/mm/damon/vaddr.c +++ b/mm/damon/vaddr.c @@ -328,10 +328,8 @@ static int damon_mkold_pmd_entry(pmd_t *pmd, unsigned long addr, } pte = pte_offset_map_lock(walk->mm, pmd, addr, &ptl); - if (!pte) { - walk->action = ACTION_AGAIN; + if (!pte) return 0; - } if (!pte_present(ptep_get(pte))) goto out; damon_ptep_mkold(pte, walk->vma, addr); @@ -481,10 +479,8 @@ static int damon_young_pmd_entry(pmd_t *pmd, unsigned long addr, #endif /* CONFIG_TRANSPARENT_HUGEPAGE */ pte = pte_offset_map_lock(walk->mm, pmd, addr, &ptl); - if (!pte) { - walk->action = ACTION_AGAIN; + if (!pte) return 0; - } ptent = ptep_get(pte); if (!pte_present(ptent)) goto out; -- 2.39.5 ^ permalink raw reply [flat|nested] 8+ messages in thread
* Re: [BUG REPORT] mm/damon: softlockup when kdamond walk page with cpu hotplug 2025-09-20 10:42 ` SeongJae Park @ 2025-09-22 3:48 ` Xinyu Zheng 2025-09-25 11:11 ` Xinyu Zheng 1 sibling, 0 replies; 8+ messages in thread From: Xinyu Zheng @ 2025-09-22 3:48 UTC (permalink / raw) To: SeongJae Park, Hugh Dickins Cc: Andrew Morton, Paul E . McKenney, Peter Zijlstra, damon, linux-mm, linux-kernel, zouyipeng On 9/20/2025 6:42 PM, SeongJae Park wrote: > On Fri, 19 Sep 2025 20:56:56 -0700 (PDT) Hugh Dickins <hughd@google.com> wrote: > >> On Thu, 18 Sep 2025, SeongJae Park wrote: >> >>> Hello, >>> >>> On Thu, 18 Sep 2025 03:00:29 +0000 Xinyu Zheng <zhengxinyu6@huawei.com> wrote: >>> >>>> A softlockup issue was found with stress test: > [...] >> This had me worried for a while: thought we might be needing to change >> lots of other places, and scatter cond_resched()s here and there. >> >> But no: no need for cond_resched()'s, this is all just a confusion about >> where pmd migration entries are handled: a pmd migration entry is accepted >> by pmd_trans_huge_lock(), but is not accepted by pmd_trans_huge(). >> >> See fs/proc/task_mmu.c for mm_walk examples of trying pmd_trans_huge_lock(), >> then pte_offset_map_lock() if it failed, or ACTION_AGAIN if that failed too. >> >> When I ACTION_AGAINed damon_mkold_pmd_entry() and damon_young_pmd_entry() >> in 6.5, I didn't realize that the pmd migration entries were reaching the >> pte_offset_map_lock(), with corrupt results (or did pmd_bad() filter them >> out? I didn't think so, but it'll take me too long now to work out whether >> a pmd migration entry counts as pmd_bad or not); but knew that the new >> pte_offset_map_lock() filtered them out safely if there was a race. >> >> But they've been reaching it without any race, so yes the ACTION_AGAIN >> would send the mm_walk back again and again for as long as the pmd >> migration entry remained there: not good, and Xinyu finds a lockup >> when hotplugging CPU without preemption. > > Thank you for your detailed and kind explanation, Hugh! > >> >> My suggested patch below (please take it over SJ, and do with it what >> you will), converting damon_mkold_pmd_entry() and damon_young_pmd_entry() >> to use pmd_trans_huge_lock() as I'd been expecting, so handling the >> pmd migration entry up in that block. (Side note: patch against 6.17-rc, >> but I see mm.git adds also a damos_va_stat_pmd_entry(), which would >> better be converted to the same pmd_trans_huge_lock() pattern - >> though I notice you're not setting ACTION_AGAIN in that one.) >> >> But I have to admit, there's very little gained by using ACTION_AGAIN >> in these functions: it helps not to miss the range when racing against >> THP collapse or split, but you're already content to miss the extent >> if it has a pmd migration entry, and there can still be an instant when >> the range which used to have a page table does not yet show the THP. >> >> So if you prefer a smaller fix (but a larger source file!), just >> dropping the walk->action = ACTION_AGAIN lines should be good enough. Hi, Hugh. Thank you for your patient explanation and solution! > > I agree all your points. > > I'd prefer the smaller source file following your suggested change below (using > pmd_trans_huge_lock()) in long term. But, for a short term, I'd prefer the > smaller fix (dropping walk->action = ACTION_AGAIN) since it should also be > merged into stable@, up to 6.5.y. > > So, I'd like to suggest as following. Let's drop the > 'walk->action = ACTION_AGAIN' like the below attached one, for now. After it > is confirmed to fix the issue and merged into relevant trees including stable > trees, let's revisit the code to cleanup following pmd_trans_huge_lock() > pattern. > > Please let me know if I'm missing something, or you have other opinions. > > Xinyu, could you please test if the below attached patch fixes your issue and > let us know the result? Thanks for all the reply and suggestion, SJ! I am trying to reproduce the scenario now and will send the result back very soon! > > If Xinyu confirms the validity of the fix and no one objects to the above plan, > I will post the fix as a formal one with a better commit message. > > > Thanks, > SJ > > [...] > > ==== >8 ==== > From 743cafda8982624229541741dbfe5ff252328ac0 Mon Sep 17 00:00:00 2001 > From: SeongJae Park <sj@kernel.org> > Date: Sat, 20 Sep 2025 03:35:34 -0700 > Subject: [PATCH] mm/damon/vaddr: do not try page table walk again > > For a quick fix of a softlockup issue: > https://lore.kernel.org/20250918030029.2652607-1-zhengxinyu6@huawei.com > > Signed-off-by: SeongJae Park <sj@kernel.org> > --- > From 743cafda8982624229541741dbfe5ff252328ac0 Mon Sep 17 00:00:00 2001 > From: SeongJae Park <sj@kernel.org> > Date: Sat, 20 Sep 2025 03:35:34 -0700 > Subject: [PATCH] mm/damon/vaddr: do not try page table walk again > > For a quick fix of a softlockup issue: > https://lore.kernel.org/20250918030029.2652607-1-zhengxinyu6@huawei.com > > Signed-off-by: SeongJae Park <sj@kernel.org> > --- > mm/damon/vaddr.c | 8 ++------ > 1 file changed, 2 insertions(+), 6 deletions(-) > > diff --git a/mm/damon/vaddr.c b/mm/damon/vaddr.c > index 8c048f9b129e..7e834467b2d8 100644 > --- a/mm/damon/vaddr.c > +++ b/mm/damon/vaddr.c > @@ -328,10 +328,8 @@ static int damon_mkold_pmd_entry(pmd_t *pmd, unsigned long addr, > } > > pte = pte_offset_map_lock(walk->mm, pmd, addr, &ptl); > - if (!pte) { > - walk->action = ACTION_AGAIN; > + if (!pte) > return 0; > - } > if (!pte_present(ptep_get(pte))) > goto out; > damon_ptep_mkold(pte, walk->vma, addr); > @@ -481,10 +479,8 @@ static int damon_young_pmd_entry(pmd_t *pmd, unsigned long addr, > #endif /* CONFIG_TRANSPARENT_HUGEPAGE */ > > pte = pte_offset_map_lock(walk->mm, pmd, addr, &ptl); > - if (!pte) { > - walk->action = ACTION_AGAIN; > + if (!pte) > return 0; > - } > ptent = ptep_get(pte); > if (!pte_present(ptent)) > goto out; -- Xinyu Zheng ^ permalink raw reply [flat|nested] 8+ messages in thread
* Re: [BUG REPORT] mm/damon: softlockup when kdamond walk page with cpu hotplug 2025-09-20 10:42 ` SeongJae Park 2025-09-22 3:48 ` Xinyu Zheng @ 2025-09-25 11:11 ` Xinyu Zheng 2025-09-30 0:49 ` SeongJae Park 1 sibling, 1 reply; 8+ messages in thread From: Xinyu Zheng @ 2025-09-25 11:11 UTC (permalink / raw) To: SeongJae Park Cc: Andrew Morton, Paul E . McKenney, Peter Zijlstra, damon, linux-mm, linux-kernel, zouyipeng, Hugh Dickins On 9/20/2025 6:42 PM, SeongJae Park wrote: > On Fri, 19 Sep 2025 20:56:56 -0700 (PDT) Hugh Dickins <hughd@google.com> wrote: > >> On Thu, 18 Sep 2025, SeongJae Park wrote: >> >>> Hello, >>> >>> On Thu, 18 Sep 2025 03:00:29 +0000 Xinyu Zheng <zhengxinyu6@huawei.com> wrote: >>> >>>> A softlockup issue was found with stress test: > [...] >> This had me worried for a while: thought we might be needing to change >> lots of other places, and scatter cond_resched()s here and there. >> >> But no: no need for cond_resched()'s, this is all just a confusion about >> where pmd migration entries are handled: a pmd migration entry is accepted >> by pmd_trans_huge_lock(), but is not accepted by pmd_trans_huge(). >> >> See fs/proc/task_mmu.c for mm_walk examples of trying pmd_trans_huge_lock(), >> then pte_offset_map_lock() if it failed, or ACTION_AGAIN if that failed too. >> >> When I ACTION_AGAINed damon_mkold_pmd_entry() and damon_young_pmd_entry() >> in 6.5, I didn't realize that the pmd migration entries were reaching the >> pte_offset_map_lock(), with corrupt results (or did pmd_bad() filter them >> out? I didn't think so, but it'll take me too long now to work out whether >> a pmd migration entry counts as pmd_bad or not); but knew that the new >> pte_offset_map_lock() filtered them out safely if there was a race. >> >> But they've been reaching it without any race, so yes the ACTION_AGAIN >> would send the mm_walk back again and again for as long as the pmd >> migration entry remained there: not good, and Xinyu finds a lockup >> when hotplugging CPU without preemption. > > Thank you for your detailed and kind explanation, Hugh! > >> >> My suggested patch below (please take it over SJ, and do with it what >> you will), converting damon_mkold_pmd_entry() and damon_young_pmd_entry() >> to use pmd_trans_huge_lock() as I'd been expecting, so handling the >> pmd migration entry up in that block. (Side note: patch against 6.17-rc, >> but I see mm.git adds also a damos_va_stat_pmd_entry(), which would >> better be converted to the same pmd_trans_huge_lock() pattern - >> though I notice you're not setting ACTION_AGAIN in that one.) >> >> But I have to admit, there's very little gained by using ACTION_AGAIN >> in these functions: it helps not to miss the range when racing against >> THP collapse or split, but you're already content to miss the extent >> if it has a pmd migration entry, and there can still be an instant when >> the range which used to have a page table does not yet show the THP. >> >> So if you prefer a smaller fix (but a larger source file!), just >> dropping the walk->action = ACTION_AGAIN lines should be good enough. > > I agree all your points. > > I'd prefer the smaller source file following your suggested change below (using > pmd_trans_huge_lock()) in long term. But, for a short term, I'd prefer the > smaller fix (dropping walk->action = ACTION_AGAIN) since it should also be > merged into stable@, up to 6.5.y. > > So, I'd like to suggest as following. Let's drop the > 'walk->action = ACTION_AGAIN' like the below attached one, for now. After it > is confirmed to fix the issue and merged into relevant trees including stable > trees, let's revisit the code to cleanup following pmd_trans_huge_lock() > pattern. > > Please let me know if I'm missing something, or you have other opinions. > > Xinyu, could you please test if the below attached patch fixes your issue and > let us know the result? > > If Xinyu confirms the validity of the fix and no one objects to the above plan, > I will post the fix as a formal one with a better commit message. > Hi, SJ. The below attached patch is works! These days, I reproduced the same softlockup scenario for three times. It can be reproduced around an hour when I increase the memory pressure. Then I patched the below fix, and ran the same testcase whole day, softlock didn't trigger again. Sorry for the long wait. Thanks again for your solution and explanation! > > Thanks, > SJ > > [...] > > ==== >8 ==== > From 743cafda8982624229541741dbfe5ff252328ac0 Mon Sep 17 00:00:00 2001 > From: SeongJae Park <sj@kernel.org> > Date: Sat, 20 Sep 2025 03:35:34 -0700 > Subject: [PATCH] mm/damon/vaddr: do not try page table walk again > > For a quick fix of a softlockup issue: > https://lore.kernel.org/20250918030029.2652607-1-zhengxinyu6@huawei.com > > Signed-off-by: SeongJae Park <sj@kernel.org> > --- > From 743cafda8982624229541741dbfe5ff252328ac0 Mon Sep 17 00:00:00 2001 > From: SeongJae Park <sj@kernel.org> > Date: Sat, 20 Sep 2025 03:35:34 -0700 > Subject: [PATCH] mm/damon/vaddr: do not try page table walk again > > For a quick fix of a softlockup issue: > https://lore.kernel.org/20250918030029.2652607-1-zhengxinyu6@huawei.com > > Signed-off-by: SeongJae Park <sj@kernel.org> > --- > mm/damon/vaddr.c | 8 ++------ > 1 file changed, 2 insertions(+), 6 deletions(-) > > diff --git a/mm/damon/vaddr.c b/mm/damon/vaddr.c > index 8c048f9b129e..7e834467b2d8 100644 > --- a/mm/damon/vaddr.c > +++ b/mm/damon/vaddr.c > @@ -328,10 +328,8 @@ static int damon_mkold_pmd_entry(pmd_t *pmd, unsigned long addr, > } > > pte = pte_offset_map_lock(walk->mm, pmd, addr, &ptl); > - if (!pte) { > - walk->action = ACTION_AGAIN; > + if (!pte) > return 0; > - } > if (!pte_present(ptep_get(pte))) > goto out; > damon_ptep_mkold(pte, walk->vma, addr); > @@ -481,10 +479,8 @@ static int damon_young_pmd_entry(pmd_t *pmd, unsigned long addr, > #endif /* CONFIG_TRANSPARENT_HUGEPAGE */ > > pte = pte_offset_map_lock(walk->mm, pmd, addr, &ptl); > - if (!pte) { > - walk->action = ACTION_AGAIN; > + if (!pte) > return 0; > - } > ptent = ptep_get(pte); > if (!pte_present(ptent)) > goto out; Xinyu Zheng ^ permalink raw reply [flat|nested] 8+ messages in thread
* Re: [BUG REPORT] mm/damon: softlockup when kdamond walk page with cpu hotplug 2025-09-25 11:11 ` Xinyu Zheng @ 2025-09-30 0:49 ` SeongJae Park 0 siblings, 0 replies; 8+ messages in thread From: SeongJae Park @ 2025-09-30 0:49 UTC (permalink / raw) To: Xinyu Zheng Cc: SeongJae Park, Andrew Morton, Paul E . McKenney, Peter Zijlstra, damon, linux-mm, linux-kernel, zouyipeng, Hugh Dickins On Thu, 25 Sep 2025 19:11:13 +0800 Xinyu Zheng <zhengxinyu6@huawei.com> wrote: > > > On 9/20/2025 6:42 PM, SeongJae Park wrote: > > On Fri, 19 Sep 2025 20:56:56 -0700 (PDT) Hugh Dickins <hughd@google.com> wrote: > > > >> On Thu, 18 Sep 2025, SeongJae Park wrote: > >> > >>> Hello, > >>> > >>> On Thu, 18 Sep 2025 03:00:29 +0000 Xinyu Zheng <zhengxinyu6@huawei.com> wrote: > >>> > >>>> A softlockup issue was found with stress test: [...] > > Xinyu, could you please test if the below attached patch fixes your issue and > > let us know the result? > > > > If Xinyu confirms the validity of the fix and no one objects to the above plan, > > I will post the fix as a formal one with a better commit message. > > > > Hi, SJ. The below attached patch is works! > > These days, I reproduced the same softlockup scenario for three times. > It can be reproduced around an hour when I increase the memory pressure. > Then I patched the below fix, and ran the same testcase whole day, > softlock didn't trigger again. > > Sorry for the long wait. Thanks again for your solution and explanation! Sorry for late response, and thank you for confirming that the fix works. I just posted it as a formal patch: https://lore.kernel.org/20250930004410.55228-1-sj@kernel.org Thanks, SJ [...] ^ permalink raw reply [flat|nested] 8+ messages in thread
end of thread, other threads:[~2025-09-30 0:49 UTC | newest] Thread overview: 8+ messages (download: mbox.gz / follow: Atom feed) -- links below jump to the message on this page -- 2025-09-18 3:00 [BUG REPORT] mm/damon: softlockup when kdamond walk page with cpu hotplug Xinyu Zheng 2025-09-18 8:52 ` SeongJae Park 2025-09-18 8:57 ` SeongJae Park 2025-09-20 3:56 ` Hugh Dickins 2025-09-20 10:42 ` SeongJae Park 2025-09-22 3:48 ` Xinyu Zheng 2025-09-25 11:11 ` Xinyu Zheng 2025-09-30 0:49 ` SeongJae Park
This is a public inbox, see mirroring instructions for how to clone and mirror all data and code used for this inbox