linux-mm.kvack.org archive mirror
 help / color / mirror / Atom feed
* [Bug report][xfstests generic/751] hang on nfs writeback
@ 2025-10-21  5:14 Zorro Lang
  2025-10-22 16:24 ` Mike Snitzer
  0 siblings, 1 reply; 3+ messages in thread
From: Zorro Lang @ 2025-10-21  5:14 UTC (permalink / raw)
  To: linux-nfs; +Cc: linux-mm

Hi,

When I did xfstests regression test on nfs, it hang on generic/751 many times,
refer to [1](on x86_64) and [2](on aarch64). I've hit it 8 times until now.

I tested on latest linux v6.18-rc2. My underlying test device isn't loopback device,
it's general disk, and make xfs on it:
  meta-data=/dev/sda5              isize=512    agcount=4, agsize=983040 blks
           =                       sectsz=4096  attr=2, projid32bit=1
           =                       crc=1        finobt=1, sparse=1, rmapbt=1
           =                       reflink=1    bigtime=1 inobtcount=1 nrext64=1
           =                       exchange=0   metadir=0
  data     =                       bsize=4096   blocks=3932160, imaxpct=25
           =                       sunit=0      swidth=0 blks
  naming   =version 2              bsize=4096   ascii-ci=0, ftype=1, parent=0
  log      =internal log           bsize=4096   blocks=54324, version=2
           =                       sectsz=4096  sunit=1 blks, lazy-count=1
  realtime =none                   extsz=4096   blocks=0, rtextents=0
           =                       rgcount=0    rgsize=0 extents
           =                       zoned=0      start=0 reserved=0

Two xfs are mounted on /mnt/xfstests/test and /mnt/xfstests/scratch seperately,
then export as:
  # cat /etc/exports
  /mnt/xfstests/test/nfs-server *(rw,insecure,no_root_squash)
  /mnt/xfstests/scratch/nfs-server *(rw,insecure,no_root_squash)

The nfs mount option is only "-o vers=4.2". BTW, nfs server and client are
in same machine/system.

# cat local.config
export FSTYP=nfs
export TEST_DEV=xxxx-xxx-xxxx.xxxx.xxx.xxx:/mnt/xfstests/test/nfs-server
export TEST_DIR=/mnt/xfstests/test/nfs-client
export SCRATCH_DEV=xxxx-xxx-xxxx.xxxx.xxx.xxx:/mnt/xfstests/scratch/nfs-server
export SCRATCH_MNT=/mnt/xfstests/scratch/nfs-client
export MOUNT_OPTIONS="-o vers=4.2"
export TEST_FS_MOUNT_OPTS="-o vers=4.2"


[1]
[23369.572660] run fstests generic/751 at 2025-10-20 23:15:24 
[-- MARK -- Tue Oct 21 03:20:00 2025] 
[-- MARK -- Tue Oct 21 03:25:00 2025] 
[-- MARK -- Tue Oct 21 03:30:00 2025] 
[-- MARK -- Tue Oct 21 03:35:00 2025] 
[-- MARK -- Tue Oct 21 03:40:00 2025] 
[25069.500900] INFO: task kworker/u9:73:825900 blocked for more than 122 seconds. 
[25069.501484]       Not tainted 6.18.0-rc2 #1 
[25069.501822] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. 
[25069.502385] task:kworker/u9:73   state:D stack:24712 pid:825900 tgid:825900 ppid:2      task_flags:0x4208060 flags:0x00080000 
[25069.503258] Workqueue: writeback wb_workfn (flush-0:48) 
[25069.782725] Call Trace: 
[25069.787191]  <TASK> 
[25069.787368]  __schedule+0x838/0x1890 
[25069.799533]  ? __pfx___schedule+0x10/0x10 
[25069.799852]  ? __blk_flush_plug+0x27b/0x4d0 
[25069.800617]  ? find_held_lock+0x32/0x90 
[25069.801292]  ? __lock_release.isra.0+0x1a4/0x2c0 
[25069.801639]  schedule+0xd4/0x260 
[25069.801902]  io_schedule+0x8f/0xf0 
[25069.802162]  folio_wait_bit_common+0x2d9/0x780 
[25069.802919]  ? folio_wait_bit_common+0x1dd/0x780 
[25069.803324]  ? __pfx_folio_wait_bit_common+0x10/0x10 
[25069.803684]  ? nfs_page_clear_headlock+0x31/0x80 [nfs] 
[25069.804254]  ? __pfx_wake_page_function+0x10/0x10 
[25069.804602]  ? __pfx___might_resched+0x10/0x10 
[25069.805160]  writeback_get_folio+0x3f9/0x500 
[25069.806221]  writeback_iter+0x136/0x720 
[25069.806628]  nfs_writepages+0x4f8/0x9b0 [nfs] 
[25069.807100]  ? mark_held_locks+0x40/0x70 
[25069.807403]  ? __pfx_nfs_writepages+0x10/0x10 [nfs] 
[25069.807846]  ? virtqueue_notify+0x68/0xc0 
[25069.808346]  ? virtio_queue_rq+0x2b1/0x650 [virtio_blk] 
[25069.808925]  ? __lock_acquire+0x57c/0xbd0 
[25069.809295]  do_writepages+0x21f/0x560 
[25069.809576]  ? __pfx_do_writepages+0x10/0x10 
[25069.809953]  ? rcu_is_watching+0x15/0xb0 
[25069.810587]  __writeback_single_inode+0xe2/0x5f0 
[25069.810982]  ? __lock_release.isra.0+0x1a4/0x2c0 
[25069.811373]  ? __pfx___writeback_single_inode+0x10/0x10 
[25069.811815]  ? writeback_sb_inodes+0x416/0xd00 
[25069.812194]  writeback_sb_inodes+0x535/0xd00 
[25069.812533]  ? __pfx_stack_trace_save+0x10/0x10 
[25069.814193]  ? local_clock_noinstr+0xd/0xe0 
[25069.814506]  ? __pfx_writeback_sb_inodes+0x10/0x10 
[25069.814910]  ? __lock_release.isra.0+0x1a4/0x2c0 
[25069.815407]  ? lock_acquire+0x10b/0x150 
[25069.815703]  ? down_read_trylock+0x4b/0x60 
[25069.816077]  __writeback_inodes_wb+0xf4/0x270 
[25069.816441]  ? __pfx___writeback_inodes_wb+0x10/0x10 
[25069.816864]  ? queue_io+0x329/0x510 
[25069.817148]  wb_writeback+0x70a/0x9c0 
[25069.817448]  ? __pfx_wb_writeback+0x10/0x10 
[25069.817831]  ? get_nr_dirty_inodes+0xcb/0x180 
[25069.818264]  wb_do_writeback+0x5d4/0x8e0 
[25069.818575]  ? __pfx_wb_do_writeback+0x10/0x10 
[25069.818975]  ? set_worker_desc+0x16e/0x190 
[25069.819313]  ? __pfx_set_worker_desc+0x10/0x10 
[25069.819669]  wb_workfn+0x7c/0x200 
[25069.819997]  process_one_work+0xd8b/0x1320 
[25069.820342]  ? __pfx_process_one_work+0x10/0x10 
[25069.820694]  ? assign_work+0x16c/0x240 
[25069.821053]  worker_thread+0x5f3/0xfe0 
[25069.821364]  ? __pfx_worker_thread+0x10/0x10 
[25069.821697]  kthread+0x3b4/0x770 
[25069.822122]  ? kvm_sched_clock_read+0x11/0x20 
[25069.822457]  ? local_clock_noinstr+0xd/0xe0 
[25069.822790]  ? __pfx_kthread+0x10/0x10 
[25069.823079]  ? __lock_release.isra.0+0x1a4/0x2c0 
[25069.823430]  ? rcu_is_watching+0x15/0xb0 
[25069.823729]  ? __pfx_kthread+0x10/0x10 
[25069.824035]  ret_from_fork+0x393/0x480 
[25069.864794]  ? __pfx_kthread+0x10/0x10 
[25069.865076]  ? __pfx_kthread+0x10/0x10 
[25069.865371]  ret_from_fork_asm+0x1a/0x30 
[25069.879905]  </TASK> 
[25069.880145]  
[25069.880145] Showing all locks held in the system: 
[25069.889550] 1 lock held by khungtaskd/37: 
[25069.889886]  #0: ffffffffa29309e0 (rcu_read_lock){....}-{1:3}, at: rcu_lock_acquire.constprop.0+0x7/0x30 
[25069.890603] 2 locks held by 751/825733: 
[25069.890942]  #0: ffff888101e36440 (sb_writers#16){.+.+}-{0:0}, at: ksys_write+0xf9/0x1d0 
[25069.891589]  #1: ffffffffa2b97fd0 (split_debug_mutex){+.+.}-{4:4}, at: split_huge_pages_write+0x124/0x430 
[25069.892449] 3 locks held by kworker/u9:73/825900: 
[25069.892818]  #0: ffff8881029df958 ((wq_completion)writeback){+.+.}-{0:0}, at: process_one_work+0x7f5/0x1320 
[25069.893530]  #1: ffffc9000ad9fd10 ((work_completion)(&(&wb->dwork)->work)){+.+.}-{0:0}, at: process_one_work+0xd3f/0x1320 
[25069.894333]  #2: ffff88810b1480e8 (&type->s_umount_key#68){++++}-{4:4}, at: super_trylock_shared+0x1c/0xa0 
[25069.895056] 2 locks held by kworker/u10:2/826054: 
[25069.895418]  #0: ffff888103ec2158 ((wq_completion)nfsiod){+.+.}-{0:0}, at: process_one_work+0x7f5/0x1320 
[25069.896123]  #1: ffffc90001e1fd10 ((work_completion)(&ctx->work)){+.+.}-{0:0}, at: process_one_work+0xd3f/0x1320 
[25069.896952]  
[25069.897085] ============================================= 
[25069.897085]  
[-- MARK -- Tue Oct 21 03:45:00 2025] 
[25192.380157] INFO: task kworker/u9:73:825900 blocked for more than 245 seconds. 
[25192.380707]       Not tainted 6.18.0-rc2 #1 
[25192.381042] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. 
[25192.381605] task:kworker/u9:73   state:D stack:24712 pid:825900 tgid:825900 ppid:2      task_flags:0x4208060 flags:0x00080000 
[25192.382516] Workqueue: writeback wb_workfn (flush-0:48) 
[25192.382914] Call Trace: 
[25192.383126]  <TASK> 
[25192.383304]  __schedule+0x838/0x1890 
[25192.383586]  ? __pfx___schedule+0x10/0x10 
[25192.383888]  ? __blk_flush_plug+0x27b/0x4d0 
[25192.384228]  ? find_held_lock+0x32/0x90 
[25192.384527]  ? __lock_release.isra.0+0x1a4/0x2c0 
[25192.384880]  schedule+0xd4/0x260 
[25192.385150]  io_schedule+0x8f/0xf0 
[25192.385416]  folio_wait_bit_common+0x2d9/0x780 
[25192.385816]  ? folio_wait_bit_common+0x1dd/0x780 
[25192.386196]  ? __pfx_folio_wait_bit_common+0x10/0x10 
[25192.386572]  ? nfs_page_clear_headlock+0x31/0x80 [nfs] 
[25192.387045]  ? __pfx_wake_page_function+0x10/0x10 
[25192.387400]  ? __pfx___might_resched+0x10/0x10 
[25192.387742]  writeback_get_folio+0x3f9/0x500 
[25192.388092]  writeback_iter+0x136/0x720 
[25192.388389]  nfs_writepages+0x4f8/0x9b0 [nfs] 
[25192.388788]  ? mark_held_locks+0x40/0x70 
[25192.389102]  ? __pfx_nfs_writepages+0x10/0x10 [nfs] 
[25192.389530]  ? virtqueue_notify+0x68/0xc0 
[25192.389831]  ? virtio_queue_rq+0x2b1/0x650 [virtio_blk] 
[25192.390255]  ? __lock_acquire+0x57c/0xbd0 
[25192.390574]  do_writepages+0x21f/0x560 
[25192.390861]  ? __pfx_do_writepages+0x10/0x10 
[25192.391200]  ? rcu_is_watching+0x15/0xb0 
[25192.391502]  __writeback_single_inode+0xe2/0x5f0 
[25192.391848]  ? __lock_release.isra.0+0x1a4/0x2c0 
[25192.392210]  ? __pfx___writeback_single_inode+0x10/0x10 
[25192.392596]  ? writeback_sb_inodes+0x416/0xd00 
[25192.392935]  writeback_sb_inodes+0x535/0xd00 
[25192.393286]  ? __pfx_stack_trace_save+0x10/0x10 
[25192.393628]  ? local_clock_noinstr+0xd/0xe0 
[25192.393940]  ? __pfx_writeback_sb_inodes+0x10/0x10 
[25192.394310]  ? __lock_release.isra.0+0x1a4/0x2c0 
[25192.394731]  ? lock_acquire+0x10b/0x150 
[25192.395048]  ? down_read_trylock+0x4b/0x60 
[25192.395360]  __writeback_inodes_wb+0xf4/0x270 
[25192.395698]  ? __pfx___writeback_inodes_wb+0x10/0x10 
[25192.396081]  ? queue_io+0x329/0x510 
[25192.396355]  wb_writeback+0x70a/0x9c0 
[25192.396645]  ? __pfx_wb_writeback+0x10/0x10 
[25192.396963]  ? get_nr_dirty_inodes+0xcb/0x180 
[25192.397314]  wb_do_writeback+0x5d4/0x8e0 
[25192.397622]  ? __pfx_wb_do_writeback+0x10/0x10 
[25192.397950]  ? set_worker_desc+0x16e/0x190 
[25192.398272]  ? __pfx_set_worker_desc+0x10/0x10 
[25192.398628]  wb_workfn+0x7c/0x200 
[25192.398886]  process_one_work+0xd8b/0x1320 
[25192.399222]  ? __pfx_process_one_work+0x10/0x10 
[25192.399574]  ? assign_work+0x16c/0x240 
[25192.399861]  worker_thread+0x5f3/0xfe0 
[25192.400176]  ? __pfx_worker_thread+0x10/0x10 
[25192.400497]  kthread+0x3b4/0x770 
[25192.400747]  ? kvm_sched_clock_read+0x11/0x20 
[25192.401089]  ? local_clock_noinstr+0xd/0xe0 
[25192.401401]  ? __pfx_kthread+0x10/0x10 
[25192.401684]  ? __lock_release.isra.0+0x1a4/0x2c0 
[25192.402046]  ? rcu_is_watching+0x15/0xb0 
[25192.402343]  ? __pfx_kthread+0x10/0x10 
[25192.402629]  ret_from_fork+0x393/0x480 
[25192.402910]  ? __pfx_kthread+0x10/0x10 
[25192.403208]  ? __pfx_kthread+0x10/0x10 
[25192.403493]  ret_from_fork_asm+0x1a/0x30 
[25192.403808]  </TASK> 
[25192.403994]  
[25192.403994] Showing all locks held in the system: 
[25192.404474] 1 lock held by khungtaskd/37: 
[25192.404774]  #0: ffffffffa29309e0 (rcu_read_lock){....}-{1:3}, at: rcu_lock_acquire.constprop.0+0x7/0x30 
[25192.405487] 2 locks held by 751/825733: 
[25192.405779]  #0: ffff888101e36440 (sb_writers#16){.+.+}-{0:0}, at: ksys_write+0xf9/0x1d0 
[25192.406385]  #1: ffffffffa2b97fd0 (split_debug_mutex){+.+.}-{4:4}, at: split_huge_pages_write+0x124/0x430 
[25192.407092] 4 locks held by kworker/u10:40/825850: 
[25192.407443]  #0: ffff888103ec5158 ((wq_completion)nfslocaliod){+.+.}-{0:0}, at: process_one_work+0x7f5/0x1320 
[25192.408175]  #1: ffffc9000b0cfd10 ((work_completion)(&iocb->work)#2){+.+.}-{0:0}, at: process_one_work+0xd3f/0x1320 
[25192.408924]  #2: ffff88815c2c2440 (sb_writers#13){++++}-{0:0}, at: process_one_work+0xd8b/0x1320 
[25192.409579]  #3: ffff8881a1324b58 (&sb->s_type->i_mutex_key#13){++++}-{4:4}, at: xfs_ilock+0x360/0x460 [xfs] 
[25192.410660] 3 locks held by kworker/u9:73/825900: 
[25192.411026]  #0: ffff8881029df958 ((wq_completion)writeback){+.+.}-{0:0}, at: process_one_work+0x7f5/0x1320 
[25192.411724]  #1: ffffc9000ad9fd10 ((work_completion)(&(&wb->dwork)->work)){+.+.}-{0:0}, at: process_one_work+0xd3f/0x1320 
[25192.412518]  #2: ffff88810b1480e8 (&type->s_umount_key#68){++++}-{4:4}, at: super_trylock_shared+0x1c/0xa0 
[25192.413237] 4 locks held by kworker/1:1/826103: 
[25192.413576]  
[25192.413706] ============================================= 
[25192.413706]  
[25315.259312] INFO: task kworker/u9:73:825900 blocked for more than 368 seconds. 
[25315.259973]       Not tainted 6.18.0-rc2 #1 
[25315.260324] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. 
[25315.260894] task:kworker/u9:73   state:D stack:24712 pid:825900 tgid:825900 ppid:2      task_flags:0x4208060 flags:0x00080000 
[25315.261752] Workqueue: writeback wb_workfn (flush-0:48) 
[25315.262153] Call Trace: 
[25315.262369]  <TASK> 
[25315.262551]  __schedule+0x838/0x1890 
[25315.262840]  ? __pfx___schedule+0x10/0x10 
[25315.263150]  ? __blk_flush_plug+0x27b/0x4d0 
[25315.263493]  ? find_held_lock+0x32/0x90 
[25315.263794]  ? __lock_release.isra.0+0x1a4/0x2c0 
[25315.264151]  schedule+0xd4/0x260 
[25315.264425]  io_schedule+0x8f/0xf0 
[25315.264694]  folio_wait_bit_common+0x2d9/0x780 
[25315.265100]  ? folio_wait_bit_common+0x1dd/0x780 
[25315.265473]  ? __pfx_folio_wait_bit_common+0x10/0x10 
[25315.265856]  ? nfs_page_clear_headlock+0x31/0x80 [nfs] 
[25315.266332]  ? __pfx_wake_page_function+0x10/0x10 
[25315.266692]  ? __pfx___might_resched+0x10/0x10 
[25315.267040]  writeback_get_folio+0x3f9/0x500 
[25315.267393]  writeback_iter+0x136/0x720 
[25315.267693]  nfs_writepages+0x4f8/0x9b0 [nfs] 
[25315.268106]  ? mark_held_locks+0x40/0x70 
[25315.268425]  ? __pfx_nfs_writepages+0x10/0x10 [nfs] 
[25315.268861]  ? virtqueue_notify+0x68/0xc0 
[25315.269168]  ? virtio_queue_rq+0x2b1/0x650 [virtio_blk] 
[25315.269597]  ? __lock_acquire+0x57c/0xbd0 
[25315.269921]  do_writepages+0x21f/0x560 
[25315.270212]  ? __pfx_do_writepages+0x10/0x10 
[25315.270555]  ? rcu_is_watching+0x15/0xb0 
[25315.270864]  __writeback_single_inode+0xe2/0x5f0 
[25315.271212]  ? __lock_release.isra.0+0x1a4/0x2c0 
[25315.271584]  ? __pfx___writeback_single_inode+0x10/0x10 
[25315.271975]  ? writeback_sb_inodes+0x416/0xd00 
[25315.272336]  writeback_sb_inodes+0x535/0xd00 
[25315.272675]  ? __pfx_stack_trace_save+0x10/0x10 
[25315.273018]  ? local_clock_noinstr+0xd/0xe0 
[25315.273354]  ? __pfx_writeback_sb_inodes+0x10/0x10 
[25315.273713]  ? __lock_release.isra.0+0x1a4/0x2c0 
[25315.274107]  ? lock_acquire+0x10b/0x150 
[25315.274426]  ? down_read_trylock+0x4b/0x60 
[25315.274743]  __writeback_inodes_wb+0xf4/0x270 
[25315.275083]  ? __pfx___writeback_inodes_wb+0x10/0x10 
[25315.275470]  ? queue_io+0x329/0x510 
[25315.275749]  wb_writeback+0x70a/0x9c0 
[25315.276046]  ? __pfx_wb_writeback+0x10/0x10 
[25315.276389]  ? get_nr_dirty_inodes+0xcb/0x180 
[25315.276729]  wb_do_writeback+0x5d4/0x8e0 
[25315.277042]  ? __pfx_wb_do_writeback+0x10/0x10 
[25315.277401]  ? set_worker_desc+0x16e/0x190 
[25315.277713]  ? __pfx_set_worker_desc+0x10/0x10 
[25315.278072]  wb_workfn+0x7c/0x200 
[25315.278352]  process_one_work+0xd8b/0x1320 
[25315.278678]  ? __pfx_process_one_work+0x10/0x10 
[25315.279034]  ? assign_work+0x16c/0x240 
[25315.279346]  worker_thread+0x5f3/0xfe0 
[25315.279650]  ? __pfx_worker_thread+0x10/0x10 
[25315.279980]  kthread+0x3b4/0x770 
[25315.280229]  ? kvm_sched_clock_read+0x11/0x20 
[25315.280577]  ? local_clock_noinstr+0xd/0xe0 
[25315.280899]  ? __pfx_kthread+0x10/0x10 
[25315.281183]  ? __lock_release.isra.0+0x1a4/0x2c0 
[25315.281551]  ? rcu_is_watching+0x15/0xb0 
[25315.281886]  ? __pfx_kthread+0x10/0x10 
[25315.282176]  ret_from_fork+0x393/0x480 
[25315.282480]  ? __pfx_kthread+0x10/0x10 
[25315.282770]  ? __pfx_kthread+0x10/0x10 
[25315.283058]  ret_from_fork_asm+0x1a/0x30 
[25315.283401]  </TASK> 
[25315.283585]  
[25315.283585] Showing all locks held in the system: 
[25315.284041] 1 lock held by khungtaskd/37: 
[25315.284360]  #0: ffffffffa29309e0 (rcu_read_lock){....}-{1:3}, at: rcu_lock_acquire.constprop.0+0x7/0x30 
[25315.285066] 2 locks held by 751/825733: 
[25315.285370]  #0: ffff888101e36440 (sb_writers#16){.+.+}-{0:0}, at: ksys_write+0xf9/0x1d0 
[25315.285974]  #1: ffffffffa2b97fd0 (split_debug_mutex){+.+.}-{4:4}, at: split_huge_pages_write+0x124/0x430 
[25315.286683] 3 locks held by kworker/u9:73/825900: 
[25315.287039]  #0: ffff8881029df958 ((wq_completion)writeback){+.+.}-{0:0}, at: process_one_work+0x7f5/0x1320 
[25315.287744]  #1: ffffc9000ad9fd10 ((work_completion)(&(&wb->dwork)->work)){+.+.}-{0:0}, at: process_one_work+0xd3f/0x1320 
[25315.288534]  #2: ffff88810b1480e8 (&type->s_umount_key#68){++++}-{4:4}, at: super_trylock_shared+0x1c/0xa0 
[25315.289261] 2 locks held by kworker/u10:2/826054: 
[25315.289612]  #0: ffff888103ec2158 ((wq_completion)nfsiod){+.+.}-{0:0}, at: process_one_work+0x7f5/0x1320 
[25315.290324]  #1: ffffc90001e1fd10 ((work_completion)(&ctx->work)){+.+.}-{0:0}, at: process_one_work+0xd3f/0x1320 
[25315.291064]  
[25315.291199] ============================================= 
[25315.291199]  
[25438.138585] INFO: task kworker/u9:73:825900 blocked for more than 491 seconds. 
[25438.139167]       Not tainted 6.18.0-rc2 #1 
[25438.139486] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. 
[25438.140085] task:kworker/u9:73   state:D stack:24712 pid:825900 tgid:825900 ppid:2      task_flags:0x4208060 flags:0x00080000 
[25438.140963] Workqueue: writeback wb_workfn (flush-0:48) 
[25438.141360] Call Trace: 
[25438.141576]  <TASK> 
[25438.141757]  __schedule+0x838/0x1890 
[25438.142043]  ? __pfx___schedule+0x10/0x10 
[25438.142347]  ? __blk_flush_plug+0x27b/0x4d0 
[25438.142689]  ? find_held_lock+0x32/0x90 
[25438.142987]  ? __lock_release.isra.0+0x1a4/0x2c0 
[25438.143349]  schedule+0xd4/0x260 
[25438.143622]  io_schedule+0x8f/0xf0 
[25438.143891]  folio_wait_bit_common+0x2d9/0x780 
[25438.144263]  ? folio_wait_bit_common+0x1dd/0x780 
[25438.144639]  ? __pfx_folio_wait_bit_common+0x10/0x10 
[25438.145021]  ? nfs_page_clear_headlock+0x31/0x80 [nfs] 
[25438.145479]  ? __pfx_wake_page_function+0x10/0x10 
[25438.145854]  ? __pfx___might_resched+0x10/0x10 
[25438.146205]  writeback_get_folio+0x3f9/0x500 
[25438.146561]  writeback_iter+0x136/0x720 
[25438.146860]  nfs_writepages+0x4f8/0x9b0 [nfs] 
[25438.147265]  ? mark_held_locks+0x40/0x70 
[25438.147582]  ? __pfx_nfs_writepages+0x10/0x10 [nfs] 
[25438.148015]  ? virtqueue_notify+0x68/0xc0 
[25438.148319]  ? virtio_queue_rq+0x2b1/0x650 [virtio_blk] 
[25438.148749]  ? __lock_acquire+0x57c/0xbd0 
[25438.149072]  do_writepages+0x21f/0x560 
[25438.149364]  ? __pfx_do_writepages+0x10/0x10 
[25438.149705]  ? rcu_is_watching+0x15/0xb0 
[25438.150014]  __writeback_single_inode+0xe2/0x5f0 
[25438.150362]  ? __lock_release.isra.0+0x1a4/0x2c0 
[25438.150731]  ? __pfx___writeback_single_inode+0x10/0x10 
[25438.151123]  ? writeback_sb_inodes+0x416/0xd00 
[25438.151466]  writeback_sb_inodes+0x535/0xd00 
[25438.151825]  ? __pfx_stack_trace_save+0x10/0x10 
[25438.152171]  ? local_clock_noinstr+0xd/0xe0 
[25438.152504]  ? __pfx_writeback_sb_inodes+0x10/0x10 
[25438.152864]  ? __lock_release.isra.0+0x1a4/0x2c0 
[25438.153262]  ? lock_acquire+0x10b/0x150 
[25438.153577]  ? down_read_trylock+0x4b/0x60 
[25438.153894]  __writeback_inodes_wb+0xf4/0x270 
[25438.154231]  ? __pfx___writeback_inodes_wb+0x10/0x10 
[25438.154622]  ? queue_io+0x329/0x510 
[25438.154899]  wb_writeback+0x70a/0x9c0 
[25438.155193]  ? __pfx_wb_writeback+0x10/0x10 
[25438.155536]  ? get_nr_dirty_inodes+0xcb/0x180 
[25438.155877]  wb_do_writeback+0x5d4/0x8e0 
[25438.156190]  ? __pfx_wb_do_writeback+0x10/0x10 
[25438.156554]  ? set_worker_desc+0x16e/0x190 
[25438.156865]  ? __pfx_set_worker_desc+0x10/0x10 
[25438.157225]  wb_workfn+0x7c/0x200 
[25438.157486]  process_one_work+0xd8b/0x1320 
[25438.157830]  ? __pfx_process_one_work+0x10/0x10 
[25438.158185]  ? assign_work+0x16c/0x240 
[25438.158478]  worker_thread+0x5f3/0xfe0 
[25438.158800]  ? __pfx_worker_thread+0x10/0x10 
[25438.159129]  kthread+0x3b4/0x770 
[25438.159382]  ? kvm_sched_clock_read+0x11/0x20 
[25438.159730]  ? local_clock_noinstr+0xd/0xe0 
[25438.160049]  ? __pfx_kthread+0x10/0x10 
[25438.160335]  ? __lock_release.isra.0+0x1a4/0x2c0 
[25438.160703]  ? rcu_is_watching+0x15/0xb0 
[25438.161010]  ? __pfx_kthread+0x10/0x10 
[25438.161303]  ret_from_fork+0x393/0x480 
[25438.161605]  ? __pfx_kthread+0x10/0x10 
[25438.161892]  ? __pfx_kthread+0x10/0x10 
[25438.162184]  ret_from_fork_asm+0x1a/0x30 
[25438.162521]  </TASK> 
[25438.162705]  
[25438.162705] Showing all locks held in the system: 
[25438.163161] 1 lock held by khungtaskd/37: 
[25438.163461]  #0: ffffffffa29309e0 (rcu_read_lock){....}-{1:3}, at: rcu_lock_acquire.constprop.0+0x7/0x30 
[25438.164190] 2 locks held by 751/825733: 
[25438.164483]  #0: ffff888101e36440 (sb_writers#16){.+.+}-{0:0}, at: ksys_write+0xf9/0x1d0 
[25438.165101]  #1: ffffffffa2b97fd0 (split_debug_mutex){+.+.}-{4:4}, at: split_huge_pages_write+0x124/0x430 
[25438.165820] 3 locks held by kworker/u9:73/825900: 
[25438.166177]  #0: ffff8881029df958 ((wq_completion)writeback){+.+.}-{0:0}, at: process_one_work+0x7f5/0x1320 
[25438.166895]  #1: ffffc9000ad9fd10 ((work_completion)(&(&wb->dwork)->work)){+.+.}-{0:0}, at: process_one_work+0xd3f/0x1320 
[25438.167703]  #2: ffff88810b1480e8 (&type->s_umount_key#68){++++}-{4:4}, at: super_trylock_shared+0x1c/0xa0 
[25438.168412] 2 locks held by kworker/u10:2/826054: 
[25438.168783]  #0: ffff888103ec2158 ((wq_completion)nfsiod){+.+.}-{0:0}, at: process_one_work+0x7f5/0x1320 
[25438.169473]  #1: ffffc90001e1fd10 ((work_completion)(&ctx->work)){+.+.}-{0:0}, at: process_one_work+0xd3f/0x1320 
[25438.170230]  
[25438.170364] ============================================= 
[25438.170364]  
[-- MARK -- Tue Oct 21 03:50:00 2025] 
[25561.017812] INFO: task kworker/u9:73:825900 blocked for more than 614 seconds. 
[25561.018375]       Not tainted 6.18.0-rc2 #1 
[25561.018694] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. 
[25561.019310] task:kworker/u9:73   state:D stack:24712 pid:825900 tgid:825900 ppid:2      task_flags:0x4208060 flags:0x00080000 
[25561.020172] Workqueue: writeback wb_workfn (flush-0:48) 
[25561.020573] Call Trace: 
[25561.020786]  <TASK> 
[25561.020966]  __schedule+0x838/0x1890 
[25561.021254]  ? __pfx___schedule+0x10/0x10 
[25561.021560]  ? __blk_flush_plug+0x27b/0x4d0 
[25561.021906]  ? find_held_lock+0x32/0x90 
[25561.022200]  ? __lock_release.isra.0+0x1a4/0x2c0 
[25561.022560]  schedule+0xd4/0x260 
[25561.022833]  io_schedule+0x8f/0xf0 
[25561.023102]  folio_wait_bit_common+0x2d9/0x780 
[25561.023475]  ? folio_wait_bit_common+0x1dd/0x780 
[25561.023844]  ? __pfx_folio_wait_bit_common+0x10/0x10 
[25561.024216]  ? nfs_page_clear_headlock+0x31/0x80 [nfs] 
[25561.024686]  ? __pfx_wake_page_function+0x10/0x10 
[25561.025050]  ? __pfx___might_resched+0x10/0x10 
[25561.025389]  writeback_get_folio+0x3f9/0x500 
[25561.025711]  writeback_iter+0x136/0x720 
[25561.026014]  nfs_writepages+0x4f8/0x9b0 [nfs] 
[25561.026402]  ? mark_held_locks+0x40/0x70 
[25561.026698]  ? __pfx_nfs_writepages+0x10/0x10 [nfs] 
[25561.027155]  ? virtqueue_notify+0x68/0xc0 
[25561.027462]  ? virtio_queue_rq+0x2b1/0x650 [virtio_blk] 
[25561.027894]  ? __lock_acquire+0x57c/0xbd0 
[25561.028213]  do_writepages+0x21f/0x560 
[25561.028507]  ? __pfx_do_writepages+0x10/0x10 
[25561.028849]  ? rcu_is_watching+0x15/0xb0 
[25561.029154]  __writeback_single_inode+0xe2/0x5f0 
[25561.029608]  ? __lock_release.isra.0+0x1a4/0x2c0 
[25561.029982]  ? __pfx___writeback_single_inode+0x10/0x10 
[25561.030374]  ? writeback_sb_inodes+0x416/0xd00 
[25561.030715]  writeback_sb_inodes+0x535/0xd00 
[25561.031072]  ? __pfx_stack_trace_save+0x10/0x10 
[25561.031420]  ? local_clock_noinstr+0xd/0xe0 
[25561.031761]  ? __pfx_writeback_sb_inodes+0x10/0x10 
[25561.032121]  ? __lock_release.isra.0+0x1a4/0x2c0 
[25561.032516]  ? lock_acquire+0x10b/0x150 
[25561.032832]  ? down_read_trylock+0x4b/0x60 
[25561.033149]  __writeback_inodes_wb+0xf4/0x270 
[25561.033487]  ? __pfx___writeback_inodes_wb+0x10/0x10 
[25561.033881]  ? queue_io+0x329/0x510 
[25561.034157]  wb_writeback+0x70a/0x9c0 
[25561.034452]  ? __pfx_wb_writeback+0x10/0x10 
[25561.034794]  ? get_nr_dirty_inodes+0xcb/0x180 
[25561.035134]  wb_do_writeback+0x5d4/0x8e0 
[25561.035443]  ? __pfx_wb_do_writeback+0x10/0x10 
[25561.035796]  ? set_worker_desc+0x16e/0x190 
[25561.036105]  ? __pfx_set_worker_desc+0x10/0x10 
[25561.036463]  wb_workfn+0x7c/0x200 
[25561.036724]  process_one_work+0xd8b/0x1320 
[25561.037065]  ? __pfx_process_one_work+0x10/0x10 
[25561.037422]  ? assign_work+0x16c/0x240 
[25561.037718]  worker_thread+0x5f3/0xfe0 
[25561.038035]  ? __pfx_worker_thread+0x10/0x10 
[25561.038355]  kthread+0x3b4/0x770 
[25561.038603]  ? kvm_sched_clock_read+0x11/0x20 
[25561.038951]  ? local_clock_noinstr+0xd/0xe0 
[25561.039272]  ? __pfx_kthread+0x10/0x10 
[25561.039555]  ? __lock_release.isra.0+0x1a4/0x2c0 
[25561.039921]  ? rcu_is_watching+0x15/0xb0 
[25561.040221]  ? __pfx_kthread+0x10/0x10 
[25561.040511]  ret_from_fork+0x393/0x480 
[25561.040813]  ? __pfx_kthread+0x10/0x10 
[25561.041101]  ? __pfx_kthread+0x10/0x10 
[25561.041392]  ret_from_fork_asm+0x1a/0x30 
[25561.041707]  </TASK> 
[25561.041908]  
[25561.041908] Showing all locks held in the system: 
[25561.042365] 1 lock held by khungtaskd/37: 
[25561.042662]  #0: ffffffffa29309e0 (rcu_read_lock){....}-{1:3}, at: rcu_lock_acquire.constprop.0+0x7/0x30 
[25561.043384] 2 locks held by 751/825733: 
[25561.043677]  #0: ffff888101e36440 (sb_writers#16){.+.+}-{0:0}, at: ksys_write+0xf9/0x1d0 
[25561.044296]  #1: ffffffffa2b97fd0 (split_debug_mutex){+.+.}-{4:4}, at: split_huge_pages_write+0x124/0x430 
[25561.045009] 3 locks held by kworker/u9:73/825900: 
[25561.045362]  #0: ffff8881029df958 ((wq_completion)writeback){+.+.}-{0:0}, at: process_one_work+0x7f5/0x1320 
[25561.046085]  #1: ffffc9000ad9fd10 ((work_completion)(&(&wb->dwork)->work)){+.+.}-{0:0}, at: process_one_work+0xd3f/0x1320 
[25561.046895]  #2: ffff88810b1480e8 (&type->s_umount_key#68){++++}-{4:4}, at: super_trylock_shared+0x1c/0xa0 
[25561.047600] 2 locks held by kworker/u10:13/826064: 
[25561.047970]  #0: ffff888103ec2158 ((wq_completion)nfsiod){+.+.}-{0:0}, at: process_one_work+0x7f5/0x1320 
[25561.048657]  #1: ffffc90001e9fd10 ((work_completion)(&ctx->work)){+.+.}-{0:0}, at: process_one_work+0xd3f/0x1320 
[25561.049410]  
[25561.049538] ============================================= 
[25561.049538]  
[25683.897277] INFO: task kworker/u9:73:825900 blocked for more than 737 seconds. 
[25683.897831]       Not tainted 6.18.0-rc2 #1 
[25683.898182] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. 
[25683.898754] task:kworker/u9:73   state:D stack:24712 pid:825900 tgid:825900 ppid:2      task_flags:0x4208060 flags:0x00080000 
[25683.899587] Workqueue: writeback wb_workfn (flush-0:48) 
[25683.900001] Call Trace: 
[25683.900200]  <TASK> 
[25683.900379]  __schedule+0x838/0x1890 
[25683.900663]  ? __pfx___schedule+0x10/0x10 
[25683.900967]  ? __blk_flush_plug+0x27b/0x4d0 
[25683.901308]  ? find_held_lock+0x32/0x90 
[25683.901609]  ? __lock_release.isra.0+0x1a4/0x2c0 
[25683.901964]  schedule+0xd4/0x260 
[25683.902235]  io_schedule+0x8f/0xf0 
[25683.902507]  folio_wait_bit_common+0x2d9/0x780 
[25683.902874]  ? folio_wait_bit_common+0x1dd/0x780 
[25683.903256]  ? __pfx_folio_wait_bit_common+0x10/0x10 
[25683.903636]  ? nfs_page_clear_headlock+0x31/0x80 [nfs] 
[25683.904111]  ? __pfx_wake_page_function+0x10/0x10 
[25683.904469]  ? __pfx___might_resched+0x10/0x10 
[25683.904819]  writeback_get_folio+0x3f9/0x500 
[25683.905170]  writeback_iter+0x136/0x720 
[25683.905469]  nfs_writepages+0x4f8/0x9b0 [nfs] 
[25683.905878]  ? mark_held_locks+0x40/0x70 
[25683.906194]  ? __pfx_nfs_writepages+0x10/0x10 [nfs] 
[25683.906631]  ? virtqueue_notify+0x68/0xc0 
[25683.906936]  ? virtio_queue_rq+0x2b1/0x650 [virtio_blk] 
[25683.907368]  ? __lock_acquire+0x57c/0xbd0 
[25683.907691]  do_writepages+0x21f/0x560 
[25683.907999]  ? __pfx_do_writepages+0x10/0x10 
[25683.908325]  ? rcu_is_watching+0x15/0xb0 
[25683.908635]  __writeback_single_inode+0xe2/0x5f0 
[25683.909013]  ? __lock_release.isra.0+0x1a4/0x2c0 
[25683.909368]  ? __pfx___writeback_single_inode+0x10/0x10 
[25683.909756]  ? writeback_sb_inodes+0x416/0xd00 
[25683.910118]  writeback_sb_inodes+0x535/0xd00 
[25683.910459]  ? __pfx_stack_trace_save+0x10/0x10 
[25683.910807]  ? local_clock_noinstr+0xd/0xe0 
[25683.911138]  ? __pfx_writeback_sb_inodes+0x10/0x10 
[25683.911500]  ? __lock_release.isra.0+0x1a4/0x2c0 
[25683.911896]  ? lock_acquire+0x10b/0x150 
[25683.912211]  ? down_read_trylock+0x4b/0x60 
[25683.912532]  __writeback_inodes_wb+0xf4/0x270 
[25683.912869]  ? __pfx___writeback_inodes_wb+0x10/0x10 
[25683.913259]  ? queue_io+0x329/0x510 
[25683.913540]  wb_writeback+0x70a/0x9c0 
[25683.913832]  ? __pfx_wb_writeback+0x10/0x10 
[25683.914174]  ? get_nr_dirty_inodes+0xcb/0x180 
[25683.914516]  wb_do_writeback+0x5d4/0x8e0 
[25683.914831]  ? __pfx_wb_do_writeback+0x10/0x10 
[25683.915185]  ? set_worker_desc+0x16e/0x190 
[25683.915501]  ? __pfx_set_worker_desc+0x10/0x10 
[25683.915860]  wb_workfn+0x7c/0x200 
[25683.916139]  process_one_work+0xd8b/0x1320 
[25683.916465]  ? __pfx_process_one_work+0x10/0x10 
[25683.916822]  ? assign_work+0x16c/0x240 
[25683.917133]  worker_thread+0x5f3/0xfe0 
[25683.917437]  ? __pfx_worker_thread+0x10/0x10 
[25683.917767]  kthread+0x3b4/0x770 
[25683.918037]  ? kvm_sched_clock_read+0x11/0x20 
[25683.918368]  ? local_clock_noinstr+0xd/0xe0 
[25683.918687]  ? __pfx_kthread+0x10/0x10 
[25683.918969]  ? __lock_release.isra.0+0x1a4/0x2c0 
[25683.919334]  ? rcu_is_watching+0x15/0xb0 
[25683.919640]  ? __pfx_kthread+0x10/0x10 
[25683.919929]  ret_from_fork+0x393/0x480 
[25683.920234]  ? __pfx_kthread+0x10/0x10 
[25683.920523]  ? __pfx_kthread+0x10/0x10 
[25683.920815]  ret_from_fork_asm+0x1a/0x30 
[25683.921152]  </TASK> 
[25683.921333] Future hung task reports are suppressed, see sysctl kernel.hung_task_warnings 
[25683.921921]  
[25683.921921] Showing all locks held in the system: 
[25683.922392] 1 lock held by khungtaskd/37: 
[25683.922698]  #0: ffffffffa29309e0 (rcu_read_lock){....}-{1:3}, at: rcu_lock_acquire.constprop.0+0x7/0x30 
[25683.923423] 2 locks held by 751/825733: 
[25683.923720]  #0: ffff888101e36440 (sb_writers#16){.+.+}-{0:0}, at: ksys_write+0xf9/0x1d0 
[25683.924331]  #1: ffffffffa2b97fd0 (split_debug_mutex){+.+.}-{4:4}, at: split_huge_pages_write+0x124/0x430 
[25683.925045] 3 locks held by kworker/u9:73/825900: 
[25683.925398]  #0: ffff8881029df958 ((wq_completion)writeback){+.+.}-{0:0}, at: process_one_work+0x7f5/0x1320 
[25683.926121]  #1: ffffc9000ad9fd10 ((work_completion)(&(&wb->dwork)->work)){+.+.}-{0:0}, at: process_one_work+0xd3f/0x1320 
[25683.926918]  #2: ffff88810b1480e8 (&type->s_umount_key#68){++++}-{4:4}, at: super_trylock_shared+0x1c/0xa0 
[25683.927641] 4 locks held by kworker/u9:143/826036: 
[25683.928022]  #0: ffff888103ec5158 ((wq_completion)nfslocaliod){+.+.}-{0:0}, at: process_one_work+0x7f5/0x1320 
[25683.928743]  #1: ffffc90001cc7d10 ((work_completion)(&iocb->work)#2){+.+.}-{0:0}, at: process_one_work+0xd3f/0x1320 
[25683.929514]  #2: ffff88815c2c2440 (sb_writers#13){++++}-{0:0}, at: process_one_work+0xd8b/0x1320 
[25683.930175]  #3: ffff8881a1324b58 (&sb->s_type->i_mutex_key#13){++++}-{4:4}, at: xfs_ilock+0x360/0x460 [xfs] 
[25683.931255]  
[25683.931391] ============================================= 
[25683.931391]  
[-- MARK -- Tue Oct 21 03:55:00 2025] 
[-- MARK -- Tue Oct 21 04:00:00 2025] 
[-- MARK -- Tue Oct 21 04:05:00 2025] 


[2]
[12082.409406] run fstests generic/751 at 2025-10-20 19:04:12 
[-- MARK -- Mon Oct 20 23:05:00 2025] 
[-- MARK -- Mon Oct 20 23:10:00 2025] 
[-- MARK -- Mon Oct 20 23:15:00 2025] 
[13023.321998] INFO: task kworker/u17:206:858774 blocked for more than 122 seconds. 
[13023.322058]       Tainted: G    B               6.18.0-rc2 #1 
[13023.322062] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. 
[13023.322064] task:kworker/u17:206 state:D stack:23184 pid:858774 tgid:858774 ppid:2      task_flags:0x4208060 flags:0x00000210 
[13023.322077] Workqueue: writeback wb_workfn (flush-0:56) 
[13023.322089] Call trace: 
[13023.322091]  __switch_to+0x1e0/0x488 (T) 
[13023.322098]  __schedule+0x748/0x1430 
[13023.322103]  schedule+0xd0/0x240 
[13023.322107]  io_schedule+0xb4/0x120 
[13023.322112]  folio_wait_bit_common+0x2c4/0x708 
[13023.322117]  __folio_lock+0x24/0x38 
[13023.322120]  writeback_get_folio+0x37c/0x480 
[13023.322125]  writeback_iter+0x128/0x6d0 
[13023.322130]  nfs_writepages+0x504/0x930 [nfs] 
[13023.322194]  do_writepages+0x204/0x4c0 
[13023.322198]  __writeback_single_inode+0xec/0x4f8 
[13023.322204]  writeback_sb_inodes+0x4a4/0xab8 
[13023.322296]  __writeback_inodes_wb+0x104/0x260 
[13023.322301]  wb_writeback+0x840/0xce0 
[13023.322305]  wb_do_writeback+0x69c/0x940 
[13023.322308]  wb_workfn+0x80/0x1c0 
[13023.322312]  process_one_work+0x774/0x12d0 
[13023.322318]  worker_thread+0x434/0xca0 
[13023.322322]  kthread+0x2ec/0x390 
[13023.322326]  ret_from_fork+0x10/0x20 
[13023.322409] INFO: lockdep is turned off. 
[-- MARK -- Mon Oct 20 23:20:00 2025] 
[13146.212082] INFO: task kworker/u17:206:858774 blocked for more than 245 seconds. 
[13146.212109]       Tainted: G    B               6.18.0-rc2 #1 
[13146.212120] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. 
[13146.212125] task:kworker/u17:206 state:D stack:23184 pid:858774 tgid:858774 ppid:2      task_flags:0x4208060 flags:0x00000210 
[13146.212141] Workqueue: writeback wb_workfn (flush-0:56) 
[13146.212160] Call trace: 
[13146.212165]  __switch_to+0x1e0/0x488 (T) 
[13146.212175]  __schedule+0x748/0x1430 
[13146.212182]  schedule+0xd0/0x240 
[13146.212189]  io_schedule+0xb4/0x120 
[13146.212196]  folio_wait_bit_common+0x2c4/0x708 
[13146.212203]  __folio_lock+0x24/0x38 
[13146.212209]  writeback_get_folio+0x37c/0x480 
[13146.212217]  writeback_iter+0x128/0x6d0 
[13146.212222]  nfs_writepages+0x504/0x930 [nfs] 
[13146.212287]  do_writepages+0x204/0x4c0 
[13146.212291]  __writeback_single_inode+0xec/0x4f8 
[13146.212294]  writeback_sb_inodes+0x4a4/0xab8 
[13146.212298]  __writeback_inodes_wb+0x104/0x260 
[13146.212302]  wb_writeback+0x840/0xce0 
[13146.212306]  wb_do_writeback+0x69c/0x940 
[13146.212309]  wb_workfn+0x80/0x1c0 
[13146.212313]  process_one_work+0x774/0x12d0 
[13146.212318]  worker_thread+0x434/0xca0 
[13146.212321]  kthread+0x2ec/0x390 
[13146.212325]  ret_from_fork+0x10/0x20 
[13146.212360] INFO: lockdep is turned off. 
[13269.082400] INFO: task kworker/u17:206:858774 blocked for more than 368 seconds. 
[13269.082553]       Tainted: G    B               6.18.0-rc2 #1 
[13269.082557] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. 
[13269.082559] task:kworker/u17:206 state:D stack:23184 pid:858774 tgid:858774 ppid:2      task_flags:0x4208060 flags:0x00000210 
[13269.082570] Workqueue: writeback wb_workfn (flush-0:56) 
[13269.082581] Call trace: 
[13269.082583]  __switch_to+0x1e0/0x488 (T) 
[13269.082591]  __schedule+0x748/0x1430 
[13269.082596]  schedule+0xd0/0x240 
[13269.082600]  io_schedule+0xb4/0x120 
[13269.082604]  folio_wait_bit_common+0x2c4/0x708 
[13269.082610]  __folio_lock+0x24/0x38 
[13269.082613]  writeback_get_folio+0x37c/0x480 
[13269.082618]  writeback_iter+0x128/0x6d0 
[13269.082622]  nfs_writepages+0x504/0x930 [nfs] 
[13269.082749]  do_writepages+0x204/0x4c0 
[13269.082754]  __writeback_single_inode+0xec/0x4f8 
[13269.082758]  writeback_sb_inodes+0x4a4/0xab8 
[13269.082762]  __writeback_inodes_wb+0x104/0x260 
[13269.082766]  wb_writeback+0x840/0xce0 
[13269.082769]  wb_do_writeback+0x69c/0x940 
[13269.082773]  wb_workfn+0x80/0x1c0 
[13269.082777]  process_one_work+0x774/0x12d0 
[13269.082782]  worker_thread+0x434/0xca0 
[13269.082785]  kthread+0x2ec/0x390 
[13269.082789]  ret_from_fork+0x10/0x20 
[13269.082805] INFO: lockdep is turned off. 
[-- MARK -- Mon Oct 20 23:25:00 2025] 
[13391.972505] INFO: task kworker/u17:206:858774 blocked for more than 491 seconds. 
[13391.972535]       Tainted: G    B               6.18.0-rc2 #1 
[13391.972549] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. 
[13391.972555] task:kworker/u17:206 state:D stack:23184 pid:858774 tgid:858774 ppid:2      task_flags:0x4208060 flags:0x00000210 
[13391.972571] Workqueue: writeback wb_workfn (flush-0:56) 
[13391.972585] Call trace: 
[13391.972587]  __switch_to+0x1e0/0x488 (T) 
[13391.972595]  __schedule+0x748/0x1430 
[13391.972600]  schedule+0xd0/0x240 
[13391.972605]  io_schedule+0xb4/0x120 
[13391.972609]  folio_wait_bit_common+0x2c4/0x708 
[13391.972614]  __folio_lock+0x24/0x38 
[13391.972618]  writeback_get_folio+0x37c/0x480 
[13391.972624]  writeback_iter+0x128/0x6d0 
[13391.972628]  nfs_writepages+0x504/0x930 [nfs] 
[13391.972700]  do_writepages+0x204/0x4c0 
[13391.972704]  __writeback_single_inode+0xec/0x4f8 
[13391.972708]  writeback_sb_inodes+0x4a4/0xab8 
[13391.972712]  __writeback_inodes_wb+0x104/0x260 
[13391.972716]  wb_writeback+0x840/0xce0 
[13391.972720]  wb_do_writeback+0x69c/0x940 
[13391.972724]  wb_workfn+0x80/0x1c0 
[13391.972727]  process_one_work+0x774/0x12d0 
[13391.972733]  worker_thread+0x434/0xca0 
[13391.972736]  kthread+0x2ec/0x390 
[13391.972740]  ret_from_fork+0x10/0x20 
[13391.972761] INFO: lockdep is turned off. 
[13514.852712] INFO: task kworker/u17:206:858774 blocked for more than 614 seconds. 
[13514.852737]       Tainted: G    B               6.18.0-rc2 #1 
[13514.852749] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. 
[13514.852756] task:kworker/u17:206 state:D stack:23184 pid:858774 tgid:858774 ppid:2      task_flags:0x4208060 flags:0x00000210 
[13514.852768] Workqueue: writeback wb_workfn (flush-0:56) 
[13514.852779] Call trace: 
[13514.852781]  __switch_to+0x1e0/0x488 (T) 
[13514.852790]  __schedule+0x748/0x1430 
[13514.852795]  schedule+0xd0/0x240 
[13514.852808]  io_schedule+0xb4/0x120 
[13514.852812]  folio_wait_bit_common+0x2c4/0x708 
[13514.852818]  __folio_lock+0x24/0x38 
[13514.852822]  writeback_get_folio+0x37c/0x480 
[13514.852827]  writeback_iter+0x128/0x6d0 
[13514.852831]  nfs_writepages+0x504/0x930 [nfs] 
[13514.852897]  do_writepages+0x204/0x4c0 
[13514.852901]  __writeback_single_inode+0xec/0x4f8 
[13514.852905]  writeback_sb_inodes+0x4a4/0xab8 
[13514.852908]  __writeback_inodes_wb+0x104/0x260 
[13514.852912]  wb_writeback+0x840/0xce0 
[13514.852916]  wb_do_writeback+0x69c/0x940 
[13514.852919]  wb_workfn+0x80/0x1c0 
[13514.852923]  process_one_work+0x774/0x12d0 
[13514.852928]  worker_thread+0x434/0xca0 
[13514.852932]  kthread+0x2ec/0x390 
[13514.852936]  ret_from_fork+0x10/0x20 
[13514.852955] INFO: lockdep is turned off. 
[-- MARK -- Mon Oct 20 23:30:00 2025] 
[13637.723028] INFO: task kworker/u17:206:858774 blocked for more than 737 seconds. 
[13637.723055]       Tainted: G    B               6.18.0-rc2 #1 
[13637.723067] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. 
[13637.723072] task:kworker/u17:206 state:D stack:23184 pid:858774 tgid:858774 ppid:2      task_flags:0x4208060 flags:0x00000210 
[13637.723086] Workqueue: writeback wb_workfn (flush-0:56) 
[13637.723099] Call trace: 
[13637.723101]  __switch_to+0x1e0/0x488 (T) 
[13637.723110]  __schedule+0x748/0x1430 
[13637.723115]  schedule+0xd0/0x240 
[13637.723119]  io_schedule+0xb4/0x120 
[13637.723124]  folio_wait_bit_common+0x2c4/0x708 
[13637.723129]  __folio_lock+0x24/0x38 
[13637.723133]  writeback_get_folio+0x37c/0x480 
[13637.723138]  writeback_iter+0x128/0x6d0 
[13637.723142]  nfs_writepages+0x504/0x930 [nfs] 
[13637.723210]  do_writepages+0x204/0x4c0 
[13637.723213]  __writeback_single_inode+0xec/0x4f8 
[13637.723217]  writeback_sb_inodes+0x4a4/0xab8 
[13637.723221]  __writeback_inodes_wb+0x104/0x260 
[13637.723225]  wb_writeback+0x840/0xce0 
[13637.723229]  wb_do_writeback+0x69c/0x940 
[13637.723232]  wb_workfn+0x80/0x1c0 
[13637.723237]  process_one_work+0x774/0x12d0 
[13637.723242]  worker_thread+0x434/0xca0 
[13637.723246]  kthread+0x2ec/0x390 
[13637.723250]  ret_from_fork+0x10/0x20 
[13637.723271] INFO: lockdep is turned off. 
[13760.603116] INFO: task kworker/u17:206:858774 blocked for more than 860 seconds. 
[13760.603144]       Tainted: G    B               6.18.0-rc2 #1 
[13760.603158] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. 
[13760.603163] task:kworker/u17:206 state:D stack:23184 pid:858774 tgid:858774 ppid:2      task_flags:0x4208060 flags:0x00000210 
[13760.603180] Workqueue: writeback wb_workfn (flush-0:56) 
[13760.603199] Call trace: 
[13760.603203]  __switch_to+0x1e0/0x488 (T) 
[13760.603214]  __schedule+0x748/0x1430 
[13760.603221]  schedule+0xd0/0x240 
[13760.603228]  io_schedule+0xb4/0x120 
[13760.603234]  folio_wait_bit_common+0x2c4/0x708 
[13760.603242]  __folio_lock+0x24/0x38 
[13760.603248]  writeback_get_folio+0x37c/0x480 
[13760.603341]  writeback_iter+0x128/0x6d0 
[13760.603349]  nfs_writepages+0x504/0x930 [nfs] 
[13760.603421]  do_writepages+0x204/0x4c0 
[13760.603427]  __writeback_single_inode+0xec/0x4f8 
[13760.603434]  writeback_sb_inodes+0x4a4/0xab8 
[13760.603440]  __writeback_inodes_wb+0x104/0x260 
[13760.603447]  wb_writeback+0x840/0xce0 
[13760.603453]  wb_do_writeback+0x69c/0x940 
[13760.603459]  wb_workfn+0x80/0x1c0 
[13760.603465]  process_one_work+0x774/0x12d0 
[13760.603472]  worker_thread+0x434/0xca0 
[13760.603478]  kthread+0x2ec/0x390 
[13760.603485]  ret_from_fork+0x10/0x20 
[13760.603507] INFO: lockdep is turned off. 
[13883.483295] INFO: task kworker/u17:206:858774 blocked for more than 983 seconds. 
[13883.483320]       Tainted: G    B               6.18.0-rc2 #1 
[13883.483340] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. 
[13883.483343] task:kworker/u17:206 state:D stack:23184 pid:858774 tgid:858774 ppid:2      task_flags:0x4208060 flags:0x00000210 
[13883.483356] Workqueue: writeback wb_workfn (flush-0:56) 
[13883.483368] Call trace: 
[13883.483371]  __switch_to+0x1e0/0x488 (T) 
[13883.483379]  __schedule+0x748/0x1430 
[13883.483384]  schedule+0xd0/0x240 
[13883.483388]  io_schedule+0xb4/0x120 
[13883.483392]  folio_wait_bit_common+0x2c4/0x708 
[13883.483397]  __folio_lock+0x24/0x38 
[13883.483401]  writeback_get_folio+0x37c/0x480 
[13883.483406]  writeback_iter+0x128/0x6d0 
[13883.483411]  nfs_writepages+0x504/0x930 [nfs] 
[13883.483477]  do_writepages+0x204/0x4c0 
[13883.483481]  __writeback_single_inode+0xec/0x4f8 
[13883.483485]  writeback_sb_inodes+0x4a4/0xab8 
[13883.483489]  __writeback_inodes_wb+0x104/0x260 
[13883.483493]  wb_writeback+0x840/0xce0 
[13883.483496]  wb_do_writeback+0x69c/0x940 
[13883.483500]  wb_workfn+0x80/0x1c0 
[13883.483504]  process_one_work+0x774/0x12d0 
[13883.483509]  worker_thread+0x434/0xca0 
[13883.483512]  kthread+0x2ec/0x390 
[13883.483517]  ret_from_fork+0x10/0x20 
[13883.483538] INFO: lockdep is turned off. 
[-- MARK -- Mon Oct 20 23:35:00 2025] 
[14006.363497] INFO: task kworker/u17:206:858774 blocked for more than 1105 seconds. 
[14006.363521]       Tainted: G    B               6.18.0-rc2 #1 
[14006.363542] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. 
[14006.363545] task:kworker/u17:206 state:D stack:23184 pid:858774 tgid:858774 ppid:2      task_flags:0x4208060 flags:0x00000210 
[14006.363557] Workqueue: writeback wb_workfn (flush-0:56) 
[14006.363569] Call trace: 
[14006.363571]  __switch_to+0x1e0/0x488 (T) 
[14006.363579]  __schedule+0x748/0x1430 
[14006.363584]  schedule+0xd0/0x240 
[14006.363588]  io_schedule+0xb4/0x120 
[14006.363592]  folio_wait_bit_common+0x2c4/0x708 
[14006.363597]  __folio_lock+0x24/0x38 
[14006.363601]  writeback_get_folio+0x37c/0x480 
[14006.363606]  writeback_iter+0x128/0x6d0 
[14006.363610]  nfs_writepages+0x504/0x930 [nfs] 
[14006.363676]  do_writepages+0x204/0x4c0 
[14006.363680]  __writeback_single_inode+0xec/0x4f8 
[14006.363684]  writeback_sb_inodes+0x4a4/0xab8 
[14006.363687]  __writeback_inodes_wb+0x104/0x260 
[14006.363691]  wb_writeback+0x840/0xce0 
[14006.363695]  wb_do_writeback+0x69c/0x940 
[14006.363699]  wb_workfn+0x80/0x1c0 
[14006.363702]  process_one_work+0x774/0x12d0 
[14006.363707]  worker_thread+0x434/0xca0 
[14006.363711]  kthread+0x2ec/0x390 
[14006.363715]  ret_from_fork+0x10/0x20 
[14006.363735] INFO: lockdep is turned off. 
[14129.243869] INFO: task kworker/u17:206:858774 blocked for more than 1228 seconds. 
[14129.244026]       Tainted: G    B               6.18.0-rc2 #1 
[14129.244030] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. 
[14129.244033] task:kworker/u17:206 state:D stack:23184 pid:858774 tgid:858774 ppid:2      task_flags:0x4208060 flags:0x00000210 
[14129.244045] Workqueue: writeback wb_workfn (flush-0:56) 
[14129.244059] Call trace: 
[14129.244061]  __switch_to+0x1e0/0x488 (T) 
[14129.244070]  __schedule+0x748/0x1430 
[14129.244075]  schedule+0xd0/0x240 
[14129.244079]  io_schedule+0xb4/0x120 
[14129.244084]  folio_wait_bit_common+0x2c4/0x708 
[14129.244089]  __folio_lock+0x24/0x38 
[14129.244092]  writeback_get_folio+0x37c/0x480 
[14129.244097]  writeback_iter+0x128/0x6d0 
[14129.244102]  nfs_writepages+0x504/0x930 [nfs] 
[14129.244173]  do_writepages+0x204/0x4c0 
[14129.244176]  __writeback_single_inode+0xec/0x4f8 
[14129.244180]  writeback_sb_inodes+0x4a4/0xab8 
[14129.244184]  __writeback_inodes_wb+0x104/0x260 
[14129.244188]  wb_writeback+0x840/0xce0 
[14129.244192]  wb_do_writeback+0x69c/0x940 
[14129.244195]  wb_workfn+0x80/0x1c0 
[14129.244199]  process_one_work+0x774/0x12d0 
[14129.244205]  worker_thread+0x434/0xca0 
[14129.244209]  kthread+0x2ec/0x390 
[14129.244213]  ret_from_fork+0x10/0x20 
[14129.244217] Future hung task reports are suppressed, see sysctl kernel.hung_task_warnings 
[14129.244242] INFO: lockdep is turned off. 
[-- MARK -- Mon Oct 20 23:40:00 2025] 
[-- MARK -- Mon Oct 20 23:45:00 2025] 
[-- MARK -- Mon Oct 20 23:50:00 2025]



^ permalink raw reply	[flat|nested] 3+ messages in thread

* Re: [Bug report][xfstests generic/751] hang on nfs writeback
  2025-10-21  5:14 [Bug report][xfstests generic/751] hang on nfs writeback Zorro Lang
@ 2025-10-22 16:24 ` Mike Snitzer
  2025-10-29 20:50   ` Mike Snitzer
  0 siblings, 1 reply; 3+ messages in thread
From: Mike Snitzer @ 2025-10-22 16:24 UTC (permalink / raw)
  To: Zorro Lang; +Cc: linux-nfs, linux-mm, linux-xfs

On Tue, Oct 21, 2025 at 01:14:08PM +0800, Zorro Lang wrote:
> Hi,
> 
> When I did xfstests regression test on nfs, it hang on generic/751 many times,
> refer to [1](on x86_64) and [2](on aarch64). I've hit it 8 times until now.

You've hit it 8 times, all on the same kernel?  Any idea if this is a
regression specific to 6.18-rc1 and later?
 
> I tested on latest linux v6.18-rc2. My underlying test device isn't loopback device,
> it's general disk, and make xfs on it:
>   meta-data=/dev/sda5              isize=512    agcount=4, agsize=983040 blks
>            =                       sectsz=4096  attr=2, projid32bit=1
>            =                       crc=1        finobt=1, sparse=1, rmapbt=1
>            =                       reflink=1    bigtime=1 inobtcount=1 nrext64=1
>            =                       exchange=0   metadir=0
>   data     =                       bsize=4096   blocks=3932160, imaxpct=25
>            =                       sunit=0      swidth=0 blks
>   naming   =version 2              bsize=4096   ascii-ci=0, ftype=1, parent=0
>   log      =internal log           bsize=4096   blocks=54324, version=2
>            =                       sectsz=4096  sunit=1 blks, lazy-count=1
>   realtime =none                   extsz=4096   blocks=0, rtextents=0
>            =                       rgcount=0    rgsize=0 extents
>            =                       zoned=0      start=0 reserved=0
> 
> Two xfs are mounted on /mnt/xfstests/test and /mnt/xfstests/scratch seperately,
> then export as:
>   # cat /etc/exports
>   /mnt/xfstests/test/nfs-server *(rw,insecure,no_root_squash)
>   /mnt/xfstests/scratch/nfs-server *(rw,insecure,no_root_squash)
> 
> The nfs mount option is only "-o vers=4.2". BTW, nfs server and client are
> in same machine/system.

Being that NFS client and server are on the same localhost, I see
NFS's LOCALIO is being used (nfslocaliod mentioned in "locks held"
section, so IO is bypassing going over the wire to NFSD and is instead
being issued directly from NFS client through LOCALIO workqueue and
then down to XFS).

I'll try to make sense of the locking (reported below) but this will
take some time.

Cc'ing XFS list because this is NFS LOCALIO ontop of XFS, could be
there is something fundamental in LOCALIO's nfslocaliod workqueue that
is setting us up for deadlock if XFS takes a specific path...

NOTE: fs/nfs/inode.c:nfsiod_start(), XFS's nature has influenced
LOCALIO workqueue config hacks already.. could be dragons lurking:

/*
 * Start the nfsiod workqueues
 */
static int nfsiod_start(void)
{
        dprintk("RPC:       creating workqueue nfsiod\n");
        nfsiod_workqueue = alloc_workqueue("nfsiod", WQ_MEM_RECLAIM | WQ_UNBOUND, 0);
        if (nfsiod_workqueue == NULL)
                return -ENOMEM;
#if IS_ENABLED(CONFIG_NFS_LOCALIO)
        /*
         * localio writes need to use a normal (non-memreclaim) workqueue.
         * When we start getting low on space, XFS goes and calls flush_work() on
         * a non-memreclaim work queue, which causes a priority inversion problem.
         */
        dprintk("RPC:       creating workqueue nfslocaliod\n");
        nfslocaliod_workqueue = alloc_workqueue("nfslocaliod", WQ_UNBOUND, 0);
        if (unlikely(nfslocaliod_workqueue == NULL)) {
                nfsiod_stop();
                return -ENOMEM;
        }
#endif /* CONFIG_NFS_LOCALIO */
        return 0;
}

Mike

 
> # cat local.config
> export FSTYP=nfs
> export TEST_DEV=xxxx-xxx-xxxx.xxxx.xxx.xxx:/mnt/xfstests/test/nfs-server
> export TEST_DIR=/mnt/xfstests/test/nfs-client
> export SCRATCH_DEV=xxxx-xxx-xxxx.xxxx.xxx.xxx:/mnt/xfstests/scratch/nfs-server
> export SCRATCH_MNT=/mnt/xfstests/scratch/nfs-client
> export MOUNT_OPTIONS="-o vers=4.2"
> export TEST_FS_MOUNT_OPTS="-o vers=4.2"
> 
> 
> [1]
> [23369.572660] run fstests generic/751 at 2025-10-20 23:15:24 
> [-- MARK -- Tue Oct 21 03:20:00 2025] 
> [-- MARK -- Tue Oct 21 03:25:00 2025] 
> [-- MARK -- Tue Oct 21 03:30:00 2025] 
> [-- MARK -- Tue Oct 21 03:35:00 2025] 
> [-- MARK -- Tue Oct 21 03:40:00 2025] 
> [25069.500900] INFO: task kworker/u9:73:825900 blocked for more than 122 seconds. 
> [25069.501484]       Not tainted 6.18.0-rc2 #1 
> [25069.501822] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. 
> [25069.502385] task:kworker/u9:73   state:D stack:24712 pid:825900 tgid:825900 ppid:2      task_flags:0x4208060 flags:0x00080000 
> [25069.503258] Workqueue: writeback wb_workfn (flush-0:48) 
> [25069.782725] Call Trace: 
> [25069.787191]  <TASK> 
> [25069.787368]  __schedule+0x838/0x1890 
> [25069.799533]  ? __pfx___schedule+0x10/0x10 
> [25069.799852]  ? __blk_flush_plug+0x27b/0x4d0 
> [25069.800617]  ? find_held_lock+0x32/0x90 
> [25069.801292]  ? __lock_release.isra.0+0x1a4/0x2c0 
> [25069.801639]  schedule+0xd4/0x260 
> [25069.801902]  io_schedule+0x8f/0xf0 
> [25069.802162]  folio_wait_bit_common+0x2d9/0x780 
> [25069.802919]  ? folio_wait_bit_common+0x1dd/0x780 
> [25069.803324]  ? __pfx_folio_wait_bit_common+0x10/0x10 
> [25069.803684]  ? nfs_page_clear_headlock+0x31/0x80 [nfs] 
> [25069.804254]  ? __pfx_wake_page_function+0x10/0x10 
> [25069.804602]  ? __pfx___might_resched+0x10/0x10 
> [25069.805160]  writeback_get_folio+0x3f9/0x500 
> [25069.806221]  writeback_iter+0x136/0x720 
> [25069.806628]  nfs_writepages+0x4f8/0x9b0 [nfs] 
> [25069.807100]  ? mark_held_locks+0x40/0x70 
> [25069.807403]  ? __pfx_nfs_writepages+0x10/0x10 [nfs] 
> [25069.807846]  ? virtqueue_notify+0x68/0xc0 
> [25069.808346]  ? virtio_queue_rq+0x2b1/0x650 [virtio_blk] 
> [25069.808925]  ? __lock_acquire+0x57c/0xbd0 
> [25069.809295]  do_writepages+0x21f/0x560 
> [25069.809576]  ? __pfx_do_writepages+0x10/0x10 
> [25069.809953]  ? rcu_is_watching+0x15/0xb0 
> [25069.810587]  __writeback_single_inode+0xe2/0x5f0 
> [25069.810982]  ? __lock_release.isra.0+0x1a4/0x2c0 
> [25069.811373]  ? __pfx___writeback_single_inode+0x10/0x10 
> [25069.811815]  ? writeback_sb_inodes+0x416/0xd00 
> [25069.812194]  writeback_sb_inodes+0x535/0xd00 
> [25069.812533]  ? __pfx_stack_trace_save+0x10/0x10 
> [25069.814193]  ? local_clock_noinstr+0xd/0xe0 
> [25069.814506]  ? __pfx_writeback_sb_inodes+0x10/0x10 
> [25069.814910]  ? __lock_release.isra.0+0x1a4/0x2c0 
> [25069.815407]  ? lock_acquire+0x10b/0x150 
> [25069.815703]  ? down_read_trylock+0x4b/0x60 
> [25069.816077]  __writeback_inodes_wb+0xf4/0x270 
> [25069.816441]  ? __pfx___writeback_inodes_wb+0x10/0x10 
> [25069.816864]  ? queue_io+0x329/0x510 
> [25069.817148]  wb_writeback+0x70a/0x9c0 
> [25069.817448]  ? __pfx_wb_writeback+0x10/0x10 
> [25069.817831]  ? get_nr_dirty_inodes+0xcb/0x180 
> [25069.818264]  wb_do_writeback+0x5d4/0x8e0 
> [25069.818575]  ? __pfx_wb_do_writeback+0x10/0x10 
> [25069.818975]  ? set_worker_desc+0x16e/0x190 
> [25069.819313]  ? __pfx_set_worker_desc+0x10/0x10 
> [25069.819669]  wb_workfn+0x7c/0x200 
> [25069.819997]  process_one_work+0xd8b/0x1320 
> [25069.820342]  ? __pfx_process_one_work+0x10/0x10 
> [25069.820694]  ? assign_work+0x16c/0x240 
> [25069.821053]  worker_thread+0x5f3/0xfe0 
> [25069.821364]  ? __pfx_worker_thread+0x10/0x10 
> [25069.821697]  kthread+0x3b4/0x770 
> [25069.822122]  ? kvm_sched_clock_read+0x11/0x20 
> [25069.822457]  ? local_clock_noinstr+0xd/0xe0 
> [25069.822790]  ? __pfx_kthread+0x10/0x10 
> [25069.823079]  ? __lock_release.isra.0+0x1a4/0x2c0 
> [25069.823430]  ? rcu_is_watching+0x15/0xb0 
> [25069.823729]  ? __pfx_kthread+0x10/0x10 
> [25069.824035]  ret_from_fork+0x393/0x480 
> [25069.864794]  ? __pfx_kthread+0x10/0x10 
> [25069.865076]  ? __pfx_kthread+0x10/0x10 
> [25069.865371]  ret_from_fork_asm+0x1a/0x30 
> [25069.879905]  </TASK> 
> [25069.880145]  
> [25069.880145] Showing all locks held in the system: 
> [25069.889550] 1 lock held by khungtaskd/37: 
> [25069.889886]  #0: ffffffffa29309e0 (rcu_read_lock){....}-{1:3}, at: rcu_lock_acquire.constprop.0+0x7/0x30 
> [25069.890603] 2 locks held by 751/825733: 
> [25069.890942]  #0: ffff888101e36440 (sb_writers#16){.+.+}-{0:0}, at: ksys_write+0xf9/0x1d0 
> [25069.891589]  #1: ffffffffa2b97fd0 (split_debug_mutex){+.+.}-{4:4}, at: split_huge_pages_write+0x124/0x430 
> [25069.892449] 3 locks held by kworker/u9:73/825900: 
> [25069.892818]  #0: ffff8881029df958 ((wq_completion)writeback){+.+.}-{0:0}, at: process_one_work+0x7f5/0x1320 
> [25069.893530]  #1: ffffc9000ad9fd10 ((work_completion)(&(&wb->dwork)->work)){+.+.}-{0:0}, at: process_one_work+0xd3f/0x1320 
> [25069.894333]  #2: ffff88810b1480e8 (&type->s_umount_key#68){++++}-{4:4}, at: super_trylock_shared+0x1c/0xa0 
> [25069.895056] 2 locks held by kworker/u10:2/826054: 
> [25069.895418]  #0: ffff888103ec2158 ((wq_completion)nfsiod){+.+.}-{0:0}, at: process_one_work+0x7f5/0x1320 
> [25069.896123]  #1: ffffc90001e1fd10 ((work_completion)(&ctx->work)){+.+.}-{0:0}, at: process_one_work+0xd3f/0x1320 
> [25069.896952]  
> [25069.897085] ============================================= 
> [25069.897085]  
> [-- MARK -- Tue Oct 21 03:45:00 2025] 
> [25192.380157] INFO: task kworker/u9:73:825900 blocked for more than 245 seconds. 
> [25192.380707]       Not tainted 6.18.0-rc2 #1 
> [25192.381042] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. 
> [25192.381605] task:kworker/u9:73   state:D stack:24712 pid:825900 tgid:825900 ppid:2      task_flags:0x4208060 flags:0x00080000 
> [25192.382516] Workqueue: writeback wb_workfn (flush-0:48) 
> [25192.382914] Call Trace: 
> [25192.383126]  <TASK> 
> [25192.383304]  __schedule+0x838/0x1890 
> [25192.383586]  ? __pfx___schedule+0x10/0x10 
> [25192.383888]  ? __blk_flush_plug+0x27b/0x4d0 
> [25192.384228]  ? find_held_lock+0x32/0x90 
> [25192.384527]  ? __lock_release.isra.0+0x1a4/0x2c0 
> [25192.384880]  schedule+0xd4/0x260 
> [25192.385150]  io_schedule+0x8f/0xf0 
> [25192.385416]  folio_wait_bit_common+0x2d9/0x780 
> [25192.385816]  ? folio_wait_bit_common+0x1dd/0x780 
> [25192.386196]  ? __pfx_folio_wait_bit_common+0x10/0x10 
> [25192.386572]  ? nfs_page_clear_headlock+0x31/0x80 [nfs] 
> [25192.387045]  ? __pfx_wake_page_function+0x10/0x10 
> [25192.387400]  ? __pfx___might_resched+0x10/0x10 
> [25192.387742]  writeback_get_folio+0x3f9/0x500 
> [25192.388092]  writeback_iter+0x136/0x720 
> [25192.388389]  nfs_writepages+0x4f8/0x9b0 [nfs] 
> [25192.388788]  ? mark_held_locks+0x40/0x70 
> [25192.389102]  ? __pfx_nfs_writepages+0x10/0x10 [nfs] 
> [25192.389530]  ? virtqueue_notify+0x68/0xc0 
> [25192.389831]  ? virtio_queue_rq+0x2b1/0x650 [virtio_blk] 
> [25192.390255]  ? __lock_acquire+0x57c/0xbd0 
> [25192.390574]  do_writepages+0x21f/0x560 
> [25192.390861]  ? __pfx_do_writepages+0x10/0x10 
> [25192.391200]  ? rcu_is_watching+0x15/0xb0 
> [25192.391502]  __writeback_single_inode+0xe2/0x5f0 
> [25192.391848]  ? __lock_release.isra.0+0x1a4/0x2c0 
> [25192.392210]  ? __pfx___writeback_single_inode+0x10/0x10 
> [25192.392596]  ? writeback_sb_inodes+0x416/0xd00 
> [25192.392935]  writeback_sb_inodes+0x535/0xd00 
> [25192.393286]  ? __pfx_stack_trace_save+0x10/0x10 
> [25192.393628]  ? local_clock_noinstr+0xd/0xe0 
> [25192.393940]  ? __pfx_writeback_sb_inodes+0x10/0x10 
> [25192.394310]  ? __lock_release.isra.0+0x1a4/0x2c0 
> [25192.394731]  ? lock_acquire+0x10b/0x150 
> [25192.395048]  ? down_read_trylock+0x4b/0x60 
> [25192.395360]  __writeback_inodes_wb+0xf4/0x270 
> [25192.395698]  ? __pfx___writeback_inodes_wb+0x10/0x10 
> [25192.396081]  ? queue_io+0x329/0x510 
> [25192.396355]  wb_writeback+0x70a/0x9c0 
> [25192.396645]  ? __pfx_wb_writeback+0x10/0x10 
> [25192.396963]  ? get_nr_dirty_inodes+0xcb/0x180 
> [25192.397314]  wb_do_writeback+0x5d4/0x8e0 
> [25192.397622]  ? __pfx_wb_do_writeback+0x10/0x10 
> [25192.397950]  ? set_worker_desc+0x16e/0x190 
> [25192.398272]  ? __pfx_set_worker_desc+0x10/0x10 
> [25192.398628]  wb_workfn+0x7c/0x200 
> [25192.398886]  process_one_work+0xd8b/0x1320 
> [25192.399222]  ? __pfx_process_one_work+0x10/0x10 
> [25192.399574]  ? assign_work+0x16c/0x240 
> [25192.399861]  worker_thread+0x5f3/0xfe0 
> [25192.400176]  ? __pfx_worker_thread+0x10/0x10 
> [25192.400497]  kthread+0x3b4/0x770 
> [25192.400747]  ? kvm_sched_clock_read+0x11/0x20 
> [25192.401089]  ? local_clock_noinstr+0xd/0xe0 
> [25192.401401]  ? __pfx_kthread+0x10/0x10 
> [25192.401684]  ? __lock_release.isra.0+0x1a4/0x2c0 
> [25192.402046]  ? rcu_is_watching+0x15/0xb0 
> [25192.402343]  ? __pfx_kthread+0x10/0x10 
> [25192.402629]  ret_from_fork+0x393/0x480 
> [25192.402910]  ? __pfx_kthread+0x10/0x10 
> [25192.403208]  ? __pfx_kthread+0x10/0x10 
> [25192.403493]  ret_from_fork_asm+0x1a/0x30 
> [25192.403808]  </TASK> 
> [25192.403994]  
> [25192.403994] Showing all locks held in the system: 
> [25192.404474] 1 lock held by khungtaskd/37: 
> [25192.404774]  #0: ffffffffa29309e0 (rcu_read_lock){....}-{1:3}, at: rcu_lock_acquire.constprop.0+0x7/0x30 
> [25192.405487] 2 locks held by 751/825733: 
> [25192.405779]  #0: ffff888101e36440 (sb_writers#16){.+.+}-{0:0}, at: ksys_write+0xf9/0x1d0 
> [25192.406385]  #1: ffffffffa2b97fd0 (split_debug_mutex){+.+.}-{4:4}, at: split_huge_pages_write+0x124/0x430 
> [25192.407092] 4 locks held by kworker/u10:40/825850: 
> [25192.407443]  #0: ffff888103ec5158 ((wq_completion)nfslocaliod){+.+.}-{0:0}, at: process_one_work+0x7f5/0x1320 
> [25192.408175]  #1: ffffc9000b0cfd10 ((work_completion)(&iocb->work)#2){+.+.}-{0:0}, at: process_one_work+0xd3f/0x1320 
> [25192.408924]  #2: ffff88815c2c2440 (sb_writers#13){++++}-{0:0}, at: process_one_work+0xd8b/0x1320 
> [25192.409579]  #3: ffff8881a1324b58 (&sb->s_type->i_mutex_key#13){++++}-{4:4}, at: xfs_ilock+0x360/0x460 [xfs] 
> [25192.410660] 3 locks held by kworker/u9:73/825900: 
> [25192.411026]  #0: ffff8881029df958 ((wq_completion)writeback){+.+.}-{0:0}, at: process_one_work+0x7f5/0x1320 
> [25192.411724]  #1: ffffc9000ad9fd10 ((work_completion)(&(&wb->dwork)->work)){+.+.}-{0:0}, at: process_one_work+0xd3f/0x1320 
> [25192.412518]  #2: ffff88810b1480e8 (&type->s_umount_key#68){++++}-{4:4}, at: super_trylock_shared+0x1c/0xa0 
> [25192.413237] 4 locks held by kworker/1:1/826103: 
> [25192.413576]  
> [25192.413706] ============================================= 
> [25192.413706]  
> [25315.259312] INFO: task kworker/u9:73:825900 blocked for more than 368 seconds. 
> [25315.259973]       Not tainted 6.18.0-rc2 #1 
> [25315.260324] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. 
> [25315.260894] task:kworker/u9:73   state:D stack:24712 pid:825900 tgid:825900 ppid:2      task_flags:0x4208060 flags:0x00080000 
> [25315.261752] Workqueue: writeback wb_workfn (flush-0:48) 
> [25315.262153] Call Trace: 
> [25315.262369]  <TASK> 
> [25315.262551]  __schedule+0x838/0x1890 
> [25315.262840]  ? __pfx___schedule+0x10/0x10 
> [25315.263150]  ? __blk_flush_plug+0x27b/0x4d0 
> [25315.263493]  ? find_held_lock+0x32/0x90 
> [25315.263794]  ? __lock_release.isra.0+0x1a4/0x2c0 
> [25315.264151]  schedule+0xd4/0x260 
> [25315.264425]  io_schedule+0x8f/0xf0 
> [25315.264694]  folio_wait_bit_common+0x2d9/0x780 
> [25315.265100]  ? folio_wait_bit_common+0x1dd/0x780 
> [25315.265473]  ? __pfx_folio_wait_bit_common+0x10/0x10 
> [25315.265856]  ? nfs_page_clear_headlock+0x31/0x80 [nfs] 
> [25315.266332]  ? __pfx_wake_page_function+0x10/0x10 
> [25315.266692]  ? __pfx___might_resched+0x10/0x10 
> [25315.267040]  writeback_get_folio+0x3f9/0x500 
> [25315.267393]  writeback_iter+0x136/0x720 
> [25315.267693]  nfs_writepages+0x4f8/0x9b0 [nfs] 
> [25315.268106]  ? mark_held_locks+0x40/0x70 
> [25315.268425]  ? __pfx_nfs_writepages+0x10/0x10 [nfs] 
> [25315.268861]  ? virtqueue_notify+0x68/0xc0 
> [25315.269168]  ? virtio_queue_rq+0x2b1/0x650 [virtio_blk] 
> [25315.269597]  ? __lock_acquire+0x57c/0xbd0 
> [25315.269921]  do_writepages+0x21f/0x560 
> [25315.270212]  ? __pfx_do_writepages+0x10/0x10 
> [25315.270555]  ? rcu_is_watching+0x15/0xb0 
> [25315.270864]  __writeback_single_inode+0xe2/0x5f0 
> [25315.271212]  ? __lock_release.isra.0+0x1a4/0x2c0 
> [25315.271584]  ? __pfx___writeback_single_inode+0x10/0x10 
> [25315.271975]  ? writeback_sb_inodes+0x416/0xd00 
> [25315.272336]  writeback_sb_inodes+0x535/0xd00 
> [25315.272675]  ? __pfx_stack_trace_save+0x10/0x10 
> [25315.273018]  ? local_clock_noinstr+0xd/0xe0 
> [25315.273354]  ? __pfx_writeback_sb_inodes+0x10/0x10 
> [25315.273713]  ? __lock_release.isra.0+0x1a4/0x2c0 
> [25315.274107]  ? lock_acquire+0x10b/0x150 
> [25315.274426]  ? down_read_trylock+0x4b/0x60 
> [25315.274743]  __writeback_inodes_wb+0xf4/0x270 
> [25315.275083]  ? __pfx___writeback_inodes_wb+0x10/0x10 
> [25315.275470]  ? queue_io+0x329/0x510 
> [25315.275749]  wb_writeback+0x70a/0x9c0 
> [25315.276046]  ? __pfx_wb_writeback+0x10/0x10 
> [25315.276389]  ? get_nr_dirty_inodes+0xcb/0x180 
> [25315.276729]  wb_do_writeback+0x5d4/0x8e0 
> [25315.277042]  ? __pfx_wb_do_writeback+0x10/0x10 
> [25315.277401]  ? set_worker_desc+0x16e/0x190 
> [25315.277713]  ? __pfx_set_worker_desc+0x10/0x10 
> [25315.278072]  wb_workfn+0x7c/0x200 
> [25315.278352]  process_one_work+0xd8b/0x1320 
> [25315.278678]  ? __pfx_process_one_work+0x10/0x10 
> [25315.279034]  ? assign_work+0x16c/0x240 
> [25315.279346]  worker_thread+0x5f3/0xfe0 
> [25315.279650]  ? __pfx_worker_thread+0x10/0x10 
> [25315.279980]  kthread+0x3b4/0x770 
> [25315.280229]  ? kvm_sched_clock_read+0x11/0x20 
> [25315.280577]  ? local_clock_noinstr+0xd/0xe0 
> [25315.280899]  ? __pfx_kthread+0x10/0x10 
> [25315.281183]  ? __lock_release.isra.0+0x1a4/0x2c0 
> [25315.281551]  ? rcu_is_watching+0x15/0xb0 
> [25315.281886]  ? __pfx_kthread+0x10/0x10 
> [25315.282176]  ret_from_fork+0x393/0x480 
> [25315.282480]  ? __pfx_kthread+0x10/0x10 
> [25315.282770]  ? __pfx_kthread+0x10/0x10 
> [25315.283058]  ret_from_fork_asm+0x1a/0x30 
> [25315.283401]  </TASK> 
> [25315.283585]  
> [25315.283585] Showing all locks held in the system: 
> [25315.284041] 1 lock held by khungtaskd/37: 
> [25315.284360]  #0: ffffffffa29309e0 (rcu_read_lock){....}-{1:3}, at: rcu_lock_acquire.constprop.0+0x7/0x30 
> [25315.285066] 2 locks held by 751/825733: 
> [25315.285370]  #0: ffff888101e36440 (sb_writers#16){.+.+}-{0:0}, at: ksys_write+0xf9/0x1d0 
> [25315.285974]  #1: ffffffffa2b97fd0 (split_debug_mutex){+.+.}-{4:4}, at: split_huge_pages_write+0x124/0x430 
> [25315.286683] 3 locks held by kworker/u9:73/825900: 
> [25315.287039]  #0: ffff8881029df958 ((wq_completion)writeback){+.+.}-{0:0}, at: process_one_work+0x7f5/0x1320 
> [25315.287744]  #1: ffffc9000ad9fd10 ((work_completion)(&(&wb->dwork)->work)){+.+.}-{0:0}, at: process_one_work+0xd3f/0x1320 
> [25315.288534]  #2: ffff88810b1480e8 (&type->s_umount_key#68){++++}-{4:4}, at: super_trylock_shared+0x1c/0xa0 
> [25315.289261] 2 locks held by kworker/u10:2/826054: 
> [25315.289612]  #0: ffff888103ec2158 ((wq_completion)nfsiod){+.+.}-{0:0}, at: process_one_work+0x7f5/0x1320 
> [25315.290324]  #1: ffffc90001e1fd10 ((work_completion)(&ctx->work)){+.+.}-{0:0}, at: process_one_work+0xd3f/0x1320 
> [25315.291064]  
> [25315.291199] ============================================= 
> [25315.291199]  
> [25438.138585] INFO: task kworker/u9:73:825900 blocked for more than 491 seconds. 
> [25438.139167]       Not tainted 6.18.0-rc2 #1 
> [25438.139486] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. 
> [25438.140085] task:kworker/u9:73   state:D stack:24712 pid:825900 tgid:825900 ppid:2      task_flags:0x4208060 flags:0x00080000 
> [25438.140963] Workqueue: writeback wb_workfn (flush-0:48) 
> [25438.141360] Call Trace: 
> [25438.141576]  <TASK> 
> [25438.141757]  __schedule+0x838/0x1890 
> [25438.142043]  ? __pfx___schedule+0x10/0x10 
> [25438.142347]  ? __blk_flush_plug+0x27b/0x4d0 
> [25438.142689]  ? find_held_lock+0x32/0x90 
> [25438.142987]  ? __lock_release.isra.0+0x1a4/0x2c0 
> [25438.143349]  schedule+0xd4/0x260 
> [25438.143622]  io_schedule+0x8f/0xf0 
> [25438.143891]  folio_wait_bit_common+0x2d9/0x780 
> [25438.144263]  ? folio_wait_bit_common+0x1dd/0x780 
> [25438.144639]  ? __pfx_folio_wait_bit_common+0x10/0x10 
> [25438.145021]  ? nfs_page_clear_headlock+0x31/0x80 [nfs] 
> [25438.145479]  ? __pfx_wake_page_function+0x10/0x10 
> [25438.145854]  ? __pfx___might_resched+0x10/0x10 
> [25438.146205]  writeback_get_folio+0x3f9/0x500 
> [25438.146561]  writeback_iter+0x136/0x720 
> [25438.146860]  nfs_writepages+0x4f8/0x9b0 [nfs] 
> [25438.147265]  ? mark_held_locks+0x40/0x70 
> [25438.147582]  ? __pfx_nfs_writepages+0x10/0x10 [nfs] 
> [25438.148015]  ? virtqueue_notify+0x68/0xc0 
> [25438.148319]  ? virtio_queue_rq+0x2b1/0x650 [virtio_blk] 
> [25438.148749]  ? __lock_acquire+0x57c/0xbd0 
> [25438.149072]  do_writepages+0x21f/0x560 
> [25438.149364]  ? __pfx_do_writepages+0x10/0x10 
> [25438.149705]  ? rcu_is_watching+0x15/0xb0 
> [25438.150014]  __writeback_single_inode+0xe2/0x5f0 
> [25438.150362]  ? __lock_release.isra.0+0x1a4/0x2c0 
> [25438.150731]  ? __pfx___writeback_single_inode+0x10/0x10 
> [25438.151123]  ? writeback_sb_inodes+0x416/0xd00 
> [25438.151466]  writeback_sb_inodes+0x535/0xd00 
> [25438.151825]  ? __pfx_stack_trace_save+0x10/0x10 
> [25438.152171]  ? local_clock_noinstr+0xd/0xe0 
> [25438.152504]  ? __pfx_writeback_sb_inodes+0x10/0x10 
> [25438.152864]  ? __lock_release.isra.0+0x1a4/0x2c0 
> [25438.153262]  ? lock_acquire+0x10b/0x150 
> [25438.153577]  ? down_read_trylock+0x4b/0x60 
> [25438.153894]  __writeback_inodes_wb+0xf4/0x270 
> [25438.154231]  ? __pfx___writeback_inodes_wb+0x10/0x10 
> [25438.154622]  ? queue_io+0x329/0x510 
> [25438.154899]  wb_writeback+0x70a/0x9c0 
> [25438.155193]  ? __pfx_wb_writeback+0x10/0x10 
> [25438.155536]  ? get_nr_dirty_inodes+0xcb/0x180 
> [25438.155877]  wb_do_writeback+0x5d4/0x8e0 
> [25438.156190]  ? __pfx_wb_do_writeback+0x10/0x10 
> [25438.156554]  ? set_worker_desc+0x16e/0x190 
> [25438.156865]  ? __pfx_set_worker_desc+0x10/0x10 
> [25438.157225]  wb_workfn+0x7c/0x200 
> [25438.157486]  process_one_work+0xd8b/0x1320 
> [25438.157830]  ? __pfx_process_one_work+0x10/0x10 
> [25438.158185]  ? assign_work+0x16c/0x240 
> [25438.158478]  worker_thread+0x5f3/0xfe0 
> [25438.158800]  ? __pfx_worker_thread+0x10/0x10 
> [25438.159129]  kthread+0x3b4/0x770 
> [25438.159382]  ? kvm_sched_clock_read+0x11/0x20 
> [25438.159730]  ? local_clock_noinstr+0xd/0xe0 
> [25438.160049]  ? __pfx_kthread+0x10/0x10 
> [25438.160335]  ? __lock_release.isra.0+0x1a4/0x2c0 
> [25438.160703]  ? rcu_is_watching+0x15/0xb0 
> [25438.161010]  ? __pfx_kthread+0x10/0x10 
> [25438.161303]  ret_from_fork+0x393/0x480 
> [25438.161605]  ? __pfx_kthread+0x10/0x10 
> [25438.161892]  ? __pfx_kthread+0x10/0x10 
> [25438.162184]  ret_from_fork_asm+0x1a/0x30 
> [25438.162521]  </TASK> 
> [25438.162705]  
> [25438.162705] Showing all locks held in the system: 
> [25438.163161] 1 lock held by khungtaskd/37: 
> [25438.163461]  #0: ffffffffa29309e0 (rcu_read_lock){....}-{1:3}, at: rcu_lock_acquire.constprop.0+0x7/0x30 
> [25438.164190] 2 locks held by 751/825733: 
> [25438.164483]  #0: ffff888101e36440 (sb_writers#16){.+.+}-{0:0}, at: ksys_write+0xf9/0x1d0 
> [25438.165101]  #1: ffffffffa2b97fd0 (split_debug_mutex){+.+.}-{4:4}, at: split_huge_pages_write+0x124/0x430 
> [25438.165820] 3 locks held by kworker/u9:73/825900: 
> [25438.166177]  #0: ffff8881029df958 ((wq_completion)writeback){+.+.}-{0:0}, at: process_one_work+0x7f5/0x1320 
> [25438.166895]  #1: ffffc9000ad9fd10 ((work_completion)(&(&wb->dwork)->work)){+.+.}-{0:0}, at: process_one_work+0xd3f/0x1320 
> [25438.167703]  #2: ffff88810b1480e8 (&type->s_umount_key#68){++++}-{4:4}, at: super_trylock_shared+0x1c/0xa0 
> [25438.168412] 2 locks held by kworker/u10:2/826054: 
> [25438.168783]  #0: ffff888103ec2158 ((wq_completion)nfsiod){+.+.}-{0:0}, at: process_one_work+0x7f5/0x1320 
> [25438.169473]  #1: ffffc90001e1fd10 ((work_completion)(&ctx->work)){+.+.}-{0:0}, at: process_one_work+0xd3f/0x1320 
> [25438.170230]  
> [25438.170364] ============================================= 
> [25438.170364]  
> [-- MARK -- Tue Oct 21 03:50:00 2025] 
> [25561.017812] INFO: task kworker/u9:73:825900 blocked for more than 614 seconds. 
> [25561.018375]       Not tainted 6.18.0-rc2 #1 
> [25561.018694] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. 
> [25561.019310] task:kworker/u9:73   state:D stack:24712 pid:825900 tgid:825900 ppid:2      task_flags:0x4208060 flags:0x00080000 
> [25561.020172] Workqueue: writeback wb_workfn (flush-0:48) 
> [25561.020573] Call Trace: 
> [25561.020786]  <TASK> 
> [25561.020966]  __schedule+0x838/0x1890 
> [25561.021254]  ? __pfx___schedule+0x10/0x10 
> [25561.021560]  ? __blk_flush_plug+0x27b/0x4d0 
> [25561.021906]  ? find_held_lock+0x32/0x90 
> [25561.022200]  ? __lock_release.isra.0+0x1a4/0x2c0 
> [25561.022560]  schedule+0xd4/0x260 
> [25561.022833]  io_schedule+0x8f/0xf0 
> [25561.023102]  folio_wait_bit_common+0x2d9/0x780 
> [25561.023475]  ? folio_wait_bit_common+0x1dd/0x780 
> [25561.023844]  ? __pfx_folio_wait_bit_common+0x10/0x10 
> [25561.024216]  ? nfs_page_clear_headlock+0x31/0x80 [nfs] 
> [25561.024686]  ? __pfx_wake_page_function+0x10/0x10 
> [25561.025050]  ? __pfx___might_resched+0x10/0x10 
> [25561.025389]  writeback_get_folio+0x3f9/0x500 
> [25561.025711]  writeback_iter+0x136/0x720 
> [25561.026014]  nfs_writepages+0x4f8/0x9b0 [nfs] 
> [25561.026402]  ? mark_held_locks+0x40/0x70 
> [25561.026698]  ? __pfx_nfs_writepages+0x10/0x10 [nfs] 
> [25561.027155]  ? virtqueue_notify+0x68/0xc0 
> [25561.027462]  ? virtio_queue_rq+0x2b1/0x650 [virtio_blk] 
> [25561.027894]  ? __lock_acquire+0x57c/0xbd0 
> [25561.028213]  do_writepages+0x21f/0x560 
> [25561.028507]  ? __pfx_do_writepages+0x10/0x10 
> [25561.028849]  ? rcu_is_watching+0x15/0xb0 
> [25561.029154]  __writeback_single_inode+0xe2/0x5f0 
> [25561.029608]  ? __lock_release.isra.0+0x1a4/0x2c0 
> [25561.029982]  ? __pfx___writeback_single_inode+0x10/0x10 
> [25561.030374]  ? writeback_sb_inodes+0x416/0xd00 
> [25561.030715]  writeback_sb_inodes+0x535/0xd00 
> [25561.031072]  ? __pfx_stack_trace_save+0x10/0x10 
> [25561.031420]  ? local_clock_noinstr+0xd/0xe0 
> [25561.031761]  ? __pfx_writeback_sb_inodes+0x10/0x10 
> [25561.032121]  ? __lock_release.isra.0+0x1a4/0x2c0 
> [25561.032516]  ? lock_acquire+0x10b/0x150 
> [25561.032832]  ? down_read_trylock+0x4b/0x60 
> [25561.033149]  __writeback_inodes_wb+0xf4/0x270 
> [25561.033487]  ? __pfx___writeback_inodes_wb+0x10/0x10 
> [25561.033881]  ? queue_io+0x329/0x510 
> [25561.034157]  wb_writeback+0x70a/0x9c0 
> [25561.034452]  ? __pfx_wb_writeback+0x10/0x10 
> [25561.034794]  ? get_nr_dirty_inodes+0xcb/0x180 
> [25561.035134]  wb_do_writeback+0x5d4/0x8e0 
> [25561.035443]  ? __pfx_wb_do_writeback+0x10/0x10 
> [25561.035796]  ? set_worker_desc+0x16e/0x190 
> [25561.036105]  ? __pfx_set_worker_desc+0x10/0x10 
> [25561.036463]  wb_workfn+0x7c/0x200 
> [25561.036724]  process_one_work+0xd8b/0x1320 
> [25561.037065]  ? __pfx_process_one_work+0x10/0x10 
> [25561.037422]  ? assign_work+0x16c/0x240 
> [25561.037718]  worker_thread+0x5f3/0xfe0 
> [25561.038035]  ? __pfx_worker_thread+0x10/0x10 
> [25561.038355]  kthread+0x3b4/0x770 
> [25561.038603]  ? kvm_sched_clock_read+0x11/0x20 
> [25561.038951]  ? local_clock_noinstr+0xd/0xe0 
> [25561.039272]  ? __pfx_kthread+0x10/0x10 
> [25561.039555]  ? __lock_release.isra.0+0x1a4/0x2c0 
> [25561.039921]  ? rcu_is_watching+0x15/0xb0 
> [25561.040221]  ? __pfx_kthread+0x10/0x10 
> [25561.040511]  ret_from_fork+0x393/0x480 
> [25561.040813]  ? __pfx_kthread+0x10/0x10 
> [25561.041101]  ? __pfx_kthread+0x10/0x10 
> [25561.041392]  ret_from_fork_asm+0x1a/0x30 
> [25561.041707]  </TASK> 
> [25561.041908]  
> [25561.041908] Showing all locks held in the system: 
> [25561.042365] 1 lock held by khungtaskd/37: 
> [25561.042662]  #0: ffffffffa29309e0 (rcu_read_lock){....}-{1:3}, at: rcu_lock_acquire.constprop.0+0x7/0x30 
> [25561.043384] 2 locks held by 751/825733: 
> [25561.043677]  #0: ffff888101e36440 (sb_writers#16){.+.+}-{0:0}, at: ksys_write+0xf9/0x1d0 
> [25561.044296]  #1: ffffffffa2b97fd0 (split_debug_mutex){+.+.}-{4:4}, at: split_huge_pages_write+0x124/0x430 
> [25561.045009] 3 locks held by kworker/u9:73/825900: 
> [25561.045362]  #0: ffff8881029df958 ((wq_completion)writeback){+.+.}-{0:0}, at: process_one_work+0x7f5/0x1320 
> [25561.046085]  #1: ffffc9000ad9fd10 ((work_completion)(&(&wb->dwork)->work)){+.+.}-{0:0}, at: process_one_work+0xd3f/0x1320 
> [25561.046895]  #2: ffff88810b1480e8 (&type->s_umount_key#68){++++}-{4:4}, at: super_trylock_shared+0x1c/0xa0 
> [25561.047600] 2 locks held by kworker/u10:13/826064: 
> [25561.047970]  #0: ffff888103ec2158 ((wq_completion)nfsiod){+.+.}-{0:0}, at: process_one_work+0x7f5/0x1320 
> [25561.048657]  #1: ffffc90001e9fd10 ((work_completion)(&ctx->work)){+.+.}-{0:0}, at: process_one_work+0xd3f/0x1320 
> [25561.049410]  
> [25561.049538] ============================================= 
> [25561.049538]  
> [25683.897277] INFO: task kworker/u9:73:825900 blocked for more than 737 seconds. 
> [25683.897831]       Not tainted 6.18.0-rc2 #1 
> [25683.898182] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. 
> [25683.898754] task:kworker/u9:73   state:D stack:24712 pid:825900 tgid:825900 ppid:2      task_flags:0x4208060 flags:0x00080000 
> [25683.899587] Workqueue: writeback wb_workfn (flush-0:48) 
> [25683.900001] Call Trace: 
> [25683.900200]  <TASK> 
> [25683.900379]  __schedule+0x838/0x1890 
> [25683.900663]  ? __pfx___schedule+0x10/0x10 
> [25683.900967]  ? __blk_flush_plug+0x27b/0x4d0 
> [25683.901308]  ? find_held_lock+0x32/0x90 
> [25683.901609]  ? __lock_release.isra.0+0x1a4/0x2c0 
> [25683.901964]  schedule+0xd4/0x260 
> [25683.902235]  io_schedule+0x8f/0xf0 
> [25683.902507]  folio_wait_bit_common+0x2d9/0x780 
> [25683.902874]  ? folio_wait_bit_common+0x1dd/0x780 
> [25683.903256]  ? __pfx_folio_wait_bit_common+0x10/0x10 
> [25683.903636]  ? nfs_page_clear_headlock+0x31/0x80 [nfs] 
> [25683.904111]  ? __pfx_wake_page_function+0x10/0x10 
> [25683.904469]  ? __pfx___might_resched+0x10/0x10 
> [25683.904819]  writeback_get_folio+0x3f9/0x500 
> [25683.905170]  writeback_iter+0x136/0x720 
> [25683.905469]  nfs_writepages+0x4f8/0x9b0 [nfs] 
> [25683.905878]  ? mark_held_locks+0x40/0x70 
> [25683.906194]  ? __pfx_nfs_writepages+0x10/0x10 [nfs] 
> [25683.906631]  ? virtqueue_notify+0x68/0xc0 
> [25683.906936]  ? virtio_queue_rq+0x2b1/0x650 [virtio_blk] 
> [25683.907368]  ? __lock_acquire+0x57c/0xbd0 
> [25683.907691]  do_writepages+0x21f/0x560 
> [25683.907999]  ? __pfx_do_writepages+0x10/0x10 
> [25683.908325]  ? rcu_is_watching+0x15/0xb0 
> [25683.908635]  __writeback_single_inode+0xe2/0x5f0 
> [25683.909013]  ? __lock_release.isra.0+0x1a4/0x2c0 
> [25683.909368]  ? __pfx___writeback_single_inode+0x10/0x10 
> [25683.909756]  ? writeback_sb_inodes+0x416/0xd00 
> [25683.910118]  writeback_sb_inodes+0x535/0xd00 
> [25683.910459]  ? __pfx_stack_trace_save+0x10/0x10 
> [25683.910807]  ? local_clock_noinstr+0xd/0xe0 
> [25683.911138]  ? __pfx_writeback_sb_inodes+0x10/0x10 
> [25683.911500]  ? __lock_release.isra.0+0x1a4/0x2c0 
> [25683.911896]  ? lock_acquire+0x10b/0x150 
> [25683.912211]  ? down_read_trylock+0x4b/0x60 
> [25683.912532]  __writeback_inodes_wb+0xf4/0x270 
> [25683.912869]  ? __pfx___writeback_inodes_wb+0x10/0x10 
> [25683.913259]  ? queue_io+0x329/0x510 
> [25683.913540]  wb_writeback+0x70a/0x9c0 
> [25683.913832]  ? __pfx_wb_writeback+0x10/0x10 
> [25683.914174]  ? get_nr_dirty_inodes+0xcb/0x180 
> [25683.914516]  wb_do_writeback+0x5d4/0x8e0 
> [25683.914831]  ? __pfx_wb_do_writeback+0x10/0x10 
> [25683.915185]  ? set_worker_desc+0x16e/0x190 
> [25683.915501]  ? __pfx_set_worker_desc+0x10/0x10 
> [25683.915860]  wb_workfn+0x7c/0x200 
> [25683.916139]  process_one_work+0xd8b/0x1320 
> [25683.916465]  ? __pfx_process_one_work+0x10/0x10 
> [25683.916822]  ? assign_work+0x16c/0x240 
> [25683.917133]  worker_thread+0x5f3/0xfe0 
> [25683.917437]  ? __pfx_worker_thread+0x10/0x10 
> [25683.917767]  kthread+0x3b4/0x770 
> [25683.918037]  ? kvm_sched_clock_read+0x11/0x20 
> [25683.918368]  ? local_clock_noinstr+0xd/0xe0 
> [25683.918687]  ? __pfx_kthread+0x10/0x10 
> [25683.918969]  ? __lock_release.isra.0+0x1a4/0x2c0 
> [25683.919334]  ? rcu_is_watching+0x15/0xb0 
> [25683.919640]  ? __pfx_kthread+0x10/0x10 
> [25683.919929]  ret_from_fork+0x393/0x480 
> [25683.920234]  ? __pfx_kthread+0x10/0x10 
> [25683.920523]  ? __pfx_kthread+0x10/0x10 
> [25683.920815]  ret_from_fork_asm+0x1a/0x30 
> [25683.921152]  </TASK> 
> [25683.921333] Future hung task reports are suppressed, see sysctl kernel.hung_task_warnings 
> [25683.921921]  
> [25683.921921] Showing all locks held in the system: 
> [25683.922392] 1 lock held by khungtaskd/37: 
> [25683.922698]  #0: ffffffffa29309e0 (rcu_read_lock){....}-{1:3}, at: rcu_lock_acquire.constprop.0+0x7/0x30 
> [25683.923423] 2 locks held by 751/825733: 
> [25683.923720]  #0: ffff888101e36440 (sb_writers#16){.+.+}-{0:0}, at: ksys_write+0xf9/0x1d0 
> [25683.924331]  #1: ffffffffa2b97fd0 (split_debug_mutex){+.+.}-{4:4}, at: split_huge_pages_write+0x124/0x430 
> [25683.925045] 3 locks held by kworker/u9:73/825900: 
> [25683.925398]  #0: ffff8881029df958 ((wq_completion)writeback){+.+.}-{0:0}, at: process_one_work+0x7f5/0x1320 
> [25683.926121]  #1: ffffc9000ad9fd10 ((work_completion)(&(&wb->dwork)->work)){+.+.}-{0:0}, at: process_one_work+0xd3f/0x1320 
> [25683.926918]  #2: ffff88810b1480e8 (&type->s_umount_key#68){++++}-{4:4}, at: super_trylock_shared+0x1c/0xa0 
> [25683.927641] 4 locks held by kworker/u9:143/826036: 
> [25683.928022]  #0: ffff888103ec5158 ((wq_completion)nfslocaliod){+.+.}-{0:0}, at: process_one_work+0x7f5/0x1320 
> [25683.928743]  #1: ffffc90001cc7d10 ((work_completion)(&iocb->work)#2){+.+.}-{0:0}, at: process_one_work+0xd3f/0x1320 
> [25683.929514]  #2: ffff88815c2c2440 (sb_writers#13){++++}-{0:0}, at: process_one_work+0xd8b/0x1320 
> [25683.930175]  #3: ffff8881a1324b58 (&sb->s_type->i_mutex_key#13){++++}-{4:4}, at: xfs_ilock+0x360/0x460 [xfs] 
> [25683.931255]  
> [25683.931391] ============================================= 
> [25683.931391]  
> [-- MARK -- Tue Oct 21 03:55:00 2025] 
> [-- MARK -- Tue Oct 21 04:00:00 2025] 
> [-- MARK -- Tue Oct 21 04:05:00 2025] 
> 
> 
> [2]
> [12082.409406] run fstests generic/751 at 2025-10-20 19:04:12 
> [-- MARK -- Mon Oct 20 23:05:00 2025] 
> [-- MARK -- Mon Oct 20 23:10:00 2025] 
> [-- MARK -- Mon Oct 20 23:15:00 2025] 
> [13023.321998] INFO: task kworker/u17:206:858774 blocked for more than 122 seconds. 
> [13023.322058]       Tainted: G    B               6.18.0-rc2 #1 
> [13023.322062] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. 
> [13023.322064] task:kworker/u17:206 state:D stack:23184 pid:858774 tgid:858774 ppid:2      task_flags:0x4208060 flags:0x00000210 
> [13023.322077] Workqueue: writeback wb_workfn (flush-0:56) 
> [13023.322089] Call trace: 
> [13023.322091]  __switch_to+0x1e0/0x488 (T) 
> [13023.322098]  __schedule+0x748/0x1430 
> [13023.322103]  schedule+0xd0/0x240 
> [13023.322107]  io_schedule+0xb4/0x120 
> [13023.322112]  folio_wait_bit_common+0x2c4/0x708 
> [13023.322117]  __folio_lock+0x24/0x38 
> [13023.322120]  writeback_get_folio+0x37c/0x480 
> [13023.322125]  writeback_iter+0x128/0x6d0 
> [13023.322130]  nfs_writepages+0x504/0x930 [nfs] 
> [13023.322194]  do_writepages+0x204/0x4c0 
> [13023.322198]  __writeback_single_inode+0xec/0x4f8 
> [13023.322204]  writeback_sb_inodes+0x4a4/0xab8 
> [13023.322296]  __writeback_inodes_wb+0x104/0x260 
> [13023.322301]  wb_writeback+0x840/0xce0 
> [13023.322305]  wb_do_writeback+0x69c/0x940 
> [13023.322308]  wb_workfn+0x80/0x1c0 
> [13023.322312]  process_one_work+0x774/0x12d0 
> [13023.322318]  worker_thread+0x434/0xca0 
> [13023.322322]  kthread+0x2ec/0x390 
> [13023.322326]  ret_from_fork+0x10/0x20 
> [13023.322409] INFO: lockdep is turned off. 
> [-- MARK -- Mon Oct 20 23:20:00 2025] 
> [13146.212082] INFO: task kworker/u17:206:858774 blocked for more than 245 seconds. 
> [13146.212109]       Tainted: G    B               6.18.0-rc2 #1 
> [13146.212120] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. 
> [13146.212125] task:kworker/u17:206 state:D stack:23184 pid:858774 tgid:858774 ppid:2      task_flags:0x4208060 flags:0x00000210 
> [13146.212141] Workqueue: writeback wb_workfn (flush-0:56) 
> [13146.212160] Call trace: 
> [13146.212165]  __switch_to+0x1e0/0x488 (T) 
> [13146.212175]  __schedule+0x748/0x1430 
> [13146.212182]  schedule+0xd0/0x240 
> [13146.212189]  io_schedule+0xb4/0x120 
> [13146.212196]  folio_wait_bit_common+0x2c4/0x708 
> [13146.212203]  __folio_lock+0x24/0x38 
> [13146.212209]  writeback_get_folio+0x37c/0x480 
> [13146.212217]  writeback_iter+0x128/0x6d0 
> [13146.212222]  nfs_writepages+0x504/0x930 [nfs] 
> [13146.212287]  do_writepages+0x204/0x4c0 
> [13146.212291]  __writeback_single_inode+0xec/0x4f8 
> [13146.212294]  writeback_sb_inodes+0x4a4/0xab8 
> [13146.212298]  __writeback_inodes_wb+0x104/0x260 
> [13146.212302]  wb_writeback+0x840/0xce0 
> [13146.212306]  wb_do_writeback+0x69c/0x940 
> [13146.212309]  wb_workfn+0x80/0x1c0 
> [13146.212313]  process_one_work+0x774/0x12d0 
> [13146.212318]  worker_thread+0x434/0xca0 
> [13146.212321]  kthread+0x2ec/0x390 
> [13146.212325]  ret_from_fork+0x10/0x20 
> [13146.212360] INFO: lockdep is turned off. 
> [13269.082400] INFO: task kworker/u17:206:858774 blocked for more than 368 seconds. 
> [13269.082553]       Tainted: G    B               6.18.0-rc2 #1 
> [13269.082557] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. 
> [13269.082559] task:kworker/u17:206 state:D stack:23184 pid:858774 tgid:858774 ppid:2      task_flags:0x4208060 flags:0x00000210 
> [13269.082570] Workqueue: writeback wb_workfn (flush-0:56) 
> [13269.082581] Call trace: 
> [13269.082583]  __switch_to+0x1e0/0x488 (T) 
> [13269.082591]  __schedule+0x748/0x1430 
> [13269.082596]  schedule+0xd0/0x240 
> [13269.082600]  io_schedule+0xb4/0x120 
> [13269.082604]  folio_wait_bit_common+0x2c4/0x708 
> [13269.082610]  __folio_lock+0x24/0x38 
> [13269.082613]  writeback_get_folio+0x37c/0x480 
> [13269.082618]  writeback_iter+0x128/0x6d0 
> [13269.082622]  nfs_writepages+0x504/0x930 [nfs] 
> [13269.082749]  do_writepages+0x204/0x4c0 
> [13269.082754]  __writeback_single_inode+0xec/0x4f8 
> [13269.082758]  writeback_sb_inodes+0x4a4/0xab8 
> [13269.082762]  __writeback_inodes_wb+0x104/0x260 
> [13269.082766]  wb_writeback+0x840/0xce0 
> [13269.082769]  wb_do_writeback+0x69c/0x940 
> [13269.082773]  wb_workfn+0x80/0x1c0 
> [13269.082777]  process_one_work+0x774/0x12d0 
> [13269.082782]  worker_thread+0x434/0xca0 
> [13269.082785]  kthread+0x2ec/0x390 
> [13269.082789]  ret_from_fork+0x10/0x20 
> [13269.082805] INFO: lockdep is turned off. 
> [-- MARK -- Mon Oct 20 23:25:00 2025] 
> [13391.972505] INFO: task kworker/u17:206:858774 blocked for more than 491 seconds. 
> [13391.972535]       Tainted: G    B               6.18.0-rc2 #1 
> [13391.972549] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. 
> [13391.972555] task:kworker/u17:206 state:D stack:23184 pid:858774 tgid:858774 ppid:2      task_flags:0x4208060 flags:0x00000210 
> [13391.972571] Workqueue: writeback wb_workfn (flush-0:56) 
> [13391.972585] Call trace: 
> [13391.972587]  __switch_to+0x1e0/0x488 (T) 
> [13391.972595]  __schedule+0x748/0x1430 
> [13391.972600]  schedule+0xd0/0x240 
> [13391.972605]  io_schedule+0xb4/0x120 
> [13391.972609]  folio_wait_bit_common+0x2c4/0x708 
> [13391.972614]  __folio_lock+0x24/0x38 
> [13391.972618]  writeback_get_folio+0x37c/0x480 
> [13391.972624]  writeback_iter+0x128/0x6d0 
> [13391.972628]  nfs_writepages+0x504/0x930 [nfs] 
> [13391.972700]  do_writepages+0x204/0x4c0 
> [13391.972704]  __writeback_single_inode+0xec/0x4f8 
> [13391.972708]  writeback_sb_inodes+0x4a4/0xab8 
> [13391.972712]  __writeback_inodes_wb+0x104/0x260 
> [13391.972716]  wb_writeback+0x840/0xce0 
> [13391.972720]  wb_do_writeback+0x69c/0x940 
> [13391.972724]  wb_workfn+0x80/0x1c0 
> [13391.972727]  process_one_work+0x774/0x12d0 
> [13391.972733]  worker_thread+0x434/0xca0 
> [13391.972736]  kthread+0x2ec/0x390 
> [13391.972740]  ret_from_fork+0x10/0x20 
> [13391.972761] INFO: lockdep is turned off. 
> [13514.852712] INFO: task kworker/u17:206:858774 blocked for more than 614 seconds. 
> [13514.852737]       Tainted: G    B               6.18.0-rc2 #1 
> [13514.852749] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. 
> [13514.852756] task:kworker/u17:206 state:D stack:23184 pid:858774 tgid:858774 ppid:2      task_flags:0x4208060 flags:0x00000210 
> [13514.852768] Workqueue: writeback wb_workfn (flush-0:56) 
> [13514.852779] Call trace: 
> [13514.852781]  __switch_to+0x1e0/0x488 (T) 
> [13514.852790]  __schedule+0x748/0x1430 
> [13514.852795]  schedule+0xd0/0x240 
> [13514.852808]  io_schedule+0xb4/0x120 
> [13514.852812]  folio_wait_bit_common+0x2c4/0x708 
> [13514.852818]  __folio_lock+0x24/0x38 
> [13514.852822]  writeback_get_folio+0x37c/0x480 
> [13514.852827]  writeback_iter+0x128/0x6d0 
> [13514.852831]  nfs_writepages+0x504/0x930 [nfs] 
> [13514.852897]  do_writepages+0x204/0x4c0 
> [13514.852901]  __writeback_single_inode+0xec/0x4f8 
> [13514.852905]  writeback_sb_inodes+0x4a4/0xab8 
> [13514.852908]  __writeback_inodes_wb+0x104/0x260 
> [13514.852912]  wb_writeback+0x840/0xce0 
> [13514.852916]  wb_do_writeback+0x69c/0x940 
> [13514.852919]  wb_workfn+0x80/0x1c0 
> [13514.852923]  process_one_work+0x774/0x12d0 
> [13514.852928]  worker_thread+0x434/0xca0 
> [13514.852932]  kthread+0x2ec/0x390 
> [13514.852936]  ret_from_fork+0x10/0x20 
> [13514.852955] INFO: lockdep is turned off. 
> [-- MARK -- Mon Oct 20 23:30:00 2025] 
> [13637.723028] INFO: task kworker/u17:206:858774 blocked for more than 737 seconds. 
> [13637.723055]       Tainted: G    B               6.18.0-rc2 #1 
> [13637.723067] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. 
> [13637.723072] task:kworker/u17:206 state:D stack:23184 pid:858774 tgid:858774 ppid:2      task_flags:0x4208060 flags:0x00000210 
> [13637.723086] Workqueue: writeback wb_workfn (flush-0:56) 
> [13637.723099] Call trace: 
> [13637.723101]  __switch_to+0x1e0/0x488 (T) 
> [13637.723110]  __schedule+0x748/0x1430 
> [13637.723115]  schedule+0xd0/0x240 
> [13637.723119]  io_schedule+0xb4/0x120 
> [13637.723124]  folio_wait_bit_common+0x2c4/0x708 
> [13637.723129]  __folio_lock+0x24/0x38 
> [13637.723133]  writeback_get_folio+0x37c/0x480 
> [13637.723138]  writeback_iter+0x128/0x6d0 
> [13637.723142]  nfs_writepages+0x504/0x930 [nfs] 
> [13637.723210]  do_writepages+0x204/0x4c0 
> [13637.723213]  __writeback_single_inode+0xec/0x4f8 
> [13637.723217]  writeback_sb_inodes+0x4a4/0xab8 
> [13637.723221]  __writeback_inodes_wb+0x104/0x260 
> [13637.723225]  wb_writeback+0x840/0xce0 
> [13637.723229]  wb_do_writeback+0x69c/0x940 
> [13637.723232]  wb_workfn+0x80/0x1c0 
> [13637.723237]  process_one_work+0x774/0x12d0 
> [13637.723242]  worker_thread+0x434/0xca0 
> [13637.723246]  kthread+0x2ec/0x390 
> [13637.723250]  ret_from_fork+0x10/0x20 
> [13637.723271] INFO: lockdep is turned off. 
> [13760.603116] INFO: task kworker/u17:206:858774 blocked for more than 860 seconds. 
> [13760.603144]       Tainted: G    B               6.18.0-rc2 #1 
> [13760.603158] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. 
> [13760.603163] task:kworker/u17:206 state:D stack:23184 pid:858774 tgid:858774 ppid:2      task_flags:0x4208060 flags:0x00000210 
> [13760.603180] Workqueue: writeback wb_workfn (flush-0:56) 
> [13760.603199] Call trace: 
> [13760.603203]  __switch_to+0x1e0/0x488 (T) 
> [13760.603214]  __schedule+0x748/0x1430 
> [13760.603221]  schedule+0xd0/0x240 
> [13760.603228]  io_schedule+0xb4/0x120 
> [13760.603234]  folio_wait_bit_common+0x2c4/0x708 
> [13760.603242]  __folio_lock+0x24/0x38 
> [13760.603248]  writeback_get_folio+0x37c/0x480 
> [13760.603341]  writeback_iter+0x128/0x6d0 
> [13760.603349]  nfs_writepages+0x504/0x930 [nfs] 
> [13760.603421]  do_writepages+0x204/0x4c0 
> [13760.603427]  __writeback_single_inode+0xec/0x4f8 
> [13760.603434]  writeback_sb_inodes+0x4a4/0xab8 
> [13760.603440]  __writeback_inodes_wb+0x104/0x260 
> [13760.603447]  wb_writeback+0x840/0xce0 
> [13760.603453]  wb_do_writeback+0x69c/0x940 
> [13760.603459]  wb_workfn+0x80/0x1c0 
> [13760.603465]  process_one_work+0x774/0x12d0 
> [13760.603472]  worker_thread+0x434/0xca0 
> [13760.603478]  kthread+0x2ec/0x390 
> [13760.603485]  ret_from_fork+0x10/0x20 
> [13760.603507] INFO: lockdep is turned off. 
> [13883.483295] INFO: task kworker/u17:206:858774 blocked for more than 983 seconds. 
> [13883.483320]       Tainted: G    B               6.18.0-rc2 #1 
> [13883.483340] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. 
> [13883.483343] task:kworker/u17:206 state:D stack:23184 pid:858774 tgid:858774 ppid:2      task_flags:0x4208060 flags:0x00000210 
> [13883.483356] Workqueue: writeback wb_workfn (flush-0:56) 
> [13883.483368] Call trace: 
> [13883.483371]  __switch_to+0x1e0/0x488 (T) 
> [13883.483379]  __schedule+0x748/0x1430 
> [13883.483384]  schedule+0xd0/0x240 
> [13883.483388]  io_schedule+0xb4/0x120 
> [13883.483392]  folio_wait_bit_common+0x2c4/0x708 
> [13883.483397]  __folio_lock+0x24/0x38 
> [13883.483401]  writeback_get_folio+0x37c/0x480 
> [13883.483406]  writeback_iter+0x128/0x6d0 
> [13883.483411]  nfs_writepages+0x504/0x930 [nfs] 
> [13883.483477]  do_writepages+0x204/0x4c0 
> [13883.483481]  __writeback_single_inode+0xec/0x4f8 
> [13883.483485]  writeback_sb_inodes+0x4a4/0xab8 
> [13883.483489]  __writeback_inodes_wb+0x104/0x260 
> [13883.483493]  wb_writeback+0x840/0xce0 
> [13883.483496]  wb_do_writeback+0x69c/0x940 
> [13883.483500]  wb_workfn+0x80/0x1c0 
> [13883.483504]  process_one_work+0x774/0x12d0 
> [13883.483509]  worker_thread+0x434/0xca0 
> [13883.483512]  kthread+0x2ec/0x390 
> [13883.483517]  ret_from_fork+0x10/0x20 
> [13883.483538] INFO: lockdep is turned off. 
> [-- MARK -- Mon Oct 20 23:35:00 2025] 
> [14006.363497] INFO: task kworker/u17:206:858774 blocked for more than 1105 seconds. 
> [14006.363521]       Tainted: G    B               6.18.0-rc2 #1 
> [14006.363542] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. 
> [14006.363545] task:kworker/u17:206 state:D stack:23184 pid:858774 tgid:858774 ppid:2      task_flags:0x4208060 flags:0x00000210 
> [14006.363557] Workqueue: writeback wb_workfn (flush-0:56) 
> [14006.363569] Call trace: 
> [14006.363571]  __switch_to+0x1e0/0x488 (T) 
> [14006.363579]  __schedule+0x748/0x1430 
> [14006.363584]  schedule+0xd0/0x240 
> [14006.363588]  io_schedule+0xb4/0x120 
> [14006.363592]  folio_wait_bit_common+0x2c4/0x708 
> [14006.363597]  __folio_lock+0x24/0x38 
> [14006.363601]  writeback_get_folio+0x37c/0x480 
> [14006.363606]  writeback_iter+0x128/0x6d0 
> [14006.363610]  nfs_writepages+0x504/0x930 [nfs] 
> [14006.363676]  do_writepages+0x204/0x4c0 
> [14006.363680]  __writeback_single_inode+0xec/0x4f8 
> [14006.363684]  writeback_sb_inodes+0x4a4/0xab8 
> [14006.363687]  __writeback_inodes_wb+0x104/0x260 
> [14006.363691]  wb_writeback+0x840/0xce0 
> [14006.363695]  wb_do_writeback+0x69c/0x940 
> [14006.363699]  wb_workfn+0x80/0x1c0 
> [14006.363702]  process_one_work+0x774/0x12d0 
> [14006.363707]  worker_thread+0x434/0xca0 
> [14006.363711]  kthread+0x2ec/0x390 
> [14006.363715]  ret_from_fork+0x10/0x20 
> [14006.363735] INFO: lockdep is turned off. 
> [14129.243869] INFO: task kworker/u17:206:858774 blocked for more than 1228 seconds. 
> [14129.244026]       Tainted: G    B               6.18.0-rc2 #1 
> [14129.244030] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. 
> [14129.244033] task:kworker/u17:206 state:D stack:23184 pid:858774 tgid:858774 ppid:2      task_flags:0x4208060 flags:0x00000210 
> [14129.244045] Workqueue: writeback wb_workfn (flush-0:56) 
> [14129.244059] Call trace: 
> [14129.244061]  __switch_to+0x1e0/0x488 (T) 
> [14129.244070]  __schedule+0x748/0x1430 
> [14129.244075]  schedule+0xd0/0x240 
> [14129.244079]  io_schedule+0xb4/0x120 
> [14129.244084]  folio_wait_bit_common+0x2c4/0x708 
> [14129.244089]  __folio_lock+0x24/0x38 
> [14129.244092]  writeback_get_folio+0x37c/0x480 
> [14129.244097]  writeback_iter+0x128/0x6d0 
> [14129.244102]  nfs_writepages+0x504/0x930 [nfs] 
> [14129.244173]  do_writepages+0x204/0x4c0 
> [14129.244176]  __writeback_single_inode+0xec/0x4f8 
> [14129.244180]  writeback_sb_inodes+0x4a4/0xab8 
> [14129.244184]  __writeback_inodes_wb+0x104/0x260 
> [14129.244188]  wb_writeback+0x840/0xce0 
> [14129.244192]  wb_do_writeback+0x69c/0x940 
> [14129.244195]  wb_workfn+0x80/0x1c0 
> [14129.244199]  process_one_work+0x774/0x12d0 
> [14129.244205]  worker_thread+0x434/0xca0 
> [14129.244209]  kthread+0x2ec/0x390 
> [14129.244213]  ret_from_fork+0x10/0x20 
> [14129.244217] Future hung task reports are suppressed, see sysctl kernel.hung_task_warnings 
> [14129.244242] INFO: lockdep is turned off. 
> [-- MARK -- Mon Oct 20 23:40:00 2025] 
> [-- MARK -- Mon Oct 20 23:45:00 2025] 
> [-- MARK -- Mon Oct 20 23:50:00 2025]
> 
> 


^ permalink raw reply	[flat|nested] 3+ messages in thread

* Re: [Bug report][xfstests generic/751] hang on nfs writeback
  2025-10-22 16:24 ` Mike Snitzer
@ 2025-10-29 20:50   ` Mike Snitzer
  0 siblings, 0 replies; 3+ messages in thread
From: Mike Snitzer @ 2025-10-29 20:50 UTC (permalink / raw)
  To: Zorro Lang; +Cc: linux-nfs, linux-mm, linux-xfs

On Wed, Oct 22, 2025 at 12:24:12PM -0400, Mike Snitzer wrote:
> On Tue, Oct 21, 2025 at 01:14:08PM +0800, Zorro Lang wrote:
> > Hi,
> > 
> > When I did xfstests regression test on nfs, it hang on generic/751 many times,
> > refer to [1](on x86_64) and [2](on aarch64). I've hit it 8 times until now.
> 
> You've hit it 8 times, all on the same kernel?  Any idea if this is a
> regression specific to 6.18-rc1 and later?
>  
> > I tested on latest linux v6.18-rc2. My underlying test device isn't loopback device,
> > it's general disk, and make xfs on it:
> >   meta-data=/dev/sda5              isize=512    agcount=4, agsize=983040 blks
> >            =                       sectsz=4096  attr=2, projid32bit=1
> >            =                       crc=1        finobt=1, sparse=1, rmapbt=1
> >            =                       reflink=1    bigtime=1 inobtcount=1 nrext64=1
> >            =                       exchange=0   metadir=0
> >   data     =                       bsize=4096   blocks=3932160, imaxpct=25
> >            =                       sunit=0      swidth=0 blks
> >   naming   =version 2              bsize=4096   ascii-ci=0, ftype=1, parent=0
> >   log      =internal log           bsize=4096   blocks=54324, version=2
> >            =                       sectsz=4096  sunit=1 blks, lazy-count=1
> >   realtime =none                   extsz=4096   blocks=0, rtextents=0
> >            =                       rgcount=0    rgsize=0 extents
> >            =                       zoned=0      start=0 reserved=0
> > 
> > Two xfs are mounted on /mnt/xfstests/test and /mnt/xfstests/scratch seperately,
> > then export as:
> >   # cat /etc/exports
> >   /mnt/xfstests/test/nfs-server *(rw,insecure,no_root_squash)
> >   /mnt/xfstests/scratch/nfs-server *(rw,insecure,no_root_squash)
> > 
> > The nfs mount option is only "-o vers=4.2". BTW, nfs server and client are
> > in same machine/system.
> 
> Being that NFS client and server are on the same localhost, I see
> NFS's LOCALIO is being used (nfslocaliod mentioned in "locks held"
> section, so IO is bypassing going over the wire to NFSD and is instead
> being issued directly from NFS client through LOCALIO workqueue and
> then down to XFS).
> 
> I'll try to make sense of the locking (reported below) but this will
> take some time.
> 
> Cc'ing XFS list because this is NFS LOCALIO ontop of XFS, could be
> there is something fundamental in LOCALIO's nfslocaliod workqueue that
> is setting us up for deadlock if XFS takes a specific path...
> 
> NOTE: fs/nfs/inode.c:nfsiod_start(), XFS's nature has influenced
> LOCALIO workqueue config hacks already.. could be dragons lurking:
> 
> /*
>  * Start the nfsiod workqueues
>  */
> static int nfsiod_start(void)
> {
>         dprintk("RPC:       creating workqueue nfsiod\n");
>         nfsiod_workqueue = alloc_workqueue("nfsiod", WQ_MEM_RECLAIM | WQ_UNBOUND, 0);
>         if (nfsiod_workqueue == NULL)
>                 return -ENOMEM;
> #if IS_ENABLED(CONFIG_NFS_LOCALIO)
>         /*
>          * localio writes need to use a normal (non-memreclaim) workqueue.
>          * When we start getting low on space, XFS goes and calls flush_work() on
>          * a non-memreclaim work queue, which causes a priority inversion problem.
>          */
>         dprintk("RPC:       creating workqueue nfslocaliod\n");
>         nfslocaliod_workqueue = alloc_workqueue("nfslocaliod", WQ_UNBOUND, 0);
>         if (unlikely(nfslocaliod_workqueue == NULL)) {
>                 nfsiod_stop();
>                 return -ENOMEM;
>         }
> #endif /* CONFIG_NFS_LOCALIO */
>         return 0;
> }
> 
> Mike
> 
>  
> > # cat local.config
> > export FSTYP=nfs
> > export TEST_DEV=xxxx-xxx-xxxx.xxxx.xxx.xxx:/mnt/xfstests/test/nfs-server
> > export TEST_DIR=/mnt/xfstests/test/nfs-client
> > export SCRATCH_DEV=xxxx-xxx-xxxx.xxxx.xxx.xxx:/mnt/xfstests/scratch/nfs-server
> > export SCRATCH_MNT=/mnt/xfstests/scratch/nfs-client
> > export MOUNT_OPTIONS="-o vers=4.2"
> > export TEST_FS_MOUNT_OPTS="-o vers=4.2"
> > 
> > 
> > [1]
> > [23369.572660] run fstests generic/751 at 2025-10-20 23:15:24 
> > [-- MARK -- Tue Oct 21 03:20:00 2025] 
> > [-- MARK -- Tue Oct 21 03:25:00 2025] 
> > [-- MARK -- Tue Oct 21 03:30:00 2025] 
> > [-- MARK -- Tue Oct 21 03:35:00 2025] 
> > [-- MARK -- Tue Oct 21 03:40:00 2025] 
> > [25069.500900] INFO: task kworker/u9:73:825900 blocked for more than 122 seconds. 
> > [25069.501484]       Not tainted 6.18.0-rc2 #1 
> > [25069.501822] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. 
> > [25069.502385] task:kworker/u9:73   state:D stack:24712 pid:825900 tgid:825900 ppid:2      task_flags:0x4208060 flags:0x00080000 
> > [25069.503258] Workqueue: writeback wb_workfn (flush-0:48) 
> > [25069.782725] Call Trace: 
> > [25069.787191]  <TASK> 
> > [25069.787368]  __schedule+0x838/0x1890 
> > [25069.799533]  ? __pfx___schedule+0x10/0x10 
> > [25069.799852]  ? __blk_flush_plug+0x27b/0x4d0 
> > [25069.800617]  ? find_held_lock+0x32/0x90 
> > [25069.801292]  ? __lock_release.isra.0+0x1a4/0x2c0 
> > [25069.801639]  schedule+0xd4/0x260 
> > [25069.801902]  io_schedule+0x8f/0xf0 
> > [25069.802162]  folio_wait_bit_common+0x2d9/0x780 
> > [25069.802919]  ? folio_wait_bit_common+0x1dd/0x780 
> > [25069.803324]  ? __pfx_folio_wait_bit_common+0x10/0x10 
> > [25069.803684]  ? nfs_page_clear_headlock+0x31/0x80 [nfs] 
> > [25069.804254]  ? __pfx_wake_page_function+0x10/0x10 
> > [25069.804602]  ? __pfx___might_resched+0x10/0x10 
> > [25069.805160]  writeback_get_folio+0x3f9/0x500 
> > [25069.806221]  writeback_iter+0x136/0x720 
> > [25069.806628]  nfs_writepages+0x4f8/0x9b0 [nfs] 
> > [25069.807100]  ? mark_held_locks+0x40/0x70 
> > [25069.807403]  ? __pfx_nfs_writepages+0x10/0x10 [nfs] 
> > [25069.807846]  ? virtqueue_notify+0x68/0xc0 
> > [25069.808346]  ? virtio_queue_rq+0x2b1/0x650 [virtio_blk] 
> > [25069.808925]  ? __lock_acquire+0x57c/0xbd0 
> > [25069.809295]  do_writepages+0x21f/0x560 
> > [25069.809576]  ? __pfx_do_writepages+0x10/0x10 
> > [25069.809953]  ? rcu_is_watching+0x15/0xb0 
> > [25069.810587]  __writeback_single_inode+0xe2/0x5f0 
> > [25069.810982]  ? __lock_release.isra.0+0x1a4/0x2c0 
> > [25069.811373]  ? __pfx___writeback_single_inode+0x10/0x10 
> > [25069.811815]  ? writeback_sb_inodes+0x416/0xd00 
> > [25069.812194]  writeback_sb_inodes+0x535/0xd00 
> > [25069.812533]  ? __pfx_stack_trace_save+0x10/0x10 
> > [25069.814193]  ? local_clock_noinstr+0xd/0xe0 
> > [25069.814506]  ? __pfx_writeback_sb_inodes+0x10/0x10 
> > [25069.814910]  ? __lock_release.isra.0+0x1a4/0x2c0 
> > [25069.815407]  ? lock_acquire+0x10b/0x150 
> > [25069.815703]  ? down_read_trylock+0x4b/0x60 
> > [25069.816077]  __writeback_inodes_wb+0xf4/0x270 
> > [25069.816441]  ? __pfx___writeback_inodes_wb+0x10/0x10 
> > [25069.816864]  ? queue_io+0x329/0x510 
> > [25069.817148]  wb_writeback+0x70a/0x9c0 
> > [25069.817448]  ? __pfx_wb_writeback+0x10/0x10 
> > [25069.817831]  ? get_nr_dirty_inodes+0xcb/0x180 
> > [25069.818264]  wb_do_writeback+0x5d4/0x8e0 
> > [25069.818575]  ? __pfx_wb_do_writeback+0x10/0x10 
> > [25069.818975]  ? set_worker_desc+0x16e/0x190 
> > [25069.819313]  ? __pfx_set_worker_desc+0x10/0x10 
> > [25069.819669]  wb_workfn+0x7c/0x200 
> > [25069.819997]  process_one_work+0xd8b/0x1320 
> > [25069.820342]  ? __pfx_process_one_work+0x10/0x10 
> > [25069.820694]  ? assign_work+0x16c/0x240 
> > [25069.821053]  worker_thread+0x5f3/0xfe0 
> > [25069.821364]  ? __pfx_worker_thread+0x10/0x10 
> > [25069.821697]  kthread+0x3b4/0x770 
> > [25069.822122]  ? kvm_sched_clock_read+0x11/0x20 
> > [25069.822457]  ? local_clock_noinstr+0xd/0xe0 
> > [25069.822790]  ? __pfx_kthread+0x10/0x10 
> > [25069.823079]  ? __lock_release.isra.0+0x1a4/0x2c0 
> > [25069.823430]  ? rcu_is_watching+0x15/0xb0 
> > [25069.823729]  ? __pfx_kthread+0x10/0x10 
> > [25069.824035]  ret_from_fork+0x393/0x480 
> > [25069.864794]  ? __pfx_kthread+0x10/0x10 
> > [25069.865076]  ? __pfx_kthread+0x10/0x10 
> > [25069.865371]  ret_from_fork_asm+0x1a/0x30 
> > [25069.879905]  </TASK> 
> > [25069.880145]  
> > [25069.880145] Showing all locks held in the system: 
> > [25069.889550] 1 lock held by khungtaskd/37: 
> > [25069.889886]  #0: ffffffffa29309e0 (rcu_read_lock){....}-{1:3}, at: rcu_lock_acquire.constprop.0+0x7/0x30 
> > [25069.890603] 2 locks held by 751/825733: 
> > [25069.890942]  #0: ffff888101e36440 (sb_writers#16){.+.+}-{0:0}, at: ksys_write+0xf9/0x1d0 
> > [25069.891589]  #1: ffffffffa2b97fd0 (split_debug_mutex){+.+.}-{4:4}, at: split_huge_pages_write+0x124/0x430 
> > [25069.892449] 3 locks held by kworker/u9:73/825900: 
> > [25069.892818]  #0: ffff8881029df958 ((wq_completion)writeback){+.+.}-{0:0}, at: process_one_work+0x7f5/0x1320 
> > [25069.893530]  #1: ffffc9000ad9fd10 ((work_completion)(&(&wb->dwork)->work)){+.+.}-{0:0}, at: process_one_work+0xd3f/0x1320 
> > [25069.894333]  #2: ffff88810b1480e8 (&type->s_umount_key#68){++++}-{4:4}, at: super_trylock_shared+0x1c/0xa0 
> > [25069.895056] 2 locks held by kworker/u10:2/826054: 
> > [25069.895418]  #0: ffff888103ec2158 ((wq_completion)nfsiod){+.+.}-{0:0}, at: process_one_work+0x7f5/0x1320 
> > [25069.896123]  #1: ffffc90001e1fd10 ((work_completion)(&ctx->work)){+.+.}-{0:0}, at: process_one_work+0xd3f/0x1320 
> > [25069.896952]  
> > [25069.897085] ============================================= 

I was able to test xfstests generic/751 on a server in the Red Hat
lab, but I used a different baseline kernel so far (its 6.12.53-based
but has all of NFSD, NFS and SUNRPC changes from 6.18-rc1 backported).

The test runs without deadlock but there is one lockdep splat due to
missing this commit:

commit 45f69d091bab64a332fe751da9829dcd136348fd
Author: Long Li <leo.lilong@huawei.com>
Date:   Sat Jun 22 16:26:31 2024 +0800

    xfs: eliminate lockdep false positives in xfs_attr_shortform_list

    xfs_attr_shortform_list() only called from a non-transactional context, it
    hold ilock before alloc memory and maybe trapped in memory reclaim. Since
    commit 204fae32d5f7("xfs: clean up remaining GFP_NOFS users") removed
    GFP_NOFS flag, lockdep warning will be report as [1]. Eliminate lockdep
    false positives by use __GFP_NOLOCKDEP to alloc memory
    in xfs_attr_shortform_list().

    [1] https://lore.kernel.org/linux-xfs/000000000000e33add0616358204@google.com/

    Reported-by: syzbot+4248e91deb3db78358a2@syzkaller.appspotmail.com
    Signed-off-by: Long Li <leo.lilong@huawei.com>
    Reviewed-by: Dave Chinner <dchinner@redhat.com>
    Signed-off-by: Carlos Maiolino <cem@kernel.org>

After applying that commit, a generic/751 retest comes up clean. I'll
now switch to testing latest v6.18-rcX kernel.

Mike


^ permalink raw reply	[flat|nested] 3+ messages in thread

end of thread, other threads:[~2025-10-29 20:50 UTC | newest]

Thread overview: 3+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2025-10-21  5:14 [Bug report][xfstests generic/751] hang on nfs writeback Zorro Lang
2025-10-22 16:24 ` Mike Snitzer
2025-10-29 20:50   ` Mike Snitzer

This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox