linux-mm.kvack.org archive mirror
 help / color / mirror / Atom feed
From: "Pulavarty, Badari" <badari.pulavarty@intel.com>
To: "david@fromorbit.com" <david@fromorbit.com>
Cc: "akpm@linux-foundation.org" <akpm@linux-foundation.org>,
	"bfoster@redhat.com" <bfoster@redhat.com>,
	"huangzhaoyang@gmail.com" <huangzhaoyang@gmail.com>,
	"ke.wang@unisoc.com" <ke.wang@unisoc.com>,
	"linux-fsdevel@vger.kernel.org" <linux-fsdevel@vger.kernel.org>,
	"inux-kernel@vger.kernel.org" <inux-kernel@vger.kernel.org>,
	"linux-mm@kvack.org" <linux-mm@kvack.org>,
	"willy@infradead.org" <willy@infradead.org>,
	"zhaoyang.huang@unisoc.com" <zhaoyang.huang@unisoc.com>,
	"Shutemov, Kirill" <kirill.shutemov@intel.com>,
	"Tang, Feng" <feng.tang@intel.com>,
	"Huang, Ying" <ying.huang@intel.com>,
	"Yin, Fengwei" <fengwei.yin@intel.com>,
	"Hansen, Dave" <dave.hansen@intel.com>,
	"Zanussi, Tom" <tom.zanussi@intel.com>
Subject: Re: [RFC PATCH] mm: move xa forward when run across zombie page
Date: Fri, 21 Oct 2022 21:37:36 +0000	[thread overview]
Message-ID: <DM6PR11MB3978E31FE5149BA89D371E079C2D9@DM6PR11MB3978.namprd11.prod.outlook.com> (raw)

Hi All,

I have been tracking similar issue(s) with soft lockup or panics on my system consistently with my workload.
Tried multiple kernel versions. Issue seem to happen consistently on 6.1-rc1 (while it seem to happen on 5.17, 5.19, 6.0.X)

PANIC: "Kernel panic - not syncing: softlockup: hung tasks"

    RIP: 0000000000000001  RSP: ff3d8e7f0d9978ea  RFLAGS: ff3d8e7f0d9978e8
    RAX: 0000000000000000  RBX: 0000000000000000  RCX: 0000000000000000
    RDX: 000000006b9c66f1  RSI: ff506ca15ff33c20  RDI: 0000000000000000
    RBP: ffffffff84bc64cc   R8: ff3d8e412cabdff0   R9: ffffffff84c00e8b
    R10: ff506ca15ff33b69  R11: 0000000000000000  R12: ff506ca15ff33b58
    R13: ffffffff84bc79a3  R14: ff506ca15ff33b38  R15: 0000000000000000
    ORIG_RAX: ff506ca15ff33a80  CS: ff506ca15ff33c78  SS: 0000
#9 [ff506ca15ff33c18] xas_load at ffffffff84b49a7f
#10 [ff506ca15ff33c28] __filemap_get_folio at ffffffff840985da
#11 [ff506ca15ff33ce8] swap_cache_get_folio at ffffffff841119db
#12 [ff506ca15ff33d18] do_swap_page at ffffffff840dbd21
#13 [ff506ca15ff33db8] __handle_mm_fault at ffffffff840ddee3
#14 [ff506ca15ff33e88] handle_mm_fault at ffffffff840de55d
#15 [ff506ca15ff33ec8] do_user_addr_fault at ffffffff83e93247
#16 [ff506ca15ff33f20] exc_page_fault at ffffffff84bc711d
#17 [ff506ca15ff33f50] asm_exc_page_fault at ffffffff84c00b77

Tried various patches proposed on this thread chain.. but no luck so far.

Looks like its stuck in following loop forever causing softlockup/panic:

 if (!folio_try_get_rcu(folio)) 
                goto repeat;

Looking at the crash dump, mapping->host became NULL. Not sure what exactly is happening.
Welcome any ideas to track it down further.

struct address_space {
  host = 0x0,
  i_pages = {
    xa_lock = {
      {
        rlock = {
          raw_lock = {
            {
              val = {
                counter = 0
              },
              {
                locked = 0 '\000',
                pending = 0 '\000'
              },
              {
                locked_pending = 0,
                tail = 0
              }
            }
          }
        }
      }
    },
    xa_flags = 1,
    xa_head = 0xff3d8e7f9ca41daa
  },
  invalidate_lock = {
    count = {
      counter = 0
    },
    owner = {
      counter = 0
    },
    osq = {
      tail = {
        counter = 0
      }
    },
    wait_lock = {
      raw_lock = {
        {
          val = {
            counter = 0
          },
          {
            locked = 0 '\000',
            pending = 0 '\000'
          },
          {
            locked_pending = 0,
            tail = 0
          }
        }
      }
    },
    wait_list = {
      next = 0x0,
      prev = 0x0
    }
  },
  gfp_mask = 0,
  i_mmap_writable = {
    counter = 0
  },
  i_mmap = {
    rb_root = {
      rb_node = 0x0
    },
    rb_leftmost = 0x0
  },
  i_mmap_rwsem = {
    count = {
      counter = 0
    },
    owner = {
      counter = 0
    },
    osq = {
      tail = {
        counter = 0
      }
    },
    wait_lock = {
      raw_lock = {
        {
          val = {
            counter = 0
          },
          {
            locked = 0 '\000',
            pending = 0 '\000'
          },
          {
            locked_pending = 0,
            tail = 0
          }
        }
      }
    },
    wait_list = {
      next = 0x0,
      prev = 0x0
    }
  },
  nrpages = 1897,
  writeback_index = 0,
  a_ops = 0xffffffff85044560,
  flags = 32,
  wb_err = 0,
  private_lock = {
    {
      rlock = {
        raw_lock = {
          {
            val = {
              counter = 0
            },
            {
              locked = 0 '\000',
              pending = 0 '\000'
            },
            {
              locked_pending = 0,
              tail = 0
            }
          }
        }
      }
    }
  },
  private_list = {
    next = 0x0,
    prev = 0x0
  },
  private_data = 0x0
}



Thanks,
Badari


             reply	other threads:[~2022-10-21 21:37 UTC|newest]

Thread overview: 35+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2022-10-21 21:37 Pulavarty, Badari [this message]
2022-10-21 22:31 ` Matthew Wilcox
2022-10-21 22:40   ` Pulavarty, Badari
2022-10-31 19:25   ` Pulavarty, Badari
2022-10-31 19:39     ` Hugh Dickins
2022-10-31 21:33       ` Pulavarty, Badari
  -- strict thread matches above, loose matches on Subject: below --
2022-10-14  5:30 zhaoyang.huang
2022-10-14 12:11 ` Matthew Wilcox
2022-10-17  5:34   ` Zhaoyang Huang
2022-10-17  6:58     ` Zhaoyang Huang
2022-10-17 15:55     ` Matthew Wilcox
2022-10-18  2:52       ` Zhaoyang Huang
2022-10-18  3:09         ` Matthew Wilcox
2022-10-18 22:30           ` Dave Chinner
2022-10-19  1:16             ` Dave Chinner
2022-10-19  4:47               ` Dave Chinner
2022-10-19  5:48                 ` Zhaoyang Huang
2022-10-19 13:06                   ` Matthew Wilcox
2022-10-20  1:27                     ` Zhaoyang Huang
2022-10-26 19:49                   ` Matthew Wilcox
2022-10-27  1:57                     ` Zhaoyang Huang
2022-10-19 11:49             ` Brian Foster
2022-10-20  2:04               ` Dave Chinner
2022-10-20  3:12                 ` Zhaoyang Huang
2022-10-19 15:23             ` Matthew Wilcox
2022-10-19 22:04               ` Dave Chinner
2022-10-19 22:46                 ` Dave Chinner
2022-10-19 23:42                   ` Dave Chinner
2022-10-20 21:52                 ` Matthew Wilcox
2022-10-26  8:38                   ` Zhaoyang Huang
2022-10-26 14:38                     ` Matthew Wilcox
2022-10-26 16:01                   ` Matthew Wilcox
2022-10-28  4:05                     ` Dave Chinner
2022-11-01  7:17                   ` Dave Chinner
2024-04-11  7:04                     ` Zhaoyang Huang

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=DM6PR11MB3978E31FE5149BA89D371E079C2D9@DM6PR11MB3978.namprd11.prod.outlook.com \
    --to=badari.pulavarty@intel.com \
    --cc=akpm@linux-foundation.org \
    --cc=bfoster@redhat.com \
    --cc=dave.hansen@intel.com \
    --cc=david@fromorbit.com \
    --cc=feng.tang@intel.com \
    --cc=fengwei.yin@intel.com \
    --cc=huangzhaoyang@gmail.com \
    --cc=inux-kernel@vger.kernel.org \
    --cc=ke.wang@unisoc.com \
    --cc=kirill.shutemov@intel.com \
    --cc=linux-fsdevel@vger.kernel.org \
    --cc=linux-mm@kvack.org \
    --cc=tom.zanussi@intel.com \
    --cc=willy@infradead.org \
    --cc=ying.huang@intel.com \
    --cc=zhaoyang.huang@unisoc.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox