From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org Received: from kanga.kvack.org (kanga.kvack.org [205.233.56.17]) by smtp.lore.kernel.org (Postfix) with ESMTP id 4C058C6FD18 for ; Thu, 20 Apr 2023 02:07:59 +0000 (UTC) Received: by kanga.kvack.org (Postfix) id DCF516B0074; Wed, 19 Apr 2023 22:07:58 -0400 (EDT) Received: by kanga.kvack.org (Postfix, from userid 40) id D7E946B0075; Wed, 19 Apr 2023 22:07:58 -0400 (EDT) X-Delivered-To: int-list-linux-mm@kvack.org Received: by kanga.kvack.org (Postfix, from userid 63042) id BF86E6B0078; Wed, 19 Apr 2023 22:07:58 -0400 (EDT) X-Delivered-To: linux-mm@kvack.org Received: from relay.hostedemail.com (smtprelay0015.hostedemail.com [216.40.44.15]) by kanga.kvack.org (Postfix) with ESMTP id ACFF96B0074 for ; Wed, 19 Apr 2023 22:07:58 -0400 (EDT) Received: from smtpin01.hostedemail.com (a10.router.float.18 [10.200.18.1]) by unirelay04.hostedemail.com (Postfix) with ESMTP id 7937F1A0370 for ; Thu, 20 Apr 2023 02:07:58 +0000 (UTC) X-FDA: 80700133836.01.CD70A7A Received: from mail1.bemta37.messagelabs.com (mail1.bemta37.messagelabs.com [85.158.142.1]) by imf14.hostedemail.com (Postfix) with ESMTP id 97191100013 for ; Thu, 20 Apr 2023 02:07:55 +0000 (UTC) Authentication-Results: imf14.hostedemail.com; dkim=pass header.d=fujitsu.com header.s=170520fj header.b=u+xeus85; spf=pass (imf14.hostedemail.com: domain of ruansy.fnst@fujitsu.com designates 85.158.142.1 as permitted sender) smtp.mailfrom=ruansy.fnst@fujitsu.com; dmarc=pass (policy=none) header.from=fujitsu.com ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=hostedemail.com; s=arc-20220608; t=1681956476; h=from:from:sender:reply-to:subject:subject:date:date: message-id:message-id:to:to:cc:cc:mime-version:mime-version: content-type:content-type: content-transfer-encoding:content-transfer-encoding: in-reply-to:in-reply-to:references:references:dkim-signature; bh=PfWgfbWMhubqztmsyQFlopaTe+XA8QMuhGDUA2zZUbI=; b=Lfxm2XHoQjo+OX244tJN/O891ZeccWHq1jwaliKFgQOTAFe8zw4VwnmM28Z+A+L2TYl1KR Ga+mV+1MtdSqdhnZ4wflbdIirSdZvge/KbXZYuWljoXMfszYsp8tMkJijCWp85/tCM9Yxe ldyrEL+CEpk0lb+8RTxcJ5/gkoA/CX4= ARC-Authentication-Results: i=1; imf14.hostedemail.com; dkim=pass header.d=fujitsu.com header.s=170520fj header.b=u+xeus85; spf=pass (imf14.hostedemail.com: domain of ruansy.fnst@fujitsu.com designates 85.158.142.1 as permitted sender) smtp.mailfrom=ruansy.fnst@fujitsu.com; dmarc=pass (policy=none) header.from=fujitsu.com ARC-Seal: i=1; s=arc-20220608; d=hostedemail.com; t=1681956476; a=rsa-sha256; cv=none; b=WB8jU4ib+VeVrm2TnT9gHwytS0nj+AT457MUeIOMLjeZoWGiL2yAoXHpkVFRGly4jWXkAr 0tsVyWs0epNlgWAVUFX/CDpvnV1GjfhNgP3lfjLu8fn/hh6aut7XWtPykaP8bvpyFs9Y3w FGj1wQq2O67P1hA5QNYzoLt9OCttQqc= DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=fujitsu.com; s=170520fj; t=1681956473; i=@fujitsu.com; bh=PfWgfbWMhubqztmsyQFlopaTe+XA8QMuhGDUA2zZUbI=; h=Message-ID:Date:MIME-Version:Subject:From:To:CC:References: In-Reply-To:Content-Type:Content-Transfer-Encoding; b=u+xeus85Adzw0Ms8GenUXlFeyOWqSpKnlflp3YH9d1nuyM7PPBwRataCTAZmHsnvq lrmBQblMyXoCU/2RRxYNW72eOAFMlZyO8IHA8wfC+GV/VFK4nE/7SEul77x+L8+hC2 aIGHDcPOehiD5LTeSrbyCybhfKiNMtRG/T5aYItdKIGhYuFcWlO4bjr9ugxZZilBH+ 7s+ViSXp+xjPWuto12axTo/SV45O04wm4xmpKJMgarctWxDL2wqvoaOgPeol0FQSNX D5uUnzdc+malJWEE3hNh+cbRH3zaLJuRBW9SB1CecXLtrs03zZ7yGIu9BfaNxs6oLs IkUT4u7pLgxkw== X-Brightmail-Tracker: H4sIAAAAAAAAA+NgFtrEKsWRWlGSWpSXmKPExsViZ8ORpFs+zyH FYMN8GYs569ewWUyfeoHR4vITPovZ05uZLPbsPclicW/Nf1aLXX92sFus/PGH1eL3jzlsDpwe m1doeSze85LJY9OqTjaPTZ8msXucmPGbxePF5pmMHmcWHGH3+LxJLoAjijUzLym/IoE1o2HuP KaC6fEVnx78Y2pgvOTTxcjJISSwkVFiaqcChL2USeLkK88uRi4gezujRPf7aWwgCV4BO4mNf/ cyg9gsAqoS01ddZIKIC0qcnPmEBcQWFUiWOLapDaxeWCBUYuGfOWA1bAI6EhcW/GUFsUUECiX 2LH0HVs8sUCHRuOgfM8TiGonGpxfBejkFXCReLp8BVWMhsfjNQXYIW16ieetssHoJASWJi1/v sELYlRKtH36xQNhqElfPbWKewCg0C8l5s5CMmoVk1AJG5lWM5sWpRWWpRbqGBnpJRZnpGSW5i Zk5eolVuol6qaW6eflFJRm6hnqJ5cV6qcXFesWVuck5KXp5qSWbGIHRl1KcOH8H4+W+v3qHGC U5mJREeZU326cI8SXlp1RmJBZnxBeV5qQWH2KU4eBQkuCtnOGQIiRYlJqeWpGWmQNMBDBpCQ4 eJRFe816gNG9xQWJucWY6ROoUoy7H2oYDe5mFWPLy81KlxHmV5wIVCYAUZZTmwY2AJaVLjLJS wryMDAwMQjwFqUW5mSWo8q8YxTkYlYR5i0Cm8GTmlcBtegV0BBPQEQ/FwI4oSURISTUwbd7ur cQnIDzZUWJRmFVKe1T75cP3mfKu6+dq/bdqm5l4y/NE7qwaW6awtTv3/Dp4eN0LuU7vjVtVmZ nUNx5YwLb9mZSV2F7ve/kNXXJJER8vpbDUaMcI3Sn2OFjuU+B2Kqr06CqpV4kaKSfMJmVsjc/ KmOHB6a/IdCykaN6Cy7fvXPtp2OB5Yv1KPv+32rsMOTee2KLY0SSccsU79dXRezPKbU+ZCvIE GDaotdu9/2r0YnbVxcX6uiZsSmEHxEO0d73xnpH7se/6rtTgXY+041h/qeeG/ftt/qHy6sst2 SpcOxacTjht+YD3vaK44cs9nTpp3RPv3buVvjntwvF999+sU1t+91HPXjfGlo87lViKMxINtZ iLihMBtEJDwcUDAAA= X-Env-Sender: ruansy.fnst@fujitsu.com X-Msg-Ref: server-6.tower-745.messagelabs.com!1681956471!10184!1 X-Originating-IP: [62.60.8.98] X-SYMC-ESS-Client-Auth: outbound-route-from=pass X-StarScan-Received: X-StarScan-Version: 9.104.2; banners=-,-,- X-VirusChecked: Checked Received: (qmail 7634 invoked from network); 20 Apr 2023 02:07:51 -0000 Received: from unknown (HELO n03ukasimr03.n03.fujitsu.local) (62.60.8.98) by server-6.tower-745.messagelabs.com with ECDHE-RSA-AES256-GCM-SHA384 encrypted SMTP; 20 Apr 2023 02:07:51 -0000 Received: from n03ukasimr03.n03.fujitsu.local (localhost [127.0.0.1]) by n03ukasimr03.n03.fujitsu.local (Postfix) with ESMTP id C53201C3; Thu, 20 Apr 2023 03:07:50 +0100 (BST) Received: from R01UKEXCASM223.r01.fujitsu.local (R01UKEXCASM223 [10.182.185.121]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-SHA384 (256/256 bits)) (No client certificate requested) by n03ukasimr03.n03.fujitsu.local (Postfix) with ESMTPS id B7E671B9; Thu, 20 Apr 2023 03:07:50 +0100 (BST) Received: from [192.168.50.5] (10.167.234.230) by R01UKEXCASM223.r01.fujitsu.local (10.182.185.121) with Microsoft SMTP Server (TLS) id 15.0.1497.42; Thu, 20 Apr 2023 03:07:46 +0100 Message-ID: <0a53ee26-5771-0808-ccdc-d1739c9dacac@fujitsu.com> Date: Thu, 20 Apr 2023 10:07:39 +0800 MIME-Version: 1.0 User-Agent: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:102.0) Gecko/20100101 Thunderbird/102.9.0 Subject: Re: [RFC PATCH v11.1 2/2] mm, pmem, xfs: Introduce MF_MEM_REMOVE for unbind From: Shiyang Ruan To: , , , CC: , , , , References: <1679996506-2-3-git-send-email-ruansy.fnst@fujitsu.com> <1681296735-2-1-git-send-email-ruansy.fnst@fujitsu.com> In-Reply-To: <1681296735-2-1-git-send-email-ruansy.fnst@fujitsu.com> Content-Type: text/plain; charset="UTF-8"; format=flowed Content-Transfer-Encoding: 8bit X-Originating-IP: [10.167.234.230] X-ClientProxiedBy: G08CNEXCHPEKD07.g08.fujitsu.local (10.167.33.80) To R01UKEXCASM223.r01.fujitsu.local (10.182.185.121) X-Virus-Scanned: ClamAV using ClamSMTP X-Rspamd-Server: rspam07 X-Rspamd-Queue-Id: 97191100013 X-Rspam-User: X-Stat-Signature: k3wbzk7tj7d6uyy4cukcq18rmx5h6wto X-HE-Tag: 1681956475-60843 X-HE-Meta: U2FsdGVkX19XtxQ1xTqivdhf/tZsiJKvfh+8h0cEJz/ITFFgc3TpYAe9n+jGvsvh2DfUQsF56dW36M/BWOx0skQHYvNDwVgYNqzSYOX30DKb7PdZ4LbeDyfZ3p1EnSWjYRnqbdkbllTgDesgKSgXndoqlYfcQYiFskzOUIl+38kLEmeWHO+0GJvH3lG7Qec/+H27mwCPbRLh2OrYPN8USVTM4XK8aca8Gw/kTUV/6prcos4DvSGvfXDlKcMOg0Pcy6kF+tfE77FkZhs1LT3qjMmpFWVfgPDRZi/0M3JSWLLbCKvVfj9jBxu4ZvM5KIfZNMvYiGp9pDLQPfQAkPCrfeLc9vl7InoXndlgtXItJOA4cUMg7LT7Zf2lcepccp/T+xibu73bAbQOdjiA2FBJdG74U6Qw7YgbSK4jcI2R3t8n/g1qAfz1V9fWhPOplCqeBFHnVgig8eqvBhHd9JrWK5gQjrwIVplVDAkun/fn/IcmAzTcG08TDyOdS+seOMBiP6WolIAxBy7a62a79APbtKEuFfBfV50+JKpmbpg9JB/nMKEUAf2XDactXvdnPc5Nuutel9Q4eU4SGCzWbN/dqENB9N42tixye8P+ra8XTdEw3oUafuPzIQPj7B8siub1/x1Tl1IH3uGcs6tE/O+C3br/SuTnHXNuoDioCb7cH3SoS9mBP+Y0+iOUFom9iobZbrAaiwc8eGXNF9/2i4CBuCH8gQBai8Oj9Esw+hb4FbGtohQydw8hqYlGOGF2jMlKpj49lYNStEl0mmWc2qnUPxNqrciLBtAKFHHy3ipaSnLT84Up9EYUT4P+2P+tjYds//IMMXMCl7r4bLewsI101udZviAesM4Fo/oS5zXiwjP94oju8s93vEB2RjIFVV6pj/dsPnW23i4MxYbi2DLcwkKJ3SDq7BPSevRdzo66ArSIr97FlF6ZbFhoPdW3kqit0aMAvCfjm0BMl3IgfWj HnGVSyYO 5n9f77EF2OjedGteBDGDhXSw4Pl/TWWPah31/OR+SPe4jlo89LdcX1dbwqX4HQsdbav+isqe0Of/nC/cfYSkBFwt9gr5lVKTXCncHo34GXmn0lvB3HOXj8z4fKiVicyCMEJhlYZMXQ0yeT4SwKWQ4+53r55KXGb6Bb/laT70gYsbEqyPJU9HCY9mjOKDVlfUCWpTTYa1k7GObxcH0P5jRHCdonqbnSqlIpq+A8Mp5MvaRUMdySgJ8P3VwytP42r2NmvMuqsiA3tNsPeYty8+vC93paZBcgm/rY24kudElxoMk4SUpYDRIM6tbxVvbkcJkj4R4eSTUmKEdXm8e6KMSwRiiZaM4sW2qeQmZVW5a6uF/VseBSHu+RoTgAA== X-Bogosity: Ham, tests=bogofilter, spamicity=0.000000, version=1.2.4 Sender: owner-linux-mm@kvack.org Precedence: bulk X-Loop: owner-majordomo@kvack.org List-ID: Ping~ 在 2023/4/12 18:52, Shiyang Ruan 写道: > This is a RFC HOTFIX. > > This hotfix adds a exclusive forzen state to make sure any others won't > thaw the fs during xfs_dax_notify_failure(): > > #define SB_FREEZE_EXCLUSIVE (SB_FREEZE_COMPLETE + 2) > Using +2 here is because Darrick's patch[0] is using +1. So, should we > make these definitions global? > > Another thing I can't make up my mind is: when another freezer has freeze > the fs, should we wait unitl it finish, or print a warning in dmesg and > return -EBUSY? > > Since there are at least 2 places needs exclusive forzen state, I think > we can refactor helper functions of freeze/thaw for them. e.g. > int freeze_super_exclusive(struct super_block *sb, int frozen); > int thaw_super_exclusive(struct super_block *sb, int frozen); > > [0] https://git.kernel.org/pub/scm/linux/kernel/git/djwong/xfs-linux.git/commit/?h=repair-fscounters&id=c3a0d1de4d54ffb565dbc7092dfe1fb851940669 > > > --- Original commit message --- > This patch is inspired by Dan's "mm, dax, pmem: Introduce > dev_pagemap_failure()"[1]. With the help of dax_holder and > ->notify_failure() mechanism, the pmem driver is able to ask filesystem > (or mapped device) on it to unmap all files in use and notify processes > who are using those files. > > Call trace: > trigger unbind > -> unbind_store() > -> ... (skip) > -> devres_release_all() > -> kill_dax() > -> dax_holder_notify_failure(dax_dev, 0, U64_MAX, MF_MEM_PRE_REMOVE) > -> xfs_dax_notify_failure() > `-> freeze_super() > `-> do xfs rmap > ` -> mf_dax_kill_procs() > ` -> collect_procs_fsdax() // all associated > ` -> unmap_and_kill() > ` -> invalidate_inode_pages2() // drop file's cache > `-> thaw_super() > > Introduce MF_MEM_PRE_REMOVE to let filesystem know this is a remove > event. Also introduce a exclusive freeze/thaw to lock the filesystem to > prevent new dax mapping from being created. And do not shutdown > filesystem directly if something not supported, or if failure range > includes metadata area. Make sure all files and processes are handled > correctly. Also drop the cache of associated files before pmem is > removed. > > [1]: https://lore.kernel.org/linux-mm/161604050314.1463742.14151665140035795571.stgit@dwillia2-desk3.amr.corp.intel.com/ > > Signed-off-by: Shiyang Ruan > --- > drivers/dax/super.c | 3 +- > fs/xfs/xfs_notify_failure.c | 151 ++++++++++++++++++++++++++++++++++-- > include/linux/mm.h | 1 + > mm/memory-failure.c | 17 +++- > 4 files changed, 162 insertions(+), 10 deletions(-) > > diff --git a/drivers/dax/super.c b/drivers/dax/super.c > index c4c4728a36e4..2e1a35e82fce 100644 > --- a/drivers/dax/super.c > +++ b/drivers/dax/super.c > @@ -323,7 +323,8 @@ void kill_dax(struct dax_device *dax_dev) > return; > > if (dax_dev->holder_data != NULL) > - dax_holder_notify_failure(dax_dev, 0, U64_MAX, 0); > + dax_holder_notify_failure(dax_dev, 0, U64_MAX, > + MF_MEM_PRE_REMOVE); > > clear_bit(DAXDEV_ALIVE, &dax_dev->flags); > synchronize_srcu(&dax_srcu); > diff --git a/fs/xfs/xfs_notify_failure.c b/fs/xfs/xfs_notify_failure.c > index 1e2eddb8f90f..796dd954d33a 100644 > --- a/fs/xfs/xfs_notify_failure.c > +++ b/fs/xfs/xfs_notify_failure.c > @@ -22,6 +22,7 @@ > > #include > #include > +#include > > struct xfs_failure_info { > xfs_agblock_t startblock; > @@ -73,10 +74,16 @@ xfs_dax_failure_fn( > struct xfs_mount *mp = cur->bc_mp; > struct xfs_inode *ip; > struct xfs_failure_info *notify = data; > + struct address_space *mapping; > + pgoff_t pgoff; > + unsigned long pgcnt; > int error = 0; > > if (XFS_RMAP_NON_INODE_OWNER(rec->rm_owner) || > (rec->rm_flags & (XFS_RMAP_ATTR_FORK | XFS_RMAP_BMBT_BLOCK))) { > + /* The device is about to be removed. Not a really failure. */ > + if (notify->mf_flags & MF_MEM_PRE_REMOVE) > + return 0; > notify->want_shutdown = true; > return 0; > } > @@ -92,14 +99,120 @@ xfs_dax_failure_fn( > return 0; > } > > - error = mf_dax_kill_procs(VFS_I(ip)->i_mapping, > - xfs_failure_pgoff(mp, rec, notify), > - xfs_failure_pgcnt(mp, rec, notify), > - notify->mf_flags); > + mapping = VFS_I(ip)->i_mapping; > + pgoff = xfs_failure_pgoff(mp, rec, notify); > + pgcnt = xfs_failure_pgcnt(mp, rec, notify); > + > + /* Continue the rmap query if the inode isn't a dax file. */ > + if (dax_mapping(mapping)) > + error = mf_dax_kill_procs(mapping, pgoff, pgcnt, > + notify->mf_flags); > + > + /* Invalidate the cache anyway. */ > + invalidate_inode_pages2_range(mapping, pgoff, pgoff + pgcnt - 1); > + > xfs_irele(ip); > return error; > } > > +#define SB_FREEZE_EXCLUSIVE (SB_FREEZE_COMPLETE + 2) > + > +static int > +xfs_dax_notify_failure_freeze( > + struct xfs_mount *mp) > +{ > + struct super_block *sb = mp->m_super; > + int error = 0; > + int level; > + > + /* Wait until we're ready to freeze. */ > + down_write(&sb->s_umount); > + while (sb->s_writers.frozen != SB_UNFROZEN) { > + up_write(&sb->s_umount); > + > + // just wait, or print warning in dmesg then return -EBUSY? > + > + delay(HZ / 10); > + down_write(&sb->s_umount); > + } > + > + if (sb_rdonly(sb)) { > + sb->s_writers.frozen = SB_FREEZE_EXCLUSIVE; > + goto out; > + } > + > + sb->s_writers.frozen = SB_FREEZE_WRITE; > + /* Release s_umount to preserve sb_start_write -> s_umount ordering */ > + up_write(&sb->s_umount); > + percpu_down_write(sb->s_writers.rw_sem + SB_FREEZE_WRITE - 1); > + down_write(&sb->s_umount); > + > + /* Now we go and block page faults... */ > + sb->s_writers.frozen = SB_FREEZE_PAGEFAULT; > + percpu_down_write(sb->s_writers.rw_sem + SB_FREEZE_PAGEFAULT - 1); > + > + /* All writers are done so after syncing there won't be dirty data */ > + error = sync_filesystem(sb); > + if (error) { > + sb->s_writers.frozen = SB_UNFROZEN; > + for (level = SB_FREEZE_PAGEFAULT - 1; level >= 0; level--) > + percpu_up_write(sb->s_writers.rw_sem + level); > + wake_up(&sb->s_writers.wait_unfrozen); > + goto out; > + } > + > + /* Now wait for internal filesystem counter */ > + sb->s_writers.frozen = SB_FREEZE_FS; > + percpu_down_write(sb->s_writers.rw_sem + SB_FREEZE_FS - 1); > + > + /* > + * To prevent anyone else from unfreezing us, set the VFS freeze level > + * to one higher than SB_FREEZE_COMPLETE. > + */ > + sb->s_writers.frozen = SB_FREEZE_EXCLUSIVE; > + for (level = SB_FREEZE_LEVELS - 1; level >= 0; level--) > + percpu_rwsem_release(sb->s_writers.rw_sem + level, 0, > + _THIS_IP_); > + > +out: > + up_write(&sb->s_umount); > + return error; > +} > + > +static void > +xfs_dax_notify_failure_thaw( > + struct xfs_mount *mp) > +{ > + struct super_block *sb = mp->m_super; > + int level; > + > + down_write(&sb->s_umount); > + if (sb->s_writers.frozen != SB_FREEZE_EXCLUSIVE) { > + /* somebody snuck in and unfroze us? */ > + ASSERT(0); > + up_write(&sb->s_umount); > + return; > + } > + > + if (sb_rdonly(sb)) { > + sb->s_writers.frozen = SB_UNFROZEN; > + goto out; > + } > + > + for (level = 0; level < SB_FREEZE_LEVELS; ++level) > + percpu_rwsem_acquire(sb->s_writers.rw_sem + level, 0, > + _THIS_IP_); > + > + sb->s_writers.frozen = SB_UNFROZEN; > + for (level = SB_FREEZE_LEVELS - 1; level >= 0; level--) > + percpu_up_write(sb->s_writers.rw_sem + level); > + > +out: > + wake_up(&sb->s_writers.wait_unfrozen); > + up_write(&sb->s_umount); > +} > + > static int > xfs_dax_notify_ddev_failure( > struct xfs_mount *mp, > @@ -164,11 +277,22 @@ xfs_dax_notify_ddev_failure( > } > > xfs_trans_cancel(tp); > + > + /* Thaw the fs if it is freezed before. */ > + if (mf_flags & MF_MEM_PRE_REMOVE) > + xfs_dax_notify_failure_thaw(mp); > + > + /* > + * Determine how to shutdown the filesystem according to the > + * error code and flags. > + */ > if (error || notify.want_shutdown) { > xfs_force_shutdown(mp, SHUTDOWN_CORRUPT_ONDISK); > if (!error) > error = -EFSCORRUPTED; > - } > + } else if (mf_flags & MF_MEM_PRE_REMOVE) > + xfs_force_shutdown(mp, SHUTDOWN_FORCE_UMOUNT); > + > return error; > } > > @@ -182,6 +306,7 @@ xfs_dax_notify_failure( > struct xfs_mount *mp = dax_holder(dax_dev); > u64 ddev_start; > u64 ddev_end; > + int error; > > if (!(mp->m_super->s_flags & SB_BORN)) { > xfs_warn(mp, "filesystem is not ready for notify_failure()!"); > @@ -196,6 +321,8 @@ xfs_dax_notify_failure( > > if (mp->m_logdev_targp && mp->m_logdev_targp->bt_daxdev == dax_dev && > mp->m_logdev_targp != mp->m_ddev_targp) { > + if (mf_flags & MF_MEM_PRE_REMOVE) > + return 0; > xfs_err(mp, "ondisk log corrupt, shutting down fs!"); > xfs_force_shutdown(mp, SHUTDOWN_CORRUPT_ONDISK); > return -EFSCORRUPTED; > @@ -209,6 +336,12 @@ xfs_dax_notify_failure( > ddev_start = mp->m_ddev_targp->bt_dax_part_off; > ddev_end = ddev_start + bdev_nr_bytes(mp->m_ddev_targp->bt_bdev) - 1; > > + /* Notify failure on the whole device. */ > + if (offset == 0 && len == U64_MAX) { > + offset = ddev_start; > + len = bdev_nr_bytes(mp->m_ddev_targp->bt_bdev); > + } > + > /* Ignore the range out of filesystem area */ > if (offset + len - 1 < ddev_start) > return -ENXIO; > @@ -225,6 +358,14 @@ xfs_dax_notify_failure( > if (offset + len - 1 > ddev_end) > len = ddev_end - offset + 1; > > + if (mf_flags & MF_MEM_PRE_REMOVE) { > + xfs_info(mp, "device is about to be removed!"); > + /* Freeze fs to prevent new mappings from being created. */ > + error = xfs_dax_notify_failure_freeze(mp); > + if (error) > + return error; > + } > + > return xfs_dax_notify_ddev_failure(mp, BTOBB(offset), BTOBB(len), > mf_flags); > } > diff --git a/include/linux/mm.h b/include/linux/mm.h > index 1f79667824eb..ac3f22c20e1d 100644 > --- a/include/linux/mm.h > +++ b/include/linux/mm.h > @@ -3436,6 +3436,7 @@ enum mf_flags { > MF_UNPOISON = 1 << 4, > MF_SW_SIMULATED = 1 << 5, > MF_NO_RETRY = 1 << 6, > + MF_MEM_PRE_REMOVE = 1 << 7, > }; > int mf_dax_kill_procs(struct address_space *mapping, pgoff_t index, > unsigned long count, int mf_flags); > diff --git a/mm/memory-failure.c b/mm/memory-failure.c > index fae9baf3be16..6e6acec45568 100644 > --- a/mm/memory-failure.c > +++ b/mm/memory-failure.c > @@ -623,7 +623,7 @@ static void collect_procs_file(struct page *page, struct list_head *to_kill, > */ > static void collect_procs_fsdax(struct page *page, > struct address_space *mapping, pgoff_t pgoff, > - struct list_head *to_kill) > + struct list_head *to_kill, bool pre_remove) > { > struct vm_area_struct *vma; > struct task_struct *tsk; > @@ -631,8 +631,15 @@ static void collect_procs_fsdax(struct page *page, > i_mmap_lock_read(mapping); > read_lock(&tasklist_lock); > for_each_process(tsk) { > - struct task_struct *t = task_early_kill(tsk, true); > + struct task_struct *t = tsk; > > + /* > + * Search for all tasks while MF_MEM_PRE_REMOVE, because the > + * current may not be the one accessing the fsdax page. > + * Otherwise, search for the current task. > + */ > + if (!pre_remove) > + t = task_early_kill(tsk, true); > if (!t) > continue; > vma_interval_tree_foreach(vma, &mapping->i_mmap, pgoff, pgoff) { > @@ -1732,6 +1739,7 @@ int mf_dax_kill_procs(struct address_space *mapping, pgoff_t index, > dax_entry_t cookie; > struct page *page; > size_t end = index + count; > + bool pre_remove = mf_flags & MF_MEM_PRE_REMOVE; > > mf_flags |= MF_ACTION_REQUIRED | MF_MUST_KILL; > > @@ -1743,9 +1751,10 @@ int mf_dax_kill_procs(struct address_space *mapping, pgoff_t index, > if (!page) > goto unlock; > > - SetPageHWPoison(page); > + if (!pre_remove) > + SetPageHWPoison(page); > > - collect_procs_fsdax(page, mapping, index, &to_kill); > + collect_procs_fsdax(page, mapping, index, &to_kill, pre_remove); > unmap_and_kill(&to_kill, page_to_pfn(page), mapping, > index, mf_flags); > unlock: