linux-mm.kvack.org archive mirror
 help / color / mirror / Atom feed
From: zhiguojiang <justinjiang@vivo.com>
To: Barry Song <21cnbao@gmail.com>
Cc: Andrew Morton <akpm@linux-foundation.org>,
	linux-mm@kvack.org, linux-kernel@vger.kernel.org,
	Chris Li <chrisl@kernel.org>,
	opensource.kernel@vivo.com
Subject: Re: [PATCH] mm: swap: mTHP frees entries as a whole
Date: Tue, 6 Aug 2024 10:01:51 +0800	[thread overview]
Message-ID: <3699860f-3887-4a99-b9ef-10e3f86ec3bb@vivo.com> (raw)
In-Reply-To: <CAGsJ_4wqENiGf4FoEKA2yO5pmu3SfJD9qsjHD0E7eHPZG1+PuA@mail.gmail.com>



在 2024/8/6 6:09, Barry Song 写道:
> On Tue, Aug 6, 2024 at 4:08 AM Zhiguo Jiang <justinjiang@vivo.com> wrote:
>> Support mTHP's attempt to free swap entries as a whole, which can avoid
>> frequent swap_info locking for every individual entry in
>> swapcache_free_entries(). When the swap_map count values corresponding
>> to all contiguous entries are all zero excluding SWAP_HAS_CACHE, the
>> entries will be freed directly by skippping percpu swp_slots caches.
>>
> No, this isn't quite good. Please review the work done by Chris and Kairui[1];
> they have handled it better. On a different note, I have a patch that can
> handle zap_pte_range() for swap entries in batches[2][3].
I'm glad to see your optimized submission about batch freeing swap 
entries for
zap_pte_range(), sorry, I didn't see it before. My this patch can be 
ignored.

Thanks
Zhiguo

>
> [1] https://lore.kernel.org/linux-mm/20240730-swap-allocator-v5-5-cb9c148b9297@kernel.org/
> [2] https://lore.kernel.org/linux-mm/20240803091118.84274-1-21cnbao@gmail.com/
> [3] https://lore.kernel.org/linux-mm/CAGsJ_4wPnQqKOHx6iQcwO8bQzoBXKr2qY2AgSxMwTQCj3-8YWw@mail.gmail.com/
>
>> Signed-off-by: Zhiguo Jiang <justinjiang@vivo.com>
>> ---
>>   mm/swapfile.c | 61 +++++++++++++++++++++++++++++++++++++++++++++++++++
>>   1 file changed, 61 insertions(+)
>>
>> diff --git a/mm/swapfile.c b/mm/swapfile.c
>> index ea023fc25d08..829fb4cfb6ec
>> --- a/mm/swapfile.c
>> +++ b/mm/swapfile.c
>> @@ -1493,6 +1493,58 @@ static void swap_entry_range_free(struct swap_info_struct *p, swp_entry_t entry,
>>          swap_range_free(p, offset, nr_pages);
>>   }
>>
>> +/*
>> + * Free the contiguous swap entries as a whole, caller have to
>> + * ensure all entries belong to the same folio.
>> + */
>> +static void swap_entry_range_check_and_free(struct swap_info_struct *p,
>> +                                 swp_entry_t entry, int nr, bool *any_only_cache)
>> +{
>> +       const unsigned long start_offset = swp_offset(entry);
>> +       const unsigned long end_offset = start_offset + nr;
>> +       unsigned long offset;
>> +       DECLARE_BITMAP(to_free, SWAPFILE_CLUSTER) = { 0 };
>> +       struct swap_cluster_info *ci;
>> +       int i = 0, nr_setbits = 0;
>> +       unsigned char count;
>> +
>> +       /*
>> +        * Free and check swap_map count values corresponding to all contiguous
>> +        * entries in the whole folio range.
>> +        */
>> +       WARN_ON_ONCE(nr > SWAPFILE_CLUSTER);
>> +       ci = lock_cluster_or_swap_info(p, start_offset);
>> +       for (offset = start_offset; offset < end_offset; offset++, i++) {
>> +               if (data_race(p->swap_map[offset])) {
>> +                       count = __swap_entry_free_locked(p, offset, 1);
>> +                       if (!count) {
>> +                               bitmap_set(to_free, i, 1);
>> +                               nr_setbits++;
>> +                       } else if (count == SWAP_HAS_CACHE) {
>> +                               *any_only_cache = true;
>> +                       }
>> +               } else {
>> +                       WARN_ON_ONCE(1);
>> +               }
>> +       }
>> +       unlock_cluster_or_swap_info(p, ci);
>> +
>> +       /*
>> +        * If the swap_map count values corresponding to all contiguous entries are
>> +        * all zero excluding SWAP_HAS_CACHE, the entries will be freed directly by
>> +        * skippping percpu swp_slots caches, which can avoid frequent swap_info
>> +        * locking for every individual entry.
>> +        */
>> +       if (nr > 1 && nr_setbits == nr) {
>> +               spin_lock(&p->lock);
>> +               swap_entry_range_free(p, entry, nr);
>> +               spin_unlock(&p->lock);
>> +       } else {
>> +               for_each_set_bit(i, to_free, SWAPFILE_CLUSTER)
>> +                       free_swap_slot(swp_entry(p->type, start_offset + i));
>> +       }
>> +}
>> +
>>   static void cluster_swap_free_nr(struct swap_info_struct *sis,
>>                  unsigned long offset, int nr_pages,
>>                  unsigned char usage)
>> @@ -1808,6 +1860,14 @@ void free_swap_and_cache_nr(swp_entry_t entry, int nr)
>>          if (WARN_ON(end_offset > si->max))
>>                  goto out;
>>
>> +       /*
>> +        * Try to free all contiguous entries about mTHP as a whole.
>> +        */
>> +       if (IS_ENABLED(CONFIG_THP_SWAP) && nr > 1) {
>> +               swap_entry_range_check_and_free(si, entry, nr, &any_only_cache);
>> +               goto free_cache;
>> +       }
>> +
>>          /*
>>           * First free all entries in the range.
>>           */
>> @@ -1821,6 +1881,7 @@ void free_swap_and_cache_nr(swp_entry_t entry, int nr)
>>                  }
>>          }
>>
>> +free_cache:
>>          /*
>>           * Short-circuit the below loop if none of the entries had their
>>           * reference drop to zero.
>> --
>> 2.39.0
>>



  reply	other threads:[~2024-08-06  2:02 UTC|newest]

Thread overview: 7+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2024-08-05 16:07 Zhiguo Jiang
2024-08-05 22:09 ` Barry Song
2024-08-06  2:01   ` zhiguojiang [this message]
2024-08-06  2:07     ` Barry Song
2024-08-06  7:40       ` zhiguojiang
2024-08-06  6:48         ` Barry Song
2024-08-06  8:12           ` zhiguojiang

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=3699860f-3887-4a99-b9ef-10e3f86ec3bb@vivo.com \
    --to=justinjiang@vivo.com \
    --cc=21cnbao@gmail.com \
    --cc=akpm@linux-foundation.org \
    --cc=chrisl@kernel.org \
    --cc=linux-kernel@vger.kernel.org \
    --cc=linux-mm@kvack.org \
    --cc=opensource.kernel@vivo.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox