From: Zi Yan <ziy@nvidia.com>
To: Wei Yang <richard.weiyang@gmail.com>
Cc: akpm@linux-foundation.org, david@redhat.com,
lorenzo.stoakes@oracle.com, baolin.wang@linux.alibaba.com,
Liam.Howlett@oracle.com, npache@redhat.com, ryan.roberts@arm.com,
dev.jain@arm.com, baohua@kernel.org, lance.yang@linux.dev,
linux-mm@kvack.org
Subject: Re: [PATCH] mm/huge_memory: merge uniform_split_supported() and non_uniform_split_supported()
Date: Tue, 04 Nov 2025 21:14:31 -0500 [thread overview]
Message-ID: <E1772F62-C1FC-4B67-A47F-7A1A6581C89D@nvidia.com> (raw)
In-Reply-To: <20251104075326.hqktuvois66j3cdk@master>
On 4 Nov 2025, at 2:53, Wei Yang wrote:
> On Mon, Nov 03, 2025 at 09:30:03PM -0500, Zi Yan wrote:
>> On 3 Nov 2025, at 19:36, Wei Yang wrote:
>>
>>> On Mon, Nov 03, 2025 at 11:34:47AM -0500, Zi Yan wrote:
>>>> On 31 Oct 2025, at 22:11, Wei Yang wrote:
>>>>
>>>>> The functions uniform_split_supported() and
>>>>> non_uniform_split_supported() share significantly similar logic.
>>>>>
>>>>> The only functional difference is that uniform_split_supported()
>>>>> includes an additional check on the requested @new_order before
>>>>
>>>> Please elaborate on what the check is for.
>>>>
>>>>> proceeding with further validation.
>>>
>>> How about this:
>>>
>>> The only functional difference is that uniform_split_supported() includes an
>>> additional check on the requested @new_order and split type to confirm support
>>> from file system or swap cache.
>>
>> You are describing what the code does instead of its actual meaning.
>> You need to describe:
>> 1. what is the difference between uniform split and non-uniform split?
>> 2. what order does what file systems support? Only order-0.
>> 3. what order does swap cache support? Only order-0.
>> 4. why can uniform split be used to split large folios from 2 or 3 to
>> order-0?
>> 5. why can non uniform split not be used to split large folios from 2
>> or 3 to order-0?
>> 6. The logic similarity between uniform_split_supported() and
>> non_uniform_split_supported() and they can be combined with detailed
>> comment.
>>
>
> Here is the updated version:
>
> The only functional difference is that uniform_split_supported() includes an
> additional check on the requested @new_order.
>
> The reason for this check comes from the following two aspects:
>
> * some file system or swap cache just supports order-0 folio
> * the behavioral difference between uniform/non-uniform split
>
> The behavioral difference between uniform split and non-uniform:
>
> * uniform split splits folio directly to @new_order
> * non-uniform split creates after-split folios with orders from
> folio_order(folio) - 1 to new_order.
>
> This means for non-uniform split or !new_order split we should check the file
> system and swap cache respectively.
>
LGTM.
>>>
>>>>>
>>>>> This commit unifies the logic by introducing a new variable,
>>>>> @need_check, which is conditionally set based on whether a uniform
>>>>> split is requested. This allows us to merge the two functions into
>>>>> a single, combined helper, removing redundant code and simplifying
>>>>> the split support checking mechanism.
>>>>>
>>>>> Signed-off-by: Wei Yang <richard.weiyang@gmail.com>
>>>>> Cc: Zi Yan <ziy@nvidia.com>
>>>>> ---
>>>>> include/linux/huge_mm.h | 8 +++---
>>>>> mm/huge_memory.c | 55 +++++++++++------------------------------
>>>>> 2 files changed, 18 insertions(+), 45 deletions(-)
>>>>>
>>>>> diff --git a/include/linux/huge_mm.h b/include/linux/huge_mm.h
>>>>> index cbb2243f8e56..79343809a7be 100644
>>>>> --- a/include/linux/huge_mm.h
>>>>> +++ b/include/linux/huge_mm.h
>>>>> @@ -369,10 +369,8 @@ int __split_huge_page_to_list_to_order(struct page *page, struct list_head *list
>>>>> unsigned int new_order, bool unmapped);
>>>>> int min_order_for_split(struct folio *folio);
>>>>> int split_folio_to_list(struct folio *folio, struct list_head *list);
>>>>> -bool uniform_split_supported(struct folio *folio, unsigned int new_order,
>>>>> - bool warns);
>>>>> -bool non_uniform_split_supported(struct folio *folio, unsigned int new_order,
>>>>> - bool warns);
>>>>> +bool folio_split_supported(struct folio *folio, unsigned int new_order,
>>>>> + bool uniform_split, bool warns);
>>>>> int folio_split(struct folio *folio, unsigned int new_order, struct page *page,
>>>>> struct list_head *list);
>>>>>
>>>>> @@ -403,7 +401,7 @@ static inline int split_huge_page_to_order(struct page *page, unsigned int new_o
>>>>> static inline int try_folio_split_to_order(struct folio *folio,
>>>>> struct page *page, unsigned int new_order)
>>>>> {
>>>>> - if (!non_uniform_split_supported(folio, new_order, /* warns= */ false))
>>>>> + if (!folio_split_supported(folio, new_order, /* uniform_split = */ false, /* warns= */ false))
>>>>> return split_huge_page_to_order(&folio->page, new_order);
>>>>> return folio_split(folio, new_order, page, NULL);
>>>>> }
>>>>> diff --git a/mm/huge_memory.c b/mm/huge_memory.c
>>>>> index d1fa0d2d9b44..f6d2cb2a5ca0 100644
>>>>> --- a/mm/huge_memory.c
>>>>> +++ b/mm/huge_memory.c
>>>>> @@ -3673,55 +3673,34 @@ static int __split_unmapped_folio(struct folio *folio, int new_order,
>>>>> return 0;
>>>>> }
>>>>>
>>>>> -bool non_uniform_split_supported(struct folio *folio, unsigned int new_order,
>>>>> - bool warns)
>>>>> +bool folio_split_supported(struct folio *folio, unsigned int new_order,
>>>>> + bool uniform_split, bool warns)
>>>>> {
>>>>> - if (folio_test_anon(folio)) {
>>>>> - /* order-1 is not supported for anonymous THP. */
>>>>> - VM_WARN_ONCE(warns && new_order == 1,
>>>>> - "Cannot split to order-1 folio");
>>>>> - return new_order != 1;
>>>>> - } else if (IS_ENABLED(CONFIG_READ_ONLY_THP_FOR_FS) &&
>>>>> - !mapping_large_folio_support(folio->mapping)) {
>>>>> - /*
>>>>> - * No split if the file system does not support large folio.
>>>>> - * Note that we might still have THPs in such mappings due to
>>>>> - * CONFIG_READ_ONLY_THP_FOR_FS. But in that case, the mapping
>>>>> - * does not actually support large folios properly.
>>>>> - */
>>>>> - VM_WARN_ONCE(warns,
>>>>> - "Cannot split file folio to non-0 order");
>>>>> - return false;
>>>>> - }
>>>>> -
>>>>> - /* Only swapping a whole PMD-mapped folio is supported */
>>>>> - if (folio_test_swapcache(folio)) {
>>>>> - VM_WARN_ONCE(warns,
>>>>> - "Cannot split swapcache folio to non-0 order");
>>>>> - return false;
>>>>> - }
>>>>> + bool need_check = uniform_split ? new_order : true;
>>>>>
>>>>> - return true;
>>>>> -}
>>>>> -
>>>>> -/* See comments in non_uniform_split_supported() */
>>>>> -bool uniform_split_supported(struct folio *folio, unsigned int new_order,
>>>>> - bool warns)
>>>>> -{
>>>>> if (folio_test_anon(folio)) {
>>>>> + /* order-1 is not supported for anonymous THP. */
>>>>> VM_WARN_ONCE(warns && new_order == 1,
>>>>> "Cannot split to order-1 folio");
>>>>> return new_order != 1;
>>>>> - } else if (new_order) {
>>>>> + } else if (need_check) {
>>>>> if (IS_ENABLED(CONFIG_READ_ONLY_THP_FOR_FS) &&
>>>>> !mapping_large_folio_support(folio->mapping)) {
>>>>> + /*
>>>>> + * No split if the file system does not support large
>>>>> + * folio. Note that we might still have THPs in such
>>>>> + * mappings due to CONFIG_READ_ONLY_THP_FOR_FS. But in
>>>>> + * that case, the mapping does not actually support
>>>>> + * large folios properly.
>>>>> + */
>>>>
>>>> Blindly copying the comment here causes fusion. The checks for
>>>> uniform and non uniform look similar but this comment is specific
>>>> for non uniform split. The “No split” only applies to non uniform
>>>> split, but for uniform split as long as order is 0, the folio
>>>> can be split.
>>>>
>>>
>>> Per my understanding, "no split" applies to both uniform/non uniform split
>>> when new_order is not 0.
>>
>> Not exactly. For non uniform split, any new_order value is not allowed.
>>
>>>
>>> So the logic here is:
>>>
>>> * uniform split && !new_order: no more check
>>> * non uniform split: do the check regardless of the new_order
>>>
>>> But I am lack of some background knowledge, if it is wrong, please correct me.
>>
>> You are changing the code, please do your homework first. Or you can
>> ask. After go through the above 6 bullet points, you should get the
>> background knowledge.
>>
>>>
>>>> Please rewrite this comment to clarify both uniform and non uniform
>>>> cases.
>>>
>>> Not sure this one would be better?
>>>
>>> We can always split a folio down to a single page (new_order == 0) directly.
>>
>> Not always, the exceptions are listed below.
>>
>
> I mean uniform split to order-0, maybe above line misleading to non-uniform
> split?
Right. If you mean uniform split, then write uniform split instead.
>
>>>
>>> For any other scenario
>>> * uniform split targeting a large folio (new_order > 0)
>>> * any non-uniform split
>>> we must confirm that the file system supports large folios.
>>>
>>> Note that we might still have THPs in such mappings due to
>>> CONFIG_READ_ONLY_THP_FOR_FS. But in that case, the mapping does not actually
>>> support large folios properly.
>>
>> These filesystems do not support large folios except THPs created from khugepaged when CONFIG_READ_ONLY_THP_FOR_FS is enabled.
>>
>
> Want to confirm to see whether I understand correctly.
>
> We have two kinds of file system:
>
> a) support large folio
> b) not support large folio
>
> For a), we can split large folio to min_order_for_split(), uniform or
> non-uniform.
>
> For b), normally there is no large folio. The large folio is collapsed by
> khugepaged when CONFIG_READ_ONLY_THP_FOR_FS is enabled. So we can only split
> it to order-0 folio for this case.
Exactly.
>
> Not sure this one would be better?
>
> We can always split a folio down to a single page (new_order == 0)
> uniformly.
>
> For any other scenario
> * uniform split targeting a large folio (new_order > 0)
> * any non-uniform split
> we must confirm that the file system supports large folios.
>
> Note that we might still have THPs in such mappings, which is created from
> khugepaged when CONFIG_READ_ONLY_THP_FOR_FS is enabled. But in that case,
> the mapping does not actually support large folios properly.
LGTM.
>
>
>>>>> VM_WARN_ONCE(warns,
>>>>> "Cannot split file folio to non-0 order");
>>>>> return false;
>>>>> }
>>>>> }
>>>>>
>>>>> - if (new_order && folio_test_swapcache(folio)) {
>>>>> + /* Only swapping a whole PMD-mapped folio is supported */
>>>>
>>>> The same issue like the above one. Please rewrite this comment as well.
>>>>
>>>
>>> How about this one:
>>>
>>> swapcache folio could only be split to order 0
>>
>> This looks good.
>>
>>>
>>> For non-uniform split or uniform split targeting a large folio, return
>>> false.
>>
>> You are just describing the code.
>>
>> non-uniform split creates after-split folios with orders from
>> folio_order(folio) - 1 to new_order, making it not suitable for any swapcache
>> folio split. Only uniform split to order-0 can be used here.
>>
>
> Below is the updated version:
>
> swapcache folio could only be split to order 0
>
> non-uniform split creates after-split folios with orders from
> folio_order(folio) - 1 to new_order, making it not suitable for any
> swapcache folio split. Only uniform split to order-0 can be used here.
LGTM.
Thank you for updating the comments. Looking forward to your updated patch.
>
>>>
>>>>> + if (need_check && folio_test_swapcache(folio)) {
>>>>> VM_WARN_ONCE(warns,
>>>>> "Cannot split swapcache folio to non-0 order");
>>>>> return false;
>>>>> @@ -3779,11 +3758,7 @@ static int __folio_split(struct folio *folio, unsigned int new_order,
>>>>> if (new_order >= old_order)
>>>>> return -EINVAL;
>>>>>
>>>>> - if (uniform_split && !uniform_split_supported(folio, new_order, true))
>>>>> - return -EINVAL;
>>>>> -
>>>>> - if (!uniform_split &&
>>>>> - !non_uniform_split_supported(folio, new_order, true))
>>>>> + if (!folio_split_supported(folio, new_order, uniform_split, /* warn = */ true))
>>>>> return -EINVAL;
>>>>>
>>>>> is_hzp = is_huge_zero_folio(folio);
>>>>> --
>>>>> 2.34.1
>>>>
>>>>
>>>> Best Regards,
>>>> Yan, Zi
>>>
>>> --
>>> Wei Yang
>>> Help you, Help me
>>
>>
>> Best Regards,
>> Yan, Zi
>
> --
> Wei Yang
> Help you, Help me
Best Regards,
Yan, Zi
next prev parent reply other threads:[~2025-11-05 2:14 UTC|newest]
Thread overview: 13+ messages / expand[flat|nested] mbox.gz Atom feed top
2025-11-01 2:11 Wei Yang
2025-11-03 9:04 ` Dev Jain
2025-11-03 16:19 ` Zi Yan
2025-11-03 11:50 ` David Hildenbrand (Red Hat)
2025-11-04 0:41 ` Wei Yang
2025-11-04 9:05 ` David Hildenbrand (Red Hat)
2025-11-04 13:31 ` Wei Yang
2025-11-03 16:34 ` Zi Yan
2025-11-04 0:36 ` Wei Yang
2025-11-04 2:30 ` Zi Yan
2025-11-04 7:53 ` Wei Yang
2025-11-05 2:14 ` Zi Yan [this message]
2025-11-05 2:44 ` Wei Yang
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=E1772F62-C1FC-4B67-A47F-7A1A6581C89D@nvidia.com \
--to=ziy@nvidia.com \
--cc=Liam.Howlett@oracle.com \
--cc=akpm@linux-foundation.org \
--cc=baohua@kernel.org \
--cc=baolin.wang@linux.alibaba.com \
--cc=david@redhat.com \
--cc=dev.jain@arm.com \
--cc=lance.yang@linux.dev \
--cc=linux-mm@kvack.org \
--cc=lorenzo.stoakes@oracle.com \
--cc=npache@redhat.com \
--cc=richard.weiyang@gmail.com \
--cc=ryan.roberts@arm.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox