From: David Hildenbrand <david@redhat.com>
To: Dong Aisheng <aisheng.dong@nxp.com>, linux-mm@kvack.org
Cc: linux-kernel@vger.kernel.org,
linux-arm-kernel@lists.infradead.org, dongas86@gmail.com,
jason.hui.liu@nxp.com, leoyang.li@nxp.com, abel.vesa@nxp.com,
shawnguo@kernel.org, linux-imx@nxp.com,
akpm@linux-foundation.org, m.szyprowski@samsung.com,
lecopzer.chen@mediatek.com, vbabka@suse.cz,
stable@vger.kernel.org, shijie.qin@nxp.com
Subject: Re: [PATCH v2 2/2] mm: cma: try next MAX_ORDER_NR_PAGES during retry
Date: Tue, 25 Jan 2022 17:33:49 +0100 [thread overview]
Message-ID: <517e1ea1-f826-228b-16a0-da1dc76017cc@redhat.com> (raw)
In-Reply-To: <20220112131552.3329380-3-aisheng.dong@nxp.com>
On 12.01.22 14:15, Dong Aisheng wrote:
> On an ARMv7 platform with 32M pageblock(MAX_ORDER 14), we observed a
Did you actually intend to talk about pageblocks here (and below)?
I assume you have to be clearer here that you talk about the maximum
allocation granularity, which is usually bigger than actual pageblock size.
> huge number of repeat retries of CMA allocation (1k+) during booting
> when allocating one page for each of 3 mmc instance probe.
>
> This is caused by CMA now supports cocurrent allocation since commit
> a4efc174b382 ("mm/cma.c: remove redundant cma_mutex lock").
> The pageblock or (MAX_ORDER -1) from which we are trying to allocate
> memory may have already been acquired and isolated by others.
> Current cma_alloc() will then retry the next area by the step of
> bitmap_no + mask + 1 which are very likely within the same isolated range
> and fail again. So when the pageblock or MAX_ORDER is big (e.g. 8192),
> keep retrying in a small step become meaningless because it will be known
> to fail at a huge number of times due to the pageblock has been isolated
> by others, especially when allocating only one or two pages.
>
> Instread of looping in the same pageblock and wasting CPU mips a lot,
> especially for big pageblock system (e.g. 16M or 32M),
> we try the next MAX_ORDER_NR_PAGES directly.
>
> Doing this way can greatly mitigate the situtation.
>
> Below is the original error log during booting:
> [ 2.004804] cma: cma_alloc(cma (ptrval), count 1, align 0)
> [ 2.010318] cma: cma_alloc(cma (ptrval), count 1, align 0)
> [ 2.010776] cma: cma_alloc(): memory range at (ptrval) is busy, retrying
> [ 2.010785] cma: cma_alloc(): memory range at (ptrval) is busy, retrying
> [ 2.010793] cma: cma_alloc(): memory range at (ptrval) is busy, retrying
> [ 2.010800] cma: cma_alloc(): memory range at (ptrval) is busy, retrying
> [ 2.010807] cma: cma_alloc(): memory range at (ptrval) is busy, retrying
> [ 2.010814] cma: cma_alloc(): memory range at (ptrval) is busy, retrying
> .... (+1K retries)
>
> After fix, the 1200+ reties can be reduced to 0.
> Another test running 8 VPU decoder in parallel shows that 1500+ retries
> dropped to ~145.
>
> IOW this patch can improve the CMA allocation speed a lot when there're
> enough CMA memory by reducing retries significantly.
>
> Cc: Andrew Morton <akpm@linux-foundation.org>
> Cc: Marek Szyprowski <m.szyprowski@samsung.com>
> Cc: Lecopzer Chen <lecopzer.chen@mediatek.com>
> Cc: David Hildenbrand <david@redhat.com>
> Cc: Vlastimil Babka <vbabka@suse.cz>
> CC: stable@vger.kernel.org # 5.11+
> Fixes: a4efc174b382 ("mm/cma.c: remove redundant cma_mutex lock")
> Signed-off-by: Dong Aisheng <aisheng.dong@nxp.com>
> ---
> v1->v2:
> * change to align with MAX_ORDER_NR_PAGES instead of pageblock_nr_pages
> ---
> mm/cma.c | 4 +++-
> 1 file changed, 3 insertions(+), 1 deletion(-)
>
> diff --git a/mm/cma.c b/mm/cma.c
> index 1c13a729d274..1251f65e2364 100644
> --- a/mm/cma.c
> +++ b/mm/cma.c
> @@ -500,7 +500,9 @@ struct page *cma_alloc(struct cma *cma, unsigned long count,
> trace_cma_alloc_busy_retry(cma->name, pfn, pfn_to_page(pfn),
> count, align);
> /* try again with a bit different memory target */
> - start = bitmap_no + mask + 1;
> + start = ALIGN(bitmap_no + mask + 1,
> + MAX_ORDER_NR_PAGES >> cma->order_per_bit);
Mind giving the reader a hint in the code why we went for
MAX_ORDER_NR_PAGES?
What would happen if the CMA granularity is bigger than
MAX_ORDER_NR_PAGES? I'd assume no harm done, as we'd try aligning to 0.
--
Thanks,
David / dhildenb
next prev parent reply other threads:[~2022-01-25 16:33 UTC|newest]
Thread overview: 5+ messages / expand[flat|nested] mbox.gz Atom feed top
2022-01-12 13:15 [PATCH v2 0/2] mm: fix cma allocation fail sometimes Dong Aisheng
2022-01-12 13:15 ` [PATCH v2 1/2] mm: cma: fix allocation may " Dong Aisheng
2022-01-24 14:34 ` [PATCH v2 0/2] mm: fix cma allocation " Dong Aisheng
[not found] ` <20220112131552.3329380-3-aisheng.dong@nxp.com>
2022-01-25 16:33 ` David Hildenbrand [this message]
2022-01-28 12:20 ` [PATCH v2 2/2] mm: cma: try next MAX_ORDER_NR_PAGES during retry Dong Aisheng
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=517e1ea1-f826-228b-16a0-da1dc76017cc@redhat.com \
--to=david@redhat.com \
--cc=abel.vesa@nxp.com \
--cc=aisheng.dong@nxp.com \
--cc=akpm@linux-foundation.org \
--cc=dongas86@gmail.com \
--cc=jason.hui.liu@nxp.com \
--cc=lecopzer.chen@mediatek.com \
--cc=leoyang.li@nxp.com \
--cc=linux-arm-kernel@lists.infradead.org \
--cc=linux-imx@nxp.com \
--cc=linux-kernel@vger.kernel.org \
--cc=linux-mm@kvack.org \
--cc=m.szyprowski@samsung.com \
--cc=shawnguo@kernel.org \
--cc=shijie.qin@nxp.com \
--cc=stable@vger.kernel.org \
--cc=vbabka@suse.cz \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox