From: Sourav Panda <souravpanda@google.com>
To: David Rientjes <rientjes@google.com>
Cc: Kent Overstreet <kent.overstreet@linux.dev>,
Casey Chen <cachen@purestorage.com>,
Andrew Morton <akpm@linux-foundation.org>,
surenb@google.com, corbet@lwn.net, dennis@kernel.org,
tj@kernel.org, cl@gentwo.org, Vlastimil Babka <vbabka@suse.cz>,
mhocko@suse.com, jackmanb@google.com, hannes@cmpxchg.org,
ziy@nvidia.com, roman.gushchin@linux.dev, harry.yoo@oracle.com,
linux-mm@kvack.org, linux-kernel@vger.kernel.org,
linux-doc@vger.kernel.org, yzhong@purestorage.com
Subject: Re: [PATCH] alloc_tag: add per-NUMA node stats
Date: Wed, 9 Jul 2025 22:54:00 -0700 [thread overview]
Message-ID: <CANruzcRE_EtczNpwus65tpfqLuwSEU+xyRG-hs10PAsj=cP4ow@mail.gmail.com> (raw)
In-Reply-To: <3c9b5773-83ed-4f13-11a8-fcc162c8c483@google.com>
On Tue, Jul 8, 2025 at 2:52 PM David Rientjes <rientjes@google.com> wrote:
>
> On Wed, 18 Jun 2025, Kent Overstreet wrote:
>
> > On Tue, Jun 10, 2025 at 05:30:53PM -0600, Casey Chen wrote:
> > > Add support for tracking per-NUMA node statistics in /proc/allocinfo.
> > > Previously, each alloc_tag had a single set of counters (bytes and
> > > calls), aggregated across all CPUs. With this change, each CPU can
> > > maintain separate counters for each NUMA node, allowing finer-grained
> > > memory allocation profiling.
> > >
> > > This feature is controlled by the new
> > > CONFIG_MEM_ALLOC_PROFILING_PER_NUMA_STATS option:
> > >
> > > * When enabled (=y), the output includes per-node statistics following
> > > the total bytes/calls:
> > >
> > > <size> <calls> <tag info>
> > > ...
> > > 315456 9858 mm/dmapool.c:338 func:pool_alloc_page
> > > nid0 94912 2966
> > > nid1 220544 6892
> > > 7680 60 mm/dmapool.c:254 func:dma_pool_create
> > > nid0 4224 33
> > > nid1 3456 27
> >
> > I just received a report of memory reclaim issues where it seems DMA32
> > is stuffed full.
> >
> > So naturally, instrumenting to see what's consuming DMA32 is going to be
> > the first thing to do, which made me think of your patchset.
> >
> > I wonder if we should think about something a bit more general, so it's
> > easy to break out accounting different ways depending on what we want to
> > debug.
> >
>
> Right, per-node memory attribution, or per zone, is very useful.
>
> Casey, what's the latest status of your patch? Using alloc_tag for
> attributing memory overheads has been exceedingly useful for Google Cloud
> and adding better insight it for per-node breakdown would be even better.
>
> Our use case is quite simple: we sell guest memory to the customer as
> persistent hugetlb and keep some memory on the host for ourselves (VMM,
> host userspace, host kernel). We track every page of that overhead memory
> because memory pressure here can cause all sorts of issues like userspace
> unresponsiveness. We also want to sell as much guest memory as possible
> to avoid stranding cpus.
>
> To do that, per-node breakdown of memory allocations would be a tremendous
> help. We have memory that is asymmetric for NUMA, even for memory that
> has affinity to the NIC. Being able to inspect the origins of memory for
> a specific NUMA node that is under memory pressure where other NUMA nodes
> are not under memory pressure would be excellent.
>
> Adding Sourav Panda as well as he may have additional thoughts on this.
I agree with David, especially the point regarding NIC affinity. I was
dealing with a similar bug today, but pertaining to SSD where this
patchset would have helped in the investigation.
That being said, I think pgalloc_tag_swap() has to be modified as
well, which gets called by __migrate_folio().
next prev parent reply other threads:[~2025-07-10 5:54 UTC|newest]
Thread overview: 15+ messages / expand[flat|nested] mbox.gz Atom feed top
2025-06-10 23:30 Casey Chen
2025-06-11 1:21 ` Andrew Morton
2025-06-11 1:33 ` Casey Chen
2025-06-11 3:47 ` Kent Overstreet
2025-06-11 3:41 ` Kent Overstreet
2025-06-12 5:36 ` David Wang
2025-06-12 15:37 ` Kent Overstreet
2025-06-18 22:16 ` Kent Overstreet
2025-07-08 21:52 ` David Rientjes
2025-07-08 22:38 ` Christoph Lameter (Ampere)
2025-07-09 19:14 ` David Rientjes
2025-07-08 22:53 ` Casey Chen
2025-07-08 23:07 ` Casey Chen
2025-07-10 5:54 ` Sourav Panda [this message]
-- strict thread matches above, loose matches on Subject: below --
2025-05-30 0:39 [PATCH 0/1] alloc_tag: add per-numa " Casey Chen
2025-05-30 0:39 ` [PATCH] " Casey Chen
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to='CANruzcRE_EtczNpwus65tpfqLuwSEU+xyRG-hs10PAsj=cP4ow@mail.gmail.com' \
--to=souravpanda@google.com \
--cc=akpm@linux-foundation.org \
--cc=cachen@purestorage.com \
--cc=cl@gentwo.org \
--cc=corbet@lwn.net \
--cc=dennis@kernel.org \
--cc=hannes@cmpxchg.org \
--cc=harry.yoo@oracle.com \
--cc=jackmanb@google.com \
--cc=kent.overstreet@linux.dev \
--cc=linux-doc@vger.kernel.org \
--cc=linux-kernel@vger.kernel.org \
--cc=linux-mm@kvack.org \
--cc=mhocko@suse.com \
--cc=rientjes@google.com \
--cc=roman.gushchin@linux.dev \
--cc=surenb@google.com \
--cc=tj@kernel.org \
--cc=vbabka@suse.cz \
--cc=yzhong@purestorage.com \
--cc=ziy@nvidia.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox