From: Muchun Song <songmuchun@bytedance.com>
To: Shakeel Butt <shakeelb@google.com>
Cc: Johannes Weiner <hannes@cmpxchg.org>,
Michal Hocko <mhocko@kernel.org>,
Vladimir Davydov <vdavydov.dev@gmail.com>,
Andrew Morton <akpm@linux-foundation.org>,
Cgroups <cgroups@vger.kernel.org>, Linux MM <linux-mm@kvack.org>,
LKML <linux-kernel@vger.kernel.org>
Subject: Re: [External] Re: [PATCH] mm: memcontrol: Add the missing numa stat of anon and file for cgroup v2
Date: Fri, 11 Sep 2020 11:51:42 +0800 [thread overview]
Message-ID: <CAMZfGtUncFZZe2RDb54ALMt1DAVqBxAD1TQrpjei1H5nRbH6jg@mail.gmail.com> (raw)
In-Reply-To: <CALvZod5JQWGHUAPnj9S0pKFQreLPST441mZnp+h=fue_nnh1yQ@mail.gmail.com>
On Fri, Sep 11, 2020 at 12:02 AM Shakeel Butt <shakeelb@google.com> wrote:
>
> On Thu, Sep 10, 2020 at 1:46 AM Muchun Song <songmuchun@bytedance.com> wrote:
> >
> > In the cgroup v1, we have a numa_stat interface. This is useful for
> > providing visibility into the numa locality information within an
> > memcg since the pages are allowed to be allocated from any physical
> > node. One of the use cases is evaluating application performance by
> > combining this information with the application's CPU allocation.
> > But the cgroup v2 does not. So this patch adds the missing information.
> >
> > Signed-off-by: Muchun Song <songmuchun@bytedance.com>
> > ---
>
> I am actually working on exposing this info on v2 as well.
>
> > mm/memcontrol.c | 46 ++++++++++++++++++++++++++++++++++++++++++++--
> > 1 file changed, 44 insertions(+), 2 deletions(-)
> >
> > diff --git a/mm/memcontrol.c b/mm/memcontrol.c
> > index 75cd1a1e66c8..c779673f29b2 100644
> > --- a/mm/memcontrol.c
> > +++ b/mm/memcontrol.c
> > @@ -1492,10 +1492,34 @@ static bool mem_cgroup_wait_acct_move(struct mem_cgroup *memcg)
> > return false;
> > }
> >
> > +#ifdef CONFIG_NUMA
> > +static unsigned long memcg_node_page_state(struct mem_cgroup *memcg,
> > + unsigned int nid,
> > + enum node_stat_item idx)
> > +{
> > + long x;
> > + struct mem_cgroup_per_node *pn;
> > + struct lruvec *lruvec = mem_cgroup_lruvec(memcg, NODE_DATA(nid));
> > +
> > + VM_BUG_ON(nid >= nr_node_ids);
> > +
> > + pn = container_of(lruvec, struct mem_cgroup_per_node, lruvec);
> > + x = atomic_long_read(&pn->lruvec_stat[idx]);
> > +#ifdef CONFIG_SMP
> > + if (x < 0)
> > + x = 0;
> > +#endif
> > + return x;
> > +}
> > +#endif
> > +
> > static char *memory_stat_format(struct mem_cgroup *memcg)
> > {
> > struct seq_buf s;
> > int i;
> > +#ifdef CONFIG_NUMA
> > + int nid;
> > +#endif
> >
> > seq_buf_init(&s, kmalloc(PAGE_SIZE, GFP_KERNEL), PAGE_SIZE);
> > if (!s.buffer)
> > @@ -1512,12 +1536,30 @@ static char *memory_stat_format(struct mem_cgroup *memcg)
> > * Current memory state:
> > */
> >
>
> Let's not break the parsers of memory.stat. I would prefer a separate
> interface like v1 i.e. memory.numa_stat.
It is also a good idea to expose a new interface like memory.numa_stat.
>
> > - seq_buf_printf(&s, "anon %llu\n",
> > + seq_buf_printf(&s, "anon %llu",
> > (u64)memcg_page_state(memcg, NR_ANON_MAPPED) *
> > PAGE_SIZE);
> > - seq_buf_printf(&s, "file %llu\n",
> > +#ifdef CONFIG_NUMA
> > + for_each_node_state(nid, N_MEMORY)
> > + seq_buf_printf(&s, " N%d=%llu", nid,
> > + (u64)memcg_node_page_state(memcg, nid,
> > + NR_ANON_MAPPED) *
> > + PAGE_SIZE);
> > +#endif
> > + seq_buf_putc(&s, '\n');
> > +
> > + seq_buf_printf(&s, "file %llu",
> > (u64)memcg_page_state(memcg, NR_FILE_PAGES) *
> > PAGE_SIZE);
> > +#ifdef CONFIG_NUMA
> > + for_each_node_state(nid, N_MEMORY)
> > + seq_buf_printf(&s, " N%d=%llu", nid,
> > + (u64)memcg_node_page_state(memcg, nid,
> > + NR_FILE_PAGES) *
> > + PAGE_SIZE);
> > +#endif
> > + seq_buf_putc(&s, '\n');
> > +
>
> The v1's numa_stat exposes the LRUs, why NR_ANON_MAPPED and NR_FILE_PAGES?
If we want to expose the anon per node, we need to add inactive anon and
active anon together. Why not use NR_ANON_MAPPED directly?
>
> Also I think exposing slab_[un]reclaimable per node would be beneficial as well.
Yeah, I agree with you. Maybe kernel_stack and percpu also should
be exposed.
>
> > seq_buf_printf(&s, "kernel_stack %llu\n",
> > (u64)memcg_page_state(memcg, NR_KERNEL_STACK_KB) *
> > 1024);
> > --
> > 2.20.1
> >
--
Yours,
Muchun
next prev parent reply other threads:[~2020-09-11 3:52 UTC|newest]
Thread overview: 7+ messages / expand[flat|nested] mbox.gz Atom feed top
2020-09-10 8:42 Muchun Song
2020-09-10 16:01 ` Shakeel Butt
2020-09-11 3:51 ` Muchun Song [this message]
2020-09-11 14:55 ` [External] " Shakeel Butt
2020-09-11 15:47 ` Muchun Song
2020-09-11 15:55 ` Shakeel Butt
2020-09-11 21:51 ` Roman Gushchin
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=CAMZfGtUncFZZe2RDb54ALMt1DAVqBxAD1TQrpjei1H5nRbH6jg@mail.gmail.com \
--to=songmuchun@bytedance.com \
--cc=akpm@linux-foundation.org \
--cc=cgroups@vger.kernel.org \
--cc=hannes@cmpxchg.org \
--cc=linux-kernel@vger.kernel.org \
--cc=linux-mm@kvack.org \
--cc=mhocko@kernel.org \
--cc=shakeelb@google.com \
--cc=vdavydov.dev@gmail.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox