From: SeongJae Park <sj38.park@gmail.com>
To: Shakeel Butt <shakeelb@google.com>
Cc: SeongJae Park <sj38.park@gmail.com>,
SeongJae Park <sjpark@amazon.de>,
Jonathan.Cameron@huawei.com, acme@kernel.org,
alexander.shishkin@linux.intel.com, amit@kernel.org,
benh@kernel.crashing.org,
Brendan Higgins <brendanhiggins@google.com>,
Jonathan Corbet <corbet@lwn.net>,
David Hildenbrand <david@redhat.com>,
dwmw@amazon.com, Marco Elver <elver@google.com>,
"Du, Fan" <fan.du@intel.com>,
foersleo@amazon.de, greg@kroah.com,
Greg Thelen <gthelen@google.com>,
guoju.fgj@alibaba-inc.com, jgowans@amazon.com,
Mel Gorman <mgorman@suse.de>,
mheyne@amazon.de, Minchan Kim <minchan@kernel.org>,
Ingo Molnar <mingo@redhat.com>,
namhyung@kernel.org,
"Peter Zijlstra (Intel)" <peterz@infradead.org>,
Rik van Riel <riel@surriel.com>,
David Rientjes <rientjes@google.com>,
Steven Rostedt <rostedt@goodmis.org>,
Mike Rapoport <rppt@kernel.org>, Shuah Khan <shuah@kernel.org>,
sieberf@amazon.com, snu@zelle79.org,
Vlastimil Babka <vbabka@suse.cz>,
Vladimir Davydov <vdavydov.dev@gmail.com>,
zgf574564920@gmail.com, linux-damon@amazon.com,
Linux MM <linux-mm@kvack.org>,
linux-doc@vger.kernel.org, LKML <linux-kernel@vger.kernel.org>
Subject: Re: [PATCH v31 05/13] mm/damon: Implement primitives for the virtual memory address spaces
Date: Thu, 24 Jun 2021 10:26:21 +0000 [thread overview]
Message-ID: <20210624102623.24563-3-sjpark@amazon.de> (raw)
In-Reply-To: <20210624102623.24563-1-sjpark@amazon.de>
In-Reply-To: <CALvZod4C9KJPnzOUYhnrkvW_fq+WAKODH5czCLMxLvwpRm2Bxg@mail.gmail.com>
From: SeongJae Park <sjpark@amazon.de>
On Tue, 22 Jun 2021 08:00:58 -0700 Shakeel Butt <shakeelb@google.com> wrote:
> On Mon, Jun 21, 2021 at 1:31 AM SeongJae Park <sj38.park@gmail.com> wrote:
> >
> > From: SeongJae Park <sjpark@amazon.de>
> >
> > This commit introduces a reference implementation of the address space
> > specific low level primitives for the virtual address space, so that
> > users of DAMON can easily monitor the data accesses on virtual address
> > spaces of specific processes by simply configuring the implementation to
> > be used by DAMON.
> >
> > The low level primitives for the fundamental access monitoring are
> > defined in two parts:
> >
> > 1. Identification of the monitoring target address range for the address
> > space.
> > 2. Access check of specific address range in the target space.
> >
> > The reference implementation for the virtual address space does the
> > works as below.
> >
> > PTE Accessed-bit Based Access Check
> > -----------------------------------
> >
> > The implementation uses PTE Accessed-bit for basic access checks. That
> > is, it clears the bit for the next sampling target page and checks
> > whether it is set again after one sampling period. This could disturb
> > the reclaim logic. DAMON uses ``PG_idle`` and ``PG_young`` page flags
> > to solve the conflict, as Idle page tracking does.
> >
> > VMA-based Target Address Range Construction
> > -------------------------------------------
> >
> > Only small parts in the super-huge virtual address space of the
> > processes are mapped to physical memory and accessed. Thus, tracking
> > the unmapped address regions is just wasteful. However, because DAMON
> > can deal with some level of noise using the adaptive regions adjustment
> > mechanism, tracking every mapping is not strictly required but could
> > even incur a high overhead in some cases. That said, too huge unmapped
> > areas inside the monitoring target should be removed to not take the
> > time for the adaptive mechanism.
> >
> > For the reason, this implementation converts the complex mappings to
> > three distinct regions that cover every mapped area of the address
> > space. Also, the two gaps between the three regions are the two biggest
> > unmapped areas in the given address space. The two biggest unmapped
> > areas would be the gap between the heap and the uppermost mmap()-ed
> > region, and the gap between the lowermost mmap()-ed region and the stack
> > in most of the cases. Because these gaps are exceptionally huge in
> > usual address spaces, excluding these will be sufficient to make a
> > reasonable trade-off. Below shows this in detail::
> >
> > <heap>
> > <BIG UNMAPPED REGION 1>
> > <uppermost mmap()-ed region>
> > (small mmap()-ed regions and munmap()-ed regions)
> > <lowermost mmap()-ed region>
> > <BIG UNMAPPED REGION 2>
> > <stack>
> >
> > Signed-off-by: SeongJae Park <sjpark@amazon.de>
> > Reviewed-by: Leonard Foerster <foersleo@amazon.de>
> > Reviewed-by: Fernand Sieber <sieberf@amazon.com>
>
> Couple of nits below and one concern on the default value of
> primitive_update_interval of virtual address space primitive.
> Otherwise looks good to me.
Thank you!
>
> [...]
>
> > +
> > +/*
> > + * Size-evenly split a region into 'nr_pieces' small regions
> > + *
> > + * Returns 0 on success, or negative error code otherwise.
> > + */
> > +static int damon_va_evenly_split_region(struct damon_ctx *ctx,
>
> I don't see ctx being used in this function.
Good point, will remove that from the next spin.
>
> > + struct damon_region *r, unsigned int nr_pieces)
> > +{
> > + unsigned long sz_orig, sz_piece, orig_end;
> > + struct damon_region *n = NULL, *next;
> > + unsigned long start;
> > +
> > + if (!r || !nr_pieces)
> > + return -EINVAL;
> > +
> > + orig_end = r->ar.end;
> > + sz_orig = r->ar.end - r->ar.start;
> > + sz_piece = ALIGN_DOWN(sz_orig / nr_pieces, DAMON_MIN_REGION);
> > +
> > + if (!sz_piece)
> > + return -EINVAL;
> > +
> > + r->ar.end = r->ar.start + sz_piece;
> > + next = damon_next_region(r);
> > + for (start = r->ar.end; start + sz_piece <= orig_end;
> > + start += sz_piece) {
> > + n = damon_new_region(start, start + sz_piece);
> > + if (!n)
> > + return -ENOMEM;
> > + damon_insert_region(n, r, next);
> > + r = n;
> > + }
> > + /* complement last region for possible rounding error */
> > + if (n)
> > + n->ar.end = orig_end;
> > +
> > + return 0;
> > +}
>
> [...]
>
> > +/*
> > + * Get the three regions in the given target (task)
> > + *
> > + * Returns 0 on success, negative error code otherwise.
> > + */
> > +static int damon_va_three_regions(struct damon_target *t,
> > + struct damon_addr_range regions[3])
> > +{
> > + struct mm_struct *mm;
> > + int rc;
> > +
> > + mm = damon_get_mm(t);
> > + if (!mm)
> > + return -EINVAL;
> > +
> > + mmap_read_lock(mm);
> > + rc = __damon_va_three_regions(mm->mmap, regions);
> > + mmap_read_unlock(mm);
>
> This is being called for each target every second by default. Seems
> too aggressive. Applications don't change their address space every
> second. I would recommend to default ctx->primitive_update_interval to
> a higher default value.
Good point. If there are many targets and each target has a huge number of
VMAs, the overhead could be high. Nevertheless, I couldn't find the overhead
in my test setup. Also, it seems someone are already started exploring DAMON
patchset with the default value. and usages from others. Silently changing the
default value could distract such people. So, if you think it's ok, I'd like
to change the default value only after someone finds the overhead from their
usages and asks a change.
If you disagree or you found the overhead from your usage, please feel free to
let me know.
>
> > +
> > + mmput(mm);
> > + return rc;
> > +}
> > +
>
> [...]
>
> > +static void __damon_va_init_regions(struct damon_ctx *c,
>
> Keep the convention of naming damon_ctx ctx.
Ok, I will do so from the next spin.
>
> > + struct damon_target *t)
> > +{
> > + struct damon_region *r;
> > + struct damon_addr_range regions[3];
> > + unsigned long sz = 0, nr_pieces;
> > + int i;
> > +
> > + if (damon_va_three_regions(t, regions)) {
> > + pr_err("Failed to get three regions of target %lu\n", t->id);
> > + return;
> > + }
> > +
> > + for (i = 0; i < 3; i++)
> > + sz += regions[i].end - regions[i].start;
> > + if (c->min_nr_regions)
> > + sz /= c->min_nr_regions;
> > + if (sz < DAMON_MIN_REGION)
> > + sz = DAMON_MIN_REGION;
> > +
> > + /* Set the initial three regions of the target */
> > + for (i = 0; i < 3; i++) {
> > + r = damon_new_region(regions[i].start, regions[i].end);
> > + if (!r) {
> > + pr_err("%d'th init region creation failed\n", i);
> > + return;
> > + }
> > + damon_add_region(r, t);
> > +
> > + nr_pieces = (regions[i].end - regions[i].start) / sz;
> > + damon_va_evenly_split_region(c, r, nr_pieces);
> > + }
> > +}
>
> [...]
>
> > +/*
> > + * Update damon regions for the three big regions of the given target
> > + *
> > + * t the given target
> > + * bregions the three big regions of the target
> > + */
> > +static void damon_va_apply_three_regions(struct damon_ctx *ctx,
>
> ctx not used in this function.
Good eye, will remove that from the next version.
>
>
> > + struct damon_target *t, struct damon_addr_range bregions[3])
> > +{
> > + struct damon_region *r, *next;
> > + unsigned int i = 0;
> > +
> > + /* Remove regions which are not in the three big regions now */
> > + damon_for_each_region_safe(r, next, t) {
> > + for (i = 0; i < 3; i++) {
> > + if (damon_intersect(r, &bregions[i]))
> > + break;
> > + }
> > + if (i == 3)
> > + damon_destroy_region(r);
> > + }
> > +
> > + /* Adjust intersecting regions to fit with the three big regions */
> > + for (i = 0; i < 3; i++) {
> > + struct damon_region *first = NULL, *last;
> > + struct damon_region *newr;
> > + struct damon_addr_range *br;
> > +
> > + br = &bregions[i];
> > + /* Get the first and last regions which intersects with br */
> > + damon_for_each_region(r, t) {
> > + if (damon_intersect(r, br)) {
> > + if (!first)
> > + first = r;
> > + last = r;
> > + }
> > + if (r->ar.start >= br->end)
> > + break;
> > + }
> > + if (!first) {
> > + /* no damon_region intersects with this big region */
> > + newr = damon_new_region(
> > + ALIGN_DOWN(br->start,
> > + DAMON_MIN_REGION),
> > + ALIGN(br->end, DAMON_MIN_REGION));
> > + if (!newr)
> > + continue;
> > + damon_insert_region(newr, damon_prev_region(r), r);
> > + } else {
> > + first->ar.start = ALIGN_DOWN(br->start,
> > + DAMON_MIN_REGION);
> > + last->ar.end = ALIGN(br->end, DAMON_MIN_REGION);
> > + }
> > + }
> > +}
Thanks,
SeongJae Park
next prev parent reply other threads:[~2021-06-24 10:26 UTC|newest]
Thread overview: 34+ messages / expand[flat|nested] mbox.gz Atom feed top
2021-06-21 8:30 [PATCH v31 00/13] Introduce Data Access MONitor (DAMON) SeongJae Park
2021-06-21 8:30 ` [PATCH v31 01/13] mm: " SeongJae Park
2021-06-22 14:59 ` Shakeel Butt
2021-06-24 10:26 ` SeongJae Park
2021-06-24 14:34 ` Shakeel Butt
2021-06-21 8:30 ` [PATCH v31 02/13] mm/damon/core: Implement region-based sampling SeongJae Park
2021-06-22 14:59 ` Shakeel Butt
2021-06-21 8:30 ` [PATCH v31 03/13] mm/damon: Adaptively adjust regions SeongJae Park
2021-06-22 14:59 ` Shakeel Butt
2021-06-24 10:26 ` SeongJae Park
2021-06-21 8:30 ` [PATCH v31 04/13] mm/idle_page_tracking: Make PG_idle reusable SeongJae Park
2021-06-21 8:31 ` [PATCH v31 05/13] mm/damon: Implement primitives for the virtual memory address spaces SeongJae Park
2021-06-22 15:00 ` Shakeel Butt
2021-06-24 10:26 ` SeongJae Park [this message]
2021-06-24 14:42 ` Shakeel Butt
2021-06-24 15:21 ` SeongJae Park
2021-06-24 16:33 ` Shakeel Butt
2021-06-24 17:38 ` SeongJae Park
2021-07-01 0:18 ` Shakeel Butt
2021-07-01 0:19 ` Shakeel Butt
2021-06-21 8:31 ` [PATCH v31 06/13] mm/damon: Add a tracepoint SeongJae Park
2021-06-22 15:01 ` Shakeel Butt
2021-06-21 8:31 ` [PATCH v31 07/13] mm/damon: Implement a debugfs-based user space interface SeongJae Park
2021-06-22 18:12 ` Shakeel Butt
2021-06-24 10:26 ` SeongJae Park
2021-06-24 14:52 ` Shakeel Butt
2021-06-21 8:31 ` [PATCH v31 08/13] mm/damon/dbgfs: Export kdamond pid to the user space SeongJae Park
2021-06-22 18:23 ` Shakeel Butt
2021-06-24 10:26 ` SeongJae Park
2021-06-21 8:31 ` [PATCH v31 09/13] mm/damon/dbgfs: Support multiple contexts SeongJae Park
2021-06-21 8:31 ` [PATCH v31 10/13] Documentation: Add documents for DAMON SeongJae Park
2021-06-21 8:31 ` [PATCH v31 11/13] mm/damon: Add kunit tests SeongJae Park
2021-06-21 8:31 ` [PATCH v31 12/13] mm/damon: Add user space selftests SeongJae Park
2021-06-21 8:31 ` [PATCH v31 13/13] MAINTAINERS: Update for DAMON SeongJae Park
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20210624102623.24563-3-sjpark@amazon.de \
--to=sj38.park@gmail.com \
--cc=Jonathan.Cameron@huawei.com \
--cc=acme@kernel.org \
--cc=alexander.shishkin@linux.intel.com \
--cc=amit@kernel.org \
--cc=benh@kernel.crashing.org \
--cc=brendanhiggins@google.com \
--cc=corbet@lwn.net \
--cc=david@redhat.com \
--cc=dwmw@amazon.com \
--cc=elver@google.com \
--cc=fan.du@intel.com \
--cc=foersleo@amazon.de \
--cc=greg@kroah.com \
--cc=gthelen@google.com \
--cc=guoju.fgj@alibaba-inc.com \
--cc=jgowans@amazon.com \
--cc=linux-damon@amazon.com \
--cc=linux-doc@vger.kernel.org \
--cc=linux-kernel@vger.kernel.org \
--cc=linux-mm@kvack.org \
--cc=mgorman@suse.de \
--cc=mheyne@amazon.de \
--cc=minchan@kernel.org \
--cc=mingo@redhat.com \
--cc=namhyung@kernel.org \
--cc=peterz@infradead.org \
--cc=riel@surriel.com \
--cc=rientjes@google.com \
--cc=rostedt@goodmis.org \
--cc=rppt@kernel.org \
--cc=shakeelb@google.com \
--cc=shuah@kernel.org \
--cc=sieberf@amazon.com \
--cc=sjpark@amazon.de \
--cc=snu@zelle79.org \
--cc=vbabka@suse.cz \
--cc=vdavydov.dev@gmail.com \
--cc=zgf574564920@gmail.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox