From: Michal Hocko <mhocko@kernel.org>
To: Jesper Dangaard Brouer <netdev@brouer.com>
Cc: Pekka Enberg <penberg@iki.fi>, "Tobin C. Harding" <me@tobin.cc>,
Vlastimil Babka <vbabka@suse.cz>,
"Tobin C. Harding" <tobin@kernel.org>,
Andrew Morton <akpm@linux-foundation.org>,
Christoph Lameter <cl@linux.com>,
Pekka Enberg <penberg@kernel.org>,
David Rientjes <rientjes@google.com>,
Joonsoo Kim <iamjoonsoo.kim@lge.com>, Tejun Heo <tj@kernel.org>,
Qian Cai <cai@lca.pw>,
Linus Torvalds <torvalds@linux-foundation.org>,
linux-mm@kvack.org, linux-kernel@vger.kernel.org,
Mel Gorman <mgorman@techsingularity.net>,
"netdev@vger.kernel.org" <netdev@vger.kernel.org>,
Alexander Duyck <alexander.duyck@gmail.com>
Subject: Re: [PATCH 0/1] mm: Remove the SLAB allocator
Date: Wed, 17 Apr 2019 15:38:52 +0200 [thread overview]
Message-ID: <20190417133852.GL5878@dhcp22.suse.cz> (raw)
In-Reply-To: <20190417105018.78604ad8@carbon>
On Wed 17-04-19 10:50:18, Jesper Dangaard Brouer wrote:
> On Thu, 11 Apr 2019 11:27:26 +0300
> Pekka Enberg <penberg@iki.fi> wrote:
>
> > Hi,
> >
> > On 4/11/19 10:55 AM, Michal Hocko wrote:
> > > Please please have it more rigorous then what happened when SLUB was
> > > forced to become a default
> >
> > This is the hard part.
> >
> > Even if you are able to show that SLUB is as fast as SLAB for all the
> > benchmarks you run, there's bound to be that one workload where SLUB
> > regresses. You will then have people complaining about that (rightly so)
> > and you're again stuck with two allocators.
> >
> > To move forward, I think we should look at possible *pathological* cases
> > where we think SLAB might have an advantage. For example, SLUB had much
> > more difficulties with remote CPU frees than SLAB. Now I don't know if
> > this is the case, but it should be easy to construct a synthetic
> > benchmark to measure this.
>
> I do think SLUB have a number of pathological cases where SLAB is
> faster. If was significantly more difficult to get good bulk-free
> performance for SLUB. SLUB is only fast as long as objects belong to
> the same page. To get good bulk-free performance if objects are
> "mixed", I coded this[1] way-too-complex fast-path code to counter
> act this (joined work with Alex Duyck).
>
> [1] https://github.com/torvalds/linux/blob/v5.1-rc5/mm/slub.c#L3033-L3113
How often is this a real problem for real workloads?
> > For example, have a userspace process that does networking, which is
> > often memory allocation intensive, so that we know that SKBs traverse
> > between CPUs. You can do this by making sure that the NIC queues are
> > mapped to CPU N (so that network softirqs have to run on that CPU) but
> > the process is pinned to CPU M.
>
> If someone want to test this with SKBs then be-aware that we netdev-guys
> have a number of optimizations where we try to counter act this. (As
> minimum disable TSO and GRO).
>
> It might also be possible for people to get inspired by and adapt the
> micro benchmarking[2] kernel modules that I wrote when developing the
> SLUB and SLAB optimizations:
>
> [2] https://github.com/netoptimizer/prototype-kernel/tree/master/kernel/mm
While microbenchmarks are good to see pathological behavior, I would be
really interested to see some numbers for real world usecases.
> > It's, of course, worth thinking about other pathological cases too.
> > Workloads that cause large allocations is one. Workloads that cause lots
> > of slab cache shrinking is another.
>
> I also worry about long uptimes when SLUB objects/pages gets too
> fragmented... as I said SLUB is only efficient when objects are
> returned to the same page, while SLAB is not.
Is this something that has been actually measured in a real deployment?
--
Michal Hocko
SUSE Labs
next prev parent reply other threads:[~2019-04-17 13:38 UTC|newest]
Thread overview: 14+ messages / expand[flat|nested] mbox.gz Atom feed top
2019-04-10 2:47 Tobin C. Harding
2019-04-10 2:47 ` [PATCH 1/1] mm: Remove " Tobin C. Harding
2019-04-10 8:02 ` [PATCH 0/1] mm: Remove the " Vlastimil Babka
2019-04-10 8:16 ` Tobin C. Harding
2019-04-11 7:55 ` Michal Hocko
2019-04-11 8:27 ` Pekka Enberg
2019-04-17 8:50 ` Jesper Dangaard Brouer
2019-04-17 13:27 ` Christopher Lameter
2019-04-17 13:38 ` Michal Hocko [this message]
2019-04-22 14:43 ` Jesper Dangaard Brouer
2019-04-11 8:44 ` Mel Gorman
2019-04-10 21:53 ` David Rientjes
2019-04-12 11:28 ` Mel Gorman
2019-04-17 3:52 ` Andrew Morton
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20190417133852.GL5878@dhcp22.suse.cz \
--to=mhocko@kernel.org \
--cc=akpm@linux-foundation.org \
--cc=alexander.duyck@gmail.com \
--cc=cai@lca.pw \
--cc=cl@linux.com \
--cc=iamjoonsoo.kim@lge.com \
--cc=linux-kernel@vger.kernel.org \
--cc=linux-mm@kvack.org \
--cc=me@tobin.cc \
--cc=mgorman@techsingularity.net \
--cc=netdev@brouer.com \
--cc=netdev@vger.kernel.org \
--cc=penberg@iki.fi \
--cc=penberg@kernel.org \
--cc=rientjes@google.com \
--cc=tj@kernel.org \
--cc=tobin@kernel.org \
--cc=torvalds@linux-foundation.org \
--cc=vbabka@suse.cz \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox