From: Luis Chamberlain <mcgrof@kernel.org>
To: Alexei Starovoitov <alexei.starovoitov@gmail.com>
Cc: Song Liu <song@kernel.org>, bpf <bpf@vger.kernel.org>,
linux-mm <linux-mm@kvack.org>,
Andrew Morton <akpm@linux-foundation.org>,
X86 ML <x86@kernel.org>, Peter Zijlstra <peterz@infradead.org>,
Christoph Hellwig <hch@lst.de>,
Rick Edgecombe <rick.p.edgecombe@intel.com>,
aaron.lu@intel.com, Mike Rapoport <rppt@kernel.org>
Subject: Re: [PATCH bpf-next v3 4/6] bpf: use execmem_alloc for bpf program and bpf dispatcher
Date: Thu, 17 Nov 2022 12:01:30 -0800 [thread overview]
Message-ID: <Y3aTGs9rtWiHpoea@bombadil.infradead.org> (raw)
In-Reply-To: <CAADnVQL6AiQLaURNbchVtUNK2nGFUYCSPZk5dZbScW=iKp1bYw@mail.gmail.com>
On Wed, Nov 16, 2022 at 06:10:23PM -0800, Alexei Starovoitov wrote:
> On Wed, Nov 16, 2022 at 6:04 PM Luis Chamberlain <mcgrof@kernel.org> wrote:
> >
> > On Wed, Nov 16, 2022 at 05:06:19PM -0800, Song Liu wrote:
> > > Use execmem_alloc, execmem_free, and execmem_fill instead of
> > > bpf_prog_pack_alloc, bpf_prog_pack_free, and bpf_arch_text_copy.
> > >
> > > execmem_free doesn't require extra size information. Therefore, the free
> > > and error handling path can be simplified.
> > >
> > > There are some tests that show the benefit of execmem_alloc.
> > >
> > > Run 100 instances of the following benchmark from bpf selftests:
> > > tools/testing/selftests/bpf/bench -w2 -d100 -a trig-kprobe
> > > which loads 7 BPF programs, and triggers one of them.
> > >
> > > Then use perf to monitor TLB related counters:
> > > perf stat -e iTLB-load-misses,itlb_misses.walk_completed_4k, \
> > > itlb_misses.walk_completed_2m_4m -a
> > >
> > > The following results are from a qemu VM with 32 cores.
> > >
> > > Before bpf_prog_pack:
> > > iTLB-load-misses: 350k/s
> > > itlb_misses.walk_completed_4k: 90k/s
> > > itlb_misses.walk_completed_2m_4m: 0.1/s
> > >
> > > With bpf_prog_pack (current upstream):
> > > iTLB-load-misses: 220k/s
> > > itlb_misses.walk_completed_4k: 68k/s
> > > itlb_misses.walk_completed_2m_4m: 0.2/s
> > >
> > > With execmem_alloc (with this set):
> > > iTLB-load-misses: 185k/s
> > > itlb_misses.walk_completed_4k: 58k/s
> > > itlb_misses.walk_completed_2m_4m: 1/s
> >
> > Wonderful.
> >
> > It would be nice to have this integrated into the bpf selftest,
>
>
> No. Luis please stop suggesting things that don't make sense.
> selftest/bpf are not doing performance benchmarking.
> We have the 'bench' tool for that.
> That's what Song used and it's only running standalone
> and not part of any CI.
I'm not suggesting to instantiate the VM or crap like that, I'm just
asking for the simple script to run 100 instances. This allows folks
to reproduce results in an easy way.
Whether or not you don't want that for selftests/bpf -- fine, a simple
in commit script can easily represent a loop in bash if that's all
that was done.
Luis
next prev parent reply other threads:[~2022-11-17 20:01 UTC|newest]
Thread overview: 16+ messages / expand[flat|nested] mbox.gz Atom feed top
2022-11-17 1:06 [PATCH bpf-next v3 0/6] execmem_alloc for BPF programs Song Liu
2022-11-17 1:06 ` [PATCH bpf-next v3 1/6] vmalloc: introduce execmem_alloc, execmem_free, and execmem_fill Song Liu
2022-11-17 1:36 ` Luis Chamberlain
2022-11-17 6:48 ` Song Liu
2022-11-17 1:06 ` [PATCH bpf-next v3 2/6] x86/alternative: support execmem_alloc() and execmem_free() Song Liu
2022-11-17 1:06 ` [PATCH bpf-next v3 3/6] selftests/vm: extend test_vmalloc to test execmem_* APIs Song Liu
2022-11-17 1:49 ` Luis Chamberlain
2022-11-17 6:41 ` Song Liu
2022-11-17 20:04 ` Luis Chamberlain
2022-11-17 1:06 ` [PATCH bpf-next v3 4/6] bpf: use execmem_alloc for bpf program and bpf dispatcher Song Liu
2022-11-17 1:52 ` Luis Chamberlain
2022-11-17 2:10 ` Alexei Starovoitov
2022-11-17 20:01 ` Luis Chamberlain [this message]
2022-11-17 20:03 ` Luis Chamberlain
2022-11-17 1:06 ` [PATCH bpf-next v3 5/6] vmalloc: introduce register_text_tail_vm() Song Liu
2022-11-17 1:06 ` [PATCH bpf-next v3 6/6] x86: use register_text_tail_vm Song Liu
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=Y3aTGs9rtWiHpoea@bombadil.infradead.org \
--to=mcgrof@kernel.org \
--cc=aaron.lu@intel.com \
--cc=akpm@linux-foundation.org \
--cc=alexei.starovoitov@gmail.com \
--cc=bpf@vger.kernel.org \
--cc=hch@lst.de \
--cc=linux-mm@kvack.org \
--cc=peterz@infradead.org \
--cc=rick.p.edgecombe@intel.com \
--cc=rppt@kernel.org \
--cc=song@kernel.org \
--cc=x86@kernel.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox