linux-mm.kvack.org archive mirror
 help / color / mirror / Atom feed
From: Shakeel Butt <shakeelb@google.com>
To: Cathy Zhang <cathy.zhang@intel.com>
Cc: edumazet@google.com, davem@davemloft.net, kuba@kernel.org,
	 pabeni@redhat.com, jesse.brandeburg@intel.com,
	suresh.srinivas@intel.com,  tim.c.chen@intel.com,
	lizhen.you@intel.com, eric.dumazet@gmail.com,
	 netdev@vger.kernel.org, linux-mm@kvack.org,
	cgroups@vger.kernel.org
Subject: Re: [PATCH net-next 1/2] net: Keep sk->sk_forward_alloc as a proper size
Date: Tue, 9 May 2023 17:19:10 +0000	[thread overview]
Message-ID: <20230509171910.yka3hucbwfnnq5fq@google.com> (raw)
In-Reply-To: <20230508020801.10702-2-cathy.zhang@intel.com>

On Sun, May 07, 2023 at 07:08:00PM -0700, Cathy Zhang wrote:
> Before commit 4890b686f408 ("net: keep sk->sk_forward_alloc as small as
> possible"), each TCP can forward allocate up to 2 MB of memory and
> tcp_memory_allocated might hit tcp memory limitation quite soon.

Not just the system level tcp memory limit but we have actually seen in
production unneeded and unexpected memcg OOMs and the commit 4890b686f408
fixes those OOMs as well.

> To
> reduce the memory pressure, that commit keeps sk->sk_forward_alloc as
> small as possible, which will be less than 1 page size if SO_RESERVE_MEM
> is not specified.
> 
> However, with commit 4890b686f408 ("net: keep sk->sk_forward_alloc as
> small as possible"), memcg charge hot paths are observed while system is
> stressed with a large amount of connections. That is because
> sk->sk_forward_alloc is too small and it's always less than
> sk->truesize, network handlers like tcp_rcv_established() should jump to
> slow path more frequently to increase sk->sk_forward_alloc. Each memory
> allocation will trigger memcg charge, then perf top shows the following
> contention paths on the busy system.
> 
>     16.77%  [kernel]            [k] page_counter_try_charge
>     16.56%  [kernel]            [k] page_counter_cancel
>     15.65%  [kernel]            [k] try_charge_memcg
> 
> In order to avoid the memcg overhead and performance penalty,

IMO this is not the right place to fix memcg performance overhead,
specifically because it will re-introduce the memcg OOMs issue. Please
fix the memcg overhead in the memcg code.

Please share the detail profile of the memcg code. I can help in
brainstorming and reviewing the fix.

> sk->sk_forward_alloc should be kept with a proper size instead of as
> small as possible. Keep memory up to 64KB from reclaims when uncharging
> sk_buff memory, which is closer to the maximum size of sk_buff. It will
> help reduce the frequency of allocating memory during TCP connection.
> The original reclaim threshold for reserved memory per-socket is 2MB, so
> the extraneous memory reserved now is about 32 times less than before
> commit 4890b686f408 ("net: keep sk->sk_forward_alloc as small as
> possible").
> 
> Run memcached with memtier_benchamrk to verify the optimization fix. 8
> server-client pairs are created with bridge network on localhost, server
> and client of the same pair share 28 logical CPUs.
> 
> Results (Average for 5 run)
> RPS (with/without patch)	+2.07x
> 

Do you have regression data from any production workload? Please keep in
mind that many times we (MM subsystem) accepts the regressions of
microbenchmarks over complicated optimizations. So, if there is a real
production regression, please be very explicit about it.


       reply	other threads:[~2023-05-09 17:19 UTC|newest]

Thread overview: 39+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
     [not found] <20230508020801.10702-1-cathy.zhang@intel.com>
     [not found] ` <20230508020801.10702-2-cathy.zhang@intel.com>
2023-05-09 17:19   ` Shakeel Butt [this message]
2023-05-09 18:04     ` Chen, Tim C
2023-05-09 18:17       ` Shakeel Butt
2023-05-10  7:03         ` Zhang, Cathy
2023-05-10  7:32           ` Zhang, Cathy
     [not found]   ` <3887b08ac0e55e27a24d2f66afcfff1961ed9b13.camel@redhat.com>
     [not found]     ` <CH3PR11MB73459006FCE3887E1EA3B82FFC769@CH3PR11MB7345.namprd11.prod.outlook.com>
     [not found]       ` <CH3PR11MB73456D792EC6E7614E2EF14DFC769@CH3PR11MB7345.namprd11.prod.outlook.com>
     [not found]         ` <CANn89iL6Ckuu9vOEvc7A9CBLGuh-EpbwFRxRAchV-6VFyhTUpg@mail.gmail.com>
     [not found]           ` <CH3PR11MB73458BB403D537CFA96FD8DDFC769@CH3PR11MB7345.namprd11.prod.outlook.com>
     [not found]             ` <CANn89iJvpgXTwGEiXAkFwY3j3RqVhNzJ_6_zmuRb4w7rUA_8Ug@mail.gmail.com>
2023-05-09 16:09               ` Shakeel Butt
2023-05-10  6:54                 ` Zhang, Cathy
2023-05-10 11:11                 ` Zhang, Cathy
2023-05-10 11:24                   ` Eric Dumazet
2023-05-10 13:52                     ` Zhang, Cathy
2023-05-10 15:07                       ` Eric Dumazet
2023-05-10 16:09                         ` Zhang, Cathy
2023-05-10 19:00                           ` Shakeel Butt
2023-05-11  0:53                             ` Zhang, Cathy
2023-05-11  6:59                               ` Zhang, Cathy
2023-05-11  7:50                                 ` Eric Dumazet
2023-05-11  9:26                                   ` Zhang, Cathy
2023-05-11 16:23                                     ` Shakeel Butt
2023-05-11 16:35                                       ` Eric Dumazet
2023-05-11 17:10                                         ` Shakeel Butt
2023-05-11 21:18                                     ` Shakeel Butt
2023-05-12  2:38                                       ` Zhang, Cathy
2023-05-12  3:23                                         ` Zhang, Cathy
2023-05-12  5:06                                           ` Shakeel Butt
2023-05-12  5:51                                             ` Zhang, Cathy
2023-05-12 17:17                                               ` Shakeel Butt
2023-05-15  3:46                                                 ` Zhang, Cathy
2023-05-15  4:13                                                   ` Shakeel Butt
2023-05-15  6:27                                                     ` Zhang, Cathy
2023-05-15 19:50                                                       ` Shakeel Butt
2023-05-16  5:46                                                         ` Oliver Sang
2023-05-17 16:24                                                           ` Shakeel Butt
2023-05-17 16:33                                                             ` Eric Dumazet
2023-05-17 17:04                                                               ` Shakeel Butt
2023-07-28  2:26                                                                 ` Zhang, Cathy
2023-05-19  2:53                                                             ` Oliver Sang
2023-05-31  8:46                                                             ` Oliver Sang
2023-05-09 17:58             ` Shakeel Butt
2023-05-10  7:21               ` Zhang, Cathy

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=20230509171910.yka3hucbwfnnq5fq@google.com \
    --to=shakeelb@google.com \
    --cc=cathy.zhang@intel.com \
    --cc=cgroups@vger.kernel.org \
    --cc=davem@davemloft.net \
    --cc=edumazet@google.com \
    --cc=eric.dumazet@gmail.com \
    --cc=jesse.brandeburg@intel.com \
    --cc=kuba@kernel.org \
    --cc=linux-mm@kvack.org \
    --cc=lizhen.you@intel.com \
    --cc=netdev@vger.kernel.org \
    --cc=pabeni@redhat.com \
    --cc=suresh.srinivas@intel.com \
    --cc=tim.c.chen@intel.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox