From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org Received: from kanga.kvack.org (kanga.kvack.org [205.233.56.17]) by smtp.lore.kernel.org (Postfix) with ESMTP id 9F74EC3601A for ; Fri, 4 Apr 2025 18:55:45 +0000 (UTC) Received: by kanga.kvack.org (Postfix) id D72126B0007; Fri, 4 Apr 2025 14:55:43 -0400 (EDT) Received: by kanga.kvack.org (Postfix, from userid 40) id D21FD6B000C; Fri, 4 Apr 2025 14:55:43 -0400 (EDT) X-Delivered-To: int-list-linux-mm@kvack.org Received: by kanga.kvack.org (Postfix, from userid 63042) id BC22C6B000D; Fri, 4 Apr 2025 14:55:43 -0400 (EDT) X-Delivered-To: linux-mm@kvack.org Received: from relay.hostedemail.com (smtprelay0012.hostedemail.com [216.40.44.12]) by kanga.kvack.org (Postfix) with ESMTP id 9FF716B0007 for ; Fri, 4 Apr 2025 14:55:43 -0400 (EDT) Received: from smtpin14.hostedemail.com (a10.router.float.18 [10.200.18.1]) by unirelay06.hostedemail.com (Postfix) with ESMTP id E9698B89E3 for ; Fri, 4 Apr 2025 18:55:43 +0000 (UTC) X-FDA: 83297265366.14.2993A78 Received: from us-smtp-delivery-124.mimecast.com (us-smtp-delivery-124.mimecast.com [170.10.133.124]) by imf22.hostedemail.com (Postfix) with ESMTP id 9825EC0011 for ; Fri, 4 Apr 2025 18:55:41 +0000 (UTC) Authentication-Results: imf22.hostedemail.com; dkim=pass header.d=redhat.com header.s=mimecast20190719 header.b=EfUYWEwA; spf=pass (imf22.hostedemail.com: domain of llong@redhat.com designates 170.10.133.124 as permitted sender) smtp.mailfrom=llong@redhat.com; dmarc=pass (policy=quarantine) header.from=redhat.com ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=hostedemail.com; s=arc-20220608; t=1743792941; h=from:from:sender:reply-to:subject:subject:date:date: message-id:message-id:to:to:cc:cc:mime-version:mime-version: content-type:content-type: content-transfer-encoding:content-transfer-encoding: in-reply-to:in-reply-to:references:references:dkim-signature; bh=RJ+WDACAsHxbVWot9DqTlvwAc/91z/vLEjznERJRGEA=; b=ZNP/Itp6u2XZNFWzYUxnnlnCXnSeIDRs3vpjA99Ux7g96ptDSTrLiFqYklB4h0wCnX+3rI JVmEv9A3EvGjuf38CwG9nkWn8o/yDF8xhEBCETPTNEJ87h/Fg7mOI0FA7aC31k1yesSaCN U/6TEz9woJgYSOMes05FBiMQLmjUdsA= ARC-Seal: i=1; s=arc-20220608; d=hostedemail.com; t=1743792941; a=rsa-sha256; cv=none; b=Hx9zGUhSK1BDUWrmfuH4wRWW2MgBBpIZ57DrLniyfhKZlQNftaYvABpCfumfFYHHNeneQ7 GPvrxYjdfwYu6cHADp/ci5Uy/NQOWB1EDsrQE0wjmdCf4DeirxYjSyRop+lGkBIVgJyLXp 1pqF5lQzCOZa+UDvKJFgdlCMT522F4U= ARC-Authentication-Results: i=1; imf22.hostedemail.com; dkim=pass header.d=redhat.com header.s=mimecast20190719 header.b=EfUYWEwA; spf=pass (imf22.hostedemail.com: domain of llong@redhat.com designates 170.10.133.124 as permitted sender) smtp.mailfrom=llong@redhat.com; dmarc=pass (policy=quarantine) header.from=redhat.com DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=redhat.com; s=mimecast20190719; t=1743792941; h=from:from:reply-to:subject:subject:date:date:message-id:message-id: to:to:cc:cc:mime-version:mime-version:content-type:content-type: content-transfer-encoding:content-transfer-encoding: in-reply-to:in-reply-to:references:references; bh=RJ+WDACAsHxbVWot9DqTlvwAc/91z/vLEjznERJRGEA=; b=EfUYWEwAzxarHgqAUT5fSmL6LqulWs27+i3mSPkj1APUF4fkW9e/pTr1wBvbLGfP30QPwT cMYv0t4HE6xnjNQOXKQWYS5v5BySoDUcDx0bIRBhB+BXTFvKc9h6USH+86ZhSH9YtxteFc /g1y2KbN1zXFDg9gHoD7Do3K1ZUR95c= Received: from mail-qk1-f200.google.com (mail-qk1-f200.google.com [209.85.222.200]) by relay.mimecast.com with ESMTP with STARTTLS (version=TLSv1.3, cipher=TLS_AES_256_GCM_SHA384) id us-mta-616-opkdmvYkOOWrV-FkU9118w-1; Fri, 04 Apr 2025 14:55:38 -0400 X-MC-Unique: opkdmvYkOOWrV-FkU9118w-1 X-Mimecast-MFC-AGG-ID: opkdmvYkOOWrV-FkU9118w_1743792937 Received: by mail-qk1-f200.google.com with SMTP id af79cd13be357-7c5d608e6f5so561878185a.0 for ; Fri, 04 Apr 2025 11:55:38 -0700 (PDT) X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20230601; t=1743792937; x=1744397737; h=content-transfer-encoding:in-reply-to:content-language:references :cc:to:subject:user-agent:mime-version:date:message-id:from :x-gm-message-state:from:to:cc:subject:date:message-id:reply-to; bh=RJ+WDACAsHxbVWot9DqTlvwAc/91z/vLEjznERJRGEA=; b=MOcGatGPCnemD+Dp3X1uf12w08k2V6pw1XkatBHAHixtkjTKptqH0KqXfN69ymY/V5 LCud6G2u/HxQPRVH3VnErUQbdRYiPRz7VG8Ov28cnZYQJSL40rUhSfUPKJkv9cC1KaOR 7zGNc6P+1hh+MRtJ9a6AXDirZwm4V2N/yq4QVmEzNesBZPlG3RtT+5wuC4fcv4Au1q2k 8K6z4ormg/R5Mj2pd5b8xWFG/9D129IiwTQFtYcPDt6GQGU/DDNqcVul4dagOTGqG30N It7Br37jP3jBW7NOHZ8j9yRZPULGlPHFjpxqCvbQxIhbV87uKvU7KbyGfhK6OuixJP6P IFOA== X-Forwarded-Encrypted: i=1; AJvYcCXtR5Ix0WtAifQ/4uKfZ0ShfZ2mPLYMoZe2sZ8YBHrkINzHQfTTk7V9YwSTkbfcFgc2pcHHqjBppw==@kvack.org X-Gm-Message-State: AOJu0YwZvLtl/MxYne/K2QRUOCo/lYYxCX6NMY5AugGOurTnXB7gIrgt ueVSgWBptxI27qdnnS9AqH8BRkqOyKLBrXVFc9AdOUvmEhzl7V1GZ38PgsXZeCPMgv93QCnwC3g w4lXWh3M03U7Y0kMPPdHEw4w67qBZXJhy9e7lGrvk4jVLEJ0p X-Gm-Gg: ASbGncsxv+H1JHOZDfEFaDHdL/NRw+Eo7FhbhFpCVY3GgoKcFy/i2H2HraO9XTjk8qz 69rBYrpKuKvmjRyBaLqxX6LvRknXx1U3E7I822wX53cDgoN/8CkV1xvxBW4mIFbG7PFuVtlgWbo TnKBY7wy3BHSAL2JtMbIVtDMAWbq0b9MZrSdEwOuoUcON/tQmCuw6n9iQYAilYxTJLA+RFJ5ueg jBrPWG9aZO8ag/ZzP4QaZwKh8jOo2Rd+5Ay1Ll6A6y94WkGbGEr1rKt6AU7W/ffHcZlYJFkiDRY lpJhIZk3V9noTFryJqye+u57ZN+PHfAWl3BCUbguBnZEMub4mXu9/2ndrc9xsQ== X-Received: by 2002:a05:620a:d93:b0:7c5:6dc7:7e7c with SMTP id af79cd13be357-7c774d6fcd5mr576731485a.26.1743792937395; Fri, 04 Apr 2025 11:55:37 -0700 (PDT) X-Google-Smtp-Source: AGHT+IG1GWz7ZoZU1D4Cwc3QUuGir5Wj6Kdn7kWMfIbYb7f8OvFkQTHaKFBY0OhVoifakryu8wvgkg== X-Received: by 2002:a05:620a:d93:b0:7c5:6dc7:7e7c with SMTP id af79cd13be357-7c774d6fcd5mr576727485a.26.1743792936962; Fri, 04 Apr 2025 11:55:36 -0700 (PDT) Received: from ?IPV6:2601:188:c100:5710:315f:57b3:b997:5fca? ([2601:188:c100:5710:315f:57b3:b997:5fca]) by smtp.gmail.com with ESMTPSA id 6a1803df08f44-6ef0f1393fdsm24772476d6.89.2025.04.04.11.55.35 (version=TLS1_3 cipher=TLS_AES_128_GCM_SHA256 bits=128/128); Fri, 04 Apr 2025 11:55:36 -0700 (PDT) From: Waiman Long X-Google-Original-From: Waiman Long Message-ID: Date: Fri, 4 Apr 2025 14:55:35 -0400 MIME-Version: 1.0 User-Agent: Mozilla Thunderbird Subject: Re: [PATCH v2 1/2] memcg: Don't generate low/min events if either low/min or elow/emin is 0 To: Johannes Weiner , Waiman Long Cc: Tejun Heo , Michal Hocko , Roman Gushchin , Shakeel Butt , Muchun Song , Andrew Morton , =?UTF-8?Q?Michal_Koutn=C3=BD?= , Shuah Khan , linux-kernel@vger.kernel.org, cgroups@vger.kernel.org, linux-mm@kvack.org, linux-kselftest@vger.kernel.org References: <20250404012435.656045-1-longman@redhat.com> <1ac51e8e-8dc0-4cd8-9414-f28125061bb3@redhat.com> <20250404181308.GA300138@cmpxchg.org> In-Reply-To: <20250404181308.GA300138@cmpxchg.org> X-Mimecast-Spam-Score: 0 X-Mimecast-MFC-PROC-ID: sSuYasFfHNclhX6-5lO-wotmzGq-ZQWSoUbuaQx-TpY_1743792937 X-Mimecast-Originator: redhat.com Content-Language: en-US Content-Type: text/plain; charset=UTF-8; format=flowed Content-Transfer-Encoding: 8bit X-Stat-Signature: d3gxyfwbr9mczhe13gnafwneq5ogfxzg X-Rspam-User: X-Rspamd-Queue-Id: 9825EC0011 X-Rspamd-Server: rspam08 X-HE-Tag: 1743792941-988082 X-HE-Meta: U2FsdGVkX1+coDTi5yXLfbqXzCuOU3wFtcvjTk3PsmZIfjNoVqQyj+gtjJ0SiHpWaVouJ3ueGrms6NwsgcDW3VuxP6aFp2rYZ7SZAhoMSJbnwIPemNCYKeDw8R7aNYgEIiPNkBb7tCW5n/faGYMRTLkwWc1vPLyST9zpKilmQouwZOymALlML3anQlDWjpTHig8Q7ghc5j2m5oHRPIGVo8ljW8GQrbDtcoZf9TiU0izjaxUrdfg0fxQVhHFJO6bWonPejjT7Y31GJ7b1Zv4pzdzVRiHRLGjmOleaxqtAOHlvQuCvy13xSV75v7+aLidrmfkqdSVTATLcdyFjOvtCgYWAY7Mvg8BTd2QI4QbPaccBZx7HW62g90eWcgnHYwzYnrYSbjwNxXKVOVjpm1Kicy3auKA8stIAmJqHnGizDoEuLfFKz3jbIJV7cqmLRTLD/YCvWAXULhTtWp4IFQ9GhwswnGDayUuM+NHb/cEQJf92fHfFxVd4J9rySnaiyK315nzcoo9Q/lJdsZbguuIufq1smbLEZcrgvzSliGWGx9RLr0Dfe1bWRytCbSlhnrTfxozhtTAt+oU7vBWmaXYVcFz8GJr97za7sdIDOgE/9mYaLDO9g0Qk6bU4VDvTeGKzjAdfR/ELLrOgYd71e86l0uACO2eFJmqsarM3kw7BlK25WytG2NXiI6cBmO16SMQ4Fc585cVHXRHu8HUNNyeT4OwgcDYA8JW7ofYd57eJLVbC59D1QrEEJj1IWFJkqj0Yc1ZflCelRP5oz5f13UWtEHFLq4kUbjTJkUDaeEsetbnAno5BiOPl4Ha1Lfq+Zu9gqJbAmFfwnuJqAsQr88uj7TRmTYklEaG3ULt3yZNzR2GlJQ/YXHCHc0eEGj0mM7Dp6XKe1DIC9Na6q6XcH4T4x4N78ZlZBIeEH69kXClo/IZVJN0sGjXRtUKtcri+Cf0DzknISEFs9yuRldk6JdF zHZgGgJX NS5iXdgwPP+tMs9t5igDZ/kP1lsixWsR/PJb19taeegDC9gysokBVTHT+eUIcwNapeKdJavpmvIw8NgOEEi9qRtcVjKB19U+TvGrEC6mvQDFEZtQqQvLSHkZF2TsLga20EGBE1b12InSpybWwmegQOn5h9eDVc3ndIXq68TyF9Ca1kNf7KPGioX3zmxuZiM6Ab3iF8wi+TQW7+uHL/MwrTlbm8QZKnhlorqY9wB4K0DhSKOtFdUH2NaIV6WUlOS2upqZOuiHk+JOQbcWY1c5SAi46gxdM0we/v6mcIq1nQBKnREp0JBeikk5oi0auPMquix8jVAmRSOJnORG/SF3NSZ12N7CqhYap7P+yO5jUWzdhKg6ZSLhtRuDqVU0VQc/SPjtJyKaZafuoVYdGeo37LbYcWUTGJfAMN25P8XaRvfVi6mtsd3Kshrqq/UjbXxKDkRv9I86vTxP16OI= X-Bogosity: Ham, tests=bogofilter, spamicity=0.000000, version=1.2.4 Sender: owner-linux-mm@kvack.org Precedence: bulk X-Loop: owner-majordomo@kvack.org List-ID: List-Subscribe: List-Unsubscribe: On 4/4/25 2:13 PM, Johannes Weiner wrote: > On Fri, Apr 04, 2025 at 01:25:33PM -0400, Waiman Long wrote: >> On 4/4/25 1:12 PM, Tejun Heo wrote: >>> Hello, >>> >>> On Thu, Apr 03, 2025 at 09:24:34PM -0400, Waiman Long wrote: >>> ... >>>> The simple and naive fix of changing the operator to ">", however, >>>> changes the memory reclaim behavior which can lead to other failures >>>> as low events are needed to facilitate memory reclaim. So we can't do >>>> that without some relatively riskier changes in memory reclaim. >>> I'm doubtful using ">" would change reclaim behavior in a meaningful way and >>> that'd be more straightforward. What do mm people think? > The knob documentation uses "within low" and "above low" to > distinguish whether you are protected or not, so at least from a code > clarity pov, >= makes more sense to me: if your protection is N and > you use exactly N, you're considered protected. > > That also means that by definition an empty cgroup is protected. It's > not in excess of its protection. The test result isn't wrong. > > The real weirdness is issuing a "low reclaim" event when no reclaim is > going to happen*. > > The patch effectively special cases "empty means in excess" to avoid > the event and fall through to reclaim, which then does nothing as a > result of its own scan target calculations. That seems convoluted. > > Why not skip empty cgroups before running inapplicable checks? > > diff --git a/mm/vmscan.c b/mm/vmscan.c > index b620d74b0f66..260ab238ec22 100644 > --- a/mm/vmscan.c > +++ b/mm/vmscan.c > @@ -5963,6 +5963,9 @@ static void shrink_node_memcgs(pg_data_t *pgdat, struct scan_control *sc) > > mem_cgroup_calculate_protection(target_memcg, memcg); > > + if (!mem_cgroup_usage(memcg, false)) > + continue; > + > if (mem_cgroup_below_min(target_memcg, memcg)) { > /* > * Hard protection. Yes, that should take care of the memcg with no task case. > >> I haven't looked deeply into why that is the case, but >> test_memcg_low/min tests had other failures when I made this change. > It surprises me as well that it makes any practical difference. I looked at it again and failure is the same expected memory.current check in test_memcontrol. If I remove the equal sign, I got errors like: values_close: child 0 = 8339456, 29MB = 30408704 failed with err = 21 not ok 1 test_memcg_min So the test is expecting memory.current to have around 29MB, but it got a lot less (~8MB) in this case. Before removing the equality sign, I usually got about 25 MB and above for child 0. That is a pretty big change in behavior, so I didn't make it. > > * Waiman points out that the weirdness is seeing low events without > having a low configured. Eh, this isn't really true with recursive > propagation; you may or may not have an elow depending on parental > configuration and sibling behavior. > Do you mind if we just don't update the low event count if low isn't set, but leave the rest the same like diff --git a/include/linux/memcontrol.h b/include/linux/memcontrol.h index 91721c8862c3..48a8bfa7d337 100644 --- a/include/linux/memcontrol.h +++ b/include/linux/memcontrol.h @@ -659,21 +659,25 @@ static inline bool mem_cgroup_unprotected(struct mem_cgro>  static inline bool mem_cgroup_below_low(struct mem_cgroup *target,                                         struct mem_cgroup *memcg)  { +       unsigned long elow; +         if (mem_cgroup_unprotected(target, memcg))                 return false; -       return READ_ONCE(memcg->memory.elow) >= -               page_counter_read(&memcg->memory); +       elow = READ_ONCE(memcg->memory.elow); +       return elow && (page_counter_read(&memcg->memory) <= elow);  }  static inline bool mem_cgroup_below_min(struct mem_cgroup *target,                                         struct mem_cgroup *memcg)  { +       unsigned long emin; +         if (mem_cgroup_unprotected(target, memcg))                 return false; -       return READ_ONCE(memcg->memory.emin) >= -               page_counter_read(&memcg->memory); +       emin = READ_ONCE(memcg->memory.emin); +       return emin && (page_counter_read(&memcg->memory) <= emin);  }  void mem_cgroup_commit_charge(struct folio *folio, struct mem_cgroup *memcg); diff --git a/mm/vmscan.c b/mm/vmscan.c index 77d015d5db0c..e8c1838c7962 100644 --- a/mm/vmscan.c +++ b/mm/vmscan.c @@ -4827,7 +4827,8 @@ static int shrink_one(struct lruvec *lruvec, struct scan_>                 if (READ_ONCE(lruvec->lrugen.seg) != MEMCG_LRU_TAIL)                         return MEMCG_LRU_TAIL; -               memcg_memory_event(memcg, MEMCG_LOW); +               if (memcg->memory.low) +                       memcg_memory_event(memcg, MEMCG_LOW);         }         success = try_to_shrink_lruvec(lruvec, sc); @@ -5902,6 +5903,9 @@ static void shrink_node_memcgs(pg_data_t *pgdat, struct s>                 mem_cgroup_calculate_protection(target_memcg, memcg); +               if (!mem_cgroup_usage(memcg, false)) +                       continue; +                 if (mem_cgroup_below_min(target_memcg, memcg)) {                         /*                          * Hard protection. @@ -5919,7 +5923,8 @@ static void shrink_node_memcgs(pg_data_t *pgdat, struct s>                                 sc->memcg_low_skipped = 1;                                 continue;                         } -                       memcg_memory_event(memcg, MEMCG_LOW); +                       if (memcg->memory.low) +                               memcg_memory_event(memcg, MEMCG_LOW);                 }                 reclaimed = sc->nr_reclaimed; Cheers, Longman