From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org Received: from kanga.kvack.org (kanga.kvack.org [205.233.56.17]) by smtp.lore.kernel.org (Postfix) with ESMTP id AA6E0C02181 for ; Mon, 20 Jan 2025 21:04:08 +0000 (UTC) Received: by kanga.kvack.org (Postfix) id DF3336B0082; Mon, 20 Jan 2025 16:04:07 -0500 (EST) Received: by kanga.kvack.org (Postfix, from userid 40) id DA2D86B0083; Mon, 20 Jan 2025 16:04:07 -0500 (EST) X-Delivered-To: int-list-linux-mm@kvack.org Received: by kanga.kvack.org (Postfix, from userid 63042) id C69E56B0085; Mon, 20 Jan 2025 16:04:07 -0500 (EST) X-Delivered-To: linux-mm@kvack.org Received: from relay.hostedemail.com (smtprelay0011.hostedemail.com [216.40.44.11]) by kanga.kvack.org (Postfix) with ESMTP id AD7C36B0082 for ; Mon, 20 Jan 2025 16:04:07 -0500 (EST) Received: from smtpin28.hostedemail.com (a10.router.float.18 [10.200.18.1]) by unirelay03.hostedemail.com (Postfix) with ESMTP id 2469CA03DC for ; Mon, 20 Jan 2025 21:04:07 +0000 (UTC) X-FDA: 83029057734.28.412F502 Received: from mail-qt1-f172.google.com (mail-qt1-f172.google.com [209.85.160.172]) by imf03.hostedemail.com (Postfix) with ESMTP id 36F0220013 for ; Mon, 20 Jan 2025 21:04:05 +0000 (UTC) Authentication-Results: imf03.hostedemail.com; dkim=pass header.d=google.com header.s=20230601 header.b=CDuEmvNs; spf=pass (imf03.hostedemail.com: domain of surenb@google.com designates 209.85.160.172 as permitted sender) smtp.mailfrom=surenb@google.com; dmarc=pass (policy=reject) header.from=google.com ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=hostedemail.com; s=arc-20220608; t=1737407045; h=from:from:sender:reply-to:subject:subject:date:date: message-id:message-id:to:to:cc:cc:mime-version:mime-version: content-type:content-type: content-transfer-encoding:content-transfer-encoding: in-reply-to:in-reply-to:references:references:dkim-signature; bh=FvD7Un8xWMnwZqQf1PvdUdzRzsQlazu9kusmBWVVJJs=; b=VDwvmbOwz8r0U/lejyHvFVFsUaoKm9y2wtHjhT3OVK4qcYVC/NhRKfQvN49znJJ8PXyytd FN9O79Aa6RIt4jtxAddNuWrNXQELcQlg7cJYc2t8BBZoZ0rbxKqy85015V1eS+Jl/JCXJi WFCSJiRir95MYVzRrgdggnGxC8Rhqw0= ARC-Seal: i=1; s=arc-20220608; d=hostedemail.com; t=1737407045; a=rsa-sha256; cv=none; b=wfzXYGs8iX/m5uyLQdctq1y1RJgf59xLimQmHgqLdn1WUWDeg8LL2KKSgewGXH7/NyqAeX ySCU+CztNXW4QMfFR7d7sggclSuZjiZe+5V2yz4gMhOByY9ob0XrVupqD6OOw4gpWrxv8U kgSpSeoOxsjLNqdpnkseG0Wi+s6Y0gE= ARC-Authentication-Results: i=1; imf03.hostedemail.com; dkim=pass header.d=google.com header.s=20230601 header.b=CDuEmvNs; spf=pass (imf03.hostedemail.com: domain of surenb@google.com designates 209.85.160.172 as permitted sender) smtp.mailfrom=surenb@google.com; dmarc=pass (policy=reject) header.from=google.com Received: by mail-qt1-f172.google.com with SMTP id d75a77b69052e-4678c9310afso502901cf.1 for ; Mon, 20 Jan 2025 13:04:04 -0800 (PST) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=20230601; t=1737407044; x=1738011844; darn=kvack.org; h=content-transfer-encoding:cc:to:subject:message-id:date:from :in-reply-to:references:mime-version:from:to:cc:subject:date :message-id:reply-to; bh=FvD7Un8xWMnwZqQf1PvdUdzRzsQlazu9kusmBWVVJJs=; b=CDuEmvNsgPyPxNqhovpPVN0xbBdYRq3ycPGB1IufPhN9do+kqF5LUEkCT1cTJtR2Vk CmO7bMgWP85i2CCiz44tJB0LQT4yTLiS7DBszWRahr0c1+QWyepr+V7iyiNAaFW9Lvw/ kvCb7IxJgyYYohb+yYkDi6I/6x2qw+NOLXjzTbd2Xcq3PNYQpHuJ+m7d1riQG9rgi8K8 HO6M5lLzNLoQwNUNKM+glGYEcKZS+9Jjlpk7Qad3R9ncDWVzXGsxbcKCZymzZkoD2U94 cZDQQpH0f0JC9XyyoiBBRy1HPNkcsvTUN1+TjbXgEji0wtQZ2T92yYozygu3blDMLVKs /eGg== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20230601; t=1737407044; x=1738011844; h=content-transfer-encoding:cc:to:subject:message-id:date:from :in-reply-to:references:mime-version:x-gm-message-state:from:to:cc :subject:date:message-id:reply-to; bh=FvD7Un8xWMnwZqQf1PvdUdzRzsQlazu9kusmBWVVJJs=; b=Sm7EFYcaMynx85oQYyY4Mw3deyZpD4eNT+ZWa33Ehsyg6YxE7H65PGDmF21EzAVsSE oGRQNI0QnAnWLyISheJIX/fYJ8gDYu9G3N6ZE+79k6maCcl7lawGtcRucCpSvbWyr0ty 2O1FOmm8GDVIQ+0zZUc1SRKLsFgub/DtPSKVjEKBuMSqJa7fGM3qHfwB67FojFy9phYs ArrXRnNeFyk9IW+YJomSUOo4+Z5v/4ye2Qhfcxhhx9xIwFnZGMjCGkWvsq3LlD5n6qaX Bl4h0xicsiStUueIe1HFilRs0k+h3Giuu6+xkfRlp44imV/YFn0G4pa9ShfVtsnMMY6m Nx/Q== X-Forwarded-Encrypted: i=1; AJvYcCXUBEzfCucnMUCjOdsvBdv+Svp83UmyZwoJFDyJV4jtfyxhiXUfxR07h7BJYkAxTLwx7fAym8ATKw==@kvack.org X-Gm-Message-State: AOJu0Yw8fQUlx1DWlWGGUeNdR8eWyYsRVQ6fXQmWcQie2y+wae5l3ks2 DJFH9mYhVa1MPUslXkou2LzAjoluipszqJCjfnssW5lj5XyUM8X4LOErist3NepAQxHAUDvWEDr lGPhvkbpEptF3OSWXQ5uPdcv5Z0NOYleMpUvV X-Gm-Gg: ASbGncsoocRzDq96vQLOG63AcAKtn8Jm0syfI90hBydi+YCuLmCxAe5dpFyl7ohhKSz sbpnNtYgeZX61x4wE3gUaEbOYIe8Z2EHTAjncQ+jFwKtudSGXLip6l7ujwcqcqiRv3bQDt8+R5S Mk5Zw= X-Google-Smtp-Source: AGHT+IEZ3ZSJbn6fo5hGT+VujNc378Vpx9U0DBHim9iL8P1zQ7P4+C4bZzXVtcklQK145Z3i1KO60ZVJwCC8K6Jqt9U= X-Received: by 2002:a05:622a:6201:b0:466:8c7c:3663 with SMTP id d75a77b69052e-46e21081ab9mr6965131cf.5.1737407043918; Mon, 20 Jan 2025 13:04:03 -0800 (PST) MIME-Version: 1.0 References: <20250106112103.25401-1-hao.ge@linux.dev> <48f208b6.32ab.19455c70dbe.Coremail.00107082@163.com> <254a4857.b2b.19458d0dbc2.Coremail.00107082@163.com> <213ff7d2.7c6c.1945eb0c2ff.Coremail.00107082@163.com> <961050d.3c22.19462e1e30d.Coremail.00107082@163.com> <4c0ef846.1055.19467928c70.Coremail.00107082@163.com> In-Reply-To: <4c0ef846.1055.19467928c70.Coremail.00107082@163.com> From: Suren Baghdasaryan Date: Mon, 20 Jan 2025 13:03:52 -0800 X-Gm-Features: AbW1kvY7wz3yEurvO9KP82BfhUFyNzmRhwwE3xNqO5yjS_10fo5EOwMsP0OQv0I Message-ID: Subject: Re: memory alloc profiling seems not work properly during bootup? To: David Wang <00107082@163.com> Cc: kent.overstreet@linux.dev, Hao Ge , akpm@linux-foundation.org, linux-kernel@vger.kernel.org, linux-mm@kvack.org, Hao Ge , Alessio Balsini , Pasha Tatashin , Sourav Panda Content-Type: text/plain; charset="UTF-8" Content-Transfer-Encoding: quoted-printable X-Stat-Signature: 984uau8hznokzn4x88zjrw9ud3kq5gfj X-Rspam-User: X-Rspamd-Queue-Id: 36F0220013 X-Rspamd-Server: rspam03 X-HE-Tag: 1737407045-944976 X-HE-Meta: U2FsdGVkX1+ynjTxe3QlRoFB/FYifJXomQZMp8FTUT+9Ri9VS7d2SUFyLBf8aBQhTVAUFCd8ToMwxhZ5nJh+rubNiavcIZzjyLkT4QsRLlOUkawfgC5xEC0gEMg/2wId8/aKzI4pcQuK3mhTovaCvdTiuqPZSbtPdRDkpA3bIm+GD8T9aiiimEfWO1+ipP3riEwI3WN117IARjAT2EqOvF0ozkQK9fsdfLUidFwLbOPmkAro6AsF4Iea71ETD6scKz+7Ia++dPoahxann6Yk/xiy9QXK/tMYa2GZdf3MAiUOKn0uj+U7LaXkhoZr3FCa/pRWc8RVrXod6QfoWbQVUEY+mTdDgwxE0gFQwfHZslCpEuMl85pyp9oHQRxYsKzDPKAQ6cYTXx+Xpg/CYYdbVKlLqlakgWlHumKS3vQ/dYwhQIOsmJybBXjJHXfZcJSLzRcQFdZ8qmRiKYNyDE+Ty14CO4oBCWRVh2JufRfUFoWp38RgCdiessgCwKA1BrWb3gQliKpjVBb6lZy3INvqYHPKtoEkZiN8lB9WN/N/qxyY9qbsGJ6COK2MuvOl1UycIaeUCn0zvCN7IZMi5aARez9Z7GvTVc8VQ6B2HuUcBjAJDUReGXYvYvQrZAMhUWSOrAL5F1k9zbYtBAvMuNy9HfS5ST4VnIIMqNAD+9K2+EsfWZO0v1aT5+qnF//fmAQP5CPX2KumVZH5PCs6NUhbUnw90Mp4OPrbaEg+qYIy+ZE/HiO6yVOTxN3mFIvnzuxeIOoi0oJCf+p3yX8+d/Zlb3nY/sNtW5zk1IYAAr2nQXg5l+rEOR0OcxqutRO81yAxAD87JvVBWJzsYyIkTDiYC702GI+CCbaVPMAJPjvmyo+xsvsNjWD3roeJMWFAuVRlmiWx2UoA6p3uZb/KxcqDavQnthMVXhTFox4QgV2Kgc+RIgqFgwtN2i/cGfZIxwaBy5qPJ7EAIsdTFeJHXhb rChTpv/u twTZ3pAAj0cNaDEs4yY2ay2Rfb+0bDlSR0wXL4mHJWZ0CntiXq5wlocVNZx9ro8Xiot39yFOpbFr6twIqVMZKnnBhxVgLH+WCI8YqNouHoJK8V5q5tT9mBGCNYRh0Hq8w5/SKQHeC9MUJPaDE88Yn8UOkjzQrp3gPlM/OIb/bEkw10iI6HcMK44m/qGEMf2JXv+aAU5IhJFuTm6+Eqo14CBBltKs0Edid2bluWg8OKityrz+HXh3OsDhkBfB7t/F66g3dMXM6a4zMfxSjUKUn188kZX1zKyUfTXGqNeAjbiF+do38T919i8Qb8sMFmCRzQnRAhAlz3eSkUUrYFwT1EcgY0u16VZ3lGkLNBkzReFD/NcsL3NIi6K/Vnw== X-Bogosity: Ham, tests=bogofilter, spamicity=0.024780, version=1.2.4 Sender: owner-linux-mm@kvack.org Precedence: bulk X-Loop: owner-majordomo@kvack.org List-ID: List-Subscribe: List-Unsubscribe: On Tue, Jan 14, 2025 at 5:27=E2=80=AFPM David Wang <00107082@163.com> wrote= : > > > At 2025-01-15 02:48:13, "Suren Baghdasaryan" wrote: > >On Mon, Jan 13, 2025 at 7:36=E2=80=AFPM David Wang <00107082@163.com> wr= ote: > >> > > >> >> I have my accumulative counter patch and filter out items with 0 ac= cumulative counter, > >> >> I am almost sure the patch would not cause this accounting issue, b= ut not 100%..... > >> > > >> >Have you tested this without your accumulative counter patch? > >> >IIUC, that patch filters out any allocation which has never been hit. > >> >So, if suspend/resume path contains allocations which were never hit > >> >before then those allocations would become suddenly visible, like in > >> >your case. That's why I'm against filtering allocinfo data in the > >> >kernel. Please try this without your patch and see if the data become= s > >> >more consistent. > >> > >> I remove all my patch and build a 6.13.0-rc7 kernel, > >> After boot up, > >> 64 1 kernel/sched/topology.c:2579 func:alloc_sched_do= mains > >> 896 14 kernel/sched/topology.c:2275 func:__sdt_alloc > >> 896 14 kernel/sched/topology.c:2266 func:__sdt_alloc > >> 96 6 kernel/sched/topology.c:2259 func:__sdt_alloc > >> 12288 24 kernel/sched/topology.c:2252 func:__sdt_alloc > >> 0 0 kernel/sched/topology.c:2242 func:__sdt_alloc > >> 0 0 kernel/sched/topology.c:2238 func:__sdt_alloc > >> 0 0 kernel/sched/topology.c:2234 func:__sdt_alloc > >> 0 0 kernel/sched/topology.c:2230 func:__sdt_alloc > >> 512 1 kernel/sched/topology.c:1961 func:sched_init_num= a > >> > >> And after suspend/resume, no change detected: > >> 64 1 kernel/sched/topology.c:2579 func:alloc_sched_do= mains > >> 896 14 kernel/sched/topology.c:2275 func:__sdt_alloc > >> 896 14 kernel/sched/topology.c:2266 func:__sdt_alloc > >> 96 6 kernel/sched/topology.c:2259 func:__sdt_alloc > >> 12288 24 kernel/sched/topology.c:2252 func:__sdt_alloc > >> 0 0 kernel/sched/topology.c:2242 func:__sdt_alloc > >> 0 0 kernel/sched/topology.c:2238 func:__sdt_alloc > >> 0 0 kernel/sched/topology.c:2234 func:__sdt_alloc > >> 0 0 kernel/sched/topology.c:2230 func:__sdt_alloc > >> 512 1 kernel/sched/topology.c:1961 func:sched_init_num= a > >> > >> I also build a image with accumulative counter, but no filter. > >> > >> After boot up: > >> 64 1 kernel/sched/topology.c:2579 func:alloc_sched_do= mains 2 > >> 896 14 kernel/sched/topology.c:2275 func:__sdt_alloc 80 > >> 896 14 kernel/sched/topology.c:2266 func:__sdt_alloc 80 > >> 96 6 kernel/sched/topology.c:2259 func:__sdt_alloc 80 > >> 12288 24 kernel/sched/topology.c:2252 func:__sdt_alloc 80 > >> 0 0 kernel/sched/topology.c:2242 func:__sdt_alloc 0 = <---this *0* seems wrong > >> 0 0 kernel/sched/topology.c:2238 func:__sdt_alloc 0 > >> 0 0 kernel/sched/topology.c:2234 func:__sdt_alloc 0 > >> 0 0 kernel/sched/topology.c:2230 func:__sdt_alloc 0 > >> 512 1 kernel/sched/topology.c:1961 func:sched_init_num= a 1 > >> > >> And then suspend/resume: > >> 64 1 kernel/sched/topology.c:2579 func:alloc_sched_do= mains 17 > >> 896 14 kernel/sched/topology.c:2275 func:__sdt_alloc 39= 5 > >> 896 14 kernel/sched/topology.c:2266 func:__sdt_alloc 39= 5 > >> 96 6 kernel/sched/topology.c:2259 func:__sdt_alloc 39= 5 > >> 12288 24 kernel/sched/topology.c:2252 func:__sdt_alloc 39= 5 > >> 0 0 kernel/sched/topology.c:2242 func:__sdt_alloc 70 > >> 0 0 kernel/sched/topology.c:2238 func:__sdt_alloc 70 > >> 0 0 kernel/sched/topology.c:2234 func:__sdt_alloc 70 > >> 0 0 kernel/sched/topology.c:2230 func:__sdt_alloc 70 > >> 512 1 kernel/sched/topology.c:1961 func:sched_init_num= a 1> > >> Reading the code, those allocation behaviors should be tied together: > >> if kzalloc_node at line#2252 happened, then alloc_percpu at line#2230 = should also happened. > > > >Hmm, ok. Looks like early calls to alloc_percpu() are not being > >registered somehow. Could you please share your cumulative counter > >patch with me? I'll try to reproduce this locally and see if I can > > >spot the issue. > > Sure, here is the patch base on 6.13.0-rc7. Thanks and sorry for the delay. It looks like the per-cpu allocations you pointed out happen early enough in the boot process that chunk->obj_exts was not allocated yet. Therefore the check inside pcpu_alloc_tag_alloc_hook() for chunk->obj_exts fails and accounting gets skipped. Allocating obj_exts earlier is not trivial because slab is not available yet. I'll need to look closer into per-cpu code to see how this can be fixed. > > > > > diff --git a/include/linux/alloc_tag.h b/include/linux/alloc_tag.h > index 0bbbe537c5f9..6ca680604c6d 100644 > --- a/include/linux/alloc_tag.h > +++ b/include/linux/alloc_tag.h > @@ -18,6 +18,7 @@ > struct alloc_tag_counters { > u64 bytes; > u64 calls; > + u64 accu_calls; > }; > > /* > @@ -124,7 +125,7 @@ static inline bool mem_alloc_profiling_enabled(void) > > static inline struct alloc_tag_counters alloc_tag_read(struct alloc_tag = *tag) > { > - struct alloc_tag_counters v =3D { 0, 0 }; > + struct alloc_tag_counters v =3D { 0, 0, 0 }; > struct alloc_tag_counters *counter; > int cpu; > > @@ -132,6 +133,7 @@ static inline struct alloc_tag_counters alloc_tag_rea= d(struct alloc_tag *tag) > counter =3D per_cpu_ptr(tag->counters, cpu); > v.bytes +=3D counter->bytes; > v.calls +=3D counter->calls; > + v.accu_calls +=3D counter->accu_calls; > } > > return v; > @@ -179,6 +181,7 @@ static inline bool alloc_tag_ref_set(union codetag_re= f *ref, struct alloc_tag *t > * counter because when we free each part the counter will be decrem= ented. > */ > this_cpu_inc(tag->counters->calls); > + this_cpu_inc(tag->counters->accu_calls); > return true; > } > > diff --git a/lib/alloc_tag.c b/lib/alloc_tag.c > index 7dcebf118a3e..615833d4fbd7 100644 > --- a/lib/alloc_tag.c > +++ b/lib/alloc_tag.c > @@ -97,6 +97,7 @@ static void alloc_tag_to_text(struct seq_buf *out, stru= ct codetag *ct) > > seq_buf_printf(out, "%12lli %8llu ", bytes, counter.calls); > codetag_to_text(out, ct); > + seq_buf_printf(out, " %llu", counter.accu_calls); > seq_buf_putc(out, ' '); > seq_buf_putc(out, '\n'); > } > > > > David > > > > >> > >> kernel/sched/topology.c > >> 2230 sdd->sd =3D alloc_percpu(struct sched_domain *); > >> 2231 if (!sdd->sd) > >> 2232 return -ENOMEM; > >> ... > >> 2246 for_each_cpu(j, cpu_map) { > >> ... > >> 2252 sd =3D kzalloc_node(sizeof(struct sched_d= omain) + cpumask_size(), > >> 2253 GFP_KERNEL, cpu_to_node(j= )); > >> ... > >> 2257 *per_cpu_ptr(sdd->sd, j) =3D sd; > >> > >> > >> But somehow during bootup, those alloc_percpu in kernel/sched/topology= .c:__sdt_alloc were missed in profiling. > >> (I am not meant to sell the idea of accumulative counter again here, b= ut it dose help sometimes. :). > >> > >> >Thanks, > >> >Suren. > >> > > >> > > >> >> > >> > >> Thanks > >> David