From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org Received: from kanga.kvack.org (kanga.kvack.org [205.233.56.17]) by smtp.lore.kernel.org (Postfix) with ESMTP id 7769CE7718F for ; Mon, 30 Dec 2024 17:47:17 +0000 (UTC) Received: by kanga.kvack.org (Postfix) id 04D1D6B00A0; Mon, 30 Dec 2024 12:47:17 -0500 (EST) Received: by kanga.kvack.org (Postfix, from userid 40) id F3DEB6B00A1; Mon, 30 Dec 2024 12:47:16 -0500 (EST) X-Delivered-To: int-list-linux-mm@kvack.org Received: by kanga.kvack.org (Postfix, from userid 63042) id DB6B86B00A2; Mon, 30 Dec 2024 12:47:16 -0500 (EST) X-Delivered-To: linux-mm@kvack.org Received: from relay.hostedemail.com (smtprelay0013.hostedemail.com [216.40.44.13]) by kanga.kvack.org (Postfix) with ESMTP id B562F6B00A0 for ; Mon, 30 Dec 2024 12:47:16 -0500 (EST) Received: from smtpin03.hostedemail.com (a10.router.float.18 [10.200.18.1]) by unirelay05.hostedemail.com (Postfix) with ESMTP id 69026444E6 for ; Mon, 30 Dec 2024 17:47:16 +0000 (UTC) X-FDA: 82952354520.03.2FEA500 Received: from mail-pl1-f169.google.com (mail-pl1-f169.google.com [209.85.214.169]) by imf04.hostedemail.com (Postfix) with ESMTP id E478F40006 for ; Mon, 30 Dec 2024 17:46:25 +0000 (UTC) Authentication-Results: imf04.hostedemail.com; dkim=pass header.d=gmail.com header.s=20230601 header.b=SKAk8kD+; spf=pass (imf04.hostedemail.com: domain of ryncsn@gmail.com designates 209.85.214.169 as permitted sender) smtp.mailfrom=ryncsn@gmail.com; dmarc=pass (policy=none) header.from=gmail.com ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=hostedemail.com; s=arc-20220608; t=1735580813; h=from:from:sender:reply-to:reply-to:subject:subject:date:date: message-id:message-id:to:to:cc:cc:mime-version:mime-version: content-type:content-transfer-encoding:content-transfer-encoding: in-reply-to:in-reply-to:references:references:dkim-signature; bh=kMKA8ivFUMjT0SUJNutwCBcko1EjPqMzoc0GZ2wQpbM=; b=zNLxowPykmT/YAnk8RUPw/ZAkE6nVxMXSY2sFSPPd64OrI8qSba5LSIMH98vhpfcYz7ZXc NGkRcvnQrMsyR7f3EhTD9b/Xlo3bDmGSeS2OdTiYcN8k1igLhDwMMhk7dBN2qUARbctxnA NPGesYQoGSPMr11B/GT+xRrQXvWpHKg= ARC-Authentication-Results: i=1; imf04.hostedemail.com; dkim=pass header.d=gmail.com header.s=20230601 header.b=SKAk8kD+; spf=pass (imf04.hostedemail.com: domain of ryncsn@gmail.com designates 209.85.214.169 as permitted sender) smtp.mailfrom=ryncsn@gmail.com; dmarc=pass (policy=none) header.from=gmail.com ARC-Seal: i=1; s=arc-20220608; d=hostedemail.com; t=1735580813; a=rsa-sha256; cv=none; b=cWxj7k7Md52qtceZU7HrolduthkiKnXxcR7zuidO3ACUFQIWYIOa4RlkSdAwN3UL5bqvPB mtFl2DsJDEYVDNRWhDz/8wEsYfhuL5bV6XYbwmggmAiByA7VHWgIHmx9vDZR6OH/lBRTNq qCm9rdtWW9HzMeccXigeWxuDraCmIz4= Received: by mail-pl1-f169.google.com with SMTP id d9443c01a7336-2163b0c09afso120032285ad.0 for ; Mon, 30 Dec 2024 09:47:14 -0800 (PST) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=20230601; t=1735580832; x=1736185632; darn=kvack.org; h=content-transfer-encoding:mime-version:reply-to:references :in-reply-to:message-id:date:subject:cc:to:from:from:to:cc:subject :date:message-id:reply-to; bh=kMKA8ivFUMjT0SUJNutwCBcko1EjPqMzoc0GZ2wQpbM=; b=SKAk8kD+9tmN422//upaHvd8bEDlZlg0/bANpTcO73X6Juu+WHQ+yxR632yn80aLN9 1MTJyRKKuvy+15//3mFnjF6ktfl28xNbLMK7lgdi2BQgaz39+MrEMPtXQ+X2LraQw6xw F+C9NpIdQUlcgi15TRkjMP7/uph1Ucdn2hn36MtuXF3bUVZpe+fObaf0RU7VeM1yosoa WgRvDt+u0aSsV5UJ2j6FFwtSoXDTgMIRCAxF+DHvN7EWa0f6f4l1RvUQ8Yy0zLjJP+Hv l5BNL7NZTSD2YzYSwb/EvfupZnbCReDNxlPZudJG66CfOctOFNguRWe4IHJ5EFD2dHtO 9ZFQ== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20230601; t=1735580832; x=1736185632; h=content-transfer-encoding:mime-version:reply-to:references :in-reply-to:message-id:date:subject:cc:to:from:x-gm-message-state :from:to:cc:subject:date:message-id:reply-to; bh=kMKA8ivFUMjT0SUJNutwCBcko1EjPqMzoc0GZ2wQpbM=; b=lkB6GmXcjF+g5htsjkUNHlxXvJr5ggujcljLl0w7TnZahRsgrJoLshbLWKcgd2yaLk iDdMaA4BVSZnSqYbN3eA6sftRkYTRVeKieWBufI2qj631XardP3hDfn9v3oOGt+riXKX NGkiu0yJEJsDMrwF7/VASCMS3ywAgNoz5QyNCaChKGCYe5UqhAQu3QzMDnjg9hp+L53V 4uRYPhQqHowb6ih6679We4DtskQUzjPVT7sNfi3uY2lGJOw1NmqLZlCkkYoxjxnuwEms vMR0XAVV9AM3BJvs8O2egWWAU22CFl0DwGhQ8Gs8GQ+7HEe0hu+yiH7/FIkRakFp50ob wYxw== X-Gm-Message-State: AOJu0Yxl9PyUODVCBNx4iVVc01u4VSQPM1tfXQEuWVrKxj+C4ngsmUKX YwDKx6DDCKk/YJKAAnhDa7WRN1cJgyyh1ZrFITXZEsn+o80sskWZby8CFjhkSamGEA== X-Gm-Gg: ASbGncuBbG3Ou/L8/vLQvCDMVyUDrZtZu0mSBkiu2pIIrfMRAXCtPBLpofu8lBr5b+J NrcsJ2su74E1dN6dQr7O1GMD89r9JeDKAlT+KxLuW35XMyt3ORhLqXUJvunJESI+pCWXhS9ph/U i+cCG0KRjSnx+WwMFKFf9KP7D4K7zJAmpRdhYG/zyosNKqj0bOPzKfDYObCsMnnRYXBSnULuDgq XucddIEewdg+I7LKIzqEROCQhGTdPhadI3KBHu5pBzgez8LZCrNRXvJdsNHi++9/FtezOltX0dU X-Google-Smtp-Source: AGHT+IH4UroL9TLEghz5+DxcxLZh6o0DshITcoANLbD1cbTnpb/W/I4XNhbFO6b4xObMW4xleBerCA== X-Received: by 2002:a17:902:c947:b0:20c:9821:6998 with SMTP id d9443c01a7336-219e6e858b4mr465264605ad.10.1735580832386; Mon, 30 Dec 2024 09:47:12 -0800 (PST) Received: from KASONG-MC4.tencent.com ([1.203.117.231]) by smtp.gmail.com with ESMTPSA id d9443c01a7336-219dc9cdf25sm180687695ad.118.2024.12.30.09.47.08 (version=TLS1_3 cipher=TLS_CHACHA20_POLY1305_SHA256 bits=256/256); Mon, 30 Dec 2024 09:47:11 -0800 (PST) From: Kairui Song To: linux-mm@kvack.org Cc: Andrew Morton , Chris Li , Barry Song , Ryan Roberts , Hugh Dickins , Yosry Ahmed , "Huang, Ying" , Nhat Pham , Johannes Weiner , Kalesh Singh , linux-kernel@vger.kernel.org, Kairui Song Subject: [PATCH v3 10/13] mm, swap: simplify percpu cluster updating Date: Tue, 31 Dec 2024 01:46:18 +0800 Message-ID: <20241230174621.61185-11-ryncsn@gmail.com> X-Mailer: git-send-email 2.47.1 In-Reply-To: <20241230174621.61185-1-ryncsn@gmail.com> References: <20241230174621.61185-1-ryncsn@gmail.com> Reply-To: Kairui Song MIME-Version: 1.0 Content-Transfer-Encoding: 8bit X-Rspamd-Queue-Id: E478F40006 X-Stat-Signature: iao859ru5aksg1qhdhqdr5te9xp6xkqn X-Rspam-User: X-Rspamd-Server: rspam11 X-HE-Tag: 1735580785-493802 X-HE-Meta: U2FsdGVkX19uUFMDMQFC53BKO7d8cNOPnIBvEIeI7gFLDRd5robpCdi3z1Ufb1I0bR7uXyep8iJl4uDAGceiKEt39jYlog2jhnPJ/T7t917l58enIGxyIFEnAtP4Q2YSHKwaapn0qziNVFtTAmIzuGROJ5IkRVQYNikkFerZYXjcfIbcgyOwM+4UUwCSQQkWQ9tBU8QUd6JaFAI3fxI8ZT8jyzt6JCjiAU6Lk2DNSdHs2ao3W57TiyaTnFZAb+aEbE2VkW3gI/bhgT578099hUDRaJxdI4+q/43dj6HXJBcCKUi/Cx+W3vp7Dhn9IHBxjMSKCCFvk07p6yUUVk4tnSBdqjl80iEd5tpqRiqRIIUs1Kr8DrdD8ecksLfwtC6a885Lgp7FRzXmE90yUzEXMxTB5Jk+dPRSHh74/vgybIw6qS2Wl/KmLlB9/1oQyVhjSLXcy1n3aGcKGorGu/LdWLvAQUgGL1ApsdhzBxhrhlC4RQACz3haRHKD26ASEtcHtxAkGMjCjtD7tTlcJ5Mh9wHGPT9ncOYLhE+F6JeGnvoXAHhT3pfZejtHuhCvMJnOIWDtMkUOWnmtbb0h4zZhHvlDCyk1/sHoUc6XIJw6MmcJj9zvsnNuvTNANzvNDXzkWzAxLNi8ALmU09JESMVVMatCjUzSFkezRfMBW/6IJqa6dDhVXd2X0mdhoy1hH3clj22vlgby1I4pmdkVl0PlYqPyONt0JbBcsiots/uXiHpFV/SyIub/1kP8smULA5Q5sYFfSe4d2fJ5KXCC66OakzMQfgRM6OLtyckaSYUvcw4nifqIy9bjdOpSqgCQfMva+ZHCnEYi0XDAFPGKPgSFfsvQhy8OyclMuJKdX+or4coAY5k75TXHmTuKHdXkaER9JC9kYc+gkDhVBxoD+RbDTeccUgImAoY2S1ddYklxdkTMIEG30czzzzBhWFg1feyHNrrioRtc38zmjNkDbQp QV9vmDla gtPkeXqlSe7MAYt96ykibNDNKvd0aoCzWpkidaH8HG8KoEE5n9taTEliaf+8lM803skeHEDoQDw97u3XXzqPFHjai/wNK0tTv8z0deZzikTm9Un+BxiLnZCxjlaLJC9SmqP+IGumQ+sxJeW9jEHre5OVxBiCBOUqWKZiaRetFFPuaBQRuK/3jTTRLI8PrwTvQXOAbjypVOsh8aNKuRLQTG3TnzollZWoOfeuqfWlAyWEMxQa3v8golah4zxVuEaGeJw126/tSs+jEnAb+eAOqnt3rCjhtr1J9nCtq7L+TFQweBfctwwZxjvf0Nvi4+Ys21U4j0nQRIPA7jkFFMA4Zdbz5FIAZcAq6cxypdFo+4f9XQIxV3d+wwlgBpI/W8YBDU72aNCHEQwPMKoN5tMHSLozQvCpnV3vaB1Zh7WmXtUhlsHHueyz6d1b2N92k5drIKw02+CYbhkd2JIUUzjyHVCIWC77GKLn1FK0zoTCTKX2cqk3XZOqvgYEfD1KC7zDw1Cb9krNROzic9ejtBf31AZ00LA== X-Bogosity: Ham, tests=bogofilter, spamicity=0.000002, version=1.2.4 Sender: owner-linux-mm@kvack.org Precedence: bulk X-Loop: owner-majordomo@kvack.org List-ID: List-Subscribe: List-Unsubscribe: From: Kairui Song Instead of using a returning argument, we can simply store the next cluster offset to the fixed percpu location, which reduce the stack usage and simplify the function: Object size: ./scripts/bloat-o-meter mm/swapfile.o mm/swapfile.o.new add/remove: 0/0 grow/shrink: 0/2 up/down: 0/-271 (-271) Function old new delta get_swap_pages 2847 2733 -114 alloc_swap_scan_cluster 894 737 -157 Total: Before=30833, After=30562, chg -0.88% Stack usage: Before: swapfile.c:1190:5:get_swap_pages 240 static After: swapfile.c:1185:5:get_swap_pages 216 static Signed-off-by: Kairui Song --- include/linux/swap.h | 4 +-- mm/swapfile.c | 66 +++++++++++++++++++------------------------- 2 files changed, 31 insertions(+), 39 deletions(-) diff --git a/include/linux/swap.h b/include/linux/swap.h index c4ff31cb6bde..4c1d2e69689f 100644 --- a/include/linux/swap.h +++ b/include/linux/swap.h @@ -275,9 +275,9 @@ enum swap_cluster_flags { * The first page in the swap file is the swap header, which is always marked * bad to prevent it from being allocated as an entry. This also prevents the * cluster to which it belongs being marked free. Therefore 0 is safe to use as - * a sentinel to indicate next is not valid in percpu_cluster. + * a sentinel to indicate an entry is not valid. */ -#define SWAP_NEXT_INVALID 0 +#define SWAP_ENTRY_INVALID 0 #ifdef CONFIG_THP_SWAP #define SWAP_NR_ORDERS (PMD_ORDER + 1) diff --git a/mm/swapfile.c b/mm/swapfile.c index dadd4fead689..60a650ba88fd 100644 --- a/mm/swapfile.c +++ b/mm/swapfile.c @@ -759,23 +759,23 @@ static bool cluster_alloc_range(struct swap_info_struct *si, struct swap_cluster return true; } -static unsigned int alloc_swap_scan_cluster(struct swap_info_struct *si, unsigned long offset, - unsigned int *foundp, unsigned int order, +/* Try use a new cluster for current CPU and allocate from it. */ +static unsigned int alloc_swap_scan_cluster(struct swap_info_struct *si, + struct swap_cluster_info *ci, + unsigned long offset, + unsigned int order, unsigned char usage) { - unsigned long start = offset & ~(SWAPFILE_CLUSTER - 1); + unsigned int next = SWAP_ENTRY_INVALID, found = SWAP_ENTRY_INVALID; + unsigned long start = ALIGN_DOWN(offset, SWAPFILE_CLUSTER); unsigned long end = min(start + SWAPFILE_CLUSTER, si->max); unsigned int nr_pages = 1 << order; bool need_reclaim, ret; - struct swap_cluster_info *ci; - ci = &si->cluster_info[offset / SWAPFILE_CLUSTER]; lockdep_assert_held(&ci->lock); - if (end < nr_pages || ci->count + nr_pages > SWAPFILE_CLUSTER) { - offset = SWAP_NEXT_INVALID; + if (end < nr_pages || ci->count + nr_pages > SWAPFILE_CLUSTER) goto out; - } for (end -= nr_pages; offset <= end; offset += nr_pages) { need_reclaim = false; @@ -789,34 +789,27 @@ static unsigned int alloc_swap_scan_cluster(struct swap_info_struct *si, unsigne * cluster has no flag set, and change of list * won't cause fragmentation. */ - if (!cluster_is_usable(ci, order)) { - offset = SWAP_NEXT_INVALID; + if (!cluster_is_usable(ci, order)) goto out; - } if (cluster_is_free(ci)) offset = start; /* Reclaim failed but cluster is usable, try next */ if (!ret) continue; } - if (!cluster_alloc_range(si, ci, offset, usage, order)) { - offset = SWAP_NEXT_INVALID; - goto out; - } - *foundp = offset; - if (ci->count == SWAPFILE_CLUSTER) { - offset = SWAP_NEXT_INVALID; - goto out; - } + if (!cluster_alloc_range(si, ci, offset, usage, order)) + break; + found = offset; offset += nr_pages; + if (ci->count < SWAPFILE_CLUSTER && offset <= end) + next = offset; break; } - if (offset > end) - offset = SWAP_NEXT_INVALID; out: relocate_cluster(si, ci); unlock_cluster(ci); - return offset; + __this_cpu_write(si->percpu_cluster->next[order], next); + return found; } /* Return true if reclaimed a whole cluster */ @@ -885,8 +878,8 @@ static unsigned long cluster_alloc_swap_entry(struct swap_info_struct *si, int o if (cluster_is_usable(ci, order)) { if (cluster_is_free(ci)) offset = cluster_offset(si, ci); - offset = alloc_swap_scan_cluster(si, offset, &found, - order, usage); + found = alloc_swap_scan_cluster(si, ci, offset, + order, usage); } else { unlock_cluster(ci); } @@ -897,8 +890,8 @@ static unsigned long cluster_alloc_swap_entry(struct swap_info_struct *si, int o new_cluster: ci = cluster_isolate_lock(si, &si->free_clusters); if (ci) { - offset = alloc_swap_scan_cluster(si, cluster_offset(si, ci), - &found, order, usage); + found = alloc_swap_scan_cluster(si, ci, cluster_offset(si, ci), + order, usage); if (found) goto done; } @@ -911,8 +904,8 @@ static unsigned long cluster_alloc_swap_entry(struct swap_info_struct *si, int o unsigned int frags = 0, frags_existing; while ((ci = cluster_isolate_lock(si, &si->nonfull_clusters[order]))) { - offset = alloc_swap_scan_cluster(si, cluster_offset(si, ci), - &found, order, usage); + found = alloc_swap_scan_cluster(si, ci, cluster_offset(si, ci), + order, usage); /* * With `fragmenting` set to true, it will surely take * the cluster off nonfull list @@ -932,8 +925,8 @@ static unsigned long cluster_alloc_swap_entry(struct swap_info_struct *si, int o * per-CPU usage, but they could contain newly released * reclaimable (eg. lazy-freed swap cache) slots. */ - offset = alloc_swap_scan_cluster(si, cluster_offset(si, ci), - &found, order, usage); + found = alloc_swap_scan_cluster(si, ci, cluster_offset(si, ci), + order, usage); if (found) goto done; frags++; @@ -959,21 +952,20 @@ static unsigned long cluster_alloc_swap_entry(struct swap_info_struct *si, int o */ while ((ci = cluster_isolate_lock(si, &si->frag_clusters[o]))) { atomic_long_dec(&si->frag_cluster_nr[o]); - offset = alloc_swap_scan_cluster(si, cluster_offset(si, ci), - &found, order, usage); + found = alloc_swap_scan_cluster(si, ci, cluster_offset(si, ci), + 0, usage); if (found) goto done; } while ((ci = cluster_isolate_lock(si, &si->nonfull_clusters[o]))) { - offset = alloc_swap_scan_cluster(si, cluster_offset(si, ci), - &found, order, usage); + found = alloc_swap_scan_cluster(si, ci, cluster_offset(si, ci), + 0, usage); if (found) goto done; } } done: - __this_cpu_write(si->percpu_cluster->next[order], offset); local_unlock(&si->percpu_cluster->lock); return found; @@ -3194,7 +3186,7 @@ static struct swap_cluster_info *setup_clusters(struct swap_info_struct *si, cluster = per_cpu_ptr(si->percpu_cluster, cpu); for (i = 0; i < SWAP_NR_ORDERS; i++) - cluster->next[i] = SWAP_NEXT_INVALID; + cluster->next[i] = SWAP_ENTRY_INVALID; local_lock_init(&cluster->lock); } -- 2.47.1