From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org Received: from kanga.kvack.org (kanga.kvack.org [205.233.56.17]) by smtp.lore.kernel.org (Postfix) with ESMTP id CF85AC021AA for ; Wed, 19 Feb 2025 01:12:56 +0000 (UTC) Received: by kanga.kvack.org (Postfix) id 689F62801C5; Tue, 18 Feb 2025 20:12:56 -0500 (EST) Received: by kanga.kvack.org (Postfix, from userid 40) id 661182801BB; Tue, 18 Feb 2025 20:12:56 -0500 (EST) X-Delivered-To: int-list-linux-mm@kvack.org Received: by kanga.kvack.org (Postfix, from userid 63042) id 48C512801C5; Tue, 18 Feb 2025 20:12:56 -0500 (EST) X-Delivered-To: linux-mm@kvack.org Received: from relay.hostedemail.com (smtprelay0012.hostedemail.com [216.40.44.12]) by kanga.kvack.org (Postfix) with ESMTP id 1FDFB2801BB for ; Tue, 18 Feb 2025 20:12:56 -0500 (EST) Received: from smtpin06.hostedemail.com (a10.router.float.18 [10.200.18.1]) by unirelay08.hostedemail.com (Postfix) with ESMTP id 8840C1408C9 for ; Wed, 19 Feb 2025 01:12:55 +0000 (UTC) X-FDA: 83134919910.06.A14EB96 Received: from NAM10-MW2-obe.outbound.protection.outlook.com (mail-mw2nam10on2135.outbound.protection.outlook.com [40.107.94.135]) by imf21.hostedemail.com (Postfix) with ESMTP id 5CA861C0012 for ; Wed, 19 Feb 2025 01:12:52 +0000 (UTC) Authentication-Results: imf21.hostedemail.com; dkim=pass header.d=os.amperecomputing.com header.s=selector2 header.b=fIhyYjHW; spf=pass (imf21.hostedemail.com: domain of yang@os.amperecomputing.com designates 40.107.94.135 as permitted sender) smtp.mailfrom=yang@os.amperecomputing.com; arc=pass ("microsoft.com:s=arcselector10001:i=1"); dmarc=pass (policy=quarantine) header.from=amperecomputing.com ARC-Message-Signature: i=2; a=rsa-sha256; c=relaxed/relaxed; d=hostedemail.com; s=arc-20220608; t=1739927572; h=from:from:sender:reply-to:subject:subject:date:date: message-id:message-id:to:to:cc:cc:mime-version:mime-version: content-type:content-type: content-transfer-encoding:content-transfer-encoding: in-reply-to:in-reply-to:references:references:dkim-signature; bh=AbOhFxaF/DlCgkheSGeJwrNSEr/k4tiSt54xTbgEBhw=; b=t0FoDgdXDvKZTVKKLu1IBa7hphPzrzeXKUm9LPFziNku3oebi5af1z+bqCHaUj+Tz7lB+E McImE/koNsM+J6mvHEiYgaG4gufeDanNmMXedDcKWjDHsZ3sZuPUxH4dndKhbCPbEGaepA +QEDPJ/WSsnwWVsAATjIJ9R1BxcYH6U= ARC-Authentication-Results: i=2; imf21.hostedemail.com; dkim=pass header.d=os.amperecomputing.com header.s=selector2 header.b=fIhyYjHW; spf=pass (imf21.hostedemail.com: domain of yang@os.amperecomputing.com designates 40.107.94.135 as permitted sender) smtp.mailfrom=yang@os.amperecomputing.com; arc=pass ("microsoft.com:s=arcselector10001:i=1"); dmarc=pass (policy=quarantine) header.from=amperecomputing.com ARC-Seal: i=2; s=arc-20220608; d=hostedemail.com; t=1739927572; a=rsa-sha256; cv=pass; b=owBwQvu19EhE6YEG3pLiqBm20xxMdaqGaecxR87Qv/6jj+T1SSCpiaCnZYwzxygeEHRE5L BYKT0Yt1pxrbagni6UDVPaNKIParfbFxmlVEea5mKyfOPWUcpnk49Rda1mgAT27mIcRq5m VMG5mQ7iO0Pje9H37RtDWlSK/9dYPEw= ARC-Seal: i=1; a=rsa-sha256; s=arcselector10001; d=microsoft.com; cv=none; b=SE0Ihe8IUGW3W3/gG5AH0Ca8ZdFLH8H+zbcTX4HkObL3T9g/+aiaUamjWfvc+tylx77VZUwjAEc++oFaU2s/BxnZIrBIqGGhqytJqYvnaHDxEFlWUTxwGwhqpH+S2sBigcIccV91g5f5B5cHOddB6tofIVdJteX7QeImyHuT++YbiSnNAJlQ+/JjK57eVPCt9PsvP1skX/3GwxY4nC2Jcay9EU3E09WJJRApMV1lwq8RCCkqfR1BFZSVugrUcKPxFWsMDJjslg8rFrVc7I9QAA+NBq4OVILm/zyXwSMcuQlcrqQOiAFcIszVGWnCK5CyGIu2fLUxVU0kf9m0P3Gzhg== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=microsoft.com; s=arcselector10001; h=From:Date:Subject:Message-ID:Content-Type:MIME-Version:X-MS-Exchange-AntiSpam-MessageData-ChunkCount:X-MS-Exchange-AntiSpam-MessageData-0:X-MS-Exchange-AntiSpam-MessageData-1; bh=AbOhFxaF/DlCgkheSGeJwrNSEr/k4tiSt54xTbgEBhw=; b=G8V96qelhXDt2VK6mA3v5EYYKjbHeWp+nqU8u7vmSnHl9HhHVtQUVa7aFztyzYbxY9qxRAOIVNF9ye0BYOzIUVp9ytcQ27Ypq25bUDl2C7LPhWqVaWvQt/EUe8I8D/KhsgN4/U4YpVtBb45DfrRQmGi8pHuU65M1tudV5prtLwSC3ebQKL7uViGlGgj1deHFucNECN8dh+87e4aPRWVLQyVDA5v/Eo1b1cl8s0GSDmOwgaHPOOPuVD/QuXvxYv1jxaWOyW+ZSgiIiYd45SID6zhtXNYjluyD1c043ArX2E4v2mFMrxkjK0MXZBuhkP+3c+YigqirtU9CqhgfOAT/1Q== ARC-Authentication-Results: i=1; mx.microsoft.com 1; spf=pass smtp.mailfrom=os.amperecomputing.com; dmarc=pass action=none header.from=os.amperecomputing.com; dkim=pass header.d=os.amperecomputing.com; arc=none DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=os.amperecomputing.com; s=selector2; h=From:Date:Subject:Message-ID:Content-Type:MIME-Version:X-MS-Exchange-SenderADCheck; bh=AbOhFxaF/DlCgkheSGeJwrNSEr/k4tiSt54xTbgEBhw=; b=fIhyYjHWNeU84jxcsfnPdx6cidHvvjqr7vgz8SWibzojStofGFyzp2lcEUuw5m4v2CuQOIGPVrL/5WKfLjvvLA3ODW5iwvvwrtaHeWcHKaia/YvbTTuwBe7a6NDdpChXJH8oKebwqDm3fzeOw5V4F1KGpY70wlx03U81IeUxh2Q= Received: from CH0PR01MB6873.prod.exchangelabs.com (2603:10b6:610:112::22) by SJ2PR01MB8584.prod.exchangelabs.com (2603:10b6:a03:548::6) with Microsoft SMTP Server (version=TLS1_2, cipher=TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384) id 15.20.8466.12; Wed, 19 Feb 2025 01:12:46 +0000 Received: from CH0PR01MB6873.prod.exchangelabs.com ([fe80::3850:9112:f3bf:6460]) by CH0PR01MB6873.prod.exchangelabs.com ([fe80::3850:9112:f3bf:6460%6]) with mapi id 15.20.8466.009; Wed, 19 Feb 2025 01:12:44 +0000 Message-ID: Date: Tue, 18 Feb 2025 17:12:41 -0800 User-Agent: Mozilla Thunderbird Subject: Re: [PATCH] /dev/zero: make private mapping full anonymous mapping To: Oliver Sang Cc: oe-lkp@lists.linux.dev, lkp@intel.com, linux-kernel@vger.kernel.org, arnd@arndb.de, gregkh@linuxfoundation.org, Liam.Howlett@oracle.com, lorenzo.stoakes@oracle.com, vbabka@suse.cz, jannh@google.com, willy@infradead.org, liushixin2@huawei.com, akpm@linux-foundation.org, linux-mm@kvack.org References: <202501281038.617c6b60-lkp@intel.com> <01b5c112-480f-40dd-892f-5eb7c53e2f68@os.amperecomputing.com> <6f110208-eca3-42fa-aa7b-7d93f61ef750@os.amperecomputing.com> Content-Language: en-US From: Yang Shi In-Reply-To: Content-Type: text/plain; charset=UTF-8; format=flowed Content-Transfer-Encoding: 8bit X-ClientProxiedBy: SJ0PR03CA0186.namprd03.prod.outlook.com (2603:10b6:a03:2ef::11) To CH0PR01MB6873.prod.exchangelabs.com (2603:10b6:610:112::22) MIME-Version: 1.0 X-MS-PublicTrafficType: Email X-MS-TrafficTypeDiagnostic: CH0PR01MB6873:EE_|SJ2PR01MB8584:EE_ X-MS-Office365-Filtering-Correlation-Id: e3214924-f86b-43ba-aab6-08dd5082836d X-MS-Exchange-SenderADCheck: 1 X-MS-Exchange-AntiSpam-Relay: 0 X-Microsoft-Antispam: BCL:0;ARA:13230040|7416014|376014|1800799024|366016; X-Microsoft-Antispam-Message-Info: =?utf-8?B?T0MvbjVZUm5kSkFGK3dGS25uVUFwbjFINWlYTmZDTzhlcDhwSWNXZU5Fbmw3?= =?utf-8?B?VERsUkxUWndEeG5CWkRyS0VUUzR0VitFWU1xSEtxZ1piQTJnc09xeUxLeENZ?= =?utf-8?B?aUMyTndyWDNZbC9uUUJlRWU1eDBaUDEyYytOM2MrVytMak5kWjlFM3NFb2Ru?= =?utf-8?B?SGgvSjEycXdibktjNExRRjhSTlhFWDgzVXZvcUQrWkJ1b0lpaGg1djhMRVMx?= =?utf-8?B?TEg4Skd1K2RpaFJNYlBTTWhyUTI2T0hPaGZFK0M4WXhVbER6QVZLZzN2aVFa?= =?utf-8?B?YnVnQ01JNkx4alI3YktCbm0yVHNHaXRHc01pT3AwN3pDY0VwaW9IeGMwS0NC?= =?utf-8?B?WnZGdWswMElrTkpiUWJsRkIyQy9Bb2JDdG91TGVjQVF5NGRBWWwzZHg4YWtt?= =?utf-8?B?SUdvbmx4b0hvd0JuVk1OYWpNWjFqT3RmOFliMkxPVkJQQzNPWnZ1c2NISXow?= =?utf-8?B?OW55V29hWWtXYnh4TEQyZzl1dkNqN2JyWWpiNlp1K2dGWjdka1NVK29qbGQy?= =?utf-8?B?cW9YLzRta3MzN1IwRnFscWllbGRXd1hZeGR2c1BZd3h4c3g5VmhPYUlmT2Nl?= =?utf-8?B?SElCVlVEWTlOc3JxM1ZuVjFkS1hyV1ltRmwvUkx3bDU4L1dKVytVZXRsbmdE?= =?utf-8?B?Tm9MQi8ydC9KTURDRWpuSmRTNGhQVmJaU2owMkZwaElybk9mMzFpcnJxcmhu?= =?utf-8?B?YnN1WkYvVksvMFJxb1ZxNmxoZFIrRERBVkJBZUZiaGNQUlJnOHVoWjV4V0lJ?= =?utf-8?B?bnpYUjFLZ0IrQWRFVzdmZzJ4TU9CVVhPZFdwZjFHRkpaVHExZW1hRC9JWEV5?= =?utf-8?B?bmJPc2s5cnV4cFJSTGp2TkhUeVBNenZKRlI1NkllSEg2U0dIbFNhZWRNaWdz?= =?utf-8?B?OXpWeWtzN2RHek5hN2s1bEFMK1VhbHg1dUxtYjF3M0lsRVdUN0FEYjJ0THBk?= =?utf-8?B?MThqRnBNb201dUJ2dDVRZ0I3WHZsTTh0Vlkzb2szMEN3aU9STkQ1ekhXWGhY?= =?utf-8?B?VlQ2Q3htaWdWYmdVVVAvdU4zdDBhRHpKSERKTytEQ0JBM2xxYUVYQUJWbHB2?= =?utf-8?B?aE96ZGZQeUFTODQ2Mm5tSUxNVWZKRktabFJnYnNLQnNrUTlPM0t2V21ndWhH?= =?utf-8?B?elBzQmJ0NExLQzdaRS9lWmc0eDVWMVdCSXluQkRWSi8yYjFLQkRlMDB3YzY1?= =?utf-8?B?aFhuOTFHZjJOOVJVa01sZ2FvSUVZQXM5Sm1mSW94WE5aMGxHeDNDTjFDSEZu?= =?utf-8?B?OTJLUWg0cnJaaXNiSEk2Y0hqTzFYbE0rWGgrbkE0aUpwVk82a0xqS2lvVUx4?= =?utf-8?B?MkZoNjliSCt2c2V5SUlXOWJoZzIrYWtteVEzRWYwb2wvbXdDdVU3UE1QYUYx?= =?utf-8?B?SmRjSnROQVJKMmZ5NVBkQXJpMlIyd0VPckw3Z2ozQjdiam5tejJIYThEZ3A0?= =?utf-8?B?UnFtdE1lVU52NWVzWVlscjh2c3hGbitmTHJ4OWphQzJ6UEhadUE0QXYwRTg2?= =?utf-8?B?SUV0VVl2Zis3cTdqcUdTODROZDdsNy9DL2ZEQVBQbjZKSkM3dkFRNVNVWnY3?= =?utf-8?B?LzVOSXFsOFRPMmJBQTMwVUZpY2ZMQnVMTE9iU1lHRFo4TnRaU0p5Q0Z2SnF5?= =?utf-8?B?bjNlbHE3WTRoWXE3M0R6RW1ESHU1UE1KVUl2NzgwNjhhM09GdjZEc1FpUC9Z?= =?utf-8?B?djQ2Wm9QeGJPQkxINUtENjNiSmtlblh5NW93aFpuM05LOEVyZWp0UHYwSU9B?= =?utf-8?B?UnIvZEVVOG4rNWZEMUhRUFVuVkNiVm5KQkh2SjI4Qko4ZGdWbjZCWHFCczBz?= =?utf-8?B?dXNTd1lWa1NjUUowdjFISEdBR2wzbzQySFV1bVQzYUNCMjRmdXZUNGY3VnZn?= =?utf-8?Q?y4fTeaMNtFbUD?= X-Forefront-Antispam-Report: CIP:255.255.255.255;CTRY:;LANG:en;SCL:1;SRV:;IPV:NLI;SFV:NSPM;H:CH0PR01MB6873.prod.exchangelabs.com;PTR:;CAT:NONE;SFS:(13230040)(7416014)(376014)(1800799024)(366016);DIR:OUT;SFP:1102; X-MS-Exchange-AntiSpam-MessageData-ChunkCount: 1 X-MS-Exchange-AntiSpam-MessageData-0: =?utf-8?B?ellOcmpVUGlXNHVQVU5oK1Uxbm0zcUFLaVgzVm9tZ2t3VHRPeG9Qd3Y1Z0xp?= =?utf-8?B?bHdJbVlJazdKck40MHdjR2d6ZGo1S3JKV0JSakZwc3pwT2hrRTcydHlrcDBM?= =?utf-8?B?dHMvelhSSE1Lbmw3M0Y2VTl5bUZOK2JJcERqUDhTMm1UK2FtcnlJenBBVlpr?= =?utf-8?B?cis4ZVF1ajRBM0RRWk9pakRZc0NCY2tQMjZ4RmxSZ0piTTVLdHlqYXlDRlFH?= =?utf-8?B?TDE2QmorMG4wWWx4QlJtdm4vR0RGVENKLzM0QjJ4Y0RnSGNuNG44ZHhHQWNV?= =?utf-8?B?K25hTCt0QTBpT1FXczJhTHRqYjJIR3NRUmdEZjBKcm9jYzRLTzdNdGkyMDdm?= =?utf-8?B?OHBvbWt5OGtRT1M4VVI2N2xMc3NncE90UmtlZ3JwUlJSYXQyc0tvSW9qWVlq?= =?utf-8?B?bFVLNUZEcU4wM0xCLzhUaElSU2tGQU5iQm45em9xQm1BZVNyc3pqa0MvUlNU?= =?utf-8?B?NUxQVkUrdGJDTFhwdFpoc2VUY3FlVGN1T1lBN0JOUm54TFRaNWRHdUxRZE9I?= =?utf-8?B?bmRGVVgrZytsQ1E2YU1CclE0dGhCUjNzdzc2a1pwdDhPa3gyNFpvMFllRXNv?= =?utf-8?B?bkwwSE5XKytCWFVEZWRDNitpRTd6ZGZGQjVnVEJ1K0ZNa1VnVFM1c01BaC9Y?= =?utf-8?B?eEEySVJZTHNTMUdhZ2krTi8wclR2aXVMcWNTdk9EWU1DY0pQdk9sMkNoNDdL?= =?utf-8?B?bTVkYmp4alhGa0t0cXBtakdIOUtNSlJYb0tCV3RvOEVkcUYxOXVOb0g2aFE5?= =?utf-8?B?bTJJbnEycCtJYkhibWNDRDFqb0U4NVlXbDJHODlWZGlLbHRQeGsyMTJNOUlR?= =?utf-8?B?T1Z3NitKSklBZDh3b3NNOHBma1NDWjlOYTRYZWpyRU1CR2MyTUxZbHdxRSsr?= =?utf-8?B?ZnJ4Q0FYaERudk51ZTErZzd1U2NINDFtcGFtR1RDSmFMMXRXd0d3eWEvdW1y?= =?utf-8?B?dk1HRGZ5L2RlNjBUTDk5dFlwV05ocVVPa2hLSjZkWEpJRW1ycFMrc1B4Tkw4?= =?utf-8?B?VjVBVHJTZ1AzWGZDMU9RRHB3Yk5qRmY4VUZZRzRmQS8wUmZTbGpQcno2cVhM?= =?utf-8?B?cUFLTmcwZDJZWUxjdGxyeE1LR2N5amg4dnY4b0pNNjBqcjZsZFYxVXNwZnp3?= =?utf-8?B?NzVncmM0UWkzcnpwN1gxdTRBWkZONGN5a21YMWdnaUZ4c1VkeFNNNFhiYk91?= =?utf-8?B?aW1LK2p1UXR0SlM3clJxWHNZM21FN0swTVJwSCtqY3E3ekQ5U1VpY0U1Tmo5?= =?utf-8?B?cHRyUnFibEIwZGxteWI1TElycjdiK3NJZG1LalRmcFNrK29iemh2ZWp4MU90?= =?utf-8?B?U1pnd0VnU0htQ2RHR2hSUnNoeW5MdUZtWWlSMEdWSHZuUFY2dlY2VU9UbXBu?= =?utf-8?B?SlJ4Q3BPcGFDdlRzSm9RY3dDa25rbmhtSWRUcktReWYzZmsrRHVLeVZ6SEVw?= =?utf-8?B?ZmdROGlHVFdYdkpRcE8xQzcwVUVwbGJUUDNPeDFKT2U3M29FRUlrSXZZNGZs?= =?utf-8?B?NjR5VGdXRE1zVldxdEhUS1J2Lzkyb2F0TmxxdmRwMzA2OUZMc1RIc0ZBdkdH?= =?utf-8?B?cmo5cldpd3lLNEVUVjB5T0luVURtNUhManNwVVY4aFNIdkMxNkIwVU1NcnYy?= =?utf-8?B?Q3FDTEkyQWFMN1BYa0h3MUxTaUdWclNSZGwzcDdtQXA2UDgyNTlIOU90eWYx?= =?utf-8?B?dFl3M0dFRVI0N09IaEwzK3A2eUlLOUg4T3NTZXdpSzVvcS9LaXFXaUtFUmdi?= =?utf-8?B?eWNEVTJsdlVxSkE3ellEZUxvNUhSUnB5QURMREw0RzZCMkUzRWs1SjBnMkt4?= =?utf-8?B?QVZXL1pENHo0WE90blVkUUF3U1RWU2hPQWdnYjFkRjYyYy9JTHI0TlBIOWlU?= =?utf-8?B?NllEd05zcWtvZTJtSzlXcHZ2VER3VXBGU3JxSnpsOUJWWHJDUS9VQjArQXpF?= =?utf-8?B?dE05ZEdPdjVOU21ydlVOMExTVy9LT2ZyeUxFb2xBYWM3djdxMUxMYkRZeHhZ?= =?utf-8?B?Q0ZidHNmOWlzc0lTVjFQaWZPQzk0WmlaNVZtUVFtT3M4NUw2VWdhd2VpSFkr?= =?utf-8?B?NThXaFhscVNiQTNTVXJEVW4wNGkzYlFhZHRpcm9KaUN5eU9NYnpKczdlK0tS?= =?utf-8?B?dnZGMW9iUXdQczZLQjEzVm9DbkZDY2lkU2NIL1poblJGZTc2RSs2QUNCR3R3?= =?utf-8?B?R0E9PQ==?= X-OriginatorOrg: os.amperecomputing.com X-MS-Exchange-CrossTenant-Network-Message-Id: e3214924-f86b-43ba-aab6-08dd5082836d X-MS-Exchange-CrossTenant-AuthSource: CH0PR01MB6873.prod.exchangelabs.com X-MS-Exchange-CrossTenant-AuthAs: Internal X-MS-Exchange-CrossTenant-OriginalArrivalTime: 19 Feb 2025 01:12:43.9847 (UTC) X-MS-Exchange-CrossTenant-FromEntityHeader: Hosted X-MS-Exchange-CrossTenant-Id: 3bc2b170-fd94-476d-b0ce-4229bdc904a7 X-MS-Exchange-CrossTenant-MailboxType: HOSTED X-MS-Exchange-CrossTenant-UserPrincipalName: 9eDw1SdciV07PyMu4XBtRK141rojC8DhLhdU+NPljdc25DoUG7bSAY0PumYVLYByQZrzJhYtVXuwtUpBWBJ95gHQv61HZ0kNt+vAp0cmak4= X-MS-Exchange-Transport-CrossTenantHeadersStamped: SJ2PR01MB8584 X-Rspam-User: X-Rspamd-Server: rspam09 X-Rspamd-Queue-Id: 5CA861C0012 X-Stat-Signature: yt75to3k98ikjf81fn39xx8gbes57uit X-HE-Tag: 1739927572-620856 X-HE-Meta: U2FsdGVkX1/7y7qWUofMI8MN37IrjQocUYB4/jzp1zuIQ0MXt3ofIUsOf+/JDKepPYJyzccFbwaRK9B5x46olFWuvagRnJJjaUh9EbIitwyUY8/vFcvJbfaQ3YNKZrrEWfFc+kYe3tXK1itTVy9WJB/ebFhiV/lRlH2kRxd4UwN1FWmZpApUY4Ei1tyYYLjOtCpNJuKxyX4wPSTSUYuRFzN1CDZ7/EC8QX8WKiC8T7vX+uECXE5D/xXyl+iA4HJxaifovFPsJR9BErF+GJPzfeCCRVLr1CMDNrryzIUqUyPmrY3dI7+UWeLbE6+E8bJgfaF+3U4Cc7t58xTLQDxEBRjuBwPaWQDoAdQSHIxzf1pEAy7v2KLeOGZgjHZsKaj4/k/9wXgOXJQ2ZNtqEXRnxLSKrUqGpxDjxJwY4p4wfRUDRYkx8AFRTiWca4HbxEZY6aDwfmokiDrEZ6OeljTEuOYXCQsu1k7ILctmhc2qWX1TMnikF8s62Durc8nUBscnPjYjrbJeLgVxN8JW5nYkVU7/xHUyJhbQRrSM9RrwHPqAMb2ipLqWFsBLaKN+m+DGICzEa4nzQ9mfujHOZefLNpfxUMu5T7aXDi5opKwTZrnCCu0SI7rkAljTwXot894Fe4JEQvuOm2rj2Rvc7OMHsN3j/voge2BViiHOHKCif3qwBYOdHlz+uFHb+TC++MciMRWDKRNbnlUU4S1iz/0/lPxCcmOl7Bh8vFHEkl6WhYfVIDTbS+FcEn5zRbqY1x40s6lxD8/aJ5UdW7/tHLXQTc/9xMxWRBkoZmIqH+LGfgNqUIjYy/dYTAk23p532Ig/Hf57PifuWr4KsLJ+fwUJzeU195Xgtecyn0BLEu15kJ7zS+H3LNxDrOjZ0s9vkdkJxk8DSKif6NLqJfiKCggB4tBcPgKf0yr3J15+/+BiALY4QIW7QW9nSwXUXUk2tRw7F9SPvs6mWsWiX1ewahS 3pShmjZh xCSg57ZZ6SgUFAQ9osfDlT6O6i7uvnAnXUvpsxXoHFvRedSgg/jpRxCq3K9VFNeNL5iT1zq20+6ELg7AgeE9YuW6Ngfh4WmC4ZAF9iqlBewTx3tBTuIWR3YbSci0DCOSHONuQmoNIcmdmvQm0E8HFlbR8u20DOFb0X3z5KDNn57ol76ig0t7t5ZheDu1cKA1OICtZOBERDwGZIOh1E1HxT9hxPbt1n6sb3720S0FWJhnKVShKlo0WE/WMok0UxXayXOX4S/+AM8iIX8el46fW3/Mmxf1SDtUjjgQa/4J/XpMl6Ze6gb4mS4TpPBaXhrN2NMkJJM2p23/bj3iXCs0WczEb57nKz9OUemNNI1CjKgevV8+1ZsjEkk8+IGsoC63zH3GZu9695KsivAeBaOyQscSFfZF84noptx4J76yn3Fs9kz9TI7pBmSDNtQZ3Eu+7IcY6+E8sfN1oKWmjiNIuJpUjle+gufJTgVHw X-Bogosity: Ham, tests=bogofilter, spamicity=0.000000, version=1.2.4 Sender: owner-linux-mm@kvack.org Precedence: bulk X-Loop: owner-majordomo@kvack.org List-ID: List-Subscribe: List-Unsubscribe: On 2/17/25 10:30 PM, Oliver Sang wrote: > hi, Yang Shi, > > On Fri, Feb 14, 2025 at 02:53:37PM -0800, Yang Shi wrote: >> On 2/12/25 6:04 PM, Oliver Sang wrote: >>> hi, Yang Shi, >>> >>> On Fri, Feb 07, 2025 at 10:10:37AM -0800, Yang Shi wrote: >>>> On 2/6/25 12:02 AM, Oliver Sang wrote: >>> [...] >>> >>>>> since we applied your "/dev/zero: make private mapping full anonymous mapping" >>>>> patch upon a68d3cbfad like below: >>>>> >>>>> * 7143ee2391f1e /dev/zero: make private mapping full anonymous mapping >>>>> * a68d3cbfade64 memstick: core: fix kernel-doc notation >>>>> >>>>> so I applied below patch also upon a68d3cbfad. >>>>> >>>>> we saw big improvement but not that big. >>>>> >>>>> ========================================================================================= >>>>> compiler/cpufreq_governor/kconfig/rootfs/runtime/tbox_group/test/testcase: >>>>> gcc-12/performance/x86_64-rhel-9.4/debian-12-x86_64-20240206.cgz/300s/lkp-cpl-4sp2/small-allocs/vm-scalability >>>>> >>>>> commit: >>>>> a68d3cbfad ("memstick: core: fix kernel-doc notation") >>>>> 52ec85cb99 <--- your patch >>>>> >>>>> >>>>> a68d3cbfade64392 52ec85cb99e9b31dc304eae965a >>>>> ---------------- --------------------------- >>>>> %stddev %change %stddev >>>>> \ | \ >>>>> 14364828 ± 4% +410.6% 73349239 ± 3% vm-scalability.throughput >>>>> >>>>> full comparison as below [1] just FYI. >>>> Thanks for the update. I stared at the profiling report for a whole day, but >>>> I didn't figure out where that 400% lost. I just saw the number of page >>>> faults was fewer. And it seems like the reduction of page faults match the >>>> 400% loss. So I did more trace and profiling. >>>> >>>> The test case did the below stuff in a tight loop: >>>>   mmap 40K memory from /dev/zero (read only) >>>>   read the area >>>> >>>> So two major factors to the performance: mmap and page fault. The >>>> alternative patch did reduce the overhead of mmap to the same level as the >>>> original patch. >>>> >>>> The further perf profiling showed the cost of page fault is higher than the >>>> original patch. But the profiling of page fault was interesting: >>>> >>>> -   44.87%     0.01%  usemem [kernel.kallsyms]                   [k] >>>> do_translation_fault >>>>    - 44.86% do_translation_fault >>>>       - 44.83% do_page_fault >>>>          - 44.53% handle_mm_fault >>>>               9.04% __handle_mm_fault >>>> >>>> Page fault consumed 40% of cpu time in handle_mm_fault, but >>>> __handle_mm_fault just consumed 9%, I expected it should be the major >>>> consumer. >>>> >>>> So I annotated handle_mm_fault, then found the most time was consumed by >>>> lru_gen_enter_fault() -> vma_has_recency() (my kernel has multi-gen LRU >>>> enabled): >>>> >>>>       │     if (vma->vm_file && (vma->vm_file->f_mode & FMODE_NOREUSE)) >>>>        │     ↓ cbz     x1, b4 >>>>   0.00 │       ldr     w0, [x1, #12] >>>>  99.59 │       eor     x0, x0, #0x800000 >>>>   0.00 │       ubfx    w0, w0, #23, #1 >>>>        │     current->in_lru_fault = vma_has_recency(vma); >>>>   0.00 │ b4:   ldrh    w1, [x2, #1992] >>>>   0.01 │       bfi     w1, w0, #5, #1 >>>>   0.00 │       strh    w1, [x2, #1992] >>>> >>>> >>>> vma_has_recency() read vma->vm_file->f_mode if vma->vm_file is not NULL. But >>>> that load took a long time. So I inspected struct file and saw: >>>> >>>> struct file { >>>>     file_ref_t            f_ref; >>>>     spinlock_t            f_lock; >>>>     fmode_t                f_mode; >>>>     const struct file_operations    *f_op; >>>>     ... >>>> } >>>> >>>> The f_mode is in the same cache line with f_ref (my kernel does NOT have >>>> spin lock debug enabled). The test case mmap /dev/zero in a tight loop, so >>>> the refcount is modified (fget/fput) very frequently, this resulted in >>>> somehow false sharing. >>>> >>>> So I tried the below patch on top of the alternative patch: >>>> >>>> diff --git a/include/linux/mm_inline.h b/include/linux/mm_inline.h >>>> index f9157a0c42a5..ba11dc0b1c7c 100644 >>>> --- a/include/linux/mm_inline.h >>>> +++ b/include/linux/mm_inline.h >>>> @@ -608,6 +608,9 @@ static inline bool vma_has_recency(struct vm_area_struct >>>> *vma) >>>>         if (vma->vm_flags & (VM_SEQ_READ | VM_RAND_READ)) >>>>                 return false; >>>> >>>> +       if (vma_is_anonymous(vma)) >>>> +               return true; >>>> + >>>>         if (vma->vm_file && (vma->vm_file->f_mode & FMODE_NOREUSE)) >>>>                 return false; >>>> >>>> This made the profiling of page fault look normal: >>>> >>>>                         - 1.90% do_translation_fault >>>>                            - 1.87% do_page_fault >>>>                               - 1.49% handle_mm_fault >>>>                                  - 1.36% __handle_mm_fault >>>> >>>> Please try this in your test. >>>> >>>> But AFAICT I have never seen performance issue reported due to the false >>>> sharing of refcount and other fields in struct file. This benchmark stressed >>>> this quite badly. >>> I applied your above patch upon alternative patch last time, then saw more >>> improvement (+445.2% vs a68d3cbfad), but still not that big as in our original >>> report. >> Thanks for the update. It looks like the problem is still in page faults. I >> did my test on arm64 machine. I also noticed struct file has >> "__randomize_layout", so it may have different layout on x86 than arm64? >> >> The page fault handler may also access other fields of struct file that may >> cause false sharing, for example, accessing f_mapping to read gfp flags. >> This may not be a problem on my machine, but may be more costly on yours >> depending on the real layout of struct file on the machines, >> >> Can you please try the below patch on top of the current patches? Thank you >> so much for your patience. > you are welcome! > > now has more improvements. I just list "a68d3cbfad + 3 patches so far" vs > a68d3cbfad below, if you want more data, please let me know. > > ========================================================================================= > compiler/cpufreq_governor/kconfig/rootfs/runtime/tbox_group/test/testcase: > gcc-12/performance/x86_64-rhel-9.4/debian-12-x86_64-20240206.cgz/300s/lkp-cpl-4sp2/small-allocs/vm-scalability > > commit: > a68d3cbfad ("memstick: core: fix kernel-doc notation") > edc84ea79f <--- a68d3cbfad + 3 patches so far > > a68d3cbfade64392 edc84ea79f8dc11853076b96ad5 > ---------------- --------------------------- > %stddev %change %stddev > \ | \ > 14364828 ± 4% +685.6% 1.129e+08 ± 5% vm-scalability.throughput > > full data is as below [1] FYI. Thank you for the update. It is close to the 800% target, and it looks like there may be still some overhead in page fault handler due to the false sharing. For example, the vma_is_dax() call in __thp_vma_allowable_orders() which is called if pmd is null. I'm not sure how much the impact could be. However, I'm not sure whether we should continue chasing it or not. Because the false sharing in struct file should be very rare for real life workload. The workload has to map the same file then do page fault again and again in a tight loop, and the struct file is shared by multiple processes. Such behavior should be rare in real life. And changing the layout of struct file to avoid the false sharing sounds better than adding vma_is_anonymous() call in all the possible places. But it may introduce new false sharing. Having refcount in a dedicated cache line is doable too, however it will increase the size of struct file (from 192 bytes to 256 bytes). So neither seems worth it. We can split all the patches into two parts, the first part is to avoid i_mmap_rwsem contention, the second part is the struct file false sharing. IMHO the first part is more real. I can come up with a formal patch then send to the mailing list Thanks, Yang >> diff --git a/mm/memory.c b/mm/memory.c >> index 539c0f7c6d54..1fa9dbce0f66 100644 >> --- a/mm/memory.c >> +++ b/mm/memory.c >> @@ -3214,6 +3214,9 @@ static gfp_t __get_fault_gfp_mask(struct >> vm_area_struct *vma) >>  { >>         struct file *vm_file = vma->vm_file; >> >> +       if (vma_is_anonymous(vma)) >> +               return GFP_KERNEL; >> + >>         if (vm_file) >>                 return mapping_gfp_mask(vm_file->f_mapping) | __GFP_FS | >> __GFP_IO; >> > [1] > ========================================================================================= > compiler/cpufreq_governor/kconfig/rootfs/runtime/tbox_group/test/testcase: > gcc-12/performance/x86_64-rhel-9.4/debian-12-x86_64-20240206.cgz/300s/lkp-cpl-4sp2/small-allocs/vm-scalability > > commit: > a68d3cbfad ("memstick: core: fix kernel-doc notation") > edc84ea79f <--- a68d3cbfad + 3 patches so far > > a68d3cbfade64392 edc84ea79f8dc11853076b96ad5 > ---------------- --------------------------- > %stddev %change %stddev > \ | \ > 5.262e+09 ± 3% -59.8% 2.114e+09 ± 2% cpuidle..time > 7924008 ± 3% -83.9% 1275131 ± 5% cpuidle..usage > 1871164 ± 4% -16.8% 1557233 ± 8% numa-numastat.node3.local_node > 1952164 ± 3% -14.8% 1663189 ± 7% numa-numastat.node3.numa_hit > 399.52 -75.0% 99.77 ± 2% uptime.boot > 14507 -22.1% 11296 uptime.idle > 3408 ± 5% -99.8% 7.25 ± 46% perf-c2c.DRAM.local > 18076 ± 3% -99.8% 43.00 ±100% perf-c2c.DRAM.remote > 8082 ± 5% -99.8% 12.50 ± 63% perf-c2c.HITM.local > 6544 ± 6% -99.7% 22.88 ±151% perf-c2c.HITM.remote > 14627 ± 4% -99.8% 35.38 ±114% perf-c2c.HITM.total > 6.99 ± 3% +177.6% 19.41 ± 3% vmstat.cpu.id > 91.35 -28.5% 65.31 vmstat.cpu.sy > 1.71 +793.1% 15.25 ± 4% vmstat.cpu.us > 34204 ± 5% -64.1% 12271 ± 9% vmstat.system.cs > 266575 -21.2% 210049 vmstat.system.in > 6.49 ± 3% +10.0 16.46 ± 3% mpstat.cpu.all.idle% > 0.63 -0.3 0.34 ± 3% mpstat.cpu.all.irq% > 0.03 ± 2% +0.3 0.31 ± 4% mpstat.cpu.all.soft% > 91.17 -24.1 67.09 mpstat.cpu.all.sys% > 1.68 ± 2% +14.1 15.80 ± 4% mpstat.cpu.all.usr% > 337.33 -98.7% 4.25 ± 10% mpstat.max_utilization.seconds > 352.76 -84.7% 53.95 ± 4% time.elapsed_time > 352.76 -84.7% 53.95 ± 4% time.elapsed_time.max > 225965 ± 7% -17.1% 187329 ± 12% time.involuntary_context_switches > 9.592e+08 ± 4% +11.9% 1.074e+09 time.minor_page_faults > 20852 -10.0% 18761 time.percent_of_cpu_this_job_got > 72302 -88.6% 8227 ± 6% time.system_time > 1260 ± 3% +50.7% 1899 time.user_time > 5393707 ± 5% -98.8% 66895 ± 21% time.voluntary_context_switches > 1609925 -50.7% 793216 meminfo.Active > 1609925 -50.7% 793216 meminfo.Active(anon) > 160837 ± 33% -72.5% 44155 ± 9% meminfo.AnonHugePages > 4435665 -18.7% 3608195 meminfo.Cached > 1775547 -44.2% 990889 meminfo.Committed_AS > 148539 -47.4% 78096 meminfo.Mapped > 4245538 ± 4% -24.6% 3202495 meminfo.PageTables > 929777 -88.9% 102759 meminfo.Shmem > 25676018 ± 3% +14.3% 29335678 meminfo.max_used_kB > 64129 ± 4% +706.8% 517389 ± 7% vm-scalability.median > 45.40 ± 5% +2248.9 2294 ± 2% vm-scalability.stddev% > 14364828 ± 4% +685.6% 1.129e+08 ± 5% vm-scalability.throughput > 352.76 -84.7% 53.95 ± 4% vm-scalability.time.elapsed_time > 352.76 -84.7% 53.95 ± 4% vm-scalability.time.elapsed_time.max > 225965 ± 7% -17.1% 187329 ± 12% vm-scalability.time.involuntary_context_switches > 9.592e+08 ± 4% +11.9% 1.074e+09 vm-scalability.time.minor_page_faults > 20852 -10.0% 18761 vm-scalability.time.percent_of_cpu_this_job_got > 72302 -88.6% 8227 ± 6% vm-scalability.time.system_time > 1260 ± 3% +50.7% 1899 vm-scalability.time.user_time > 5393707 ± 5% -98.8% 66895 ± 21% vm-scalability.time.voluntary_context_switches > 4.316e+09 ± 4% +11.9% 4.832e+09 vm-scalability.workload > 1063552 ± 4% -24.9% 799008 ± 3% numa-meminfo.node0.PageTables > 125455 ±106% -85.5% 18164 ±165% numa-meminfo.node0.Shmem > 1062709 ± 4% -25.7% 789746 ± 4% numa-meminfo.node1.PageTables > 176171 ± 71% -92.4% 13303 ±230% numa-meminfo.node1.Shmem > 35515 ± 91% -97.3% 976.55 ± 59% numa-meminfo.node2.Mapped > 1058901 ± 4% -25.3% 791392 ± 4% numa-meminfo.node2.PageTables > 770405 ± 30% -79.2% 160245 ±101% numa-meminfo.node3.Active > 770405 ± 30% -79.2% 160245 ±101% numa-meminfo.node3.Active(anon) > 380096 ± 50% -62.5% 142513 ± 98% numa-meminfo.node3.AnonPages.max > 1146977 ±108% -92.8% 82894 ± 60% numa-meminfo.node3.FilePages > 52663 ± 47% -97.2% 1488 ± 39% numa-meminfo.node3.Mapped > 1058539 ± 4% -22.3% 821992 ± 3% numa-meminfo.node3.PageTables > 558943 ± 14% -93.7% 35227 ±124% numa-meminfo.node3.Shmem > 265763 ± 4% -24.9% 199601 ± 3% numa-vmstat.node0.nr_page_table_pages > 31364 ±106% -85.5% 4539 ±165% numa-vmstat.node0.nr_shmem > 265546 ± 4% -25.5% 197854 ± 5% numa-vmstat.node1.nr_page_table_pages > 44052 ± 71% -92.5% 3323 ±230% numa-vmstat.node1.nr_shmem > 8961 ± 91% -97.3% 244.02 ± 59% numa-vmstat.node2.nr_mapped > 264589 ± 4% -25.2% 197920 ± 3% numa-vmstat.node2.nr_page_table_pages > 192683 ± 30% -79.2% 40126 ±101% numa-vmstat.node3.nr_active_anon > 286819 ±108% -92.8% 20761 ± 60% numa-vmstat.node3.nr_file_pages > 13124 ± 49% -97.2% 372.02 ± 39% numa-vmstat.node3.nr_mapped > 264499 ± 4% -22.4% 205376 ± 3% numa-vmstat.node3.nr_page_table_pages > 139810 ± 14% -93.7% 8844 ±124% numa-vmstat.node3.nr_shmem > 192683 ± 30% -79.2% 40126 ±101% numa-vmstat.node3.nr_zone_active_anon > 1951359 ± 3% -14.9% 1661427 ± 7% numa-vmstat.node3.numa_hit > 1870359 ± 4% -16.8% 1555470 ± 8% numa-vmstat.node3.numa_local > 402515 -50.7% 198246 proc-vmstat.nr_active_anon > 170568 +1.8% 173591 proc-vmstat.nr_anon_pages > 1109246 -18.7% 902238 proc-vmstat.nr_file_pages > 37525 -47.3% 19768 proc-vmstat.nr_mapped > 1059932 ± 4% -24.2% 803105 ± 2% proc-vmstat.nr_page_table_pages > 232507 -89.0% 25623 proc-vmstat.nr_shmem > 37297 -5.4% 35299 proc-vmstat.nr_slab_reclaimable > 402515 -50.7% 198246 proc-vmstat.nr_zone_active_anon > 61931 ± 8% -83.9% 9948 ± 59% proc-vmstat.numa_hint_faults > 15755 ± 21% -96.6% 541.38 ± 36% proc-vmstat.numa_hint_faults_local > 6916516 ± 3% -8.0% 6360040 proc-vmstat.numa_hit > 6568542 ± 3% -8.5% 6012265 proc-vmstat.numa_local > 293942 ± 3% -68.8% 91724 ± 48% proc-vmstat.numa_pte_updates > 9.608e+08 ± 4% +11.8% 1.074e+09 proc-vmstat.pgfault > 55981 ± 2% -68.7% 17541 ± 2% proc-vmstat.pgreuse > 0.82 ± 4% -51.0% 0.40 ± 8% perf-stat.i.MPKI > 2.714e+10 ± 2% +378.3% 1.298e+11 ± 9% perf-stat.i.branch-instructions > 0.11 ± 3% +0.1 0.24 ± 8% perf-stat.i.branch-miss-rate% > 24932893 +306.8% 1.014e+08 ± 9% perf-stat.i.branch-misses > 64.93 -7.5 57.48 perf-stat.i.cache-miss-rate% > 88563288 ± 3% +35.0% 1.196e+08 ± 7% perf-stat.i.cache-misses > 1.369e+08 ± 3% +43.7% 1.968e+08 ± 7% perf-stat.i.cache-references > 34508 ± 4% -47.3% 18199 ± 9% perf-stat.i.context-switches > 7.67 -75.7% 1.87 ± 3% perf-stat.i.cpi > 224605 +22.5% 275084 ± 6% perf-stat.i.cpu-clock > 696.35 ± 2% -53.5% 323.77 ± 2% perf-stat.i.cpu-migrations > 10834 ± 4% -24.1% 8224 ± 11% perf-stat.i.cycles-between-cache-misses > 1.102e+11 +282.2% 4.212e+11 ± 9% perf-stat.i.instructions > 0.14 +334.6% 0.62 ± 5% perf-stat.i.ipc > 24.25 ± 3% +626.9% 176.25 ± 4% perf-stat.i.metric.K/sec > 2722043 ± 3% +803.8% 24600740 ± 9% perf-stat.i.minor-faults > 2722043 ± 3% +803.8% 24600739 ± 9% perf-stat.i.page-faults > 224605 +22.5% 275084 ± 6% perf-stat.i.task-clock > 0.81 ± 3% -62.2% 0.31 ± 11% perf-stat.overall.MPKI > 0.09 -0.0 0.08 ± 2% perf-stat.overall.branch-miss-rate% > 64.81 -2.4 62.37 perf-stat.overall.cache-miss-rate% > 7.24 -70.7% 2.12 ± 5% perf-stat.overall.cpi > 8933 ± 4% -21.9% 6978 ± 7% perf-stat.overall.cycles-between-cache-misses > 0.14 +242.2% 0.47 ± 5% perf-stat.overall.ipc > 9012 ± 2% -57.8% 3806 perf-stat.overall.path-length > 2.701e+10 ± 2% +285.4% 1.041e+11 ± 5% perf-stat.ps.branch-instructions > 24708939 +215.8% 78042343 ± 4% perf-stat.ps.branch-misses > 89032538 ± 3% +15.9% 1.032e+08 ± 8% perf-stat.ps.cache-misses > 1.374e+08 ± 3% +20.6% 1.656e+08 ± 9% perf-stat.ps.cache-references > 34266 ± 5% -66.2% 11570 ± 10% perf-stat.ps.context-switches > 223334 -1.6% 219861 perf-stat.ps.cpu-clock > 7.941e+11 -9.9% 7.157e+11 perf-stat.ps.cpu-cycles > 693.54 ± 2% -67.2% 227.38 ± 4% perf-stat.ps.cpu-migrations > 1.097e+11 +208.3% 3.381e+11 ± 5% perf-stat.ps.instructions > 2710577 ± 3% +626.7% 19698901 ± 5% perf-stat.ps.minor-faults > 2710577 ± 3% +626.7% 19698902 ± 5% perf-stat.ps.page-faults > 223334 -1.6% 219861 perf-stat.ps.task-clock > 3.886e+13 ± 2% -52.7% 1.839e+13 perf-stat.total.instructions > 64052898 ± 5% -99.9% 81213 ± 23% sched_debug.cfs_rq:/.avg_vruntime.avg > 95701822 ± 7% -96.4% 3425672 ± 7% sched_debug.cfs_rq:/.avg_vruntime.max > 43098762 ± 6% -100.0% 153.42 ± 36% sched_debug.cfs_rq:/.avg_vruntime.min > 9223270 ± 9% -95.9% 380347 ± 16% sched_debug.cfs_rq:/.avg_vruntime.stddev > 0.00 ± 22% -100.0% 0.00 sched_debug.cfs_rq:/.h_nr_delayed.avg > 0.69 ± 8% -100.0% 0.00 sched_debug.cfs_rq:/.h_nr_delayed.max > 0.05 ± 12% -100.0% 0.00 sched_debug.cfs_rq:/.h_nr_delayed.stddev > 0.78 ± 2% -94.5% 0.04 ± 21% sched_debug.cfs_rq:/.h_nr_running.avg > 1.97 ± 5% -49.3% 1.00 sched_debug.cfs_rq:/.h_nr_running.max > 0.28 ± 7% -29.1% 0.20 ± 10% sched_debug.cfs_rq:/.h_nr_running.stddev > 411536 ± 58% -100.0% 1.15 ±182% sched_debug.cfs_rq:/.left_deadline.avg > 43049468 ± 22% -100.0% 258.27 ±182% sched_debug.cfs_rq:/.left_deadline.max > 3836405 ± 37% -100.0% 17.22 ±182% sched_debug.cfs_rq:/.left_deadline.stddev > 411536 ± 58% -100.0% 1.06 ±191% sched_debug.cfs_rq:/.left_vruntime.avg > 43049467 ± 22% -100.0% 236.56 ±191% sched_debug.cfs_rq:/.left_vruntime.max > 3836405 ± 37% -100.0% 15.77 ±191% sched_debug.cfs_rq:/.left_vruntime.stddev > 64052901 ± 5% -99.9% 81213 ± 23% sched_debug.cfs_rq:/.min_vruntime.avg > 95701822 ± 7% -96.4% 3425672 ± 7% sched_debug.cfs_rq:/.min_vruntime.max > 43098762 ± 6% -100.0% 153.42 ± 36% sched_debug.cfs_rq:/.min_vruntime.min > 9223270 ± 9% -95.9% 380347 ± 16% sched_debug.cfs_rq:/.min_vruntime.stddev > 0.77 ± 2% -94.4% 0.04 ± 21% sched_debug.cfs_rq:/.nr_running.avg > 1.50 ± 9% -33.3% 1.00 sched_debug.cfs_rq:/.nr_running.max > 0.26 ± 10% -22.7% 0.20 ± 10% sched_debug.cfs_rq:/.nr_running.stddev > 1.61 ± 24% +413.4% 8.24 ± 60% sched_debug.cfs_rq:/.removed.runnable_avg.avg > 86.69 +508.6% 527.62 ± 4% sched_debug.cfs_rq:/.removed.runnable_avg.max > 11.14 ± 13% +428.4% 58.87 ± 32% sched_debug.cfs_rq:/.removed.runnable_avg.stddev > 1.61 ± 24% +413.3% 8.24 ± 60% sched_debug.cfs_rq:/.removed.util_avg.avg > 86.69 +508.6% 527.62 ± 4% sched_debug.cfs_rq:/.removed.util_avg.max > 11.14 ± 13% +428.4% 58.87 ± 32% sched_debug.cfs_rq:/.removed.util_avg.stddev > 411536 ± 58% -100.0% 1.06 ±191% sched_debug.cfs_rq:/.right_vruntime.avg > 43049467 ± 22% -100.0% 236.56 ±191% sched_debug.cfs_rq:/.right_vruntime.max > 3836405 ± 37% -100.0% 15.77 ±191% sched_debug.cfs_rq:/.right_vruntime.stddev > 769.03 -84.7% 117.79 ± 3% sched_debug.cfs_rq:/.util_avg.avg > 1621 ± 5% -32.7% 1092 ± 16% sched_debug.cfs_rq:/.util_avg.max > 159.12 ± 8% +33.2% 211.88 ± 7% sched_debug.cfs_rq:/.util_avg.stddev > 724.17 ± 2% -98.6% 10.41 ± 32% sched_debug.cfs_rq:/.util_est.avg > 1360 ± 15% -51.5% 659.38 ± 10% sched_debug.cfs_rq:/.util_est.max > 234.34 ± 9% -68.2% 74.43 ± 18% sched_debug.cfs_rq:/.util_est.stddev > 766944 ± 3% +18.9% 912012 sched_debug.cpu.avg_idle.avg > 1067639 ± 5% +25.5% 1339736 ± 9% sched_debug.cpu.avg_idle.max > 3799 ± 7% -38.3% 2346 ± 23% sched_debug.cpu.avg_idle.min > 321459 ± 2% -36.6% 203909 ± 7% sched_debug.cpu.avg_idle.stddev > 195573 -76.9% 45144 sched_debug.cpu.clock.avg > 195596 -76.9% 45160 sched_debug.cpu.clock.max > 195548 -76.9% 45123 sched_debug.cpu.clock.min > 13.79 ± 3% -36.0% 8.83 ± 2% sched_debug.cpu.clock.stddev > 194424 -76.8% 45019 sched_debug.cpu.clock_task.avg > 194608 -76.8% 45145 sched_debug.cpu.clock_task.max > 181834 -82.1% 32559 sched_debug.cpu.clock_task.min > 4241 ± 2% -96.8% 136.38 ± 21% sched_debug.cpu.curr->pid.avg > 9799 ± 2% -59.8% 3934 sched_debug.cpu.curr->pid.max > 1365 ± 10% -49.1% 695.11 ± 10% sched_debug.cpu.curr->pid.stddev > 537665 ± 4% +28.3% 690006 ± 6% sched_debug.cpu.max_idle_balance_cost.max > 3119 ± 56% +479.5% 18078 ± 29% sched_debug.cpu.max_idle_balance_cost.stddev > 0.00 ± 12% -68.3% 0.00 ± 17% sched_debug.cpu.next_balance.stddev > 0.78 ± 2% -95.3% 0.04 ± 20% sched_debug.cpu.nr_running.avg > 2.17 ± 8% -53.8% 1.00 sched_debug.cpu.nr_running.max > 0.29 ± 8% -35.4% 0.19 ± 9% sched_debug.cpu.nr_running.stddev > 25773 ± 5% -97.0% 764.82 ± 3% sched_debug.cpu.nr_switches.avg > 48669 ± 10% -77.2% 11080 ± 12% sched_debug.cpu.nr_switches.max > 19006 ± 7% -99.2% 151.12 ± 15% sched_debug.cpu.nr_switches.min > 4142 ± 8% -69.5% 1264 ± 6% sched_debug.cpu.nr_switches.stddev > 0.07 ± 23% -93.3% 0.01 ± 53% sched_debug.cpu.nr_uninterruptible.avg > 240.19 ± 16% -80.2% 47.50 ± 44% sched_debug.cpu.nr_uninterruptible.max > -77.92 -88.1% -9.25 sched_debug.cpu.nr_uninterruptible.min > 37.87 ± 5% -84.7% 5.78 ± 13% sched_debug.cpu.nr_uninterruptible.stddev > 195549 -76.9% 45130 sched_debug.cpu_clk > 194699 -77.3% 44280 sched_debug.ktime > 0.00 -100.0% 0.00 sched_debug.rt_rq:.rt_nr_running.avg > 0.17 -100.0% 0.00 sched_debug.rt_rq:.rt_nr_running.max > 0.01 -100.0% 0.00 sched_debug.rt_rq:.rt_nr_running.stddev > 196368 -76.6% 45975 sched_debug.sched_clk > 95.59 -95.6 0.00 perf-profile.calltrace.cycles-pp.__mmap > 95.54 -95.5 0.00 perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.__mmap > 95.54 -95.5 0.00 perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.__mmap > 94.54 -94.5 0.00 perf-profile.calltrace.cycles-pp.ksys_mmap_pgoff.do_syscall_64.entry_SYSCALL_64_after_hwframe.__mmap > 94.46 -94.4 0.07 ±264% perf-profile.calltrace.cycles-pp.vm_mmap_pgoff.ksys_mmap_pgoff.do_syscall_64.entry_SYSCALL_64_after_hwframe.__mmap > 94.45 -94.0 0.41 ±158% perf-profile.calltrace.cycles-pp.do_mmap.vm_mmap_pgoff.ksys_mmap_pgoff.do_syscall_64.entry_SYSCALL_64_after_hwframe > 94.14 -93.9 0.29 ±134% perf-profile.calltrace.cycles-pp.__mmap_new_vma.__mmap_region.do_mmap.vm_mmap_pgoff.ksys_mmap_pgoff > 94.25 -93.8 0.41 ±158% perf-profile.calltrace.cycles-pp.__mmap_region.do_mmap.vm_mmap_pgoff.ksys_mmap_pgoff.do_syscall_64 > 93.79 -93.7 0.07 ±264% perf-profile.calltrace.cycles-pp.vma_link_file.__mmap_new_vma.__mmap_region.do_mmap.vm_mmap_pgoff > 93.44 -93.4 0.00 perf-profile.calltrace.cycles-pp.down_write.vma_link_file.__mmap_new_vma.__mmap_region.do_mmap > 93.40 -93.4 0.00 perf-profile.calltrace.cycles-pp.rwsem_down_write_slowpath.down_write.vma_link_file.__mmap_new_vma.__mmap_region > 93.33 -93.3 0.00 perf-profile.calltrace.cycles-pp.rwsem_optimistic_spin.rwsem_down_write_slowpath.down_write.vma_link_file.__mmap_new_vma > 92.89 -92.9 0.00 perf-profile.calltrace.cycles-pp.osq_lock.rwsem_optimistic_spin.rwsem_down_write_slowpath.down_write.vma_link_file > 0.00 +1.7 1.69 ± 65% perf-profile.calltrace.cycles-pp.dup_mm.copy_process.kernel_clone.__do_sys_clone.do_syscall_64 > 0.00 +1.9 1.90 ± 55% perf-profile.calltrace.cycles-pp.__mmput.exit_mm.do_exit.do_group_exit.__x64_sys_exit_group > 0.00 +1.9 1.90 ± 55% perf-profile.calltrace.cycles-pp.exit_mm.do_exit.do_group_exit.__x64_sys_exit_group.x64_sys_call > 0.00 +1.9 1.93 ± 53% perf-profile.calltrace.cycles-pp.proc_reg_read_iter.vfs_read.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe > 0.00 +1.9 1.93 ± 53% perf-profile.calltrace.cycles-pp.seq_read_iter.proc_reg_read_iter.vfs_read.ksys_read.do_syscall_64 > 0.00 +2.0 1.99 ± 53% perf-profile.calltrace.cycles-pp.copy_process.kernel_clone.__do_sys_clone.do_syscall_64.entry_SYSCALL_64_after_hwframe > 0.00 +2.0 2.02 ± 64% perf-profile.calltrace.cycles-pp.do_pte_missing.__handle_mm_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault > 0.00 +2.3 2.27 ± 56% perf-profile.calltrace.cycles-pp.__do_sys_clone.do_syscall_64.entry_SYSCALL_64_after_hwframe._Fork > 0.00 +2.3 2.27 ± 56% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe._Fork > 0.00 +2.3 2.27 ± 56% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe._Fork > 0.00 +2.3 2.27 ± 56% perf-profile.calltrace.cycles-pp.kernel_clone.__do_sys_clone.do_syscall_64.entry_SYSCALL_64_after_hwframe._Fork > 0.00 +2.4 2.45 ± 53% perf-profile.calltrace.cycles-pp._Fork > 0.00 +2.5 2.51 ± 52% perf-profile.calltrace.cycles-pp.__x64_sys_exit_group.x64_sys_call.do_syscall_64.entry_SYSCALL_64_after_hwframe > 0.00 +2.5 2.51 ± 52% perf-profile.calltrace.cycles-pp.do_exit.do_group_exit.__x64_sys_exit_group.x64_sys_call.do_syscall_64 > 0.00 +2.5 2.51 ± 52% perf-profile.calltrace.cycles-pp.do_group_exit.__x64_sys_exit_group.x64_sys_call.do_syscall_64.entry_SYSCALL_64_after_hwframe > 0.00 +2.5 2.51 ± 52% perf-profile.calltrace.cycles-pp.x64_sys_call.do_syscall_64.entry_SYSCALL_64_after_hwframe > 0.00 +3.2 3.17 ± 42% perf-profile.calltrace.cycles-pp.__handle_mm_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault > 0.00 +3.3 3.28 ± 52% perf-profile.calltrace.cycles-pp.perf_mmap__push.record__mmap_read_evlist.__cmd_record.cmd_record.run_builtin > 0.00 +3.3 3.28 ± 52% perf-profile.calltrace.cycles-pp.record__mmap_read_evlist.__cmd_record.cmd_record.run_builtin.handle_internal_command > 0.00 +4.1 4.10 ± 45% perf-profile.calltrace.cycles-pp.__cmd_record.cmd_record.run_builtin.handle_internal_command.main > 0.00 +4.1 4.10 ± 45% perf-profile.calltrace.cycles-pp.cmd_record.run_builtin.handle_internal_command.main > 0.00 +4.8 4.80 ± 61% perf-profile.calltrace.cycles-pp.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.acpi_safe_halt.acpi_idle_do_entry.acpi_idle_enter > 0.00 +5.0 4.98 ± 69% perf-profile.calltrace.cycles-pp.generic_perform_write.shmem_file_write_iter.vfs_write.ksys_write.do_syscall_64 > 0.00 +5.1 5.07 ± 71% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.write.writen.record__pushfn > 0.00 +5.1 5.07 ± 71% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.write.writen.record__pushfn.perf_mmap__push > 0.00 +5.1 5.07 ± 71% perf-profile.calltrace.cycles-pp.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe.write.writen > 0.00 +5.1 5.07 ± 71% perf-profile.calltrace.cycles-pp.shmem_file_write_iter.vfs_write.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe > 0.00 +5.1 5.07 ± 71% perf-profile.calltrace.cycles-pp.write.writen.record__pushfn.perf_mmap__push.record__mmap_read_evlist > 0.00 +5.1 5.07 ± 71% perf-profile.calltrace.cycles-pp.writen.record__pushfn.perf_mmap__push.record__mmap_read_evlist.__cmd_record > 0.00 +5.1 5.11 ± 47% perf-profile.calltrace.cycles-pp.exit_mmap.__mmput.exit_mm.do_exit.do_group_exit > 0.00 +5.1 5.12 ± 70% perf-profile.calltrace.cycles-pp.record__pushfn.perf_mmap__push.record__mmap_read_evlist.__cmd_record.cmd_record > 0.00 +6.1 6.08 ± 50% perf-profile.calltrace.cycles-pp.acpi_safe_halt.acpi_idle_do_entry.acpi_idle_enter.cpuidle_enter_state.cpuidle_enter > 0.00 +7.8 7.84 ± 21% perf-profile.calltrace.cycles-pp.vfs_read.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe.read > 0.00 +7.9 7.88 ± 20% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.read > 0.00 +7.9 7.88 ± 20% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.read > 0.00 +7.9 7.88 ± 20% perf-profile.calltrace.cycles-pp.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe.read > 0.00 +7.9 7.88 ± 20% perf-profile.calltrace.cycles-pp.read > 0.00 +11.1 11.10 ± 41% perf-profile.calltrace.cycles-pp.handle_internal_command.main > 0.00 +11.1 11.10 ± 41% perf-profile.calltrace.cycles-pp.main > 0.00 +11.1 11.10 ± 41% perf-profile.calltrace.cycles-pp.run_builtin.handle_internal_command.main > 0.00 +11.2 11.18 ± 73% perf-profile.calltrace.cycles-pp.vfs_write.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe.write > 0.00 +15.9 15.94 ± 41% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe > 0.00 +15.9 15.94 ± 41% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe > 0.00 +19.5 19.54 ± 41% perf-profile.calltrace.cycles-pp.asm_sysvec_reschedule_ipi.acpi_safe_halt.acpi_idle_do_entry.acpi_idle_enter.cpuidle_enter_state > 1.21 ± 3% +36.7 37.86 ± 7% perf-profile.calltrace.cycles-pp.acpi_idle_do_entry.acpi_idle_enter.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call > 1.21 ± 3% +36.7 37.86 ± 7% perf-profile.calltrace.cycles-pp.acpi_idle_enter.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call.do_idle > 1.21 ± 3% +37.0 38.24 ± 7% perf-profile.calltrace.cycles-pp.cpuidle_enter.cpuidle_idle_call.do_idle.cpu_startup_entry.start_secondary > 1.21 ± 3% +37.2 38.41 ± 7% perf-profile.calltrace.cycles-pp.cpuidle_idle_call.do_idle.cpu_startup_entry.start_secondary.common_startup_64 > 1.21 ± 3% +37.4 38.57 ± 6% perf-profile.calltrace.cycles-pp.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call.do_idle.cpu_startup_entry > 1.22 ± 3% +38.5 39.67 ± 7% perf-profile.calltrace.cycles-pp.cpu_startup_entry.start_secondary.common_startup_64 > 1.22 ± 3% +38.5 39.67 ± 7% perf-profile.calltrace.cycles-pp.do_idle.cpu_startup_entry.start_secondary.common_startup_64 > 1.22 ± 3% +38.5 39.67 ± 7% perf-profile.calltrace.cycles-pp.start_secondary.common_startup_64 > 1.22 ± 3% +38.9 40.09 ± 6% perf-profile.calltrace.cycles-pp.common_startup_64 > 2.19 ± 3% +45.2 47.41 ± 14% perf-profile.calltrace.cycles-pp.asm_sysvec_apic_timer_interrupt.acpi_safe_halt.acpi_idle_do_entry.acpi_idle_enter.cpuidle_enter_state > 95.60 -95.4 0.22 ±135% perf-profile.children.cycles-pp.__mmap > 94.55 -93.9 0.60 ±103% perf-profile.children.cycles-pp.ksys_mmap_pgoff > 94.14 -93.7 0.44 ±112% perf-profile.children.cycles-pp.__mmap_new_vma > 93.79 -93.7 0.10 ±264% perf-profile.children.cycles-pp.vma_link_file > 94.46 -93.5 0.96 ± 76% perf-profile.children.cycles-pp.vm_mmap_pgoff > 94.45 -93.5 0.96 ± 76% perf-profile.children.cycles-pp.do_mmap > 94.25 -93.4 0.86 ± 87% perf-profile.children.cycles-pp.__mmap_region > 93.40 -93.4 0.00 perf-profile.children.cycles-pp.rwsem_down_write_slowpath > 93.33 -93.3 0.00 perf-profile.children.cycles-pp.rwsem_optimistic_spin > 93.44 -93.2 0.22 ±149% perf-profile.children.cycles-pp.down_write > 92.91 -92.9 0.00 perf-profile.children.cycles-pp.osq_lock > 95.58 -45.4 50.16 ± 8% perf-profile.children.cycles-pp.entry_SYSCALL_64_after_hwframe > 95.58 -45.4 50.16 ± 8% perf-profile.children.cycles-pp.do_syscall_64 > 0.00 +1.1 1.12 ± 74% perf-profile.children.cycles-pp.filemap_map_pages > 0.00 +1.1 1.12 ± 76% perf-profile.children.cycles-pp.vfs_fstatat > 0.00 +1.2 1.19 ± 35% perf-profile.children.cycles-pp.vsnprintf > 0.00 +1.2 1.20 ± 46% perf-profile.children.cycles-pp.seq_printf > 0.00 +1.3 1.28 ± 78% perf-profile.children.cycles-pp.__do_sys_newfstatat > 0.00 +1.5 1.54 ± 75% perf-profile.children.cycles-pp.folios_put_refs > 0.00 +1.6 1.56 ± 52% perf-profile.children.cycles-pp.__cond_resched > 0.00 +1.6 1.60 ± 32% perf-profile.children.cycles-pp.sched_balance_newidle > 0.00 +1.7 1.69 ± 65% perf-profile.children.cycles-pp.dup_mm > 0.00 +1.9 1.93 ± 53% perf-profile.children.cycles-pp.proc_reg_read_iter > 0.00 +2.0 1.99 ± 53% perf-profile.children.cycles-pp.copy_process > 0.00 +2.1 2.06 ± 51% perf-profile.children.cycles-pp.__x64_sys_ioctl > 0.00 +2.1 2.08 ± 45% perf-profile.children.cycles-pp.proc_single_show > 0.00 +2.1 2.14 ± 45% perf-profile.children.cycles-pp.seq_read > 0.00 +2.2 2.16 ± 47% perf-profile.children.cycles-pp.ioctl > 0.00 +2.2 2.17 ± 33% perf-profile.children.cycles-pp.schedule > 0.00 +2.2 2.20 ± 28% perf-profile.children.cycles-pp.__pick_next_task > 0.00 +2.2 2.21 ± 47% perf-profile.children.cycles-pp.perf_evsel__run_ioctl > 0.00 +2.3 2.26 ± 58% perf-profile.children.cycles-pp.do_read_fault > 0.00 +2.3 2.27 ± 56% perf-profile.children.cycles-pp.__do_sys_clone > 0.00 +2.3 2.27 ± 56% perf-profile.children.cycles-pp.kernel_clone > 0.00 +2.4 2.37 ± 58% perf-profile.children.cycles-pp.zap_present_ptes > 0.00 +2.4 2.45 ± 53% perf-profile.children.cycles-pp._Fork > 0.00 +2.6 2.59 ± 53% perf-profile.children.cycles-pp.__x64_sys_exit_group > 0.00 +2.6 2.59 ± 53% perf-profile.children.cycles-pp.x64_sys_call > 0.00 +2.6 2.64 ± 44% perf-profile.children.cycles-pp.do_pte_missing > 0.00 +3.1 3.13 ± 59% perf-profile.children.cycles-pp.zap_pte_range > 0.00 +3.2 3.21 ± 58% perf-profile.children.cycles-pp.zap_pmd_range > 0.00 +3.4 3.40 ± 56% perf-profile.children.cycles-pp.unmap_page_range > 0.00 +3.4 3.43 ± 55% perf-profile.children.cycles-pp.unmap_vmas > 0.19 ± 23% +3.9 4.06 ± 45% perf-profile.children.cycles-pp.__handle_mm_fault > 0.51 ± 6% +4.0 4.49 ± 38% perf-profile.children.cycles-pp.handle_mm_fault > 0.04 ± 44% +4.0 4.04 ± 28% perf-profile.children.cycles-pp.__schedule > 0.77 ± 3% +4.4 5.18 ± 39% perf-profile.children.cycles-pp.exc_page_fault > 0.76 ± 3% +4.4 5.18 ± 39% perf-profile.children.cycles-pp.do_user_addr_fault > 0.58 ± 2% +4.7 5.26 ± 53% perf-profile.children.cycles-pp.sysvec_apic_timer_interrupt > 0.00 +5.1 5.07 ± 71% perf-profile.children.cycles-pp.writen > 0.00 +5.1 5.07 ± 69% perf-profile.children.cycles-pp.generic_perform_write > 0.00 +5.1 5.12 ± 47% perf-profile.children.cycles-pp.exit_mm > 0.00 +5.1 5.12 ± 70% perf-profile.children.cycles-pp.record__pushfn > 0.00 +5.1 5.12 ± 70% perf-profile.children.cycles-pp.shmem_file_write_iter > 1.18 +5.5 6.69 ± 33% perf-profile.children.cycles-pp.asm_exc_page_fault > 0.00 +6.2 6.24 ± 43% perf-profile.children.cycles-pp.__mmput > 0.00 +6.2 6.24 ± 43% perf-profile.children.cycles-pp.exit_mmap > 0.00 +7.0 7.00 ± 51% perf-profile.children.cycles-pp.perf_mmap__push > 0.00 +7.0 7.00 ± 51% perf-profile.children.cycles-pp.record__mmap_read_evlist > 0.00 +7.2 7.25 ± 52% perf-profile.children.cycles-pp.__fput > 0.00 +7.3 7.35 ± 20% perf-profile.children.cycles-pp.seq_read_iter > 0.00 +7.8 7.84 ± 21% perf-profile.children.cycles-pp.vfs_read > 0.00 +7.9 7.88 ± 20% perf-profile.children.cycles-pp.ksys_read > 0.00 +7.9 7.88 ± 20% perf-profile.children.cycles-pp.read > 0.00 +9.9 9.93 ± 41% perf-profile.children.cycles-pp.asm_sysvec_reschedule_ipi > 0.02 ±141% +11.1 11.10 ± 41% perf-profile.children.cycles-pp.__cmd_record > 0.02 ±141% +11.1 11.10 ± 41% perf-profile.children.cycles-pp.cmd_record > 0.02 ±141% +11.1 11.10 ± 41% perf-profile.children.cycles-pp.handle_internal_command > 0.02 ±141% +11.1 11.10 ± 41% perf-profile.children.cycles-pp.main > 0.02 ±141% +11.1 11.10 ± 41% perf-profile.children.cycles-pp.run_builtin > 0.00 +11.2 11.18 ± 73% perf-profile.children.cycles-pp.vfs_write > 0.00 +11.2 11.23 ± 73% perf-profile.children.cycles-pp.ksys_write > 0.00 +11.2 11.23 ± 73% perf-profile.children.cycles-pp.write > 0.00 +13.6 13.61 ± 44% perf-profile.children.cycles-pp.do_exit > 0.00 +13.6 13.61 ± 44% perf-profile.children.cycles-pp.do_group_exit > 1.70 ± 2% +25.0 26.72 ± 15% perf-profile.children.cycles-pp.asm_sysvec_apic_timer_interrupt > 1.21 ± 3% +36.6 37.81 ± 7% perf-profile.children.cycles-pp.acpi_safe_halt > 1.21 ± 3% +36.6 37.86 ± 7% perf-profile.children.cycles-pp.acpi_idle_do_entry > 1.21 ± 3% +36.6 37.86 ± 7% perf-profile.children.cycles-pp.acpi_idle_enter > 1.21 ± 3% +37.4 38.57 ± 6% perf-profile.children.cycles-pp.cpuidle_enter_state > 1.21 ± 3% +37.4 38.66 ± 6% perf-profile.children.cycles-pp.cpuidle_enter > 1.22 ± 3% +37.6 38.82 ± 6% perf-profile.children.cycles-pp.cpuidle_idle_call > 1.22 ± 3% +38.5 39.67 ± 7% perf-profile.children.cycles-pp.start_secondary > 1.22 ± 3% +38.9 40.09 ± 6% perf-profile.children.cycles-pp.common_startup_64 > 1.22 ± 3% +38.9 40.09 ± 6% perf-profile.children.cycles-pp.cpu_startup_entry > 1.22 ± 3% +38.9 40.09 ± 6% perf-profile.children.cycles-pp.do_idle > 92.37 -92.4 0.00 perf-profile.self.cycles-pp.osq_lock > 1.19 ± 3% +30.7 31.90 ± 7% perf-profile.self.cycles-pp.acpi_safe_halt > 0.17 ±142% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.__do_fault.do_read_fault.do_pte_missing.__handle_mm_fault > 0.19 ± 34% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.__kmalloc_cache_noprof.perf_event_mmap_event.perf_event_mmap.__mmap_region > 0.14 ± 55% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.__wait_for_common.stop_two_cpus.migrate_swap.task_numa_migrate > 0.14 ± 73% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.change_pud_range.isra.0.change_protection_range > 0.10 ± 66% -99.9% 0.00 ±264% perf-sched.sch_delay.avg.ms.__cond_resched.down_write.__mmap_new_vma.__mmap_region.do_mmap > 0.11 ± 59% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.down_write.vma_link_file.__mmap_new_vma.__mmap_region > 0.04 ±132% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.down_write_killable.exec_mmap.begin_new_exec.load_elf_binary > 0.07 ±101% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.down_write_killable.vm_mmap_pgoff.ksys_mmap_pgoff.do_syscall_64 > 0.02 ± 31% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.dput.__fput.__x64_sys_close.do_syscall_64 > 0.02 ±143% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.kmem_cache_alloc_noprof.mas_alloc_nodes.mas_preallocate.__mmap_new_vma > 0.10 ± 44% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.kmem_cache_alloc_noprof.vm_area_alloc.__mmap_new_vma.__mmap_region > 0.12 ±145% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.mutex_lock.perf_poll.do_poll.constprop > 0.04 ± 55% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm > 0.25 ± 41% -98.5% 0.00 ±105% perf-sched.sch_delay.avg.ms.__cond_resched.stop_one_cpu.migrate_task_to.task_numa_migrate.isra > 0.11 ± 59% -97.1% 0.00 ± 61% perf-sched.sch_delay.avg.ms.__cond_resched.stop_one_cpu.sched_exec.bprm_execve.part > 0.40 ± 50% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown] > 0.32 ±104% -100.0% 0.00 perf-sched.sch_delay.avg.ms.do_nanosleep.hrtimer_nanosleep.common_nsleep.__x64_sys_clock_nanosleep > 0.01 ± 12% -100.0% 0.00 perf-sched.sch_delay.avg.ms.irq_thread.kthread.ret_from_fork.ret_from_fork_asm > 0.08 ± 28% -99.5% 0.00 ±264% perf-sched.sch_delay.avg.ms.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown] > 0.18 ± 57% -96.8% 0.01 ±193% perf-sched.sch_delay.avg.ms.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt.[unknown] > 0.03 ± 83% -100.0% 0.00 perf-sched.sch_delay.avg.ms.irqentry_exit_to_user_mode.asm_sysvec_call_function_single.[unknown] > 0.01 ± 20% -100.0% 0.00 perf-sched.sch_delay.avg.ms.irqentry_exit_to_user_mode.asm_sysvec_call_function_single.[unknown].[unknown] > 0.02 ± 65% -100.0% 0.00 perf-sched.sch_delay.avg.ms.irqentry_exit_to_user_mode.asm_sysvec_reschedule_ipi.[unknown] > 0.32 ± 47% -98.2% 0.01 ± 42% perf-sched.sch_delay.avg.ms.schedule_hrtimeout_range.do_poll.constprop.0.do_sys_poll > 0.19 ±185% -96.5% 0.01 ± 33% perf-sched.sch_delay.avg.ms.schedule_hrtimeout_range.ep_poll.do_epoll_wait.__x64_sys_epoll_wait > 0.07 ± 20% -100.0% 0.00 perf-sched.sch_delay.avg.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write.vma_link_file > 0.26 ± 17% -100.0% 0.00 perf-sched.sch_delay.avg.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone > 0.02 ± 60% -94.2% 0.00 ±264% perf-sched.sch_delay.avg.ms.schedule_timeout.kcompactd.kthread.ret_from_fork > 0.01 ±128% -100.0% 0.00 perf-sched.sch_delay.avg.ms.wait_for_partner.fifo_open.do_dentry_open.vfs_open > 1.00 ±151% -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.__do_fault.do_read_fault.do_pte_missing.__handle_mm_fault > 25.45 ± 94% -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.__kmalloc_cache_noprof.perf_event_mmap_event.perf_event_mmap.__mmap_region > 4.56 ± 67% -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.__wait_for_common.stop_two_cpus.migrate_swap.task_numa_migrate > 3.55 ± 97% -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.change_pud_range.isra.0.change_protection_range > 2.13 ± 67% -100.0% 0.00 ±264% perf-sched.sch_delay.max.ms.__cond_resched.down_write.__mmap_new_vma.__mmap_region.do_mmap > 3.16 ± 78% -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.down_write.vma_link_file.__mmap_new_vma.__mmap_region > 0.30 ±159% -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.down_write_killable.exec_mmap.begin_new_exec.load_elf_binary > 1.61 ±100% -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.down_write_killable.vm_mmap_pgoff.ksys_mmap_pgoff.do_syscall_64 > 0.03 ± 86% -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.dput.__fput.__x64_sys_close.do_syscall_64 > 0.20 ±182% -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.kmem_cache_alloc_noprof.mas_alloc_nodes.mas_preallocate.__mmap_new_vma > 3.51 ± 21% -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.kmem_cache_alloc_noprof.vm_area_alloc.__mmap_new_vma.__mmap_region > 0.83 ±160% -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.mutex_lock.perf_poll.do_poll.constprop > 0.09 ± 31% -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm > 3.59 ± 11% -99.9% 0.00 ±105% perf-sched.sch_delay.max.ms.__cond_resched.stop_one_cpu.migrate_task_to.task_numa_migrate.isra > 1.60 ± 69% -99.6% 0.01 ±129% perf-sched.sch_delay.max.ms.__cond_resched.stop_one_cpu.sched_exec.bprm_execve.part > 0.81 ± 43% -100.0% 0.00 perf-sched.sch_delay.max.ms.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown] > 1.02 ± 88% -100.0% 0.00 perf-sched.sch_delay.max.ms.do_nanosleep.hrtimer_nanosleep.common_nsleep.__x64_sys_clock_nanosleep > 0.02 ± 7% -100.0% 0.00 perf-sched.sch_delay.max.ms.irq_thread.kthread.ret_from_fork.ret_from_fork_asm > 9.68 ± 32% -100.0% 0.00 ±264% perf-sched.sch_delay.max.ms.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown] > 12.26 ±109% -100.0% 0.01 ±193% perf-sched.sch_delay.max.ms.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt.[unknown] > 5.60 ±139% -100.0% 0.00 perf-sched.sch_delay.max.ms.irqentry_exit_to_user_mode.asm_sysvec_call_function_single.[unknown] > 0.03 ±106% -100.0% 0.00 perf-sched.sch_delay.max.ms.irqentry_exit_to_user_mode.asm_sysvec_call_function_single.[unknown].[unknown] > 2.11 ± 61% -100.0% 0.00 perf-sched.sch_delay.max.ms.irqentry_exit_to_user_mode.asm_sysvec_reschedule_ipi.[unknown] > 3.67 ± 25% -99.8% 0.01 ± 16% perf-sched.sch_delay.max.ms.schedule_hrtimeout_range.do_poll.constprop.0.do_sys_poll > 1.65 ±187% -99.3% 0.01 ± 23% perf-sched.sch_delay.max.ms.schedule_hrtimeout_range.ep_poll.do_epoll_wait.__x64_sys_epoll_wait > 37.84 ± 47% -100.0% 0.00 perf-sched.sch_delay.max.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write.vma_link_file > 4.68 ± 36% -100.0% 0.00 perf-sched.sch_delay.max.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone > 0.21 ±169% -99.6% 0.00 ±264% perf-sched.sch_delay.max.ms.schedule_timeout.kcompactd.kthread.ret_from_fork > 7.92 ±131% -99.2% 0.06 ± 92% perf-sched.sch_delay.max.ms.syscall_exit_to_user_mode.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown] > 0.36 ±186% -100.0% 0.00 perf-sched.sch_delay.max.ms.wait_for_partner.fifo_open.do_dentry_open.vfs_open > 33.45 ± 3% -91.6% 2.81 ± 90% perf-sched.total_wait_and_delay.average.ms > 97903 ± 4% -98.2% 1776 ± 28% perf-sched.total_wait_and_delay.count.ms > 2942 ± 23% -95.2% 141.09 ± 36% perf-sched.total_wait_and_delay.max.ms > 33.37 ± 3% -91.9% 2.69 ± 95% perf-sched.total_wait_time.average.ms > 2942 ± 23% -96.7% 97.14 ± 19% perf-sched.total_wait_time.max.ms > 3.97 ± 6% -100.0% 0.00 perf-sched.wait_and_delay.avg.ms.__cond_resched.__kmalloc_cache_noprof.perf_event_mmap_event.perf_event_mmap.__mmap_region > 3.08 ± 4% -94.3% 0.18 ± 92% perf-sched.wait_and_delay.avg.ms.__cond_resched.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr.__sched_setaffinity > 119.91 ± 38% -100.0% 0.00 perf-sched.wait_and_delay.avg.ms.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm > 433.73 ± 41% -100.0% 0.00 perf-sched.wait_and_delay.avg.ms.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown] > 302.41 ± 5% -100.0% 0.00 perf-sched.wait_and_delay.avg.ms.irq_thread.kthread.ret_from_fork.ret_from_fork_asm > 1.48 ± 6% -100.0% 0.00 perf-sched.wait_and_delay.avg.ms.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown] > 23.24 ± 25% -96.7% 0.76 ± 27% perf-sched.wait_and_delay.avg.ms.pipe_read.vfs_read.ksys_read.do_syscall_64 > 327.16 ± 9% -99.8% 0.76 ±188% perf-sched.wait_and_delay.avg.ms.schedule_hrtimeout_range.do_poll.constprop.0.do_sys_poll > 369.37 ± 2% -98.9% 4.03 ±204% perf-sched.wait_and_delay.avg.ms.schedule_hrtimeout_range.ep_poll.do_epoll_wait.__x64_sys_epoll_wait > 0.96 ± 6% -100.0% 0.00 perf-sched.wait_and_delay.avg.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write.vma_link_file > 453.60 -100.0% 0.00 perf-sched.wait_and_delay.avg.ms.schedule_timeout.kcompactd.kthread.ret_from_fork > 187.66 -96.7% 6.11 ±109% perf-sched.wait_and_delay.avg.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm > 2.37 ± 29% -99.6% 0.01 ±264% perf-sched.wait_and_delay.avg.ms.syscall_exit_to_user_mode.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown] > 750.07 -99.3% 5.10 ± 84% perf-sched.wait_and_delay.avg.ms.worker_thread.kthread.ret_from_fork.ret_from_fork_asm > 1831 ± 9% -100.0% 0.00 perf-sched.wait_and_delay.count.__cond_resched.__kmalloc_cache_noprof.perf_event_mmap_event.perf_event_mmap.__mmap_region > 1269 ± 8% -45.8% 688.12 ± 21% perf-sched.wait_and_delay.count.__cond_resched.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr.__sched_setaffinity > 6.17 ± 45% -100.0% 0.00 perf-sched.wait_and_delay.count.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm > 5.00 -100.0% 0.00 perf-sched.wait_and_delay.count.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown] > 14.33 ± 5% -100.0% 0.00 perf-sched.wait_and_delay.count.irq_thread.kthread.ret_from_fork.ret_from_fork_asm > 810.00 ± 10% -100.0% 0.00 perf-sched.wait_and_delay.count.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown] > 3112 ± 24% -97.9% 65.75 ±106% perf-sched.wait_and_delay.count.pipe_read.vfs_read.ksys_read.do_syscall_64 > 40.50 ± 8% -98.8% 0.50 ±173% perf-sched.wait_and_delay.count.schedule_hrtimeout_range.do_poll.constprop.0.do_sys_poll > 73021 ± 3% -100.0% 0.00 perf-sched.wait_and_delay.count.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write.vma_link_file > 40.00 -100.0% 0.00 perf-sched.wait_and_delay.count.schedule_timeout.kcompactd.kthread.ret_from_fork > 1122 -99.0% 10.88 ± 98% perf-sched.wait_and_delay.count.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread > 11323 ± 3% -93.6% 722.25 ± 20% perf-sched.wait_and_delay.count.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm > 1887 ± 45% -100.0% 0.88 ±264% perf-sched.wait_and_delay.count.syscall_exit_to_user_mode.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown] > 1238 -93.9% 75.62 ± 79% perf-sched.wait_and_delay.count.worker_thread.kthread.ret_from_fork.ret_from_fork_asm > 35.19 ± 57% -100.0% 0.00 perf-sched.wait_and_delay.max.ms.__cond_resched.__kmalloc_cache_noprof.perf_event_mmap_event.perf_event_mmap.__mmap_region > 1002 -91.0% 89.82 ± 93% perf-sched.wait_and_delay.max.ms.__cond_resched.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr.__sched_setaffinity > 318.48 ± 65% -100.0% 0.00 perf-sched.wait_and_delay.max.ms.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm > 1000 -100.0% 0.00 perf-sched.wait_and_delay.max.ms.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown] > 966.90 ± 7% -100.0% 0.00 perf-sched.wait_and_delay.max.ms.irq_thread.kthread.ret_from_fork.ret_from_fork_asm > 20.79 ± 19% -100.0% 0.00 perf-sched.wait_and_delay.max.ms.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown] > 1043 -98.4% 16.64 ±214% perf-sched.wait_and_delay.max.ms.pipe_read.vfs_read.ksys_read.do_syscall_64 > 1240 ± 20% -99.9% 1.52 ±188% perf-sched.wait_and_delay.max.ms.schedule_hrtimeout_range.do_poll.constprop.0.do_sys_poll > 500.34 -96.9% 15.38 ±232% perf-sched.wait_and_delay.max.ms.schedule_hrtimeout_range.ep_poll.do_epoll_wait.__x64_sys_epoll_wait > 58.83 ± 39% -100.0% 0.00 perf-sched.wait_and_delay.max.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write.vma_link_file > 505.17 -100.0% 0.00 perf-sched.wait_and_delay.max.ms.schedule_timeout.kcompactd.kthread.ret_from_fork > 19.77 ± 55% -62.8% 7.36 ± 85% perf-sched.wait_and_delay.max.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread > 1237 ± 34% -91.7% 102.88 ± 33% perf-sched.wait_and_delay.max.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm > 1001 -100.0% 0.05 ±264% perf-sched.wait_and_delay.max.ms.syscall_exit_to_user_mode.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown] > 2794 ± 24% -97.9% 59.20 ± 61% perf-sched.wait_and_delay.max.ms.worker_thread.kthread.ret_from_fork.ret_from_fork_asm > 49.27 ±119% -100.0% 0.01 ±264% perf-sched.wait_time.avg.ms.__cond_resched.__alloc_pages_noprof.alloc_pages_mpol_noprof.folio_alloc_mpol_noprof.shmem_alloc_folio > 58.17 ±187% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.__do_fault.do_read_fault.do_pte_missing.__handle_mm_fault > 3.78 ± 5% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.__kmalloc_cache_noprof.perf_event_mmap_event.perf_event_mmap.__mmap_region > 2.99 ± 4% -97.0% 0.09 ± 91% perf-sched.wait_time.avg.ms.__cond_resched.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr.__sched_setaffinity > 3.92 ± 5% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.__wait_for_common.stop_two_cpus.migrate_swap.task_numa_migrate > 4.71 ± 8% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.change_pud_range.isra.0.change_protection_range > 1.67 ± 20% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.down_write.__mmap_new_vma.__mmap_region.do_mmap > 2.10 ± 27% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.down_write.vma_link_file.__mmap_new_vma.__mmap_region > 0.01 ± 44% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.down_write_killable.exec_mmap.begin_new_exec.load_elf_binary > 1.67 ± 21% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.down_write_killable.vm_mmap_pgoff.ksys_mmap_pgoff.do_syscall_64 > 0.04 ±133% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.dput.__fput.__x64_sys_close.do_syscall_64 > 67.14 ± 73% -99.5% 0.32 ±177% perf-sched.wait_time.avg.ms.__cond_resched.generic_perform_write.shmem_file_write_iter.vfs_write.ksys_write > 1.65 ± 67% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc_noprof.mas_alloc_nodes.mas_preallocate.__mmap_new_vma > 2.30 ± 14% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc_noprof.vm_area_alloc.__mmap_new_vma.__mmap_region > 42.44 ±200% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.mutex_lock.perf_poll.do_poll.constprop > 152.73 ±152% -100.0% 0.06 ±249% perf-sched.wait_time.avg.ms.__cond_resched.shmem_inode_acct_blocks.shmem_alloc_and_add_folio.shmem_get_folio_gfp.shmem_write_begin > 119.87 ± 38% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm > 3.80 ± 18% -99.9% 0.00 ±105% perf-sched.wait_time.avg.ms.__cond_resched.stop_one_cpu.migrate_task_to.task_numa_migrate.isra > 433.32 ± 41% -100.0% 0.00 perf-sched.wait_time.avg.ms.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown] > 250.23 ±107% -100.0% 0.00 perf-sched.wait_time.avg.ms.do_nanosleep.hrtimer_nanosleep.common_nsleep.__x64_sys_clock_nanosleep > 29.19 ± 5% -99.2% 0.25 ± 24% perf-sched.wait_time.avg.ms.do_task_dead.do_exit.do_group_exit.__x64_sys_exit_group.x64_sys_call > 302.40 ± 5% -100.0% 0.00 perf-sched.wait_time.avg.ms.irq_thread.kthread.ret_from_fork.ret_from_fork_asm > 1.40 ± 6% -100.0% 0.00 perf-sched.wait_time.avg.ms.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown] > 4.03 ± 8% -99.9% 0.01 ±193% perf-sched.wait_time.avg.ms.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt.[unknown] > 35.38 ±192% -100.0% 0.00 ±264% perf-sched.wait_time.avg.ms.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt.[unknown].[unknown] > 0.05 ± 40% -100.0% 0.00 perf-sched.wait_time.avg.ms.irqentry_exit_to_user_mode.asm_sysvec_call_function_single.[unknown] > 0.72 ±220% -100.0% 0.00 perf-sched.wait_time.avg.ms.irqentry_exit_to_user_mode.asm_sysvec_call_function_single.[unknown].[unknown] > 1.00 ±120% -99.9% 0.00 ±264% perf-sched.wait_time.avg.ms.irqentry_exit_to_user_mode.asm_sysvec_reschedule_ipi.[unknown] > 23.07 ± 24% -97.1% 0.67 ± 10% perf-sched.wait_time.avg.ms.pipe_read.vfs_read.ksys_read.do_syscall_64 > 326.84 ± 9% -99.6% 1.19 ±108% perf-sched.wait_time.avg.ms.schedule_hrtimeout_range.do_poll.constprop.0.do_sys_poll > 369.18 ± 2% -98.7% 4.72 ±167% perf-sched.wait_time.avg.ms.schedule_hrtimeout_range.ep_poll.do_epoll_wait.__x64_sys_epoll_wait > 0.89 ± 6% -100.0% 0.00 perf-sched.wait_time.avg.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write.vma_link_file > 1.17 ± 16% -99.7% 0.00 ±264% perf-sched.wait_time.avg.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone > 453.58 -100.0% 0.00 perf-sched.wait_time.avg.ms.schedule_timeout.kcompactd.kthread.ret_from_fork > 4.42 -25.4% 3.30 ± 17% perf-sched.wait_time.avg.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread > 187.58 -96.8% 6.05 ±110% perf-sched.wait_time.avg.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm > 2.36 ± 29% -99.1% 0.02 ± 84% perf-sched.wait_time.avg.ms.syscall_exit_to_user_mode.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown] > 0.01 ±156% -100.0% 0.00 perf-sched.wait_time.avg.ms.wait_for_partner.fifo_open.do_dentry_open.vfs_open > 750.01 -99.5% 3.45 ±141% perf-sched.wait_time.avg.ms.worker_thread.kthread.ret_from_fork.ret_from_fork_asm > 340.69 ±135% -100.0% 0.01 ±264% perf-sched.wait_time.max.ms.__cond_resched.__alloc_pages_noprof.alloc_pages_mpol_noprof.folio_alloc_mpol_noprof.shmem_alloc_folio > 535.09 ±128% -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.__do_fault.do_read_fault.do_pte_missing.__handle_mm_fault > 22.04 ± 32% -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.__kmalloc_cache_noprof.perf_event_mmap_event.perf_event_mmap.__mmap_region > 1001 -95.5% 44.91 ± 93% perf-sched.wait_time.max.ms.__cond_resched.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr.__sched_setaffinity > 13.57 ± 17% -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.__wait_for_common.stop_two_cpus.migrate_swap.task_numa_migrate > 13.54 ± 10% -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.change_pud_range.isra.0.change_protection_range > 10.17 ± 19% -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.down_write.__mmap_new_vma.__mmap_region.do_mmap > 11.35 ± 25% -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.down_write.vma_link_file.__mmap_new_vma.__mmap_region > 0.01 ± 32% -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.down_write_killable.exec_mmap.begin_new_exec.load_elf_binary > 10.62 ± 9% -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.down_write_killable.vm_mmap_pgoff.ksys_mmap_pgoff.do_syscall_64 > 0.20 ±199% -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.dput.__fput.__x64_sys_close.do_syscall_64 > 1559 ± 64% -100.0% 0.44 ±167% perf-sched.wait_time.max.ms.__cond_resched.generic_perform_write.shmem_file_write_iter.vfs_write.ksys_write > 6.93 ± 53% -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc_noprof.mas_alloc_nodes.mas_preallocate.__mmap_new_vma > 14.42 ± 22% -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc_noprof.vm_area_alloc.__mmap_new_vma.__mmap_region > 159.10 ±148% -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.mutex_lock.perf_poll.do_poll.constprop > 391.02 ±171% -100.0% 0.12 ±256% perf-sched.wait_time.max.ms.__cond_resched.shmem_inode_acct_blocks.shmem_alloc_and_add_folio.shmem_get_folio_gfp.shmem_write_begin > 318.43 ± 65% -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm > 13.14 ± 21% -100.0% 0.00 ±105% perf-sched.wait_time.max.ms.__cond_resched.stop_one_cpu.migrate_task_to.task_numa_migrate.isra > 1000 -100.0% 0.00 perf-sched.wait_time.max.ms.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown] > 500.84 ± 99% -100.0% 0.00 perf-sched.wait_time.max.ms.do_nanosleep.hrtimer_nanosleep.common_nsleep.__x64_sys_clock_nanosleep > 641.50 ± 23% -99.2% 5.27 ± 76% perf-sched.wait_time.max.ms.do_task_dead.do_exit.do_group_exit.__x64_sys_exit_group.x64_sys_call > 10.75 ± 98% -89.8% 1.10 ± 78% perf-sched.wait_time.max.ms.do_wait.kernel_wait4.do_syscall_64.entry_SYSCALL_64_after_hwframe > 966.89 ± 7% -100.0% 0.00 perf-sched.wait_time.max.ms.irq_thread.kthread.ret_from_fork.ret_from_fork_asm > 15.80 ± 8% -100.0% 0.00 perf-sched.wait_time.max.ms.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown] > 16.69 ± 10% -100.0% 0.01 ±193% perf-sched.wait_time.max.ms.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt.[unknown] > 41.71 ±158% -100.0% 0.00 ±264% perf-sched.wait_time.max.ms.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt.[unknown].[unknown] > 11.64 ± 61% -100.0% 0.00 perf-sched.wait_time.max.ms.irqentry_exit_to_user_mode.asm_sysvec_call_function_single.[unknown] > 2.94 ±213% -100.0% 0.00 perf-sched.wait_time.max.ms.irqentry_exit_to_user_mode.asm_sysvec_call_function_single.[unknown].[unknown] > 175.70 ±210% -100.0% 0.00 ±264% perf-sched.wait_time.max.ms.irqentry_exit_to_user_mode.asm_sysvec_reschedule_ipi.[unknown] > 1043 -99.6% 4.46 ±105% perf-sched.wait_time.max.ms.pipe_read.vfs_read.ksys_read.do_syscall_64 > 1240 ± 20% -99.8% 2.37 ±108% perf-sched.wait_time.max.ms.schedule_hrtimeout_range.do_poll.constprop.0.do_sys_poll > 500.11 -96.5% 17.32 ±201% perf-sched.wait_time.max.ms.schedule_hrtimeout_range.ep_poll.do_epoll_wait.__x64_sys_epoll_wait > 32.65 ± 33% -100.0% 0.00 perf-sched.wait_time.max.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write.vma_link_file > 22.94 ± 56% -100.0% 0.00 ±264% perf-sched.wait_time.max.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone > 505.00 -100.0% 0.00 perf-sched.wait_time.max.ms.schedule_timeout.kcompactd.kthread.ret_from_fork > 12.20 ± 43% -59.2% 4.98 perf-sched.wait_time.max.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread > 1237 ± 34% -92.5% 92.94 ± 20% perf-sched.wait_time.max.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm > 1000 -100.0% 0.09 ±111% perf-sched.wait_time.max.ms.syscall_exit_to_user_mode.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown] > 0.36 ±190% -100.0% 0.00 perf-sched.wait_time.max.ms.wait_for_partner.fifo_open.do_dentry_open.vfs_open > 2794 ± 24% -98.9% 30.12 ±114% perf-sched.wait_time.max.ms.worker_thread.kthread.ret_from_fork.ret_from_fork_asm