From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org Received: from kanga.kvack.org (kanga.kvack.org [205.233.56.17]) by smtp.lore.kernel.org (Postfix) with ESMTP id 3C443C02196 for ; Thu, 6 Feb 2025 08:02:46 +0000 (UTC) Received: by kanga.kvack.org (Postfix) id D04DD280001; Thu, 6 Feb 2025 03:02:45 -0500 (EST) Received: by kanga.kvack.org (Postfix, from userid 40) id CB43B6B0093; Thu, 6 Feb 2025 03:02:45 -0500 (EST) X-Delivered-To: int-list-linux-mm@kvack.org Received: by kanga.kvack.org (Postfix, from userid 63042) id ADF93280001; Thu, 6 Feb 2025 03:02:45 -0500 (EST) X-Delivered-To: linux-mm@kvack.org Received: from relay.hostedemail.com (smtprelay0012.hostedemail.com [216.40.44.12]) by kanga.kvack.org (Postfix) with ESMTP id 873836B0092 for ; Thu, 6 Feb 2025 03:02:45 -0500 (EST) Received: from smtpin14.hostedemail.com (a10.router.float.18 [10.200.18.1]) by unirelay03.hostedemail.com (Postfix) with ESMTP id 45F31A0F55 for ; Thu, 6 Feb 2025 08:02:45 +0000 (UTC) X-FDA: 83088778290.14.01A2EC2 Received: from mgamail.intel.com (mgamail.intel.com [192.198.163.11]) by imf10.hostedemail.com (Postfix) with ESMTP id 07A43C000E for ; Thu, 6 Feb 2025 08:02:40 +0000 (UTC) Authentication-Results: imf10.hostedemail.com; dkim=pass header.d=intel.com header.s=Intel header.b=Bs+mNcQ+; spf=pass (imf10.hostedemail.com: domain of oliver.sang@intel.com designates 192.198.163.11 as permitted sender) smtp.mailfrom=oliver.sang@intel.com; dmarc=pass (policy=none) header.from=intel.com; arc=reject ("signature check failed: fail, {[1] = sig:microsoft.com:reject}") ARC-Message-Signature: i=2; a=rsa-sha256; c=relaxed/relaxed; d=hostedemail.com; s=arc-20220608; t=1738828961; h=from:from:sender:reply-to:subject:subject:date:date: message-id:message-id:to:to:cc:cc:mime-version:mime-version: content-type:content-type: content-transfer-encoding:content-transfer-encoding: in-reply-to:in-reply-to:references:references:dkim-signature; bh=rR7k87Q6NXBvur3rDFnIO2ZJV/gprw+E+LzGccoveU4=; b=gmvWwy4hANXf+M8NU99xa6gC9k2KVGBqfrur2x8Dd6he1DgCaFsN2vobCSDm27MQG4wDVL IfLCpHo3Y+SzozdaEQgGqbQcJAGkp9ZVAV5GWYonl+81A60Ez8Q9TPghVwVCdGIx2ukcNO 5R5Mo6HxU5B/E8F2VgCSoohZQSe9ErE= ARC-Authentication-Results: i=2; imf10.hostedemail.com; dkim=pass header.d=intel.com header.s=Intel header.b=Bs+mNcQ+; spf=pass (imf10.hostedemail.com: domain of oliver.sang@intel.com designates 192.198.163.11 as permitted sender) smtp.mailfrom=oliver.sang@intel.com; dmarc=pass (policy=none) header.from=intel.com; arc=reject ("signature check failed: fail, {[1] = sig:microsoft.com:reject}") ARC-Seal: i=2; s=arc-20220608; d=hostedemail.com; t=1738828961; a=rsa-sha256; cv=fail; b=4duFZqb69qqgOVMX6GEXCPi1BfmoMM9ZiP8MdGV8OixWw1bkO20cBA9uJfQLWxaxvH3ubw jJEEyNdk1nlwkoD8aA7TMYlrsQA2SS7FbbMP8cbgMoufqtQvJlCl66PVVZxYP5oc0jaJke fDbh2uLoPPRMQKIS3/oDXl+UeVxGsvM= DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/simple; d=intel.com; i=@intel.com; q=dns/txt; s=Intel; t=1738828961; x=1770364961; h=date:from:to:cc:subject:message-id:references: content-transfer-encoding:in-reply-to:mime-version; bh=dTLWayqfpQSaJH8SIXWvfEyYuBpFtbDyUpdRfuz1s0w=; b=Bs+mNcQ+8EDbDSMBmFf3ERJQWNk/vX6uZZKFih080rLMX6UrcNYPVSr+ KHREaUSe7S0DkxM9i5Ac23SOpZLkK4q5QgyffA/4Q/VvkEZG+keLL7H3e iFYhqo91guUL2Ax6AutXdeMbLk9N/Oj0KikP9lsluy5rkhqCPhViD+iY6 aV6lC4fGKm1Ysr1CNZOPtfsc7UuiPSlYMgUlFeLyxigQYZU557LTh1r8J HK3S/OPXWOQylXE4sUStTZRlZpAQnG8xm42w+pY+YEu1ZrCF+yNEcLrLq +1bPDiEp3x6hY9UiHdcJZKxIT47gZR7KasTzGO2I56WB8AvHSWCMwONzS A==; X-CSE-ConnectionGUID: lGR1b4xITPmuZyzJHCqPxg== X-CSE-MsgGUID: XNtkW5hvQPiR4cXPg1UWCQ== X-IronPort-AV: E=McAfee;i="6700,10204,11336"; a="50039994" X-IronPort-AV: E=Sophos;i="6.13,263,1732608000"; d="scan'208";a="50039994" Received: from fmviesa010.fm.intel.com ([10.60.135.150]) by fmvoesa105.fm.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 06 Feb 2025 00:02:38 -0800 X-CSE-ConnectionGUID: 0sNmUDrLQz6QgKNV8Qc3gQ== X-CSE-MsgGUID: TIg9uXlpRre5YPnlLEqQKg== X-ExtLoop1: 1 X-IronPort-AV: E=Sophos;i="6.13,263,1732608000"; d="scan'208";a="111672513" Received: from orsmsx603.amr.corp.intel.com ([10.22.229.16]) by fmviesa010.fm.intel.com with ESMTP/TLS/AES256-GCM-SHA384; 06 Feb 2025 00:02:35 -0800 Received: from orsmsx601.amr.corp.intel.com (10.22.229.14) by ORSMSX603.amr.corp.intel.com (10.22.229.16) with Microsoft SMTP Server (version=TLS1_2, cipher=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256) id 15.1.2507.44; Thu, 6 Feb 2025 00:02:33 -0800 Received: from ORSEDG601.ED.cps.intel.com (10.7.248.6) by orsmsx601.amr.corp.intel.com (10.22.229.14) with Microsoft SMTP Server (version=TLS1_2, cipher=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256) id 15.1.2507.44 via Frontend Transport; Thu, 6 Feb 2025 00:02:33 -0800 Received: from NAM10-DM6-obe.outbound.protection.outlook.com (104.47.58.44) by edgegateway.intel.com (134.134.137.102) with Microsoft SMTP Server (version=TLS1_2, cipher=TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384) id 15.1.2507.44; Thu, 6 Feb 2025 00:02:30 -0800 ARC-Seal: i=1; a=rsa-sha256; s=arcselector10001; d=microsoft.com; cv=none; b=nEWgayuD3Hw8CaIaA17eJ1FMLC7x6MHJVfZ/aSGgQctjz0e7EiyHdTbAK5uBfBHJxLmhVgbewdq2kbg3gkxCIkgqBUm06FQ5KTRT6vORzMtp+KaNt12aQJRi9ociTJ/qI/3G7JRJ8zBa49j/+1UMvoKPQXH/FN116KMIBTY96V/P1JDWbFEdFgUkRyUSHhZXgPEuwnCgeR9HBbsWHTuya8XSYQHxLV0owmeCcMZrWuucPjQgA2tKcUYkZJfk1M6L6eDsEKzV1k9NETdpAIMbSilOkoN5+BjJsiQokwspTyWq2fH1Je8pjgKcDnYFgX5AtvHXx1VnaoT7vIUCjuyZXA== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=microsoft.com; s=arcselector10001; h=From:Date:Subject:Message-ID:Content-Type:MIME-Version:X-MS-Exchange-AntiSpam-MessageData-ChunkCount:X-MS-Exchange-AntiSpam-MessageData-0:X-MS-Exchange-AntiSpam-MessageData-1; bh=jrBFyhlxDbqoCfJ8thmz8x2SziAHAY97NBWKIGykUgg=; b=vb6swlF0f0GzCY2P6GHLBMpFYyGecf3BZI/N4XS/QHM9tgx9KKDRcLaq3MGWKx2vTOo2EaxFUoPSS+nS491EhhYpiGvyhyyvf7cunVwX80Xgqwy5feDHL50MJocY2Jc+iug2+8RoXSMbXVqX86qu0zBKB4UJcOX5sWeOxQx8u7KqIrArWcwFwsc7lpN5nmyB+HiaaiWqY+5qb30Wc4HAfkR2al8hFzvSJHcBLUTQ/lR0SMsVnHdWuMhHqu2KnFXcmpcTJ0/HRa74pD2gJkwiAYfPZdtgenXhAn0Bl9BxgefUj7OUE3c579wnlDMtL/lhikKuLxq7jn96BhMw5dnuuw== ARC-Authentication-Results: i=1; mx.microsoft.com 1; spf=pass smtp.mailfrom=intel.com; dmarc=pass action=none header.from=intel.com; dkim=pass header.d=intel.com; arc=none Received: from LV3PR11MB8603.namprd11.prod.outlook.com (2603:10b6:408:1b6::9) by PH0PR11MB4837.namprd11.prod.outlook.com (2603:10b6:510:41::7) with Microsoft SMTP Server (version=TLS1_2, cipher=TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384) id 15.20.8422.11; Thu, 6 Feb 2025 08:02:27 +0000 Received: from LV3PR11MB8603.namprd11.prod.outlook.com ([fe80::4622:29cf:32b:7e5c]) by LV3PR11MB8603.namprd11.prod.outlook.com ([fe80::4622:29cf:32b:7e5c%3]) with mapi id 15.20.8398.025; Thu, 6 Feb 2025 08:02:27 +0000 Date: Thu, 6 Feb 2025 16:02:15 +0800 From: Oliver Sang To: Yang Shi CC: , , , , , , , , , , , , , Subject: Re: [PATCH] /dev/zero: make private mapping full anonymous mapping Message-ID: References: <202501281038.617c6b60-lkp@intel.com> <01b5c112-480f-40dd-892f-5eb7c53e2f68@os.amperecomputing.com> Content-Type: text/plain; charset="iso-8859-1" Content-Disposition: inline Content-Transfer-Encoding: 8bit In-Reply-To: <01b5c112-480f-40dd-892f-5eb7c53e2f68@os.amperecomputing.com> X-ClientProxiedBy: SI1PR02CA0046.apcprd02.prod.outlook.com (2603:1096:4:1f5::14) To LV3PR11MB8603.namprd11.prod.outlook.com (2603:10b6:408:1b6::9) MIME-Version: 1.0 X-MS-PublicTrafficType: Email X-MS-TrafficTypeDiagnostic: LV3PR11MB8603:EE_|PH0PR11MB4837:EE_ X-MS-Office365-Filtering-Correlation-Id: 7a573c58-652f-4d4c-6904-08dd46849899 X-MS-Exchange-SenderADCheck: 1 X-MS-Exchange-AntiSpam-Relay: 0 X-Microsoft-Antispam: BCL:0;ARA:13230040|1800799024|366016|7416014|376014|7053199007; X-Microsoft-Antispam-Message-Info: =?iso-8859-1?Q?iHPLAcwBKkgF2OaZPxK+3vmymX9ImiZMxCJb3nTaZLNmaaSQCUMFJHQPn6?= =?iso-8859-1?Q?2yX23bRaPUqO1d+pMG8tUsFqoa/af28MErh2WUK11FPFlYOhuVP15Lh2L+?= =?iso-8859-1?Q?P0yD4L62e+0ns2TzRi7HRl/u33sugjfPG5u/ijhPxXe5A5EJb81bGF5jxW?= =?iso-8859-1?Q?zpRoJ3nq6Ymc2RIo4KRu/nGzWiStGOk4Yl/qVfzrolzri81p6oGAOa+12H?= =?iso-8859-1?Q?8Xym2J0xPxD6C1VOnZhKdk5pE9TKzZwCWjTiV55+ew2JsTZaYQZxQx7lO7?= =?iso-8859-1?Q?bbgx+wp+iG2ASycSOOSHF1D5BaBKXAhP3DfT4SoDI4BfV+pIjCTJ7OWgNi?= =?iso-8859-1?Q?kbIGtVhqvRlwRWUPDMaC4xRz2bxPnCLATSuiXzHL5yYw2HY6ieUIGm6X0V?= =?iso-8859-1?Q?kQ7eMUpksT9xdDCIho+mlZCTyXcJygzAhgKYidydFq/ArbbK9l5plHFVme?= =?iso-8859-1?Q?Ot4oV1MCgeHartB+Uj1quzwXg7SRExZ2pq/uYv823HxYFGMlJURtxDVpfw?= =?iso-8859-1?Q?5EsgI7Pwjuz/YgyjaZtc8DXHhtEKOzIlZbKdgK5xIXwo1B1Z5/Nt4/LSs4?= =?iso-8859-1?Q?fOqAH7k1hXfV72f1bkq0GW5tkjzUG2z93QqNWfkGtGnGi7TYSHrqhbLkIA?= =?iso-8859-1?Q?WQ13MrpkLUKn2rvHzeCzOmRE5YRaFxzcAVc+sJ6bt8zFRsnP2PogozWLj8?= =?iso-8859-1?Q?xUZx+UradRAHUOTZoQOorBqkjQQYdFihz8jOnj7Gnv5nFs+EkBDFUPT86J?= =?iso-8859-1?Q?6p0pi+99EOsxLlAHQO5sK+fu7E2tM6YDQkoAj1jKzn7KAuARn2PvJOOvxk?= =?iso-8859-1?Q?sd/2ycEtG9RcLBIZrJeUNoHE3pUcGd5J2ipLWftZGt4LoXeWBF5NgGz5Un?= =?iso-8859-1?Q?Por34POeSGXXxtSJ7QPkNq3MDA7a0Au7IGkutHftuk/gzmasAhI7kTqwOI?= =?iso-8859-1?Q?cReqsRtl+/bz86AWXNK/7XxNdYCL2vpUBviIP1PjcXMN31EiQ2pNHbhRjA?= =?iso-8859-1?Q?3LV7dFpBHm90QkayNsfEfZjaqmU9eZcm+sbv4EqVn2hZHYW+OHUW7o3j23?= =?iso-8859-1?Q?EbXwUE86QqF13ZmPlUFkTuFs/X6qwDH+zFbd5hFhXJ3mbAVmuzqairedw/?= =?iso-8859-1?Q?ikE3jSLwcM+Xk2ZM3XyTEga4c/h4Y9o7mMF2swDwQQgdBef7h7JKjCFP7Z?= =?iso-8859-1?Q?YEHx0LOIZJfA11C9eFR6CAn3dXfcH+echHWN7VRCKAh9hbk/2FrwJAnSVg?= =?iso-8859-1?Q?/xkXDS3D9ViNOqH/yXxCAQC2keB97wsLuTrO3jpwBZ75+xo94NxwxsS83N?= =?iso-8859-1?Q?42SxCXN/VtF6nsBrRJfmRiaKJDFOLzNMnR0EkDeo1tpIRxjdY/e3t/1rRC?= =?iso-8859-1?Q?OKc9QX9lGdKxnMKIsH42aHhmwZxjh96w=3D=3D?= X-Forefront-Antispam-Report: CIP:255.255.255.255;CTRY:;LANG:en;SCL:1;SRV:;IPV:NLI;SFV:NSPM;H:LV3PR11MB8603.namprd11.prod.outlook.com;PTR:;CAT:NONE;SFS:(13230040)(1800799024)(366016)(7416014)(376014)(7053199007);DIR:OUT;SFP:1101; X-MS-Exchange-AntiSpam-MessageData-ChunkCount: 1 X-MS-Exchange-AntiSpam-MessageData-0: =?iso-8859-1?Q?pop5IuOIpcXt7KtTQxCV5KHxEm3ETqNfx7Rz2chEFwSSe2sCg3HTTxZ/rN?= =?iso-8859-1?Q?jZWBwADXP6FBzZWUAX2rsxB7e7AlPXCjyBuST9HgZsRlJeee+1x13X+KSo?= =?iso-8859-1?Q?KntDN489YdGTYelvkNV7H23Ac2Fjzf6Fzb7KyKZaaXlCbC72pB4uS5dOr1?= =?iso-8859-1?Q?zzB6K21xZlW2VHRiZypE/AMTsm2xGM26qJYAjaEJxbvf7f+LBvlbi0KOZC?= =?iso-8859-1?Q?tHMWx50SpqPUWHm99EL391sdC2pCLF7LVVD648lz8BYiP3UqyGQCdmYeZL?= =?iso-8859-1?Q?G9nbFzfPEqbB9Rle1OoQp/NcN75jbqt7OUOcXKoui0THcdA7hO9uVweG1+?= =?iso-8859-1?Q?LGmiH8RrlvfrS7e4P83F99dvdyGRciEPx9B9sz1YloPJeDVok8krIZvlht?= =?iso-8859-1?Q?1Gsi9O6eL1SUHRkg22Q6CHWccKH3Acn3S9JEKcQbWYNbis1QSTbsqtqBuu?= =?iso-8859-1?Q?PONV3ViPA9fr3wnmxch1ejsrPtl12K7zMn2LG39Hyvd0oySeHCMT8tCa2z?= =?iso-8859-1?Q?EaQVhIbPxjjVej36sjiD6Akj2ON2hKATZNdW2l/JZiIkOYIAGR+Q6WkL0m?= =?iso-8859-1?Q?YxezJp37R7/35B0I7bR8hyscqDzRlaWxB0QsUTvqBHzyWqmoCQEu+D9NRN?= =?iso-8859-1?Q?rQBVltJjWlTWUPh7zS//XdYo0O/CHYAuEKmeNCxHQAQkBtddx25hScSPbI?= =?iso-8859-1?Q?jkOKh9lmBsGcozy18ezKBcNtaa+pzsxU+enFoIzhJZt2JlUGBhkW0cFDMx?= =?iso-8859-1?Q?/LcSuvQVx0d3J5R5vfMQM2id1gBM2XOl19SFDfJiQ7SAIE8VZ8JZOjPZIS?= =?iso-8859-1?Q?vmREQYXL4xUg9hT58fDkkSNQknUC3GKQE4/jo6845mx74Fis/KhgfxNyYD?= =?iso-8859-1?Q?AlvR9W87GP2lMxVjZYuJYgVfDKNmnP284MLw5jlmwMFvXl7D8quaLUIZq2?= =?iso-8859-1?Q?uJXGWKmmpL6lr8pS7d923AznBV6iCnf4FACnhDxZQv3nUlvxVh3W2cJgk9?= =?iso-8859-1?Q?+YbwqQDSrIHkE7TL3ZhIhBDQ1OTPMCSQtjQWORzNt0Za6oH49nTy7E20Jf?= =?iso-8859-1?Q?1aOpxhQoUKGZr2dSHhfdZ1FCUwCox+NWJDE3yGv9VtdRmcPc+1dOEO5oXV?= =?iso-8859-1?Q?wv5hEpGM3ARax7IoWv0Mhg4wiGADm5eudTYXATJWdbfreYycVdRXDN4hoC?= =?iso-8859-1?Q?QgNgEOAB8Qy4bgWSMJPKPRrd4yuGfBwvsnhf5hWtCLm6xwSWeQIxbEZTBi?= =?iso-8859-1?Q?DG77t9nR1aMiXVJTpn9A4k8rQalCSYDvQG8pj6Ep6UbCYE9wlUz7p+aMaI?= =?iso-8859-1?Q?jxO0dQBLj5KZg5lqmO6byFQnC5qRpRFOhUe0BE8wWgj9stBGrar/RxnXaZ?= =?iso-8859-1?Q?7Lpb0gU0xhbroj6r6H+SFEo5f+k0fKThdHPhKnGZhWT2j/WyOuUzswM1u5?= =?iso-8859-1?Q?xEOI7OZuwrhuAlSaYFxh8cZ7fqlcvtfqNf7mLH+7VYtYSXV50BR1QeSQ/z?= =?iso-8859-1?Q?bJqhjTMXd1/S6axtnw4KHIw+fZv41F6/3Na1TVAxY0AnUtqdedLpfSHLaB?= =?iso-8859-1?Q?vxUFoiJXUJik4VDrc4GUjl7aas0raSUeN5Sk/qOMTmyg9eN89gzCuQBqfa?= =?iso-8859-1?Q?dsCR/pkXq1zKImph6Vq2N7zlVfGhvZEosgzll42Ua10j8Z9fpL0xxa7w?= =?iso-8859-1?Q?=3D=3D?= X-MS-Exchange-CrossTenant-Network-Message-Id: 7a573c58-652f-4d4c-6904-08dd46849899 X-MS-Exchange-CrossTenant-AuthSource: LV3PR11MB8603.namprd11.prod.outlook.com X-MS-Exchange-CrossTenant-AuthAs: Internal X-MS-Exchange-CrossTenant-OriginalArrivalTime: 06 Feb 2025 08:02:26.9898 (UTC) X-MS-Exchange-CrossTenant-FromEntityHeader: Hosted X-MS-Exchange-CrossTenant-Id: 46c98d88-e344-4ed4-8496-4ed7712e255d X-MS-Exchange-CrossTenant-MailboxType: HOSTED X-MS-Exchange-CrossTenant-UserPrincipalName: bJn+vgq2x0dMWapGRxlZXKsKeH4sxWjyoN1vPJ9Ad8Xg+mtVKjBWE7kUAMABnrySBbMIrROJ0b1pilktuZqdDA== X-MS-Exchange-Transport-CrossTenantHeadersStamped: PH0PR11MB4837 X-OriginatorOrg: intel.com X-Rspamd-Queue-Id: 07A43C000E X-Stat-Signature: 3u9yf8kpsfkgqddcdih5jm8j9hidh1ji X-Rspamd-Server: rspam08 X-Rspam-User: X-HE-Tag: 1738828960-485301 X-HE-Meta: U2FsdGVkX19GPkEg7A91QFSlJDNVzQ02aOiwSv/E2xc3KH0NP6QxEUXD61jD7pHjzAXSJK1T7Sy4Jz1HaFzMob0hCEzvy2gXtgo/B7o8ogiR5NQLzUVEWTyJP+/c3vXeQXTKuodoLK+AokqeJindAF7WJMjoxsydu6QAudhh069Jd+0fFyV4hDjRBMRoHkJ7S+//jJvEMtg8IQ3RN6NX6MGl0I0QSOinel5y6J1H93f5gLfgTyPs6aZNa5TIIiXoout6FvTBtDGi9b2EZO4j3Ciw7qYV9yHzY+8goGZzuHhfjmGRELtpzMWGYoonfmnP89TlrKbVEfxQZU3fpQDCEjj4SQWNXg0n6q/wSya4hASFZRSpot4Nh8SW1mfXR62UWUlj2jRSoikBk/bZPtH9N/PjKTvBV9v/UnQz6iLoFdNPNAi1Udra0/ATQHlCJZOdyJkEJ6RRjp4EKhNlsA6rG5Tlqh+Wb+9/EEG6Og+15HpfnIa7FfxHpo5CZHqVfpCjN0bkJNp0zRwBycrW38DYCOAtFbmA9ZmGzYHROS/BHp9tfJDFN64SL8NPocDfeeFxVSA/lIhIffXx3JpNSd0BWuiTxbTSEiD81DtMflz/gPfQVOn4UmfbJVp22u3lCii/Lm1E+nQlR/NOudOtXvxO1I+GM80kY7JCT/CQ3iGPQbN71nWuNGLPS+jed4x7S0T4WVYExlusXS/9DdFiVLrTvtoe930qyKs2gsQiCYo+L+OOrrYjY61gFaPGNq0gkMCDzPeMF4rgi7405Hu6FYJrcMq+tG5J1hB4fAbQr/BJj2BJMYTHaTupDZgA1LiZv27PBi3XPaAq4qt/ec/Lt8ETtJfGCDGK3++3QqaHFbOWXQmvy7S8UCFcO4W4+hhg9/NF5fOUWCtx5Qn6HduA5a1VscRG1Kk5eHdeJd0nxs7651V3ir/9W6WFWGhOtAA+Ya1AXbWOkIazQt8+isZScGu uPQlBzrt w2Nj/FTXPXOA7II/FNTyRgA3qlnXpS+4pH9FX8Slm6MavuMGS2dde1eCc60LDbhOFstIW3oaLOSmNYkfKvlvAQWXoqqKu0b7cJLHauuf6+ghD68Z8H5zZflS4ucwMGiaOaXuCUy1TeY6vnlFyH0eqDqcLCT4QaD+tMF7UqRjC1tLYqSOvk1yc6TaL9qWuKN4HysYn20vQMVsIK0AhtAvwJmz4oLUqExvJcJU8P0epcz7ooZL+JU89pFDjCIefUUANEGtVnowK5wMOS1us1OvGrK+mTJoCyDABHUJph/Yt6lmOsVqUJOZrtmIsuZK9EWitatpVfCAXstrRqMZLgwErGQEeWu4oo/+F4N19++clSHwvaGXMqS0nPKUb9xx9KygZvY/WlTaauqP4ofaw1FaxtaC6X2uHKjaFp67hRiabNhoOaVDSSFcyyCrPWKV3wUSU+C+VH+bz2uL/V5OxwyARFmnFuiEQF48qHI8i327vS2lfLZpkvNhSmTvBQjjdQ7dNgpnA16FMPOuYrTyP6yID/tTA6Un6G7AlcvI0f9y7OhLg5f7aXMbcD68TJe38ACZxo3FSGECrfxQHWTyv66jAFFsdnsRZZP2Tv3hp X-Bogosity: Ham, tests=bogofilter, spamicity=0.000000, version=1.2.4 Sender: owner-linux-mm@kvack.org Precedence: bulk X-Loop: owner-majordomo@kvack.org List-ID: List-Subscribe: List-Unsubscribe: hi, Yang Shi, On Fri, Jan 31, 2025 at 10:38:03AM -0800, Yang Shi wrote: > > > > On 1/27/25 7:14 PM, kernel test robot wrote: > > hi, All, > > > > we don't have enough knowledge to understand fully the discussion for this > > patch, we saw "NACK" but there were more discussions later. > > so below report is just FYI what we observed in our tests. thanks > > Thanks for the report. It was nack'ed because of the change to smaps/maps > files in proc. > > > > > Hello, > > > > kernel test robot noticed a 858.5% improvement of vm-scalability.throughput on: > > > > > > commit: 7143ee2391f1ea15e6791e129870473543634de2 ("[PATCH] /dev/zero: make private mapping full anonymous mapping") > > url: https://github.com/intel-lab-lkp/linux/commits/Yang-Shi/dev-zero-make-private-mapping-full-anonymous-mapping/20250114-063339 > > base: https://git.kernel.org/cgit/linux/kernel/git/gregkh/char-misc.git a68d3cbfade64392507302f3a920113b60dc811f > > patch link: https://lore.kernel.org/all/20250113223033.4054534-1-yang@os.amperecomputing.com/ > > patch subject: [PATCH] /dev/zero: make private mapping full anonymous mapping > > > > testcase: vm-scalability > > config: x86_64-rhel-9.4 > > compiler: gcc-12 > > test machine: 224 threads 4 sockets Intel(R) Xeon(R) Platinum 8380H CPU @ 2.90GHz (Cooper Lake) with 192G memory > > parameters: > > > > runtime: 300s > > test: small-allocs > > It seems this benchmark allocates huge amount of small areas (each area is > as big as 40K) by mmap'ing /dev/zero. > > This patch makes /dev/zero mapping a full anonymous mapping, so the later > vma_link_file() is actually skipped, which needs acquire file rmap lock then > insert the mapping into file rmap tree. The below profiling also showed > this. > Quoted here so that we don't have to scroll down: > > > 95.60 -95.2 0.42 ±113% perf-profile.children.cycles-pp.__mmap > > 94.14 -93.6 0.54 ±106% perf-profile.children.cycles-pp.__mmap_new_vma > > 93.79 -93.6 0.21 ±171% perf-profile.children.cycles-pp.vma_link_file > > 93.40 -93.4 0.00 perf-profile.children.cycles-pp.rwsem_down_write_slowpath > > 93.33 -93.3 0.00 perf-profile.children.cycles-pp.rwsem_optimistic_spin > > 93.44 -93.2 0.24 ±178% perf-profile.children.cycles-pp.down_write > > 94.55 -93.1 1.40 ± 51% perf-profile.children.cycles-pp.ksys_mmap_pgoff > > 94.25 -93.0 1.30 ± 59% perf-profile.children.cycles-pp.__mmap_region > > 92.91 -92.9 0.00 perf-profile.children.cycles-pp.osq_lock > > 94.45 -92.7 1.72 ± 34% perf-profile.children.cycles-pp.do_mmap > > 94.46 -92.6 1.83 ± 31% perf-profile.children.cycles-pp.vm_mmap_pgoff > > It significantly speed up mmap for this benchmark and the rmap lock > contention is reduced significantly for both multi-processes and > multi-threads. > > The benchmark itself may exaggerate the improvement, but it may really speed > up some real life workloads. For example, multiple applications which may > allocate anonymous mapping by mmap'ing /dev/zero, then they may have > contention on /dev/zero's rmap lock. > > It doesn't make too much sense to link /dev/zero anonymous vmas to the file > rmap tree. So the below patch should be able to speed up the benchmark too. sorry for late and thanks a lot for information! > > Oliver, can you please give this patch a try? it seems this is an alternative patch? since we applied your "/dev/zero: make private mapping full anonymous mapping" patch upon a68d3cbfad like below: * 7143ee2391f1e /dev/zero: make private mapping full anonymous mapping * a68d3cbfade64 memstick: core: fix kernel-doc notation so I applied below patch also upon a68d3cbfad. we saw big improvement but not that big. ========================================================================================= compiler/cpufreq_governor/kconfig/rootfs/runtime/tbox_group/test/testcase: gcc-12/performance/x86_64-rhel-9.4/debian-12-x86_64-20240206.cgz/300s/lkp-cpl-4sp2/small-allocs/vm-scalability commit: a68d3cbfad ("memstick: core: fix kernel-doc notation") 52ec85cb99 <--- your patch a68d3cbfade64392 52ec85cb99e9b31dc304eae965a ---------------- --------------------------- %stddev %change %stddev \ | \ 14364828 ± 4% +410.6% 73349239 ± 3% vm-scalability.throughput full comparison as below [1] just FYI. > > > diff --git a/mm/vma.c b/mm/vma.c > index bb2119e5a0d0..1092222c40ae 100644 > --- a/mm/vma.c > +++ b/mm/vma.c > @@ -1633,6 +1633,9 @@ static void unlink_file_vma_batch_process(struct > unlink_vma_file_batch *vb) >  void unlink_file_vma_batch_add(struct unlink_vma_file_batch *vb, >                                struct vm_area_struct *vma) >  { > +       if (vma_is_anonymous(vma)) > +               return; > + >         if (vma->vm_file == NULL) >                 return; > > @@ -1658,6 +1661,9 @@ void unlink_file_vma(struct vm_area_struct *vma) >  { >         struct file *file = vma->vm_file; > > +       if (vma_is_anonymous(vma)) > +               return; > + >         if (file) { >                 struct address_space *mapping = file->f_mapping; > > @@ -1672,6 +1678,9 @@ void vma_link_file(struct vm_area_struct *vma) >         struct file *file = vma->vm_file; >         struct address_space *mapping; > > +       if (vma_is_anonymous(vma)) > +               return; > + >         if (file) { >                 mapping = file->f_mapping; >                 i_mmap_lock_write(mapping); > > > Because /dev/zero's private mapping is an anonymous mapping with valid > vm_file, so we need to bail out early if the vma is anonymous even though it > has vm_file. IMHO, making /dev/zero private mapping a full anonymous mapping > looks more clean. > [1] ========================================================================================= compiler/cpufreq_governor/kconfig/rootfs/runtime/tbox_group/test/testcase: gcc-12/performance/x86_64-rhel-9.4/debian-12-x86_64-20240206.cgz/300s/lkp-cpl-4sp2/small-allocs/vm-scalability commit: a68d3cbfad ("memstick: core: fix kernel-doc notation") 52ec85cb99 <--- your patch a68d3cbfade64392 52ec85cb99e9b31dc304eae965a ---------------- --------------------------- %stddev %change %stddev \ | \ 5.262e+09 ± 3% -45.0% 2.896e+09 ± 6% cpuidle..time 7924008 ± 3% -79.3% 1643339 ± 11% cpuidle..usage 1871164 ± 4% -22.4% 1452554 ± 12% numa-numastat.node3.local_node 1952164 ± 3% -20.1% 1560294 ± 12% numa-numastat.node3.numa_hit 399.52 -68.2% 126.86 uptime.boot 14507 -15.7% 12232 uptime.idle 6.99 ± 3% +147.9% 17.34 ± 4% vmstat.cpu.id 1.71 +473.6% 9.79 ± 2% vmstat.cpu.us 34204 ± 5% -72.9% 9272 ± 7% vmstat.system.cs 266575 -21.2% 210191 vmstat.system.in 3408 ± 5% -99.8% 8.38 ± 48% perf-c2c.DRAM.local 18076 ± 3% -99.8% 32.25 ± 27% perf-c2c.DRAM.remote 8082 ± 5% -99.8% 15.50 ± 64% perf-c2c.HITM.local 6544 ± 6% -99.8% 13.62 ± 51% perf-c2c.HITM.remote 14627 ± 4% -99.8% 29.12 ± 53% perf-c2c.HITM.total 6.49 ± 3% +8.8 15.24 ± 5% mpstat.cpu.all.idle% 0.63 -0.3 0.32 ± 4% mpstat.cpu.all.irq% 0.03 ± 2% +0.2 0.26 ± 2% mpstat.cpu.all.soft% 91.17 -17.0 74.15 mpstat.cpu.all.sys% 1.68 ± 2% +8.3 10.03 ± 2% mpstat.cpu.all.usr% 337.33 -97.4% 8.88 ± 75% mpstat.max_utilization.seconds 352.76 -77.3% 79.95 ± 2% time.elapsed_time 352.76 -77.3% 79.95 ± 2% time.elapsed_time.max 225965 ± 7% -16.0% 189844 ± 6% time.involuntary_context_switches 9.592e+08 ± 4% +11.9% 1.074e+09 time.minor_page_faults 20852 -8.8% 19012 time.percent_of_cpu_this_job_got 72302 -81.4% 13425 ± 3% time.system_time 1260 ± 3% +41.0% 1777 time.user_time 5393707 ± 5% -98.4% 86880 ± 17% time.voluntary_context_switches 1609925 -50.3% 800493 meminfo.Active 1609925 -50.3% 800493 meminfo.Active(anon) 160837 ± 33% -63.9% 58119 ± 13% meminfo.AnonHugePages 4435665 -18.5% 3614714 meminfo.Cached 1775547 -43.8% 998415 meminfo.Committed_AS 148539 -43.7% 83699 ± 4% meminfo.Mapped 4245538 ± 4% -20.9% 3356561 meminfo.PageTables 14166291 ± 4% -9.6% 12806082 meminfo.SUnreclaim 929777 -88.2% 109274 ± 3% meminfo.Shmem 14315492 ± 4% -9.6% 12947821 meminfo.Slab 64129 ± 4% +418.9% 332751 ± 3% vm-scalability.median 45.40 ± 5% +1961.8 2007 ± 8% vm-scalability.stddev% 14364828 ± 4% +410.6% 73349239 ± 3% vm-scalability.throughput 352.76 -77.3% 79.95 ± 2% vm-scalability.time.elapsed_time 352.76 -77.3% 79.95 ± 2% vm-scalability.time.elapsed_time.max 225965 ± 7% -16.0% 189844 ± 6% vm-scalability.time.involuntary_context_switches 9.592e+08 ± 4% +11.9% 1.074e+09 vm-scalability.time.minor_page_faults 20852 -8.8% 19012 vm-scalability.time.percent_of_cpu_this_job_got 72302 -81.4% 13425 ± 3% vm-scalability.time.system_time 1260 ± 3% +41.0% 1777 vm-scalability.time.user_time 5393707 ± 5% -98.4% 86880 ± 17% vm-scalability.time.voluntary_context_switches 4.316e+09 ± 4% +11.9% 4.832e+09 vm-scalability.workload 265763 ± 4% -20.5% 211398 ± 4% numa-vmstat.node0.nr_page_table_pages 31364 ±106% -85.0% 4690 ±169% numa-vmstat.node0.nr_shmem 12205 ± 67% -74.1% 3161 ±199% numa-vmstat.node1.nr_mapped 265546 ± 4% -21.8% 207742 ± 4% numa-vmstat.node1.nr_page_table_pages 44052 ± 71% -86.0% 6163 ±161% numa-vmstat.node1.nr_shmem 885590 ± 4% -9.9% 797649 ± 4% numa-vmstat.node1.nr_slab_unreclaimable 264589 ± 4% -21.2% 208598 ± 4% numa-vmstat.node2.nr_page_table_pages 881598 ± 4% -10.0% 793829 ± 4% numa-vmstat.node2.nr_slab_unreclaimable 192683 ± 30% -61.0% 75078 ± 70% numa-vmstat.node3.nr_active_anon 286819 ±108% -93.0% 19993 ± 39% numa-vmstat.node3.nr_file_pages 13124 ± 49% -92.3% 1006 ± 57% numa-vmstat.node3.nr_mapped 264499 ± 4% -22.1% 206135 ± 2% numa-vmstat.node3.nr_page_table_pages 139810 ± 14% -90.5% 13229 ± 89% numa-vmstat.node3.nr_shmem 880199 ± 4% -11.8% 776210 ± 5% numa-vmstat.node3.nr_slab_unreclaimable 192683 ± 30% -61.0% 75077 ± 70% numa-vmstat.node3.nr_zone_active_anon 1951359 ± 3% -20.1% 1558936 ± 12% numa-vmstat.node3.numa_hit 1870359 ± 4% -22.4% 1451195 ± 12% numa-vmstat.node3.numa_local 402515 -50.3% 200150 proc-vmstat.nr_active_anon 170568 +1.9% 173746 proc-vmstat.nr_anon_pages 4257257 +0.9% 4296664 proc-vmstat.nr_dirty_background_threshold 8524925 +0.9% 8603835 proc-vmstat.nr_dirty_threshold 1109246 -18.5% 903959 proc-vmstat.nr_file_pages 42815276 +0.9% 43210344 proc-vmstat.nr_free_pages 37525 -43.6% 21164 ± 4% proc-vmstat.nr_mapped 1059932 ± 4% -21.1% 836810 proc-vmstat.nr_page_table_pages 232507 -88.2% 27341 ± 3% proc-vmstat.nr_shmem 37297 -5.0% 35436 proc-vmstat.nr_slab_reclaimable 3537843 ± 4% -9.8% 3192506 proc-vmstat.nr_slab_unreclaimable 402515 -50.3% 200150 proc-vmstat.nr_zone_active_anon 61931 ± 8% -83.8% 10023 ± 45% proc-vmstat.numa_hint_faults 15755 ± 21% -87.1% 2039 ± 97% proc-vmstat.numa_hint_faults_local 6916516 ± 3% -7.1% 6425430 proc-vmstat.numa_hit 6568542 ± 3% -7.5% 6077764 proc-vmstat.numa_local 293942 ± 3% -69.6% 89435 ± 49% proc-vmstat.numa_pte_updates 9.608e+08 ± 4% +11.8% 1.074e+09 proc-vmstat.pgfault 55981 ± 2% -63.1% 20641 ± 2% proc-vmstat.pgreuse 1063552 ± 4% -20.3% 847673 ± 4% numa-meminfo.node0.PageTables 3565610 ± 4% -8.0% 3279375 ± 3% numa-meminfo.node0.SUnreclaim 125455 ±106% -85.2% 18620 ±168% numa-meminfo.node0.Shmem 3592377 ± 4% -7.1% 3336072 ± 4% numa-meminfo.node0.Slab 48482 ± 67% -74.3% 12475 ±199% numa-meminfo.node1.Mapped 1062709 ± 4% -21.7% 831966 ± 4% numa-meminfo.node1.PageTables 3543793 ± 4% -10.0% 3189589 ± 4% numa-meminfo.node1.SUnreclaim 176171 ± 71% -86.0% 24677 ±161% numa-meminfo.node1.Shmem 3593431 ± 4% -10.4% 3220352 ± 4% numa-meminfo.node1.Slab 1058901 ± 4% -21.3% 833124 ± 4% numa-meminfo.node2.PageTables 3527862 ± 4% -10.2% 3168666 ± 5% numa-meminfo.node2.SUnreclaim 3565750 ± 4% -10.3% 3200248 ± 5% numa-meminfo.node2.Slab 770405 ± 30% -61.0% 300435 ± 70% numa-meminfo.node3.Active 770405 ± 30% -61.0% 300435 ± 70% numa-meminfo.node3.Active(anon) 1146977 ±108% -93.0% 80110 ± 40% numa-meminfo.node3.FilePages 52663 ± 47% -91.6% 4397 ± 56% numa-meminfo.node3.Mapped 6368902 ± 20% -21.2% 5021246 ± 2% numa-meminfo.node3.MemUsed 1058539 ± 4% -22.2% 823061 ± 3% numa-meminfo.node3.PageTables 3522496 ± 4% -12.1% 3096728 ± 6% numa-meminfo.node3.SUnreclaim 558943 ± 14% -90.5% 53054 ± 89% numa-meminfo.node3.Shmem 3557392 ± 4% -12.3% 3119454 ± 6% numa-meminfo.node3.Slab 0.82 ± 4% -39.7% 0.50 ± 12% perf-stat.i.MPKI 2.714e+10 ± 2% +185.7% 7.755e+10 ± 6% perf-stat.i.branch-instructions 0.11 ± 3% +0.1 0.20 ± 5% perf-stat.i.branch-miss-rate% 24932893 +156.6% 63980942 ± 5% perf-stat.i.branch-misses 64.93 -10.1 54.87 ± 2% perf-stat.i.cache-miss-rate% 34508 ± 4% -61.4% 13315 ± 10% perf-stat.i.context-switches 7.67 -63.7% 2.79 ± 6% perf-stat.i.cpi 224605 +10.8% 248972 ± 4% perf-stat.i.cpu-clock 696.35 ± 2% -57.4% 296.79 ± 3% perf-stat.i.cpu-migrations 1.102e+11 +128.5% 2.518e+11 ± 6% perf-stat.i.instructions 0.14 +198.2% 0.42 ± 5% perf-stat.i.ipc 24.25 ± 3% +375.8% 115.36 ± 3% perf-stat.i.metric.K/sec 2722043 ± 3% +439.7% 14690226 ± 6% perf-stat.i.minor-faults 2722043 ± 3% +439.7% 14690226 ± 6% perf-stat.i.page-faults 224605 +10.8% 248972 ± 4% perf-stat.i.task-clock 0.81 ± 3% -52.5% 0.39 ± 14% perf-stat.overall.MPKI 0.09 -0.0 0.08 ± 2% perf-stat.overall.branch-miss-rate% 64.81 -6.4 58.40 perf-stat.overall.cache-miss-rate% 7.24 -56.3% 3.17 ± 3% perf-stat.overall.cpi 0.14 +129.0% 0.32 ± 3% perf-stat.overall.ipc 9012 ± 2% -57.5% 3827 perf-stat.overall.path-length 2.701e+10 ± 2% +159.6% 7.012e+10 ± 2% perf-stat.ps.branch-instructions 24708939 +119.2% 54173035 perf-stat.ps.branch-misses 34266 ± 5% -73.9% 8949 ± 7% perf-stat.ps.context-switches 7.941e+11 -9.1% 7.219e+11 perf-stat.ps.cpu-cycles 693.54 ± 2% -68.6% 217.73 ± 5% perf-stat.ps.cpu-migrations 1.097e+11 +108.1% 2.282e+11 ± 2% perf-stat.ps.instructions 2710577 ± 3% +388.7% 13246535 ± 2% perf-stat.ps.minor-faults 2710577 ± 3% +388.7% 13246536 ± 2% perf-stat.ps.page-faults 3.886e+13 ± 2% -52.4% 1.849e+13 perf-stat.total.instructions 64052898 ± 5% -96.2% 2460331 ±166% sched_debug.cfs_rq:/.avg_vruntime.avg 95701822 ± 7% -85.1% 14268127 ±116% sched_debug.cfs_rq:/.avg_vruntime.max 43098762 ± 6% -96.0% 1715136 ±173% sched_debug.cfs_rq:/.avg_vruntime.min 9223270 ± 9% -84.2% 1457904 ±122% sched_debug.cfs_rq:/.avg_vruntime.stddev 0.78 ± 2% -77.0% 0.18 ±130% sched_debug.cfs_rq:/.h_nr_running.avg 43049468 ± 22% -89.3% 4590302 ±180% sched_debug.cfs_rq:/.left_deadline.max 3836405 ± 37% -85.6% 550773 ±176% sched_debug.cfs_rq:/.left_deadline.stddev 43049467 ± 22% -89.3% 4590279 ±180% sched_debug.cfs_rq:/.left_vruntime.max 3836405 ± 37% -85.6% 550772 ±176% sched_debug.cfs_rq:/.left_vruntime.stddev 64052901 ± 5% -96.2% 2460341 ±166% sched_debug.cfs_rq:/.min_vruntime.avg 95701822 ± 7% -85.1% 14268127 ±116% sched_debug.cfs_rq:/.min_vruntime.max 43098762 ± 6% -96.0% 1715136 ±173% sched_debug.cfs_rq:/.min_vruntime.min 9223270 ± 9% -84.2% 1457902 ±122% sched_debug.cfs_rq:/.min_vruntime.stddev 0.77 ± 2% -77.4% 0.17 ±128% sched_debug.cfs_rq:/.nr_running.avg 1.61 ± 24% +396.0% 7.96 ± 62% sched_debug.cfs_rq:/.removed.runnable_avg.avg 86.69 +424.4% 454.62 ± 24% sched_debug.cfs_rq:/.removed.runnable_avg.max 11.14 ± 13% +409.8% 56.79 ± 35% sched_debug.cfs_rq:/.removed.runnable_avg.stddev 1.61 ± 24% +396.0% 7.96 ± 62% sched_debug.cfs_rq:/.removed.util_avg.avg 86.69 +424.4% 454.62 ± 24% sched_debug.cfs_rq:/.removed.util_avg.max 11.14 ± 13% +409.8% 56.79 ± 35% sched_debug.cfs_rq:/.removed.util_avg.stddev 43049467 ± 22% -89.3% 4590282 ±180% sched_debug.cfs_rq:/.right_vruntime.max 3836405 ± 37% -85.6% 550772 ±176% sched_debug.cfs_rq:/.right_vruntime.stddev 286633 ± 43% +262.3% 1038592 ± 36% sched_debug.cfs_rq:/.runnable_avg.avg 34728895 ± 30% +349.2% 1.56e+08 ± 26% sched_debug.cfs_rq:/.runnable_avg.max 2845573 ± 30% +325.9% 12119045 ± 26% sched_debug.cfs_rq:/.runnable_avg.stddev 769.03 -69.9% 231.86 ± 84% sched_debug.cfs_rq:/.util_avg.avg 1621 ± 5% -31.5% 1111 ± 8% sched_debug.cfs_rq:/.util_avg.max 724.17 ± 2% -89.6% 75.66 ±147% sched_debug.cfs_rq:/.util_est.avg 1360 ± 15% -39.2% 826.88 ± 37% sched_debug.cfs_rq:/.util_est.max 766944 ± 3% +18.1% 905901 sched_debug.cpu.avg_idle.avg 321459 ± 2% -35.6% 207172 ± 10% sched_debug.cpu.avg_idle.stddev 195573 -72.7% 53401 ± 24% sched_debug.cpu.clock.avg 195596 -72.7% 53442 ± 24% sched_debug.cpu.clock.max 195548 -72.7% 53352 ± 24% sched_debug.cpu.clock.min 194424 -72.6% 53229 ± 24% sched_debug.cpu.clock_task.avg 194608 -72.6% 53383 ± 24% sched_debug.cpu.clock_task.max 181834 -77.5% 40964 ± 31% sched_debug.cpu.clock_task.min 4241 ± 2% -80.6% 821.65 ±142% sched_debug.cpu.curr->pid.avg 9799 ± 2% -55.4% 4365 ± 17% sched_debug.cpu.curr->pid.max 1365 ± 10% -48.0% 709.44 ± 5% sched_debug.cpu.curr->pid.stddev 537665 ± 4% +31.2% 705318 ± 14% sched_debug.cpu.max_idle_balance_cost.max 3119 ± 56% +579.1% 21184 ± 39% sched_debug.cpu.max_idle_balance_cost.stddev 0.78 ± 2% -76.3% 0.18 ±135% sched_debug.cpu.nr_running.avg 25773 ± 5% -96.1% 1007 ± 41% sched_debug.cpu.nr_switches.avg 48669 ± 10% -76.5% 11448 ± 13% sched_debug.cpu.nr_switches.max 19006 ± 7% -98.6% 258.81 ± 64% sched_debug.cpu.nr_switches.min 4142 ± 8% -66.3% 1396 ± 17% sched_debug.cpu.nr_switches.stddev 0.07 ± 23% -92.9% 0.01 ± 41% sched_debug.cpu.nr_uninterruptible.avg 240.19 ± 16% -82.1% 42.94 ± 41% sched_debug.cpu.nr_uninterruptible.max -77.92 -88.1% -9.25 sched_debug.cpu.nr_uninterruptible.min 37.87 ± 5% -85.8% 5.36 ± 13% sched_debug.cpu.nr_uninterruptible.stddev 195549 -72.7% 53356 ± 24% sched_debug.cpu_clk 194699 -73.0% 52506 ± 25% sched_debug.ktime 0.00 -100.0% 0.00 sched_debug.rt_rq:.rt_nr_running.avg 0.17 -100.0% 0.00 sched_debug.rt_rq:.rt_nr_running.max 0.01 -100.0% 0.00 sched_debug.rt_rq:.rt_nr_running.stddev 196368 -72.4% 54191 ± 24% sched_debug.sched_clk 0.17 ±142% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.__do_fault.do_read_fault.do_pte_missing.__handle_mm_fault 0.19 ± 34% -51.3% 0.09 ± 37% perf-sched.sch_delay.avg.ms.__cond_resched.__kmalloc_cache_noprof.perf_event_mmap_event.perf_event_mmap.__mmap_region 0.14 ± 55% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.__wait_for_common.stop_two_cpus.migrate_swap.task_numa_migrate 0.14 ± 73% -82.5% 0.03 ±168% perf-sched.sch_delay.avg.ms.__cond_resched.change_pud_range.isra.0.change_protection_range 0.11 ± 59% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.down_write.vma_link_file.__mmap_new_vma.__mmap_region 0.04 ±132% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.down_write_killable.exec_mmap.begin_new_exec.load_elf_binary 0.02 ± 31% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.dput.__fput.__x64_sys_close.do_syscall_64 0.00 ±223% +51950.0% 0.26 ±212% perf-sched.sch_delay.avg.ms.__cond_resched.exit_mmap.__mmput.exit_mm.do_exit 0.25 ± 59% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.generic_perform_write.shmem_file_write_iter.vfs_write.ksys_write 0.12 ±145% -99.1% 0.00 ±141% perf-sched.sch_delay.avg.ms.__cond_resched.mutex_lock.perf_poll.do_poll.constprop 0.25 ± 41% -81.6% 0.05 ± 69% perf-sched.sch_delay.avg.ms.__cond_resched.stop_one_cpu.migrate_task_to.task_numa_migrate.isra 0.11 ± 59% -87.1% 0.01 ±198% perf-sched.sch_delay.avg.ms.__cond_resched.stop_one_cpu.sched_exec.bprm_execve.part 0.40 ± 50% -97.8% 0.01 ± 30% perf-sched.sch_delay.avg.ms.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown] 2.25 ±138% -99.6% 0.01 ± 7% perf-sched.sch_delay.avg.ms.devkmsg_read.vfs_read.ksys_read.do_syscall_64 0.32 ±104% -97.3% 0.01 ± 38% perf-sched.sch_delay.avg.ms.do_nanosleep.hrtimer_nanosleep.common_nsleep.__x64_sys_clock_nanosleep 0.01 ± 12% -34.9% 0.01 ± 18% perf-sched.sch_delay.avg.ms.irq_thread.kthread.ret_from_fork.ret_from_fork_asm 0.01 ± 20% -100.0% 0.00 perf-sched.sch_delay.avg.ms.irqentry_exit_to_user_mode.asm_sysvec_call_function_single.[unknown].[unknown] 0.19 ±185% -95.6% 0.01 ± 44% perf-sched.sch_delay.avg.ms.schedule_hrtimeout_range.ep_poll.do_epoll_wait.__x64_sys_epoll_wait 0.07 ± 20% -100.0% 0.00 perf-sched.sch_delay.avg.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write.vma_link_file 0.26 ± 17% -98.8% 0.00 ± 10% perf-sched.sch_delay.avg.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone 0.03 ± 51% -69.7% 0.01 ± 67% perf-sched.sch_delay.avg.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread 0.01 ± 55% +721.9% 0.10 ± 29% perf-sched.sch_delay.avg.ms.syscall_exit_to_user_mode.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown] 0.01 ±128% -83.6% 0.00 ± 20% perf-sched.sch_delay.avg.ms.wait_for_partner.fifo_open.do_dentry_open.vfs_open 0.06 ± 31% +1921.5% 1.23 ±165% perf-sched.sch_delay.avg.ms.worker_thread.kthread.ret_from_fork.ret_from_fork_asm 1.00 ±151% -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.__do_fault.do_read_fault.do_pte_missing.__handle_mm_fault 25.45 ± 94% -98.6% 0.36 ± 61% perf-sched.sch_delay.max.ms.__cond_resched.__kmalloc_cache_noprof.perf_event_mmap_event.perf_event_mmap.__mmap_region 4.56 ± 67% -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.__wait_for_common.stop_two_cpus.migrate_swap.task_numa_migrate 3.55 ± 97% -98.9% 0.04 ±189% perf-sched.sch_delay.max.ms.__cond_resched.change_pud_range.isra.0.change_protection_range 3.16 ± 78% -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.down_write.vma_link_file.__mmap_new_vma.__mmap_region 0.30 ±159% -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.down_write_killable.exec_mmap.begin_new_exec.load_elf_binary 0.03 ± 86% -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.dput.__fput.__x64_sys_close.do_syscall_64 0.00 ±223% +3.2e+06% 15.79 ±259% perf-sched.sch_delay.max.ms.__cond_resched.exit_mmap.__mmput.exit_mm.do_exit 3.09 ± 45% -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.generic_perform_write.shmem_file_write_iter.vfs_write.ksys_write 3.51 ± 21% -86.1% 0.49 ± 72% perf-sched.sch_delay.max.ms.__cond_resched.kmem_cache_alloc_noprof.vm_area_alloc.__mmap_new_vma.__mmap_region 3.59 ± 11% -92.0% 0.29 ±165% perf-sched.sch_delay.max.ms.__cond_resched.stop_one_cpu.migrate_task_to.task_numa_migrate.isra 1.60 ± 69% -95.7% 0.07 ±243% perf-sched.sch_delay.max.ms.__cond_resched.stop_one_cpu.sched_exec.bprm_execve.part 0.81 ± 43% -98.5% 0.01 ± 43% perf-sched.sch_delay.max.ms.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown] 1.02 ± 88% -98.1% 0.02 ± 47% perf-sched.sch_delay.max.ms.do_nanosleep.hrtimer_nanosleep.common_nsleep.__x64_sys_clock_nanosleep 9.68 ± 32% -92.2% 0.76 ± 72% perf-sched.sch_delay.max.ms.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown] 12.26 ±109% -92.9% 0.87 ±101% perf-sched.sch_delay.max.ms.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt.[unknown] 0.03 ±106% -100.0% 0.00 perf-sched.sch_delay.max.ms.irqentry_exit_to_user_mode.asm_sysvec_call_function_single.[unknown].[unknown] 37.84 ± 47% -100.0% 0.00 perf-sched.sch_delay.max.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write.vma_link_file 4.68 ± 36% -99.8% 0.01 ± 65% perf-sched.sch_delay.max.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone 0.36 ±186% -96.3% 0.01 ± 90% perf-sched.sch_delay.max.ms.wait_for_partner.fifo_open.do_dentry_open.vfs_open 97903 ± 4% -38.3% 60433 ± 29% perf-sched.total_wait_and_delay.count.ms 3.97 ± 6% -100.0% 0.00 perf-sched.wait_and_delay.avg.ms.__cond_resched.__kmalloc_cache_noprof.perf_event_mmap_event.perf_event_mmap.__mmap_region 302.41 ± 5% -27.4% 219.54 ± 14% perf-sched.wait_and_delay.avg.ms.irq_thread.kthread.ret_from_fork.ret_from_fork_asm 1.48 ± 6% -90.9% 0.14 ± 79% perf-sched.wait_and_delay.avg.ms.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown] 327.16 ± 9% -46.6% 174.81 ± 24% perf-sched.wait_and_delay.avg.ms.schedule_hrtimeout_range.do_poll.constprop.0.do_sys_poll 369.37 ± 2% -75.3% 91.05 ± 35% perf-sched.wait_and_delay.avg.ms.schedule_hrtimeout_range.ep_poll.do_epoll_wait.__x64_sys_epoll_wait 0.96 ± 6% -100.0% 0.00 perf-sched.wait_and_delay.avg.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write.vma_link_file 187.66 +120.6% 413.97 ± 14% perf-sched.wait_and_delay.avg.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm 1831 ± 9% -100.0% 0.00 perf-sched.wait_and_delay.count.__cond_resched.__kmalloc_cache_noprof.perf_event_mmap_event.perf_event_mmap.__mmap_region 6.17 ± 45% -79.7% 1.25 ±142% perf-sched.wait_and_delay.count.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm 40.50 ± 8% +245.7% 140.00 ± 23% perf-sched.wait_and_delay.count.schedule_hrtimeout_range.do_poll.constprop.0.do_sys_poll 13.17 ± 2% +624.4% 95.38 ± 19% perf-sched.wait_and_delay.count.schedule_hrtimeout_range.ep_poll.do_epoll_wait.__x64_sys_epoll_wait 73021 ± 3% -100.0% 0.00 perf-sched.wait_and_delay.count.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write.vma_link_file 11323 ± 3% -75.9% 2725 ± 28% perf-sched.wait_and_delay.count.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm 1887 ± 45% -96.1% 73.88 ± 78% perf-sched.wait_and_delay.count.syscall_exit_to_user_mode.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown] 1238 -34.5% 811.25 ± 13% perf-sched.wait_and_delay.count.worker_thread.kthread.ret_from_fork.ret_from_fork_asm 35.19 ± 57% -100.0% 0.00 perf-sched.wait_and_delay.max.ms.__cond_resched.__kmalloc_cache_noprof.perf_event_mmap_event.perf_event_mmap.__mmap_region 20.79 ± 19% -95.9% 0.84 ± 93% perf-sched.wait_and_delay.max.ms.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown] 1240 ± 20% -14.4% 1062 ± 10% perf-sched.wait_and_delay.max.ms.schedule_hrtimeout_range.do_poll.constprop.0.do_sys_poll 500.34 +31.2% 656.38 ± 39% perf-sched.wait_and_delay.max.ms.schedule_hrtimeout_range.ep_poll.do_epoll_wait.__x64_sys_epoll_wait 58.83 ± 39% -100.0% 0.00 perf-sched.wait_and_delay.max.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write.vma_link_file 1237 ± 34% +151.7% 3114 ± 25% perf-sched.wait_and_delay.max.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm 49.27 ±119% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.__alloc_pages_noprof.alloc_pages_mpol_noprof.folio_alloc_mpol_noprof.shmem_alloc_folio 58.17 ±187% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.__do_fault.do_read_fault.do_pte_missing.__handle_mm_fault 3.78 ± 5% -97.6% 0.09 ± 37% perf-sched.wait_time.avg.ms.__cond_resched.__kmalloc_cache_noprof.perf_event_mmap_event.perf_event_mmap.__mmap_region 2.99 ± 4% +15.4% 3.45 ± 10% perf-sched.wait_time.avg.ms.__cond_resched.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr.__sched_setaffinity 3.92 ± 5% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.__wait_for_common.stop_two_cpus.migrate_swap.task_numa_migrate 4.71 ± 8% -99.5% 0.02 ±170% perf-sched.wait_time.avg.ms.__cond_resched.change_pud_range.isra.0.change_protection_range 1.67 ± 20% -92.7% 0.12 ± 30% perf-sched.wait_time.avg.ms.__cond_resched.down_write.__mmap_new_vma.__mmap_region.do_mmap 2.10 ± 27% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.down_write.vma_link_file.__mmap_new_vma.__mmap_region 0.01 ± 44% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.down_write_killable.exec_mmap.begin_new_exec.load_elf_binary 1.67 ± 21% -94.3% 0.10 ± 35% perf-sched.wait_time.avg.ms.__cond_resched.down_write_killable.vm_mmap_pgoff.ksys_mmap_pgoff.do_syscall_64 0.04 ±133% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.dput.__fput.__x64_sys_close.do_syscall_64 2.30 ± 14% -95.5% 0.10 ± 42% perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc_noprof.vm_area_alloc.__mmap_new_vma.__mmap_region 2.00 ± 74% +2917.4% 60.44 ± 33% perf-sched.wait_time.avg.ms.devkmsg_read.vfs_read.ksys_read.do_syscall_64 29.19 ± 5% -38.5% 17.96 ± 28% perf-sched.wait_time.avg.ms.do_task_dead.do_exit.do_group_exit.__x64_sys_exit_group.x64_sys_call 0.37 ± 30% +5524.5% 20.95 ± 30% perf-sched.wait_time.avg.ms.do_wait.kernel_wait4.do_syscall_64.entry_SYSCALL_64_after_hwframe 302.40 ± 5% -27.4% 219.53 ± 14% perf-sched.wait_time.avg.ms.irq_thread.kthread.ret_from_fork.ret_from_fork_asm 1.40 ± 6% -92.7% 0.10 ± 18% perf-sched.wait_time.avg.ms.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown] 0.72 ±220% -100.0% 0.00 perf-sched.wait_time.avg.ms.irqentry_exit_to_user_mode.asm_sysvec_call_function_single.[unknown].[unknown] 326.84 ± 9% -46.6% 174.54 ± 24% perf-sched.wait_time.avg.ms.schedule_hrtimeout_range.do_poll.constprop.0.do_sys_poll 369.18 ± 2% -75.3% 91.04 ± 35% perf-sched.wait_time.avg.ms.schedule_hrtimeout_range.ep_poll.do_epoll_wait.__x64_sys_epoll_wait 0.89 ± 6% -100.0% 0.00 perf-sched.wait_time.avg.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write.vma_link_file 187.58 +120.6% 413.77 ± 14% perf-sched.wait_time.avg.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm 2.36 ± 29% +1759.6% 43.80 ± 33% perf-sched.wait_time.avg.ms.syscall_exit_to_user_mode.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown] 0.01 ±156% -97.9% 0.00 ±264% perf-sched.wait_time.avg.ms.wait_for_partner.fifo_open.do_dentry_open.vfs_open 340.69 ±135% -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.__alloc_pages_noprof.alloc_pages_mpol_noprof.folio_alloc_mpol_noprof.shmem_alloc_folio 535.09 ±128% -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.__do_fault.do_read_fault.do_pte_missing.__handle_mm_fault 22.04 ± 32% -98.4% 0.36 ± 61% perf-sched.wait_time.max.ms.__cond_resched.__kmalloc_cache_noprof.perf_event_mmap_event.perf_event_mmap.__mmap_region 13.57 ± 17% -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.__wait_for_common.stop_two_cpus.migrate_swap.task_numa_migrate 13.54 ± 10% -99.7% 0.04 ±189% perf-sched.wait_time.max.ms.__cond_resched.change_pud_range.isra.0.change_protection_range 10.17 ± 19% -95.2% 0.49 ± 56% perf-sched.wait_time.max.ms.__cond_resched.down_write.__mmap_new_vma.__mmap_region.do_mmap 11.35 ± 25% -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.down_write.vma_link_file.__mmap_new_vma.__mmap_region 0.01 ± 32% -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.down_write_killable.exec_mmap.begin_new_exec.load_elf_binary 10.62 ± 9% -96.5% 0.38 ± 72% perf-sched.wait_time.max.ms.__cond_resched.down_write_killable.vm_mmap_pgoff.ksys_mmap_pgoff.do_syscall_64 0.20 ±199% -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.dput.__fput.__x64_sys_close.do_syscall_64 14.42 ± 22% -96.6% 0.49 ± 72% perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc_noprof.vm_area_alloc.__mmap_new_vma.__mmap_region 4.00 ± 74% +19182.5% 772.23 ± 40% perf-sched.wait_time.max.ms.devkmsg_read.vfs_read.ksys_read.do_syscall_64 10.75 ± 98% +6512.2% 710.88 ± 56% perf-sched.wait_time.max.ms.do_wait.kernel_wait4.do_syscall_64.entry_SYSCALL_64_after_hwframe 15.80 ± 8% -95.2% 0.76 ± 72% perf-sched.wait_time.max.ms.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown] 11.64 ± 61% -98.9% 0.13 ±132% perf-sched.wait_time.max.ms.irqentry_exit_to_user_mode.asm_sysvec_call_function_single.[unknown] 2.94 ±213% -100.0% 0.00 perf-sched.wait_time.max.ms.irqentry_exit_to_user_mode.asm_sysvec_call_function_single.[unknown].[unknown] 1240 ± 20% -14.3% 1062 ± 10% perf-sched.wait_time.max.ms.schedule_hrtimeout_range.do_poll.constprop.0.do_sys_poll 500.11 +31.2% 656.37 ± 39% perf-sched.wait_time.max.ms.schedule_hrtimeout_range.ep_poll.do_epoll_wait.__x64_sys_epoll_wait 32.65 ± 33% -100.0% 0.00 perf-sched.wait_time.max.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write.vma_link_file 1237 ± 34% +151.6% 3113 ± 25% perf-sched.wait_time.max.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm 95.59 -95.6 0.00 perf-profile.calltrace.cycles-pp.__mmap 95.54 -95.5 0.00 perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.__mmap 95.54 -95.5 0.00 perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.__mmap 94.54 -94.5 0.00 perf-profile.calltrace.cycles-pp.ksys_mmap_pgoff.do_syscall_64.entry_SYSCALL_64_after_hwframe.__mmap 94.46 -94.0 0.41 ±138% perf-profile.calltrace.cycles-pp.vm_mmap_pgoff.ksys_mmap_pgoff.do_syscall_64.entry_SYSCALL_64_after_hwframe.__mmap 94.14 -93.7 0.40 ±136% perf-profile.calltrace.cycles-pp.__mmap_new_vma.__mmap_region.do_mmap.vm_mmap_pgoff.ksys_mmap_pgoff 93.79 -93.5 0.31 ±134% perf-profile.calltrace.cycles-pp.vma_link_file.__mmap_new_vma.__mmap_region.do_mmap.vm_mmap_pgoff 93.40 -93.4 0.00 perf-profile.calltrace.cycles-pp.rwsem_down_write_slowpath.down_write.vma_link_file.__mmap_new_vma.__mmap_region 93.33 -93.3 0.00 perf-profile.calltrace.cycles-pp.rwsem_optimistic_spin.rwsem_down_write_slowpath.down_write.vma_link_file.__mmap_new_vma 93.44 -93.3 0.14 ±264% perf-profile.calltrace.cycles-pp.down_write.vma_link_file.__mmap_new_vma.__mmap_region.do_mmap 94.45 -93.0 1.42 ± 60% perf-profile.calltrace.cycles-pp.do_mmap.vm_mmap_pgoff.ksys_mmap_pgoff.do_syscall_64.entry_SYSCALL_64_after_hwframe 94.25 -92.9 1.33 ± 61% perf-profile.calltrace.cycles-pp.__mmap_region.do_mmap.vm_mmap_pgoff.ksys_mmap_pgoff.do_syscall_64 92.89 -92.9 0.00 perf-profile.calltrace.cycles-pp.osq_lock.rwsem_optimistic_spin.rwsem_down_write_slowpath.down_write.vma_link_file 0.00 +1.1 1.09 ± 33% perf-profile.calltrace.cycles-pp.dup_mmap.dup_mm.copy_process.kernel_clone.__do_sys_clone 0.00 +1.4 1.37 ± 49% perf-profile.calltrace.cycles-pp.setlocale 0.00 +1.6 1.64 ± 47% perf-profile.calltrace.cycles-pp.__irq_exit_rcu.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.acpi_safe_halt.acpi_idle_do_entry 0.00 +1.6 1.64 ± 47% perf-profile.calltrace.cycles-pp.handle_softirqs.__irq_exit_rcu.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.acpi_safe_halt 0.00 +1.6 1.65 ± 43% perf-profile.calltrace.cycles-pp.do_sys_openat2.__x64_sys_openat.do_syscall_64.entry_SYSCALL_64_after_hwframe 0.00 +1.8 1.76 ± 44% perf-profile.calltrace.cycles-pp.__x64_sys_openat.do_syscall_64.entry_SYSCALL_64_after_hwframe 0.00 +1.9 1.93 ± 26% perf-profile.calltrace.cycles-pp.dup_mm.copy_process.kernel_clone.__do_sys_clone.do_syscall_64 0.00 +2.2 2.16 ± 44% perf-profile.calltrace.cycles-pp.do_pte_missing.__handle_mm_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault 0.00 +2.2 2.23 ± 33% perf-profile.calltrace.cycles-pp.copy_process.kernel_clone.__do_sys_clone.do_syscall_64.entry_SYSCALL_64_after_hwframe 0.00 +2.4 2.37 ± 36% perf-profile.calltrace.cycles-pp.zap_present_ptes.zap_pte_range.zap_pmd_range.unmap_page_range.unmap_vmas 0.00 +2.5 2.48 ± 32% perf-profile.calltrace.cycles-pp.get_cpu_sleep_time_us.get_idle_time.uptime_proc_show.seq_read_iter.vfs_read 0.00 +2.5 2.50 ± 45% perf-profile.calltrace.cycles-pp.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault 0.00 +2.5 2.54 ± 47% perf-profile.calltrace.cycles-pp.__mmput.exit_mm.do_exit.do_group_exit.__x64_sys_exit_group 0.00 +2.5 2.54 ± 47% perf-profile.calltrace.cycles-pp.exit_mm.do_exit.do_group_exit.__x64_sys_exit_group.x64_sys_call 0.00 +2.6 2.62 ± 35% perf-profile.calltrace.cycles-pp.__do_sys_clone.do_syscall_64.entry_SYSCALL_64_after_hwframe._Fork 0.00 +2.6 2.62 ± 35% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe._Fork 0.00 +2.6 2.62 ± 35% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe._Fork 0.00 +2.6 2.62 ± 35% perf-profile.calltrace.cycles-pp.kernel_clone.__do_sys_clone.do_syscall_64.entry_SYSCALL_64_after_hwframe._Fork 0.00 +2.7 2.68 ± 35% perf-profile.calltrace.cycles-pp.get_idle_time.uptime_proc_show.seq_read_iter.vfs_read.ksys_read 0.00 +2.8 2.77 ± 33% perf-profile.calltrace.cycles-pp.uptime_proc_show.seq_read_iter.vfs_read.ksys_read.do_syscall_64 0.00 +2.8 2.82 ± 32% perf-profile.calltrace.cycles-pp._Fork 0.00 +2.8 2.84 ± 45% perf-profile.calltrace.cycles-pp.do_user_addr_fault.exc_page_fault.asm_exc_page_fault 0.00 +2.8 2.84 ± 45% perf-profile.calltrace.cycles-pp.exc_page_fault.asm_exc_page_fault 0.00 +2.9 2.89 ± 39% perf-profile.calltrace.cycles-pp.event_function_call.perf_event_release_kernel.perf_release.__fput.task_work_run 0.00 +2.9 2.89 ± 39% perf-profile.calltrace.cycles-pp.smp_call_function_single.event_function_call.perf_event_release_kernel.perf_release.__fput 0.00 +3.1 3.10 ± 64% perf-profile.calltrace.cycles-pp.proc_reg_read_iter.vfs_read.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe 0.00 +3.1 3.10 ± 64% perf-profile.calltrace.cycles-pp.seq_read_iter.proc_reg_read_iter.vfs_read.ksys_read.do_syscall_64 0.00 +3.1 3.13 ± 33% perf-profile.calltrace.cycles-pp.asm_exc_page_fault 0.00 +3.2 3.18 ± 37% perf-profile.calltrace.cycles-pp.seq_read_iter.vfs_read.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe 0.00 +3.2 3.20 ± 28% perf-profile.calltrace.cycles-pp.mutex_unlock.sw_perf_event_destroy._free_event.perf_event_release_kernel.perf_release 0.00 +3.2 3.24 ± 39% perf-profile.calltrace.cycles-pp.bprm_execve.do_execveat_common.__x64_sys_execve.do_syscall_64.entry_SYSCALL_64_after_hwframe 0.00 +3.2 3.24 ± 36% perf-profile.calltrace.cycles-pp.__x64_sys_exit_group.x64_sys_call.do_syscall_64.entry_SYSCALL_64_after_hwframe 0.00 +3.2 3.24 ± 36% perf-profile.calltrace.cycles-pp.do_exit.do_group_exit.__x64_sys_exit_group.x64_sys_call.do_syscall_64 0.00 +3.2 3.24 ± 36% perf-profile.calltrace.cycles-pp.do_group_exit.__x64_sys_exit_group.x64_sys_call.do_syscall_64.entry_SYSCALL_64_after_hwframe 0.00 +3.2 3.24 ± 36% perf-profile.calltrace.cycles-pp.x64_sys_call.do_syscall_64.entry_SYSCALL_64_after_hwframe 0.00 +3.8 3.85 ± 39% perf-profile.calltrace.cycles-pp.__x64_sys_execve.do_syscall_64.entry_SYSCALL_64_after_hwframe.execve 0.00 +3.8 3.85 ± 39% perf-profile.calltrace.cycles-pp.do_execveat_common.__x64_sys_execve.do_syscall_64.entry_SYSCALL_64_after_hwframe.execve 0.00 +3.8 3.85 ± 39% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.execve 0.00 +3.8 3.85 ± 39% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.execve 0.00 +3.8 3.85 ± 39% perf-profile.calltrace.cycles-pp.execve 0.00 +4.0 4.04 ± 43% perf-profile.calltrace.cycles-pp.do_filp_open.do_sys_openat2.__x64_sys_openat.do_syscall_64.entry_SYSCALL_64_after_hwframe 0.00 +4.0 4.04 ± 43% perf-profile.calltrace.cycles-pp.path_openat.do_filp_open.do_sys_openat2.__x64_sys_openat.do_syscall_64 0.00 +4.1 4.10 ± 30% perf-profile.calltrace.cycles-pp.unmap_page_range.unmap_vmas.exit_mmap.__mmput.exit_mm 0.00 +4.2 4.18 ± 31% perf-profile.calltrace.cycles-pp.zap_pmd_range.unmap_page_range.unmap_vmas.exit_mmap.__mmput 0.00 +4.2 4.18 ± 31% perf-profile.calltrace.cycles-pp.zap_pte_range.zap_pmd_range.unmap_page_range.unmap_vmas.exit_mmap 0.00 +4.2 4.20 ± 28% perf-profile.calltrace.cycles-pp.unmap_vmas.exit_mmap.__mmput.exit_mm.do_exit 0.00 +4.2 4.25 ± 65% perf-profile.calltrace.cycles-pp.generic_perform_write.shmem_file_write_iter.vfs_write.ksys_write.do_syscall_64 0.00 +4.3 4.27 ± 26% perf-profile.calltrace.cycles-pp.__handle_mm_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault 0.00 +4.3 4.30 ± 22% perf-profile.calltrace.cycles-pp.__mmput.exit_mm.do_exit.do_group_exit.get_signal 0.00 +4.3 4.30 ± 22% perf-profile.calltrace.cycles-pp.exit_mm.do_exit.do_group_exit.get_signal.arch_do_signal_or_restart 0.00 +4.5 4.46 ± 59% perf-profile.calltrace.cycles-pp.shmem_file_write_iter.vfs_write.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe 0.00 +4.6 4.57 ± 58% perf-profile.calltrace.cycles-pp.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe.write.writen 0.00 +4.7 4.68 ± 55% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.write.writen.record__pushfn 0.00 +4.7 4.68 ± 55% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.write.writen.record__pushfn.perf_mmap__push 0.00 +4.7 4.68 ± 55% perf-profile.calltrace.cycles-pp.record__pushfn.perf_mmap__push.record__mmap_read_evlist.__cmd_record.cmd_record 0.00 +4.7 4.68 ± 55% perf-profile.calltrace.cycles-pp.write.writen.record__pushfn.perf_mmap__push.record__mmap_read_evlist 0.00 +4.7 4.68 ± 55% perf-profile.calltrace.cycles-pp.writen.record__pushfn.perf_mmap__push.record__mmap_read_evlist.__cmd_record 0.00 +4.9 4.90 ± 57% perf-profile.calltrace.cycles-pp.vfs_write.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe.write 0.00 +4.9 4.92 ± 26% perf-profile.calltrace.cycles-pp.sw_perf_event_destroy._free_event.perf_event_release_kernel.perf_release.__fput 0.00 +5.0 4.99 ±100% perf-profile.calltrace.cycles-pp.__intel_pmu_enable_all.perf_rotate_context.perf_mux_hrtimer_handler.__hrtimer_run_queues.hrtimer_interrupt 0.00 +5.0 4.99 ±100% perf-profile.calltrace.cycles-pp.perf_rotate_context.perf_mux_hrtimer_handler.__hrtimer_run_queues.hrtimer_interrupt.__sysvec_apic_timer_interrupt 0.00 +5.1 5.08 ±102% perf-profile.calltrace.cycles-pp.perf_mux_hrtimer_handler.__hrtimer_run_queues.hrtimer_interrupt.__sysvec_apic_timer_interrupt.sysvec_apic_timer_interrupt 0.00 +5.1 5.14 ± 28% perf-profile.calltrace.cycles-pp.perf_mmap__push.record__mmap_read_evlist.__cmd_record.cmd_record.run_builtin 0.00 +5.1 5.14 ± 28% perf-profile.calltrace.cycles-pp.record__mmap_read_evlist.__cmd_record.cmd_record.run_builtin.handle_internal_command 0.00 +5.4 5.43 ± 25% perf-profile.calltrace.cycles-pp._free_event.perf_event_release_kernel.perf_release.__fput.task_work_run 0.00 +5.8 5.82 ± 94% perf-profile.calltrace.cycles-pp.__sysvec_apic_timer_interrupt.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.acpi_safe_halt.acpi_idle_do_entry 0.00 +5.8 5.82 ± 94% perf-profile.calltrace.cycles-pp.hrtimer_interrupt.__sysvec_apic_timer_interrupt.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.acpi_safe_halt 0.00 +6.1 6.07 ± 90% perf-profile.calltrace.cycles-pp.__hrtimer_run_queues.hrtimer_interrupt.__sysvec_apic_timer_interrupt.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt 0.00 +6.6 6.62 ± 24% perf-profile.calltrace.cycles-pp.__cmd_record.cmd_record.run_builtin.handle_internal_command.main 0.00 +6.6 6.62 ± 24% perf-profile.calltrace.cycles-pp.cmd_record.run_builtin.handle_internal_command.main 0.00 +6.8 6.76 ± 18% perf-profile.calltrace.cycles-pp.exit_mmap.__mmput.exit_mm.do_exit.do_group_exit 0.00 +7.6 7.56 ± 76% perf-profile.calltrace.cycles-pp.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.acpi_safe_halt.acpi_idle_do_entry.acpi_idle_enter 0.00 +8.0 8.03 ± 27% perf-profile.calltrace.cycles-pp.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe.read 0.00 +8.0 8.03 ± 27% perf-profile.calltrace.cycles-pp.vfs_read.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe.read 0.00 +8.0 8.05 ± 68% perf-profile.calltrace.cycles-pp.acpi_safe_halt.acpi_idle_do_entry.acpi_idle_enter.cpuidle_enter_state.cpuidle_enter 0.00 +8.1 8.13 ± 28% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.read 0.00 +8.1 8.13 ± 28% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.read 0.00 +8.1 8.13 ± 28% perf-profile.calltrace.cycles-pp.read 0.00 +9.1 9.05 ± 35% perf-profile.calltrace.cycles-pp.handle_internal_command.main 0.00 +9.1 9.05 ± 35% perf-profile.calltrace.cycles-pp.main 0.00 +9.1 9.05 ± 35% perf-profile.calltrace.cycles-pp.run_builtin.handle_internal_command.main 0.00 +9.3 9.26 ± 30% perf-profile.calltrace.cycles-pp.perf_event_release_kernel.perf_release.__fput.task_work_run.do_exit 0.00 +9.3 9.26 ± 30% perf-profile.calltrace.cycles-pp.perf_release.__fput.task_work_run.do_exit.do_group_exit 0.00 +10.1 10.14 ± 28% perf-profile.calltrace.cycles-pp.__fput.task_work_run.do_exit.do_group_exit.get_signal 0.00 +10.2 10.23 ± 27% perf-profile.calltrace.cycles-pp.task_work_run.do_exit.do_group_exit.get_signal.arch_do_signal_or_restart 0.00 +11.0 10.98 ± 55% perf-profile.calltrace.cycles-pp.asm_sysvec_reschedule_ipi.acpi_safe_halt.acpi_idle_do_entry.acpi_idle_enter.cpuidle_enter_state 0.00 +20.6 20.64 ± 30% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe 0.00 +20.6 20.64 ± 30% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe 1.21 ± 3% +36.6 37.80 ± 12% perf-profile.calltrace.cycles-pp.cpuidle_enter.cpuidle_idle_call.do_idle.cpu_startup_entry.start_secondary 1.21 ± 3% +36.6 37.80 ± 12% perf-profile.calltrace.cycles-pp.cpuidle_idle_call.do_idle.cpu_startup_entry.start_secondary.common_startup_64 1.22 ± 3% +36.8 38.00 ± 13% perf-profile.calltrace.cycles-pp.do_idle.cpu_startup_entry.start_secondary.common_startup_64 1.22 ± 3% +36.9 38.10 ± 13% perf-profile.calltrace.cycles-pp.cpu_startup_entry.start_secondary.common_startup_64 1.22 ± 3% +36.9 38.10 ± 13% perf-profile.calltrace.cycles-pp.start_secondary.common_startup_64 1.21 ± 3% +37.2 38.43 ± 11% perf-profile.calltrace.cycles-pp.acpi_idle_do_entry.acpi_idle_enter.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call 1.21 ± 3% +37.2 38.43 ± 11% perf-profile.calltrace.cycles-pp.acpi_idle_enter.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call.do_idle 1.21 ± 3% +37.3 38.54 ± 12% perf-profile.calltrace.cycles-pp.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call.do_idle.cpu_startup_entry 1.22 ± 3% +37.6 38.84 ± 12% perf-profile.calltrace.cycles-pp.common_startup_64 2.19 ± 3% +53.9 56.10 ± 19% perf-profile.calltrace.cycles-pp.asm_sysvec_apic_timer_interrupt.acpi_safe_halt.acpi_idle_do_entry.acpi_idle_enter.cpuidle_enter_state 95.60 -95.2 0.41 ±138% perf-profile.children.cycles-pp.__mmap 94.14 -93.7 0.49 ±130% perf-profile.children.cycles-pp.__mmap_new_vma 93.79 -93.5 0.31 ±134% perf-profile.children.cycles-pp.vma_link_file 93.40 -93.4 0.00 perf-profile.children.cycles-pp.rwsem_down_write_slowpath 93.33 -93.3 0.00 perf-profile.children.cycles-pp.rwsem_optimistic_spin 94.55 -93.1 1.42 ± 60% perf-profile.children.cycles-pp.ksys_mmap_pgoff 92.91 -92.9 0.00 perf-profile.children.cycles-pp.osq_lock 93.44 -92.7 0.75 ±109% perf-profile.children.cycles-pp.down_write 94.46 -92.6 1.84 ± 34% perf-profile.children.cycles-pp.vm_mmap_pgoff 94.45 -92.6 1.84 ± 34% perf-profile.children.cycles-pp.do_mmap 94.25 -92.6 1.66 ± 37% perf-profile.children.cycles-pp.__mmap_region 95.58 -44.8 50.78 ± 11% perf-profile.children.cycles-pp.entry_SYSCALL_64_after_hwframe 95.58 -44.8 50.78 ± 11% perf-profile.children.cycles-pp.do_syscall_64 0.00 +1.1 1.09 ± 33% perf-profile.children.cycles-pp.dup_mmap 0.00 +1.4 1.37 ± 49% perf-profile.children.cycles-pp.setlocale 0.00 +1.9 1.93 ± 26% perf-profile.children.cycles-pp.dup_mm 0.03 ± 70% +2.0 1.99 ± 36% perf-profile.children.cycles-pp.handle_softirqs 0.00 +2.0 1.99 ± 36% perf-profile.children.cycles-pp.__irq_exit_rcu 0.00 +2.0 2.02 ± 38% perf-profile.children.cycles-pp.folios_put_refs 0.00 +2.1 2.06 ± 52% perf-profile.children.cycles-pp._raw_spin_lock 0.00 +2.2 2.16 ± 44% perf-profile.children.cycles-pp.do_pte_missing 0.00 +2.2 2.21 ± 68% perf-profile.children.cycles-pp.link_path_walk 0.00 +2.2 2.23 ± 33% perf-profile.children.cycles-pp.copy_process 0.00 +2.3 2.30 ± 40% perf-profile.children.cycles-pp.__tlb_batch_free_encoded_pages 0.00 +2.3 2.30 ± 40% perf-profile.children.cycles-pp.free_pages_and_swap_cache 0.00 +2.3 2.34 ± 46% perf-profile.children.cycles-pp.walk_component 0.00 +2.4 2.37 ± 36% perf-profile.children.cycles-pp.zap_present_ptes 0.00 +2.5 2.48 ± 32% perf-profile.children.cycles-pp.get_cpu_sleep_time_us 0.00 +2.6 2.62 ± 35% perf-profile.children.cycles-pp.__do_sys_clone 0.00 +2.6 2.62 ± 35% perf-profile.children.cycles-pp.kernel_clone 0.00 +2.7 2.68 ± 35% perf-profile.children.cycles-pp.get_idle_time 0.00 +2.8 2.77 ± 33% perf-profile.children.cycles-pp.uptime_proc_show 0.00 +2.9 2.91 ± 32% perf-profile.children.cycles-pp._Fork 0.00 +3.1 3.10 ± 64% perf-profile.children.cycles-pp.proc_reg_read_iter 0.00 +3.2 3.24 ± 39% perf-profile.children.cycles-pp.bprm_execve 0.00 +3.2 3.24 ± 36% perf-profile.children.cycles-pp.__x64_sys_exit_group 0.00 +3.2 3.24 ± 36% perf-profile.children.cycles-pp.x64_sys_call 0.00 +3.8 3.85 ± 39% perf-profile.children.cycles-pp.__x64_sys_execve 0.00 +3.8 3.85 ± 39% perf-profile.children.cycles-pp.do_execveat_common 0.00 +3.8 3.85 ± 39% perf-profile.children.cycles-pp.execve 0.00 +4.0 3.99 ± 38% perf-profile.children.cycles-pp.mutex_unlock 0.00 +4.2 4.19 ± 31% perf-profile.children.cycles-pp.zap_pte_range 0.00 +4.2 4.25 ± 65% perf-profile.children.cycles-pp.generic_perform_write 0.00 +4.3 4.29 ± 29% perf-profile.children.cycles-pp.unmap_page_range 0.00 +4.3 4.29 ± 29% perf-profile.children.cycles-pp.zap_pmd_range 0.00 +4.3 4.31 ± 51% perf-profile.children.cycles-pp.do_filp_open 0.00 +4.3 4.31 ± 51% perf-profile.children.cycles-pp.path_openat 0.19 ± 23% +4.4 4.60 ± 26% perf-profile.children.cycles-pp.__handle_mm_fault 0.00 +4.5 4.46 ± 59% perf-profile.children.cycles-pp.shmem_file_write_iter 0.00 +4.5 4.55 ± 24% perf-profile.children.cycles-pp.event_function_call 0.00 +4.5 4.55 ± 24% perf-profile.children.cycles-pp.smp_call_function_single 0.00 +4.6 4.58 ± 30% perf-profile.children.cycles-pp.unmap_vmas 0.51 ± 6% +4.6 5.14 ± 24% perf-profile.children.cycles-pp.handle_mm_fault 0.00 +4.7 4.68 ± 55% perf-profile.children.cycles-pp.record__pushfn 0.00 +4.7 4.68 ± 55% perf-profile.children.cycles-pp.writen 0.00 +4.8 4.80 ± 48% perf-profile.children.cycles-pp.do_sys_openat2 0.77 ± 3% +4.8 5.59 ± 21% perf-profile.children.cycles-pp.exc_page_fault 0.76 ± 3% +4.8 5.59 ± 21% perf-profile.children.cycles-pp.do_user_addr_fault 0.00 +4.9 4.90 ± 57% perf-profile.children.cycles-pp.ksys_write 0.00 +4.9 4.90 ± 57% perf-profile.children.cycles-pp.vfs_write 0.00 +4.9 4.90 ± 48% perf-profile.children.cycles-pp.__x64_sys_openat 0.00 +4.9 4.92 ± 26% perf-profile.children.cycles-pp.sw_perf_event_destroy 0.00 +5.0 4.99 ±100% perf-profile.children.cycles-pp.perf_rotate_context 0.00 +5.0 5.01 ± 54% perf-profile.children.cycles-pp.write 0.00 +5.1 5.09 ±102% perf-profile.children.cycles-pp.perf_mux_hrtimer_handler 0.00 +5.4 5.43 ± 25% perf-profile.children.cycles-pp._free_event 1.18 +5.6 6.78 ± 20% perf-profile.children.cycles-pp.asm_exc_page_fault 0.46 +5.6 6.07 ± 90% perf-profile.children.cycles-pp.__hrtimer_run_queues 0.00 +5.7 5.75 ± 39% perf-profile.children.cycles-pp.perf_mmap__push 0.00 +5.7 5.75 ± 39% perf-profile.children.cycles-pp.record__mmap_read_evlist 0.53 +5.8 6.28 ± 89% perf-profile.children.cycles-pp.__sysvec_apic_timer_interrupt 0.53 +5.8 6.28 ± 89% perf-profile.children.cycles-pp.hrtimer_interrupt 0.00 +6.6 6.65 ± 77% perf-profile.children.cycles-pp.__intel_pmu_enable_all 0.00 +6.8 6.85 ± 20% perf-profile.children.cycles-pp.exit_mm 0.58 ± 2% +7.6 8.14 ± 75% perf-profile.children.cycles-pp.sysvec_apic_timer_interrupt 0.00 +7.7 7.67 ± 23% perf-profile.children.cycles-pp.exit_mmap 0.00 +7.7 7.67 ± 30% perf-profile.children.cycles-pp.seq_read_iter 0.00 +7.7 7.72 ± 80% perf-profile.children.cycles-pp.asm_sysvec_reschedule_ipi 0.00 +7.8 7.75 ± 23% perf-profile.children.cycles-pp.__mmput 0.00 +8.0 8.03 ± 27% perf-profile.children.cycles-pp.ksys_read 0.00 +8.0 8.03 ± 27% perf-profile.children.cycles-pp.vfs_read 0.00 +8.1 8.13 ± 28% perf-profile.children.cycles-pp.read 0.02 ±141% +9.0 9.05 ± 35% perf-profile.children.cycles-pp.__cmd_record 0.02 ±141% +9.0 9.05 ± 35% perf-profile.children.cycles-pp.cmd_record 0.02 ±141% +9.0 9.05 ± 35% perf-profile.children.cycles-pp.handle_internal_command 0.02 ±141% +9.0 9.05 ± 35% perf-profile.children.cycles-pp.main 0.02 ±141% +9.0 9.05 ± 35% perf-profile.children.cycles-pp.run_builtin 0.00 +9.3 9.26 ± 30% perf-profile.children.cycles-pp.perf_event_release_kernel 0.00 +9.3 9.26 ± 30% perf-profile.children.cycles-pp.perf_release 1.02 ± 4% +9.3 10.33 ± 27% perf-profile.children.cycles-pp.task_work_run 0.00 +11.0 11.05 ± 28% perf-profile.children.cycles-pp.__fput 0.00 +15.8 15.85 ± 25% perf-profile.children.cycles-pp.arch_do_signal_or_restart 0.00 +15.8 15.85 ± 25% perf-profile.children.cycles-pp.get_signal 0.00 +19.1 19.09 ± 19% perf-profile.children.cycles-pp.do_exit 0.00 +19.1 19.09 ± 19% perf-profile.children.cycles-pp.do_group_exit 1.70 ± 2% +30.7 32.41 ± 21% perf-profile.children.cycles-pp.asm_sysvec_apic_timer_interrupt 1.22 ± 3% +36.9 38.10 ± 13% perf-profile.children.cycles-pp.start_secondary 1.21 ± 3% +37.2 38.43 ± 11% perf-profile.children.cycles-pp.acpi_idle_do_entry 1.21 ± 3% +37.2 38.43 ± 11% perf-profile.children.cycles-pp.acpi_idle_enter 1.21 ± 3% +37.2 38.43 ± 11% perf-profile.children.cycles-pp.acpi_safe_halt 1.22 ± 3% +37.3 38.54 ± 12% perf-profile.children.cycles-pp.cpuidle_idle_call 1.21 ± 3% +37.3 38.54 ± 12% perf-profile.children.cycles-pp.cpuidle_enter 1.21 ± 3% +37.3 38.54 ± 12% perf-profile.children.cycles-pp.cpuidle_enter_state 1.22 ± 3% +37.6 38.84 ± 12% perf-profile.children.cycles-pp.common_startup_64 1.22 ± 3% +37.6 38.84 ± 12% perf-profile.children.cycles-pp.cpu_startup_entry 1.22 ± 3% +37.6 38.84 ± 12% perf-profile.children.cycles-pp.do_idle 92.37 -92.4 0.00 perf-profile.self.cycles-pp.osq_lock 0.00 +2.1 2.06 ± 52% perf-profile.self.cycles-pp._raw_spin_lock 0.00 +2.6 2.61 ± 36% perf-profile.self.cycles-pp.smp_call_function_single 0.00 +3.7 3.68 ± 37% perf-profile.self.cycles-pp.mutex_unlock 0.00 +6.6 6.65 ± 77% perf-profile.self.cycles-pp.__intel_pmu_enable_all 1.19 ± 3% +29.2 30.38 ± 15% perf-profile.self.cycles-pp.acpi_safe_halt