From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org Received: from kanga.kvack.org (kanga.kvack.org [205.233.56.17]) (using TLSv1 with cipher DHE-RSA-AES256-SHA (256/256 bits)) (No client certificate requested) by smtp.lore.kernel.org (Postfix) with ESMTPS id B9DABCAC5A5 for ; Thu, 25 Sep 2025 11:00:57 +0000 (UTC) Received: by kanga.kvack.org (Postfix) id F146C8E000A; Thu, 25 Sep 2025 07:00:56 -0400 (EDT) Received: by kanga.kvack.org (Postfix, from userid 40) id EEC048E0001; Thu, 25 Sep 2025 07:00:56 -0400 (EDT) X-Delivered-To: int-list-linux-mm@kvack.org Received: by kanga.kvack.org (Postfix, from userid 63042) id DDAC58E000A; Thu, 25 Sep 2025 07:00:56 -0400 (EDT) X-Delivered-To: linux-mm@kvack.org Received: from relay.hostedemail.com (smtprelay0014.hostedemail.com [216.40.44.14]) by kanga.kvack.org (Postfix) with ESMTP id CB0C48E0001 for ; Thu, 25 Sep 2025 07:00:56 -0400 (EDT) Received: from smtpin22.hostedemail.com (a10.router.float.18 [10.200.18.1]) by unirelay10.hostedemail.com (Postfix) with ESMTP id 80E23C0460 for ; Thu, 25 Sep 2025 11:00:56 +0000 (UTC) X-FDA: 83927480112.22.03EC6B7 Received: from us-smtp-delivery-124.mimecast.com (us-smtp-delivery-124.mimecast.com [170.10.133.124]) by imf12.hostedemail.com (Postfix) with ESMTP id C17D84000B for ; Thu, 25 Sep 2025 11:00:53 +0000 (UTC) Authentication-Results: imf12.hostedemail.com; dkim=pass header.d=redhat.com header.s=mimecast20190719 header.b=YSBVZ+3y; spf=pass (imf12.hostedemail.com: domain of david@redhat.com designates 170.10.133.124 as permitted sender) smtp.mailfrom=david@redhat.com; dmarc=pass (policy=quarantine) header.from=redhat.com ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=hostedemail.com; s=arc-20220608; t=1758798054; h=from:from:sender:reply-to:subject:subject:date:date: message-id:message-id:to:to:cc:cc:mime-version:mime-version: content-type:content-type: content-transfer-encoding:content-transfer-encoding: in-reply-to:in-reply-to:references:references:dkim-signature; bh=LsFjckaZRSZSgvIBBGozNR9QSfDm9ZdOZyuvNGkLoik=; b=AqpeVnuccq91GXSZmYGAQPl1gCo09K2f1sW+QSwP/DxKWzwQgJNHOozaFGOluHnJtNVnCq 0wiAQrGpCvkSYMU5elxZbljweiW2EZH4bb9SvQLOE0Ylc4o0Z0CzkUIxVwuP+p5xZBXiPa pl8CAjwvucIqkKeGQBIgTP3z+03V+Lw= ARC-Authentication-Results: i=1; imf12.hostedemail.com; dkim=pass header.d=redhat.com header.s=mimecast20190719 header.b=YSBVZ+3y; spf=pass (imf12.hostedemail.com: domain of david@redhat.com designates 170.10.133.124 as permitted sender) smtp.mailfrom=david@redhat.com; dmarc=pass (policy=quarantine) header.from=redhat.com ARC-Seal: i=1; s=arc-20220608; d=hostedemail.com; t=1758798054; a=rsa-sha256; cv=none; b=uf62wbn135zUNqR9GtlmU4ahJDrP3+Qy1QZpZX3PpfmoW1ik3gtHO80ZkrvmHBzNH4noUj CzPGxonsqYKAmR5gud9NhbVNwe+/6C6l8yeUDMkCOz9Yk04K9gLabWd98ykTiTyoYAF72N UCFZWxnpFwmbf7KOv6xHL6xhLmK7Ubs= DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=redhat.com; s=mimecast20190719; t=1758798053; h=from:from:reply-to:subject:subject:date:date:message-id:message-id: to:to:cc:cc:mime-version:mime-version:content-type:content-type: content-transfer-encoding:content-transfer-encoding: in-reply-to:in-reply-to:references:references:autocrypt:autocrypt; bh=LsFjckaZRSZSgvIBBGozNR9QSfDm9ZdOZyuvNGkLoik=; b=YSBVZ+3ynHMzQ6isRHnb4Cw32CrMrNNX9jfwnVxUuYVjnX35B23Syw1LwBikenmr1GXUWi xRzBtVMh9joA7LMSInPL7YMAk6xAY/mY+AitvBgt0no0KsonJMLBhwn/Rli/JcDabcUrcY n/qdjlWV+ERchNoZA0mikQ5ZDu1i36Y= Received: from mail-wm1-f71.google.com (mail-wm1-f71.google.com [209.85.128.71]) by relay.mimecast.com with ESMTP with STARTTLS (version=TLSv1.3, cipher=TLS_AES_256_GCM_SHA384) id us-mta-634-b6LY_QkcMoiAE3Rkhdd6vg-1; Thu, 25 Sep 2025 07:00:52 -0400 X-MC-Unique: b6LY_QkcMoiAE3Rkhdd6vg-1 X-Mimecast-MFC-AGG-ID: b6LY_QkcMoiAE3Rkhdd6vg_1758798051 Received: by mail-wm1-f71.google.com with SMTP id 5b1f17b1804b1-46da436df64so9527815e9.0 for ; Thu, 25 Sep 2025 04:00:51 -0700 (PDT) X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20230601; t=1758798051; x=1759402851; h=content-transfer-encoding:in-reply-to:autocrypt:content-language :from:references:cc:to:subject:user-agent:mime-version:date :message-id:x-gm-message-state:from:to:cc:subject:date:message-id :reply-to; bh=LsFjckaZRSZSgvIBBGozNR9QSfDm9ZdOZyuvNGkLoik=; b=AqEZC/ZekDkScRGiN7p+YeXOhr+paVa+swMDe4gyL1NL5yk5TjjEN3H1ilO9BskD6M Be94kqN314R4XscDNIkV2fi1VCub3iQY2KDjr+kWydslJ5Pzaq7EgkdyaUPxfc6HNrA0 oAKv+4yzLqLcMI6CpQLGWGQDiZ7qWPe9wNfUfwnwyTY4CH22/DlIzMfMRexTX9zdA9Pr k/DQJUquofyQUhtZi6JeAQM/+9KKewLH7FqrUAVOY6Zf9qD5d8SxXNQnrsVXtVv5lo2o AuTv02jw6buRlx9tDGZwa7nfG7fxBFbrdKtonGQWCv9UfHOeLBM+GW/q64z3KYaXwTiz HpOw== X-Forwarded-Encrypted: i=1; AJvYcCXEUvCC+twUUtSJW3rkXHPkpTzeBlIY7ij9gsIfxf505u9z0O7vepZfb99l5YoSvFHnib9zsFn2yw==@kvack.org X-Gm-Message-State: AOJu0Yyuj/nBZQYUpgM3O5QC2wOYu5+e4tNlZcuMwfIj5l9XCXBggulE CITtAJS0i6zaAaVHJFIPL3vXwoXhDzsSy0N4fFADZDBYr6gzhnnyRmfwV3gmEHDnyH14zL1VplE UmBLRrw8I3SJcVvcfVnPnD+Y45kzonSBalhpH3Hd0n/ZpzJtFwnLV X-Gm-Gg: ASbGnctx6q8X+TQOmP0Zr2MgMyVVPpLSfdhT1rGTuaRF24MPxmyFyfsj7MalzdwDuVU BF2UvdQIS/qypklWlsAgYd/5R9djOZKR2V+gMTYd0bo+Lcsv6tdC264hDQKNmRYfSD9W+Y4L1Qk IlEgU4UOaxbt6wUHHGzgQKrtXY2tBChmrrD8HiTzpHmy2s3mqtxXFZGOcuWQ7IDZy0Z6QRX8zzk /pw2u2Fuu5+8v1s2qNTS4GuX+Gck6pdKUoUIbZn5af4mQXe7/irb6VtrUb4UNJUBz+MGFzFS5NK QBk6i+DkjkNgOrkyTW52K++tm4ajvSH6uoFWBAkl+lpzf5/KI3qVhISoNLKMT13Fe4HhR6EXrmB 8WGz9INCXEmV/7VrrSVdhwWlLnUvjMY64fTB7NQDJCR0Jzn/8AK9QJhr73mIFlVJHaGOn X-Received: by 2002:a05:600c:1554:b0:45d:d6fc:24f7 with SMTP id 5b1f17b1804b1-46e32a158camr31228595e9.32.1758798050157; Thu, 25 Sep 2025 04:00:50 -0700 (PDT) X-Google-Smtp-Source: AGHT+IHBJES16mG0dvjX/+8YEMFt/mXYaxImgtqz08XZYHmHVSOGfz52OI8Lf1kSYhL10lrl8HbeQA== X-Received: by 2002:a05:600c:1554:b0:45d:d6fc:24f7 with SMTP id 5b1f17b1804b1-46e32a158camr31227445e9.32.1758798049427; Thu, 25 Sep 2025 04:00:49 -0700 (PDT) Received: from ?IPV6:2003:d8:2f3f:f800:c101:5c9f:3bc9:3d08? (p200300d82f3ff800c1015c9f3bc93d08.dip0.t-ipconnect.de. [2003:d8:2f3f:f800:c101:5c9f:3bc9:3d08]) by smtp.gmail.com with ESMTPSA id ffacd0b85a97d-40fc6de90desm2659224f8f.47.2025.09.25.04.00.45 (version=TLS1_3 cipher=TLS_AES_128_GCM_SHA256 bits=128/128); Thu, 25 Sep 2025 04:00:48 -0700 (PDT) Message-ID: Date: Thu, 25 Sep 2025 13:00:45 +0200 MIME-Version: 1.0 User-Agent: Mozilla Thunderbird Subject: Re: [PATCH v7 05/12] KVM: guest_memfd: Add flag to remove from direct map To: "Roy, Patrick" Cc: "pbonzini@redhat.com" , "corbet@lwn.net" , "maz@kernel.org" , "oliver.upton@linux.dev" , "joey.gouly@arm.com" , "suzuki.poulose@arm.com" , "yuzenghui@huawei.com" , "catalin.marinas@arm.com" , "will@kernel.org" , "tglx@linutronix.de" , "mingo@redhat.com" , "bp@alien8.de" , "dave.hansen@linux.intel.com" , "x86@kernel.org" , "hpa@zytor.com" , "luto@kernel.org" , "peterz@infradead.org" , "willy@infradead.org" , "akpm@linux-foundation.org" , "lorenzo.stoakes@oracle.com" , "Liam.Howlett@oracle.com" , "vbabka@suse.cz" , "rppt@kernel.org" , "surenb@google.com" , "mhocko@suse.com" , "song@kernel.org" , "jolsa@kernel.org" , "ast@kernel.org" , "daniel@iogearbox.net" , "andrii@kernel.org" , "martin.lau@linux.dev" , "eddyz87@gmail.com" , "yonghong.song@linux.dev" , "john.fastabend@gmail.com" , "kpsingh@kernel.org" , "sdf@fomichev.me" , "haoluo@google.com" , "jgg@ziepe.ca" , "jhubbard@nvidia.com" , "peterx@redhat.com" , "jannh@google.com" , "pfalcato@suse.de" , "shuah@kernel.org" , "seanjc@google.com" , "kvm@vger.kernel.org" , "linux-doc@vger.kernel.org" , "linux-kernel@vger.kernel.org" , "linux-arm-kernel@lists.infradead.org" , "kvmarm@lists.linux.dev" , "linux-fsdevel@vger.kernel.org" , "linux-mm@kvack.org" , "bpf@vger.kernel.org" , "linux-kselftest@vger.kernel.org" , "Cali, Marco" , "Kalyazin, Nikita" , "Thomson, Jack" , "derekmn@amazon.co.uk" , "tabba@google.com" , "ackerleytng@google.com" References: <20250924151101.2225820-4-patrick.roy@campus.lmu.de> <20250924152214.7292-1-roypat@amazon.co.uk> <20250924152214.7292-2-roypat@amazon.co.uk> From: David Hildenbrand Autocrypt: addr=david@redhat.com; keydata= xsFNBFXLn5EBEAC+zYvAFJxCBY9Tr1xZgcESmxVNI/0ffzE/ZQOiHJl6mGkmA1R7/uUpiCjJ dBrn+lhhOYjjNefFQou6478faXE6o2AhmebqT4KiQoUQFV4R7y1KMEKoSyy8hQaK1umALTdL QZLQMzNE74ap+GDK0wnacPQFpcG1AE9RMq3aeErY5tujekBS32jfC/7AnH7I0v1v1TbbK3Gp XNeiN4QroO+5qaSr0ID2sz5jtBLRb15RMre27E1ImpaIv2Jw8NJgW0k/D1RyKCwaTsgRdwuK Kx/Y91XuSBdz0uOyU/S8kM1+ag0wvsGlpBVxRR/xw/E8M7TEwuCZQArqqTCmkG6HGcXFT0V9 PXFNNgV5jXMQRwU0O/ztJIQqsE5LsUomE//bLwzj9IVsaQpKDqW6TAPjcdBDPLHvriq7kGjt WhVhdl0qEYB8lkBEU7V2Yb+SYhmhpDrti9Fq1EsmhiHSkxJcGREoMK/63r9WLZYI3+4W2rAc UucZa4OT27U5ZISjNg3Ev0rxU5UH2/pT4wJCfxwocmqaRr6UYmrtZmND89X0KigoFD/XSeVv jwBRNjPAubK9/k5NoRrYqztM9W6sJqrH8+UWZ1Idd/DdmogJh0gNC0+N42Za9yBRURfIdKSb B3JfpUqcWwE7vUaYrHG1nw54pLUoPG6sAA7Mehl3nd4pZUALHwARAQABzSREYXZpZCBIaWxk ZW5icmFuZCA8ZGF2aWRAcmVkaGF0LmNvbT7CwZoEEwEIAEQCGwMCF4ACGQEFCwkIBwICIgIG FQoJCAsCBBYCAwECHgcWIQQb2cqtc1xMOkYN/MpN3hD3AP+DWgUCaJzangUJJlgIpAAKCRBN 3hD3AP+DWhAxD/9wcL0A+2rtaAmutaKTfxhTP0b4AAp1r/eLxjrbfbCCmh4pqzBhmSX/4z11 opn2KqcOsueRF1t2ENLOWzQu3Roiny2HOU7DajqB4dm1BVMaXQya5ae2ghzlJN9SIoopTWlR 0Af3hPj5E2PYvQhlcqeoehKlBo9rROJv/rjmr2x0yOM8qeTroH/ZzNlCtJ56AsE6Tvl+r7cW 3x7/Jq5WvWeudKrhFh7/yQ7eRvHCjd9bBrZTlgAfiHmX9AnCCPRPpNGNedV9Yty2Jnxhfmbv Pw37LA/jef8zlCDyUh2KCU1xVEOWqg15o1RtTyGV1nXV2O/mfuQJud5vIgzBvHhypc3p6VZJ lEf8YmT+Ol5P7SfCs5/uGdWUYQEMqOlg6w9R4Pe8d+mk8KGvfE9/zTwGg0nRgKqlQXrWRERv cuEwQbridlPAoQHrFWtwpgYMXx2TaZ3sihcIPo9uU5eBs0rf4mOERY75SK+Ekayv2ucTfjxr Kf014py2aoRJHuvy85ee/zIyLmve5hngZTTe3Wg3TInT9UTFzTPhItam6dZ1xqdTGHZYGU0O otRHcwLGt470grdiob6PfVTXoHlBvkWRadMhSuG4RORCDpq89vu5QralFNIf3EysNohoFy2A LYg2/D53xbU/aa4DDzBb5b1Rkg/udO1gZocVQWrDh6I2K3+cCs7BTQRVy5+RARAA59fefSDR 9nMGCb9LbMX+TFAoIQo/wgP5XPyzLYakO+94GrgfZjfhdaxPXMsl2+o8jhp/hlIzG56taNdt VZtPp3ih1AgbR8rHgXw1xwOpuAd5lE1qNd54ndHuADO9a9A0vPimIes78Hi1/yy+ZEEvRkHk /kDa6F3AtTc1m4rbbOk2fiKzzsE9YXweFjQvl9p+AMw6qd/iC4lUk9g0+FQXNdRs+o4o6Qvy iOQJfGQ4UcBuOy1IrkJrd8qq5jet1fcM2j4QvsW8CLDWZS1L7kZ5gT5EycMKxUWb8LuRjxzZ 3QY1aQH2kkzn6acigU3HLtgFyV1gBNV44ehjgvJpRY2cC8VhanTx0dZ9mj1YKIky5N+C0f21 zvntBqcxV0+3p8MrxRRcgEtDZNav+xAoT3G0W4SahAaUTWXpsZoOecwtxi74CyneQNPTDjNg azHmvpdBVEfj7k3p4dmJp5i0U66Onmf6mMFpArvBRSMOKU9DlAzMi4IvhiNWjKVaIE2Se9BY FdKVAJaZq85P2y20ZBd08ILnKcj7XKZkLU5FkoA0udEBvQ0f9QLNyyy3DZMCQWcwRuj1m73D sq8DEFBdZ5eEkj1dCyx+t/ga6x2rHyc8Sl86oK1tvAkwBNsfKou3v+jP/l14a7DGBvrmlYjO 59o3t6inu6H7pt7OL6u6BQj7DoMAEQEAAcLBfAQYAQgAJgIbDBYhBBvZyq1zXEw6Rg38yk3e EPcA/4NaBQJonNqrBQkmWAihAAoJEE3eEPcA/4NaKtMQALAJ8PzprBEXbXcEXwDKQu+P/vts IfUb1UNMfMV76BicGa5NCZnJNQASDP/+bFg6O3gx5NbhHHPeaWz/VxlOmYHokHodOvtL0WCC 8A5PEP8tOk6029Z+J+xUcMrJClNVFpzVvOpb1lCbhjwAV465Hy+NUSbbUiRxdzNQtLtgZzOV Zw7jxUCs4UUZLQTCuBpFgb15bBxYZ/BL9MbzxPxvfUQIPbnzQMcqtpUs21CMK2PdfCh5c4gS sDci6D5/ZIBw94UQWmGpM/O1ilGXde2ZzzGYl64glmccD8e87OnEgKnH3FbnJnT4iJchtSvx yJNi1+t0+qDti4m88+/9IuPqCKb6Stl+s2dnLtJNrjXBGJtsQG/sRpqsJz5x1/2nPJSRMsx9 5YfqbdrJSOFXDzZ8/r82HgQEtUvlSXNaXCa95ez0UkOG7+bDm2b3s0XahBQeLVCH0mw3RAQg r7xDAYKIrAwfHHmMTnBQDPJwVqxJjVNr7yBic4yfzVWGCGNE4DnOW0vcIeoyhy9vnIa3w1uZ 3iyY2Nsd7JxfKu1PRhCGwXzRw5TlfEsoRI7V9A8isUCoqE2Dzh3FvYHVeX4Us+bRL/oqareJ CIFqgYMyvHj7Q06kTKmauOe4Nf0l0qEkIuIzfoLJ3qr5UyXc2hLtWyT9Ir+lYlX9efqh7mOY qIws/H2t In-Reply-To: <20250924152214.7292-2-roypat@amazon.co.uk> X-Mimecast-Spam-Score: 0 X-Mimecast-MFC-PROC-ID: TFUBCZDY2RFOmTFPkRM9kzFvFB4MHPSENOVuPFZzJbA_1758798051 X-Mimecast-Originator: redhat.com Content-Language: en-US Content-Type: text/plain; charset=UTF-8; format=flowed Content-Transfer-Encoding: 7bit X-Rspamd-Queue-Id: C17D84000B X-Stat-Signature: htfeiq5meet1t6dhe3xg6q9driof4z61 X-Rspam-User: X-Rspamd-Server: rspam01 X-HE-Tag: 1758798053-907117 X-HE-Meta: U2FsdGVkX19Fb6UR/Y3l/I7M9cFFKpBhHaOwf2UbPwj62og4kq5XFUr6P2xwC/KZDWpeeuO1DOH74D1TnnHIaX9y6aadO/by9ucJuvr/uAD0rByxPdIbiQQjaVKcuBcCMFJpi2jlb3SvhwGcQC4vCcvv9veIHM7RYOhdBhlKYoPd+3YERztfCmsz5hkZw51TcvIrLhkAH36hGEJ3g7BFj+fhyUnseDkUSCvWMOX6STWAvJ9xH4pNtp2DOEw7JzsqTfpkL2O3U2QHG8k0Tw3ra8ggKxK8DBzo8kn4axZplCGe1/e0iTy8qjhBf1fwOozPVoBxC2CtXjljHZwhS4HnI9O6hSxd8o8mmxUhGycV/VJrNkJe7d8XKcreyt9sugJesTsgtKkcjPAak34eokDipziA9LE428YbXnW0AaSJb+mMzlHLA69nBEDV9xPPWFYVQaAofjxjeYtYGGkVmNOt+aYatIxCyDieo7blNLpDnLYgSJV0PahAw/hzJHkibAcufT5vymres578u9/QIy94qEn5ud1mg7UsYVH4rJpWtmCLN1HTbVtwzwgXTnQK8KZrTrEYU4vJAfZLVb8A0+sCriSgALMHVv/HipG9gEZ0kgfCNcOaK5rWRYF99xkQQ1Rb28e+qN39OcAgxXGFHhNQRPa6/AyNmIfpE52LbnBSy+QeF86HHUtScOsAakY/ho1NUf/3mVTYltah+We3JM2ksPzS+ArBszibHqB6QhRKomCp+Q9bbhSTaviJU9oUEgXwWJiNWaVeLuxZynqvMrh2dsOIQ1CMocEjNvWHyhyFdukmuEzkr7FBkTSNowUZJoLRfi6z0AdewNBGM5sWrQWSLcvve0otZjOcOBg3ii8m0CItbkDUWzl1RHYccR048bHxltdeg6g8Ai/9zd5kgxF1UY7WHXlnXCsuJPPlWpbhuQZM77Z5jOR7qR/tqk8vofOzXo4H/7vV7LmhIQCG5j/ 3LRQIHLi kDdIAXBjgZ8eiPOdYayzcQseXV92dIw9BYd8rY8IDqdk5ZKvGRLOmwelPj6skRKWVRSkt9DIlXCEBibU1O8aaYm9b7nqFfhj2MBlWvSGp/E5nIxZCnYNY33ZKYOPbFaXm9l2HzXvWP/+gqp0XLT5h/TLGzdr3lprOp6LJiCscwGHfAH6WeO9if/4K9jFM7Px+e2PtyVruVwFop64PF/jztt7hLcT+vEYOC8JkC2htATlgxl7/4fcl3NGjob1v53uT/EK+U2u0g2JesNJn4dU+hyU9y+VNM81qKF+fvNad1jxZMZJvh/jw1CgI/ZA6o2FeJexa/uZyfHjbH+0s1U1p7MbIyrquliuzNW4Dvs3WAcrWpZ+mGTwayvDPn39k4UBcLgm1l6SFf5Mngv84PuWHgxq85fMp3GCY3tnU2g78eDQEx6vriDakwYgTUsZvwOZvqNeHuDaMLOxfUmPrykMhVLINR19w3r6qNTKLcmAzZQRoNGPltoMIdcPAkUqymZ4DLqnb1ngGqy5dKc5mc6sDEJL/Cy7najdLpbA3yTr7/lS4wa7lAH2CA1VkmaxwPEeBmBDXkJlXgMyTHj1jV04EdvTFuak7xglADz7s74uiBOal9Fv5mjcKs9msHIpo//Ca4usC6pNBnxOaob8hgmQDlbMD4HJq/O8Hd7evhEB6A/Wd+zbARI3PUUhN/2QopZPs/aqa6vematJv5XgYOmkHzENJSlClrghd8xdf X-Bogosity: Ham, tests=bogofilter, spamicity=0.000000, version=1.2.4 Sender: owner-linux-mm@kvack.org Precedence: bulk X-Loop: owner-majordomo@kvack.org List-ID: List-Subscribe: List-Unsubscribe: On 24.09.25 17:22, Roy, Patrick wrote: > Add GUEST_MEMFD_FLAG_NO_DIRECT_MAP flag for KVM_CREATE_GUEST_MEMFD() > ioctl. When set, guest_memfd folios will be removed from the direct map > after preparation, with direct map entries only restored when the folios > are freed. > > To ensure these folios do not end up in places where the kernel cannot > deal with them, set AS_NO_DIRECT_MAP on the guest_memfd's struct > address_space if GUEST_MEMFD_FLAG_NO_DIRECT_MAP is requested. > > Add KVM_CAP_GUEST_MEMFD_NO_DIRECT_MAP to let userspace discover whether > guest_memfd supports GUEST_MEMFD_FLAG_NO_DIRECT_MAP. Support depends on > guest_memfd itself being supported, but also on whether linux supports > manipulatomg the direct map at page granularity at all (possible most of > the time, outliers being arm64 where its impossible if the direct map > has been setup using hugepages, as arm64 cannot break these apart due to > break-before-make semantics, and powerpc, which does not select > ARCH_HAS_SET_DIRECT_MAP, though also doesn't support guest_memfd > anyway). > > Note that this flag causes removal of direct map entries for all > guest_memfd folios independent of whether they are "shared" or "private" > (although current guest_memfd only supports either all folios in the > "shared" state, or all folios in the "private" state if > GUEST_MEMFD_FLAG_MMAP is not set). The usecase for removing direct map > entries of also the shared parts of guest_memfd are a special type of > non-CoCo VM where, host userspace is trusted to have access to all of > guest memory, but where Spectre-style transient execution attacks > through the host kernel's direct map should still be mitigated. In this > setup, KVM retains access to guest memory via userspace mappings of > guest_memfd, which are reflected back into KVM's memslots via > userspace_addr. This is needed for things like MMIO emulation on x86_64 > to work. > > Direct map entries are zapped right before guest or userspace mappings > of gmem folios are set up, e.g. in kvm_gmem_fault_user_mapping() or > kvm_gmem_get_pfn() [called from the KVM MMU code]. The only place where > a gmem folio can be allocated without being mapped anywhere is > kvm_gmem_populate(), where handling potential failures of direct map > removal is not possible (by the time direct map removal is attempted, > the folio is already marked as prepared, meaning attempting to re-try > kvm_gmem_populate() would just result in -EEXIST without fixing up the > direct map state). These folios are then removed form the direct map > upon kvm_gmem_get_pfn(), e.g. when they are mapped into the guest later. > > Signed-off-by: Patrick Roy > --- > Documentation/virt/kvm/api.rst | 5 +++ > arch/arm64/include/asm/kvm_host.h | 12 ++++++ > include/linux/kvm_host.h | 6 +++ > include/uapi/linux/kvm.h | 2 + > virt/kvm/guest_memfd.c | 61 ++++++++++++++++++++++++++++++- > virt/kvm/kvm_main.c | 5 +++ > 6 files changed, 90 insertions(+), 1 deletion(-) > > diff --git a/Documentation/virt/kvm/api.rst b/Documentation/virt/kvm/api.rst > index c17a87a0a5ac..b52c14d58798 100644 > --- a/Documentation/virt/kvm/api.rst > +++ b/Documentation/virt/kvm/api.rst > @@ -6418,6 +6418,11 @@ When the capability KVM_CAP_GUEST_MEMFD_MMAP is supported, the 'flags' field > supports GUEST_MEMFD_FLAG_MMAP. Setting this flag on guest_memfd creation > enables mmap() and faulting of guest_memfd memory to host userspace. > > +When the capability KVM_CAP_GMEM_NO_DIRECT_MAP is supported, the 'flags' field > +supports GUEST_MEMFG_FLAG_NO_DIRECT_MAP. Setting this flag makes the guest_memfd > +instance behave similarly to memfd_secret, and unmaps the memory backing it from > +the kernel's address space after allocation. > + Do we want to document what the implication of that is? Meaning, limitations etc. I recall that we would need the user mapping for gmem slots to be properly set up. Is that still the case in this patch set? > When the KVM MMU performs a PFN lookup to service a guest fault and the backing > guest_memfd has the GUEST_MEMFD_FLAG_MMAP set, then the fault will always be > consumed from guest_memfd, regardless of whether it is a shared or a private > diff --git a/arch/arm64/include/asm/kvm_host.h b/arch/arm64/include/asm/kvm_host.h > index 2f2394cce24e..0bfd8e5fd9de 100644 > --- a/arch/arm64/include/asm/kvm_host.h > +++ b/arch/arm64/include/asm/kvm_host.h > @@ -19,6 +19,7 @@ > #include > #include > #include > +#include > #include > #include > #include > @@ -1706,5 +1707,16 @@ void compute_fgu(struct kvm *kvm, enum fgt_group_id fgt); > void get_reg_fixed_bits(struct kvm *kvm, enum vcpu_sysreg reg, u64 *res0, u64 *res1); > void check_feature_map(void); > > +#ifdef CONFIG_KVM_GUEST_MEMFD > +static inline bool kvm_arch_gmem_supports_no_direct_map(void) > +{ > + /* > + * Without FWB, direct map access is needed in kvm_pgtable_stage2_map(), > + * as it calls dcache_clean_inval_poc(). > + */ > + return can_set_direct_map() && cpus_have_final_cap(ARM64_HAS_STAGE2_FWB); > +} > +#define kvm_arch_gmem_supports_no_direct_map kvm_arch_gmem_supports_no_direct_map > +#endif /* CONFIG_KVM_GUEST_MEMFD */ > I strongly assume that the aarch64 support should be moved to a separate patch -- if possible, see below. > #endif /* __ARM64_KVM_HOST_H__ */ > diff --git a/include/linux/kvm_host.h b/include/linux/kvm_host.h > index 1d0585616aa3..73a15cade54a 100644 > --- a/include/linux/kvm_host.h > +++ b/include/linux/kvm_host.h > @@ -731,6 +731,12 @@ static inline bool kvm_arch_has_private_mem(struct kvm *kvm) > bool kvm_arch_supports_gmem_mmap(struct kvm *kvm); > #endif > > +#ifdef CONFIG_KVM_GUEST_MEMFD > +#ifndef kvm_arch_gmem_supports_no_direct_map > +#define kvm_arch_gmem_supports_no_direct_map can_set_direct_map > +#endif Hm, wouldn't it be better to have an opt-in per arch, and really only unlock the ones we know work (tested etc), explicitly in separate patches. [...] > > #include "kvm_mm.h" > > @@ -42,6 +45,44 @@ static int __kvm_gmem_prepare_folio(struct kvm *kvm, struct kvm_memory_slot *slo > return 0; > } > > +#define KVM_GMEM_FOLIO_NO_DIRECT_MAP BIT(0) > + > +static bool kvm_gmem_folio_no_direct_map(struct folio *folio) > +{ > + return ((u64) folio->private) & KVM_GMEM_FOLIO_NO_DIRECT_MAP; > +} > + > +static int kvm_gmem_folio_zap_direct_map(struct folio *folio) > +{ > + if (kvm_gmem_folio_no_direct_map(folio)) > + return 0; > + > + int r = set_direct_map_valid_noflush(folio_page(folio, 0), folio_nr_pages(folio), > + false); > + > + if (!r) { > + unsigned long addr = (unsigned long) folio_address(folio); empty line missing. > + folio->private = (void *) ((u64) folio->private & KVM_GMEM_FOLIO_NO_DIRECT_MAP); > + flush_tlb_kernel_range(addr, addr + folio_size(folio)); > + } > + > + return r; > +} > + > +static void kvm_gmem_folio_restore_direct_map(struct folio *folio) > +{ > + /* > + * Direct map restoration cannot fail, as the only error condition > + * for direct map manipulation is failure to allocate page tables > + * when splitting huge pages, but this split would have already > + * happened in set_direct_map_invalid_noflush() in kvm_gmem_folio_zap_direct_map(). > + * Thus set_direct_map_valid_noflush() here only updates prot bits. > + */ > + if (kvm_gmem_folio_no_direct_map(folio)) > + set_direct_map_valid_noflush(folio_page(folio, 0), folio_nr_pages(folio), > + true); > +} > + > static inline void kvm_gmem_mark_prepared(struct folio *folio) > { > folio_mark_uptodate(folio); > @@ -324,13 +365,14 @@ static vm_fault_t kvm_gmem_fault_user_mapping(struct vm_fault *vmf) > struct inode *inode = file_inode(vmf->vma->vm_file); > struct folio *folio; > vm_fault_t ret = VM_FAULT_LOCKED; > + int err; > > if (((loff_t)vmf->pgoff << PAGE_SHIFT) >= i_size_read(inode)) > return VM_FAULT_SIGBUS; > > folio = kvm_gmem_get_folio(inode, vmf->pgoff); > if (IS_ERR(folio)) { > - int err = PTR_ERR(folio); > + err = PTR_ERR(folio); > > if (err == -EAGAIN) > return VM_FAULT_RETRY; > @@ -348,6 +390,13 @@ static vm_fault_t kvm_gmem_fault_user_mapping(struct vm_fault *vmf) > kvm_gmem_mark_prepared(folio); > } > > + err = kvm_gmem_folio_zap_direct_map(folio); > + I'd drop this empty line here. > + if (err) { > + ret = vmf_error(err); > + goto out_folio; > + } > + > vmf->page = folio_file_page(folio, vmf->pgoff); > > out_folio: > @@ -435,6 +484,8 @@ static void kvm_gmem_free_folio(struct folio *folio) > kvm_pfn_t pfn = page_to_pfn(page); > int order = folio_order(folio); > > + kvm_gmem_folio_restore_direct_map(folio); > + > kvm_arch_gmem_invalidate(pfn, pfn + (1ul << order)); > } > > @@ -499,6 +550,9 @@ static int __kvm_gmem_create(struct kvm *kvm, loff_t size, u64 flags) > /* Unmovable mappings are supposed to be marked unevictable as well. */ > WARN_ON_ONCE(!mapping_unevictable(inode->i_mapping)); > > + if (flags & GUEST_MEMFD_FLAG_NO_DIRECT_MAP) > + mapping_set_no_direct_map(inode->i_mapping); > + > kvm_get_kvm(kvm); > gmem->kvm = kvm; > xa_init(&gmem->bindings); > @@ -523,6 +577,9 @@ int kvm_gmem_create(struct kvm *kvm, struct kvm_create_guest_memfd *args) > if (kvm_arch_supports_gmem_mmap(kvm)) > valid_flags |= GUEST_MEMFD_FLAG_MMAP; > > + if (kvm_arch_gmem_supports_no_direct_map()) > + valid_flags |= GUEST_MEMFD_FLAG_NO_DIRECT_MAP; > + > if (flags & ~valid_flags) > return -EINVAL; > > @@ -687,6 +744,8 @@ int kvm_gmem_get_pfn(struct kvm *kvm, struct kvm_memory_slot *slot, > if (!is_prepared) > r = kvm_gmem_prepare_folio(kvm, slot, gfn, folio); > > + kvm_gmem_folio_zap_direct_map(folio); > + > folio_unlock(folio); > > if (!r) > diff --git a/virt/kvm/kvm_main.c b/virt/kvm/kvm_main.c > index 18f29ef93543..b5e702d95230 100644 > --- a/virt/kvm/kvm_main.c > +++ b/virt/kvm/kvm_main.c > @@ -65,6 +65,7 @@ > #include > > #include > +#include Likely not required here. -- Cheers David / dhildenb