From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org Received: from kanga.kvack.org (kanga.kvack.org [205.233.56.17]) by smtp.lore.kernel.org (Postfix) with ESMTP id 38E02C433FE for ; Mon, 17 Oct 2022 10:15:52 +0000 (UTC) Received: by kanga.kvack.org (Postfix) id 82AC16B0072; Mon, 17 Oct 2022 06:15:51 -0400 (EDT) Received: by kanga.kvack.org (Postfix, from userid 40) id 7DC166B0074; Mon, 17 Oct 2022 06:15:51 -0400 (EDT) X-Delivered-To: int-list-linux-mm@kvack.org Received: by kanga.kvack.org (Postfix, from userid 63042) id 6541B6B0075; Mon, 17 Oct 2022 06:15:51 -0400 (EDT) X-Delivered-To: linux-mm@kvack.org Received: from relay.hostedemail.com (smtprelay0017.hostedemail.com [216.40.44.17]) by kanga.kvack.org (Postfix) with ESMTP id 4D9B66B0072 for ; Mon, 17 Oct 2022 06:15:51 -0400 (EDT) Received: from smtpin28.hostedemail.com (a10.router.float.18 [10.200.18.1]) by unirelay04.hostedemail.com (Postfix) with ESMTP id 21C161A0C89 for ; Mon, 17 Oct 2022 10:15:51 +0000 (UTC) X-FDA: 80030035302.28.DC93C98 Received: from mail-wr1-f53.google.com (mail-wr1-f53.google.com [209.85.221.53]) by imf03.hostedemail.com (Postfix) with ESMTP id BC5C920028 for ; Mon, 17 Oct 2022 10:15:50 +0000 (UTC) Received: by mail-wr1-f53.google.com with SMTP id r13so17684389wrj.11 for ; Mon, 17 Oct 2022 03:15:50 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=20210112; h=cc:to:subject:message-id:date:from:in-reply-to:references :mime-version:from:to:cc:subject:date:message-id:reply-to; bh=FcjnuXRaDiwRJahvKb8RBCSyzSGwOC5LjhLUGe1uDng=; b=hyvfh2EkmGJkje0pk/QxmrMK3wu58HezXpCWZTir75qkSbdgybYs5f9BdEJQ8VRxZo y6aRpYY6PJTNs0keePmRwtJhkNJD/XdF8OsFUhWZ+UwwjrnzDoAYIKRkdVUgvJixyXjW nD/o6Eqw8u2mx1XAbOWJsbL1mUmxxl7uY7YrsW4c+w0aEXLsr3iTAgFzKFCcXNq9JWeD DtktHswPI1Ib69o/b5Xw5PWUIeu21lJtPIx06bfk/qBH1eUXTV/osa+5ULa0LTre2WLP s1m1famLesITF3XC4ZTWNGzXqcgELAkbqC6CXhHknsOgV91Knx2CcUJjk+K6FMwFZKrp i0nw== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20210112; h=cc:to:subject:message-id:date:from:in-reply-to:references :mime-version:x-gm-message-state:from:to:cc:subject:date:message-id :reply-to; bh=FcjnuXRaDiwRJahvKb8RBCSyzSGwOC5LjhLUGe1uDng=; b=W9DsUlIL7M6+9AzYl56u1Wb4JfeeX2KMBbXLRQaq0mEo1YBYOoFCwUCRM+HmoS4axD OxPvSTUDp9Uyxt/o3G9R+/fMm2kJ57iXSE9JzXPVjmpusOlJs5A4JcaMa+yfg8GUNN2s M0j/0QPSRjZAEIDNTIqYIIIOaov1lD6Ui1yZfbTDKC02yn8qvrP4vNm2zq/qU7/FM7iB 7BExEJlV7N5SThFONsy2c3PYydr+wnglE6S+1mUew2DlbOI+PgT0KQkihd6bCzFAkJpT N9E8gX0QWKn41LR+C7yF2ycrTtbqFrMuuFC0tdJxFR0pVQxlruD3he/RKJ+PvmIJk6k+ Ii5A== X-Gm-Message-State: ACrzQf3nKEk2/7WhELew47V/JiOGDkpEwaJhoX9pqxCXtc7rdfHg2nPL TMAySimk4rehNu4ACVLXwxGwSfCth4sOBupzaMbYD7xfi87KZA== X-Google-Smtp-Source: AMsMyM5H+U3jrxyNNUs6l1VKxbOPrLqVbM8TSFTJ1Nj0WQRe3+Q8Gepb86AIPBI4Du2SI+pEA/5+OnPmPbAp2YR5khI= X-Received: by 2002:a2e:bd12:0:b0:264:7373:3668 with SMTP id n18-20020a2ebd12000000b0026473733668mr3490403ljq.18.1666001738408; Mon, 17 Oct 2022 03:15:38 -0700 (PDT) MIME-Version: 1.0 References: <20220915142913.2213336-1-chao.p.peng@linux.intel.com> <20220915142913.2213336-6-chao.p.peng@linux.intel.com> <20221012023516.GA3218049@chaop.bj.intel.com> In-Reply-To: <20221012023516.GA3218049@chaop.bj.intel.com> From: Fuad Tabba Date: Mon, 17 Oct 2022 11:15:02 +0100 Message-ID: Subject: Re: [PATCH v8 5/8] KVM: Register/unregister the guest private memory regions To: Chao Peng Cc: kvm@vger.kernel.org, linux-kernel@vger.kernel.org, linux-mm@kvack.org, linux-fsdevel@vger.kernel.org, linux-api@vger.kernel.org, linux-doc@vger.kernel.org, qemu-devel@nongnu.org, Paolo Bonzini , Jonathan Corbet , Sean Christopherson , Vitaly Kuznetsov , Wanpeng Li , Jim Mattson , Joerg Roedel , Thomas Gleixner , Ingo Molnar , Borislav Petkov , x86@kernel.org, "H . Peter Anvin" , Hugh Dickins , Jeff Layton , "J . Bruce Fields" , Andrew Morton , Shuah Khan , Mike Rapoport , Steven Price , "Maciej S . Szmigiero" , Vlastimil Babka , Vishal Annapurve , Yu Zhang , "Kirill A . Shutemov" , luto@kernel.org, jun.nakajima@intel.com, dave.hansen@intel.com, ak@linux.intel.com, david@redhat.com, aarcange@redhat.com, ddutile@redhat.com, dhildenb@redhat.com, Quentin Perret , Michael Roth , mhocko@suse.com, Muchun Song , wei.w.wang@intel.com Content-Type: text/plain; charset="UTF-8" ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=hostedemail.com; s=arc-20220608; t=1666001750; h=from:from:sender:reply-to:subject:subject:date:date: message-id:message-id:to:to:cc:cc:mime-version:mime-version: content-type:content-type:content-transfer-encoding: in-reply-to:in-reply-to:references:references:dkim-signature; bh=FcjnuXRaDiwRJahvKb8RBCSyzSGwOC5LjhLUGe1uDng=; b=08axAyxv9cqcmR18FgCfmsI+K5WEAWX/53pOlqDY4Lu6c9y1BBKgo5xBKOeSf76oOlA00x nSwJgX+e93QzFZzyx5jpynB57fiI/FIvAaERw25vxbyUXKD2bx214EWj9KSgmod5BwfBu2 3dhqsNjG2IjAF1TPVSZ/dobpUHLgssU= ARC-Authentication-Results: i=1; imf03.hostedemail.com; dkim=pass header.d=google.com header.s=20210112 header.b=hyvfh2Ek; spf=pass (imf03.hostedemail.com: domain of tabba@google.com designates 209.85.221.53 as permitted sender) smtp.mailfrom=tabba@google.com; dmarc=pass (policy=reject) header.from=google.com ARC-Seal: i=1; s=arc-20220608; d=hostedemail.com; t=1666001750; a=rsa-sha256; cv=none; b=s7VRkh3gkcuYM07BHcYAIdtD/pDdrY/Sa4Cc2NUNy8tjS1bIOcgCyOWX3XOcghaYGV/pBO QrWxcU6IRAv+afxODtXvwuPDMC75GGJNyQIFgJ/+QD7CstfFZG858bmeTbpJgcF95kkss2 IGUEWp9HKCOa8D+yINXI4WLBOBAcxgI= X-Rspamd-Server: rspam12 X-Rspam-User: Authentication-Results: imf03.hostedemail.com; dkim=pass header.d=google.com header.s=20210112 header.b=hyvfh2Ek; spf=pass (imf03.hostedemail.com: domain of tabba@google.com designates 209.85.221.53 as permitted sender) smtp.mailfrom=tabba@google.com; dmarc=pass (policy=reject) header.from=google.com X-Stat-Signature: ifnocfni38xrok1yudnxdeacp7fcqwyy X-Rspamd-Queue-Id: BC5C920028 X-HE-Tag: 1666001750-926495 X-Bogosity: Ham, tests=bogofilter, spamicity=0.000000, version=1.2.4 Sender: owner-linux-mm@kvack.org Precedence: bulk X-Loop: owner-majordomo@kvack.org List-ID: Hi, > > > +#ifdef CONFIG_HAVE_KVM_PRIVATE_MEM > > > +#define KVM_MEM_ATTR_SHARED 0x0001 > > > +static int kvm_vm_ioctl_set_mem_attr(struct kvm *kvm, gpa_t gpa, gpa_t size, > > > + bool is_private) > > > +{ > > > > I wonder if this ioctl should be implemented as an arch-specific > > ioctl. In this patch it performs some actions that pKVM might not need > > or might want to do differently. > > I think it's doable. We can provide the mem_attr_array kind thing in > common code and let arch code decide to use it or not. Currently > mem_attr_array is defined in the struct kvm, if those bytes are > unnecessary for pKVM it can even be moved to arch definition, but that > also loses the potential code sharing for confidential usages in other > non-architectures, e.g. if ARM also supports such usage. Or it can be > provided through a different CONFIG_ instead of > CONFIG_HAVE_KVM_PRIVATE_MEM. This sounds good. Thank you. /fuad > Thanks, > Chao > > > > pKVM tracks the sharing status in the stage-2 page table's software > > bits, so it can avoid the overhead of using mem_attr_array. > > > > Also, this ioctl calls kvm_zap_gfn_range(), as does the invalidation > > notifier (introduced in patch 8). For pKVM, the kind of zapping (or > > the information conveyed to the hypervisor) might need to be different > > depending on the cause; whether it's invalidation or change of sharing > > status. > > > > > Thanks, > > /fuad > > > > > > > + gfn_t start, end; > > > + unsigned long index; > > > + void *entry; > > > + int r; > > > + > > > + if (size == 0 || gpa + size < gpa) > > > + return -EINVAL; > > > + if (gpa & (PAGE_SIZE - 1) || size & (PAGE_SIZE - 1)) > > > + return -EINVAL; > > > + > > > + start = gpa >> PAGE_SHIFT; > > > + end = (gpa + size - 1 + PAGE_SIZE) >> PAGE_SHIFT; > > > + > > > + /* > > > + * Guest memory defaults to private, kvm->mem_attr_array only stores > > > + * shared memory. > > > + */ > > > + entry = is_private ? NULL : xa_mk_value(KVM_MEM_ATTR_SHARED); > > > + > > > + for (index = start; index < end; index++) { > > > + r = xa_err(xa_store(&kvm->mem_attr_array, index, entry, > > > + GFP_KERNEL_ACCOUNT)); > > > + if (r) > > > + goto err; > > > + } > > > + > > > + kvm_zap_gfn_range(kvm, start, end); > > > + > > > + return r; > > > +err: > > > + for (; index > start; index--) > > > + xa_erase(&kvm->mem_attr_array, index); > > > + return r; > > > +} > > > +#endif /* CONFIG_HAVE_KVM_PRIVATE_MEM */ > > > + > > > #ifdef CONFIG_HAVE_KVM_PM_NOTIFIER > > > static int kvm_pm_notifier_call(struct notifier_block *bl, > > > unsigned long state, > > > @@ -1165,6 +1206,9 @@ static struct kvm *kvm_create_vm(unsigned long type, const char *fdname) > > > spin_lock_init(&kvm->mn_invalidate_lock); > > > rcuwait_init(&kvm->mn_memslots_update_rcuwait); > > > xa_init(&kvm->vcpu_array); > > > +#ifdef CONFIG_HAVE_KVM_PRIVATE_MEM > > > + xa_init(&kvm->mem_attr_array); > > > +#endif > > > > > > INIT_LIST_HEAD(&kvm->gpc_list); > > > spin_lock_init(&kvm->gpc_lock); > > > @@ -1338,6 +1382,9 @@ static void kvm_destroy_vm(struct kvm *kvm) > > > kvm_free_memslots(kvm, &kvm->__memslots[i][0]); > > > kvm_free_memslots(kvm, &kvm->__memslots[i][1]); > > > } > > > +#ifdef CONFIG_HAVE_KVM_PRIVATE_MEM > > > + xa_destroy(&kvm->mem_attr_array); > > > +#endif > > > cleanup_srcu_struct(&kvm->irq_srcu); > > > cleanup_srcu_struct(&kvm->srcu); > > > kvm_arch_free_vm(kvm); > > > @@ -1541,6 +1588,11 @@ static void kvm_replace_memslot(struct kvm *kvm, > > > } > > > } > > > > > > +bool __weak kvm_arch_has_private_mem(struct kvm *kvm) > > > +{ > > > + return false; > > > +} > > > + > > > static int check_memory_region_flags(const struct kvm_user_mem_region *mem) > > > { > > > u32 valid_flags = KVM_MEM_LOG_DIRTY_PAGES; > > > @@ -4703,6 +4755,24 @@ static long kvm_vm_ioctl(struct file *filp, > > > r = kvm_vm_ioctl_set_memory_region(kvm, &mem); > > > break; > > > } > > > +#ifdef CONFIG_HAVE_KVM_PRIVATE_MEM > > > + case KVM_MEMORY_ENCRYPT_REG_REGION: > > > + case KVM_MEMORY_ENCRYPT_UNREG_REGION: { > > > + struct kvm_enc_region region; > > > + bool set = ioctl == KVM_MEMORY_ENCRYPT_REG_REGION; > > > + > > > + if (!kvm_arch_has_private_mem(kvm)) > > > + goto arch_vm_ioctl; > > > + > > > + r = -EFAULT; > > > + if (copy_from_user(®ion, argp, sizeof(region))) > > > + goto out; > > > + > > > + r = kvm_vm_ioctl_set_mem_attr(kvm, region.addr, > > > + region.size, set); > > > + break; > > > + } > > > +#endif > > > case KVM_GET_DIRTY_LOG: { > > > struct kvm_dirty_log log; > > > > > > @@ -4856,6 +4926,9 @@ static long kvm_vm_ioctl(struct file *filp, > > > r = kvm_vm_ioctl_get_stats_fd(kvm); > > > break; > > > default: > > > +#ifdef CONFIG_HAVE_KVM_PRIVATE_MEM > > > +arch_vm_ioctl: > > > +#endif > > > r = kvm_arch_vm_ioctl(filp, ioctl, arg); > > > } > > > out: > > > -- > > > 2.25.1 > > >