From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org X-Spam-Level: X-Spam-Status: No, score=-16.1 required=3.0 tests=BAYES_00,DKIMWL_WL_HIGH, DKIM_SIGNED,DKIM_VALID,DKIM_VALID_AU,HEADER_FROM_DIFFERENT_DOMAINS, INCLUDES_CR_TRAILER,INCLUDES_PATCH,MAILING_LIST_MULTI,SPF_HELO_NONE,SPF_PASS, URIBL_BLOCKED autolearn=unavailable autolearn_force=no version=3.4.0 Received: from mail.kernel.org (mail.kernel.org [198.145.29.99]) by smtp.lore.kernel.org (Postfix) with ESMTP id 1E6B7C47089 for ; Thu, 27 May 2021 01:44:32 +0000 (UTC) Received: from kanga.kvack.org (kanga.kvack.org [205.233.56.17]) by mail.kernel.org (Postfix) with ESMTP id 9BC9B613D3 for ; Thu, 27 May 2021 01:44:31 +0000 (UTC) DMARC-Filter: OpenDMARC Filter v1.3.2 mail.kernel.org 9BC9B613D3 Authentication-Results: mail.kernel.org; dmarc=fail (p=none dis=none) header.from=redhat.com Authentication-Results: mail.kernel.org; spf=pass smtp.mailfrom=owner-linux-mm@kvack.org Received: by kanga.kvack.org (Postfix) id 4CBD46B0036; Wed, 26 May 2021 21:44:30 -0400 (EDT) Received: by kanga.kvack.org (Postfix, from userid 40) id 45B916B006E; Wed, 26 May 2021 21:44:30 -0400 (EDT) X-Delivered-To: int-list-linux-mm@kvack.org Received: by kanga.kvack.org (Postfix, from userid 63042) id E8B918D0002; Wed, 26 May 2021 21:44:29 -0400 (EDT) X-Delivered-To: linux-mm@kvack.org Received: from forelay.hostedemail.com (smtprelay0034.hostedemail.com [216.40.44.34]) by kanga.kvack.org (Postfix) with ESMTP id 9B4916B0036 for ; Wed, 26 May 2021 21:44:29 -0400 (EDT) Received: from smtpin33.hostedemail.com (10.5.19.251.rfc1918.com [10.5.19.251]) by forelay02.hostedemail.com (Postfix) with ESMTP id 2DD3F9424 for ; Thu, 27 May 2021 01:44:29 +0000 (UTC) X-FDA: 78185316258.33.83DC56A Received: from us-smtp-delivery-124.mimecast.com (us-smtp-delivery-124.mimecast.com [216.205.24.124]) by imf15.hostedemail.com (Postfix) with ESMTP id 3DE85A0001C8 for ; Thu, 27 May 2021 01:44:24 +0000 (UTC) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=redhat.com; s=mimecast20190719; t=1622079868; h=from:from:reply-to:subject:subject:date:date:message-id:message-id: to:to:cc:cc:mime-version:mime-version:content-type:content-type: in-reply-to:in-reply-to:references:references; bh=VpnQjRLTYtvem4bmBp1ngvgaZaWU0A6eMfZva3aBJoc=; b=aMTbqYws4wKiZOpRMbUFXFaWHDXoGv50dX09Q0l6OU34o/t+jlm4A+P2Lj4grICJQ1mpBO 7MhIP5XsUVtB1E0pCwutDiI/w1RV5bM0tABBSAT/q2bNgvPtPyxHNbo0G0KmYOdpOdJo/q tJm8MZawmAhVMIcc6tc2Cvafsgzgt8c= Received: from mail-qv1-f72.google.com (mail-qv1-f72.google.com [209.85.219.72]) (Using TLS) by relay.mimecast.com with ESMTP id us-mta-497-1TasySgcP3uCyy0eSxgcrA-1; Wed, 26 May 2021 21:44:26 -0400 X-MC-Unique: 1TasySgcP3uCyy0eSxgcrA-1 Received: by mail-qv1-f72.google.com with SMTP id f17-20020a0cf3d10000b02901eda24e6b92so2774152qvm.1 for ; Wed, 26 May 2021 18:44:25 -0700 (PDT) X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-gm-message-state:date:from:to:cc:subject:message-id:references :mime-version:content-disposition:in-reply-to; bh=VpnQjRLTYtvem4bmBp1ngvgaZaWU0A6eMfZva3aBJoc=; b=lSciv0f9d6NGWKBvlVeBtbF5YvHaMcuzUYNvQLs+7KXeUicezDCPxx3a2vF8zYNjYB 4X9ylboV2eEUcVfVySd9JAlTUVQ1gfrO9NtqHzS0ygDMWiu+4XC08BNpCxOqWlq7snDw JZ/mkB7f0dU6pQDGeenLZpmLTVz5j0g8CX02iP2YVHJzyoyUfifhlkn/KPEWBWnY8Bhm 2GLgf578zSMznGZMr/JXAKW7HX2YyIq3oLCPFYKfWF1ECnjBpAP5XFPN8Jec2j9bP7Tg sax9xjbcweaznna0EdyqQEfyY3tvGp8GIMl57FB76tTyEhAkIrjqhSnbNLKl038OtUJH srTA== X-Gm-Message-State: AOAM531WLW7aWJbopdGgZFtpjSXXD0ccCVDpuP9mqRQMkG4eH0x/LXNs igwtTvGoBG0TwlrObCORpJqCdld8K29SyChDt0Ql7Wm+HsXIwpvbR/oaJ6dETLz/him7CWknIAB S3QQEvVEfT6o= X-Received: by 2002:ac8:58c9:: with SMTP id u9mr1105868qta.58.1622079865059; Wed, 26 May 2021 18:44:25 -0700 (PDT) X-Google-Smtp-Source: ABdhPJzfR+dYiNTc8bzg8Cbe+4NPqyuhJDn00JrMrexnADTDq6kNzSW0QT/5/OWP1wZQyy97uuc2Sg== X-Received: by 2002:ac8:58c9:: with SMTP id u9mr1105834qta.58.1622079864805; Wed, 26 May 2021 18:44:24 -0700 (PDT) Received: from t490s (bras-base-toroon474qw-grc-72-184-145-4-219.dsl.bell.ca. [184.145.4.219]) by smtp.gmail.com with ESMTPSA id y1sm498232qkp.21.2021.05.26.18.44.22 (version=TLS1_3 cipher=TLS_AES_256_GCM_SHA384 bits=256/256); Wed, 26 May 2021 18:44:23 -0700 (PDT) Date: Wed, 26 May 2021 21:44:22 -0400 From: Peter Xu To: Alistair Popple Cc: linux-mm@kvack.org, akpm@linux-foundation.org, nouveau@lists.freedesktop.org, bskeggs@redhat.com, rcampbell@nvidia.com, linux-doc@vger.kernel.org, jhubbard@nvidia.com, bsingharora@gmail.com, linux-kernel@vger.kernel.org, dri-devel@lists.freedesktop.org, hch@infradead.org, jglisse@redhat.com, willy@infradead.org, jgg@nvidia.com, hughd@google.com Subject: Re: [PATCH v9 06/10] mm/memory.c: Allow different return codes for copy_nonpresent_pte() Message-ID: References: <20210524132725.12697-1-apopple@nvidia.com> <20210524132725.12697-7-apopple@nvidia.com> <2005328.bFqPmhE5MS@nvdebian> MIME-Version: 1.0 In-Reply-To: <2005328.bFqPmhE5MS@nvdebian> X-Mimecast-Spam-Score: 0 X-Mimecast-Originator: redhat.com Content-Type: text/plain; charset=utf-8 Content-Disposition: inline Authentication-Results: imf15.hostedemail.com; dkim=pass header.d=redhat.com header.s=mimecast20190719 header.b=aMTbqYws; dmarc=pass (policy=none) header.from=redhat.com; spf=none (imf15.hostedemail.com: domain of peterx@redhat.com has no SPF policy when checking 216.205.24.124) smtp.mailfrom=peterx@redhat.com X-Stat-Signature: wgrsb8aaku9rmkwhocf3kyze6njds71j X-Rspamd-Queue-Id: 3DE85A0001C8 X-Rspamd-Server: rspam02 X-HE-Tag: 1622079864-744639 X-Bogosity: Ham, tests=bogofilter, spamicity=0.000000, version=1.2.4 Sender: owner-linux-mm@kvack.org Precedence: bulk X-Loop: owner-majordomo@kvack.org List-ID: On Thu, May 27, 2021 at 11:20:36AM +1000, Alistair Popple wrote: > On Thursday, 27 May 2021 5:50:05 AM AEST Peter Xu wrote: > > On Mon, May 24, 2021 at 11:27:21PM +1000, Alistair Popple wrote: > > > Currently if copy_nonpresent_pte() returns a non-zero value it is > > > assumed to be a swap entry which requires further processing outside the > > > loop in copy_pte_range() after dropping locks. This prevents other > > > values being returned to signal conditions such as failure which a > > > subsequent change requires. > > > > > > Instead make copy_nonpresent_pte() return an error code if further > > > processing is required and read the value for the swap entry in the main > > > loop under the ptl. > > > > > > Signed-off-by: Alistair Popple > > > > > > --- > > > > > > v9: > > > > > > New for v9 to allow device exclusive handling to occur in > > > copy_nonpresent_pte(). > > > --- > > > > > > mm/memory.c | 12 +++++++----- > > > 1 file changed, 7 insertions(+), 5 deletions(-) > > > > > > diff --git a/mm/memory.c b/mm/memory.c > > > index 2fb455c365c2..e061cfa18c11 100644 > > > --- a/mm/memory.c > > > +++ b/mm/memory.c > > > @@ -718,7 +718,7 @@ copy_nonpresent_pte(struct mm_struct *dst_mm, struct > > > mm_struct *src_mm,> > > > if (likely(!non_swap_entry(entry))) { > > > > > > if (swap_duplicate(entry) < 0) > > > > > > - return entry.val; > > > + return -EAGAIN; > > > > > > /* make sure dst_mm is on swapoff's mmlist. */ > > > if (unlikely(list_empty(&dst_mm->mmlist))) { > > > > > > @@ -974,11 +974,13 @@ copy_pte_range(struct vm_area_struct *dst_vma, > > > struct vm_area_struct *src_vma,> > > > continue; > > > > > > } > > > if (unlikely(!pte_present(*src_pte))) { > > > > > > - entry.val = copy_nonpresent_pte(dst_mm, src_mm, > > > - dst_pte, src_pte, > > > - src_vma, addr, rss); > > > - if (entry.val) > > > + ret = copy_nonpresent_pte(dst_mm, src_mm, > > > + dst_pte, src_pte, > > > + src_vma, addr, rss); > > > + if (ret == -EAGAIN) { > > > + entry = pte_to_swp_entry(*src_pte); > > > > > > break; > > > > > > + } > > > > > > progress += 8; > > > continue; > > > > > > } > > > > Note that -EAGAIN was previously used by copy_present_page() for early cow > > use. Here later although we check entry.val first: > > > > if (entry.val) { > > if (add_swap_count_continuation(entry, GFP_KERNEL) < 0) { > > ret = -ENOMEM; > > goto out; > > } > > entry.val = 0; > > } else if (ret) { > > WARN_ON_ONCE(ret != -EAGAIN); > > prealloc = page_copy_prealloc(src_mm, src_vma, addr); > > if (!prealloc) > > return -ENOMEM; > > /* We've captured and resolved the error. Reset, try again. > > */ ret = 0; > > } > > > > We didn't reset "ret" in entry.val case (maybe we should?). Then in the next > > round of "goto again" if "ret" is unluckily untouched, it could reach the > > 2nd if check, and I think it could cause an unexpected > > page_copy_prealloc(). > > Thanks, I had considered that but saw "ret" was always set either by > copy_nonpresent_pte() or copy_present_pte(). However missed the "unlucky" case > at the start of the loop: > > if (progress >= 32) { > progress = 0; > if (need_resched() || > spin_needbreak(src_ptl) || pin_needbreak(dst_ptl)) > break; > > Looking at this again though checking different variables to figure out what > to do outside the locks and reusing error codes seems error prone. I reused - > EAGAIN for copy_nonpresent_pte() simply because that seemed the most sensible > error code, but I don't think that aids readability and it might be better to > use a unique error code for each case needing extra handling. > > So it might be better if I update this patch to: > 1) Use unique error codes for each case requiring special handling outside the > lock. > 2) Only check "ret" to determine what to do outside locks (ie. not entry.val) > 3) Document these. > 4) Always reset ret after handling. > > Thoughts? Looks good to me. Thanks, -- Peter Xu