From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org X-Spam-Level: X-Spam-Status: No, score=-8.5 required=3.0 tests=HEADER_FROM_DIFFERENT_DOMAINS, INCLUDES_PATCH,MAILING_LIST_MULTI,SIGNED_OFF_BY,SPF_HELO_NONE,SPF_PASS, URIBL_BLOCKED,USER_AGENT_MUTT autolearn=unavailable autolearn_force=no version=3.4.0 Received: from mail.kernel.org (mail.kernel.org [198.145.29.99]) by smtp.lore.kernel.org (Postfix) with ESMTP id 2C696C282E1 for ; Thu, 23 May 2019 17:28:03 +0000 (UTC) Received: from kanga.kvack.org (kanga.kvack.org [205.233.56.17]) by mail.kernel.org (Postfix) with ESMTP id B7F812175B for ; Thu, 23 May 2019 17:28:02 +0000 (UTC) DMARC-Filter: OpenDMARC Filter v1.3.2 mail.kernel.org B7F812175B Authentication-Results: mail.kernel.org; dmarc=fail (p=none dis=none) header.from=intel.com Authentication-Results: mail.kernel.org; spf=pass smtp.mailfrom=owner-linux-mm@kvack.org Received: by kanga.kvack.org (Postfix) id EBD1A6B0282; Thu, 23 May 2019 13:28:01 -0400 (EDT) Received: by kanga.kvack.org (Postfix, from userid 40) id E6EA86B028C; Thu, 23 May 2019 13:28:01 -0400 (EDT) X-Delivered-To: int-list-linux-mm@kvack.org Received: by kanga.kvack.org (Postfix, from userid 63042) id D5E116B028D; Thu, 23 May 2019 13:28:01 -0400 (EDT) X-Delivered-To: linux-mm@kvack.org Received: from mail-pl1-f197.google.com (mail-pl1-f197.google.com [209.85.214.197]) by kanga.kvack.org (Postfix) with ESMTP id 9D64B6B0282 for ; Thu, 23 May 2019 13:28:01 -0400 (EDT) Received: by mail-pl1-f197.google.com with SMTP id b69so3921527plb.9 for ; Thu, 23 May 2019 10:28:01 -0700 (PDT) X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-original-authentication-results:x-gm-message-state:date:from:to :cc:subject:message-id:references:mime-version:content-disposition :in-reply-to:user-agent; bh=VWjOv4IdjZ02Cfj3+ytDIskaKlOlSBxKlFiPSIRxOoc=; b=tiG7ey4VxBBwBzlkaTHBjrpirtHHEd+MLoXKO/Eg4ncEDZXAWWwxpKFwQTgO63fFQn K7HhsXae1Uf2PZEDflC1mK/IOOwuVjWY+GG0+ULXtuy3GGcpAVU0dQl2WmrjxIQhod8u T8QX0No42sqHWq7Oaj1bsj+HnOgzQH3Fd074u0hFA+OAtl3vMr8WckIWsq1tDNYt1Pw7 BchHTfLGB/wHZd+Sf0GUO8hQXOLt52saW7y8PolnDF/zxbNMEGcEb4iHzAO8UyFI0r0/ geRDvGsW8V3ovPceSj9tOT3JTNvafu2FRd+qw0aNskF/rc3IGC9JOHV/AiYqP6Ux8LgQ P12Q== X-Original-Authentication-Results: mx.google.com; spf=pass (google.com: domain of ira.weiny@intel.com designates 192.55.52.88 as permitted sender) smtp.mailfrom=ira.weiny@intel.com; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=intel.com X-Gm-Message-State: APjAAAXZ0ywQl3CgFwFxRKJPpgkNMx/0PS1bt/5n7BylR2BMiIOc4Ey5 B/LeE7rJ4Fgl4M8pmEy9fjPrup72orqTuwetBkNiySQIC9yg1RC1vOb3MnW2j2KSWcQ7EDaqjtn +cflhsVLaHJ7o59tF/kgrqCHzHzc2AXbGpI4VkX06p2VoEojVjN+PGVKEDTmzYQKy6g== X-Received: by 2002:a17:90a:8e86:: with SMTP id f6mr2834859pjo.66.1558632481258; Thu, 23 May 2019 10:28:01 -0700 (PDT) X-Google-Smtp-Source: APXvYqwjAp61gpk6fZ5slcdrZFU+Knnoaa86qEfPmS/hQ1Md7kT9TqU2bpuhx2bx9ubJ9HElzB4g X-Received: by 2002:a17:90a:8e86:: with SMTP id f6mr2834752pjo.66.1558632480112; Thu, 23 May 2019 10:28:00 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1558632480; cv=none; d=google.com; s=arc-20160816; b=UwyU0cQnsHwCgtAyhOwpTrKUbiZ+Xn0hVsJGCR07z8hpbjzhnVtxQS+A7RLA5hahU8 R2SRfTBw7Oool3E5ckOzrMYxcmTavT/jQ1lgsIsFuPLP6c/AnURUEwfMTwg9HwQhtzlS 43GqWW04AJlDtSWqfZ21HhzJJ6bcFvX8haPgC1he977tU/dP/u0wWdwqlbZ1SvH5Ptt7 h6cqJvE3HOLaL3P5DMWEJZFiWxvNNu8sBqnyJaiZhC/BQONTX8MVq5C0sUEStqQ8XWTw PskyK4JnjaNNWsDW2blxPNMwuCkktCnQOoqqV4nMjSIzqQFsFZMnYdyF+UShiGQCDNcX dFKw== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=user-agent:in-reply-to:content-disposition:mime-version:references :message-id:subject:cc:to:from:date; bh=VWjOv4IdjZ02Cfj3+ytDIskaKlOlSBxKlFiPSIRxOoc=; b=Fhejp8SC+Ugjw/zEAWTXqbxQ0407YZtY/ZiUN7MqHaxU1VibW967RzKPr8MlMtRUCS OJX2i3Vs+8BnUY7v8zDxOgIldKTtoq7rq5UBOK1R9XNZzv1KlzLGHRzqqJL9uBTluZsd MFp/awO+wYLDVqq1gTFwrrl4A4QvSb+dzrYvAGCMpeVAGXIHjHuOQlGwIWAwFcBfCi3+ DazsEaqJkeFQ84wldItCpGlXx2bAkLil/eroPZ7hjfnsoxE99GYxt46U0SBcmvcwMchu iPdpYXwtfz10XvPZkSeSV+qNBw+0jHPIdKkyhb4l7/oSJRD+m6QS/QQfnPFBTUsC5sn7 xxnw== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: domain of ira.weiny@intel.com designates 192.55.52.88 as permitted sender) smtp.mailfrom=ira.weiny@intel.com; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=intel.com Received: from mga01.intel.com (mga01.intel.com. [192.55.52.88]) by mx.google.com with ESMTPS id k7si150062pgm.302.2019.05.23.10.27.59 for (version=TLS1_2 cipher=ECDHE-RSA-AES128-GCM-SHA256 bits=128/128); Thu, 23 May 2019 10:28:00 -0700 (PDT) Received-SPF: pass (google.com: domain of ira.weiny@intel.com designates 192.55.52.88 as permitted sender) client-ip=192.55.52.88; Authentication-Results: mx.google.com; spf=pass (google.com: domain of ira.weiny@intel.com designates 192.55.52.88 as permitted sender) smtp.mailfrom=ira.weiny@intel.com; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=intel.com X-Amp-Result: UNKNOWN X-Amp-Original-Verdict: FILE UNKNOWN X-Amp-File-Uploaded: False Received: from fmsmga007.fm.intel.com ([10.253.24.52]) by fmsmga101.fm.intel.com with ESMTP/TLS/DHE-RSA-AES256-GCM-SHA384; 23 May 2019 10:27:59 -0700 X-ExtLoop1: 1 Received: from iweiny-desk2.sc.intel.com ([10.3.52.157]) by fmsmga007.fm.intel.com with ESMTP; 23 May 2019 10:27:59 -0700 Date: Thu, 23 May 2019 10:28:52 -0700 From: Ira Weiny To: john.hubbard@gmail.com Cc: Andrew Morton , linux-mm@kvack.org, Jason Gunthorpe , LKML , linux-rdma@vger.kernel.org, linux-fsdevel@vger.kernel.org, John Hubbard , Doug Ledford , Mike Marciniszyn , Dennis Dalessandro , Christian Benvenuti , Jan Kara , Jason Gunthorpe Subject: Re: [PATCH 1/1] infiniband/mm: convert put_page() to put_user_page*() Message-ID: <20190523172852.GA27175@iweiny-DESK2.sc.intel.com> References: <20190523072537.31940-1-jhubbard@nvidia.com> <20190523072537.31940-2-jhubbard@nvidia.com> MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline In-Reply-To: <20190523072537.31940-2-jhubbard@nvidia.com> User-Agent: Mutt/1.11.1 (2018-12-01) X-Bogosity: Ham, tests=bogofilter, spamicity=0.000000, version=1.2.4 Sender: owner-linux-mm@kvack.org Precedence: bulk X-Loop: owner-majordomo@kvack.org List-ID: On Thu, May 23, 2019 at 12:25:37AM -0700, john.hubbard@gmail.com wrote: > From: John Hubbard > > For infiniband code that retains pages via get_user_pages*(), > release those pages via the new put_user_page(), or > put_user_pages*(), instead of put_page() > > This is a tiny part of the second step of fixing the problem described > in [1]. The steps are: > > 1) Provide put_user_page*() routines, intended to be used > for releasing pages that were pinned via get_user_pages*(). > > 2) Convert all of the call sites for get_user_pages*(), to > invoke put_user_page*(), instead of put_page(). This involves dozens of > call sites, and will take some time. > > 3) After (2) is complete, use get_user_pages*() and put_user_page*() to > implement tracking of these pages. This tracking will be separate from > the existing struct page refcounting. > > 4) Use the tracking and identification of these pages, to implement > special handling (especially in writeback paths) when the pages are > backed by a filesystem. Again, [1] provides details as to why that is > desirable. > > [1] https://lwn.net/Articles/753027/ : "The Trouble with get_user_pages()" > > Cc: Doug Ledford > Cc: Jason Gunthorpe > Cc: Mike Marciniszyn > Cc: Dennis Dalessandro > Cc: Christian Benvenuti > > Reviewed-by: Jan Kara > Reviewed-by: Dennis Dalessandro > Acked-by: Jason Gunthorpe > Tested-by: Ira Weiny > Signed-off-by: John Hubbard > --- > drivers/infiniband/core/umem.c | 7 ++++--- > drivers/infiniband/core/umem_odp.c | 10 +++++----- > drivers/infiniband/hw/hfi1/user_pages.c | 11 ++++------- > drivers/infiniband/hw/mthca/mthca_memfree.c | 6 +++--- > drivers/infiniband/hw/qib/qib_user_pages.c | 11 ++++------- > drivers/infiniband/hw/qib/qib_user_sdma.c | 6 +++--- > drivers/infiniband/hw/usnic/usnic_uiom.c | 7 ++++--- > 7 files changed, 27 insertions(+), 31 deletions(-) > > diff --git a/drivers/infiniband/core/umem.c b/drivers/infiniband/core/umem.c > index e7ea819fcb11..673f0d240b3e 100644 > --- a/drivers/infiniband/core/umem.c > +++ b/drivers/infiniband/core/umem.c > @@ -54,9 +54,10 @@ static void __ib_umem_release(struct ib_device *dev, struct ib_umem *umem, int d > > for_each_sg_page(umem->sg_head.sgl, &sg_iter, umem->sg_nents, 0) { > page = sg_page_iter_page(&sg_iter); > - if (!PageDirty(page) && umem->writable && dirty) > - set_page_dirty_lock(page); > - put_page(page); > + if (umem->writable && dirty) > + put_user_pages_dirty_lock(&page, 1); > + else > + put_user_page(page); > } > > sg_free_table(&umem->sg_head); > diff --git a/drivers/infiniband/core/umem_odp.c b/drivers/infiniband/core/umem_odp.c > index f962b5bbfa40..17e46df3990a 100644 > --- a/drivers/infiniband/core/umem_odp.c > +++ b/drivers/infiniband/core/umem_odp.c > @@ -487,7 +487,7 @@ void ib_umem_odp_release(struct ib_umem_odp *umem_odp) > * The function returns -EFAULT if the DMA mapping operation fails. It returns > * -EAGAIN if a concurrent invalidation prevents us from updating the page. > * > - * The page is released via put_page even if the operation failed. For > + * The page is released via put_user_page even if the operation failed. For > * on-demand pinning, the page is released whenever it isn't stored in the > * umem. > */ > @@ -536,7 +536,7 @@ static int ib_umem_odp_map_dma_single_page( > } > > out: > - put_page(page); > + put_user_page(page); > > if (remove_existing_mapping) { > ib_umem_notifier_start_account(umem_odp); > @@ -659,7 +659,7 @@ int ib_umem_odp_map_dma_pages(struct ib_umem_odp *umem_odp, u64 user_virt, > ret = -EFAULT; > break; > } > - put_page(local_page_list[j]); > + put_user_page(local_page_list[j]); > continue; > } > > @@ -686,8 +686,8 @@ int ib_umem_odp_map_dma_pages(struct ib_umem_odp *umem_odp, u64 user_virt, > * ib_umem_odp_map_dma_single_page(). > */ > if (npages - (j + 1) > 0) > - release_pages(&local_page_list[j+1], > - npages - (j + 1)); > + put_user_pages(&local_page_list[j+1], > + npages - (j + 1)); I don't know if we discussed this before but it looks like the use of release_pages() was not entirely correct (or at least not necessary) here. So I think this is ok. As for testing, I have been running with this patch for a while but I don't have ODP hardware so that testing would not cover this code path. So you can add my: Reviewed-by: Ira Weiny > break; > } > } > diff --git a/drivers/infiniband/hw/hfi1/user_pages.c b/drivers/infiniband/hw/hfi1/user_pages.c > index 02eee8eff1db..b89a9b9aef7a 100644 > --- a/drivers/infiniband/hw/hfi1/user_pages.c > +++ b/drivers/infiniband/hw/hfi1/user_pages.c > @@ -118,13 +118,10 @@ int hfi1_acquire_user_pages(struct mm_struct *mm, unsigned long vaddr, size_t np > void hfi1_release_user_pages(struct mm_struct *mm, struct page **p, > size_t npages, bool dirty) > { > - size_t i; > - > - for (i = 0; i < npages; i++) { > - if (dirty) > - set_page_dirty_lock(p[i]); > - put_page(p[i]); > - } > + if (dirty) > + put_user_pages_dirty_lock(p, npages); > + else > + put_user_pages(p, npages); > > if (mm) { /* during close after signal, mm can be NULL */ > atomic64_sub(npages, &mm->pinned_vm); > diff --git a/drivers/infiniband/hw/mthca/mthca_memfree.c b/drivers/infiniband/hw/mthca/mthca_memfree.c > index 8ff0e90d7564..edccfd6e178f 100644 > --- a/drivers/infiniband/hw/mthca/mthca_memfree.c > +++ b/drivers/infiniband/hw/mthca/mthca_memfree.c > @@ -482,7 +482,7 @@ int mthca_map_user_db(struct mthca_dev *dev, struct mthca_uar *uar, > > ret = pci_map_sg(dev->pdev, &db_tab->page[i].mem, 1, PCI_DMA_TODEVICE); > if (ret < 0) { > - put_page(pages[0]); > + put_user_page(pages[0]); > goto out; > } > > @@ -490,7 +490,7 @@ int mthca_map_user_db(struct mthca_dev *dev, struct mthca_uar *uar, > mthca_uarc_virt(dev, uar, i)); > if (ret) { > pci_unmap_sg(dev->pdev, &db_tab->page[i].mem, 1, PCI_DMA_TODEVICE); > - put_page(sg_page(&db_tab->page[i].mem)); > + put_user_page(sg_page(&db_tab->page[i].mem)); > goto out; > } > > @@ -556,7 +556,7 @@ void mthca_cleanup_user_db_tab(struct mthca_dev *dev, struct mthca_uar *uar, > if (db_tab->page[i].uvirt) { > mthca_UNMAP_ICM(dev, mthca_uarc_virt(dev, uar, i), 1); > pci_unmap_sg(dev->pdev, &db_tab->page[i].mem, 1, PCI_DMA_TODEVICE); > - put_page(sg_page(&db_tab->page[i].mem)); > + put_user_page(sg_page(&db_tab->page[i].mem)); > } > } > > diff --git a/drivers/infiniband/hw/qib/qib_user_pages.c b/drivers/infiniband/hw/qib/qib_user_pages.c > index f712fb7fa82f..bfbfbb7e0ff4 100644 > --- a/drivers/infiniband/hw/qib/qib_user_pages.c > +++ b/drivers/infiniband/hw/qib/qib_user_pages.c > @@ -40,13 +40,10 @@ > static void __qib_release_user_pages(struct page **p, size_t num_pages, > int dirty) > { > - size_t i; > - > - for (i = 0; i < num_pages; i++) { > - if (dirty) > - set_page_dirty_lock(p[i]); > - put_page(p[i]); > - } > + if (dirty) > + put_user_pages_dirty_lock(p, num_pages); > + else > + put_user_pages(p, num_pages); > } > > /** > diff --git a/drivers/infiniband/hw/qib/qib_user_sdma.c b/drivers/infiniband/hw/qib/qib_user_sdma.c > index 0c204776263f..ac5bdb02144f 100644 > --- a/drivers/infiniband/hw/qib/qib_user_sdma.c > +++ b/drivers/infiniband/hw/qib/qib_user_sdma.c > @@ -317,7 +317,7 @@ static int qib_user_sdma_page_to_frags(const struct qib_devdata *dd, > * the caller can ignore this page. > */ > if (put) { > - put_page(page); > + put_user_page(page); > } else { > /* coalesce case */ > kunmap(page); > @@ -631,7 +631,7 @@ static void qib_user_sdma_free_pkt_frag(struct device *dev, > kunmap(pkt->addr[i].page); > > if (pkt->addr[i].put_page) > - put_page(pkt->addr[i].page); > + put_user_page(pkt->addr[i].page); > else > __free_page(pkt->addr[i].page); > } else if (pkt->addr[i].kvaddr) { > @@ -706,7 +706,7 @@ static int qib_user_sdma_pin_pages(const struct qib_devdata *dd, > /* if error, return all pages not managed by pkt */ > free_pages: > while (i < j) > - put_page(pages[i++]); > + put_user_page(pages[i++]); > > done: > return ret; > diff --git a/drivers/infiniband/hw/usnic/usnic_uiom.c b/drivers/infiniband/hw/usnic/usnic_uiom.c > index e312f522a66d..0b0237d41613 100644 > --- a/drivers/infiniband/hw/usnic/usnic_uiom.c > +++ b/drivers/infiniband/hw/usnic/usnic_uiom.c > @@ -75,9 +75,10 @@ static void usnic_uiom_put_pages(struct list_head *chunk_list, int dirty) > for_each_sg(chunk->page_list, sg, chunk->nents, i) { > page = sg_page(sg); > pa = sg_phys(sg); > - if (!PageDirty(page) && dirty) > - set_page_dirty_lock(page); > - put_page(page); > + if (dirty) > + put_user_pages_dirty_lock(&page, 1); > + else > + put_user_page(page); > usnic_dbg("pa: %pa\n", &pa); > } > kfree(chunk); > -- > 2.21.0 >