From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org X-Spam-Level: X-Spam-Status: No, score=-8.4 required=3.0 tests=DKIMWL_WL_HIGH,DKIM_SIGNED, DKIM_VALID,DKIM_VALID_AU,HEADER_FROM_DIFFERENT_DOMAINS,INCLUDES_PATCH, MAILING_LIST_MULTI,SIGNED_OFF_BY,SPF_HELO_NONE,SPF_PASS,URIBL_BLOCKED, USER_AGENT_SANE_1 autolearn=ham autolearn_force=no version=3.4.0 Received: from mail.kernel.org (mail.kernel.org [198.145.29.99]) by smtp.lore.kernel.org (Postfix) with ESMTP id 25039C0650F for ; Sat, 3 Aug 2019 00:13:46 +0000 (UTC) Received: from kanga.kvack.org (kanga.kvack.org [205.233.56.17]) by mail.kernel.org (Postfix) with ESMTP id AC0EF2086A for ; Sat, 3 Aug 2019 00:13:45 +0000 (UTC) Authentication-Results: mail.kernel.org; dkim=pass (2048-bit key) header.d=nvidia.com header.i=@nvidia.com header.b="Q3ngW/Ht" DMARC-Filter: OpenDMARC Filter v1.3.2 mail.kernel.org AC0EF2086A Authentication-Results: mail.kernel.org; dmarc=fail (p=none dis=none) header.from=nvidia.com Authentication-Results: mail.kernel.org; spf=pass smtp.mailfrom=owner-linux-mm@kvack.org Received: by kanga.kvack.org (Postfix) id 06DE36B0003; Fri, 2 Aug 2019 20:13:45 -0400 (EDT) Received: by kanga.kvack.org (Postfix, from userid 40) id 01E266B0005; Fri, 2 Aug 2019 20:13:44 -0400 (EDT) X-Delivered-To: int-list-linux-mm@kvack.org Received: by kanga.kvack.org (Postfix, from userid 63042) id E4F166B0006; Fri, 2 Aug 2019 20:13:44 -0400 (EDT) X-Delivered-To: linux-mm@kvack.org Received: from mail-pg1-f199.google.com (mail-pg1-f199.google.com [209.85.215.199]) by kanga.kvack.org (Postfix) with ESMTP id B11BB6B0003 for ; Fri, 2 Aug 2019 20:13:44 -0400 (EDT) Received: by mail-pg1-f199.google.com with SMTP id k20so48474025pgg.15 for ; Fri, 02 Aug 2019 17:13:44 -0700 (PDT) X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-gm-message-state:subject:to:cc:references:from:message-id:date :user-agent:mime-version:in-reply-to:content-language :content-transfer-encoding:dkim-signature; bh=xdJekrXrqSEkbcFhdjp7xR8+tNNoUXUqIgXLHmUnwRs=; b=JmPHhWZ/MX1FyTH7lrIJ4BKccH79U0UVrj9qtTSDddBaGdepeEAqmp3E/YpPTbrFtb xM0iZg8gDAT0B5oQjVE2Wz6CFe5xXBoydfmnhkR5laW+oQZVZKHTkOk7IP8jLxU0s3AY LMnGyjckURP0gURl5RYjNTexyeBp9LormGwNmjVvflOUZ73Kv/TLUFddAcb0HTz/lO6y clwGY+tfBz/vDFQuO74v9JZhX2hAebzH/8y9lV9K5Ut5KhFs1ua8psMQp/9ta3M6ig4w hKKHByalzYe4KaYOJ7rWtvWnKSZJjPjMCaYjF039GFwLLAu9C3UWHh0Zp7Iv1Plz6NQg EJnQ== X-Gm-Message-State: APjAAAWZpMmZkTmgaCSf06IglcWo18ajgYtagZfyJQES1La4Meig7kLU t4wHQV21MG0wefE20RvxrcbMzzjaWV3yrO1K7N3RFzWkBxCCbJSDaqq+elLS8mXEaxkliG59AEG CqZ0MY4Kb63PdXNktLTAYXSixS+cFRFpCW9aqDDUxBa2ein5RY6hy/JFERH1ny9HSkw== X-Received: by 2002:a63:790d:: with SMTP id u13mr14634238pgc.232.1564791224236; Fri, 02 Aug 2019 17:13:44 -0700 (PDT) X-Google-Smtp-Source: APXvYqz+6+i1xBffXIksobdhRd/isBBKOJBdRuQIFwZacHHi55gEdSP0mMe1x0yQKMjAxksXxOUp X-Received: by 2002:a63:790d:: with SMTP id u13mr14634197pgc.232.1564791223402; Fri, 02 Aug 2019 17:13:43 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1564791223; cv=none; d=google.com; s=arc-20160816; b=E4oyUFtA7ZmmFyGIgWeaDZapRvpt0dgTY4Wm16V4yyW3sSb0cI38YK4BOKCv/5jnNy LIR+f3gBy8HYnTcP+t3BcoGXjI2ZukPQgCiTQcRPJbHnuMy5dS1xgdjXfT4ti0ywVVLy ArkbTraVUIDc3LYHxRwcbdJ+8tqhViYXQcBU/dk39JPIIktk46TH6MiSVBpd4QHmnjDE hvHyV4dCAmKEMRtyqo/v0o2iB9ef1Bb0qCLK/azowmuA2whOGe9yOjPFRPHzlaV79LFB DEbwrRzUSyDpxG4y2prpd7oEuFoTHSHkCgDCCOaIfrW19eYLQioPjWt45X2bokg6Ou0g 7gxw== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=dkim-signature:content-transfer-encoding:content-language :in-reply-to:mime-version:user-agent:date:message-id:from:references :cc:to:subject; bh=xdJekrXrqSEkbcFhdjp7xR8+tNNoUXUqIgXLHmUnwRs=; b=C148QZ25UP+s4IGKNyCdQ2dIDgXwPEIXhdtpJmsl4kAyWT02V/JYUboSQTisDVRJoj UdewgvMWj4QSP8AxHd/MiF2KVfIjdk9VHO5enNQq/WxMhiht7wLRb9g4h7GfCSV1G5O0 iNqlt1d48tSRmjqbCj9dWXAmh1MGlhLoFCeGDRMJYR+/gpTY+2boq1Rgjg9+mBTamAJv /PE3QE4ILrQtrLy62OOhKO2jLzOmRRYOHACbZyZy7ClaYxOPg9ceYqmUUVtFx+ebRwZh mCU91prZTrcd3MsVFOEmzyitCck6QiWn4aS26HHbLlkMXMlvJjSxHO9KiRohnO1XOLaF TczA== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@nvidia.com header.s=n1 header.b="Q3ngW/Ht"; spf=pass (google.com: domain of jhubbard@nvidia.com designates 216.228.121.65 as permitted sender) smtp.mailfrom=jhubbard@nvidia.com; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=nvidia.com Received: from hqemgate16.nvidia.com (hqemgate16.nvidia.com. [216.228.121.65]) by mx.google.com with ESMTPS id r14si38120001pgm.406.2019.08.02.17.13.43 for (version=TLS1_2 cipher=ECDHE-RSA-AES128-GCM-SHA256 bits=128/128); Fri, 02 Aug 2019 17:13:43 -0700 (PDT) Received-SPF: pass (google.com: domain of jhubbard@nvidia.com designates 216.228.121.65 as permitted sender) client-ip=216.228.121.65; Authentication-Results: mx.google.com; dkim=pass header.i=@nvidia.com header.s=n1 header.b="Q3ngW/Ht"; spf=pass (google.com: domain of jhubbard@nvidia.com designates 216.228.121.65 as permitted sender) smtp.mailfrom=jhubbard@nvidia.com; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=nvidia.com Received: from hqpgpgate101.nvidia.com (Not Verified[216.228.121.13]) by hqemgate16.nvidia.com (using TLS: TLSv1.2, DES-CBC3-SHA) id ; Fri, 02 Aug 2019 17:13:43 -0700 Received: from hqmail.nvidia.com ([172.20.161.6]) by hqpgpgate101.nvidia.com (PGP Universal service); Fri, 02 Aug 2019 17:13:42 -0700 X-PGP-Universal: processed; by hqpgpgate101.nvidia.com on Fri, 02 Aug 2019 17:13:42 -0700 Received: from [10.110.48.28] (10.124.1.5) by HQMAIL107.nvidia.com (172.20.187.13) with Microsoft SMTP Server (TLS) id 15.0.1473.3; Sat, 3 Aug 2019 00:13:42 +0000 Subject: Re: [PATCH v4] staging: kpc2000: Convert put_page() to put_user_page*() To: Bharath Vedartham , , CC: Ira Weiny , =?UTF-8?B?SsOpcsO0bWUgR2xpc3Nl?= , , , References: <1564058658-3551-1-git-send-email-linux.bhar@gmail.com> X-Nvconfidentiality: public From: John Hubbard Message-ID: <4467d671-d011-0ebc-e2de-48a9745d4fe6@nvidia.com> Date: Fri, 2 Aug 2019 17:13:42 -0700 User-Agent: Mozilla/5.0 (X11; Linux x86_64; rv:60.0) Gecko/20100101 Thunderbird/60.8.0 MIME-Version: 1.0 In-Reply-To: <1564058658-3551-1-git-send-email-linux.bhar@gmail.com> X-Originating-IP: [10.124.1.5] X-ClientProxiedBy: HQMAIL104.nvidia.com (172.18.146.11) To HQMAIL107.nvidia.com (172.20.187.13) Content-Type: text/plain; charset="utf-8" Content-Language: en-US Content-Transfer-Encoding: quoted-printable DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=nvidia.com; s=n1; t=1564791223; bh=xdJekrXrqSEkbcFhdjp7xR8+tNNoUXUqIgXLHmUnwRs=; h=X-PGP-Universal:Subject:To:CC:References:X-Nvconfidentiality:From: Message-ID:Date:User-Agent:MIME-Version:In-Reply-To: X-Originating-IP:X-ClientProxiedBy:Content-Type:Content-Language: Content-Transfer-Encoding; b=Q3ngW/HtFs5cKcLOFRiDyp/sjKGrf3S/BIpPhC3sg/XAHIkhAPuOkBNsWlaIfZQm+ DyOw+GoC7Ln++nsmMD6ICDlKHFOoazO5hasrN5Se39NVf4JiW0c0sWVapHhkd1zqVJ dAgPBLhMLeWZhlQg/eS3kDbFwwREGPFjShjmIEQwOFYDWJumJqcHn31IVhlLzvXG/5 xMnItcy5iCsWOKLGXgW1x7XFgwhARbt77SFgA4QKuuOuvOspra98z+moBTUmkoKBUh xnNIF9Rn832yv7EMfep7LRQtgtB1yRzPgmhHUnfCkhoTZZSN59Pqt5f96lTYp21RKS ClwVCbnz8Yi8w== X-Bogosity: Ham, tests=bogofilter, spamicity=0.000000, version=1.2.4 Sender: owner-linux-mm@kvack.org Precedence: bulk X-Loop: owner-majordomo@kvack.org List-ID: On 7/25/19 5:44 AM, Bharath Vedartham wrote: > For pages that were retained via get_user_pages*(), release those pages > via the new put_user_page*() routines, instead of via put_page(). >=20 > This is part a tree-wide conversion, as described in commit fc1d8e7cca2d > ("mm: introduce put_user_page*(), placeholder versions"). >=20 Hi Bharath, If you like, I could re-post your patch here, modified slightly, as part of the next version of the miscellaneous call site conversion series [1]. As part of that, we should change this to use put_user_pages_dirty_lock()=20 (see below). > Cc: Ira Weiny > Cc: John Hubbard > Cc: J=C3=A9r=C3=B4me Glisse > Cc: Greg Kroah-Hartman > Cc: Matt Sickler > Cc: devel@driverdev.osuosl.org > Cc: linux-kernel@vger.kernel.org > Cc: linux-mm@kvack.org > Reviewed-by: John Hubbard > Signed-off-by: Bharath Vedartham > --- > Changes since v1 > - Improved changelog by John's suggestion. > - Moved logic to dirty pages below sg_dma_unmap > and removed PageReserved check. > Changes since v2 > - Added back PageResevered check as > suggested by John Hubbard. > Changes since v3 > - Changed the changelog as suggested by John. > - Added John's Reviewed-By tag. > Changes since v4 > - Rebased the patch on the staging tree. > - Improved commit log by fixing a line wrap. > --- > drivers/staging/kpc2000/kpc_dma/fileops.c | 17 ++++++----------- > 1 file changed, 6 insertions(+), 11 deletions(-) >=20 > diff --git a/drivers/staging/kpc2000/kpc_dma/fileops.c b/drivers/staging/= kpc2000/kpc_dma/fileops.c > index 48ca88b..f15e292 100644 > --- a/drivers/staging/kpc2000/kpc_dma/fileops.c > +++ b/drivers/staging/kpc2000/kpc_dma/fileops.c > @@ -190,9 +190,7 @@ static int kpc_dma_transfer(struct dev_private_data *= priv, > sg_free_table(&acd->sgt); > err_dma_map_sg: > err_alloc_sg_table: > - for (i =3D 0 ; i < acd->page_count ; i++) { > - put_page(acd->user_pages[i]); > - } > + put_user_pages(acd->user_pages, acd->page_count); > err_get_user_pages: > kfree(acd->user_pages); > err_alloc_userpages: > @@ -211,16 +209,13 @@ void transfer_complete_cb(struct aio_cb_data *acd,= size_t xfr_count, u32 flags) > BUG_ON(acd->ldev =3D=3D NULL); > BUG_ON(acd->ldev->pldev =3D=3D NULL); > =20 > - for (i =3D 0 ; i < acd->page_count ; i++) { > - if (!PageReserved(acd->user_pages[i])) { > - set_page_dirty(acd->user_pages[i]); > - } > - } > - > dma_unmap_sg(&acd->ldev->pldev->dev, acd->sgt.sgl, acd->sgt.nents, acd-= >ldev->dir); > =20 > - for (i =3D 0 ; i < acd->page_count ; i++) { > - put_page(acd->user_pages[i]); > + for (i =3D 0; i < acd->page_count; i++) { > + if (!PageReserved(acd->user_pages[i])) > + put_user_pages_dirty(&acd->user_pages[i], 1); This would change to: put_user_pages_dirty_lock(&acd->user_pages[i], 1, true); ...and we'd add this blurb (this time with CH's name spelled properly) to=20 the commit description: Note that this effectively changes the code's behavior in qp_release_pages(): it now ultimately calls set_page_dirty_lock(), instead of set_page_dirty(). This is probably more accurate. As Christoph Hellwig put it, "set_page_dirty() is only safe if we are dealing with a file backed page where we have reference on the inode it hangs off." [1] [1] https://lore.kernel.org/r/20190723153640.GB720@lst.de Also, future: I don't know the driver well enough to say, but maybe "true"= =20 could be replaced by "acd->ldev->dir =3D=3D DMA_FROM_DEVICE", there, but th= at would be a separate patch. thanks, --=20 John Hubbard NVIDIA > + else > + put_user_page(acd->user_pages[i]); > } > =20 > sg_free_table(&acd->sgt); >=20