From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org X-Spam-Level: X-Spam-Status: No, score=-11.4 required=3.0 tests=BAYES_00,DKIMWL_WL_MED, DKIM_SIGNED,DKIM_VALID,DKIM_VALID_AU,HEADER_FROM_DIFFERENT_DOMAINS, MAILING_LIST_MULTI,SPF_HELO_NONE,SPF_PASS,URIBL_BLOCKED,USER_IN_DEF_DKIM_WL autolearn=no autolearn_force=no version=3.4.0 Received: from mail.kernel.org (mail.kernel.org [198.145.29.99]) by smtp.lore.kernel.org (Postfix) with ESMTP id BDE58C4727F for ; Fri, 2 Oct 2020 06:40:07 +0000 (UTC) Received: from kanga.kvack.org (kanga.kvack.org [205.233.56.17]) by mail.kernel.org (Postfix) with ESMTP id 3D48C206FA for ; Fri, 2 Oct 2020 06:40:07 +0000 (UTC) Authentication-Results: mail.kernel.org; dkim=pass (2048-bit key) header.d=google.com header.i=@google.com header.b="iEkjLRgN" DMARC-Filter: OpenDMARC Filter v1.3.2 mail.kernel.org 3D48C206FA Authentication-Results: mail.kernel.org; dmarc=fail (p=reject dis=none) header.from=google.com Authentication-Results: mail.kernel.org; spf=pass smtp.mailfrom=owner-linux-mm@kvack.org Received: by kanga.kvack.org (Postfix) id 58CC56B0068; Fri, 2 Oct 2020 02:40:06 -0400 (EDT) Received: by kanga.kvack.org (Postfix, from userid 40) id 53D946B006C; Fri, 2 Oct 2020 02:40:06 -0400 (EDT) X-Delivered-To: int-list-linux-mm@kvack.org Received: by kanga.kvack.org (Postfix, from userid 63042) id 404A86B006E; Fri, 2 Oct 2020 02:40:06 -0400 (EDT) X-Delivered-To: linux-mm@kvack.org Received: from forelay.hostedemail.com (smtprelay0114.hostedemail.com [216.40.44.114]) by kanga.kvack.org (Postfix) with ESMTP id 144976B0068 for ; Fri, 2 Oct 2020 02:40:06 -0400 (EDT) Received: from smtpin25.hostedemail.com (10.5.19.251.rfc1918.com [10.5.19.251]) by forelay01.hostedemail.com (Postfix) with ESMTP id 9B771180AD801 for ; Fri, 2 Oct 2020 06:40:05 +0000 (UTC) X-FDA: 77326035570.25.shoe93_540dfcd271a2 Received: from filter.hostedemail.com (10.5.16.251.rfc1918.com [10.5.16.251]) by smtpin25.hostedemail.com (Postfix) with ESMTP id 69BCD1804E3A1 for ; Fri, 2 Oct 2020 06:40:05 +0000 (UTC) X-HE-Tag: shoe93_540dfcd271a2 X-Filterd-Recvd-Size: 9687 Received: from mail-io1-f68.google.com (mail-io1-f68.google.com [209.85.166.68]) by imf42.hostedemail.com (Postfix) with ESMTP for ; Fri, 2 Oct 2020 06:40:04 +0000 (UTC) Received: by mail-io1-f68.google.com with SMTP id z13so444705iom.8 for ; Thu, 01 Oct 2020 23:40:04 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=20161025; h=mime-version:references:in-reply-to:from:date:message-id:subject:to :cc; bh=dkXqqJWGGMD3nsu1nDQF0uD51+pRQERPlmfOf0TOTVo=; b=iEkjLRgNkKYJWF7CEttuAu/tYQ0VOufE98KsN+GWMxJUGzWnlC8BFuPq5RorTIVtcD GUz8TA+zytLWsfeghAYOXm16ewnEmWroldRiJLLUqkIpye0v5gLc92/DzuUZ3RjoC0o2 Mj+49Kak5eFgPvzj1HZE10GW3MIpMKwWStTYAIBc5xjjTuGryaebrLfyq7fyZMgbCj47 vGPoiwmHKo51+C3S2BZQ8GMGQnX0K9vkbvsbvttctT9yFjFAGrDlrgwhFItqqdUsZcp5 OTWfLuNUXiQ3QREEwcpptRzqqqoO27EicyYNmoZAuoCjFFGOVcH49Su7CdhioMsK8LU0 824A== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-gm-message-state:mime-version:references:in-reply-to:from:date :message-id:subject:to:cc; bh=dkXqqJWGGMD3nsu1nDQF0uD51+pRQERPlmfOf0TOTVo=; b=q68kMW2DnLD18H4TYtz/gS30712UTlant03WI2cmxcg4eMcYgemaKjfOUBiTS2QgZK pIQvMWOogyXRaCsfmLbkvOJZousxE64ixIpB9HQm7dL9QKcjgz67d+Y1fSpEVHUDufqq NVfzXVtQVzBGiswEJCOBt2HsAkpvPa4dEPI350s3UUIfNhGeJ0znVrS/R0udbNU03Qjw x7Wpae62XKnUIkUnwLUB5OmcH/t3aRBXNnxTJ1iEeLgbBAP2Ft2grmVfdqf32IVYOTjf y4LUGWf+eMlzyk6lDAEc2c39Ejyr92/lp+7OgRnh2d7fwDRnAYKLQRwGsEddWvPELXa0 kIsg== X-Gm-Message-State: AOAM532xiyNd0o5wVa8JNPgH7tCv3vDSF6iwb2wIJGQN7Yux0MDlQwDV vWjUiugUf2buMyyAxtaBREi0rFEBimBD0rEA29jbSw== X-Google-Smtp-Source: ABdhPJyqoA6lgzrhsEk2OHis7Bv3fG52DoLRoDsNnGgjUwRYYdWQzQoYWkvJKagd5Spw6QnrhKVAa3MxJgaS9zAd4r8= X-Received: by 2002:a6b:590c:: with SMTP id n12mr935815iob.25.1601620804015; Thu, 01 Oct 2020 23:40:04 -0700 (PDT) MIME-Version: 1.0 References: <20200930222130.4175584-1-kaleshsingh@google.com> <20200930223207.5xepuvu6wr6xw5bb@black.fi.intel.com> <20201001122706.jp2zr23a43hfomyg@black.fi.intel.com> <20201002053547.7roe7b4mpamw4uk2@black.fi.intel.com> In-Reply-To: <20201002053547.7roe7b4mpamw4uk2@black.fi.intel.com> From: Lokesh Gidra Date: Thu, 1 Oct 2020 23:39:53 -0700 Message-ID: Subject: Re: [PATCH 0/5] Speed up mremap on large regions To: "Kirill A. Shutemov" Cc: Kalesh Singh , Suren Baghdasaryan , Minchan Kim , Joel Fernandes , "Cc: Android Kernel" , Catalin Marinas , Will Deacon , Thomas Gleixner , Ingo Molnar , Borislav Petkov , "the arch/x86 maintainers" , "H. Peter Anvin" , Andrew Morton , Shuah Khan , "Aneesh Kumar K.V" , Kees Cook , Peter Zijlstra , Sami Tolvanen , Masahiro Yamada , Arnd Bergmann , Frederic Weisbecker , Krzysztof Kozlowski , Hassan Naveed , Christian Brauner , Mark Rutland , Mike Rapoport , Gavin Shan , Zhenyu Ye , Jia He , John Hubbard , William Kucharski , Sandipan Das , Ralph Campbell , Mina Almasry , Ram Pai , Dave Hansen , Kamalesh Babulal , Masami Hiramatsu , Brian Geffon , SeongJae Park , linux-kernel , "moderated list:ARM64 PORT (AARCH64 ARCHITECTURE)" , "open list:MEMORY MANAGEMENT" , "open list:KERNEL SELFTEST FRAMEWORK" Content-Type: text/plain; charset="UTF-8" X-Bogosity: Ham, tests=bogofilter, spamicity=0.000000, version=1.2.4 Sender: owner-linux-mm@kvack.org Precedence: bulk X-Loop: owner-majordomo@kvack.org List-ID: On Thu, Oct 1, 2020 at 10:36 PM Kirill A. Shutemov wrote: > > On Thu, Oct 01, 2020 at 05:09:02PM -0700, Lokesh Gidra wrote: > > On Thu, Oct 1, 2020 at 9:00 AM Kalesh Singh wrote: > > > > > > On Thu, Oct 1, 2020 at 8:27 AM Kirill A. Shutemov > > > wrote: > > > > > > > > On Wed, Sep 30, 2020 at 03:42:17PM -0700, Lokesh Gidra wrote: > > > > > On Wed, Sep 30, 2020 at 3:32 PM Kirill A. Shutemov > > > > > wrote: > > > > > > > > > > > > On Wed, Sep 30, 2020 at 10:21:17PM +0000, Kalesh Singh wrote: > > > > > > > mremap time can be optimized by moving entries at the PMD/PUD level if > > > > > > > the source and destination addresses are PMD/PUD-aligned and > > > > > > > PMD/PUD-sized. Enable moving at the PMD and PUD levels on arm64 and > > > > > > > x86. Other architectures where this type of move is supported and known to > > > > > > > be safe can also opt-in to these optimizations by enabling HAVE_MOVE_PMD > > > > > > > and HAVE_MOVE_PUD. > > > > > > > > > > > > > > Observed Performance Improvements for remapping a PUD-aligned 1GB-sized > > > > > > > region on x86 and arm64: > > > > > > > > > > > > > > - HAVE_MOVE_PMD is already enabled on x86 : N/A > > > > > > > - Enabling HAVE_MOVE_PUD on x86 : ~13x speed up > > > > > > > > > > > > > > - Enabling HAVE_MOVE_PMD on arm64 : ~ 8x speed up > > > > > > > - Enabling HAVE_MOVE_PUD on arm64 : ~19x speed up > > > > > > > > > > > > > > Altogether, HAVE_MOVE_PMD and HAVE_MOVE_PUD > > > > > > > give a total of ~150x speed up on arm64. > > > > > > > > > > > > Is there a *real* workload that benefit from HAVE_MOVE_PUD? > > > > > > > > > > > We have a Java garbage collector under development which requires > > > > > moving physical pages of multi-gigabyte heap using mremap. During this > > > > > move, the application threads have to be paused for correctness. It is > > > > > critical to keep this pause as short as possible to avoid jitters > > > > > during user interaction. This is where HAVE_MOVE_PUD will greatly > > > > > help. > > > > > > > > Any chance to quantify the effect of mremap() with and without > > > > HAVE_MOVE_PUD? > > > > > > > > I doubt it's a major contributor to the GC pause. I expect you need to > > > > move tens of gigs to get sizable effect. And if your GC routinely moves > > > > tens of gigs, maybe problem somewhere else? > > > > > > > > I'm asking for numbers, because increase in complexity comes with cost. > > > > If it doesn't provide an substantial benefit to a real workload > > > > maintaining the code forever doesn't make sense. > > > > > mremap is indeed the biggest contributor to the GC pause. It has to > > take place in what is typically known as a 'stop-the-world' pause, > > wherein all application threads are paused. During this pause the GC > > thread flips the GC roots (threads' stacks, globals etc.), and then > > resumes threads along with concurrent compaction of the heap.This > > GC-root flip differs depending on which compaction algorithm is being > > used. > > > > In our case it involves updating object references in threads' stacks > > and remapping java heap to a different location. The threads' stacks > > can be handled in parallel with the mremap. Therefore, the dominant > > factor is indeed the cost of mremap. From patches 2 and 4, it is clear > > that remapping 1GB without this optimization will take ~9ms on arm64. > > > > Although this mremap has to happen only once every GC cycle, and the > > typical size is also not going to be more than a GB or 2, pausing > > application threads for ~9ms is guaranteed to cause jitters. OTOH, > > with this optimization, mremap is reduced to ~60us, which is a totally > > acceptable pause time. > > > > Unfortunately, implementation of the new GC algorithm hasn't yet > > reached the point where I can quantify the effect of this > > optimization. But I can confirm that without this optimization the new > > GC will not be approved. > > IIUC, the 9ms -> 90us improvement attributed to combination HAVE_MOVE_PMD > and HAVE_MOVE_PUD, right? I expect HAVE_MOVE_PMD to be reasonable for some > workloads, but marginal benefit of HAVE_MOVE_PUD is in doubt. Do you see > it's useful for your workload? > Yes, 9ms -> 90us is when both are combined. The past experience has been that even ~1ms long stop-the-world pause is prone to cause jitters. HAVE_MOVE_PMD takes us only this far. So HAVE_MOVE_PUD is required to bring the mremap cost to acceptable level. Ideally, I was hoping that the functionality of HAVE_MOVE_PMD can be extended to all levels of the hierarchical page table, and in the process simplify the implementation. But unfortunately, that doesn't seem to be possible from patch 3. > -- > Kirill A. Shutemov