From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: from mail-ig0-f179.google.com (mail-ig0-f179.google.com [209.85.213.179]) by kanga.kvack.org (Postfix) with ESMTP id 6461E6B025F for ; Tue, 5 Apr 2016 10:35:34 -0400 (EDT) Received: by mail-ig0-f179.google.com with SMTP id gy3so56151645igb.0 for ; Tue, 05 Apr 2016 07:35:34 -0700 (PDT) Received: from e28smtp03.in.ibm.com (e28smtp03.in.ibm.com. [125.16.236.3]) by mx.google.com with ESMTPS id w16si13000039ioi.156.2016.04.05.07.35.30 for (version=TLS1_2 cipher=AES128-SHA bits=128/128); Tue, 05 Apr 2016 07:35:31 -0700 (PDT) Received: from localhost by e28smtp03.in.ibm.com with IBM ESMTP SMTP Gateway: Authorized Use Only! Violators will be prosecuted for from ; Tue, 5 Apr 2016 20:05:28 +0530 Received: from d28av01.in.ibm.com (d28av01.in.ibm.com [9.184.220.63]) by d28relay03.in.ibm.com (8.14.9/8.14.9/NCO v10.0) with ESMTP id u35EZPE766715682 for ; Tue, 5 Apr 2016 20:05:25 +0530 Received: from d28av01.in.ibm.com (localhost [127.0.0.1]) by d28av01.in.ibm.com (8.14.4/8.14.4/NCO v10.0 AVout) with ESMTP id u35K3UHJ007894 for ; Wed, 6 Apr 2016 01:33:32 +0530 From: "Aneesh Kumar K.V" Subject: Re: [PATCH] thp: keep huge zero pinned until tlb flush In-Reply-To: <1459814247-45614-1-git-send-email-kirill.shutemov@linux.intel.com> References: <1459814247-45614-1-git-send-email-kirill.shutemov@linux.intel.com> Date: Tue, 05 Apr 2016 20:05:06 +0530 Message-ID: <877fgcxgx1.fsf@linux.vnet.ibm.com> MIME-Version: 1.0 Content-Type: text/plain Sender: owner-linux-mm@kvack.org List-ID: To: "Kirill A. Shutemov" , Andrew Morton , Andrea Arcangeli Cc: Mel Gorman , Hugh Dickins , Johannes Weiner , Dave Hansen , Vlastimil Babka , linux-mm@kvack.org "Kirill A. Shutemov" writes: > [ text/plain ] > Andrea has found[1] a race condition on MMU-gather based TLB flush vs > split_huge_page() or shrinker which frees huge zero under us (patch 1/2 > and 2/2 respectively). > > With new THP refcounting, we don't patch 1/2: mmu_gather keeps the page > page pinned until flush is complete and the pin prevent the page from > being split under us. > > We sill need patch 2/2. This is simplified version of Andrea's patch. > We don't need fancy encoding. > > [1] http://lkml.kernel.org/r/1447938052-22165-1-git-send-email-aarcange@redhat.com May be we should add the below from the original mail put_huge_zero_page() because it can actually lead to the almost immediate freeing of the THP zero page, must be run in the MMU gather and not before the TLB flush like it is happening right now. Reviewed-by: Aneesh Kumar K.V > > Signed-off-by: Kirill A. Shutemov > Reported-by: Andrea Arcangeli > --- > include/linux/huge_mm.h | 1 + > mm/huge_memory.c | 6 +++--- > mm/swap.c | 5 +++++ > 3 files changed, 9 insertions(+), 3 deletions(-) > > diff --git a/include/linux/huge_mm.h b/include/linux/huge_mm.h > index 7008623e24b1..8232e0b8a04f 100644 > --- a/include/linux/huge_mm.h > +++ b/include/linux/huge_mm.h > @@ -152,6 +152,7 @@ static inline bool is_huge_zero_pmd(pmd_t pmd) > } > > struct page *get_huge_zero_page(void); > +void put_huge_zero_page(void); > > #else /* CONFIG_TRANSPARENT_HUGEPAGE */ > #define HPAGE_PMD_SHIFT ({ BUILD_BUG(); 0; }) > diff --git a/mm/huge_memory.c b/mm/huge_memory.c > index 820214137bc5..860c7dec197e 100644 > --- a/mm/huge_memory.c > +++ b/mm/huge_memory.c > @@ -235,7 +235,7 @@ retry: > return READ_ONCE(huge_zero_page); > } > > -static void put_huge_zero_page(void) > +void put_huge_zero_page(void) > { > /* > * Counter should never go to zero here. Only shrinker can put > @@ -1715,12 +1715,12 @@ int zap_huge_pmd(struct mmu_gather *tlb, struct vm_area_struct *vma, > if (vma_is_dax(vma)) { > spin_unlock(ptl); > if (is_huge_zero_pmd(orig_pmd)) > - put_huge_zero_page(); > + tlb_remove_page(tlb, pmd_page(orig_pmd)); > } else if (is_huge_zero_pmd(orig_pmd)) { > pte_free(tlb->mm, pgtable_trans_huge_withdraw(tlb->mm, pmd)); > atomic_long_dec(&tlb->mm->nr_ptes); > spin_unlock(ptl); > - put_huge_zero_page(); > + tlb_remove_page(tlb, pmd_page(orig_pmd)); > } else { > struct page *page = pmd_page(orig_pmd); > page_remove_rmap(page, true); > diff --git a/mm/swap.c b/mm/swap.c > index 09fe5e97714a..11915bd0f047 100644 > --- a/mm/swap.c > +++ b/mm/swap.c > @@ -728,6 +728,11 @@ void release_pages(struct page **pages, int nr, bool cold) > zone = NULL; > } > > + if (is_huge_zero_page(page)) { > + put_huge_zero_page(); > + continue; > + } > + > page = compound_head(page); > if (!put_page_testzero(page)) > continue; > -- > 2.8.0.rc3 -- To unsubscribe, send a message with 'unsubscribe linux-mm' in the body to majordomo@kvack.org. For more info on Linux MM, see: http://www.linux-mm.org/ . Don't email: email@kvack.org