From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: from mail-wm0-f72.google.com (mail-wm0-f72.google.com [74.125.82.72]) by kanga.kvack.org (Postfix) with ESMTP id 43F5C6B025F for ; Mon, 31 Jul 2017 10:37:03 -0400 (EDT) Received: by mail-wm0-f72.google.com with SMTP id 185so17358354wmk.12 for ; Mon, 31 Jul 2017 07:37:03 -0700 (PDT) Received: from mx1.suse.de (mx2.suse.de. [195.135.220.15]) by mx.google.com with ESMTPS id 63si18673089wri.231.2017.07.31.07.37.02 for (version=TLS1 cipher=AES128-SHA bits=128/128); Mon, 31 Jul 2017 07:37:02 -0700 (PDT) Date: Mon, 31 Jul 2017 16:36:53 +0200 From: Michal Hocko Subject: Re: [RFC PATCH 2/5] mm, arch: unify vmemmap_populate altmap handling Message-ID: <20170731143653.GE4829@dhcp22.suse.cz> References: <20170726083333.17754-1-mhocko@kernel.org> <20170726083333.17754-3-mhocko@kernel.org> <20170731144053.38c8b012@thinkpad> <20170731125555.GB4829@dhcp22.suse.cz> <20170731162746.60b8d98e@thinkpad> MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline In-Reply-To: <20170731162746.60b8d98e@thinkpad> Sender: owner-linux-mm@kvack.org List-ID: To: Gerald Schaefer Cc: linux-mm@kvack.org, Andrew Morton , Mel Gorman , Vlastimil Babka , Andrea Arcangeli , Jerome Glisse , Reza Arbab , Yasuaki Ishimatsu , qiuxishi@huawei.com, Kani Toshimitsu , slaoub@gmail.com, Joonsoo Kim , Andi Kleen , Daniel Kiper , Igor Mammedov , Vitaly Kuznetsov , LKML , Benjamin Herrenschmidt , Catalin Marinas , Fenghua Yu , Heiko Carstens , "H. Peter Anvin" , Ingo Molnar , Martin Schwidefsky , Michael Ellerman , Paul Mackerras , Thomas Gleixner , Tony Luck , Will Deacon On Mon 31-07-17 16:27:46, Gerald Schaefer wrote: > On Mon, 31 Jul 2017 14:55:56 +0200 > Michal Hocko wrote: > > > On Mon 31-07-17 14:40:53, Gerald Schaefer wrote: > > [...] > > > > @@ -247,12 +248,12 @@ int __meminit vmemmap_populate(unsigned long start, unsigned long end, int node) > > > > * use large frames even if they are only partially > > > > * used. > > > > * Otherwise we would have also page tables since > > > > - * vmemmap_populate gets called for each section > > > > + * __vmemmap_populate gets called for each section > > > > * separately. */ > > > > if (MACHINE_HAS_EDAT1) { > > > > void *new_page; > > > > > > > > - new_page = vmemmap_alloc_block(PMD_SIZE, node); > > > > + new_page = __vmemmap_alloc_block_buf(PMD_SIZE, node, altmap); > > > > if (!new_page) > > > > goto out; > > > > pmd_val(*pm_dir) = __pa(new_page) | sgt_prot; > > > > > > There is another call to vmemmap_alloc_block() in this function, a couple > > > of lines below, this should also be replaced by __vmemmap_alloc_block_buf(). > > > > I've noticed that one but in general I have only transformed PMD > > mappings because we shouldn't even get to pte level if the forme works > > AFAICS. Memory sections should be always 2MB aligned unless I am missing > > something. Or is this not true? > > vmemmap_populate() on s390 will only stop at pmd level if we have HW > support for large pages (MACHINE_HAS_EDAT1). In that case we will allocate > a PMD_SIZE block with vmemmap_alloc_block() and map it on pmd level as > a large page. > > Without HW large page support, we will continue to allocate a pte page, > populate the pmd entry with that, and fall through to the pte_none() > check below, with its PAGE_SIZE vmemmap_alloc_block() allocation. In this > case we should use the __vmemmap_alloc_block_buf(). OK, I see. I've considered s390 will support large pages in general. I will fold this in. Thanks! --- commit df13e3a1237c3fef399e26b0f5a015715df12ede Author: Michal Hocko Date: Mon Jul 31 16:34:18 2017 +0200 fold me "mm, arch: unify vmemmap_populate altmap handling" - use altmap even for ptes in case the HW doesn't support large pages as per Gerald Schaefer diff --git a/arch/s390/mm/vmem.c b/arch/s390/mm/vmem.c index 07120bc137a1..764b6393e66c 100644 --- a/arch/s390/mm/vmem.c +++ b/arch/s390/mm/vmem.c @@ -273,7 +273,7 @@ int __meminit __vmemmap_populate(unsigned long start, unsigned long end, int nod if (pte_none(*pt_dir)) { void *new_page; - new_page = vmemmap_alloc_block(PAGE_SIZE, node); + new_page = __vmemmap_alloc_block_buf(PAGE_SIZE, node, altmap); if (!new_page) goto out; pte_val(*pt_dir) = __pa(new_page) | pgt_prot; -- Michal Hocko SUSE Labs -- To unsubscribe, send a message with 'unsubscribe linux-mm' in the body to majordomo@kvack.org. For more info on Linux MM, see: http://www.linux-mm.org/ . Don't email: email@kvack.org