From: Ben Widawsky <ben.widawsky@intel.com>
To: linux-mm <linux-mm@kvack.org>
Subject: [PATCH 11/18] mm: Extract THP hugepage allocation
Date: Fri, 19 Jun 2020 09:24:07 -0700 [thread overview]
Message-ID: <20200619162414.1052234-12-ben.widawsky@intel.com> (raw)
In-Reply-To: <20200619162414.1052234-1-ben.widawsky@intel.com>
The next patch is going to rework this code to support
MPOL_PREFERRED_MANY. This refactor makes the that change much more
readable.
After the extraction, the resulting code makes it apparent that this can
be converted to a simple if ladder and thus allows removing the goto.
There is not meant to be any functional or behavioral changes.
Note that still at this point MPOL_PREFERRED_MANY isn't specially
handled for huge pages.
Cc: Andrew Morton <akpm@linux-foundation.org>
Cc: Dave Hansen <dave.hansen@linux.intel.com>
Cc: Michal Hocko <mhocko@kernel.org>
Signed-off-by: Ben Widawsky <ben.widawsky@intel.com>
---
mm/mempolicy.c | 96 ++++++++++++++++++++++++++------------------------
1 file changed, 49 insertions(+), 47 deletions(-)
diff --git a/mm/mempolicy.c b/mm/mempolicy.c
index 408ba78c8424..3ce2354fed44 100644
--- a/mm/mempolicy.c
+++ b/mm/mempolicy.c
@@ -2232,6 +2232,48 @@ static struct page *alloc_page_interleave(gfp_t gfp, unsigned order,
return page;
}
+static struct page *alloc_pages_vma_thp(gfp_t gfp, struct mempolicy *pol,
+ int order, int node)
+{
+ nodemask_t *nmask;
+ struct page *page;
+ int hpage_node = node;
+
+ /*
+ * For hugepage allocation and non-interleave policy which allows the
+ * current node (or other explicitly preferred node) we only try to
+ * allocate from the current/preferred node and don't fall back to other
+ * nodes, as the cost of remote accesses would likely offset THP
+ * benefits.
+ *
+ * If the policy is interleave or multiple preferred nodes, or does not
+ * allow the current node in its nodemask, we allocate the standard way.
+ */
+ if (pol->mode == MPOL_PREFERRED && !(pol->flags & MPOL_F_LOCAL))
+ hpage_node = first_node(pol->v.preferred_nodes);
+
+ nmask = policy_nodemask(gfp, pol);
+
+ /*
+ * First, try to allocate THP only on local node, but don't reclaim
+ * unnecessarily, just compact.
+ */
+ page = __alloc_pages_nodemask(gfp | __GFP_THISNODE | __GFP_NORETRY,
+ order, hpage_node, nmask);
+
+ /*
+ * If hugepage allocations are configured to always synchronous compact
+ * or the vma has been madvised to prefer hugepage backing, retry
+ * allowing remote memory with both reclaim and compact as well.
+ */
+ if (!page && (gfp & __GFP_DIRECT_RECLAIM))
+ page = __alloc_pages_nodemask(gfp, order, hpage_node, nmask);
+
+ VM_BUG_ON(page && nmask && !node_isset(page_to_nid(page), *nmask));
+
+ return page;
+}
+
/**
* alloc_pages_vma - Allocate a page for a VMA.
*
@@ -2272,57 +2314,17 @@ alloc_pages_vma(gfp_t gfp, int order, struct vm_area_struct *vma,
nid = interleave_nid(pol, vma, addr, PAGE_SHIFT + order);
mpol_cond_put(pol);
page = alloc_page_interleave(gfp, order, nid);
- goto out;
- }
-
- if (unlikely(IS_ENABLED(CONFIG_TRANSPARENT_HUGEPAGE) && hugepage)) {
- int hpage_node = node;
-
- /*
- * For hugepage allocation and non-interleave policy which
- * allows the current node (or other explicitly preferred
- * node) we only try to allocate from the current/preferred
- * node and don't fall back to other nodes, as the cost of
- * remote accesses would likely offset THP benefits.
- *
- * If the policy is interleave or multiple preferred nodes, or
- * does not allow the current node in its nodemask, we allocate
- * the standard way.
- */
- if (pol->mode == MPOL_PREFERRED && !(pol->flags & MPOL_F_LOCAL))
- hpage_node = first_node(pol->v.preferred_nodes);
-
+ } else if (unlikely(IS_ENABLED(CONFIG_TRANSPARENT_HUGEPAGE) &&
+ hugepage)) {
+ page = alloc_pages_vma_thp(gfp, pol, order, node);
+ mpol_cond_put(pol);
+ } else {
nmask = policy_nodemask(gfp, pol);
+ preferred_nid = policy_node(gfp, pol, node);
+ page = __alloc_pages_nodemask(gfp, order, preferred_nid, nmask);
mpol_cond_put(pol);
-
- /*
- * First, try to allocate THP only on local node, but
- * don't reclaim unnecessarily, just compact.
- */
- page = __alloc_pages_nodemask(gfp | __GFP_THISNODE |
- __GFP_NORETRY,
- order, hpage_node, nmask);
-
- /*
- * If hugepage allocations are configured to always synchronous
- * compact or the vma has been madvised to prefer hugepage
- * backing, retry allowing remote memory with both reclaim and
- * compact as well.
- */
- if (!page && (gfp & __GFP_DIRECT_RECLAIM))
- page = __alloc_pages_nodemask(gfp, order, hpage_node,
- nmask);
-
- VM_BUG_ON(page && nmask &&
- !node_isset(page_to_nid(page), *nmask));
- goto out;
}
- nmask = policy_nodemask(gfp, pol);
- preferred_nid = policy_node(gfp, pol, node);
- page = __alloc_pages_nodemask(gfp, order, preferred_nid, nmask);
- mpol_cond_put(pol);
-out:
return page;
}
EXPORT_SYMBOL(alloc_pages_vma);
--
2.27.0
next prev parent reply other threads:[~2020-06-19 16:24 UTC|newest]
Thread overview: 16+ messages / expand[flat|nested] mbox.gz Atom feed top
2020-06-19 16:23 [PATCH 00/18] multiple preferred nodes Ben Widawsky
2020-06-19 16:23 ` [PATCH 01/18] mm/mempolicy: Add comment for missing LOCAL Ben Widawsky
2020-06-19 16:23 ` [PATCH 02/18] mm/mempolicy: Use node_mem_id() instead of node_id() Ben Widawsky
2020-06-19 16:23 ` [PATCH 03/18] mm/page_alloc: start plumbing multi preferred node Ben Widawsky
2020-06-19 16:24 ` [PATCH 04/18] mm/page_alloc: add preferred pass to page allocation Ben Widawsky
2020-06-19 16:24 ` [PATCH 05/18] mm/mempolicy: convert single preferred_node to full nodemask Ben Widawsky
2020-06-19 16:24 ` [PATCH 06/18] mm/mempolicy: Add MPOL_PREFERRED_MANY for multiple preferred nodes Ben Widawsky
2020-06-19 16:24 ` [PATCH 07/18] mm/mempolicy: allow preferred code to take a nodemask Ben Widawsky
2020-06-19 16:24 ` [PATCH 08/18] mm/mempolicy: refactor rebind code for PREFERRED_MANY Ben Widawsky
2020-06-19 16:24 ` [PATCH 09/18] mm: Finish handling MPOL_PREFERRED_MANY Ben Widawsky
2020-06-19 16:24 ` [PATCH 10/18] mm: clean up alloc_pages_vma (thp) Ben Widawsky
2020-06-19 16:24 ` Ben Widawsky [this message]
2020-06-19 16:24 ` [PATCH 12/18] mm/mempolicy: Use __alloc_page_node for interleaved Ben Widawsky
2020-06-19 16:24 ` [PATCH 13/18] mm: kill __alloc_pages Ben Widawsky
2020-06-19 16:25 ` [PATCH 00/18] multiple preferred nodes Ben Widawsky
2020-06-19 16:24 Ben Widawsky
2020-06-19 16:24 ` [PATCH 11/18] mm: Extract THP hugepage allocation Ben Widawsky
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20200619162414.1052234-12-ben.widawsky@intel.com \
--to=ben.widawsky@intel.com \
--cc=linux-mm@kvack.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox