From: David Rientjes <rientjes@google.com>
To: Lee Schermerhorn <Lee.Schermerhorn@hp.com>
Cc: Mel Gorman <mel@csn.ul.ie>,
linux-mm@kvack.org, Andrew Morton <akpm@linux-foundation.org>,
Nishanth Aravamudan <nacc@us.ibm.com>,
Adam Litke <agl@us.ibm.com>, Andy Whitcroft <apw@canonical.com>,
eric.whitney@hp.com, Ranjit Manomohan <ranjitm@google.com>
Subject: Re: [PATCH 0/5] Huge Pages Nodes Allowed
Date: Wed, 24 Jun 2009 15:26:46 -0700 (PDT) [thread overview]
Message-ID: <alpine.DEB.2.00.0906241451460.30523@chino.kir.corp.google.com> (raw)
In-Reply-To: <1245842724.6439.19.camel@lts-notebook>
On Wed, 24 Jun 2009, Lee Schermerhorn wrote:
> David:
>
> Nish mentioned this to me a while back when I asked about his patches.
> That's one of my reasons for seeing if the simpler [IMO] nodes_allowed
> would be sufficient. I'm currently updating the nodes_allowed series
> per Mel's cleanup suggestions.
The /proc/sys/vm/hugepages_nodes_allowed support is troublesome because
it's global and can race with other writers, such as for tasks in two
disjoint cpusets attempting to allocate hugepages concurrently.
> I'll then prototype Mel's preferred
> method of using the task's mempolicy.
This proposal eliminates the aforementioned race, but now has the opposite
problem: if a single task is allocating hugepages for multiple cpusets, it
must setup the correct mempolicies to allocate (or free) for the new
cpuset mems.
> I still have reservations about
> this: static huge page allocation is currently not constrained by
> policy nor cpusets, and I can't tell whether the task's mempolicy was
> set explicitly to contstrain the huge pages or just inherited from the
> parent shell.
>
Agreed. I do think that we used to constrain hugepage allocations by
cpuset in the past, though, when the allocation was simply done via
alloc_pages_node(). We'd fallback to using nodes outside of
cpuset_current_mems_allowed when the nodes were too full or fragmented.
> Next I'll also dust off Nish's old per node hugetlb control patches and
> see what it task to update them for the multiple sizes. It will look
> pretty much as you suggest. Do you have any suggestions for a boot
> command line syntax to specify per node huge page counts at boot time
> [assuming we still want this]? Currently, for default huge page size,
> distributed across nodes, we have:
>
> hugepages=<N>
>
> I was thinking something like:
>
> hugepages=(node:count,...)
>
> using the '(' as a flag for per node counts, w/o needing to prescan for
> ':'
>
The hugepages=(node:count,...) option would still need to be interleaved
with hugepagesz= for the various sizes. This could become pretty cryptic:
hugepagesz=2M hugepages=(0:10,1:20) hugepagesz=1G \
hugepages=(2:10,3:10)
and I assume we'd use `count' of 99999 for nodes of unknown sizes where we
simply want to allocate as many hugepages as possible.
We'd still need to support hugepages=N for large NUMA machines so we don't
have to specify the same number of hugepages per node for a true
interleave, which would require an extremely large command line. And then
the behavior of
hugepagesz=1G hugepages=(0:10,1:20) hugepages=30
needs to be defined. In that case, does hugepages=30 override the
previous settings if this system only has dual nodes? If so, for SGI's 1K
node systems it's going to be difficult to specify many nodes with 10
hugepages and a few with 20. So perhaps hugepages=(node:count,...) should
increment or decrement the hugepages= value, if specified?
--
To unsubscribe, send a message with 'unsubscribe linux-mm' in
the body to majordomo@kvack.org. For more info on Linux MM,
see: http://www.linux-mm.org/ .
Don't email: <a href=mailto:"dont@kvack.org"> email@kvack.org </a>
next prev parent reply other threads:[~2009-06-24 22:26 UTC|newest]
Thread overview: 31+ messages / expand[flat|nested] mbox.gz Atom feed top
2009-06-16 13:52 Lee Schermerhorn
2009-06-16 13:52 ` [PATCH 1/5] Free huge pages round robin to balance across nodes Lee Schermerhorn
2009-06-17 13:18 ` Mel Gorman
2009-06-17 17:16 ` Lee Schermerhorn
2009-06-18 19:08 ` David Rientjes
2009-06-16 13:52 ` [PATCH 2/5] Add nodes_allowed members to hugepages hstate struct Lee Schermerhorn
2009-06-17 13:35 ` Mel Gorman
2009-06-17 17:38 ` Lee Schermerhorn
2009-06-18 9:17 ` Mel Gorman
2009-06-16 13:53 ` [PATCH 3/5] Use per hstate nodes_allowed to constrain huge page allocation Lee Schermerhorn
2009-06-17 13:39 ` Mel Gorman
2009-06-17 17:47 ` Lee Schermerhorn
2009-06-18 9:18 ` Mel Gorman
2009-06-16 13:53 ` [PATCH 4/5] Add sysctl for default hstate nodes_allowed Lee Schermerhorn
2009-06-17 13:41 ` Mel Gorman
2009-06-17 17:52 ` Lee Schermerhorn
2009-06-18 9:19 ` Mel Gorman
2009-06-16 13:53 ` [PATCH 5/5] Update huge pages kernel documentation Lee Schermerhorn
2009-06-18 18:49 ` David Rientjes
2009-06-18 19:06 ` Lee Schermerhorn
2009-06-17 13:02 ` [PATCH 0/5] Huge Pages Nodes Allowed Mel Gorman
2009-06-17 17:15 ` Lee Schermerhorn
2009-06-18 9:33 ` Mel Gorman
2009-06-18 14:46 ` Lee Schermerhorn
2009-06-18 15:00 ` Mel Gorman
2009-06-18 19:08 ` David Rientjes
2009-06-24 7:11 ` David Rientjes
2009-06-24 11:25 ` Lee Schermerhorn
2009-06-24 22:26 ` David Rientjes [this message]
2009-06-25 2:14 ` Lee Schermerhorn
2009-06-25 19:22 ` David Rientjes
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=alpine.DEB.2.00.0906241451460.30523@chino.kir.corp.google.com \
--to=rientjes@google.com \
--cc=Lee.Schermerhorn@hp.com \
--cc=agl@us.ibm.com \
--cc=akpm@linux-foundation.org \
--cc=apw@canonical.com \
--cc=eric.whitney@hp.com \
--cc=linux-mm@kvack.org \
--cc=mel@csn.ul.ie \
--cc=nacc@us.ibm.com \
--cc=ranjitm@google.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox