From: "Martin J. Bligh" <mbligh@aracnet.com>
To: Dave Hansen <haveblue@us.ibm.com>, Alok Mooley <rangdi@yahoo.com>
Cc: Linux Kernel Mailing List <linux-kernel@vger.kernel.org>,
linux-mm <linux-mm@kvack.org>
Subject: Re: Active Memory Defragmentation: Our implementation & problems
Date: Tue, 03 Feb 2004 22:05:36 -0800 [thread overview]
Message-ID: <35380000.1075874735@[10.10.2.4]> (raw)
In-Reply-To: <1075874074.14153.159.camel@nighthawk>
>> In order to move such pages, we will have to patch macros like
>> "virt_to_phys" & other related macros, so that the address
>> translation for pages moved by us will take place vmalloc style, i.e.,
>> via page tables, instead of direct +-3GB. Is it worth introducing such
>> an overhead for address translation (vmalloc does it!)? If no, then is
>> there another way out, or is it better to stick to our current
>> definition of a movable page?
>
> Low memory kernel pages are a much bigger deal to defrag. I've started
> to think about these for hotplug memory and it just makes my head hurt.
> If you want to do this, you are right, you'll have to alter virt_to_phys
> and company. The best way I've seen this is with CONFIG_NONLINEAR:
> http://lwn.net/2002/0411/a/discontig.php3
> Those lookup tables are pretty fast, and have benefits to many areas
> beyond defragmentation like NUMA and the memory hotplug projects.
I don't think that helps you really - the mappings are usually done on
chunks signficantly larger than one page, and we don't want to break
away from using large pages for the kernel mappings.
> Rather than try to defrag kernel memory now, it's probably better to
> work on schemes that keep from fragmenting memory in the first place.
Absolutely. Kernel pages are really hard (not any lowmem page is a
kernel page, of course).
>> Identifying pages moved by us may involve introducing a new page-flag.
>> A new page-flag for per-cpu pages would be great, since we have to
>> traverse the per-cpu hot & cold lists in order to identify if a page
>> is on the pcp lists.
Careful not to introduce new cacheline touches, etc whilst doing this.
The whole point of hot & cold pages is to be efficient.
If you don't need N kilobyte alignment on your N kilobyte page groups,
there's probably much more effective schemes that buddy allocator, but
that assumption may be too embedded to change.
> If the per-cpu allocator caches are your only problem, I don't see why
> we can't just flush them out when you're doing your operation. Plus,
> they aren't *that* big, so you could pretty easily go scanning them.
> Martin, can we just flush out and turn off the per-cpu hot/cold lists
> for the defrag period?
Yup, should be fairly easy to do. Just free them back with the standard
mechanisms.
>> As of now, we have adopted a failure based approach, i.e, we
>> defragment only when a higher order allocation failure has taken place
>> (just before kswapd starts swapping). We now want to defragment based
>> on thresholds kept for each allocation order. Instead of a daemon
>> kicking in on a threshold violation (as proposed by Mr. Daniel
>> Phillips), we intend to capture idle cpu cycles by inserting a new
>> process just above the idle process.
>
> I think I'd agree with Dan on that one. When kswapd is going, it's
> pretty much too late. The daemon approach would be more flexible, allow
> you to start earlier, and more easily have various levels of
> aggressiveness.
I think the policy we've taken so far is that you can't *urgently* request
large contig areas. If you need that, you should be keeping your own cache.
M.
--
To unsubscribe, send a message with 'unsubscribe linux-mm' in
the body to majordomo@kvack.org. For more info on Linux MM,
see: http://www.linux-mm.org/ .
Don't email: <a href=mailto:"aart@kvack.org"> aart@kvack.org </a>
next prev parent reply other threads:[~2004-02-04 6:05 UTC|newest]
Thread overview: 35+ messages / expand[flat|nested] mbox.gz Atom feed top
2004-02-03 4:46 Alok Mooley
2004-02-03 21:26 ` Dave Hansen
2004-02-03 22:26 ` Martin J. Bligh
2004-02-04 5:09 ` Alok Mooley
2004-02-04 5:24 ` Mike Fedyk
2004-02-04 5:54 ` Dave Hansen
2004-02-04 6:05 ` Martin J. Bligh [this message]
2004-02-04 6:22 ` Dave Hansen
2004-02-04 6:29 ` Martin J. Bligh
2004-02-04 6:40 ` Dave Hansen
2004-02-04 7:17 ` Martin J. Bligh
2004-02-04 8:30 ` Andrew Morton
2004-02-04 6:53 ` Doubt about statm_pgd_range patch Arunkumar
2004-02-04 6:57 ` Active Memory Defragmentation: Our implementation & problems IWAMOTO Toshihiro
2004-02-04 7:10 ` Dave Hansen
2004-02-04 7:50 ` IWAMOTO Toshihiro
2004-02-04 10:33 ` Hirokazu Takahashi
2004-02-04 18:33 ` Alok Mooley
2004-02-04 18:46 ` Dave Hansen
2004-02-04 18:54 ` Alok Mooley
2004-02-04 19:07 ` Richard B. Johnson
2004-02-04 19:18 ` Alok Mooley
2004-02-04 19:33 ` Richard B. Johnson
2004-02-05 5:07 ` Alok Mooley
2004-02-05 19:03 ` Pavel Machek
2004-02-04 19:35 ` Dave McCracken
2004-02-04 21:59 ` Timothy Miller
2004-02-04 23:24 ` Dave Hansen
2004-02-05 16:32 ` Dave McCracken
2004-02-04 19:37 ` Timothy Miller
2004-02-04 19:43 ` Dave Hansen
2004-02-04 19:59 ` Richard B. Johnson
2004-02-04 19:56 ` Dave Hansen
2004-02-05 5:19 ` Alok Mooley
2004-02-04 20:12 Mark_H_Johnson
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to='35380000.1075874735@[10.10.2.4]' \
--to=mbligh@aracnet.com \
--cc=haveblue@us.ibm.com \
--cc=linux-kernel@vger.kernel.org \
--cc=linux-mm@kvack.org \
--cc=rangdi@yahoo.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox