From: KAMEZAWA Hiroyuki <kamezawa.hiroyu@jp.fujitsu.com>
To: Heiko Carstens <heiko.carstens@de.ibm.com>
Cc: Andrew Morton <akpm@linux-foundation.org>,
Andy Whitcroft <apw@shadowen.org>,
Dave Hansen <haveblue@us.ibm.com>,
Gerald Schaefer <gerald.schaefer@de.ibm.com>,
linux-kernel@vger.kernel.org, linux-mm@kvack.org
Subject: Re: [PATCH] memory_hotplug: always initialize pageblock bitmap.
Date: Mon, 12 May 2008 10:55:00 +0900 [thread overview]
Message-ID: <20080512105500.ff89c0d3.kamezawa.hiroyu@jp.fujitsu.com> (raw)
In-Reply-To: <20080510124501.GA4796@osiris.boeblingen.de.ibm.com>
On Sat, 10 May 2008 14:45:01 +0200
Heiko Carstens <heiko.carstens@de.ibm.com> wrote:
> > Recently, I added a check "zone's start_pfn < pfn < zone's end"
> > to memmap_init_zone()'s usemap initialization for !SPARSEMEM case bug FIX.
> > (and I think the fix itself is sane.)
>
> Oh, you broke memory hot-add on -stable ;)
>
Ah yes, my mistake. Very sorry, (but stable was also broken if !SPARSEMEM)
> > How about calling grow_pgdat_span()/grow_zone_span() from __add_zone() ?
>
> Like this?
>
> Note that this patch is on top of the other patch I already sent, but
> reverts it...
>
> This works for me. A final version (if this is acceptable) should move
> the grow* functions also.
>
> ---
> mm/memory_hotplug.c | 28 +++++++++++++++++++---------
> mm/page_alloc.c | 3 +--
> 2 files changed, 20 insertions(+), 11 deletions(-)
>
> Index: linux-2.6/mm/memory_hotplug.c
> ===================================================================
> --- linux-2.6.orig/mm/memory_hotplug.c
> +++ linux-2.6/mm/memory_hotplug.c
> @@ -159,17 +159,33 @@ void register_page_bootmem_info_node(str
> }
> #endif /* !CONFIG_SPARSEMEM_VMEMMAP */
>
> +static void grow_zone_span(struct zone *zone, unsigned long start_pfn,
> + unsigned long end_pfn);
> +static void grow_pgdat_span(struct pglist_data *pgdat,
> + unsigned long start_pfn, unsigned long end_pfn);
> +
> static int __add_zone(struct zone *zone, unsigned long phys_start_pfn)
> {
> struct pglist_data *pgdat = zone->zone_pgdat;
> int nr_pages = PAGES_PER_SECTION;
> int nid = pgdat->node_id;
> int zone_type;
> + unsigned long flags;
>
> zone_type = zone - pgdat->node_zones;
> - if (!zone->wait_table)
> - return init_currently_empty_zone(zone, phys_start_pfn,
> - nr_pages, MEMMAP_HOTPLUG);
> + if (!zone->wait_table) {
> + int ret;
> +
> + ret = init_currently_empty_zone(zone, phys_start_pfn,
> + nr_pages, MEMMAP_HOTPLUG);
> + if (ret)
> + return ret;
> + }
> + pgdat_resize_lock(zone->zone_pgdat, &flags);
> + grow_zone_span(zone, phys_start_pfn, phys_start_pfn + nr_pages);
> + grow_pgdat_span(zone->zone_pgdat, phys_start_pfn,
> + phys_start_pfn + nr_pages);
> + pgdat_resize_unlock(zone->zone_pgdat, &flags);
> memmap_init_zone(nr_pages, nid, zone_type,
> phys_start_pfn, MEMMAP_HOTPLUG);
> return 0;
seems good. I'll try this logic on my ia64 box, which allows
NUMA-node hotplug.
Thank you!
-Kame
> @@ -363,7 +379,6 @@ static int online_pages_range(unsigned l
>
> int online_pages(unsigned long pfn, unsigned long nr_pages)
> {
> - unsigned long flags;
> unsigned long onlined_pages = 0;
> struct zone *zone;
> int need_zonelists_rebuild = 0;
> @@ -391,11 +406,6 @@ int online_pages(unsigned long pfn, unsi
> * memory_block->state_mutex.
> */
> zone = page_zone(pfn_to_page(pfn));
> - pgdat_resize_lock(zone->zone_pgdat, &flags);
> - grow_zone_span(zone, pfn, pfn + nr_pages);
> - grow_pgdat_span(zone->zone_pgdat, pfn, pfn + nr_pages);
> - pgdat_resize_unlock(zone->zone_pgdat, &flags);
> -
> /*
> * If this zone is not populated, then it is not in zonelist.
> * This means the page allocator ignores this zone.
> Index: linux-2.6/mm/page_alloc.c
> ===================================================================
> --- linux-2.6.orig/mm/page_alloc.c
> +++ linux-2.6/mm/page_alloc.c
> @@ -2862,8 +2862,6 @@ __meminit int init_currently_empty_zone(
>
> zone->zone_start_pfn = zone_start_pfn;
>
> - memmap_init(size, pgdat->node_id, zone_idx(zone), zone_start_pfn);
> -
> zone_init_free_lists(zone);
>
> return 0;
> @@ -3433,6 +3431,7 @@ static void __paginginit free_area_init_
> ret = init_currently_empty_zone(zone, zone_start_pfn,
> size, MEMMAP_EARLY);
> BUG_ON(ret);
> + memmap_init(size, nid, j, zone_start_pfn);
> zone_start_pfn += size;
> }
> }
>
--
To unsubscribe, send a message with 'unsubscribe linux-mm' in
the body to majordomo@kvack.org. For more info on Linux MM,
see: http://www.linux-mm.org/ .
Don't email: <a href=mailto:"dont@kvack.org"> email@kvack.org </a>
next prev parent reply other threads:[~2008-05-12 1:55 UTC|newest]
Thread overview: 9+ messages / expand[flat|nested] mbox.gz Atom feed top
2008-05-09 6:06 Heiko Carstens, Heiko Carstens
2008-05-09 6:39 ` KAMEZAWA Hiroyuki
2008-05-09 6:45 ` Heiko Carstens
2008-05-10 12:45 ` Heiko Carstens
2008-05-12 1:55 ` KAMEZAWA Hiroyuki [this message]
2008-05-12 9:19 ` KAMEZAWA Hiroyuki
2008-05-13 11:58 ` Heiko Carstens, Heiko Carstens
2008-05-13 18:52 ` Heiko Carstens
2008-05-14 0:52 ` KAMEZAWA Hiroyuki
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20080512105500.ff89c0d3.kamezawa.hiroyu@jp.fujitsu.com \
--to=kamezawa.hiroyu@jp.fujitsu.com \
--cc=akpm@linux-foundation.org \
--cc=apw@shadowen.org \
--cc=gerald.schaefer@de.ibm.com \
--cc=haveblue@us.ibm.com \
--cc=heiko.carstens@de.ibm.com \
--cc=linux-kernel@vger.kernel.org \
--cc=linux-mm@kvack.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox