From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: from mail-pa0-f69.google.com (mail-pa0-f69.google.com [209.85.220.69]) by kanga.kvack.org (Postfix) with ESMTP id 92B206B007E for ; Wed, 25 May 2016 22:38:03 -0400 (EDT) Received: by mail-pa0-f69.google.com with SMTP id gw7so92566276pac.0 for ; Wed, 25 May 2016 19:38:03 -0700 (PDT) Received: from mail-pf0-x241.google.com (mail-pf0-x241.google.com. [2607:f8b0:400e:c00::241]) by mx.google.com with ESMTPS id d15si3007470pfb.137.2016.05.25.19.38.02 for (version=TLS1_2 cipher=ECDHE-RSA-AES128-GCM-SHA256 bits=128/128); Wed, 25 May 2016 19:38:02 -0700 (PDT) Received: by mail-pf0-x241.google.com with SMTP id g132so446571pfb.3 for ; Wed, 25 May 2016 19:38:02 -0700 (PDT) From: js1304@gmail.com Subject: [PATCH v2 2/7] mm/page_owner: initialize page owner without holding the zone lock Date: Thu, 26 May 2016 11:37:50 +0900 Message-Id: <1464230275-25791-2-git-send-email-iamjoonsoo.kim@lge.com> In-Reply-To: <1464230275-25791-1-git-send-email-iamjoonsoo.kim@lge.com> References: <1464230275-25791-1-git-send-email-iamjoonsoo.kim@lge.com> Sender: owner-linux-mm@kvack.org List-ID: To: Andrew Morton Cc: Vlastimil Babka , mgorman@techsingularity.net, Minchan Kim , Alexander Potapenko , Hugh Dickins , Michal Hocko , linux-kernel@vger.kernel.org, linux-mm@kvack.org, Joonsoo Kim From: Joonsoo Kim It's not necessary to initialized page_owner with holding the zone lock. It would cause more contention on the zone lock although it's not a big problem since it is just debug feature. But, it is better than before so do it. This is also preparation step to use stackdepot in page owner feature. Stackdepot allocates new pages when there is no reserved space and holding the zone lock in this case will cause deadlock. Signed-off-by: Joonsoo Kim --- mm/compaction.c | 3 +++ mm/page_alloc.c | 2 -- mm/page_isolation.c | 9 ++++++--- 3 files changed, 9 insertions(+), 5 deletions(-) diff --git a/mm/compaction.c b/mm/compaction.c index 8e013eb..6043ef8 100644 --- a/mm/compaction.c +++ b/mm/compaction.c @@ -20,6 +20,7 @@ #include #include #include +#include #include "internal.h" #ifdef CONFIG_COMPACTION @@ -80,6 +81,8 @@ static void map_pages(struct list_head *list) arch_alloc_page(page, order); kernel_map_pages(page, nr_pages, 1); kasan_alloc_pages(page, order); + + set_page_owner(page, order, __GFP_MOVABLE); if (order) split_page(page, order); diff --git a/mm/page_alloc.c b/mm/page_alloc.c index 5134f46..1b1ca57 100644 --- a/mm/page_alloc.c +++ b/mm/page_alloc.c @@ -2507,8 +2507,6 @@ int __isolate_free_page(struct page *page, unsigned int order) zone->free_area[order].nr_free--; rmv_page_order(page); - set_page_owner(page, order, __GFP_MOVABLE); - /* Set the pageblock if the isolated page is at least a pageblock */ if (order >= pageblock_order - 1) { struct page *endpage = page + (1 << order) - 1; diff --git a/mm/page_isolation.c b/mm/page_isolation.c index 612122b..927f5ee 100644 --- a/mm/page_isolation.c +++ b/mm/page_isolation.c @@ -7,6 +7,7 @@ #include #include #include +#include #include "internal.h" #define CREATE_TRACE_POINTS @@ -108,8 +109,6 @@ static void unset_migratetype_isolate(struct page *page, unsigned migratetype) if (pfn_valid_within(page_to_pfn(buddy)) && !is_migrate_isolate_page(buddy)) { __isolate_free_page(page, order); - kernel_map_pages(page, (1 << order), 1); - set_page_refcounted(page); isolated_page = page; } } @@ -128,8 +127,12 @@ static void unset_migratetype_isolate(struct page *page, unsigned migratetype) zone->nr_isolate_pageblock--; out: spin_unlock_irqrestore(&zone->lock, flags); - if (isolated_page) + if (isolated_page) { + kernel_map_pages(page, (1 << order), 1); + set_page_refcounted(page); + set_page_owner(page, order, __GFP_MOVABLE); __free_pages(isolated_page, order); + } } static inline struct page * -- 1.9.1 -- To unsubscribe, send a message with 'unsubscribe linux-mm' in the body to majordomo@kvack.org. For more info on Linux MM, see: http://www.linux-mm.org/ . Don't email: email@kvack.org