From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: from mail-pa0-f49.google.com (mail-pa0-f49.google.com [209.85.220.49]) by kanga.kvack.org (Postfix) with ESMTP id CFE556B00DB for ; Mon, 3 Nov 2014 03:04:13 -0500 (EST) Received: by mail-pa0-f49.google.com with SMTP id lj1so11670499pab.36 for ; Mon, 03 Nov 2014 00:04:13 -0800 (PST) Received: from lgemrelse6q.lge.com (LGEMRELSE6Q.lge.com. [156.147.1.121]) by mx.google.com with ESMTP id z12si14856202pdi.23.2014.11.03.00.04.10 for ; Mon, 03 Nov 2014 00:04:12 -0800 (PST) Date: Mon, 3 Nov 2014 17:05:46 +0900 From: Joonsoo Kim Subject: Re: [PATCH 0/4] (CMA_AGGRESSIVE) Make CMA memory be more aggressive about allocation Message-ID: <20141103080546.GB7052@js1304-P5Q-DELUXE> References: <1413430551-22392-1-git-send-email-zhuhui@xiaomi.com> <20141024052553.GE15243@js1304-P5Q-DELUXE> MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline In-Reply-To: Sender: owner-linux-mm@kvack.org List-ID: To: Hui Zhu Cc: Hui Zhu , rjw@rjwysocki.net, len.brown@intel.com, pavel@ucw.cz, m.szyprowski@samsung.com, Andrew Morton , mina86@mina86.com, aneesh.kumar@linux.vnet.ibm.com, hannes@cmpxchg.org, Rik van Riel , mgorman@suse.de, minchan@kernel.org, nasa4836@gmail.com, ddstreet@ieee.org, Hugh Dickins , mingo@kernel.org, rientjes@google.com, Peter Zijlstra , keescook@chromium.org, atomlin@redhat.com, raistlin@linux.it, axboe@fb.com, Paul McKenney , kirill.shutemov@linux.intel.com, n-horiguchi@ah.jp.nec.com, k.khlebnikov@samsung.com, msalter@redhat.com, deller@gmx.de, tangchen@cn.fujitsu.com, ben@decadent.org.uk, akinobu.mita@gmail.com, lauraa@codeaurora.org, vbabka@suse.cz, sasha.levin@oracle.com, vdavydov@parallels.com, suleiman@google.com, "linux-kernel@vger.kernel.org" , linux-pm@vger.kernel.org, linux-mm@kvack.org On Mon, Nov 03, 2014 at 03:28:38PM +0800, Hui Zhu wrote: > On Fri, Oct 24, 2014 at 1:25 PM, Joonsoo Kim wrote: > > On Thu, Oct 16, 2014 at 11:35:47AM +0800, Hui Zhu wrote: > >> In fallbacks of page_alloc.c, MIGRATE_CMA is the fallback of > >> MIGRATE_MOVABLE. > >> MIGRATE_MOVABLE will use MIGRATE_CMA when it doesn't have a page in > >> order that Linux kernel want. > >> > >> If a system that has a lot of user space program is running, for > >> instance, an Android board, most of memory is in MIGRATE_MOVABLE and > >> allocated. Before function __rmqueue_fallback get memory from > >> MIGRATE_CMA, the oom_killer will kill a task to release memory when > >> kernel want get MIGRATE_UNMOVABLE memory because fallbacks of > >> MIGRATE_UNMOVABLE are MIGRATE_RECLAIMABLE and MIGRATE_MOVABLE. > >> This status is odd. The MIGRATE_CMA has a lot free memory but Linux > >> kernel kill some tasks to release memory. > >> > >> This patch series adds a new function CMA_AGGRESSIVE to make CMA memory > >> be more aggressive about allocation. > >> If function CMA_AGGRESSIVE is available, when Linux kernel call function > >> __rmqueue try to get pages from MIGRATE_MOVABLE and conditions allow, > >> MIGRATE_CMA will be allocated as MIGRATE_MOVABLE first. If MIGRATE_CMA > >> doesn't have enough pages for allocation, go back to allocate memory from > >> MIGRATE_MOVABLE. > >> Then the memory of MIGRATE_MOVABLE can be kept for MIGRATE_UNMOVABLE and > >> MIGRATE_RECLAIMABLE which doesn't have fallback MIGRATE_CMA. > > > > Hello, > > > > I did some work similar to this. > > Please reference following links. > > > > https://lkml.org/lkml/2014/5/28/64 > > https://lkml.org/lkml/2014/5/28/57 > > > I tested #1 approach and found the problem. Although free memory on > > meminfo can move around low watermark, there is large fluctuation on free > > memory, because too many pages are reclaimed when kswapd is invoked. > > Reason for this behaviour is that successive allocated CMA pages are > > on the LRU list in that order and kswapd reclaim them in same order. > > These memory doesn't help watermark checking from kwapd, so too many > > pages are reclaimed, I guess. > > This issue can be handle with some change around shrink code. I am > trying to integrate a patch for them. > But I am not sure we met the same issue. Do you mind give me more > info about this part? I forgot the issue because there is so big time-gap. I need sometime to bring issue back to my brain. I will answer it soon after some thinking. > > > > > And, aggressive allocation should be postponed until freepage counting > > bug is fixed, because aggressive allocation enlarge the possiblity > > of problem occurence. I tried to fix that bug, too. See following link. > > > > https://lkml.org/lkml/2014/10/23/90 > > I am following these patches. They are great! Thanks for your work. Thanks. :) -- To unsubscribe, send a message with 'unsubscribe linux-mm' in the body to majordomo@kvack.org. For more info on Linux MM, see: http://www.linux-mm.org/ . Don't email: email@kvack.org