From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: from mail-oi0-f70.google.com (mail-oi0-f70.google.com [209.85.218.70]) by kanga.kvack.org (Postfix) with ESMTP id 638006B0007 for ; Wed, 15 Aug 2018 12:55:43 -0400 (EDT) Received: by mail-oi0-f70.google.com with SMTP id w128-v6so1532462oiw.14 for ; Wed, 15 Aug 2018 09:55:43 -0700 (PDT) Received: from mx0a-00082601.pphosted.com (mx0a-00082601.pphosted.com. [67.231.145.42]) by mx.google.com with ESMTPS id b124-v6si14767346oia.425.2018.08.15.09.55.42 for (version=TLS1_2 cipher=ECDHE-RSA-AES128-GCM-SHA256 bits=128/128); Wed, 15 Aug 2018 09:55:42 -0700 (PDT) Date: Wed, 15 Aug 2018 09:55:17 -0700 From: Roman Gushchin Subject: Re: [RFC PATCH 1/2] mm: rework memcg kernel stack accounting Message-ID: <20180815165513.GA26330@castle.DHCP.thefacebook.com> References: <20180815003620.15678-1-guro@fb.com> <20180815163923.GA28953@cmpxchg.org> MIME-Version: 1.0 Content-Type: text/plain; charset="us-ascii" Content-Disposition: inline In-Reply-To: <20180815163923.GA28953@cmpxchg.org> Sender: owner-linux-mm@kvack.org List-ID: To: Johannes Weiner Cc: linux-mm@kvack.org, linux-kernel@vger.kernel.org, kernel-team@fb.com, Michal Hocko , Andy Lutomirski , Konstantin Khlebnikov , Tejun Heo On Wed, Aug 15, 2018 at 12:39:23PM -0400, Johannes Weiner wrote: > On Tue, Aug 14, 2018 at 05:36:19PM -0700, Roman Gushchin wrote: > > @@ -224,9 +224,14 @@ static unsigned long *alloc_thread_stack_node(struct task_struct *tsk, int node) > > return s->addr; > > } > > > > + /* > > + * Allocated stacks are cached and later reused by new threads, > > + * so memcg accounting is performed manually on assigning/releasing > > + * stacks to tasks. Drop __GFP_ACCOUNT. > > + */ > > stack = __vmalloc_node_range(THREAD_SIZE, THREAD_ALIGN, > > VMALLOC_START, VMALLOC_END, > > - THREADINFO_GFP, > > + THREADINFO_GFP & ~__GFP_ACCOUNT, > > PAGE_KERNEL, > > 0, node, __builtin_return_address(0)); > > > > @@ -246,12 +251,41 @@ static unsigned long *alloc_thread_stack_node(struct task_struct *tsk, int node) > > #endif > > } > > > > +static void memcg_charge_kernel_stack(struct task_struct *tsk) > > +{ > > +#ifdef CONFIG_VMAP_STACK > > + struct vm_struct *vm = task_stack_vm_area(tsk); > > + > > + if (vm) { > > + int i; > > + > > + for (i = 0; i < THREAD_SIZE / PAGE_SIZE; i++) > > + memcg_kmem_charge(vm->pages[i], __GFP_NOFAIL, > > + compound_order(vm->pages[i])); > > + > > + /* All stack pages belong to the same memcg. */ > > + mod_memcg_page_state(vm->pages[0], MEMCG_KERNEL_STACK_KB, > > + THREAD_SIZE / 1024); > > + } > > +#endif > > +} > > Before this change, the memory limit can fail the fork, but afterwards > fork() can grow memory consumption unimpeded by the cgroup settings. > > Can we continue to use try_charge() here and fail the fork? We can, but I'm not convinced we should. Kernel stack is relatively small, and it's already allocated at this point. So IMO exceeding the memcg limit for 1-2 pages isn't worse than adding complexity and handle this case (e.g. uncharge partially charged stack). Do you have an example, when it does matter?