linux-mm.kvack.org archive mirror
 help / color / mirror / Atom feed
* [PATCH] slub: add missing kmem cgroup support to kmem_cache_free_bulk
@ 2015-10-29 13:05 Jesper Dangaard Brouer
  2015-11-05  5:06 ` Joonsoo Kim
  0 siblings, 1 reply; 4+ messages in thread
From: Jesper Dangaard Brouer @ 2015-10-29 13:05 UTC (permalink / raw)
  To: linux-mm
  Cc: Joonsoo Kim, Andrew Morton, Christoph Lameter, Jesper Dangaard Brouer

Initial implementation missed support for kmem cgroup support
in kmem_cache_free_bulk() call, add this.

If CONFIG_MEMCG_KMEM is not enabled, the compiler should
be smart enough to not add any asm code.

Signed-off-by: Jesper Dangaard Brouer <brouer@redhat.com>
---
 mm/slub.c |    3 +++
 1 file changed, 3 insertions(+)

diff --git a/mm/slub.c b/mm/slub.c
index 9be12ffae9fc..9875864ad7b8 100644
--- a/mm/slub.c
+++ b/mm/slub.c
@@ -2845,6 +2845,9 @@ static int build_detached_freelist(struct kmem_cache *s, size_t size,
 	if (!object)
 		return 0;
 
+	/* Support for kmemcg */
+	s = cache_from_obj(s, object);
+
 	/* Start new detached freelist */
 	set_freepointer(s, object, NULL);
 	df->page = virt_to_head_page(object);

--
To unsubscribe, send a message with 'unsubscribe linux-mm' in
the body to majordomo@kvack.org.  For more info on Linux MM,
see: http://www.linux-mm.org/ .
Don't email: <a href=mailto:"dont@kvack.org"> email@kvack.org </a>

^ permalink raw reply	[flat|nested] 4+ messages in thread

* Re: [PATCH] slub: add missing kmem cgroup support to kmem_cache_free_bulk
  2015-10-29 13:05 [PATCH] slub: add missing kmem cgroup support to kmem_cache_free_bulk Jesper Dangaard Brouer
@ 2015-11-05  5:06 ` Joonsoo Kim
  2015-11-05  8:38   ` Vladimir Davydov
  0 siblings, 1 reply; 4+ messages in thread
From: Joonsoo Kim @ 2015-11-05  5:06 UTC (permalink / raw)
  To: Jesper Dangaard Brouer; +Cc: linux-mm, Andrew Morton, Christoph Lameter

On Thu, Oct 29, 2015 at 02:05:31PM +0100, Jesper Dangaard Brouer wrote:
> Initial implementation missed support for kmem cgroup support
> in kmem_cache_free_bulk() call, add this.
> 
> If CONFIG_MEMCG_KMEM is not enabled, the compiler should
> be smart enough to not add any asm code.
> 
> Signed-off-by: Jesper Dangaard Brouer <brouer@redhat.com>
> ---
>  mm/slub.c |    3 +++
>  1 file changed, 3 insertions(+)
> 
> diff --git a/mm/slub.c b/mm/slub.c
> index 9be12ffae9fc..9875864ad7b8 100644
> --- a/mm/slub.c
> +++ b/mm/slub.c
> @@ -2845,6 +2845,9 @@ static int build_detached_freelist(struct kmem_cache *s, size_t size,
>  	if (!object)
>  		return 0;
>  
> +	/* Support for kmemcg */
> +	s = cache_from_obj(s, object);
> +
>  	/* Start new detached freelist */
>  	set_freepointer(s, object, NULL);
>  	df->page = virt_to_head_page(object);

Hello,

It'd better to add this 's = cache_from_obj()' on kmem_cache_free_bulk().
Not only build_detached_freelist() but also slab_free() need proper
cache.

Thanks

--
To unsubscribe, send a message with 'unsubscribe linux-mm' in
the body to majordomo@kvack.org.  For more info on Linux MM,
see: http://www.linux-mm.org/ .
Don't email: <a href=mailto:"dont@kvack.org"> email@kvack.org </a>

^ permalink raw reply	[flat|nested] 4+ messages in thread

* Re: [PATCH] slub: add missing kmem cgroup support to kmem_cache_free_bulk
  2015-11-05  5:06 ` Joonsoo Kim
@ 2015-11-05  8:38   ` Vladimir Davydov
  2015-11-05 13:19     ` Jesper Dangaard Brouer
  0 siblings, 1 reply; 4+ messages in thread
From: Vladimir Davydov @ 2015-11-05  8:38 UTC (permalink / raw)
  To: Jesper Dangaard Brouer
  Cc: linux-mm, Joonsoo Kim, Andrew Morton, Christoph Lameter

On Thu, Nov 05, 2015 at 02:06:21PM +0900, Joonsoo Kim wrote:
> On Thu, Oct 29, 2015 at 02:05:31PM +0100, Jesper Dangaard Brouer wrote:
> > Initial implementation missed support for kmem cgroup support
> > in kmem_cache_free_bulk() call, add this.
> > 
> > If CONFIG_MEMCG_KMEM is not enabled, the compiler should
> > be smart enough to not add any asm code.
> > 
> > Signed-off-by: Jesper Dangaard Brouer <brouer@redhat.com>
> > ---
> >  mm/slub.c |    3 +++
> >  1 file changed, 3 insertions(+)
> > 
> > diff --git a/mm/slub.c b/mm/slub.c
> > index 9be12ffae9fc..9875864ad7b8 100644
> > --- a/mm/slub.c
> > +++ b/mm/slub.c
> > @@ -2845,6 +2845,9 @@ static int build_detached_freelist(struct kmem_cache *s, size_t size,
> >  	if (!object)
> >  		return 0;
> >  
> > +	/* Support for kmemcg */
> > +	s = cache_from_obj(s, object);
> > +
> >  	/* Start new detached freelist */
> >  	set_freepointer(s, object, NULL);
> >  	df->page = virt_to_head_page(object);
> 
> Hello,
> 
> It'd better to add this 's = cache_from_obj()' on kmem_cache_free_bulk().
> Not only build_detached_freelist() but also slab_free() need proper
> cache.

Yeah, Joonsoo is right. Besides, there's a bug in kmem_cache_alloc_bulk:

> /* Note that interrupts must be enabled when calling this function. */
> bool kmem_cache_alloc_bulk(struct kmem_cache *s, gfp_t flags, size_t size,
> 			   void **p)
> {
> 	struct kmem_cache_cpu *c;
> 	int i;
> 
> 	/*
> 	 * Drain objects in the per cpu slab, while disabling local
> 	 * IRQs, which protects against PREEMPT and interrupts
> 	 * handlers invoking normal fastpath.
> 	 */
> 	local_irq_disable();
> 	c = this_cpu_ptr(s->cpu_slab);
> 
> 	for (i = 0; i < size; i++) {
> 		void *object = c->freelist;
> 
> 		if (unlikely(!object)) {
> 			/*
> 			 * Invoking slow path likely have side-effect
> 			 * of re-populating per CPU c->freelist
> 			 */
> 			p[i] = ___slab_alloc(s, flags, NUMA_NO_NODE,
> 					    _RET_IP_, c);
> 			if (unlikely(!p[i]))
> 				goto error;
> 
> 			c = this_cpu_ptr(s->cpu_slab);
> 			continue; /* goto for-loop */
> 		}
> 
> 		/* kmem_cache debug support */
> 		s = slab_pre_alloc_hook(s, flags);

slab_pre_alloc_hook expects a global cache and returns per memcg one, so
calling this function from inside a kmemcg will result in hitting the
VM_BUG_ON in __memcg_kmem_get_cache, not saying about mis-accounting of
__slab_alloc.

memcg_kmem_get_cache should be called once, in the very beginning of
kmem_cache_alloc_bulk, and it should be matched by memcg_kmem_put_cache
when we are done.

Thanks,
Vladimir

> 		if (unlikely(!s))
> 			goto error;
> 
> 		c->freelist = get_freepointer(s, object);
> 		p[i] = object;
> 
> 		/* kmem_cache debug support */
> 		slab_post_alloc_hook(s, flags, object);
> 	}
> 	c->tid = next_tid(c->tid);
> 	local_irq_enable();
> 
> 	/* Clear memory outside IRQ disabled fastpath loop */
> 	if (unlikely(flags & __GFP_ZERO)) {
> 		int j;
> 
> 		for (j = 0; j < i; j++)
> 			memset(p[j], 0, s->object_size);
> 	}
> 
> 	return true;
> 
> error:
> 	__kmem_cache_free_bulk(s, i, p);
> 	local_irq_enable();
> 	return false;
> }

--
To unsubscribe, send a message with 'unsubscribe linux-mm' in
the body to majordomo@kvack.org.  For more info on Linux MM,
see: http://www.linux-mm.org/ .
Don't email: <a href=mailto:"dont@kvack.org"> email@kvack.org </a>

^ permalink raw reply	[flat|nested] 4+ messages in thread

* Re: [PATCH] slub: add missing kmem cgroup support to kmem_cache_free_bulk
  2015-11-05  8:38   ` Vladimir Davydov
@ 2015-11-05 13:19     ` Jesper Dangaard Brouer
  0 siblings, 0 replies; 4+ messages in thread
From: Jesper Dangaard Brouer @ 2015-11-05 13:19 UTC (permalink / raw)
  To: Vladimir Davydov
  Cc: linux-mm, Joonsoo Kim, Andrew Morton, Christoph Lameter, brouer

On Thu, 5 Nov 2015 11:38:43 +0300
Vladimir Davydov <vdavydov@virtuozzo.com> wrote:

> On Thu, Nov 05, 2015 at 02:06:21PM +0900, Joonsoo Kim wrote:
> > On Thu, Oct 29, 2015 at 02:05:31PM +0100, Jesper Dangaard Brouer wrote:
> > > Initial implementation missed support for kmem cgroup support
> > > in kmem_cache_free_bulk() call, add this.
> > > 
> > > If CONFIG_MEMCG_KMEM is not enabled, the compiler should
> > > be smart enough to not add any asm code.
> > > 
> > > Signed-off-by: Jesper Dangaard Brouer <brouer@redhat.com>
> > > ---
> > >  mm/slub.c |    3 +++
> > >  1 file changed, 3 insertions(+)
> > > 
> > > diff --git a/mm/slub.c b/mm/slub.c
> > > index 9be12ffae9fc..9875864ad7b8 100644
> > > --- a/mm/slub.c
> > > +++ b/mm/slub.c
> > > @@ -2845,6 +2845,9 @@ static int build_detached_freelist(struct kmem_cache *s, size_t size,
> > >  	if (!object)
> > >  		return 0;
> > >  
> > > +	/* Support for kmemcg */
> > > +	s = cache_from_obj(s, object);
> > > +
> > >  	/* Start new detached freelist */
> > >  	set_freepointer(s, object, NULL);
> > >  	df->page = virt_to_head_page(object);
> > 
> > Hello,
> > 
> > It'd better to add this 's = cache_from_obj()' on kmem_cache_free_bulk().
> > Not only build_detached_freelist() but also slab_free() need proper
> > cache.
> 
> Yeah, Joonsoo is right.

But cache_from_obj() takes an object as input and in kmem_cache_free_bulk()
that object is not directly available...  Could send "s" as a reference
(to build_detached_freelist) to allow re-assignment of "s" so
slab_free() gets the correct "s".  But it will not look pretty... 

Else we can get the object via: p[size -1] which also look a little
funny... but it might not be correct in-case NULL pointers in the input
p-array.


> Besides, there's a bug in kmem_cache_alloc_bulk:

Thanks for spotting this!!!

> > /* Note that interrupts must be enabled when calling this function. */
> > bool kmem_cache_alloc_bulk(struct kmem_cache *s, gfp_t flags, size_t size,
> > 			   void **p)
> > {
> > 	struct kmem_cache_cpu *c;
> > 	int i;
> > 
> > 	/*
> > 	 * Drain objects in the per cpu slab, while disabling local
> > 	 * IRQs, which protects against PREEMPT and interrupts
> > 	 * handlers invoking normal fastpath.
> > 	 */
> > 	local_irq_disable();
> > 	c = this_cpu_ptr(s->cpu_slab);
> > 
> > 	for (i = 0; i < size; i++) {
> > 		void *object = c->freelist;
> > 
> > 		if (unlikely(!object)) {
> > 			/*
> > 			 * Invoking slow path likely have side-effect
> > 			 * of re-populating per CPU c->freelist
> > 			 */
> > 			p[i] = ___slab_alloc(s, flags, NUMA_NO_NODE,
> > 					    _RET_IP_, c);
> > 			if (unlikely(!p[i]))
> > 				goto error;
> > 
> > 			c = this_cpu_ptr(s->cpu_slab);
> > 			continue; /* goto for-loop */
> > 		}
> > 
> > 		/* kmem_cache debug support */
> > 		s = slab_pre_alloc_hook(s, flags);
> 
> slab_pre_alloc_hook expects a global cache and returns per memcg one, so
> calling this function from inside a kmemcg will result in hitting the
> VM_BUG_ON in __memcg_kmem_get_cache, not saying about mis-accounting of
> __slab_alloc.
> 
> memcg_kmem_get_cache should be called once, in the very beginning of
> kmem_cache_alloc_bulk, and it should be matched by memcg_kmem_put_cache
> when we are done.

To solve this correctly it looks like I need to pull out
memcg_kmem_put_cache(s) call in the slab_post_alloc_hook() call.

> 
> > 		if (unlikely(!s))
> > 			goto error;
> > 
> > 		c->freelist = get_freepointer(s, object);
> > 		p[i] = object;
> > 
> > 		/* kmem_cache debug support */
> > 		slab_post_alloc_hook(s, flags, object);
> > 	}
> > 	c->tid = next_tid(c->tid);
> > 	local_irq_enable();
> > 
> > 	/* Clear memory outside IRQ disabled fastpath loop */
> > 	if (unlikely(flags & __GFP_ZERO)) {
> > 		int j;
> > 
> > 		for (j = 0; j < i; j++)
> > 			memset(p[j], 0, s->object_size);
> > 	}
> > 
> > 	return true;
> > 
> > error:
> > 	__kmem_cache_free_bulk(s, i, p);
> > 	local_irq_enable();
> > 	return false;
> > }



-- 
Best regards,
  Jesper Dangaard Brouer
  MSc.CS, Principal Kernel Engineer at Red Hat
  Author of http://www.iptv-analyzer.org
  LinkedIn: http://www.linkedin.com/in/brouer

--
To unsubscribe, send a message with 'unsubscribe linux-mm' in
the body to majordomo@kvack.org.  For more info on Linux MM,
see: http://www.linux-mm.org/ .
Don't email: <a href=mailto:"dont@kvack.org"> email@kvack.org </a>

^ permalink raw reply	[flat|nested] 4+ messages in thread

end of thread, other threads:[~2015-11-05 13:19 UTC | newest]

Thread overview: 4+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2015-10-29 13:05 [PATCH] slub: add missing kmem cgroup support to kmem_cache_free_bulk Jesper Dangaard Brouer
2015-11-05  5:06 ` Joonsoo Kim
2015-11-05  8:38   ` Vladimir Davydov
2015-11-05 13:19     ` Jesper Dangaard Brouer

This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox