linux-mm.kvack.org archive mirror
 help / color / mirror / Atom feed
From: Greg Thelen <gthelen@google.com>
To: Glauber Costa <glommer@parallels.com>
Cc: linux-mm@kvack.org, cgroups@vger.kernel.org,
	Dave Shrinnker <david@fromorbit.com>,
	Serge Hallyn <serge.hallyn@canonical.com>,
	kamezawa.hiroyu@jp.fujitsu.com, Michal Hocko <mhocko@suse.cz>,
	Johannes Weiner <hannes@cmpxchg.org>,
	Andrew Morton <akpm@linux-foundation.org>,
	hughd@google.com, linux-fsdevel@vger.kernel.org,
	containers@lists.linux-foundation.org,
	Dave Chinner <dchinner@redhat.com>
Subject: Re: [PATCH v3 14/32] xfs: convert buftarg LRU to generic code
Date: Sun, 14 Apr 2013 22:38:32 -0700	[thread overview]
Message-ID: <xr9361zo8iav.fsf@gthelen.mtv.corp.google.com> (raw)
In-Reply-To: <1365429659-22108-15-git-send-email-glommer@parallels.com>

On Mon, Apr 08 2013, Glauber Costa wrote:

> From: Dave Chinner <dchinner@redhat.com>
>
> Convert the buftarg LRU to use the new generic LRU list and take
> advantage of the functionality it supplies to make the buffer cache
> shrinker node aware.
>
> Signed-off-by: Glauber Costa <glommer@parallels.com>
> Signed-off-by: Dave Chinner <dchinner@redhat.com>
>
> Conflicts with 3b19034d4f:
> 	fs/xfs/xfs_buf.c
> ---
>  fs/xfs/xfs_buf.c | 167 +++++++++++++++++++++++++------------------------------
>  fs/xfs/xfs_buf.h |   5 +-
>  2 files changed, 79 insertions(+), 93 deletions(-)
>
> diff --git a/fs/xfs/xfs_buf.c b/fs/xfs/xfs_buf.c
> index 8459b5d..4cc6632 100644
> --- a/fs/xfs/xfs_buf.c
> +++ b/fs/xfs/xfs_buf.c
> @@ -85,20 +85,14 @@ xfs_buf_vmap_len(
>   * The LRU takes a new reference to the buffer so that it will only be freed
>   * once the shrinker takes the buffer off the LRU.
>   */
> -STATIC void
> +static void
>  xfs_buf_lru_add(
>  	struct xfs_buf	*bp)
>  {
> -	struct xfs_buftarg *btp = bp->b_target;
> -
> -	spin_lock(&btp->bt_lru_lock);
> -	if (list_empty(&bp->b_lru)) {
> -		atomic_inc(&bp->b_hold);
> -		list_add_tail(&bp->b_lru, &btp->bt_lru);
> -		btp->bt_lru_nr++;
> +	if (list_lru_add(&bp->b_target->bt_lru, &bp->b_lru)) {
>  		bp->b_lru_flags &= ~_XBF_LRU_DISPOSE;
> +		atomic_inc(&bp->b_hold);
>  	}
> -	spin_unlock(&btp->bt_lru_lock);
>  }
>  
>  /*
> @@ -107,24 +101,13 @@ xfs_buf_lru_add(
>   * The unlocked check is safe here because it only occurs when there are not
>   * b_lru_ref counts left on the inode under the pag->pag_buf_lock. it is there
>   * to optimise the shrinker removing the buffer from the LRU and calling
> - * xfs_buf_free(). i.e. it removes an unnecessary round trip on the
> - * bt_lru_lock.
> + * xfs_buf_free().
>   */
> -STATIC void
> +static void
>  xfs_buf_lru_del(
>  	struct xfs_buf	*bp)
>  {
> -	struct xfs_buftarg *btp = bp->b_target;
> -
> -	if (list_empty(&bp->b_lru))
> -		return;
> -
> -	spin_lock(&btp->bt_lru_lock);
> -	if (!list_empty(&bp->b_lru)) {
> -		list_del_init(&bp->b_lru);
> -		btp->bt_lru_nr--;
> -	}
> -	spin_unlock(&btp->bt_lru_lock);
> +	list_lru_del(&bp->b_target->bt_lru, &bp->b_lru);
>  }
>  
>  /*
> @@ -151,18 +134,10 @@ xfs_buf_stale(
>  	bp->b_flags &= ~_XBF_DELWRI_Q;
>  
>  	atomic_set(&(bp)->b_lru_ref, 0);
> -	if (!list_empty(&bp->b_lru)) {
> -		struct xfs_buftarg *btp = bp->b_target;
> -
> -		spin_lock(&btp->bt_lru_lock);
> -		if (!list_empty(&bp->b_lru) &&
> -		    !(bp->b_lru_flags & _XBF_LRU_DISPOSE)) {
> -			list_del_init(&bp->b_lru);
> -			btp->bt_lru_nr--;
> -			atomic_dec(&bp->b_hold);
> -		}
> -		spin_unlock(&btp->bt_lru_lock);
> -	}
> +	if (!(bp->b_lru_flags & _XBF_LRU_DISPOSE) &&
> +	    (list_lru_del(&bp->b_target->bt_lru, &bp->b_lru)))
> +		atomic_dec(&bp->b_hold);
> +
>  	ASSERT(atomic_read(&bp->b_hold) >= 1);
>  }
>  
> @@ -1498,83 +1473,95 @@ xfs_buf_iomove(
>   * returned. These buffers will have an elevated hold count, so wait on those
>   * while freeing all the buffers only held by the LRU.
>   */
> -void
> -xfs_wait_buftarg(
> -	struct xfs_buftarg	*btp)
> +static int

static enum lru_status

> +xfs_buftarg_wait_rele(
> +	struct list_head	*item,
> +	spinlock_t		*lru_lock,
> +	void			*arg)
> +
>  {
> -	struct xfs_buf		*bp;
> +	struct xfs_buf		*bp = container_of(item, struct xfs_buf, b_lru);
>  
> -restart:
> -	spin_lock(&btp->bt_lru_lock);
> -	while (!list_empty(&btp->bt_lru)) {
> -		bp = list_first_entry(&btp->bt_lru, struct xfs_buf, b_lru);
> -		if (atomic_read(&bp->b_hold) > 1) {
> -			trace_xfs_buf_wait_buftarg(bp, _RET_IP_);
> -			list_move_tail(&bp->b_lru, &btp->bt_lru);
> -			spin_unlock(&btp->bt_lru_lock);
> -			delay(100);
> -			goto restart;
> -		}
> +	if (atomic_read(&bp->b_hold) > 1) {
> +		/* need to wait */
> +		trace_xfs_buf_wait_buftarg(bp, _RET_IP_);
> +		spin_unlock(lru_lock);
> +		delay(100);
> +	} else {
>  		/*
>  		 * clear the LRU reference count so the buffer doesn't get
>  		 * ignored in xfs_buf_rele().
>  		 */
>  		atomic_set(&bp->b_lru_ref, 0);
> -		spin_unlock(&btp->bt_lru_lock);
> +		spin_unlock(lru_lock);
>  		xfs_buf_rele(bp);
> -		spin_lock(&btp->bt_lru_lock);
>  	}
> -	spin_unlock(&btp->bt_lru_lock);
> +	return 3;

	return LRU_RETRY;

>  }
>  
> -int
> -xfs_buftarg_shrink(
> +void
> +xfs_wait_buftarg(
> +	struct xfs_buftarg	*btp)
> +{
> +	while (list_lru_count(&btp->bt_lru))
> +		list_lru_walk(&btp->bt_lru, xfs_buftarg_wait_rele,
> +			      NULL, LONG_MAX);
> +}
> +
> +static int

static enum lru_status

> +xfs_buftarg_isolate(
> +	struct list_head	*item,
> +	spinlock_t		*lru_lock,
> +	void			*arg)
> +{
> +	struct xfs_buf		*bp = container_of(item, struct xfs_buf, b_lru);
> +	struct list_head	*dispose = arg;
> +
> +	/*
> +	 * Decrement the b_lru_ref count unless the value is already
> +	 * zero. If the value is already zero, we need to reclaim the
> +	 * buffer, otherwise it gets another trip through the LRU.
> +	 */
> +	if (!atomic_add_unless(&bp->b_lru_ref, -1, 0))
> +		return 1;

		return LRU_ROTATE;

> +
> +	bp->b_lru_flags |= _XBF_LRU_DISPOSE;
> +	list_move(item, dispose);
> +	return 0;

	return LRU_REMOVED;

> +}
> +
> +static long
> +xfs_buftarg_shrink_scan(
>  	struct shrinker		*shrink,
>  	struct shrink_control	*sc)
>  {
>  	struct xfs_buftarg	*btp = container_of(shrink,
>  					struct xfs_buftarg, bt_shrinker);
> -	struct xfs_buf		*bp;
> -	int nr_to_scan = sc->nr_to_scan;
>  	LIST_HEAD(dispose);
> +	long			freed;
>  
> -	if (!nr_to_scan)
> -		return btp->bt_lru_nr;
> -
> -	spin_lock(&btp->bt_lru_lock);
> -	while (!list_empty(&btp->bt_lru)) {
> -		if (nr_to_scan-- <= 0)
> -			break;
> -
> -		bp = list_first_entry(&btp->bt_lru, struct xfs_buf, b_lru);
> -
> -		/*
> -		 * Decrement the b_lru_ref count unless the value is already
> -		 * zero. If the value is already zero, we need to reclaim the
> -		 * buffer, otherwise it gets another trip through the LRU.
> -		 */
> -		if (!atomic_add_unless(&bp->b_lru_ref, -1, 0)) {
> -			list_move_tail(&bp->b_lru, &btp->bt_lru);
> -			continue;
> -		}
> -
> -		/*
> -		 * remove the buffer from the LRU now to avoid needing another
> -		 * lock round trip inside xfs_buf_rele().
> -		 */
> -		list_move(&bp->b_lru, &dispose);
> -		btp->bt_lru_nr--;
> -		bp->b_lru_flags |= _XBF_LRU_DISPOSE;
> -	}
> -	spin_unlock(&btp->bt_lru_lock);
> +	freed = list_lru_walk_nodemask(&btp->bt_lru, xfs_buftarg_isolate,
> +				       &dispose, sc->nr_to_scan,
> +				       &sc->nodes_to_scan);
>  
>  	while (!list_empty(&dispose)) {
> +		struct xfs_buf *bp;
>  		bp = list_first_entry(&dispose, struct xfs_buf, b_lru);
>  		list_del_init(&bp->b_lru);
>  		xfs_buf_rele(bp);
>  	}
>  
> -	return btp->bt_lru_nr;
> +	return freed;
> +}
> +
> +static long
> +xfs_buftarg_shrink_count(
> +	struct shrinker		*shrink,
> +	struct shrink_control	*sc)
> +{
> +	struct xfs_buftarg	*btp = container_of(shrink,
> +					struct xfs_buftarg, bt_shrinker);
> +	return list_lru_count_nodemask(&btp->bt_lru, &sc->nodes_to_scan);
>  }
>  
>  void
> @@ -1656,11 +1643,11 @@ xfs_alloc_buftarg(
>  	if (!btp->bt_bdi)
>  		goto error;
>  
> -	INIT_LIST_HEAD(&btp->bt_lru);
> -	spin_lock_init(&btp->bt_lru_lock);
> +	list_lru_init(&btp->bt_lru);
>  	if (xfs_setsize_buftarg_early(btp, bdev))
>  		goto error;
> -	btp->bt_shrinker.shrink = xfs_buftarg_shrink;
> +	btp->bt_shrinker.count_objects = xfs_buftarg_shrink_count;
> +	btp->bt_shrinker.scan_objects = xfs_buftarg_shrink_scan;
>  	btp->bt_shrinker.seeks = DEFAULT_SEEKS;
>  	register_shrinker(&btp->bt_shrinker);
>  	return btp;
> diff --git a/fs/xfs/xfs_buf.h b/fs/xfs/xfs_buf.h
> index 433a12e..5ec7d35 100644
> --- a/fs/xfs/xfs_buf.h
> +++ b/fs/xfs/xfs_buf.h
> @@ -25,6 +25,7 @@
>  #include <linux/fs.h>
>  #include <linux/buffer_head.h>
>  #include <linux/uio.h>
> +#include <linux/list_lru.h>
>  
>  /*
>   *	Base types
> @@ -92,9 +93,7 @@ typedef struct xfs_buftarg {
>  
>  	/* LRU control structures */
>  	struct shrinker		bt_shrinker;
> -	struct list_head	bt_lru;
> -	spinlock_t		bt_lru_lock;
> -	unsigned int		bt_lru_nr;
> +	struct list_lru		bt_lru;
>  } xfs_buftarg_t;
>  
>  struct xfs_buf;

--
To unsubscribe, send a message with 'unsubscribe linux-mm' in
the body to majordomo@kvack.org.  For more info on Linux MM,
see: http://www.linux-mm.org/ .
Don't email: <a href=mailto:"dont@kvack.org"> email@kvack.org </a>

  reply	other threads:[~2013-04-15  5:38 UTC|newest]

Thread overview: 45+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2013-04-08 14:00 [PATCH v3 00/32] memcg-aware slab shrinking with lasers and numbers Glauber Costa
2013-04-08 14:00 ` [PATCH v3 01/32] super: fix calculation of shrinkable objects for small numbers Glauber Costa
2013-04-08 14:00 ` [PATCH v3 02/32] vmscan: take at least one pass with shrinkers Glauber Costa
2013-04-08 14:00 ` [PATCH v3 03/32] dcache: convert dentry_stat.nr_unused to per-cpu counters Glauber Costa
2013-04-08 14:00 ` [PATCH v3 04/32] dentry: move to per-sb LRU locks Glauber Costa
2013-04-08 14:00 ` [PATCH v3 05/32] dcache: remove dentries from LRU before putting on dispose list Glauber Costa
2013-04-08 14:00 ` [PATCH v3 06/32] mm: new shrinker API Glauber Costa
2013-04-08 14:00 ` [PATCH v3 07/32] shrinker: convert superblock shrinkers to new API Glauber Costa
2013-04-08 14:00 ` [PATCH v3 08/32] list: add a new LRU list type Glauber Costa
2013-04-15  5:35   ` Greg Thelen
2013-04-15 17:56     ` Greg Thelen
2013-04-16 14:43       ` Glauber Costa
2013-04-08 14:00 ` [PATCH v3 09/32] inode: convert inode lru list to generic lru list code Glauber Costa
2013-04-08 14:00 ` [PATCH v3 10/32] dcache: convert to use new lru list infrastructure Glauber Costa
2013-04-08 14:00 ` [PATCH v3 11/32] list_lru: per-node " Glauber Costa
2013-04-15  5:37   ` Greg Thelen
2013-04-08 14:00 ` [PATCH v3 12/32] shrinker: add node awareness Glauber Costa
2013-04-15  5:38   ` Greg Thelen
2013-04-08 14:00 ` [PATCH v3 13/32] fs: convert inode and dentry shrinking to be node aware Glauber Costa
2013-04-08 14:00 ` [PATCH v3 14/32] xfs: convert buftarg LRU to generic code Glauber Costa
2013-04-15  5:38   ` Greg Thelen [this message]
2013-04-15 10:14     ` Glauber Costa
2013-04-08 14:00 ` [PATCH v3 15/32] xfs: convert dquot cache lru to list_lru Glauber Costa
2013-04-08 14:00 ` [PATCH v3 16/32] fs: convert fs shrinkers to new scan/count API Glauber Costa
2013-04-08 14:00 ` [PATCH v3 17/32] drivers: convert shrinkers to new count/scan API Glauber Costa
2013-04-08 14:00 ` [PATCH v3 18/32] shrinker: convert remaining shrinkers to " Glauber Costa
2013-04-08 14:00 ` [PATCH v3 19/32] hugepage: convert huge zero page shrinker to new shrinker API Glauber Costa
2013-04-15  5:38   ` Greg Thelen
2013-04-15  8:10   ` Kirill A. Shutemov
2013-04-08 14:00 ` [PATCH v3 20/32] shrinker: Kill old ->shrink API Glauber Costa
2013-04-15  5:38   ` Greg Thelen
2013-04-08 14:00 ` [PATCH v3 21/32] vmscan: also shrink slab in memcg pressure Glauber Costa
2013-04-08 14:00 ` [PATCH v3 22/32] memcg,list_lru: duplicate LRUs upon kmemcg creation Glauber Costa
2013-04-08 14:00 ` [PATCH v3 23/32] lru: add an element to a memcg list Glauber Costa
2013-04-08 14:00 ` [PATCH v3 24/32] list_lru: also include memcg lists in counts and scans Glauber Costa
2013-04-08 14:00 ` [PATCH v3 25/32] list_lru: per-memcg walks Glauber Costa
2013-04-08 14:00 ` [PATCH v3 26/32] memcg: per-memcg kmem shrinking Glauber Costa
2013-04-08 14:00 ` [PATCH v3 27/32] list_lru: reclaim proportionaly between memcgs and nodes Glauber Costa
2013-04-08 14:00 ` [PATCH v3 28/32] memcg: scan cache objects hierarchically Glauber Costa
2013-04-08 14:00 ` [PATCH v3 29/32] memcg: move initialization to memcg creation Glauber Costa
2013-04-08 14:00 ` [PATCH v3 30/32] memcg: shrink dead memcgs upon global memory pressure Glauber Costa
2013-04-08 14:00 ` [PATCH v3 31/32] super: targeted memcg reclaim Glauber Costa
2013-04-08 14:00 ` [PATCH v3 32/32] memcg: debugging facility to access dangling memcgs Glauber Costa
2013-04-08 20:51 ` [PATCH v3 00/32] memcg-aware slab shrinking with lasers and numbers Andrew Morton
2013-04-09  7:25   ` Glauber Costa

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=xr9361zo8iav.fsf@gthelen.mtv.corp.google.com \
    --to=gthelen@google.com \
    --cc=akpm@linux-foundation.org \
    --cc=cgroups@vger.kernel.org \
    --cc=containers@lists.linux-foundation.org \
    --cc=david@fromorbit.com \
    --cc=dchinner@redhat.com \
    --cc=glommer@parallels.com \
    --cc=hannes@cmpxchg.org \
    --cc=hughd@google.com \
    --cc=kamezawa.hiroyu@jp.fujitsu.com \
    --cc=linux-fsdevel@vger.kernel.org \
    --cc=linux-mm@kvack.org \
    --cc=mhocko@suse.cz \
    --cc=serge.hallyn@canonical.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox