From: Greg Thelen <gthelen@google.com>
To: Glauber Costa <glommer@parallels.com>
Cc: linux-mm@kvack.org, cgroups@vger.kernel.org,
Dave Shrinnker <david@fromorbit.com>,
Serge Hallyn <serge.hallyn@canonical.com>,
kamezawa.hiroyu@jp.fujitsu.com, Michal Hocko <mhocko@suse.cz>,
Johannes Weiner <hannes@cmpxchg.org>,
Andrew Morton <akpm@linux-foundation.org>,
hughd@google.com, linux-fsdevel@vger.kernel.org,
containers@lists.linux-foundation.org,
Dave Chinner <dchinner@redhat.com>
Subject: Re: [PATCH v3 14/32] xfs: convert buftarg LRU to generic code
Date: Sun, 14 Apr 2013 22:38:32 -0700 [thread overview]
Message-ID: <xr9361zo8iav.fsf@gthelen.mtv.corp.google.com> (raw)
In-Reply-To: <1365429659-22108-15-git-send-email-glommer@parallels.com>
On Mon, Apr 08 2013, Glauber Costa wrote:
> From: Dave Chinner <dchinner@redhat.com>
>
> Convert the buftarg LRU to use the new generic LRU list and take
> advantage of the functionality it supplies to make the buffer cache
> shrinker node aware.
>
> Signed-off-by: Glauber Costa <glommer@parallels.com>
> Signed-off-by: Dave Chinner <dchinner@redhat.com>
>
> Conflicts with 3b19034d4f:
> fs/xfs/xfs_buf.c
> ---
> fs/xfs/xfs_buf.c | 167 +++++++++++++++++++++++++------------------------------
> fs/xfs/xfs_buf.h | 5 +-
> 2 files changed, 79 insertions(+), 93 deletions(-)
>
> diff --git a/fs/xfs/xfs_buf.c b/fs/xfs/xfs_buf.c
> index 8459b5d..4cc6632 100644
> --- a/fs/xfs/xfs_buf.c
> +++ b/fs/xfs/xfs_buf.c
> @@ -85,20 +85,14 @@ xfs_buf_vmap_len(
> * The LRU takes a new reference to the buffer so that it will only be freed
> * once the shrinker takes the buffer off the LRU.
> */
> -STATIC void
> +static void
> xfs_buf_lru_add(
> struct xfs_buf *bp)
> {
> - struct xfs_buftarg *btp = bp->b_target;
> -
> - spin_lock(&btp->bt_lru_lock);
> - if (list_empty(&bp->b_lru)) {
> - atomic_inc(&bp->b_hold);
> - list_add_tail(&bp->b_lru, &btp->bt_lru);
> - btp->bt_lru_nr++;
> + if (list_lru_add(&bp->b_target->bt_lru, &bp->b_lru)) {
> bp->b_lru_flags &= ~_XBF_LRU_DISPOSE;
> + atomic_inc(&bp->b_hold);
> }
> - spin_unlock(&btp->bt_lru_lock);
> }
>
> /*
> @@ -107,24 +101,13 @@ xfs_buf_lru_add(
> * The unlocked check is safe here because it only occurs when there are not
> * b_lru_ref counts left on the inode under the pag->pag_buf_lock. it is there
> * to optimise the shrinker removing the buffer from the LRU and calling
> - * xfs_buf_free(). i.e. it removes an unnecessary round trip on the
> - * bt_lru_lock.
> + * xfs_buf_free().
> */
> -STATIC void
> +static void
> xfs_buf_lru_del(
> struct xfs_buf *bp)
> {
> - struct xfs_buftarg *btp = bp->b_target;
> -
> - if (list_empty(&bp->b_lru))
> - return;
> -
> - spin_lock(&btp->bt_lru_lock);
> - if (!list_empty(&bp->b_lru)) {
> - list_del_init(&bp->b_lru);
> - btp->bt_lru_nr--;
> - }
> - spin_unlock(&btp->bt_lru_lock);
> + list_lru_del(&bp->b_target->bt_lru, &bp->b_lru);
> }
>
> /*
> @@ -151,18 +134,10 @@ xfs_buf_stale(
> bp->b_flags &= ~_XBF_DELWRI_Q;
>
> atomic_set(&(bp)->b_lru_ref, 0);
> - if (!list_empty(&bp->b_lru)) {
> - struct xfs_buftarg *btp = bp->b_target;
> -
> - spin_lock(&btp->bt_lru_lock);
> - if (!list_empty(&bp->b_lru) &&
> - !(bp->b_lru_flags & _XBF_LRU_DISPOSE)) {
> - list_del_init(&bp->b_lru);
> - btp->bt_lru_nr--;
> - atomic_dec(&bp->b_hold);
> - }
> - spin_unlock(&btp->bt_lru_lock);
> - }
> + if (!(bp->b_lru_flags & _XBF_LRU_DISPOSE) &&
> + (list_lru_del(&bp->b_target->bt_lru, &bp->b_lru)))
> + atomic_dec(&bp->b_hold);
> +
> ASSERT(atomic_read(&bp->b_hold) >= 1);
> }
>
> @@ -1498,83 +1473,95 @@ xfs_buf_iomove(
> * returned. These buffers will have an elevated hold count, so wait on those
> * while freeing all the buffers only held by the LRU.
> */
> -void
> -xfs_wait_buftarg(
> - struct xfs_buftarg *btp)
> +static int
static enum lru_status
> +xfs_buftarg_wait_rele(
> + struct list_head *item,
> + spinlock_t *lru_lock,
> + void *arg)
> +
> {
> - struct xfs_buf *bp;
> + struct xfs_buf *bp = container_of(item, struct xfs_buf, b_lru);
>
> -restart:
> - spin_lock(&btp->bt_lru_lock);
> - while (!list_empty(&btp->bt_lru)) {
> - bp = list_first_entry(&btp->bt_lru, struct xfs_buf, b_lru);
> - if (atomic_read(&bp->b_hold) > 1) {
> - trace_xfs_buf_wait_buftarg(bp, _RET_IP_);
> - list_move_tail(&bp->b_lru, &btp->bt_lru);
> - spin_unlock(&btp->bt_lru_lock);
> - delay(100);
> - goto restart;
> - }
> + if (atomic_read(&bp->b_hold) > 1) {
> + /* need to wait */
> + trace_xfs_buf_wait_buftarg(bp, _RET_IP_);
> + spin_unlock(lru_lock);
> + delay(100);
> + } else {
> /*
> * clear the LRU reference count so the buffer doesn't get
> * ignored in xfs_buf_rele().
> */
> atomic_set(&bp->b_lru_ref, 0);
> - spin_unlock(&btp->bt_lru_lock);
> + spin_unlock(lru_lock);
> xfs_buf_rele(bp);
> - spin_lock(&btp->bt_lru_lock);
> }
> - spin_unlock(&btp->bt_lru_lock);
> + return 3;
return LRU_RETRY;
> }
>
> -int
> -xfs_buftarg_shrink(
> +void
> +xfs_wait_buftarg(
> + struct xfs_buftarg *btp)
> +{
> + while (list_lru_count(&btp->bt_lru))
> + list_lru_walk(&btp->bt_lru, xfs_buftarg_wait_rele,
> + NULL, LONG_MAX);
> +}
> +
> +static int
static enum lru_status
> +xfs_buftarg_isolate(
> + struct list_head *item,
> + spinlock_t *lru_lock,
> + void *arg)
> +{
> + struct xfs_buf *bp = container_of(item, struct xfs_buf, b_lru);
> + struct list_head *dispose = arg;
> +
> + /*
> + * Decrement the b_lru_ref count unless the value is already
> + * zero. If the value is already zero, we need to reclaim the
> + * buffer, otherwise it gets another trip through the LRU.
> + */
> + if (!atomic_add_unless(&bp->b_lru_ref, -1, 0))
> + return 1;
return LRU_ROTATE;
> +
> + bp->b_lru_flags |= _XBF_LRU_DISPOSE;
> + list_move(item, dispose);
> + return 0;
return LRU_REMOVED;
> +}
> +
> +static long
> +xfs_buftarg_shrink_scan(
> struct shrinker *shrink,
> struct shrink_control *sc)
> {
> struct xfs_buftarg *btp = container_of(shrink,
> struct xfs_buftarg, bt_shrinker);
> - struct xfs_buf *bp;
> - int nr_to_scan = sc->nr_to_scan;
> LIST_HEAD(dispose);
> + long freed;
>
> - if (!nr_to_scan)
> - return btp->bt_lru_nr;
> -
> - spin_lock(&btp->bt_lru_lock);
> - while (!list_empty(&btp->bt_lru)) {
> - if (nr_to_scan-- <= 0)
> - break;
> -
> - bp = list_first_entry(&btp->bt_lru, struct xfs_buf, b_lru);
> -
> - /*
> - * Decrement the b_lru_ref count unless the value is already
> - * zero. If the value is already zero, we need to reclaim the
> - * buffer, otherwise it gets another trip through the LRU.
> - */
> - if (!atomic_add_unless(&bp->b_lru_ref, -1, 0)) {
> - list_move_tail(&bp->b_lru, &btp->bt_lru);
> - continue;
> - }
> -
> - /*
> - * remove the buffer from the LRU now to avoid needing another
> - * lock round trip inside xfs_buf_rele().
> - */
> - list_move(&bp->b_lru, &dispose);
> - btp->bt_lru_nr--;
> - bp->b_lru_flags |= _XBF_LRU_DISPOSE;
> - }
> - spin_unlock(&btp->bt_lru_lock);
> + freed = list_lru_walk_nodemask(&btp->bt_lru, xfs_buftarg_isolate,
> + &dispose, sc->nr_to_scan,
> + &sc->nodes_to_scan);
>
> while (!list_empty(&dispose)) {
> + struct xfs_buf *bp;
> bp = list_first_entry(&dispose, struct xfs_buf, b_lru);
> list_del_init(&bp->b_lru);
> xfs_buf_rele(bp);
> }
>
> - return btp->bt_lru_nr;
> + return freed;
> +}
> +
> +static long
> +xfs_buftarg_shrink_count(
> + struct shrinker *shrink,
> + struct shrink_control *sc)
> +{
> + struct xfs_buftarg *btp = container_of(shrink,
> + struct xfs_buftarg, bt_shrinker);
> + return list_lru_count_nodemask(&btp->bt_lru, &sc->nodes_to_scan);
> }
>
> void
> @@ -1656,11 +1643,11 @@ xfs_alloc_buftarg(
> if (!btp->bt_bdi)
> goto error;
>
> - INIT_LIST_HEAD(&btp->bt_lru);
> - spin_lock_init(&btp->bt_lru_lock);
> + list_lru_init(&btp->bt_lru);
> if (xfs_setsize_buftarg_early(btp, bdev))
> goto error;
> - btp->bt_shrinker.shrink = xfs_buftarg_shrink;
> + btp->bt_shrinker.count_objects = xfs_buftarg_shrink_count;
> + btp->bt_shrinker.scan_objects = xfs_buftarg_shrink_scan;
> btp->bt_shrinker.seeks = DEFAULT_SEEKS;
> register_shrinker(&btp->bt_shrinker);
> return btp;
> diff --git a/fs/xfs/xfs_buf.h b/fs/xfs/xfs_buf.h
> index 433a12e..5ec7d35 100644
> --- a/fs/xfs/xfs_buf.h
> +++ b/fs/xfs/xfs_buf.h
> @@ -25,6 +25,7 @@
> #include <linux/fs.h>
> #include <linux/buffer_head.h>
> #include <linux/uio.h>
> +#include <linux/list_lru.h>
>
> /*
> * Base types
> @@ -92,9 +93,7 @@ typedef struct xfs_buftarg {
>
> /* LRU control structures */
> struct shrinker bt_shrinker;
> - struct list_head bt_lru;
> - spinlock_t bt_lru_lock;
> - unsigned int bt_lru_nr;
> + struct list_lru bt_lru;
> } xfs_buftarg_t;
>
> struct xfs_buf;
--
To unsubscribe, send a message with 'unsubscribe linux-mm' in
the body to majordomo@kvack.org. For more info on Linux MM,
see: http://www.linux-mm.org/ .
Don't email: <a href=mailto:"dont@kvack.org"> email@kvack.org </a>
next prev parent reply other threads:[~2013-04-15 5:38 UTC|newest]
Thread overview: 45+ messages / expand[flat|nested] mbox.gz Atom feed top
2013-04-08 14:00 [PATCH v3 00/32] memcg-aware slab shrinking with lasers and numbers Glauber Costa
2013-04-08 14:00 ` [PATCH v3 01/32] super: fix calculation of shrinkable objects for small numbers Glauber Costa
2013-04-08 14:00 ` [PATCH v3 02/32] vmscan: take at least one pass with shrinkers Glauber Costa
2013-04-08 14:00 ` [PATCH v3 03/32] dcache: convert dentry_stat.nr_unused to per-cpu counters Glauber Costa
2013-04-08 14:00 ` [PATCH v3 04/32] dentry: move to per-sb LRU locks Glauber Costa
2013-04-08 14:00 ` [PATCH v3 05/32] dcache: remove dentries from LRU before putting on dispose list Glauber Costa
2013-04-08 14:00 ` [PATCH v3 06/32] mm: new shrinker API Glauber Costa
2013-04-08 14:00 ` [PATCH v3 07/32] shrinker: convert superblock shrinkers to new API Glauber Costa
2013-04-08 14:00 ` [PATCH v3 08/32] list: add a new LRU list type Glauber Costa
2013-04-15 5:35 ` Greg Thelen
2013-04-15 17:56 ` Greg Thelen
2013-04-16 14:43 ` Glauber Costa
2013-04-08 14:00 ` [PATCH v3 09/32] inode: convert inode lru list to generic lru list code Glauber Costa
2013-04-08 14:00 ` [PATCH v3 10/32] dcache: convert to use new lru list infrastructure Glauber Costa
2013-04-08 14:00 ` [PATCH v3 11/32] list_lru: per-node " Glauber Costa
2013-04-15 5:37 ` Greg Thelen
2013-04-08 14:00 ` [PATCH v3 12/32] shrinker: add node awareness Glauber Costa
2013-04-15 5:38 ` Greg Thelen
2013-04-08 14:00 ` [PATCH v3 13/32] fs: convert inode and dentry shrinking to be node aware Glauber Costa
2013-04-08 14:00 ` [PATCH v3 14/32] xfs: convert buftarg LRU to generic code Glauber Costa
2013-04-15 5:38 ` Greg Thelen [this message]
2013-04-15 10:14 ` Glauber Costa
2013-04-08 14:00 ` [PATCH v3 15/32] xfs: convert dquot cache lru to list_lru Glauber Costa
2013-04-08 14:00 ` [PATCH v3 16/32] fs: convert fs shrinkers to new scan/count API Glauber Costa
2013-04-08 14:00 ` [PATCH v3 17/32] drivers: convert shrinkers to new count/scan API Glauber Costa
2013-04-08 14:00 ` [PATCH v3 18/32] shrinker: convert remaining shrinkers to " Glauber Costa
2013-04-08 14:00 ` [PATCH v3 19/32] hugepage: convert huge zero page shrinker to new shrinker API Glauber Costa
2013-04-15 5:38 ` Greg Thelen
2013-04-15 8:10 ` Kirill A. Shutemov
2013-04-08 14:00 ` [PATCH v3 20/32] shrinker: Kill old ->shrink API Glauber Costa
2013-04-15 5:38 ` Greg Thelen
2013-04-08 14:00 ` [PATCH v3 21/32] vmscan: also shrink slab in memcg pressure Glauber Costa
2013-04-08 14:00 ` [PATCH v3 22/32] memcg,list_lru: duplicate LRUs upon kmemcg creation Glauber Costa
2013-04-08 14:00 ` [PATCH v3 23/32] lru: add an element to a memcg list Glauber Costa
2013-04-08 14:00 ` [PATCH v3 24/32] list_lru: also include memcg lists in counts and scans Glauber Costa
2013-04-08 14:00 ` [PATCH v3 25/32] list_lru: per-memcg walks Glauber Costa
2013-04-08 14:00 ` [PATCH v3 26/32] memcg: per-memcg kmem shrinking Glauber Costa
2013-04-08 14:00 ` [PATCH v3 27/32] list_lru: reclaim proportionaly between memcgs and nodes Glauber Costa
2013-04-08 14:00 ` [PATCH v3 28/32] memcg: scan cache objects hierarchically Glauber Costa
2013-04-08 14:00 ` [PATCH v3 29/32] memcg: move initialization to memcg creation Glauber Costa
2013-04-08 14:00 ` [PATCH v3 30/32] memcg: shrink dead memcgs upon global memory pressure Glauber Costa
2013-04-08 14:00 ` [PATCH v3 31/32] super: targeted memcg reclaim Glauber Costa
2013-04-08 14:00 ` [PATCH v3 32/32] memcg: debugging facility to access dangling memcgs Glauber Costa
2013-04-08 20:51 ` [PATCH v3 00/32] memcg-aware slab shrinking with lasers and numbers Andrew Morton
2013-04-09 7:25 ` Glauber Costa
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=xr9361zo8iav.fsf@gthelen.mtv.corp.google.com \
--to=gthelen@google.com \
--cc=akpm@linux-foundation.org \
--cc=cgroups@vger.kernel.org \
--cc=containers@lists.linux-foundation.org \
--cc=david@fromorbit.com \
--cc=dchinner@redhat.com \
--cc=glommer@parallels.com \
--cc=hannes@cmpxchg.org \
--cc=hughd@google.com \
--cc=kamezawa.hiroyu@jp.fujitsu.com \
--cc=linux-fsdevel@vger.kernel.org \
--cc=linux-mm@kvack.org \
--cc=mhocko@suse.cz \
--cc=serge.hallyn@canonical.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox