From: Lachlan McIlroy <lachlan@sgi.com>
To: Nick Piggin <npiggin@suse.de>
Cc: Linux Memory Management List <linux-mm@kvack.org>,
xfs@oss.sgi.com, xen-devel@lists.xensource.com,
Linux Kernel Mailing List <linux-kernel@vger.kernel.org>,
Andrew Morton <akpm@linux-foundation.org>,
dri-devel@lists.sourceforge.net
Subject: Re: [rfc][patch 2/3] xfs: remove vmap cache
Date: Mon, 04 Aug 2008 16:23:44 +1000 [thread overview]
Message-ID: <4896A070.2010805@sgi.com> (raw)
In-Reply-To: <20080728123621.GB13926@wotan.suse.de>
Nick Piggin wrote:
> XFS's vmap batching simply defers a number (up to 64) of vunmaps, and keeps
> track of them in a list. To purge the batch, it just goes through the list and
> calls vunamp on each one. This is pretty poor: a global TLB flush is still
> performed on each vunmap, with the most expensive parts of the operation
> being the broadcast IPIs and locking involved in the SMP callouts, and the
> locking involved in the vmap management -- none of these are avoided by just
> batching up the calls. I'm actually surprised it ever made much difference
> at all.
So am I.
>
> Rip all this logic out of XFS completely. I improve vmap performance
> and scalability directly in the previous and subsequent patch.
Sounds good to me.
>
> Signed-off-by: Nick Piggin <npiggin@suse.de>
> ---
>
> Index: linux-2.6/fs/xfs/linux-2.6/xfs_buf.c
> ===================================================================
> --- linux-2.6.orig/fs/xfs/linux-2.6/xfs_buf.c
> +++ linux-2.6/fs/xfs/linux-2.6/xfs_buf.c
> @@ -166,75 +166,6 @@ test_page_region(
> }
>
> /*
> - * Mapping of multi-page buffers into contiguous virtual space
> - */
> -
> -typedef struct a_list {
> - void *vm_addr;
> - struct a_list *next;
> -} a_list_t;
> -
> -static a_list_t *as_free_head;
> -static int as_list_len;
> -static DEFINE_SPINLOCK(as_lock);
> -
> -/*
> - * Try to batch vunmaps because they are costly.
> - */
> -STATIC void
> -free_address(
> - void *addr)
> -{
> - a_list_t *aentry;
> -
> -#ifdef CONFIG_XEN
> - /*
> - * Xen needs to be able to make sure it can get an exclusive
> - * RO mapping of pages it wants to turn into a pagetable. If
> - * a newly allocated page is also still being vmap()ed by xfs,
> - * it will cause pagetable construction to fail. This is a
> - * quick workaround to always eagerly unmap pages so that Xen
> - * is happy.
> - */
> - vunmap(addr);
> - return;
> -#endif
> -
> - aentry = kmalloc(sizeof(a_list_t), GFP_NOWAIT);
> - if (likely(aentry)) {
> - spin_lock(&as_lock);
> - aentry->next = as_free_head;
> - aentry->vm_addr = addr;
> - as_free_head = aentry;
> - as_list_len++;
> - spin_unlock(&as_lock);
> - } else {
> - vunmap(addr);
> - }
> -}
> -
> -STATIC void
> -purge_addresses(void)
> -{
> - a_list_t *aentry, *old;
> -
> - if (as_free_head == NULL)
> - return;
> -
> - spin_lock(&as_lock);
> - aentry = as_free_head;
> - as_free_head = NULL;
> - as_list_len = 0;
> - spin_unlock(&as_lock);
> -
> - while ((old = aentry) != NULL) {
> - vunmap(aentry->vm_addr);
> - aentry = aentry->next;
> - kfree(old);
> - }
> -}
> -
> -/*
> * Internal xfs_buf_t object manipulation
> */
>
> @@ -334,7 +265,7 @@ xfs_buf_free(
> uint i;
>
> if ((bp->b_flags & XBF_MAPPED) && (bp->b_page_count > 1))
> - free_address(bp->b_addr - bp->b_offset);
> + vunmap(bp->b_addr - bp->b_offset);
>
> for (i = 0; i < bp->b_page_count; i++) {
> struct page *page = bp->b_pages[i];
> @@ -456,8 +387,6 @@ _xfs_buf_map_pages(
> bp->b_addr = page_address(bp->b_pages[0]) + bp->b_offset;
> bp->b_flags |= XBF_MAPPED;
> } else if (flags & XBF_MAPPED) {
> - if (as_list_len > 64)
> - purge_addresses();
> bp->b_addr = vmap(bp->b_pages, bp->b_page_count,
> VM_MAP, PAGE_KERNEL);
> if (unlikely(bp->b_addr == NULL))
> @@ -1739,8 +1668,6 @@ xfsbufd(
> count++;
> }
>
> - if (as_list_len > 0)
> - purge_addresses();
> if (count)
> blk_run_address_space(target->bt_mapping);
>
>
>
>
--
To unsubscribe, send a message with 'unsubscribe linux-mm' in
the body to majordomo@kvack.org. For more info on Linux MM,
see: http://www.linux-mm.org/ .
Don't email: <a href=mailto:"dont@kvack.org"> email@kvack.org </a>
next prev parent reply other threads:[~2008-08-04 6:23 UTC|newest]
Thread overview: 9+ messages / expand[flat|nested] mbox.gz Atom feed top
2008-07-28 12:34 [rfc][patch 1/3] mm: vmap rewrite Nick Piggin
2008-07-28 12:36 ` [rfc][patch 2/3] xfs: remove vmap cache Nick Piggin
2008-08-04 6:23 ` Lachlan McIlroy [this message]
2008-07-28 12:37 ` [rfc][patch 3/3] xfs: use new vmap API Nick Piggin
2008-08-04 6:28 ` Lachlan McIlroy
2008-08-04 10:57 ` Nick Piggin
2008-08-05 1:43 ` Lachlan McIlroy
2008-08-05 2:05 ` Nick Piggin
2008-07-28 23:34 ` [rfc][patch 1/3] mm: vmap rewrite Robin Holt
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=4896A070.2010805@sgi.com \
--to=lachlan@sgi.com \
--cc=akpm@linux-foundation.org \
--cc=dri-devel@lists.sourceforge.net \
--cc=linux-kernel@vger.kernel.org \
--cc=linux-mm@kvack.org \
--cc=npiggin@suse.de \
--cc=xen-devel@lists.xensource.com \
--cc=xfs@oss.sgi.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox