linux-mm.kvack.org archive mirror
 help / color / mirror / Atom feed
From: "Mark-PK Tsai (蔡沛剛)" <Mark-PK.Tsai@mediatek.com>
To: "senozhatsky@chromium.org" <senozhatsky@chromium.org>
Cc: "linux-kernel@vger.kernel.org" <linux-kernel@vger.kernel.org>,
	"linux-mediatek@lists.infradead.org"
	<linux-mediatek@lists.infradead.org>,
	"linux-mm@kvack.org" <linux-mm@kvack.org>,
	"akpm@linux-foundation.org" <akpm@linux-foundation.org>,
	"minchan@kernel.org" <minchan@kernel.org>,
	"linux-arm-kernel@lists.infradead.org"
	<linux-arm-kernel@lists.infradead.org>,
	"YJ Chiang (江英杰)" <yj.chiang@mediatek.com>,
	"matthias.bgg@gmail.com" <matthias.bgg@gmail.com>,
	"angelogioacchino.delregno@collabora.com"
	<angelogioacchino.delregno@collabora.com>
Subject: Re: [PATCH] zsmalloc: use copy_page for full page copy
Date: Sat, 7 Oct 2023 07:45:45 +0000	[thread overview]
Message-ID: <dd0fad41f5289e51efb342c1355551de782379ea.camel@mediatek.com> (raw)
In-Reply-To: <20231006111144.GB17924@google.com>

On Fri, 2023-10-06 at 20:11 +0900, Sergey Senozhatsky wrote:
>  	 
> External email : Please do not click links or open attachments until
> you have verified the sender or the content.
>  
> On (23/10/06 14:02), Mark-PK Tsai wrote:
> > Some architectures have implemented optimized
> > copy_page for full page copying, such as arm.
> > 
> > On my arm platform, use the copy_page helper
> > for single page copying is about 10 percent faster
> > than memcpy.
> > 
> > Signed-off-by: Mark-PK Tsai <mark-pk.tsai@mediatek.com>
> 
> TIL
> 
> I've never heard of arm's copy_page() before. Is it really much
> faster than memcpy()?

Based on my earlier tests, it seems better than memcpy
for full page copying.
The test code is just measures the copying time with
irq disabled.

> 
> Reviewed-by: Sergey Senozhatsky <senozhatsky@chromium.org>
> 
> 
> > diff --git a/mm/zsmalloc.c b/mm/zsmalloc.c
> > index c743ce7a5f49..b1c0dad7f4cf 100644
> > --- a/mm/zsmalloc.c
> > +++ b/mm/zsmalloc.c
> > @@ -1839,7 +1839,7 @@ static int zs_page_migrate(struct page
> *newpage, struct page *page,
> >   * Here, any user cannot access all objects in the zspage so let's
> move.
> >   */
> >  d_addr = kmap_atomic(newpage);
> > -memcpy(d_addr, s_addr, PAGE_SIZE);
> > +copy_page(d_addr, s_addr);
> 
> I guess you can also look into patching zram_drv.c, which seem to
> have
> at least one PAGE_SIZE memcpy().

Thank. I've just post another patch for zram_drv.c as below link.


https://lore.kernel.org/lkml/20231007070554.8657-1-mark-pk.tsai@mediatek.com/

      reply	other threads:[~2023-10-07  7:56 UTC|newest]

Thread overview: 3+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2023-10-06  6:02 Mark-PK Tsai
2023-10-06 11:11 ` Sergey Senozhatsky
2023-10-07  7:45   ` Mark-PK Tsai (蔡沛剛) [this message]

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=dd0fad41f5289e51efb342c1355551de782379ea.camel@mediatek.com \
    --to=mark-pk.tsai@mediatek.com \
    --cc=akpm@linux-foundation.org \
    --cc=angelogioacchino.delregno@collabora.com \
    --cc=linux-arm-kernel@lists.infradead.org \
    --cc=linux-kernel@vger.kernel.org \
    --cc=linux-mediatek@lists.infradead.org \
    --cc=linux-mm@kvack.org \
    --cc=matthias.bgg@gmail.com \
    --cc=minchan@kernel.org \
    --cc=senozhatsky@chromium.org \
    --cc=yj.chiang@mediatek.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox